diff --git "a/3917.jsonl" "b/3917.jsonl" new file mode 100644--- /dev/null +++ "b/3917.jsonl" @@ -0,0 +1,575 @@ +{"seq_id":"40338481","text":"import numpy\n\nfrom chainer import cuda\nfrom chainer import function\nfrom chainer.utils import type_check\n\n\ndef _transpose(xs, length):\n xp = cuda.get_array_module(*xs)\n lengths = numpy.zeros(length, dtype='i')\n for i, x in enumerate(xs):\n lengths[0:len(x)] = i + 1\n dtype = xs[0].dtype\n unit = xs[0].shape[1:]\n outs = tuple([xp.empty((l,) + unit, dtype=dtype) for l in lengths])\n\n for i, x in enumerate(xs):\n for p, xi in enumerate(x):\n outs[p][i] = xi\n\n return outs\n\n\nclass TransposeSequence(function.Function):\n\n \"\"\"Function that transposes a list of Variables.\"\"\"\n\n def check_type_forward(self, xs_type):\n for p, n in zip(xs_type, xs_type[1:]):\n type_check.expect(\n p.shape[0] >= n.shape[0],\n p.shape[1:] == n.shape[1:],\n )\n\n def forward(self, xs):\n if len(xs) == 0:\n return ()\n return _transpose(xs, len(xs[0]))\n\n def backward(self, xs, gs):\n return _transpose(gs, len(xs))\n\n\ndef transpose_sequence(xs):\n \"\"\"Transpose a list of Variables.\n\n This function transposes a list of :class:`~chainer.Variable` s and returns\n a list of :class:`Variable` s.\n For exampe a user gives ``[(0, 1, 2, 3), (4, 5), (6)]``, the function\n returns ``[(0, 4, 6), (1, 5), (2), (3)]``.\n Note that a given list needs to be sorted by each length of\n :class:`~chainer.Variable`.\n\n Args:\n xs (list of ~chainer.Variable): Variables to transpose.\n\n Returns:\n tuple or Variable: Transposed list.\n \"\"\"\n ys = TransposeSequence()(*xs)\n if not isinstance(ys, tuple):\n ys = (ys,)\n return ys\n","sub_path":"chainer/functions/array/transpose_sequence.py","file_name":"transpose_sequence.py","file_ext":"py","file_size_in_byte":1673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"58733142","text":"# coding: utf-8\n\nimport os\nimport re\nimport traceback\nfrom StringIO import StringIO\nimport json\nimport hashlib\nfrom ConfigParser import ConfigParser\n\nfrom tornado.gen import coroutine\nfrom tornado.httpclient import AsyncHTTPClient\nfrom tornado.options import options\n\nfrom settings import redis, root\nfrom protocol import game_pb2\nfrom utils.logger import Logger, host\n\nname_web_retry = \"web:retry\"\n# pattern = re.compile(\n# r'^http://'\n# r'(?:(?:[A-Z0-9](?:[A-Z0-9-]{0,61}[A-Z0-9])?\\.)+(?:[A-Z]{2,6}\\.?|[A-Z0-9-]{2,}\\.?)|'\n# r'\\d{1,3}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3})'\n# r'(?::\\d+)?', re.IGNORECASE)\npattern = re.compile(r'.*(?=/[^/]*/[^/]*)', re.IGNORECASE)\n\n\nclass WebRequest(object):\n def __init__(self, room_id, room_uuid, game_type, app_id, owner):\n self.room_id = room_id\n self.room_uuid = room_uuid\n self.game_type = game_type\n self.app_id = app_id\n self.owner = owner\n self.logger = Logger(room_id)\n self.web_alloc_url = None\n self.web_record_url = None\n self.reload_web_url()\n\n def reload_web_url(self):\n conf_web = ConfigParser()\n conf_web.read(os.path.join(os.path.join(root, \"confs\"), \"web.conf\"))\n self.web_alloc_url = conf_web.get(\"url\", \"alloc\")\n self.web_record_url = conf_web.get(\"url\", \"record\")\n\n @staticmethod\n def is_web_alloc(url):\n for i in (\"/room/enter\", \"/room/exit\", \"/room/dismiss\", \"/room/load_plus\", \"/room/load_minus\"):\n if i in url:\n return True\n return False\n\n @staticmethod\n def is_web_record(url):\n for i in (\"/room/refund\", \"/settle_for_round\", \"/settle_for_room\"):\n if i in url:\n return True\n return False\n\n # noinspection PyBroadException\n @coroutine\n def post(self, url, body, retry=False):\n key = hashlib.md5(url + body).hexdigest()\n if retry:\n self.reload_web_url()\n try:\n request = AsyncHTTPClient()\n response = yield request.fetch(url, method=\"POST\", body=body)\n self.logger.info(\"url: {0} code: {1} body {2}\".format(url, response.code, response))\n\n for key, value in redis.hgetall(name_web_retry).items():\n # print key, value\n redis.hdel(name_web_retry, key)\n i, j = value.split(\"|\", 1)\n m = re.findall(pattern, i)\n if self.is_web_alloc(i):\n i = i.replace(m[0], self.web_alloc_url)\n else:\n i = i.replace(m[0], self.web_record_url)\n self.post(i, j, True)\n except Exception:\n fp = StringIO()\n traceback.print_exc(file=fp)\n self.logger.fatal(\"url: {0} error: {1}\".format(url, fp.getvalue()))\n redis.hsetnx(name_web_retry, key, \"{0}|{1}\".format(url, body))\n\n @coroutine\n def enter_room(self, player):\n proto_web = game_pb2.EnterRoomWebResponse()\n proto_web.code = 1\n proto_web.room_id = self.room_id\n proto_web.player = player\n proto_web.game_type = self.game_type\n proto_web.app_id = self.app_id\n url = self.web_alloc_url + \"/room/enter\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def exit_room(self, player):\n proto_web = game_pb2.ExitRoomWebResponse()\n proto_web.player = player\n proto_web.code = 1\n proto_web.room_id = self.room_id\n proto_web.game_type = self.game_type\n proto_web.app_id = self.app_id\n url = self.web_alloc_url + \"/room/exit\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def dismiss_room(self, table):\n # print 'enter dismiss room'\n if table.conf.is_aa():\n if table.total_round <= 1:\n for p in table.player_dict.values():\n self.aa_refund(p.uuid, 0)\n # elif table.total_round < table.conf.max_rounds:\n # for p in table.player_dict.values():\n # self.aa_refund(p.uuid, 2)\n else:\n if table.total_round <= 1:\n self.refund()\n\n proto_web = game_pb2.DismissRoomWebResponse()\n proto_web.room_id = self.room_id\n proto_web.code = 1\n proto_web.game_type = self.game_type\n proto_web.app_id = self.app_id\n proto_web.owner = self.owner\n url = self.web_alloc_url + \"/room/dismiss\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def aa_refund(self, userid, refund_type):\n proto_web = game_pb2.AvgRefundRequest()\n proto_web.app_id = self.app_id\n proto_web.room_uuid = self.room_uuid\n proto_web.user_id = userid\n proto_web.refund_type = refund_type\n url = self.web_record_url + \"/aa/refund\"\n body = proto_web.SerializeToString()\n # print 'aa userid:', userid, ',refund_type:', refund_type\n self.post(url, body)\n\n @coroutine\n def refund(self):\n proto_web = game_pb2.RefundWebResponse()\n proto_web.room_id = self.room_id\n proto_web.code = 1\n proto_web.game_type = self.game_type\n proto_web.app_id = self.app_id\n proto_web.owner = self.owner\n proto_web.room_uuid = self.room_uuid\n url = self.web_record_url + \"/room/refund\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def settle_for_round(self, data):\n url = self.web_record_url + \"/settle_for_round/{0}\".format(self.game_type)\n body = json.dumps(data)\n self.post(url, body)\n\n @coroutine\n def settle_for_room(self, data):\n url = self.web_record_url + \"/settle_for_room/{0}\".format(self.game_type)\n body = json.dumps(data)\n self.post(url, body)\n\n @coroutine\n def load_plus(self):\n proto_web = game_pb2.LoadPlusWebResponse()\n proto_web.addr = host\n proto_web.port = options.server_port\n proto_web.app_id = self.app_id\n proto_web.room_uuid = self.room_uuid\n url = self.web_alloc_url + \"/room/load_plus\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def load_minus(self):\n proto_web = game_pb2.LoadPlusWebResponse()\n proto_web.addr = host\n proto_web.port = options.server_port\n proto_web.app_id = self.app_id\n proto_web.room_uuid = self.room_uuid\n url = self.web_alloc_url + \"/room/load_minus\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n\n @coroutine\n def aa_cons(self, userid):\n proto_web = game_pb2.GameConsRequest()\n proto_web.app_id = self.app_id\n proto_web.room_uuid = self.room_uuid\n proto_web.user_id = userid\n proto_web.cons_type = 1\n # print 'aa userid:', userid, ',aa_cons'\n url = self.web_record_url + \"/aa/cons\"\n body = proto_web.SerializeToString()\n self.post(url, body)\n","sub_path":"web/request.py","file_name":"request.py","file_ext":"py","file_size_in_byte":7070,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"137163696","text":"from tkinter import *\n\n# Η εντολή που θα εκτελεί το κουμπί Start\ndef start(simulation):\n simulation.run()\n simulation.has_started = True\n\n# Η εντολή που θα εκτελεί το κουμπί Pause\ndef pause(simulation, btn):\n simulation.is_paused = not simulation.is_paused\n\n if btn[\"text\"] == \"Pause\":\n btn[\"text\"] = \"Unpause\"\n elif btn[\"text\"] == \"Unpause\":\n btn[\"text\"] = \"Pause\"\n","sub_path":"uiFunctions.py","file_name":"uiFunctions.py","file_ext":"py","file_size_in_byte":449,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"50198974","text":"# 矩阵运算\nimport numpy as np\n\n# 点积运算\na = np.array([[1, 2], [3, 4]])\nb = np.array([[11, 12], [13, 14]])\n# 输出结果为\n# [[37 40]\n# [85 92]]\n# 计算公式为[[1*11+2*13, 1*12+2*14],[3*11+4*13, 3*12+4*14]]\n# print(np.dot(a, b))\n# 向量乘积\n# 输出结果为130\n# 计算公式为1*11 + 2*12 + 3*13 + 4*14 = 130\nprint(np.vdot(a, b))\n# 向量内积 numpy.inner() 函数返回一维数组的向量内积。对于更高的维度,它返回最后一个轴上的和的乘积\n# 输出结果为\n# [[35 41]\n# [81 95]]\n# 内积与矩阵乘积不一样\nprint(np.inner(a, b))\n\n# numpy.matmul 函数返回两个数组的矩阵乘积。 虽然它返回二维数组的正常乘积,但如果任一参数的维数大于2,\n# 则将其视为存在于最后两个索引的矩阵的栈,并进行相应广播\na = [[1, 0],[0, 1]]\nb = [[4, 1],[2, 2]]\n# 输出结果为\n# [[4 1]\n# [2 2]]\nprint(np.matmul(a, b))\n\na = [[1, 0], [0, 1]]\nb = [1, 2]\n# 输出结果为[1 2]\n# 将结果中附加的一行去掉\nprint(np.matmul(a, b))\n# 输出结果为[1 2]\nprint(np.matmul(b, a))\n\n# 矩阵的行列式\n# 输出结果为1.0\nprint(np.linalg.det(a))\n# 给出numpy.linalg.solve() 函数给出了矩阵形式的线性方程的解\n# numpy.linalg.inv() 函数计算矩阵的乘法逆矩阵\nx = np.array([[1, 2], [3, 4]])\ny = np.linalg.inv(x)\n# 输出结果为\n# [[-2. 1. ]\n# [ 1.5 -0.5]]\n# print(y)\n\na = np.array([[1, 1, 1], [0, 2, 5], [2, 5, -1]])\nainv = np.linalg.inv(a)\nb = np.array([[6], [-4], [27]])\nx = np.linalg.solve(a, b)\n# 方程的解为 A^-1b\n# [[ 5.]\n# [ 3.]\n# [-2.]]\nprint(x)","sub_path":"matrix.py","file_name":"matrix.py","file_ext":"py","file_size_in_byte":1581,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"585148750","text":"# Copyright (c) 2021 Microsoft Corporation. Licensed under the MIT license.\n\nimport cv2\nimport argparse\nimport json\n\nfrom scene_graph_benchmark.scene_parser import SceneParser\nfrom scene_graph_benchmark.AttrRCNN import AttrRCNN\nfrom maskrcnn_benchmark.data.transforms import build_transforms\nfrom maskrcnn_benchmark.utils.checkpoint import DetectronCheckpointer\nfrom maskrcnn_benchmark.config import cfg\nfrom scene_graph_benchmark.config import sg_cfg\nfrom maskrcnn_benchmark.data.datasets.utils.load_files import \\\n config_dataset_file\nfrom maskrcnn_benchmark.data.datasets.utils.load_files import load_labelmap_file\nfrom maskrcnn_benchmark.utils.miscellaneous import mkdir\n\n# from tools.demo.detect_utils import detect_objects_on_single_image\nfrom tools.demo.visual_utils import draw_bb, draw_rel\n\nimport os\nimport pickle\nimport threading\n\nfrom tqdm import tqdm\nimport torch\n\nfrom tools.demo.detect_utils import build_dataloader, detect_objects_on_batch\n\n\ndef postprocess_attr(dataset_attr_labelmap, label_list, conf_list):\n common_attributes = {\n 'white', 'black', 'blue', 'green', 'red', 'brown', 'yellow', 'small', 'large', 'silver', 'wooden',\n 'wood', 'orange', 'gray', 'grey', 'metal', 'pink', 'tall', 'long', 'dark', 'purple'\n }\n common_attributes_thresh = 0.1\n attr_alias_dict = {'blonde': 'blond'}\n attr_dict = {}\n for label, conf in zip(label_list, conf_list):\n label = dataset_attr_labelmap[label]\n if label in common_attributes and conf < common_attributes_thresh:\n continue\n if label in attr_alias_dict:\n label_target = attr_alias_dict[label]\n else:\n label_target = label\n if label_target in attr_dict:\n attr_dict[label_target] += conf\n else:\n attr_dict[label_target] = conf\n if len(attr_dict) > 0:\n # the most confident one comes the last\n sorted_dic = sorted(attr_dict.items(), key=lambda kv: kv[1])\n return list(zip(*sorted_dic))\n else:\n return [[], []]\n\n\ndef main():\n parser = argparse.ArgumentParser(description=\"Object Detection Demo\")\n parser.add_argument(\"--config-file\", metavar=\"FILE\",\n help=\"path to config file\")\n parser.add_argument(\"--video-dir\", required=True)\n parser.add_argument(\"--video-list\", default=None)\n parser.add_argument(\"--save-dir\", required=True)\n parser.add_argument(\"opts\", default=None, nargs=argparse.REMAINDER,\n help=\"Modify config options using the command-line\")\n\n args = parser.parse_args()\n cfg.set_new_allowed(True)\n cfg.merge_from_other_cfg(sg_cfg)\n cfg.set_new_allowed(False)\n cfg.merge_from_file(args.config_file)\n cfg.merge_from_list(args.opts)\n cfg.freeze()\n\n if not os.path.isdir(args.save_dir):\n mkdir(args.save_dir)\n\n if args.video_list is not None:\n vids = open(args.video_list).read().splitlines()\n else:\n vids = next(os.walk(args.video_dir))[1]\n\n print(f'Processing {len(vids)} videos')\n model = AttrRCNN(cfg)\n model.to('cuda')\n model.eval()\n\n checkpointer = DetectronCheckpointer(cfg, model)\n checkpointer.load(cfg.MODEL.WEIGHT)\n\n transforms = build_transforms(cfg, is_train=False)\n for i, video in enumerate(vids):\n output = os.path.join(args.save_dir, video) + '.pkl'\n if os.path.exists(output):\n continue\n\n loader = build_dataloader(os.path.join(args.video_dir, video),\n transforms, cfg.TEST.IMS_PER_BATCH)\n dets = detect_objects_on_batch(model, loader)\n pickle.dump(dets, open(output, 'wb'))\n if (i+1) % 10 == 0:\n print(f'{i+1}/{len(vids)}', flush=True)\n\n torch.cuda.empty_cache()\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"tools/demo/demo_batch.py","file_name":"demo_batch.py","file_ext":"py","file_size_in_byte":3778,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"615287843","text":"# Усредняем по 10 минут\n\nimport sqlalchemy\nfrom sqlalchemy import Column, Integer, String, Float, create_engine, DateTime\nfrom sqlalchemy import func\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker\nimport threading\nimport datetime\nimport logging\nimport time\n\nlogger = logging.getLogger(\"mitter.db_aver_10min\")\n\n# Работа с базой данных\nDB_Base = declarative_base()\nDB_Engine = create_engine(\"postgres://mitter:mitter@localhost/mitter\")\nDB_Session = sessionmaker(bind = DB_Engine)\n\nclass DataAverage10min(threading.Thread):\n\t\"\"\" \"\"\"\n\tdef __init__(self):\n\t\tthreading.Thread.__init__(self)\n\t\tDB_Base.metadata.create_all(DB_Engine)\n\n\n\tdef run(self):\n\t\tlogger.info('Поток усреднения данных за 10 минуту запущен')\n\t\ttry:\n\t\t\twhile True:\n\n\t\t\t\tch = 0 \n\t\t\t\tavr = self._get_aver_db(ch)\n\t\t\t\tif avr != None:\n\t\t\t\t\tself._put_aver_db(ch, avr)\n\n\t\t\t\tch = 1 \n\t\t\t\tavr = self._get_aver_db(ch)\n\t\t\t\tif avr != None:\n\t\t\t\t\tself._put_aver_db(ch, avr)\n\t\t\t\n\t\t\t\tch = 2 \n\t\t\t\tavr = self._get_aver_db(ch)\n\t\t\t\tif avr != None:\n\t\t\t\t\tself._put_aver_db(ch, avr)\n\n\t\t\t\ttime.sleep(10*60)\n\n\t\texcept Exception:\n\t\t\tlogger.exception(\"Исключение в потоке\")\n\n\tdef _get_aver_db(self, ch):\n\t\tsession = DB_Session()\n\t\tcurrent_time = datetime.datetime.now()\n\t\tlast_time = current_time - datetime.timedelta(seconds=10*60)\n\t\taver = session.query(func.avg(AverageCurr_1min.value)).filter(\n\t\t\t AverageCurr_1min.time > last_time).filter(\n\t\t\t AverageCurr_1min.ch == ch).all()\n\t\tsession.close()\n\t\treturn aver[0][0]\n\n\tdef _put_aver_db(self, ch, value):\n\t\tsession = DB_Session()\n\t\tAC = AverageCurr_10min(value, ch)\n\t\tsession.add(AC)\n\t\tsession.commit()\n\t\tsession.close()\n\nclass AverageCurr_1min(DB_Base):\n\t__tablename__ = 'AVERCUR1MIN'\n\tid = Column(Integer, primary_key = True)\n\ttime = Column(DateTime(timezone=True), default=datetime.datetime.now)\n\tch = Column(Integer)\n\tvalue = Column(Float)\n\n\tdef __init__(self, value, ch):\n\t\tself.value = value\n\t\tself.ch = ch\n\nclass AverageCurr_10min(DB_Base):\n\t__tablename__ = 'AVERCUR10MIN'\n\tid = Column(Integer, primary_key = True)\n\ttime = Column(DateTime(timezone=True), default=datetime.datetime.now)\n\tch = Column(Integer)\n\tvalue = Column(Float)\n\n\tdef __init__(self, value, ch):\n\t\tself.value = value\n\t\tself.ch = ch\n","sub_path":".Trash-1000/files/db_aver_10min.py","file_name":"db_aver_10min.py","file_ext":"py","file_size_in_byte":2330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"286067105","text":"from direct.directbase.DirectStart import *\nfrom pandac.PandaModules import *\nfrom Vehicle import Vehicle\n\nclass TopView(object):\n def __init__(self, vehicle):\n view = base.win.makeDisplayRegion(.65, .95, .65, .95)\n #view.setClearColor(VBase4(1, 1, 1, .5))\n #view.setClearColorActive(True)\n view.setClearDepthActive(True)\n \n self.vehicle = vehicle\n \n self.cam = render.attachNewNode(Camera('cam2'))\n view.setCamera(self.cam)\n \n self.cam.setPos(self.vehicle.chassisNP.getPos())\n #self.cam.setPos(self.vehicle.chassisNP, (0, 0, 150))\n self.cam.setZ(150)\n self.cam.setP(-90)\n \n def update(self, task):\n self.cam.setPos(self.vehicle.chassisNP.getPos())\n #self.cam.setPos(self.vehicle.chassisNP, (0, 0, 150))\n self.cam.setH(self.vehicle.chassisNP.getH())\n self.cam.setZ(250)\n self.cam.setP(-90)\n \n return task.again\n \n ","sub_path":"new-master-client-side-20151208000001/branches/dd_fix_disc_player/TopView.py","file_name":"TopView.py","file_ext":"py","file_size_in_byte":984,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"335253639","text":"import logging\r\n\r\nstdio_handler = None\r\nfacedancer_logger = None\r\n\r\n\r\ndef prepare_logging():\r\n global facedancer_logger\r\n global stdio_handler\r\n if facedancer_logger is None:\r\n def add_debug_level(num, name):\r\n def fn(self, message, *args, **kwargs):\r\n if self.isEnabledFor(num):\r\n self._log(num, message, args, **kwargs)\r\n logging.addLevelName(num, name)\r\n setattr(logging, name, num)\r\n return fn\r\n\r\n logging.Logger.verbose = add_debug_level(5, 'VERBOSE')\r\n logging.Logger.always = add_debug_level(100, 'ALWAYS')\r\n\r\n FORMAT = '[%(levelname)-6s] %(message)s'\r\n stdio_handler = logging.StreamHandler()\r\n stdio_handler.setLevel(logging.INFO)\r\n formatter = logging.Formatter(FORMAT)\r\n stdio_handler.setFormatter(formatter)\r\n facedancer_logger = logging.getLogger('facedancer')\r\n facedancer_logger.addHandler(stdio_handler)\r\n facedancer_logger.setLevel(logging.VERBOSE)\r\n return facedancer_logger\r\n\r\n\r\ndef set_default_handler_level(level):\r\n global stdio_handler\r\n stdio_handler.setLevel(level)\r\n\r\ndef get_logger(verbose):\r\n levels = {\r\n 0: logging.INFO,\r\n 1: logging.DEBUG,\r\n # verbose is added by facedancer.__init__ module\r\n 2: logging.VERBOSE,\r\n }\r\n logger = logging.getLogger('facedancer')\r\n if verbose in levels:\r\n set_default_handler_level(levels[verbose])\r\n else:\r\n set_default_handler_level(logging.VERBOSE)\r\n # if self.options.get('--quiet', False):\r\n # set_default_handler_level(logging.WARNING)\r\n return logger\r\n","sub_path":"facedancer/utils/ulogger.py","file_name":"ulogger.py","file_ext":"py","file_size_in_byte":1718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"123634769","text":"from db import post_collection, user_post\n\ndef add_user_post(title,username,img,p,location,vehicle,tipsfortravel):\n\n new_post = {\n \"Title\" :title,\n \"Username\": username,\n \"img_link\": img,\n \"content\" : p,\n \"location\": location,\n \"vehicle\": vehicle,\n \"tipsfortravel\": tipsfortravel,\n }\n user_post.insert_one(new_post)","sub_path":"Webproject-master/Webproject/up_user.py","file_name":"up_user.py","file_ext":"py","file_size_in_byte":373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"487306175","text":"import cv2\nimport numpy as np\nfrom scipy import signal\n\nfrom estimateFeatureTranslation import estimateFeatureTranslation\n\ndef return_derivatives(img):\n dx = np.array([1, -1]).reshape(1,-1)\n dy = np.array([1, -1]).reshape(-1,1)\n Ix = signal.convolve2d(img, dx)\n Iy = signal.convolve2d(img, dy)\n return Ix, Iy\n\ndef estimateAllTranslation(startXs, startYs, img1, img2):\n newXs = np.empty_like(startXs)\n newYs = np.empty_like(startYs)\n for idx1, (colX, colY) in enumerate(zip(startXs.T, startYs.T)):\n for idx2, (startX, startY) in enumerate(zip(colX, colY)):\n gray1 = cv2.cvtColor(img1, cv2.COLOR_BGR2GRAY)\n gray2 = cv2.cvtColor(img2, cv2.COLOR_BGR2GRAY)\n Ix, Iy = return_derivatives(gray1)\n if startX >= 0 and startY >= 0:\n newX, newY = estimateFeatureTranslation(startX, startY, Ix, Iy, gray1, gray2)\n newXs[idx2,idx1] = newX\n newYs[idx2,idx1] = newY\n else:\n newXs[idx2, idx1] = -1\n newYs[idx2, idx1] = -1\n\n return newXs, newYs\n","sub_path":"estimateAllTranslation.py","file_name":"estimateAllTranslation.py","file_ext":"py","file_size_in_byte":1091,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"303671851","text":"import requests\nimport re\nimport bs4\nimport sys\nheader = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/79.0.3945.117 Safari/537.36\"}\n\ntop = open(\"top.txt\", \"w\", encoding=\"utf-8\")\n\nno=1\npage=1\nwhile no<=150:\n url = \"http://bang.dangdang.com/books/fivestars/01.00.00.00.00.00-recent30-0-0-1-\"+str(page)\n response=requests.get(url)\n bsobj=bs4.BeautifulSoup(response.text.encode(encoding=\"utf-8\").decode(encoding=\"utf-8\"))\n elem=bsobj.select(\".name\")\n pattern=re.compile(\"[\\u4e00-\\u9fa5]{1,}\")\n for i in elem:\n han_word=re.search(pattern,i.getText()).group(0)\n top.write(\"top\"+str(no)+\": \"+str(han_word)+\"\\n\")\n print(\"已下载:{:.2%}\".format(no/150),end=\"\\r\")\n if no==150:\n sys.exit(0) \n no+=1\n page+=1\ntop.close()\n","sub_path":"爬取当当网排名前150的图书/sprider.py","file_name":"sprider.py","file_ext":"py","file_size_in_byte":843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"142774817","text":"import torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils.data import DataLoader, Dataset\n\n\n# Matrix Factorization With Biases\nclass LFM(torch.nn.Module):\n def __init__(self, num_user, num_item, hidden, mu):\n super(LFM, self).__init__()\n self.mu = mu\n self.user_embed = nn.Embedding(num_user, hidden)\n self.item_embed = nn.Embedding(num_item, hidden)\n self.bias_u = nn.Embedding(num_user, 1)\n self.bias_i = nn.Embedding(num_item, 1)\n self.init_params()\n\n def init_params(self):\n nn.init.normal_(self.user_embed.weight, std=0.01)\n nn.init.normal_(self.item_embed.weight, std=0.01)\n nn.init.constant_(self.bias_u.weight, 0.0)\n nn.init.constant_(self.bias_i.weight, 0.0)\n\n def forward(self, user_indexs, item_indexs):\n P = self.user_embed(user_indexs) # [batch_num,hidden]\n Q = self.item_embed(item_indexs) # [batch_num,hidden]\n interaction = torch.mul(P, Q).sum(dim=1).unsqueeze(dim=1) # element-wise product [batch_num,1]\n return self.mu + self.bias_u(user_indexs) + self.bias_i(item_indexs) + interaction # [batch_num,1]\n","sub_path":"recommend_system_homework/01_UserCF_ItemCF_MF/CF_MF_homework/MF/MF_model.py","file_name":"MF_model.py","file_ext":"py","file_size_in_byte":1161,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"362856620","text":"import sys\nimport ROOT\nimport pandas as pd\nimport numpy as np\n\n\nif __name__==\"__main__\":\n\targs=sys.argv\n\tfilename=args[1]\n\tdata=[]\n\tdata=pd.read_csv(filename,delimiter=',',dtype=None,skiprows=0)\n\n##Make Graph Data\n#\tdata.columns=['','','']\n\tx=data[\"1\"]\n\ty=data[\"2\"]\n\tn=data.shape[0]\n\tg1=ROOT.TGraph(n,x,y)\n\tg1.SetMarkerColor(2)\n\tg1.SetLineColor(2)\n\tg1.SetTitle(\"\")\n\tg1.GetXaxis().SetTitle(\"\")\n\tg1.GetYaxis().SetTitle(\"\")\n\tg1.Draw()\n\ndef stop(self):\n\tsys.stderr.write('[Read]\\tstop.\\tPress \"q\"to quit>')\n\tans=raw_input ('>')\n\tif ans in ['q','Q']:\n\t\tg1.IsA().Destructor(g1)\n\t\tsys.exit(-1)\n\telif ans in ['.','q','Q']:\n\t\treturn -1\n","sub_path":"pdGraph.py","file_name":"pdGraph.py","file_ext":"py","file_size_in_byte":627,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"309807","text":"from __future__ import division\nimport random\nimport numpy as np\nimport tensorflow as tf\nfrom collections import deque\nimport os\nimport networkx as nx\nimport time\nimport sys\n\n\nclass ReplayBuffer(object):\n \n def __init__(self, buffer_size, random_seed = 123):\n self.buffer_size = buffer_size\n self.count = 0\n self.buffer = deque()\n random.seed(random_seed)\n\n def add(self, s, a, r, t, s2):\n experience = (s, a, r, t, s2)\n if self.count < self.buffer_size: \n self.buffer.append(experience)\n self.count += 1\n else:\n self.buffer.popleft()\n self.buffer.append(experience)\n\n def size(self):\n return self.count\n\n def sample_batch(self, batch_size):\n\n batch = []\n\n if self.count < batch_size:\n ran_num = np.arange(self.count)\n batch = list(self.buffer)\n else:\n ran_num = np.random.choice(self.count, batch_size)\n batch = [self.buffer[i] for i in ran_num]\n\n s_batch = np.array([_[0] for _ in batch])\n a_batch = np.array([_[1] for _ in batch])\n r_batch = np.array([_[2] for _ in batch])\n t_batch = np.array([_[3] for _ in batch])\n s2_batch = np.array([_[4] for _ in batch])\n # index = np.array([_[0] for _ in batch])\n\n return s_batch, a_batch, r_batch, t_batch, s2_batch\n\n def clear(self):\n self.buffer.clear()\n self.count = 0\n \ndef build_summaries():\n episode_reward = tf.Variable(0.)\n tf.summary.scalar(\"Reward\", episode_reward)\n episode_ave_max_q = tf.Variable(0.)\n tf.summary.scalar(\"Qmax_Value\", episode_ave_max_q)\n exploration_rate = tf.Variable(0.)\n tf.summary.scalar(\"Exploration\", exploration_rate)\n lr_theta = tf.Variable(0.)\n tf.summary.scalar(\"LR_theta\", lr_theta)\n lr_lam = tf.Variable(0.)\n tf.summary.scalar(\"LR_lam\", lr_lam)\n\n summary_vars = [episode_reward, episode_ave_max_q, exploration_rate, lr_theta, lr_lam]\n summary_ops = tf.summary.merge_all()\n\n return summary_ops, summary_vars\n\n# class Rabin_Automaton(object):\n \n# def __init__(self, ltl, coord_dict):\n# self.ltl = ltl\n# self.coord_dict = coord_dict\n\n# with open(\"my.ltl\", \"w\") as ltlfile:\n# ltlfile.write(ltl)\n\n# if sys.platform == \"darwin\":\n# result1 = os.system(\"ltlfilt -l -F \\\"my.ltl\\\" --output=\\\"my.ltl\\\" \")\n# else:\n# result1 = os.system(\"./ltlfilt_ubuntu -l -F \\\"my.ltl\\\" --output=\\\"my.ltl\\\" \")\n# print result1\n\n# if sys.platform == \"darwin\":\n# result2 = os.system(\"./ltl2dstar --ltl2nba=spin:ltl2ba --stutter=no --output-format=dot my.ltl my.dot\")\n# else:\n# result2 = os.system(\"./ltl2dstar_ubuntu --ltl2nba=spin:ltl2ba_ubuntu --stutter=no --output-format=dot my.ltl my.dot\")\n# print result2\n\n# rabin_graph = nx.nx_agraph.read_dot(\"my.dot\")\n# rabin_graph.remove_nodes_from([\"comment\", \"type\"])\n \n# self.graph = rabin_graph\n# self.num_of_nodes = len(self.graph.node)\n \n# self.accept = [int(i) for i in self.graph.node if \"+0\" in self.graph.node[i][\"label\"]]\n# self.reject = [int(i) for i in self.graph.node if \"-0\" in self.graph.node[i][\"label\"]]\n \n# self.deadlock = []\n# for i in self.reject:\n# if str(i) in self.graph[str(i)].keys():\n# if \" true\" in [ self.graph[str(i)][str(i)][j][\"label\"] \n# for j in range(len(self.graph[str(i)][str(i)])) ]:\n# self.deadlock.append(i)\n \n# for i in self.graph.node:\n# if \"fillcolor\" in self.graph.node[i].keys():\n# if self.graph.node[i][\"fillcolor\"] == \"grey\":\n# self.init_state = int(i)\n# break\n \n# def get_graph(self):\n# return self.graph\n \n# def get_init_state(self):\n# return self.init_state\n \n# def next_state(self, current_state, next_coord):\n# ap_next = self.coord_dict[tuple(next_coord)]\n# next_states = self.possible_states(current_state[2])\n# for i in next_states:\n# next_state_aps = [self.graph[str(current_state[2])][str(i)][k][\"label\"] \n# for k in range(len(self.graph[str(current_state[2])][str(i)]))]\n# # May need to change later\n# if \" true\" in next_state_aps:\n# return current_state[-1]\n# else:\n# for j in next_state_aps:\n# if self.check_ap(ap_next, j):\n# return i\n \n# def possible_states(self, current_rabin_state):\n# return [int(i) for i in self.graph[str(current_rabin_state)].keys()]\n \n# def check_ap(self, ap_next, ap_sentence):\n# pos, neg = seperate_ap_sentence(ap_sentence)\n# if set(ap_next).issuperset(set(pos)) and self.check_neg(ap_next, neg):\n# return True\n# return False\n \n# def check_neg(self, ap, negs):\n# for i in ap:\n# if i in negs:\n# return False\n# return True\n \n# def seperate_ap_sentence(input_str):\n# if len(input_str)>1:\n# index = find_ampersand(input_str)\n# if len(index)>=1:\n# return_str = [input_str[0:index[0]]]\n# else:\n# return_str = input_str\n# if '!' in return_str:\n# return [],[return_str.replace('!','')]\n# else:\n# return [return_str],[]\n# for i in range(1,len(index)):\n# return_str += [input_str[index[i-1]+1:index[i]]]\n# return_str = return_str + [input_str[index[-1]+1:]]\n# return_str = [i.replace(' ','') for i in return_str]\n# elif len(input_str)==1:\n# return_str = input_str\n# elif len(input_str) == 0: \n# raise AttributeError('input_str has no length')\n \n# without_negs = []\n# negations = []\n# for i in range(len(return_str)):\n# if '!' in return_str[i]:\n# negations += [return_str[i].replace('!','')]\n# else:\n# without_negs += [return_str[i]]\n# return without_negs,negations\n \n# def find_ampersand(input_str):\n# index = []\n# original_length = len(input_str)\n# original_str = input_str\n# while input_str.find('&')>=0:\n# index += [input_str.index('&')-len(input_str)+original_length]\n# input_str = original_str[index[-1]+1:]\n# return index\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":6734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"266937421","text":"def decode():\n lst = ['']\n for i in string:\n if lst[-1] == i:\n lst.pop()\n else:\n lst.append(i)\n return lst\n\n\ndecode_in = open('decode.in', 'r')\ndecode_out = open('decode.out', 'w')\n\n# Читаем\nstring = decode_in.readline().rstrip()\ndecode_in.close()\n\nprint(*decode(), file=decode_out, sep='')\ndecode_out.close()\n","sub_path":"LKSH2/0703/decode.py","file_name":"decode.py","file_ext":"py","file_size_in_byte":361,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"473252374","text":"import nltk\r\n\r\n#import all books from nltk.books\r\nfrom nltk.book import *\r\n#import text 6\r\ntext_mp=[x.lower() for x in text6]\r\nfreq_mp={x:text_mp.count(x) for x in set(text_mp)}\r\n#import text 7\r\ntext_wsj=[x.lower() for x in text7]\r\nfreq_wsj={x:text_wsj.count(x) for x in set(text_wsj)}\r\n#sort text 6\r\nword_r_mp=sorted(freq_mp.items(), reverse=True, key=lambda x : x[1])\r\n#sort text 7\r\nword_r_wsj=sorted(freq_wsj.items(), reverse=True, key=lambda x : x[1])\r\n# eliminating stopwords\r\nfrom nltk.corpus import stopwords\r\nstop_words=set(stopwords.words('english'))\r\n\r\nfor(w,c) in word_r_mp:\r\n if w in stop_words:\r\n word_r_mp.remove((w,c))\r\nfor(w,c) in word_r_wsj:\r\n if w in stop_words:\r\n word_r_wsj.remove((w,c))\r\n\r\n# remove the same words\r\nfor(w,c) in word_r_mp:\r\n if w in word_r_wsj:\r\n word_r_mp.remove((w,c))\r\nfor(w,c) in word_r_wsj:\r\n if w in word_r_mp:\r\n word_r_wsj.remove((w,c))\r\n\r\n#create dictionary from text 6 and 7\r\nmp_dict=dict(word_r_mp)\r\nwsj_dict=dict(word_r_wsj)\r\n\r\n# checking how many times the word 'knight' appears\r\nprint(mp_dict)\r\nprint('-----------------')\r\nprint(wsj_dict)","sub_path":"Exercise16.py","file_name":"Exercise16.py","file_ext":"py","file_size_in_byte":1127,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"615462974","text":"import re\nimport helpers\nfrom os import remove\nfrom os.path import join\n\nfrom gestor_ficheros import GestorFicheros\nfrom config import Cliente, DataPath\n\nclass Manager:\n\n __clientes: list = []\n\n def __init__(self):\n self.__clientes = GestorFicheros.loadClient()\n print(self.__clientes)\n\n @staticmethod\n def show_client(client: Cliente):\n \"\"\" Muestra por pantalla un cliente de forma amigable \"\"\"\n print(client)\n\n def show_clients(self):\n \"\"\" Recorre la lista de clientes y los muetra uno a uno \"\"\"\n if self.__clientes == []:\n print(\"No hay clientes disponibles.\")\n else:\n for cliente in self.__clientes:\n Manager.show_client(cliente)\n\n def add(self):\n \"\"\" Añade un cliente a la lista de clientes \"\"\"\n\n print('Introduce nombre (De 2 a 30 caracteres)')\n nombre = helpers.input_text(2, 30)\n print('Introduce apellido (De 2 a 30 caracteres)')\n apellidos = helpers.input_text(2, 30)\n while True:\n print(\"Introduce DNI (2 números y 1 carácter en mayúscula)\")\n dni = helpers.input_text(3, 3)\n if self.is_valid(dni):\n GestorFicheros.writeCliente(dni, nombre, apellidos)\n self.__clientes.append(Cliente(nombre, apellidos, dni))\n break\n else:\n print(\"DNI incorrecto o en uso\")\n dni = None \n\n def is_valid(self, dni: str):\n if not re.match('[0-9]{2}[A-Z]', dni):\n return False\n\n for client in self.__clientes:\n if client.dni == dni:\n return False\n\n return True\n\n def find(self):\n \"\"\" Busca un cliente y lo devuelve junto a su índice \"\"\"\n dni = input(\"Introduce el dni del cliente\\n> \")\n for i, client in enumerate(self.__clientes):\n if client.dni == dni:\n Manager.show_client(client)\n return i, client\n print(\"No se ha encontrado ningún cliente con ese DNI\")\n return None, None\n\n def delete(self):\n \"\"\" Borra un cliente de la lista a partir del dni \"\"\"\n\n i, client = self.find()\n\n if client:\n client = self.__clientes.pop(i)\n remove(join(DataPath, '{}.json'.format(client.dni)))\n return True\n\n return False\n\n def edit(self):\n \"\"\" Modifica el nombre y apellido de un cliente a partir del dni \"\"\"\n \n i, client = self.find()\n if client:\n \n print(f\"Introduce nuevo nombre ({client.nombre})\")\n nombre = helpers.input_text(2, 30)\n self.__clientes[i].nombre = nombre\n\n print(f\"Introduce nuevo apellido ({client.apellidos})\")\n apellidos = helpers.input_text(2, 30)\n self.__clientes[i].apellidos = apellidos\n\n GestorFicheros.writeCliente(client.dni, nombre, apellidos)\n\n return True\n","sub_path":"gestor/manager.py","file_name":"manager.py","file_ext":"py","file_size_in_byte":2967,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"314869498","text":"import os\r\nimport sys\r\nimport fcntl\r\nimport json\r\nimport time\r\nimport subprocess\r\nimport traceback\r\n\r\nimport jedi\r\n\r\n\r\n# fcntl.fcntl(sys.stdin, fcntl.F_SETFL, os.O_NONBLOCK)\r\nlogfile = open(os.path.expanduser('~/.log'), 'a')\r\ndef log(s):\r\n logfile.write(str(s) + '\\n')\r\n logfile.flush()\r\n\r\n\r\ndef log_exception():\r\n log(traceback.format_exc())\r\n\r\n\r\ndef write_vim(msg):\r\n sys.stdout.write(json.dumps(msg) + '\\n')\r\n sys.stdout.flush()\r\n\r\ndef make_jedis(args):\r\n with open(args['codepath'], 'r') as f:\r\n code = f.read().strip()\r\n \r\n # set sys path\r\n sys_path = sys.path[:]\r\n tmpdir = os.path.dirname(args['filepath'])\r\n if not tmpdir.startswith('/'):\r\n while tmpdir != '':\r\n sys_path.append(tmpdir)\r\n tmpdir = os.path.dirname(tmpdir)\r\n sys_path.append(tmpdir)\r\n\r\n jedis = jedi.Script(code,\r\n sys_path=sys_path,\r\n path=str(time.time()),\r\n line=args['line'],\r\n column=args['col'] - 1\r\n )\r\n return jedis\r\n \r\n################################################################################\r\n \r\ndef jedi_complete(action, args):\r\n jedis = make_jedis(args)\r\n with open(args['rstpath'], 'w') as f:\r\n for line in jedis.completions():\r\n f.write(json.dumps({\r\n 'word': line.name,\r\n 'menu': line.description,\r\n 'icase': 1\r\n }) + '\\n')\r\n\r\n write_vim({\r\n 'action': action,\r\n 'args': {}\r\n })\r\n \r\n################################################################################ \r\n \r\ndef jedi_gotodef(action, args):\r\n jedis = make_jedis(args)\r\n with open(args['rstpath'], 'w') as f:\r\n for line in jedis.goto_definitions():\r\n f.write(json.dumps({\r\n 'filepath': line.module_path,\r\n 'line': line.line\r\n }) + '\\n')\r\n\r\n write_vim({\r\n 'action': action,\r\n 'args': {}\r\n })","sub_path":"server/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"138995211","text":"import socket\nimport asyncio\nimport sys\nimport queue\nimport threading\nclass socket1():\n\tdef __init__(self,target):\n\t\tself.target=target\n\t\t# print(target)\n\t\t\n\n\tdef scan(self,i):\n\t\t# print(\"start scan\")\n\n\t\t# print(s.connect_ex((self.target,80)))\n\t\t# for i in range(1,100):\n\t\t\t# print(i)\n\t\ts=socket.socket()\n\t\ts.settimeout(0.1)\n\t\tif s.connect_ex((self.target,i))==0:\n\t\t\tprint(i,'open')\n\t\t\ts.close()\n\tdef worker(self,q):\n\t\twhile not q.empty():\n\t\t\tport=q.get()\n\t\t\ttry:\n\t\t\t\tself.scan(port)\n\t\t\tfinally:\n\t\t\t\tq.task_done()\n\t# def main(self):\n\t# \tprint(\"start to detect \",self.target)\n\t# \tloop=asyncio.get_event_loop()\n\t# \ttasks=[asyncio.ensure_future(self.scan(port)) for port in range(1,65536)]\n\t# \tloop.run_until_complete(asyncio.wait(tasks))\nif __name__ == '__main__':\n\tprint(\"Start testing the target port\")\n\t# print(\"Example:['127.0.0.1','127.0.0.2'] 80\")\n\tif len(sys.argv)!=2:\n\t\tprint(\"format wrong\")\n\t\tsys.exit()\n\t# iplist=[]\n\ttarget=sys.argv[1]\n\tq=queue.Queue()\n\ta=socket1(target)\n\tlist(map(q.put,range(1,65535)))\n\t# for i in range(65535):\n\t# \tprint(q.get())\n\tthreads=[threading.Thread(target=a.worker,args=(q,)) for i in range(800)]\n\tlist(map(lambda x:x.start(),threads))\n\tq.join()\n\tprint(\"scan over\")\n\n\t\n\n\t\n\t# for i in sys.argv:\n\t# \tiplist.append(i)\n\t# del iplist[0]\n\t# print(iplist)\n\t# port=int(sys.argv[1])\n\t# del iplist[-1]\n\t# print(port)\n\t\n\t# a.scan(80)\n\t# for i in range(1,100):\n\t# \ta.scan(i)\n\t\n\n","sub_path":"ports(threads).py","file_name":"ports(threads).py","file_ext":"py","file_size_in_byte":1402,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"116596835","text":"import os\nfilename=input(\"filename:\")\nif os.path.exists(filename):\n\tprint (\"file exit\")\n\texit()\nelse:\n\tfd=open(\"/hcx/learn/python3/python/basis/3_file/test.txt\",'w+')\n\tfd.writelines(\"huangchangxin!\\\\n\")\n\tfd.seek(0)\n\tstr=fd.readlines()\n\tif str==\"huangchangxin!\\\\n\":\n\t\tprint(\"write correct!\")\n\telse:\n\t\tprint(\"write error!\")\n\tfd.close()\nprint(\"done\")\n","sub_path":"basis/3_file/3_1th.py","file_name":"3_1th.py","file_ext":"py","file_size_in_byte":348,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"520595382","text":"import socket\nimport _thread\nimport protocol\nimport numpy as np\nfrom scipy.io import wavfile\nimport matplotlib.pyplot as plt\nimport speech_recognition as sr\n\n# CONFIG\nframerate = 16000\ntotal = 1 # seconds\n\n# SOCKET\nsock_udp = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\nsock_udp.bind(('127.0.0.1', 1336))\n\nsock_tcp = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\nsock_tcp.bind(('127.0.0.1', 1338))\nsock_tcp.listen(5)\n\n# APPLIANCES TO SEND COMMAND\nappliances = []\nrecorders = []\n\n# SPHINX\nr = sr.Recognizer()\n\ndef appliance_name_existed(name):\n for appliance in appliances:\n if appliance['name'] == name:\n return True\n return False\n\ndef get_recorder(ip, port):\n for recorder in recorders:\n if recorder['ip'] == ip and recorder['port'] == port:\n return recorder\n\ndef process_audio_v2(wave, samplerate):\n wavfile.write('temp.wav', samplerate, wave)\n with sr.AudioFile('temp.wav') as source:\n audio = r.record(source)\n print('recognized:', r.recognize_sphinx(audio))\n\ndef process_audio(wave):\n # select random appliance\n if len(appliances) == 0:\n print('no appliance connected')\n return\n appliance = np.random.choice(appliances)\n\n # select random action\n value = np.random.randint(appliance['lower'], appliance['upper'] + 1)\n\n # construct command packet and send to appliance\n packet = protocol.command_encode(type_='CMD', value=value)\n appliance['client'].send(packet)\n\n # check if client acknowled by packet ACK\n packet = appliance['client'].recv(1024)\n if protocol.command_decode(packet):\n print('adjusted', appliance['name'], 'to', value, 'success')\n else:\n print('adjusted', appliance['name'], 'to', value, 'fail')\n appliances.pop(appliances.index(appliance))\n print('removed', appliance['name'])\n\ndef handle_audio():\n wave = np.array([])\n\n while True:\n packet, addr = sock_udp.recvfrom(1024)\n\n recorder = get_recorder(addr[0], addr[1])\n if recorder == None:\n recorder = {\n 'ip' : addr[0],\n 'port' : addr[1],\n 'wave' : np.array([])\n }\n print('accepted recorder from:', addr)\n recorders.append(recorder)\n\n sr, tmp = protocol.audio_decode(packet)\n recorder['wave'] = np.append(recorder['wave'], tmp)\n if recorder['wave'].shape[0] > framerate * total:\n # plt.plot(recorder['wave'])\n # plt.show()\n process_audio(recorder['wave'])\n # process_audio_v2(recorder['wave'], sr)\n recorder['wave'] = np.array([])\n\ndef handle_appliance():\n while True:\n client, addr = sock_tcp.accept()\n packet = client.recv(1024)\n name, upper, lower = protocol.command_decode(packet)\n\n if appliance_name_existed(name):\n packet = protocol.command_encode(type_='ERR', value='Appliance name existed')\n client.send(packet)\n client.close()\n print('rejected applicance from:', addr)\n print(' - name:', name)\n else:\n packet = protocol.command_encode(type_='ACK')\n client.send(packet)\n print('accepted applicance from:', addr)\n print(' - name:', name)\n print(' - upper:', upper)\n print(' - lower:', lower)\n\n appliances.append({\n 'ip' : addr[0],\n 'port' : addr[1],\n 'client' : client,\n 'name' : name,\n 'upper' : upper,\n 'lower' : lower,\n })\n\ndef main():\n _thread.start_new_thread(handle_appliance, ())\n handle_audio()\n\nif __name__ == '__main__':\n main()\n","sub_path":"py/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":3749,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"474276447","text":"import re\nfrom dataclasses import dataclass\n\nfrom selenium.webdriver.remote.webelement import WebElement\n\n\n@dataclass\nclass Vacancy(object):\n id: int\n title: str\n employer: str\n min_salary: int = None\n max_salary: int = None\n\n def __init__(self, webelement: WebElement) -> None:\n \"\"\"\n Args:\n webelement (WebElement):\n \"\"\"\n super().__init__()\n\n title = webelement.find_element_by_css_selector('[data-qa=\"vacancy-serp__vacancy-title\"]').text\n assert title\n\n employer = webelement.find_element_by_css_selector('[data-qa=\"vacancy-serp__vacancy-employer\"]').text\n assert employer\n\n min_salary = None\n max_salary = None\n if webelement.find_elements_by_css_selector('[data-qa=\"vacancy-serp__vacancy-compensation\"]'):\n salary_str = webelement.find_element_by_css_selector('[data-qa=\"vacancy-serp__vacancy-compensation\"]').text\n salary_data = _resolve_salary_min_max(salary_str)\n min_salary = salary_data[0]\n max_salary = salary_data[1]\n vacancy_title = webelement.find_element_by_css_selector('[data-qa=\"vacancy-serp__vacancy-title\"]')\n\n self.id = re.findall('vacancy/(.*?)\\\\?query=', vacancy_title.get_attribute('href'))[0]\n self.title = vacancy_title.text\n self.employer = employer\n self.min_salary = min_salary\n self.max_salary = max_salary\n\n\ndef _resolve_salary_min_max(salary_str: str) -> tuple:\n \"\"\"\n Args:\n salary_str (str):\n \"\"\"\n min_salary = None\n max_salary = None\n\n if salary_str.count('от'):\n # указано только от\n min_salary = re.sub(\"\\D\", \"\", salary_str)\n\n if salary_str.count('-'):\n # Указано и 'от' и 'до'\n min_salary_str = salary_str.split('-')[0]\n max_salary_str = salary_str.split('-')[1]\n min_salary = re.sub(\"\\D\", \"\", min_salary_str)\n max_salary = re.sub(\"\\D\", \"\", max_salary_str)\n\n if salary_str.count('до'):\n # указано только 'до'\n max_salary = re.sub(\"\\D\", \"\", salary_str)\n\n return (min_salary, max_salary)\n","sub_path":"hh_pages/Vacancy.py","file_name":"Vacancy.py","file_ext":"py","file_size_in_byte":2257,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"93628231","text":"# -*- coding: utf-8 -*-\n\nfrom .types import Type, NoSuchMethod, ClassOrInterface\n\n\nclass Expression(object):\n \"\"\"\n AST for simple Java expressions. Note that this package deal only with compile-time types;\n this class does not actually _evaluate_ expressions.\n \"\"\"\n\n def static_type(self):\n \"\"\"\n Returns the compile-time type of this expression, i.e. the most specific type that describes\n all the possible values it could take on at runtime. Subclasses must implement this method.\n \"\"\"\n raise NotImplementedError(type(self).__name__ + \" must implement static_type()\")\n\n def check_types(self):\n \"\"\"\n Validates the structure of this expression, checking for any logical inconsistencies in the\n child nodes and the operation this expression applies to them.\n \"\"\"\n raise NotImplementedError(type(self).__name__ + \" must implement check_types()\")\n\n\nclass Variable(Expression):\n \"\"\" An expression that reads the value of a variable, e.g. `x` in the expression `x + 5`.\n \"\"\"\n def __init__(self, name, declared_type):\n self.name = name #: The name of the variable\n self.declared_type = declared_type #: The declared type of the variable (Type)\n \n def static_type(self):\n return self.declared_type\n \n def check_types(self):\n pass\n\n\nclass Literal(Expression):\n \"\"\" A literal value entered in the code, e.g. `5` in the expression `x + 5`.\n \"\"\"\n def __init__(self, value, type):\n self.value = value #: The literal value, as a string\n self.type = type #: The type of the literal (Type)\n\n def static_type(self):\n return self.type\n \n def check_types(self):\n pass\n\n\nclass NullLiteral(Literal):\n def __init__(self):\n super().__init__(\"null\", Type.null)\n \n def static_type(self):\n return Type.null\n \n\ndef typecheck_arguments_function_call(args, types):\n for (arg, type) in zip(args, types):\n if not arg.static_type().is_subtype_of(type):\n if arg.static_type() == Type.null and type.is_subtype_of(Type.object):\n continue\n else:\n return False\n return True\n\n\nclass MethodCall(Expression):\n \"\"\"\n A Java method invocation, i.e. `foo.bar(0, 1, 2)`.\n \"\"\"\n def __init__(self, receiver, method_name, *args):\n self.receiver = receiver\n self.receiver = receiver #: The object whose method we are calling (Expression)\n self.method_name = method_name #: The name of the method to call (String)\n self.args = args #: The method arguments (list of Expressions)\n \n def check_types(self):\n for arg in self.args:\n arg.check_types()\n static_type = self.receiver.static_type()\n if static_type == Type.null:\n raise NoSuchMethod(\"Cannot invoke method \" + self.method_name + \"() on \" + static_type.name)\n if static_type.is_subtype_of(Type.object):\n method = static_type.method_named(self.method_name)\n if len(self.args) != len(method.argument_types):\n raise JavaTypeError(\"Wrong number of arguments for {}.{}(): expected {}, got {}\".format(\n static_type.name, self.method_name, \n len(method.argument_types), len(self.args)\n ))\n if not typecheck_arguments_function_call(self.args, method.argument_types):\n expected_arguments = [expectedType.name for expectedType in method.argument_types]\n received_arguments = [receivedArg.static_type().name for receivedArg in self.args]\n raise JavaTypeError(\"{}.{}() expects arguments of type ({}), but got ({})\".format(\n static_type.name, self.method_name,\n \", \".join(expected_arguments),\n \", \".join(received_arguments)\n ))\n else:\n raise JavaTypeError(\"Type {} does not have methods\".format(static_type.name))\n \n def static_type(self):\n return self.receiver.static_type().method_named(self.method_name).return_type\n\n\n\nclass ConstructorCall(Expression):\n \"\"\"\n A Java object instantiation, i.e. `new Foo(0, 1, 2)`.\n \"\"\"\n def __init__(self, instantiated_type, *args):\n self.instantiated_type = instantiated_type #: The type to instantiate (Type)\n self.args = args #: Constructor arguments (list of Expressions)\n\n def check_types(self):\n for arg in self.args:\n arg.check_types()\n if not self.instantiated_type.is_subtype_of(Type.object):\n raise JavaTypeError(\"Type \" + self.instantiated_type.name + \" is not instantiable\")\n method = self.instantiated_type.constructor\n if len(self.args) != len(method.argument_types):\n raise JavaTypeError(\"Wrong number of arguments for {} constructor: expected {}, got {}\".format(\n self.instantiated_type.name, \n len(method.argument_types), len(self.args)\n ))\n if not typecheck_arguments_function_call(self.args, method.argument_types):\n expected_arguments = [expectedType.name for expectedType in method.argument_types]\n received_arguments = [receivedArg.static_type().name for receivedArg in self.args]\n raise JavaTypeError(\"{} constructor expects arguments of type ({}), but got ({})\"\n .format(self.instantiated_type.name, \", \".join(expected_arguments), \", \".join(received_arguments)))\n \n def static_type(self):\n return self.instantiated_type\n\nclass JavaTypeError(Exception):\n \"\"\" Indicates a compile-time type error in an expression.\n \"\"\"\n pass\n\n\ndef names(named_things):\n \"\"\" Helper for formatting pretty error messages\n \"\"\"\n return \"(\" + \", \".join([e.name for e in named_things]) + \")\"\n","sub_path":"java-type-checker/java_type_checker/expressions.py","file_name":"expressions.py","file_ext":"py","file_size_in_byte":5948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"568451007","text":"import numpy as np\nimport math\nimport scipy\n\n\ndef gaussian_curvature(data: np.ndarray) -> np.ndarray:\n \"\"\"\n Computes the gaussian curvature of a numpy ndarray.\n :param data: A numpy ndarray.\n :return: Gaussian curvature.\n \"\"\"\n Zy, Zx = np.gradient(data)\n Zxy, Zxx = np.gradient(Zx)\n Zyy, _ = np.gradient(Zy)\n K = (Zxx * Zyy - (Zxy ** 2)) / (1 + (Zx ** 2) + (Zy ** 2)) ** 2\n\n return K\n\n\ndef mean_curvature(data: np.ndarray) -> np.ndarray:\n \"\"\"\n Computes the mean curvature of a numpy ndarray.\n :param data: Points ensembled from the data manifold.\n :return: Mean curvature scalar.\n \"\"\"\n Zy, Zx = numpy.gradient(data)\n Zxy, Zxx = numpy.gradient(Zx)\n Zyy, _ = numpy.gradient(Zy)\n\n H = (Zx ** 2 + 1) * Zyy - 2 * Zx * Zy * Zxy + (Zy ** 2 + 1) * Zxx\n H = -H / (2 * (Zx ** 2 + Zy ** 2 + 1) ** (1.5))\n\n return H\n\n\ndef distcorr(X, Y):\n \"\"\"\n Compute the distance correlation function on two random variables X and Y.\n :param X: Attribute column X.\n :param Y: Attribute column Y.\n :return: Distance correlation.\n \"\"\"\n X = np.atleast_1d(X)\n Y = np.atleast_1d(Y)\n\n if np.prod(X.shape) == len(X):\n X = X[:, None]\n if np.prod(Y.shape) == len(Y):\n Y = Y[:, None]\n\n X = np.atleast_2d(X)\n Y = np.atleast_2d(Y)\n n = X.shape[0]\n\n if Y.shape[0] != X.shape[0]:\n raise ValueError(\"Number of samples must match\")\n\n a = squareform(pdist(X))\n b = squareform(pdist(Y))\n A = a - a.mean(axis=0)[None, :] - a.mean(axis=1)[:, None] + a.mean()\n B = b - b.mean(axis=0)[None, :] - b.mean(axis=1)[:, None] + b.mean()\n\n dcov2_xy = (A * B).sum() / float(n * n)\n dcov2_xx = (A * A).sum() / float(n * n)\n dcov2_yy = (B * B).sum() / float(n * n)\n dcor = np.sqrt(dcov2_xy) / np.sqrt(np.sqrt(dcov2_xx) * np.sqrt(dcov2_yy))\n\n return dcor\n","sub_path":"lib/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":1848,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"350119225","text":"def application(environ, start_response):\n\toutput = 'Hello world\\n
'\n\tif environ['REQUEST_METHOD'] == 'GET':\n\t\toutput += 'Get query: '\n\t\toutput += environ['QUERY_STRING']\n\telse:\n\t\toutput += 'Post query: '\n\t\ttry:\n\t\t\trequest_body_size = int(environ.get('CONTENT_LENGTH', 0))\n\t\texcept (ValueError):\n\t\t\trequest_body_size = 0\n\t\trequest_body = environ['wsgi.input'].read(request_body_size)\n\t\toutput += request_body\n\n\tstart_response('200 OK', [('Content-type', 'text/html')])\n\treturn [output]\n","sub_path":"hello.py","file_name":"hello.py","file_ext":"py","file_size_in_byte":498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"643033676","text":"db.define_table('product',\r\n\tField('product_order', 'integer', notnull=True, unique=True),\r\n\tField('name', 'string', notnull=True, unique=True),\r\n\tField('catch', 'text'),\r\n\tField('meisho', 'string'),\r\n\tField('abv', requires=IS_MATCH('[\\d\\.]+')),\r\n\tField('volume', requires=IS_MATCH('[\\d\\.]+')),\t\r\n\tField('brewery', 'string'),\r\n\tField('prefecture', 'string'),\r\n\tField('image', 'upload'),\r\n\tField('rice', 'string'),\r\n\tField('nihonshudo', 'string'),\r\n\tField('seimai_buai', 'string'),\r\n\tField('comment', 'text'),\r\n\tField('price', 'double'),\r\n\tField('stock', 'integer'),\r\n\tformat='%(name)s')\r\n\r\n\t\r\ndb.define_table('premise',\r\n Field('first_name', length=128, default=''), \r\n Field('last_name', length=128, default=''), \r\n\tField('purchased_by', 'reference auth_user'),\r\n\tField('premise_name', length=128, default=''),\t\r\n\tField('premise_address', length=128, default=''),\r\n\tField('premise_homepage', length=128, default=''),\r\n\tField('postcode', length=128),\r\n\tField('phone_number', requires=IS_MATCH('[\\d\\-\\(\\) ]+')),\r\n format='%(first_name)s %(last_name)s')\r\n\r\n\t\r\ndb.premise.first_name.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='First Name')\r\ndb.premise.last_name.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Last Name')\r\ndb.premise.premise_name.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Restaurant/Shop Name')\r\ndb.premise.premise_homepage.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Homepage(optional): e.g. http://yourcompany.com')\r\ndb.premise.premise_address.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Restaurant/Shop Address')\r\ndb.premise.postcode.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Postcode')\r\ndb.premise.phone_number.widget = lambda f,v: SQLFORM.widgets.string.widget(f, v, _placeholder='Phone Number')\r\n\r\ndb.premise.first_name.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty) \r\ndb.premise.last_name.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty)\r\ndb.premise.premise_name.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty) \r\ndb.premise.premise_address.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty)\r\ndb.premise.postcode.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty) \r\ndb.premise.postcode.requires = [IS_MATCH('^[a-zA-Z0-9_\\- ]{5,8}$',error_message='wrong postcode')]\r\ndb.premise.phone_number.requires = IS_NOT_EMPTY(error_message=auth.messages.is_empty) \r\n\r\ndb.define_table('sales_order',\r\n\tField('order_number', requires=IS_MATCH('\\d{14}'), readable=False, writable=False), #e.g. 20160425131401\r\n\tField('ordered_by', 'reference auth_user', readable=False, writable=False),\r\n\tField('order_status', 'string', notnull=True, requires = IS_IN_SET(['in cart', 'order processing', 'invoiced', 'dispatched', 'paid', 'accounted'], zero=T('-- choose one --'), error_message='Status must be chosen')), #in cart, order processing, invoiced, dispatched, paid, accounted\r\n\tField('order_time', 'datetime', readable=False, writable=False),\r\n\tField('order_time_s', 'string', readable=False, writable=False),\r\n\tField('dispatch_time', 'string', readable=False, writable=False),\r\n\tField('delivery_premise', 'reference premise', readable=False, writable=False),\r\n\tField('payment_method', 'string', readable=False, writable=False),\t\r\n\tField('postage', 'double', readable=False, writable=False),\t\r\n\tField('subtotal', 'double', readable=False, writable=False),\r\n\tField('total', 'double', readable=False, writable=False))\r\n\r\ndb.define_table('sales_item',\r\n\tField('sales_order', 'reference sales_order'),\r\n\tField('product', 'reference product'),\r\n\tField('sales_price', 'double'),\r\n\tField('case_quantity', 'integer'))\r\n\r\n\t\r\ndb.define_table('customer_category',\r\n\tField('name', 'string'),\r\n\tField('off_rate', 'double'))\t\r\n\r\n\t\r\nme = auth.user_id","sub_path":"original files/Python code/db_custom.py","file_name":"db_custom.py","file_ext":"py","file_size_in_byte":3871,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"5605661","text":"__author__ = 'jbjose'\n\nimport itertools\nimport sys\n\ndef find_min_path(call, edges):\n if call in edges:\n return [call]\n\n if call[0] not in [edge[0] for edge in edges]:\n return None # Call failed\n\n paths = []\n for edge in filter(lambda x: x[0]==call[0], edges):\n new_edges = edges[:]\n new_edges.remove(edge)\n new_edges.remove(edge[-1::-1])\n new_path = find_min_path( (edge[1],call[1]) , new_edges )\n if new_path is not None:\n paths += [ [edge]+ new_path ]\n\n if len(paths) == 0:\n return None\n return sorted(paths, key=lambda x: len(x))[0]\n\nif __name__ == \"__main__\":\n\n network_input = line = \"\"\n while line != \"\\n\":\n network_input += line\n line = sys.stdin.readline()\n call_input = line = \"\"\n while line != \"\\n\":\n call_input += line\n line = sys.stdin.readline()\n\n# network_input=\"\"\"A B 2\n# A C 2\n# B C 2\n# B D 2\n# C E 1\n# D E 2\n# D G 1\n# E F 2\n# F G 2\"\"\"\n# call_input=\"\"\"A G\n# A G\n# C E\n# G D\n# D E\n# A B\n# A D\"\"\"\n\n edges = list( itertools.chain( *[ [(x,y),(y,x)]*int(n) for x,y,n in [edge.split(' ') for edge in network_input.split(\"\\n\",) if edge != \"\" ]] ) )\n calls = [ tuple(call.split(' ')) for call in call_input.split(\"\\n\",) if call != \"\"]\n\n for call in calls:\n shortest_path = find_min_path(call, edges)\n if shortest_path is None:\n print(\"Call {} {} -- failed\".format(*call))\n else:\n print(\"Call {} {} -- placed {} {}\".format(call[0], call[1], \" \".join([path[0] for path in shortest_path]), shortest_path[-1][-1] ))\n for edge in shortest_path:\n edges.remove(edge)\n edges.remove(edge[-1::-1])\n","sub_path":"161.Hard.Network.py","file_name":"161.Hard.Network.py","file_ext":"py","file_size_in_byte":1728,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"315748225","text":"from abc import ABC,abstractmethod\nfrom bs4 import BeautifulSoup\nimport requests\nimport re\n\nclass ScrapTrend(ABC):\n @abstractmethod\n def fetch(self):\n pass\n\nclass Web1(ScrapTrend):\n url = \"https://ajkerdeal.com/en/category/gadgets\"\n\n def fetch(self):\n name = []\n price = []\n link = []\n img = []\n response = requests.get(self.url)\n\n data = response.text\n soup = BeautifulSoup(data, 'html.parser')\n\n products = soup.find_all('div', {'class': 'deal-info-container'})\n for p in products:\n name.append(p.find('span', {'class': 'deal-title-container'}).text)\n temp = p.find('span', {'class': 'deal-price-container'}).text\n temp = re.sub('\\W+', '', temp)\n price.append(temp)\n link.append(p.find('a').get('href'))\n img.append(p.find('img', {'class': 'deal_image'}).get('src'))\n\n return name,price,link,img\n\nclass Web2(ScrapTrend):\n url = \"https://www.startech.com.bd/gadget?sort=rating&order=DESC\"\n\n def fetch(self):\n name = []\n price = []\n link = []\n img = []\n response = requests.get(self.url)\n data = response.text\n soup = BeautifulSoup(data, 'html.parser')\n\n products = soup.find_all('div', {'class': 'product-thumb'})\n\n for p in products:\n name.append(p.find('h4', {'class': 'product-name'}).text)\n temp = p.find('div', {'class': 'price space-between'}).text\n temp = re.sub('\\W+', '', temp)\n price.append(temp)\n link.append(p.find('a').get('href'))\n img.append(p.find('img').get('src'))\n\n return name, price, link, img","sub_path":"getmygadget/trend/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1712,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"547670413","text":"# -*- coding:utf-8 -*-\n\n# django library\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponse, HttpResponseRedirect,HttpResponseBadRequest\nfrom django.shortcuts import render_to_response, get_object_or_404\nfrom django.template import RequestContext\nfrom django.contrib import messages\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth.hashers import make_password\nfrom django.contrib.auth import authenticate, login as auth_login ,logout as auth_logout\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.utils import simplejson\nfrom django.core.paginator import Paginator, InvalidPage, EmptyPage\nfrom django.contrib.auth.decorators import login_required\n\n\n# our own code\nfrom authority.models import Permission\nfrom role.models import Role\nfrom log.models import Log\nfrom utils.constants import permission_type_dict\nfrom authority.decorators import permission_required\n\n# 显示用户列表\n@permission_required('role_manage')\ndef index(request):\n roles = Role.objects.all()\n paginator = Paginator(roles, 10)\n currentPage = request.POST.get('pageNum', 1)\n try:\n pager = paginator.page(currentPage)\n except InvalidPage:\n pager = paginator.page(1)\n \n return render_to_response('role/index.html',{'role_list':pager}, context_instance=RequestContext(request))\n\n# 删除记录\n@permission_required('role_manage')\ndef delete(request, role_id):\n role = None\n try:\n role = Role.objects.get(id=int(role_id))\n except BaseException:\n return HttpResponse(simplejson.dumps({\"statusCode\":400, \"message\":u'此角色不存在!'}), mimetype='application/json')\n # 删除角色和人的关联关系\n role.users.clear()\n # 删除角色和权限的关联关系\n role.permissions.clear()\n # 删除此角色\n role.delete()\n # 日志\n Log(username=request.user.username,log_type=2,relate_id=role.id,content=\"execute delete role \" + role.name + \" success!\", level=1).save()\n return HttpResponse(simplejson.dumps({\"statusCode\":200,\"url\": \"/role/index\", \"message\":u'删除成功'}), mimetype='application/json')\n \n@permission_required('role_manage')\ndef add(request):\n pdict = {}\n for key in permission_type_dict:\n pdict[permission_type_dict[key]]=Permission.objects.filter(type=key).order_by('id')\n \n if request.POST:\n role_name = request.POST.get(\"role_name\")\n role_desc = request.POST.get(\"role_desc\")\n permission_id_list = request.POST.getlist(\"permission_id\")\n # 保存角色信息\n role = Role();\n role.name = role_name\n role.desc = role_desc\n role.save()\n # 保存角色和权限对应关系\n for pid in permission_id_list:\n role.permissions.add(pid)\n \n # 日志\n Log(username=request.user.username,log_type=2,relate_id=role.id,content=\"execute add role \" + role.name + \" success!\", level=1).save()\n return HttpResponse(simplejson.dumps({\"statusCode\":200,\"url\": \"/role/index\", \"message\":u'添加成功'}), mimetype='application/json')\n return render_to_response('role/add.html',{'pdict':pdict},context_instance=RequestContext(request))\n\n# 编辑\n@permission_required('role_manage')\ndef edit(request, role_id):\n pdict = {}\n for key in permission_type_dict:\n pdict[permission_type_dict[key]]=Permission.objects.filter(type=key).order_by('id')\n role = None\n try:\n role = Role.objects.get(id=int(role_id))\n except BaseException:\n return HttpResponse(simplejson.dumps({\"statusCode\":400, \"message\":u'此角色不存在!'}), mimetype='application/json')\n \n permission_id_list = [] \n for p in role.permissions.all():\n permission_id_list.append(p.id)\n \n if request.POST:\n role_name = request.POST.get(\"role_name\")\n role_desc = request.POST.get(\"role_desc\")\n permission_id_list = request.POST.getlist(\"permission_id\")\n #保存角色信息\n role.name = role_name\n role.desc = role_desc\n role.save()\n # 保存角色和权限对应关系\n role.permissions.clear()\n for pid in permission_id_list:\n role.permissions.add(pid)\n \n # 日志\n Log(username=request.user.username,log_type=2,relate_id=role.id,content=\"execute edit role \" + role.name + \" success!\", level=1).save()\n return HttpResponse(simplejson.dumps({\"statusCode\":200,\"url\": \"/role/index\", \"message\":u'编辑成功'}), mimetype='application/json') \n return render_to_response('role/edit.html', {\"pdict\":pdict,\"role\": role,\"permission_id_list\":permission_id_list},context_instance=RequestContext(request))\n \n \n ","sub_path":"role/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"577193591","text":"import csv\r\nimport GetSnomedID\r\nimport GetCUI_Info\r\nimport Authenticate\r\n\r\nfile = csv.reader(open('C:\\\\Users\\\\jpaixao\\\\Downloads\\\\CD_16_UMNSRS_Similarity_human_judgments.csv'), delimiter=',')\r\n\r\nconverted_file = open('C:\\\\Users\\\\jpaixao\\\\repos\\\\uts-rest-api\\\\UMLS-api\\\\UMNSRS_Similarity_human_judgments.csv', 'w', newline='') \r\n\r\nfor row in file:\r\n ref_values = row[0]\r\n\r\n try:\r\n ticket = Authenticate.request_ticket()\r\n response = GetCUI_Info.get_snomed_id(row[1], ticket)\r\n snomed_code0 = response['result'][0]['code']\r\n except:\r\n snomed_code0 = \"not found\"\r\n\r\n try:\r\n ticket = Authenticate.request_ticket()\r\n response = GetCUI_Info.get_snomed_id(row[2], ticket)\r\n snomed_code1 = response['result'][0]['code']\r\n except:\r\n snomed_code1 = \"not found\"\r\n\r\n spamwriter = csv.writer(converted_file, delimiter=',')\r\n spamwriter.writerow([ref_values, snomed_code0, snomed_code1])\r\n\r\n print(row[0], row[1], row[2])","sub_path":"UMLS-api/ConvertCUItoSNOMED.py","file_name":"ConvertCUItoSNOMED.py","file_ext":"py","file_size_in_byte":987,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"224436737","text":"import pandas as pd\nimport keras.layers as layers\nimport yaml\nimport numpy as np\nfrom keras import backend as K\nimport tensorflow as tf\ndef reshape(observation):\n if len(observation.shape)==2:\n return observation\n else:\n len_=len(observation.shape)\n for i in range(2-len_):\n observation=observation[...,np.newaxis]\n return observation\ndef construct_model_from_csv(model_file, input_placeholder):\n \"\"\"\n\n :param model_file: path of csv file containing info about the design of the model\n :param input_placeholder: this is used as input to the model\n :return x: the last layer (tensor) of the model\n \"\"\"\n model_df = pd.read_csv(model_file).fillna(0)\n model_df = model_df.astype({\"layer_type\": str, \"units\": int, \"strides\":int,\n \"padding\": str, \"kernel_size\": int, \"num_filters\": int})\n x=input_placeholder\n for i in range(model_df.shape[0]):\n layer_type=model_df.layer_type[i]\n assert layer_type in [\"conv\", \"conv1D\", \"fc\", \"flatten\", \"relu\",\"elu\", \"bn\", \"pooling\"], \"invalid layer type\"\n \n if layer_type == \"conv\":\n padding=str(model_df.padding[i])\n kernel_size = int(model_df.kernel_size[i])\n num_filters = int(model_df.num_filters[i])\n strides = int(model_df.strides[i])\n \n assert padding == \"same\" or padding == \"valid\"\n assert kernel_size > 0\n assert num_filters > 0\n assert strides > 0\n\n x = layers.Conv2D(num_filters, kernel_size=kernel_size, strides=strides, padding=padding)(x)\n\n if layer_type == \"conv1D\":\n padding = str(model_df.padding[i])\n kernel_size = int(model_df.kernel_size[i])\n num_filters = int(model_df.num_filters[i])\n strides = int(model_df.strides[i])\n\n assert padding == \"same\" or padding == \"valid\"\n assert kernel_size > 0\n assert num_filters > 0\n assert strides > 0\n\n x = layers.Conv1D(num_filters, kernel_size=kernel_size, strides=strides, padding=padding)(x)\n \n elif layer_type == \"fc\":\n hidden_units = int(model_df.units[i])\n assert hidden_units > 0\n x = layers.Dense(hidden_units)(x)\n \n elif layer_type == \"flatten\":\n x = layers.Flatten()(x)\n \n elif layer_type == \"relu\":\n x = layers.Activation(\"relu\")(x)\n \n elif layer_type == \"elu\":\n x = layers.Activation(\"elu\")(x)\n \n elif layer_type == \"bn\":\n x = layers.BatchNormalization()(x)\n \n elif layer_type == \"pooling\":\n padding = int(model_df.padding[i])\n kernel_size = int(model_df.kernel_size[i])\n strides = int(model_df.strides[i])\n \n assert padding == \"same\" or padding == \"valid\"\n assert kernel_size > 0\n assert strides > 0\n x = layers.MaxPooling2D(kernel_size, strides=strides, padding=padding)(x)\n \n return x\n\n\ndef softmax(logits):\n \"\"\"\"\n :param logits: Unnormalized log probabilities\n :return Normalized probabilities\n \"\"\"\n logits-=np.max(logits)\n return np.exp(logits)/np.sum(np.exp(logits))\n\n\ndef load_config(file):\n with open(file, 'r') as stream:\n dict_ = yaml.load(stream)\n return dict_\n\ndef sample(replay_buffer,steps):\n Data=[]\n y=[]\n a=[]\n for i in range(steps):\n point=replay_buffer.get()\n Data.append(point[0][0])\n y.append(point[1])\n a.append(point[2])\n return np.array(Data),np.array(y),np.array(a)\n\ndef share_gpu():\n config1 = tf.ConfigProto()\n config1.gpu_options.allow_growth = True\n sess = tf.Session(config=config1)\n K.set_session(sess)","sub_path":"utils/tools.py","file_name":"tools.py","file_ext":"py","file_size_in_byte":4106,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"244549513","text":"import os\n\nfrom .base import * # noqa\nfrom .base import BASE_DIR\n\nDEBUG = True\n\n# test db is sqlite3\nDATABASES = {\n \"default\": {\n \"ENGINE\": \"django.db.backends.sqlite3\",\n \"NAME\": os.path.join(BASE_DIR, \"{{cookiecutter.project_slug}}-test-sqlite3.db\"),\n },\n}\n","sub_path":"{{cookiecutter.project_slug}}/src/{{cookiecutter.project_slug}}/settings/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"558460110","text":"# Copyright (C) 2002-2020 CERN for the benefit of the ATLAS collaboration\n\nfrom __future__ import print_function\n\n\nimport ROOT\n\n# later on we could define this in ROOT\nTestFile=\"myAODTTv3.root\"\nOutputFile=\"test.output.root\"\nPoolAODInput=[TestFile]\n\n# set conditions\nDetDescrVersion=\"ATLAS-R2-2015-03-01-00\"\nConditionsTag=\"OFLCOND-RUN12-SDR-31\"\n# set everything\ninclude (\"RecExCond/RecExCommon_DetFlags.py\")\nDetFlags.detdescr.all_setOff()\nDetFlags.detdescr.Calo_setOn()\ninclude (\"RecExCond/AllDet_detDescr.py\")\n\n# Set the input file\nimport AthenaPoolCnvSvc.ReadAthenaPool\n\nsvcMgr.EventSelector.InputCollections = [TestFile]\n\n# import the algorithm in python and make sure it\n# is attached to the sequence\nfrom AthenaCommon.AlgSequence import AlgSequence\ntopSequence = AlgSequence()\n#from TrigT1CaloEFex.TrigT1CaloEFexConf import TrigT1CaloEFex\n#a=TrigT1CaloEFex()\n#topSequence+=a\nfrom TrigT1CaloEFex.EFexAlgo import EFexAlgo\ntopSequence+=EFexAlgo()\n\n# initialize athena\ntheApp.initialize()\n# main loop\nev=0\nfor i in range(0,10):\n try : \n a=not theApp.nextEvent().isFailure() \n print ('Event Number',ev)\n ev=ev+1;\n except : \n print (\"except\")\n break\ntheApp.finalize().ignore();\n\nprint ('writing out file after',ev,'events')\ntheApp.exit()\n\n","sub_path":"Trigger/TrigL1Upgrade/TrigL1CaloUpgrade/share/runEFex.py","file_name":"runEFex.py","file_ext":"py","file_size_in_byte":1242,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"451783519","text":"from django.shortcuts import render\n\n\n# Module import\nfrom .models import Tweet # model class imported\n\n# Method Link\nfrom django.http import HttpResponse, Http404, JsonResponse\n\n\n# Create your views here.\ndef home_view(request, *args, **kwargs):\n # return HttpResponse(\"

WELCOME TO TWEET MAIN

\")\n return render(request, \"pages/home.html\", context={}, status=200)\n\n\n# intro to dynamic ROUT URL\n# def dynamic_route_url_view(request, num, *args, **kwargs):\n# return HttpResponse(f\"

Hello {num}

\")\n\n\n# REST API VIEW\n# which means these datas can be consumed by javaScrip, swift, etc\n# return json data by using method - JsonResponse()\n\n\ndef tweet_detail_view(request, tweet_id, *args, **kwargs):\n data = {\"id\": tweet_id}\n\n try:\n obj = Tweet.objects.get(id=tweet_id)\n data[\"content\"] = obj.content\n status = 200\n except:\n data[\"message\"] = \"This id doesn't exist\"\n status = 404\n\n return JsonResponse(data, status=status)\n\n\ndef tweet_list_view(request, *args, **kwargs):\n qs = Tweet.objects.all()\n listTweet = [{\"id\": key.id, \"content\": key.content} for key in qs]\n data = {\"response\": listTweet}\n return JsonResponse(data)","sub_path":"cloneCodingTutorial/cloneTweet/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"140684173","text":"#This is aboute how to get enable and config mode in python.\n#modification are done only on config mode so if you want to modify the device configuration\n#so how to get it.\n#Author Samiullah\n#Email: sameullah553@gmail.com\n\nimport netmiko\nfrom netmiko import ConnectHandler\nfrom pprint import pprint\nimport genie\n\nR1 ={\n \"device_type\":\"cisco_ios\",\n \"ip\":\"x.x.x.x\",\n \"username\":\"admin\",\n \"password\":\"cisco12\",\n \"secret\":\"cisco12\"\n }\nnet_connect =ConnectHandler(**R1)\n#enable_mode just use enable() function.\nnet_connect.enable()\n#For configuration mode we use config_mode() function.\nnet_connect.config_mode()\n#Here now you can send Configuration commnad. use list for a bunch of commands \n#for example\ninterface_config =[\n \"interface fastethernet0/0\",\n \"ip address x.x.x.x/24\",\n \"no shutdown\"\n ]\n#To send a bunch of commands we will use \"send_config_set()\" function\nnet_connect.send_config_set(interface_config)\nnet_connect.disconnect()\n","sub_path":"enable_config.py","file_name":"enable_config.py","file_ext":"py","file_size_in_byte":1006,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"609089528","text":"class Solution:\n def subarraySum(self, nums: List[int], k: int) -> int:\n runningsum=0\n maxlen=0\n sub={0:1}\n #Edge case\n if(nums== None or len(nums)==0):\n return 0\n for i in range(len(nums)):\n #find the running sum for each iteration\n runningsum+=nums[i]\n comp=runningsum-k\n \n #finding max count\n #if the complimemt exists in dict increase maxvalue by 1\n if(comp in sub):\n maxlen=maxlen + sub[comp]\n \n #key-->running sum in dictionary value-->1 if it occurs for first time else increment by 2\n if(runningsum not in sub):\n sub[runningsum]=1\n else:\n sub[runningsum]+=1\n return maxlen\n \n \n \n \n ","sub_path":"SubarraySumK.py","file_name":"SubarraySumK.py","file_ext":"py","file_size_in_byte":875,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"29637481","text":"from flask_restful import Resource, reqparse\r\nfrom flask_jwt import jwt_required\r\nfrom models.etudiant import EtudiantModel\r\n\r\nclass Etudiant(Resource):\r\n parser = reqparse.RequestParser()\r\n parser.add_argument('classe',\r\n type=str,\r\n required=True,\r\n help=\"This field has to be filled\"\r\n )\r\n parser.add_argument('ecole_id',\r\n type=int,\r\n required=True,\r\n help=\"Référence à la classe nécessaire\"\r\n )\r\n\r\n @jwt_required()\r\n def get(self, name):\r\n etudiant = EtudiantModel.find_by_name(name)\r\n if etudiant:\r\n return etudiant.json()\r\n return {\"message\": \"Etudiant non trouvé\"},404\r\n\r\n @jwt_required()\r\n def post(self, name):\r\n\r\n if EtudiantModel.find_by_name(name):\r\n return {\"message\":f\"Un étudiant avec le même nom existe {name}\"},400\r\n\r\n data = Etudiant.parser.parse_args()\r\n etudiant = EtudiantModel(name, **data)\r\n\r\n try:\r\n etudiant.save_to_db()\r\n except:\r\n return {\"message\":\"Une erreur est apparue à l'insertion\"}, 500\r\n\r\n return etudiant.json(), 201\r\n\r\n @jwt_required()\r\n def delete(self, name):\r\n etudiant = EtudiantModel.find_by_name(name)\r\n if etudiant:\r\n etudiant.delete_from_db()\r\n return {\"message\": \"Etudiant supprimé\"}\r\n\r\n @jwt_required()\r\n def put(self, name):\r\n\r\n data = Etudiant.parser.parse_args()\r\n etudiant = EtudiantModel.find_by_name(name)\r\n\r\n if etudiant is None:\r\n etudiant = EtudiantModel(name, **data)\r\n else:\r\n etudiant.classe = data[\"classe\"]\r\n etudiant.ecole_id = data[\"ecole_id\"]\r\n etudiant.save_to_db()\r\n\r\n return etudiant.json()\r\n\r\nclass ListeEtudiant(Resource):\r\n def get(self):\r\n return {\"etudiants\" : [x.json() for x in EtudiantModel.find_all()]}","sub_path":"ressources/etudiant.py","file_name":"etudiant.py","file_ext":"py","file_size_in_byte":2029,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"381623003","text":"from typing import Callable, List, TypeVar, overload\n\nfrom strawberry.federation.schema_directives import Key\nfrom strawberry.field import StrawberryField, field as base_field\nfrom strawberry.object_type import type as base_type\nfrom strawberry.utils.typing import __dataclass_transform__\n\nfrom .field import field\n\n\nT = TypeVar(\"T\")\n\n\n@overload\n@__dataclass_transform__(\n order_default=True, field_descriptors=(base_field, field, StrawberryField)\n)\ndef type(\n cls: T,\n *,\n name: str = None,\n description: str = None,\n keys: List[str] = None,\n extend: bool = False,\n) -> T:\n ...\n\n\n@overload\n@__dataclass_transform__(\n order_default=True, field_descriptors=(base_field, field, StrawberryField)\n)\ndef type(\n *,\n name: str = None,\n description: str = None,\n keys: List[str] = None,\n extend: bool = False,\n) -> Callable[[T], T]:\n ...\n\n\ndef type(\n cls=None,\n *,\n name=None,\n description=None,\n keys=None,\n extend=False,\n):\n directives = [Key(key) for key in keys or []]\n\n return base_type(\n cls,\n name=name,\n description=description,\n directives=directives, # type: ignore\n extend=extend,\n )\n","sub_path":"strawberry/federation/object_type.py","file_name":"object_type.py","file_ext":"py","file_size_in_byte":1195,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"322594009","text":"import pandas as pd\nfrom sklearn.cluster import KMeans\nimport KMeansTestCluster as kmtc\nimport EMTestCluster as emtc\nfrom functions import *\n\"\"\"\n@author: vincentmarois\nThis python script applies the clustering algorithms on the Letter dataset.\n\"\"\"\n\n# import dataset\nletters = pd.read_csv('letters.csv')\n\n# separate dataset into features & labels\nX = (letters.drop('lettr', axis=1)).as_matrix()\ny = letters['lettr'].as_matrix()\n\n# consider the following ranges of clusters\nclusters = range(2, 41)\n\n# search the optimal value of k for Kmeans, while considering several metrics\n# (WSS, BSS/TSS ratio, Silhouette, V-measure)\nKMeanstester = kmtc.KMeansTestCluster(X, y, clusters=clusters, plot=True, stats=True, name=\"letters\")\n_, _, Kmeans_silhouette, Kmeans_V_measure = KMeanstester.run()\n\n# search the optimal value of k for EM, while considering several metrics\n# (log-likelihood, Bayesian Information Criterion, Silhouette, V-measure)\nEMtester = emtc.EMTestCluster(X, y, clusters=clusters, plot=True, stats=True, name=\"letters\")\n_, _, EM_silhouette, EM_V_measure = EMtester.run()\n\n# Plot the Silhouette Score & V-measure metrics on same figure to compare both algorithms\nplot_silhouette_v_measure(algorithm=\"Kmeans\", clusters=clusters, silhouette=Kmeans_silhouette, v_measure=Kmeans_V_measure)\nplot_silhouette_v_measure(algorithm=\"EM\", clusters=clusters, silhouette=EM_silhouette, v_measure=EM_V_measure)\n\n# Compare the clusters found by Kmeans with the original class labels, allows to conclude on homogeneity & completeness\nn_clusters = 12\nestimator = KMeans(n_clusters=n_clusters, init='k-means++', max_iter=500, n_init=10)\nplot_clusters_labels(estimator=estimator, n_clusters=n_clusters, X=X, y=y, n_labels=26, algorithm=\"Kmeans\")\n","sub_path":"CS7641/Unsupervised Learning and Dimensionality Reduction/letters-dataset/letters_clustering.py","file_name":"letters_clustering.py","file_ext":"py","file_size_in_byte":1735,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"77784521","text":"from typing import Any, List\n\nfrom pytorch_lightning import LightningModule\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchmetrics.classification.accuracy import Accuracy\nfrom transformers.utils.dummy_pt_objects import ElectraModel\nfrom src.models.retriever_model import DualEncoder\nfrom transformers import AdamW, get_cosine_schedule_with_warmup\nfrom src.metrics.aggregation import mean, precision_at_k\n\nclass Retriever(LightningModule):\n def __init__(\n self, \n lr: float = 5e-05, \n weight_decay: float = 0.0005, \n num_negatives: int = 111095,\n encoder_momentum: float = 0.999,\n softmax_temperature: float = 0.25,\n num_workers: int = 4,\n batch_size: int = 16,\n max_epochs: int = 2,\n warmup_ratio: float = 0.1,\n **kwargs\n ):\n super().__init__()\n self.save_hyperparameters()\n \n self.encoder_q, self.encoder_c = self._load_pretrained_model()\n\n # for param_q, param_k in zip(self.encoder_q.parameters(), self.encoder_k.parameters()):\n # param_k.data.copy_(param_q.data)\n # param_k.requires_grad = False\n\n for name, param in self.encoder_c.named_parameters():\n param.requires_grad = False\n\n self.register_buffer('queue', torch.zeros(256, 112620))\n self.register_buffer('queue_ptr', torch.zeros(1, dtype=torch.long))\n \n self.criterion = nn.CrossEntropyLoss()\n\n self.train_accuracy = Accuracy()\n self.val_accuracy = Accuracy()\n\n self.metric_hist = {\n \"train/acc\": [],\n \"val/acc\": [],\n \"train/loss\": [],\n \"val/loss\": [],\n }\n \n def _load_pretrained_model(self):\n encoder = DualEncoder.load_from_checkpoint('/opt/ml/code/pytorch_lightning_examples/3_RETRIEVER/logs/runs/2021-05-18/09-21-14/checkpoints/epoch=18.ckpt')\n encoder_q = encoder.q_encoder\n encoder_c = encoder.c_encoder\n \n return encoder_q, encoder_c\n\n @torch.no_grad()\n def _momentum_update_key_encoder(self):\n for param_q, param_k in zip(self.encoder_q.parameters(), self.encoder_k.parameters()):\n em = self.hparams.encoder_momentum\n param_k.data = param_k.data * em + param_q.data * (1. - em)\n \n @torch.no_grad()\n def _dequeue_and_enqueue(self, keys):\n print('queue change')\n batch_size = keys.shape[0]\n\n ptr = int(self.queue_ptr)\n assert self.hparams.num_negatives % batch_size == 0 # for simplicity\n\n # replace the keys at ptr (dequeue and enqueue)\n self.queue[:, ptr:ptr + batch_size] = keys.T\n ptr = (ptr + batch_size) % self.hparams.num_negatives # move pointer\n\n self.queue_ptr[0] = ptr\n\n def forward(self, inputs):\n q_inputs = {\n \"input_ids\": inputs[0],\n \"attention_mask\": inputs[1],\n \"token_type_ids\": inputs[2],\n }\n c_inputs = {\n \"input_ids\": inputs[3],\n \"attention_mask\": inputs[4],\n \"token_type_ids\": inputs[5],\n }\n q = self.encoder_q(**q_inputs)\n\n with torch.no_grad():\n k = self.encoder_c(**c_inputs)\n\n l_pos = torch.einsum('nc,nc->n', [q, k]).unsqueeze(-1)\n l_neg = torch.einsum('nc,ck->nk', [q, self.queue.clone().detach()])\n\n logits = torch.cat([l_pos, l_neg], dim=1)\n logits /= self.hparams.softmax_temperature\n\n labels = torch.zeros(logits.shape[0], dtype=torch.long)\n labels = labels.type_as(logits)\n\n # self._dequeue_and_enqueue(k)\n return logits, labels\n\n def training_step(self, batch, batch_idx):\n # self._momentum_update_key_encoder()\n logits, labels = self(batch)\n\n loss = self.criterion(logits.float(), labels.long())\n\n acc1, acc5 = precision_at_k(logits, labels, top_k=(1, 5))\n\n log = {'train_loss': loss, 'train_acc1': acc1, 'train_acc5': acc5}\n self.log_dict(log)\n return loss\n\n\n def validation_step(self, batch, batch_idx):\n output, target = self(batch)\n loss = F.cross_entropy(output, target.long())\n\n acc1, acc5 = precision_at_k(output, target, top_k=(1, 5))\n\n results = {'val_loss': loss, 'val_acc1': acc1, 'val_acc5': acc5}\n return results\n\n def validation_epoch_end(self, outputs):\n val_loss = mean(outputs, 'val_loss')\n val_acc1 = mean(outputs, 'val_acc1')\n val_acc5 = mean(outputs, 'val_acc5')\n\n log = {'val_loss': val_loss, 'val_acc1': val_acc1, 'val_acc5': val_acc5}\n self.log_dict(log)\n\n def configure_optimizers(self):\n param_optimzier = list(self.encoder_q.named_parameters())\n no_decay = [\"bias\", \"LayerNorm.bias\", \"LayerNorm.weight\"]\n optimizer_grouped_parameters = [\n {\n \"params\": [\n p for n, p in param_optimzier if not any(nd in n for nd in no_decay)\n ],\n \"weight_decay\": 0.01,\n },\n {\n \"params\": [\n p for n, p in param_optimzier if any(nd in n for nd in no_decay)\n ],\n \"weight_decay\": 0.0,\n },\n ]\n optimizer = AdamW(\n optimizer_grouped_parameters, lr=self.hparams.lr, correct_bias=False\n )\n\n num_workers = self.hparams.num_workers\n data_len = len(self.train_dataloader().dataset)\n num_training_steps = int(\n data_len / (self.hparams.batch_size * num_workers) * self.hparams.max_epochs\n )\n num_warmup_steps = int(num_training_steps * self.hparams.warmup_ratio)\n scheduler = get_cosine_schedule_with_warmup(\n optimizer,\n num_warmup_steps=num_warmup_steps,\n num_training_steps=num_training_steps,\n )\n lr_scheduler = {\n \"scheduler\": scheduler,\n \"monitor\": \"loss\",\n \"interval\": \"step\",\n \"frequency\": 1,\n }\n return [optimizer], [lr_scheduler]\n \n \n","sub_path":"4_rag/src/models/retriever_moco_model.py","file_name":"retriever_moco_model.py","file_ext":"py","file_size_in_byte":6075,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"412418572","text":"import argparse\nimport asyncio\nimport logging\nimport sys\nimport threading\n\nfrom aiohttp import web\n\nfrom block_server_api.route_handler import RouteHandler\nfrom zmq.asyncio import ZMQEventLoop\nfrom block_server_api.databaseImp import DatabaseImp\n\nLOGGER = logging.getLogger(__name__)\n\n\ndef parse_args(args):\n parser = argparse.ArgumentParser(add_help=False)\n\n parser = argparse.ArgumentParser(add_help=False)\n parser.add_argument(\n '-v', '--verbose',\n action='count',\n default=0,\n help='Increase output sent to stderr')\n parser.add_argument(\n '--uri',\n type=str,\n help='database URI',\n default='mongodb://127.0.0.1:27017/')\n\n parser.add_argument(\n '-b', '--bind',\n help='identify host and port for api to run on',\n default='block-server-rest-api:9001')\n\n\n return parser.parse_args(args)\n\n\ndef init_logger(level):\n logger = logging.getLogger()\n logger.addHandler(logging.StreamHandler())\n if level == 1:\n logger.setLevel(logging.INFO)\n elif level > 1:\n logger.setLevel(logging.DEBUG)\n else:\n logger.setLevel(logging.WARN)\n\n\n\ndef start_rest_api(host, port, opts, loop):\n # start REST API\n app = web.Application(loop=loop)\n handler = RouteHandler()\n\n app.router.add_get('/height', handler.get_height)\n app.router.add_get('/block', handler.get_block_transactions)\n\n LOGGER.warning('Starting REST API on %s:%s', host, port)\n web.run_app(\n app,\n host=host,\n port=port,\n access_log=LOGGER)\n\n\ndef main():\n LOGGER.warning(\"## in api ##\")\n opts = parse_args(sys.argv[1:])\n init_logger(opts.verbose)\n try:\n host, port = opts.bind.split(\":\")\n port = int(port)\n except ValueError:\n print(\"Unable to parse binding {}: Must be in the format\"\n \" host:port\".format(opts.bind))\n sys.exit(1)\n loop = ZMQEventLoop()\n asyncio.set_event_loop(loop)\n try:\n DatabaseImp.initialize(opts.uri)\n start_rest_api(host, port, opts, loop)\n\n\n except KeyboardInterrupt:\n pass\n finally:\n print(\"Closing Loop\")\n loop.close()\n\n\nmain()\n","sub_path":"networks/sawtooth_v1_2/block_server_api/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2183,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"605673202","text":"# Copyright (c) 2016 Rackspace, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport random\n\nfrom poppy.model import ssl_certificate\nfrom poppy.storage import base\n\n\nclass CertificatesController(base.CertificatesController):\n\n def __init__(self, driver):\n super(CertificatesController, self).__init__(driver)\n\n self.certs = {}\n\n def create_certificate(self, project_id, cert_obj):\n key = (cert_obj.flavor_id, cert_obj.domain_name, cert_obj.cert_type)\n if key not in self.certs:\n self.certs[key] = cert_obj\n else:\n raise ValueError\n\n def delete_certificate(self, project_id, domain_name, cert_type):\n if \"non_exist\" in domain_name:\n raise ValueError(\"No certs on this domain\")\n\n def update_certificate(self, domain_name, cert_type, flavor_id,\n cert_details):\n key = (flavor_id, domain_name, cert_type)\n if key in self.certs:\n self.certs[key].cert_details = cert_details\n\n def get_certs_by_domain(self, domain_name, project_id=None,\n flavor_id=None,\n cert_type=None,\n status=u'create_in_progress'):\n certs = []\n for cert in self.certs:\n if domain_name in cert:\n certs.append(self.certs[cert])\n if project_id:\n if flavor_id is not None and cert_type is not None:\n return ssl_certificate.SSLCertificate(\n \"premium\",\n \"blog.testabcd.com\",\n \"san\",\n project_id=project_id,\n cert_details={\n 'Akamai': {\n u'cert_domain': u'secure2.san1.test_123.com',\n u'extra_info': {\n u'action': u'Waiting for customer domain '\n 'validation for blog.testabc.com',\n u'akamai_spsId': str(random.randint(1, 100000)\n ),\n u'create_at': u'2015-09-29 16:09:12.429147',\n u'san cert': u'secure2.san1.test_123.com',\n u'status': status}\n }\n }\n )\n return [cert for cert in certs if cert.project_id == project_id]\n else:\n if len(certs) == 1:\n return certs[0]\n else:\n raise ValueError(\"No matching certificates found for \"\n \"the domain {}\".format(domain_name))\n","sub_path":"poppy/storage/mockdb/certificates.py","file_name":"certificates.py","file_ext":"py","file_size_in_byte":3205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"41012621","text":"#:coding=utf-8:\nfrom django.core.exceptions import ImproperlyConfigured\nfrom functools import lru_cache\n\nfrom django.utils.module_loading import import_string\n\nfrom newauth.api import _get_backend_data\n\n__all__ = (\n 'User',\n 'AnonymousUser',\n 'get_user_model',\n 'get_anonymous_user_model',\n)\n\n\n@lru_cache()\ndef _get_user_models(model_name=None):\n if model_name is None:\n model_name = 'default'\n\n backend_data = _get_backend_data()\n\n try:\n model_path = backend_data[model_name]['user']\n anon_model_path = backend_data[model_name]['anon_user']\n except IndexError:\n if model_name is 'default':\n raise ImproperlyConfigured('A \"default\" user model is not specified. You must specify a \"default\" user model. Or maybe NEWAUTH_USER_MODELS isn\\'t a correctly defined dict?')\n else:\n raise ImproperlyConfigured('Error importing User model class with name \"%s\". Is NEWAUTH_USER_MODELS a correctly defined dict?' % model_name)\n\n try:\n UserCls = import_string(model_path)\n except ImportError:\n raise ImproperlyConfigured('Error importing user model class: \"%s\"' % model_path)\n except AttributeError:\n raise ImproperlyConfigured('Module does not define a class \"%s\"' % model_path)\n\n try:\n AnonUserCls = import_string(anon_model_path)\n except ImportError:\n raise ImproperlyConfigured('Error importing Anonymous user model class: \"%s\"' % anon_model_path)\n except AttributeError:\n raise ImproperlyConfigured('Module does not define a class \"%s\"' % anon_model_path)\n return UserCls, AnonUserCls\n\n\ndef get_user_model(model_name=None):\n \"\"\"\n Used to get access to defined user models.\n\n from newauth.models import User, get_user_model\n\n OtherUser = get_user_model('other')\n\n class MyProfile(models.Model):\n user = models.ForeignKey(User)\n other_user_type = models.ForeignKey(OtherUser)\n \"\"\"\n return _get_user_models(model_name)[0]\n\ndef get_anonymous_user_model(model_name=None):\n return _get_user_models(model_name)[1]\n\ntry:\n User = get_user_model('default')\n AnonymousUser = get_anonymous_user_model('default')\nexcept (IndexError, KeyError):\n raise ImproperlyConfigured('A \"default\" user model is not specified. You must specify a \"default\" user model. Or maybe NEWAUTH_USER_MODELS isn\\'t a correctly defined dict?')\n","sub_path":"newauth/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":2389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"181702455","text":"\nimport pandas as pd\nfrom sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer\nfrom sklearn.metrics.pairwise import linear_kernel, cosine_similarity\nimport warnings; warnings.simplefilter('ignore')\nimport random\nimport traceback\n\n\ndef weighted_rating(row, m, C):\n vtct = row['vote_count']\n avg = row['vote_avg']\n # Calculation based on the IMDB formula\n return (vtct / (vtct + m) * avg) + (m / (m + vtct) * C)\n\nclass Recommendation:\n\n def __init__(self, db):\n self.titles = pd.DataFrame()\n self.indices = pd.DataFrame()\n self.cosine_sim = None\n self.db = db\n self.prepareContentBasedRecomm()\n self.md = pd.DataFrame()\n\n def prepareContentBasedRecomm(self):\n try:\n # md = pd.read_csv('data/movies_metadata.csv')\n query = \"select * from movies limit 5000;\"\n self.md = pd.read_sql(query, self.db)\n print(self.md.shape)\n self.md['overview'] = self.md['overview'].fillna('')\n tf = TfidfVectorizer(analyzer='word', ngram_range=(1, 2), min_df=0, stop_words='english')\n tfidf_matrix = tf.fit_transform(self.md['overview'])\n tfidf_matrix.shape\n self.cosine_sim = linear_kernel(tfidf_matrix, tfidf_matrix)\n print(self.cosine_sim[0])\n # md = md.reset_index()\n print(self.md.head(5))\n # md.set_index('id')\n print(self.md.head(5))\n self.titles = self.md['title']\n self.indices = pd.Series(self.md.index, index=self.md['title'])\n print(self.indices.head(5))\n except Exception as e:\n print(\"DB emoty\")\n return pd.DataFrame()\n\n def getContentBasedRecomm(self, title):\n try:\n print(\"In get content based recomm\" + title)\n idx = self.indices[title]\n sim_scores = list(enumerate(self.cosine_sim[idx]))\n sim_scores = sorted(sim_scores, key=lambda x: x[1], reverse=True)\n sim_scores = sim_scores[1:31]\n indices = [i[0] for i in sim_scores]\n recommendedMovies = self.md.iloc[indices]\n recommendedMovies = recommendedMovies[['id', 'title']]\n print(recommendedMovies.head(5))\n # recommendedMovies = recommendedMovies.join(self.md, )\n return recommendedMovies\n except Exception as e:\n traceback.print_exc()\n return pd.DataFrame()\n\n # Simple Recommendation\n def getTrendingRecommendations(self):\n try:\n # md = 155514\n # pd.read_csv('data/movies_metadata.csv')\n # query = \"select * from movies;\"\n # md = pd.read_sql(query, self.db)\n # md.head()\n print(\"GEtting trending recomm\")\n vote_counts = self.md[self.md['vote_count'].notnull()]['vote_count'].astype('int')\n vote_avgs = self.md[self.md['vote_avg'].notnull()]['vote_avg'].astype('int')\n C = vote_avgs.mean()\n random_quantile = random.randint(79, 99) / 100\n print(random_quantile)\n m = vote_counts.quantile(random_quantile)\n qualified = \\\n self.md[(self.md['vote_count'] >= m) & (self.md['vote_count'].notnull()) & (self.md['vote_avg'].notnull())][\n ['id', 'title', 'releasedate', 'vote_count', 'vote_avg']]\n qualified['vote_count'] = qualified['vote_count'].astype('int')\n qualified['vote_avg'] = qualified['vote_avg'].astype('int')\n qualified.shape\n qualified['wr'] = qualified.apply(weighted_rating, args=(m, C), axis=1)\n qualified = qualified.sort_values('wr', ascending=False).head(250)\n trending = qualified.head(20)\n trending = trending.values.tolist()\n # print(md.head())\n return trending\n except Exception as e:\n traceback.print_exc()\n return pd.DataFrame()\n\n def getMoviesByGenre(self, genreid, cursor):\n try:\n cursor.execute(f\"select movieid from movie_genres where genreid='{genreid}' limit 30;\")\n movieids = cursor.fetchall()\n # print(movieids)\n allmovieids = list()\n for (i,) in movieids:\n allmovieids.append(i)\n allmovieids = tuple(allmovieids)\n # print(allmovieids)\n moviesquery = f\"select id, title from movies where id in {allmovieids} ORDER BY vote_avg DESC limit 30;\"\n df = pd.read_sql(moviesquery, self.db)\n return df.values.tolist()\n except Exception as e:\n traceback.print_exc()\n return pd.DataFrame()","sub_path":"utils/recommendations.py","file_name":"recommendations.py","file_ext":"py","file_size_in_byte":4690,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"584940199","text":"\"\"\"DatumInt.py\nImplementation of Datum (see Datum.py) for NuShellX interactions\n\"\"\"\nfrom __future__ import print_function, division, unicode_literals\n\nfrom os import path, mkdir, link\n\nfrom deprecated.int.QuantumNumbers import QuantumNumbers\nfrom deprecated.int.TwoBodyInteraction import TwoBodyInteraction\n\nfrom constants import DPATH_FILES_INT_ORG, ORG_FMT_INT_DNAME, ORG_FMT_INT_FNAME\nfrom deprecated.Datum import Datum\nfrom deprecated.int.parser import index_to_qnums_map as get_index_tuple_map\nfrom deprecated.int.parser import mass_number_from_filename as mass_from_filename\nfrom deprecated.int.parser import mass_to_index_to_energy_map as get_mie_map\nfrom deprecated.int.parser import mass_to_tbint_to_energy_map as get_miie_map\nfrom deprecated.int.parser import mass_to_zbt_map\nfrom deprecated.int.parser import name_from_filename\nfrom deprecated.int.parser import other_constants_from_filename as oc_from_filename\n\n\nclass DatumInt(Datum):\n \"\"\"Stores maps generated from *.int files and methods for generating new\n maps from this data\n \"\"\"\n def __init__(\n self, directory, exp, files, std_io_map=None,\n standardize_io_map=True, organize_files=True,\n dpath_org_files=DPATH_FILES_INT_ORG,\n dname_fmt_org=ORG_FMT_INT_DNAME,\n fname_fmt_org=ORG_FMT_INT_FNAME\n ):\n \"\"\"Initialize a particular NuShellX datum, where \"datum\" refers to\n all of the data in directory that matches the given exp\n (see ExpInt.py). In general, this will be initialized by a DataMap.\n :param directory: parent directory\n :param exp: unique identifier for the data held by this type\n :param files: list of file paths of all of the relevant files\n :param std_io_map: standard index -> orbital map to use. This fixes\n the issue of differing conventions amongst different files\n :param standardize_io_map: if true, all of the maps in this instance\n will be standardized according to the provided std_io_map\n :param organize_files: if true, all files are rewritten into\n org_file_dir according to a consistent naming scheme from which the\n ExpInt is easily read\n :param dpath_org_files: directory in which to store files that have been\n renamed\n :param dname_fmt_org: directory name to be formatted with exp\n :param fname_fmt_org: new interaction file name to be formatted with\n exp and the mass number\n \"\"\"\n super(DatumInt, self).__init__(\n directory=directory, exp=exp, files=files)\n self.name = None\n self.standardized_indexing = False\n self.files_organized = False\n # Create maps initially empty\n self.standard_index_orbital_map = std_io_map\n self._particular_index_orbital_map = None\n self._index_orbital_map = dict()\n self._mass_index_spe_map = dict()\n self._mass_interaction_index_energy_map = dict()\n self._mass_zero_body_term_map = dict()\n self._other_constants = None\n self._unorg_files = None\n # Perform setup methods\n self._set_maps()\n self._set_name()\n self._set_other_constants()\n if self.standard_index_orbital_map is not None and standardize_io_map:\n self._standardize_indexing()\n self.standardized_indexing = True\n if organize_files:\n self._organize_files(dpath_org_files, dname_fmt_org, fname_fmt_org)\n\n def _set_maps(self):\n self._set_index_orbital_map()\n self._set_mass_index_energy_map()\n self._set_mass_interaction_index_energy_map()\n self._set_zero_body_term_map()\n\n def _set_index_orbital_map(self):\n \"\"\"Retrieves the index -> orbital map from a file in the directory\n and stores it in an instance variable\n \"\"\"\n # Assuming all files characteristic have the same indexing...\n index_orbital_map = get_index_tuple_map(self.files[0])\n # Turn each tuple in the map into a named tuple\n for k in index_orbital_map.keys():\n v = index_orbital_map[k]\n nextv = QuantumNumbers(*_qnums_to_list(v))\n index_orbital_map[k] = nextv\n self._index_orbital_map = index_orbital_map\n\n def _set_mass_index_energy_map(self):\n \"\"\"Retrieves the\n mass number -> orbital index -> energy\n mapping for the directory\n \"\"\"\n self._mass_index_spe_map = (\n get_mie_map(self.dir, fpath_list=self.files))\n\n def _set_mass_interaction_index_energy_map(self):\n \"\"\"Retrieves the\n mass number -> (a, b, c, d, j) -> energy\n mapping for the directory\n \"\"\"\n miiem = (get_miie_map(self.dir, fpath_list=self.files))\n # Turn each tuple into a named tuple\n for A in miiem.keys():\n tuple_energy_map = miiem[A]\n next_tuple_energy_map = dict()\n for k in tuple_energy_map.keys():\n v = tuple_energy_map[k]\n nextk = list()\n for stritem in k:\n nextk.append(int(stritem))\n nextk = TwoBodyInteraction(*nextk)\n next_tuple_energy_map[nextk] = v\n miiem[A] = next_tuple_energy_map\n self._mass_interaction_index_energy_map = miiem\n\n def _set_zero_body_term_map(self):\n self._mass_zero_body_term_map = (\n mass_to_zbt_map(self.dir, fpath_list=self.files))\n\n def _set_name(self):\n \"\"\"Sets the incidence name variable\n \"\"\"\n self.name = name_from_filename(self.files[0])\n\n def _set_other_constants(self):\n \"\"\"Sets other heading constants. Assumes all files in a given directory\n have the same constants.\n I do not know what these are, hence the name \"other constants.\"\n They are the values that follow the single particle energies on\n the first non-comment line in the interaction files.\n \"\"\"\n self._other_constants = oc_from_filename(self.files[0])\n\n def _organize_files(self, directory, dir_fmt, file_fmt):\n \"\"\"Give the files standardized names and put them in a similarly-named\n directory\n :param dir_fmt: the string template for the directory name, should\n allow for the same number of arguments as the length of self.exp\n :param file_fmt: the string template for the file name. This should\n allow for the same number of arguments as the length of self.exp +1 for\n the mass number\n \"\"\"\n next_files = list()\n arg_list = ([self.name] +\n [str(i) if i is not None else '' for i in self.exp])\n d = path.join(directory, dir_fmt.format(*arg_list))\n if not path.exists(d):\n mkdir(d)\n for f in self.files:\n mass_num = mass_from_filename(f)\n new_f = path.join(d, file_fmt.format(*(arg_list + [mass_num])))\n next_files.append(new_f)\n if not path.exists(new_f):\n link(f, new_f)\n self._unorg_files, self.files = self.files, next_files\n\n def _standardize_indexing(self):\n self._standardize_mass_index_energy_map_indexing()\n self._standardize_mass_interaction_index_energy_map_indexing()\n self._particular_index_orbital_map = self._index_orbital_map\n self._index_orbital_map = self.standard_index_orbital_map\n\n def _standardize_mass_index_energy_map_indexing(self):\n \"\"\"Reformat the mass -> index -> energy map indices to be with respect\n to the standard io_map\n \"\"\"\n mie_map = self._mass_index_spe_map\n std_mie_map = dict()\n for m, ie_map in mie_map.items():\n std_ie_map = dict()\n for idx, energy in ie_map.items():\n next_idx = self._standard_index(idx)\n std_ie_map[next_idx] = energy\n std_mie_map[m] = std_ie_map\n self._mass_index_spe_map = std_mie_map\n\n def _standardize_mass_interaction_index_energy_map_indexing(self):\n miie_map = self._mass_interaction_index_energy_map\n std_miie_map = dict()\n for m, iie_map in miie_map.items():\n std_iie_map = dict()\n for ii, energy in iie_map.items():\n next_ii = self._standardize_interaction_index_tuple(ii)\n std_iie_map[next_ii] = energy\n std_miie_map[m] = std_iie_map\n self._mass_interaction_index_energy_map = std_miie_map\n\n def _standard_orbital_index_map(self):\n return {v: k for k, v in self.standard_index_orbital_map.items()}\n\n def _standardize_interaction_index_tuple(self, ii_tuple):\n next_tuple = [self._standard_index(i) for i in ii_tuple[0:4]]\n next_tuple += tuple(ii_tuple[4:])\n return TwoBodyInteraction(*next_tuple)\n\n def _standard_index(self, i):\n io_map = self._index_orbital_map\n soi_map = self._standard_orbital_index_map()\n return soi_map[io_map[i]]\n\n def index_orbital_map(self):\n \"\"\"Returns a map from index used for SPE's and TBME's to the\n orbital quantum numbers\n \"\"\"\n return dict(self._index_orbital_map)\n\n def mass_zero_body_term_map(self):\n \"\"\"Returns a map\n mass number -> zero body term\n \"\"\"\n return dict(self._mass_zero_body_term_map)\n\n def other_constants(self):\n \"\"\"Returns a list of the values that follow the SPE's on the top line\n \"\"\"\n return list(self._other_constants)\n\n def interaction_index_mass_energy_map(self):\n \"\"\"From the mass -> interaction index -> energy map, creates a\n mapping from interaction index -> mass -> energy\n \"\"\"\n miie_map = self._mass_interaction_index_energy_map\n iime_map = dict()\n for mass_num in miie_map.keys():\n for tup in miie_map[mass_num]:\n if tup not in iime_map.keys():\n iime_map[tup] = dict()\n iime_map[tup][mass_num] = miie_map[mass_num][tup]\n return iime_map\n\n def index_mass_energy_map(self):\n \"\"\"From the (mass -> orbital index -> energy) map produce an\n (orbital index -> mass -> energy) map\n \"\"\"\n mie_map = self._mass_index_spe_map\n ime_map = dict()\n for mass in mie_map.keys():\n for index in mie_map[mass].keys():\n if index not in ime_map.keys():\n ime_map[index] = dict()\n ime_map[index][mass] = mie_map[mass][index]\n return ime_map\n\n def interaction_indices_to_interaction_qnums(self, ii):\n \"\"\"Converts a TwoBodyInteraction of orbital indices to a\n TwoBodyInteraction of QuantumNumbers. That is, it simply replaces the\n idices (which are just labels) with the acutal quantum numbers using\n the index -> orbital map.\n \"\"\"\n next_tup = tuple()\n for index in ii[0:4]:\n qnums = self._index_orbital_map[index]\n next_tup += qnums\n next_tup += ii.j\n return TwoBodyInteraction(*next_tup)\n\n\ndef _qnums_to_list(qnums):\n qn_list = list()\n for n in qnums:\n if '/' in n:\n sn = n.split('/')\n qn_list.append(float(sn[0]) / float(sn[1]))\n else:\n qn_list.append(float(n))\n return qn_list\n","sub_path":"src/deprecated/int/DatumInt.py","file_name":"DatumInt.py","file_ext":"py","file_size_in_byte":11311,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"241173040","text":"#!/usr/bin/python\n#-*- coding: utf-8 -*-\n\n###########################################################\n# © 2011 Daniel 'grindhold' Brendle and Team\n#\n# This file is part of Skarphed.\n#\n# Skarphed is free software: you can redistribute it and/or \n# modify it under the terms of the GNU Affero General Public License \n# as published by the Free Software Foundation, either \n# version 3 of the License, or (at your option) any later \n# version.\n#\n# Skarphed is distributed in the hope that it will be \n# useful, but WITHOUT ANY WARRANTY; without even the implied \n# warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR \n# PURPOSE. See the GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public \n# License along with Skarphed. \n# If not, see http://www.gnu.org/licenses/.\n###########################################################\n\n\nimport paramiko\nimport gobject\nfrom glue.threads import Tracker, KillableThread\n\nclass SSHConnection(paramiko.SSHClient):\n def __init__(self,server):\n paramiko.SSHClient.__init__(self)\n self.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n self.server = server\n \n def getServer(self):\n return self.server\n \nclass SSHConnector(KillableThread):\n def __init__(self,server):\n KillableThread.__init__(self)\n self.connection = SSHConnection(server)\n \n def run(self):\n Tracker().addThread(self)\n server = self.connection.getServer()\n try:\n self.connection.connect(server.getIp(), 22, server.getSSHName(), server.getSSHPass())\n except paramiko.AuthenticationException:\n server.setSSHState(server.SSH_LOCKED)\n server.ssh_connection = None\n else:\n server.setSSHState(server.SSH_UNLOCKED)\n server.ssh_connection = self.connection\n server.ssh_ready = True\n Tracker().removeThread(self)\n gobject.idle_add(server.updated)\n","sub_path":"admin/src/net/SSH.py","file_name":"SSH.py","file_ext":"py","file_size_in_byte":1999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"83785458","text":"#Water Calculator Game for MC Green robot\n#Designed and written by Manas Harbola (harbolam@mcvts.net) on behalf of Middlesex County Academy\n\nimport pygame\nfrom PIL import Image\nimport threading\nimport random\n\n#Define class for water-using appliances\nclass Appliance:\n def __init__(self, attrDict):\n valid_keys = ['img_src', #location of image to display\n 'img_resize', #desired width and length resize image\n 'coordinates', #coordinates to put image\n 'name', #name of appliance\n 'sliderQuestion', #question to ask user\n 'units', #units of use (ex. times, flushes, etc.)\n 'unitRate', #number of gallons used per unit\n 'usageRange'] #min and max usage in units\n\n for key in valid_keys:\n setattr(self, key, attrDict.get(key)) #initialize attributes\n\n self.unitAmt = 0 #current value of units\n self.gallonsUsed = 0 #current amount of water (in gallons) used\n\n def generate_img(self, screen, img_dims):\n img = Image.open(self.img_src)\n img_resized = img.resize((img_dims[2], img_dims[3]))\n mode = img_resized.mode\n size = img_resized.size\n data = img_resized.tobytes()\n appliance_img = pygame.image.fromstring(data, size, mode).convert_alpha()\n \n screen.blit(appliance_img, (img_dims[0], img_dims[1]))\n\n def generate_usage(self, surface, x, y, font, color=(0,0,0)):\n TextSurf, TextRect = text_objects(self.name + ': ' + '{0:.0f}'.format(self.gallonsUsed) + ' gallons', font, color)\n TextRect.center = (x, y)\n surface.blit(TextSurf, TextRect)\n \n def menu(self, screen):\n #Button Dimensions\n button_w = 100 / 2; button_h = 125 / 2\n button_x = 0.5 * window_size[0] - 0.5 * button_w\n up_button_y = 0.25 * window_size[1]\n down_button_y = up_button_y + (3 * button_h)\n\n #Instantiate buttons\n up_button = Button(screen, darker_green, green, (button_x, up_button_y, button_w, button_h), u'\\u2191', mediumText)\n down_button = Button(screen, darker_red, red, (button_x, down_button_y, button_w, button_h), u'\\u2193', mediumText)\n okay_button = Button(screen, darker_blue, blue, (button_x, 0.75 * window_size[1], 750 / 2, 250 / 2), 'OK', mediumText)\n \n #Portions of the screen that must ONLY be updated (Improves frame rate and performance)\n setting_rect = pygame.rect.Rect(button_x, up_button_y, 512 / 2, 4 * up_button_y) #Portion of screen which adjusts value\n back_rect = pygame.rect.Rect(button_x, 0.75 * window_size[1], 750 / 2, 250 / 2) #Portion of screen for the back button\n \n updateList = [setting_rect, back_rect]\n\n #Prepare title text and location\n QuestionSurf, QuestionRect = text_objects(self.sliderQuestion, largeText, white)\n QuestionRect.center = ((window_size[0] / 2), (window_size[1] / 8))\n\n #Make entire screen white to 'clean' it\n screen.fill(white)\n \n #Write background image to buffer\n screen.blit(background, backgroundRect)\n\n #Write text and image to buffer\n screen.blit(QuestionSurf, QuestionRect)\n self.generate_img(screen, self.coordinates + self.img_resize)\n\n #Update ENTIRE screen just once\n pygame.display.update()\n\n #Store as previous value of unit rectangle\n UnitRect = pygame.rect.Rect(0,0,0,0)\n\n running = True\n\n while running:\n #Handle events here:\n for event in pygame.event.get():\n print(event)\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n\n if event.type == pygame.MOUSEBUTTONDOWN:\n touch_status = True\n\n #Check if buttons are pressed if mouse button is down\n if up_button.is_pressed(touch_status):\n if self.unitAmt < self.usageRange[1]:\n self.unitAmt += 1\n self.gallonsUsed = self.unitAmt * self.unitRate\n\n if down_button.is_pressed(touch_status):\n if self.unitAmt > self.usageRange[0]:\n self.unitAmt -= 1\n self.gallonsUsed = self.unitAmt * self.unitRate\n\n if okay_button.is_pressed(touch_status):\n game_menu(gameDisplay)\n\n else:\n touch_status = False\n\n #Used for going from text with n digit unitAmt to n - 1 digit unitAmt\n #screen.fill(white, UnitRect)\n screen.blit(background, dest=UnitRect, area=UnitRect)\n\n #Prepare to display current unitAmt between buttons\n UnitSurf, UnitRect = text_objects(str(self.unitAmt) + ' ' + self.units, mediumText, white)\n UnitRect.topleft = (button_x, up_button_y + 1.5 * button_h)\n \n #Fill the area where the unitAmt text goes with white to 'refresh' it\n #screen.fill(white, UnitRect)\n #screen.blit(background, dest=UnitRect, area=UnitRect)\n screen.blit(UnitSurf, UnitRect)\n\n up_button.generate()\n down_button.generate()\n okay_button.generate()\n\n pygame.display.update(updateList)\n clock.tick(FPS)\n\n#Render text to a surface and a corresponding rectangle\ndef text_objects(text, font, color=(0,0,0)):\n textSurface = font.render(text, True, color)\n return textSurface, textSurface.get_rect()\n\n\n#Class for generating buttons\nclass Button:\n def __init__ (self, surfaceName, ac, ic, rectVals, text, font):\n self.ac = ac #Active color of button\n self.ic = ic #Inactive color of button\n self.rectAttrs = rectVals #(x, y, w, h) of button\n self.surfaceName = surfaceName\n self.text = text\n self.font = font\n\n def generate(self):\n x, y, w, h = self.rectAttrs\n mouse = pygame.mouse.get_pos()\n\n #Check if mouse is on button\n if x + w > mouse[0] > x and y + h > mouse[1] > y:\n pygame.draw.rect(self.surfaceName, self.ac, self.rectAttrs)\n\n #Else just show darker button\n else:\n pygame.draw.rect(self.surfaceName, self.ic, self.rectAttrs)\n\n textSurf, textRect = text_objects(self.text, self.font)\n\n textRect.center = (x + (w / 2), y + (h / 2))\n self.surfaceName.blit(textSurf, textRect)\n\n def get_rect(self):\n x, y, w, h = self.rectAttrs\n return pygame.rect.Rect(x, y, w, h)\n\n def is_pressed(self, touch_status):\n x, y, w, h = self.rectAttrs\n mouse = pygame.mouse.get_pos()\n \n #Check if mouse is hovering over button or not\n if x + w > mouse[0] > x and y + h > mouse[1] > y:\n if touch_status == True:\n #print('CLICK DETECTED')\n return True\n\n elif touch_status == False:\n return False\n\n #If mouse is not hovering over button, button must obviously not be pressed\n else:\n return False\n\n#Initiate pygame\npygame.init() #SUPER IMPORTANT\n\n#Screen size of window\nwindow_size = (1920,1080)\n\n#Max FPS (frames per second) of game\nFPS = 30\n\n#Define basic colors\nblack = (0, 0, 0)\nwhite = (255, 255, 255)\nred = (255, 0, 0)\ndarker_red = (200, 0, 0)\ngreen = (0, 255, 0)\ndarker_green = (0, 200, 0)\nblue = (50, 89, 250)\ndarker_blue = (35, 67, 250)\n\n#Define basic text sizes\nlargeText = pygame.font.Font('FreeSansBold.ttf', 64) #Large text, ideal for headings\nmediumText = pygame.font.Font('FreeSansBold.ttf', 48) #Medium text, ideal for subheadings\nmediumText2 = pygame.font.Font('FreeSansBold.ttf', 24)\nsmallText = pygame.font.Font('FreeSansBold.ttf', 16) #Small text, ideal for small buttons\n\n\n#Define background\nbackground = pygame.image.load('background.jpg')\nbackground = pygame.transform.scale(background, window_size)\nbackgroundRect = background.get_rect()\n\n#Instantiate window/surface\ngameDisplay = pygame.display.set_mode(window_size)\npygame.display.set_caption('Water Calculator')\nclock = pygame.time.Clock()\n\n#ok\ndishwasher_info = {'img_src': 'dishwasher.png', 'img_resize': (256,256), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Dishwasher', 'sliderQuestion': 'How Often Do You Use The Dishwasher in a Week?',\n 'units': 'times', 'unitRate': 6, 'usageRange':(0, 10)}\n#ok\nwashing_machine_info = {'img_src': 'washing_machine.png', 'img_resize': (300,400), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Washing Machine', 'sliderQuestion': 'How Often Do You Use The Washing Machine in a Week?',\n 'units': 'times', 'unitRate': 40, 'usageRange':(0, 10)}\n#ok\nshower_info = {'img_src': 'shower.png', 'img_resize': (512 // 2,512 // 2), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Shower', 'sliderQuestion': 'How Often Do You Use The Shower in a Week?',\n 'units': 'times', 'unitRate': 17.2, 'usageRange':(0, 20)}\n#ok\ntoilet_info = {'img_src': 'toilet.png', 'img_resize': (312 // 2,512 // 2), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Toilet', 'sliderQuestion': 'How Often Do You Use The Toilet in a Week?',\n 'units': 'times', 'unitRate': 1.6, 'usageRange':(0, 30)}\n\nsink_info = {'img_src': 'sink.png', 'img_resize': (728 // 2 ,512 // 2), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Sink', 'sliderQuestion': 'How Many Hours Do You Use The Kitchen Sink in a Week?',\n 'units': 'hours', 'unitRate': 2.2*60, 'usageRange':(0, 168)}\n#ok\nfaucet_info = {'img_src': 'faucet.png', 'img_resize': (333 // 2 ,512 // 2), 'coordinates': (window_size[0] / 4, window_size[1] / 4), \n 'name': 'Faucet', 'sliderQuestion': 'How Many Hours Do You Use The Faucet in a Week?',\n 'units': 'hours', 'unitRate': 1.5*60, 'usageRange':(0, 168)}\n\n#Instantiate appliance objects\ndishwasher = Appliance(dishwasher_info)\nwashing_machine = Appliance(washing_machine_info)\nshower = Appliance(shower_info)\ntoilet = Appliance(toilet_info)\nsink = Appliance(sink_info)\nfaucet = Appliance(faucet_info)\n\n#Start Menu for Game\ndef game_intro(surface):\n #Button Dimensions\n button_w = 750 / 2; button_h = 250 / 2\n help_button_x = 270; button_y = 1300 / 2\n button_spacing = 237 / 2 #spacing between buttons in px\n play_button_x = help_button_x + button_w + button_spacing\n quit_button_x = play_button_x + button_w + button_spacing\n \n #Instantiate buttons (Only needs to be done once)\n help_button = Button(surface, blue, darker_blue, (help_button_x, button_y, button_w, button_h), 'Help', mediumText)\n play_button = Button(surface, green, darker_green, (play_button_x, button_y, button_w, button_h), 'Play', mediumText)\n quit_button = Button(surface, red, darker_red, (quit_button_x, button_y, button_w, button_h), 'Quit', mediumText)\n \n #Portion of the screen that must ONLY be updated\n help_button_rect = help_button.get_rect()\n play_button_rect = play_button.get_rect()\n quit_button_rect = quit_button.get_rect()\n updateList = [help_button_rect, play_button_rect, quit_button_rect]\n\n #Prepare title text and location\n TextSurf, TextRect = text_objects('How Much Water Do You Use At Home?', largeText, white)\n TextRect.center = ((window_size[0] / 2), (window_size[1] / 4))\n\n #Make entire screen white to 'clean' the screen\n surface.fill(white)\n\n #Write background image to buffer\n surface.blit(background, backgroundRect)\n\n #Write text to buffer\n surface.blit(TextSurf, TextRect)\n\n #Update ENTIRE screen just once\n pygame.display.update()\n\n touch_status = False #False = no touch, True = touch present\n\n running = True\n\n while running:\n\n #Handle events here:\n for event in pygame.event.get():\n print(event)\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n\n if event.type == pygame.MOUSEBUTTONDOWN:\n #print('Pressed')\n touch_status = True\n\n #Check if buttons are pressed if mouse button is down\n if quit_button.is_pressed(touch_status): #If 'Quit' button is tapped\n pygame.quit()\n quit()\n\n if play_button.is_pressed(touch_status): #If 'Play' button is tapped\n game_menu(gameDisplay)\n\n if help_button.is_pressed(touch_status): #If 'Help' button is tapped\n game_help(gameDisplay)\n else:\n touch_status = False\n\n help_button.generate()\n play_button.generate()\n quit_button.generate()\n\n #Update only the portions that need to be updated\n pygame.display.update(updateList)\n clock.tick(FPS)\n\n\n#Help Menu for Game\ndef game_help(surface):\n #Instantiate button for returning back to intro page\n back_button = Button(surface, darker_green, green, (0.5 * window_size[0] - 150 , 0.75 * window_size[1], 750 / 2, 250 / 2), 'Back', mediumText)\n\n #back_button_rect = pygame.rect.Rect(back_button.rectAttrs[0], back_button.rectAttrs[1], back_button.rectAttrs[2], back_button.rectAttrs[3])\n back_button_rect = back_button.get_rect()\n updateList = [back_button_rect]\n \n TextSurf, TextRect = text_objects('How to Play:', largeText, white)\n TextRect.center = ((window_size[0] / 2), (window_size[1] / 4))\n \n line_spacing = 75 #Spacing between each line of instructions\n\n Line1Surf, Line1Rect = text_objects('1.) Select a button under an appliance to set its value', mediumText, white)\n Line1Rect.center = ((window_size[0] / 2), (window_size[1] / 4) + 150)\n \n Line2Surf, Line2Rect = text_objects('2.) After you are done, tap the \\'Back\\' button to return to the main screen', mediumText, white)\n Line2Rect.center = ((window_size[0] / 2), (window_size[1] / 4) + 150 + (2 * line_spacing))\n\n #Make entire screen white to clean it\n surface.fill(white)\n \n #Write background image to buffer\n surface.blit(background, backgroundRect)\n\n #Write text to buffer\n surface.blit(TextSurf, TextRect)\n surface.blit(Line1Surf, Line1Rect)\n surface.blit(Line2Surf, Line2Rect)\n\n #Update ENTIRE screen just once\n pygame.display.update()\n\n running = True\n\n while running:\n\n #Handle events here:\n for event in pygame.event.get():\n print(event)\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n\n if event.type == pygame.MOUSEBUTTONDOWN:\n #print('Pressed')\n touch_status = True\n\n #Check if buttons are pressed if mouse button is down\n if back_button.is_pressed(touch_status):\n #print('Intro Activated')\n game_intro(gameDisplay)\n else:\n touch_status = False\n\n surface.fill(white)\n back_button.generate()\n\n pygame.display.update(updateList)\n clock.tick(FPS)\n\n#Provide tips on how to reduce water usage\ndef game_tips(surface):\n #Instantiate button for returning back to intro page\n back_button = Button(surface, darker_green, green, (0.5 * window_size[0] - 150 , 0.80 * window_size[1], 750 / 2, 250 / 2), 'Back', mediumText)\n\n #back_button_rect = pygame.rect.Rect(back_button.rectAttrs[0], back_button.rectAttrs[1], back_button.rectAttrs[2], back_button.rectAttrs[3])\n back_button_rect = back_button.get_rect()\n updateList = [back_button_rect]\n \n #Make entire screen white to clean it\n surface.fill(white)\n \n #Write background image to buffer\n surface.blit(background, backgroundRect)\n\n TextSurf, TextRect = text_objects('Ways to reduce your water consumption:', largeText, white)\n TextRect.center = ((window_size[0] / 2), (window_size[1] / 4))\n surface.blit(TextSurf, TextRect)\n\n line_spacing = 50 #Spacing between each line of instructions\n \n #set limits for each appliance usage\n limits = [dishwasher.unitAmt > 7, washing_machine.unitAmt > 3, shower.unitAmt > 14,\n toilet.unitAmt > 14, sink.unitAmt > 10, faucet.unitAmt > 10]\n advice = ['Use the dishwasher less often, settle for washing with your hands',\n 'Use the washing machine less often, or invest in a water-efficient one',\n 'Try investing in a water-efficient shower head', \n 'Try investing in a water-efficient toilet',\n 'Turn off the sink when you finish using it',\n 'Turn off the faucet when you finish using it']\n \n\n x, y = window_size[0] / 2, (window_size[1] / 4) + 150\n count = 1\n\n for i in range(len(limits)):\n flag = limits[i]\n if flag:\n LineSurf, LineRect = text_objects(str(count) + '.) ' + advice[i], mediumText, white)\n LineRect.center = (x, y)\n surface.blit(LineSurf, LineRect)\n y += 1.5 * line_spacing\n count += 1\n \n #Update ENTIRE screen just once\n pygame.display.update()\n\n running = True\n\n while running:\n\n #Handle events here:\n for event in pygame.event.get():\n print(event)\n\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n\n if event.type == pygame.MOUSEBUTTONDOWN:\n touch_status = True\n\n #Check if buttons are pressed if mouse button is down\n if back_button.is_pressed(touch_status):\n game_menu(surface)\n else:\n touch_status = False\n\n back_button.generate()\n\n pygame.display.update(updateList)\n clock.tick(FPS)\n\n\n\n#Main menu for game\ndef game_menu(surface):\n #Define appliance button dimensions\n button_w = 325 / 2; button_h = 125 / 2\n\n #row spacing is distance between the top left corners of two horizontally distant buttons\n row_spacing = button_w + 150\n \n #same as row spacing but for column\n column_spacing = button_h + 300\n \n #x and y values of top left button\n ref_x = 0.125 * window_size[0]; ref_y = 0.45 * window_size[1]\n\n #Top Row\n dish_dims = (ref_x, ref_y, button_w, button_h)\n wash_dims = (ref_x + row_spacing, ref_y, button_w, button_h)\n shower_dims = (ref_x + 2 * row_spacing, ref_y, button_w, button_h)\n \n #Bottom Row\n toilet_dims = (ref_x, ref_y + column_spacing, button_w, button_h)\n sink_dims = (ref_x + row_spacing, ref_y + column_spacing, button_w, button_h)\n faucet_dims = (ref_x + 2 * row_spacing, ref_y + column_spacing, button_w, button_h)\n \n #Back/Tips/Quit button dimensions\n back_dims = (0.575 * window_size[0], 0.75 * window_size[1], button_w, button_h)\n tips_dims = (0.575 * window_size[0] + (row_spacing - 60), 0.75 * window_size[1], button_w, button_h)\n quit_dims = (0.575 * window_size[0] + 2.0 * (row_spacing - 60), 0.75 * window_size[1], button_w, button_h)\n \n #INSTANTIATE BUTTONS HERE\n dish_button = Button(surface, green, darker_green, dish_dims, dishwasher.name, smallText)\n wash_button = Button(surface, green, darker_green, wash_dims, washing_machine.name, smallText)\n shower_button = Button(surface, green, darker_green, shower_dims, shower.name, smallText)\n toilet_button = Button(surface, green, darker_green, toilet_dims, toilet.name, smallText)\n sink_button = Button(surface, green, darker_green, sink_dims, sink.name, smallText)\n faucet_button = Button(surface, green, darker_green, faucet_dims, faucet.name, smallText)\n back_button = Button(surface, blue, darker_blue, back_dims, 'Back', mediumText)\n quit_button = Button(surface, red, darker_red, quit_dims, 'Quit', mediumText)\n tips_button = Button(surface, green, darker_green, tips_dims, 'Tips', mediumText)\n\n #Instantiate areas to update\n dish_rect = dish_button.get_rect()\n wash_rect = wash_button.get_rect()\n shower_rect = shower_button.get_rect()\n toilet_rect = toilet_button.get_rect()\n sink_rect = sink_button.get_rect()\n faucet_rect = faucet_button.get_rect()\n back_rect = back_button.get_rect()\n quit_rect = quit_button.get_rect()\n tips_rect = tips_button.get_rect()\n\n updateList = [dish_rect, wash_rect, shower_rect, toilet_rect,\n sink_rect, faucet_rect, back_rect, quit_rect, tips_rect]\n \n\n #Prepare titles, text and locations\n TextSurf, TextRect = text_objects('How Much Water Do You Use At Home?', largeText, white)\n TextRect.center = ((window_size[0] / 2), (window_size[1] / 8))\n \n #Prepare subheading\n TotalSurf, TotalRect = text_objects('Total Water Usage:', mediumText, white)\n TotalRect.center = ((0.75 * window_size[0]), (window_size[1] / 4))\n\n #Prepare title for total\n gallon_sum = dishwasher.gallonsUsed + washing_machine.gallonsUsed + shower.gallonsUsed + toilet.gallonsUsed + sink.gallonsUsed + faucet.gallonsUsed\n\n #SumSurf, SumRect = text_objects('Total: ' + str(gallon_sum) + ' gallons per day', mediumText, white)\n SumSurf, SumRect = text_objects('Total: ' + '{0:.0f}'.format(gallon_sum) + ' gallons per week', mediumText, white)\n\n SumRect.center = ((0.75 * window_size[0]), (window_size[1] / 4 + 450))\n\n #Make entire screen white to 'clean' it\n surface.fill(white)\n\n #Write background image to buffer\n surface.blit(background, backgroundRect)\n\n #Write text to buffer\n surface.blit(TextSurf, TextRect)\n surface.blit(TotalSurf, TotalRect)\n surface.blit(SumSurf, SumRect)\n\n #Write water usage for each item to buffer\n dishwasher.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 100), mediumText2, white)\n washing_machine.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 150), mediumText2, white)\n shower.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 200), mediumText2, white)\n toilet.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 250), mediumText2, white)\n sink.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 300), mediumText2, white)\n faucet.generate_usage(surface, (0.75 * window_size[0]), (window_size[1] / 4 + 350), mediumText2, white)\n\n\n #Write images to buffer\n dishwasher.generate_img(surface, (ref_x - 20, ref_y - 225, 200, 200))\n washing_machine.generate_img(surface, (ref_x + row_spacing + 5, ref_y - 225, 150, 200))\n shower.generate_img(surface, (ref_x + 2 * row_spacing + 20, ref_y - 225, 200, 200))\n \n toilet.generate_img(surface, (ref_x + 20, ref_y + column_spacing - 225, 120, 200))\n sink.generate_img(surface, (ref_x + row_spacing - 30, ref_y + column_spacing - 225, 280, 200))\n faucet.generate_img(surface, (ref_x + 2 * row_spacing + 20, ref_y + column_spacing - 225, 120, 200))\n\n #Refresh ENTIRE screen ONCE\n pygame.display.update()\n \n running = True\n\n while running:\n\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n quit()\n\n if event.type == pygame.MOUSEBUTTONDOWN:\n touch_status = True\n\n #Check if buttons are pressed if mouse button is down\n if dish_button.is_pressed(touch_status):\n dishwasher.menu(surface)\n if wash_button.is_pressed(touch_status):\n washing_machine.menu(surface)\n if shower_button.is_pressed(touch_status):\n shower.menu(surface)\n if toilet_button.is_pressed(touch_status):\n toilet.menu(surface)\n if sink_button.is_pressed(touch_status):\n sink.menu(surface)\n if faucet_button.is_pressed(touch_status):\n faucet.menu(surface)\n if back_button.is_pressed(touch_status):\n game_intro(surface)\n if tips_button.is_pressed(touch_status):\n game_tips(surface)\n if quit_button.is_pressed(touch_status):\n pygame.quit()\n quit()\n \n dish_button.generate()\n wash_button.generate()\n shower_button.generate()\n toilet_button.generate()\n sink_button.generate()\n faucet_button.generate()\n back_button.generate()\n quit_button.generate()\n tips_button.generate()\n\n pygame.display.update(updateList)\n\n#Execute game\ngame_intro(gameDisplay)\n\npygame.quit()\nquit()\n","sub_path":"Games/water_calculator/water_calculator.py","file_name":"water_calculator.py","file_ext":"py","file_size_in_byte":24891,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"524670166","text":"# Python 3.7\n# 001\n\nimport aidfunctions\n\n\ndef writeArchive():\n # overwrite data\n arq = open('arq_demo.txt', mode='w') # sobrescreve / overwrite\n arq.write('salvando source_text...\\n')\n arq.close()\n\n # open and add data / append\n arq = open('arq_demo.txt', mode='a')\n arq.write('adicionando conteudo ...\\n')\n arq.close()\n\n # open and read\n arq = open('arq_demo.txt', mode='r')\n conteudo_arquivo = arq.read()\n arq.close()\n\n a = 'source_text'\n\n try:\n # uso de with\n arq = open('arq_demo.txt', mode='a')\n a = 'Testa se esta escrevendo! / try write test'\n arq.write(a + '\\n')\n arq.close()\n\n except IOError:\n a = 'arquivo nao encontrado / archive not found'\n print(a)\n\n except ValueError:\n a = 'Erro de valor / value error'\n print(a)\n\n except:\n a = 'Erro desconhecido / coringa / unknown errors'\n print(a)\n\n finally:\n b = 'Cleanup'\n print(b)\n\n print('Fim.')\n\n return aidfunctions.append_elements('writeArchive()', a, b)\n","sub_path":"conceptsFundamentals/arquivos.py","file_name":"arquivos.py","file_ext":"py","file_size_in_byte":1066,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"344350793","text":"from zope import schema\nfrom zope.interface import Interface\n\nfrom adk.sitecontent import sitecontentMessageFactory as _\n\n\nclass IADKPage(Interface):\n \"\"\"A lingua plone aware documents providing preview images.\"\"\"\n\n # -*- schema definition goes here -*-\n image = schema.Bytes(\n title=_(u\"Preview Image\"),\n required=False,\n description=_(u\"Upload an image that will be displayed as a preview \"\n u\"image in listings and search results.\"),\n )\n\n text = schema.SourceText(\n title=_(u\"Text\"),\n required=True,\n description=_(u\"Field description\"),\n )\n","sub_path":"src/adk.sitecontent/adk/sitecontent/interfaces/adkpage.py","file_name":"adkpage.py","file_ext":"py","file_size_in_byte":625,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"143328439","text":"from functools import partial\n\nfrom PyQt5 import uic\nfrom PyQt5.QtCore import Qt, QTimer\nfrom PyQt5.QtWidgets import QApplication, QDialog, QGraphicsScene, QGraphicsView\nfrom PyQt5.QtGui import QPen, QColor, QBrush\n\nForm, Window = uic.loadUiType(\"toaster_main.ui\")\n\napp = QApplication([])\nwindow = Window()\nform = Form()\nform.setupUi(window)\n\nMODES = {\n # Режим №1. Заливка экрана чередующимися полосами цвета\n 1: {\n \"1x8\":\n {\"width\": 1,\n \"line1\": (85, 107, 86),\n \"line2\": (170, 148, 169),\n \"background\": (0, 0, 0), },\n\n \"2x8\":\n {\"width\": 2,\n \"line1\": (85, 107, 86),\n \"line2\": (170, 148, 169),\n \"background\": (0, 0, 0), },\n\n \"1x6\":\n {\"width\": 1,\n \"line1\": (168, 80, 164),\n \"line2\": (84, 172, 88),\n \"background\": (0, 0, 0), },\n\n \"2x6\":\n {\"width\": 2,\n \"line1\": (168, 80, 164),\n \"line2\": (84, 172, 88),\n \"background\": (0, 0, 0), },\n },\n # Режим №2. Заливка экрана чередующимися полосами цвета\n 2: {\"1x8\":\n {\"width\": 1,\n \"line1\": (255, 255, 255),\n \"line2\": (0, 0, 0),\n \"background\": (0, 0, 0), },\n\n \"2x8\":\n {\"width\": 2,\n \"line1\": (255, 255, 255),\n \"line2\": (0, 0, 0),\n \"background\": (0, 0, 0), },\n }\n}\nCURRENT_MODE = {}\n\nSCREEN_HEIGHT_DIVIDER = 1\n\nLINES_ORIENTATION = 0 # 0 - vertical, 1 - horizontal\n\n\ndef on_test_clicked(mode, lines):\n form.buttonStart.setEnabled(True)\n global CURRENT_MODE\n CURRENT_MODE = MODES[mode][lines]\n\n\ndef on_full_screen_clicked():\n global SCREEN_HEIGHT_DIVIDER\n SCREEN_HEIGHT_DIVIDER = 1\n\n\ndef on_half_screen_clicked():\n global SCREEN_HEIGHT_DIVIDER\n SCREEN_HEIGHT_DIVIDER = 2\n\n\ndef on_vertical_orientation_clicked():\n global LINES_ORIENTATION\n LINES_ORIENTATION = 0\n\n\ndef on_horizontal_orientation_clicked():\n global LINES_ORIENTATION\n LINES_ORIENTATION = 1\n\n\ndef on_slider_clicked():\n form.labelBlinking.setText(str(form.sliderBlinking.value() / 2))\n\n\ndef on_start_button_clicked():\n full_screen = QDialog()\n full_screen.scene = QGraphicsScene()\n full_screen.scene.setBackgroundBrush(QBrush(QColor(0, 0, 0)))\n full_screen.graphic_view = QGraphicsView(full_screen.scene, full_screen)\n full_screen.graphic_view.setHorizontalScrollBarPolicy(Qt.ScrollBarAlwaysOff)\n full_screen.graphic_view.setVerticalScrollBarPolicy(Qt.ScrollBarAlwaysOff)\n full_screen.graphic_view.setAlignment(Qt.AlignTop)\n\n screen = app.primaryScreen()\n size = screen.size()\n width, height = size.width(), size.height()\n # rect = screen.availableGeometry()\n # width, height = rect.width(), rect.height()\n\n full_screen.graphic_view.setGeometry(0, 0, width, height)\n full_screen.graphic_view.setStyleSheet(\"border-width: 0px; border-style: solid\")\n\n show = True\n\n def toggle_lines():\n nonlocal show\n\n if show:\n full_screen.pen = QPen(QColor(*CURRENT_MODE['line1']))\n full_screen.bg_pen = QPen(QColor(*CURRENT_MODE['line2']))\n\n full_screen.pen.setWidth(CURRENT_MODE['width'])\n\n full_screen.scene.addRect(0, 0, width, height, brush=QBrush(QColor(*CURRENT_MODE['line2'])))\n\n if LINES_ORIENTATION == 0:\n for x in range(0, width, 2 * CURRENT_MODE['width']):\n full_screen.scene.addLine(x, 0, x, height // SCREEN_HEIGHT_DIVIDER, full_screen.pen)\n\n else:\n for y in range(0, height // SCREEN_HEIGHT_DIVIDER, 2 * CURRENT_MODE['width']):\n full_screen.scene.addLine(0, y, width, y, full_screen.pen)\n else:\n full_screen.scene.clear()\n\n show = not show\n\n if form.sliderBlinking.value():\n full_screen.timer = QTimer()\n full_screen.timer.timeout.connect(toggle_lines)\n\n stopped = False\n\n def on_key(e):\n if e.key() != Qt.Key_Space:\n full_screen.close()\n\n nonlocal stopped\n if e.key() == Qt.Key_Space:\n if not stopped:\n full_screen.timer.stop()\n full_screen.scene.clear()\n if stopped:\n full_screen.timer.start(form.sliderBlinking.value() * 500)\n stopped = not stopped\n\n full_screen.keyPressEvent = on_key\n full_screen.timer.start(form.sliderBlinking.value() * 500)\n else:\n def on_key(e):\n if e.key() != Qt.Key_Space:\n full_screen.close()\n\n if e.key() == Qt.Key_Space:\n toggle_lines()\n\n full_screen.keyPressEvent = on_key\n toggle_lines()\n\n full_screen.showFullScreen()\n full_screen.exec()\n\n\nform.buttonStart.clicked.connect(on_start_button_clicked)\nform.radioTestLVDS1_1x6.clicked.connect(partial(on_test_clicked, 1, '1x6'))\nform.radioTestLVDS1_1x8.clicked.connect(partial(on_test_clicked, 1, '1x8'))\nform.radioTestLVDS1_2x6.clicked.connect(partial(on_test_clicked, 1, '2x6'))\nform.radioTestLVDS1_2x8.clicked.connect(partial(on_test_clicked, 1, '2x8'))\n\nform.radioTestLVDS2_1x8.clicked.connect(partial(on_test_clicked, 2, '1x8'))\nform.radioTestLVDS2_2x8.clicked.connect(partial(on_test_clicked, 2, '2x8'))\n\nform.radioScreenFull.clicked.connect(on_full_screen_clicked)\nform.radioScreenUpperHalf.clicked.connect(on_half_screen_clicked)\n\nform.radioOrientationVertical.clicked.connect(on_vertical_orientation_clicked)\nform.radioOrientationHorizontal.clicked.connect(on_horizontal_orientation_clicked)\n\nform.sliderBlinking.valueChanged.connect(on_slider_clicked)\n\nwindow.show()\napp.exec()\n\n# на пробел черный экран\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"164646926","text":"from . import Defaults, ModflowDefaults\nfrom ..modflow import Modflow\nimport numpy as np\nimport flopy\n\n\nclass ModflowBuilder(object):\n \"\"\"\n Class for building Modflow model objects using built in Defaults.\n\n ModflowBuilder builds a steady state model using default values\n from gsflow.builder.Defaults or a user supplied Defaults object.\n\n The user can then edit the Modflow package objects to customize\n their model runs. (ex. create transient model), etc...\n\n supported packages include : DIS, BAS, UPW, UZF, SFR, NWT, OC\n\n Parameters\n ----------\n modelgrid : gsflow.builder.FishnetGenerator\n Structured grid object from FishnetGenerator or from\n flopy.discretization.StructuredGrid\n dem_data : np.ndarray\n numpy array of dimension (nrow, ncol) of DEM elevations\n model_name : str\n model name ex. \"my_test_model\"\n defaults : gsflow.builder.Defaults\n optional parameter, user can supply a gsflow.builder.Defaults\n instance to ModflowBuilder to use a custom set of default values\n\n \"\"\"\n\n def __init__(self, modelgrid, dem_data, model_name, defaults=None):\n exe_name = \"mfnwt.exe\"\n self._ml = Modflow(model_name, exe_name=exe_name)\n\n assert (modelgrid.nrow, modelgrid.ncol) == dem_data.shape\n self._modelgrid = modelgrid\n self._dem_data = dem_data\n\n if defaults is None:\n self._defaults = Defaults().modflow.to_dict()\n elif isinstance(defaults, Defaults):\n self._defaults = defaults.modflow.to_dict()\n elif isinstance(defaults, ModflowDefaults):\n self._defaults = defaults.to_dict()\n else:\n raise TypeError(\n \"Defaults must be Defaults or ModflowDefaults object\"\n )\n\n @property\n def model(self):\n \"\"\"\n Returns the gsflow.modflow.Modflow model object\n\n \"\"\"\n return self._ml\n\n def build_all(\n self,\n reach_data,\n segment_data,\n irunbnd,\n finf=None,\n botm=None,\n ibound=None,\n iuzfbnd=None,\n ):\n \"\"\"\n Method to build all supported modflow packages\n\n Parameters\n ----------\n reach_data : np.recarray\n flopy's ModflowSfr2 reach data parameter\n segment_data : np.recarray\n flopy's ModflowSfr2 segment data parameter\n irunbnd : np.ndarray\n flopy's ModflowUZF1 irunbnd parameter\n (runoff connection to streams)\n finf : np.ndarray\n UZF1's finf array which describes precipitation for recharge\n botm : np.ndarray\n bottom elevation for single layer model\n ibound : np.ndarray\n ibound array of active model cells\n iuzfbnd : np.ndarray\n uzf ibound array of active model cells\n\n Returns\n -------\n gsflow.modflow.Modflow object\n\n \"\"\"\n self.build_dis(botm=botm)\n self.build_bas6(ibound=ibound)\n self.build_upw()\n self.build_nwt()\n self.build_oc()\n self.build_uzf(irunbnd, finf=finf, iuzfbnd=iuzfbnd)\n self.build_sfr(reach_data, segment_data)\n return self._ml\n\n def build_dis(self, botm=None):\n \"\"\"\n Method to build the dis package using defaults\n\n Parameters\n ----------\n botm : float, int, np.ndarray\n Model botm elevations for discretization file. If botm is None\n then botm elevation is set 50 length units below DEM elevation\n\n Returns\n -------\n flopy.modflow.ModflowDis object\n\n \"\"\"\n if botm is None:\n botm = self._dem_data - 50.0\n else:\n botm = botm\n\n dis_defaults = self._defaults[\"dis\"]\n\n dis = flopy.modflow.ModflowDis(\n self._ml,\n nrow=self._modelgrid.nrow,\n ncol=self._modelgrid.ncol,\n delc=self._modelgrid.delc,\n delr=self._modelgrid.delr,\n top=self._dem_data,\n botm=botm,\n **dis_defaults\n )\n return dis\n\n def build_bas6(self, ibound=None):\n \"\"\"\n Method to build the BAS6 package\n\n Parameters\n ----------\n ibound : int, np.ndarray\n array of active modflow cells within the model, >0 for active, 0\n for inactive\n\n Returns\n -------\n flopy.modflow.ModflowBas object\n\n \"\"\"\n if ibound is None:\n ibound = np.ones(\n (self._modelgrid.nrow, self._modelgrid.ncol), dtype=int\n )\n bas_defaults = self._defaults[\"bas\"]\n bas = flopy.modflow.ModflowBas(\n self._ml, ibound=ibound, strt=self._dem_data, **bas_defaults\n )\n return bas\n\n def build_upw(self):\n \"\"\"\n Method to build a default version of the UPW package\n\n Returns\n -------\n flopy.modflow.ModflowUpw\n\n \"\"\"\n upw_defaults = self._defaults[\"upw\"]\n upw = flopy.modflow.ModflowUpw(self._ml, **upw_defaults)\n return upw\n\n def build_sfr(self, reach_data, segment_data):\n \"\"\"\n Method to build a default version of the SFR package\n\n Parameters\n ----------\n reach_data : np.recarray\n reach data recarray for ModflowSfr2\n segment_data : np.recarray\n segment data recarray for ModflowSfr2\n\n Returns\n -------\n flopy.modflow.ModflowSfr2\n\n \"\"\"\n # get package defaults and build the sfr package\n sfr_defaults = self._defaults[\"sfr\"][\"pkg\"]\n nreaches = len(reach_data)\n nsegments = len(segment_data)\n\n sfr = flopy.modflow.ModflowSfr2(\n self._ml,\n nstrm=nreaches,\n nss=nsegments,\n reach_data=reach_data,\n segment_data=segment_data,\n **sfr_defaults\n )\n return sfr\n\n def build_uzf(self, irunbnd, finf=None, iuzfbnd=None):\n \"\"\"\n Method to build a UZF package object using built in defaults\n\n Parameters\n ----------\n irunbnd : np.ndarray\n flopy's ModflowUZF1 irunbnd parameter\n (runoff connection to streams)\n finf : np.ndarray\n optional finf array of precipitation\n iuzfbnd : np.ndarray\n uzf ibound array of active unsaturated zone model cells\n\n Returns\n -------\n flopy.modflow.ModflowUzf\n\n \"\"\"\n uzf_defaults = self._defaults[\"uzf\"]\n if finf is None:\n finf = 1e-08\n if iuzfbnd is None:\n iuzfbnd = 1\n\n uzf = flopy.modflow.ModflowUzf1(\n self._ml,\n irunbnd=irunbnd,\n iuzfbnd=iuzfbnd,\n finf=finf,\n **uzf_defaults\n )\n return uzf\n\n def build_nwt(self):\n \"\"\"\n Method to build a SIMPLE nwt solver instance\n\n Returns\n -------\n flopy.modflow.ModflowNwt\n\n \"\"\"\n nwt_defaults = self._defaults[\"nwt\"]\n nwt = flopy.modflow.ModflowNwt(self._ml, **nwt_defaults)\n return nwt\n\n def build_oc(self):\n \"\"\"\n Method to build a simple one stress period OC object\n\n Returns\n -------\n flopy.modflow.ModflowOc\n \"\"\"\n # build output control\n oc_defaults = self._defaults[\"oc\"]\n oc = flopy.modflow.ModflowOc(self._ml, **oc_defaults)\n return oc\n","sub_path":"gsflow/builder/modflow_builder.py","file_name":"modflow_builder.py","file_ext":"py","file_size_in_byte":7496,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"549001347","text":"import re\nimport os.path\nimport hashlib\nimport logging\nimport datetime\nfrom werkzeug.wrappers import Response\nfrom werkzeug.wsgi import wrap_file\nfrom piecrust.app import PieCrust\nfrom piecrust.rendering import QualifiedPage\nfrom piecrust.routing import RouteNotFoundError\nfrom piecrust.sources.base import MODE_PARSING\nfrom piecrust.sources.pageref import PageNotFoundError\nfrom piecrust.uriutil import split_sub_uri\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef get_app_for_server(root_dir, debug=False, sub_cache_dir=None,\n root_url='/'):\n app = PieCrust(root_dir=root_dir, debug=debug)\n if sub_cache_dir:\n app._useSubCacheDir(sub_cache_dir)\n app.config.set('site/root', root_url)\n app.config.set('server/is_serving', True)\n return app\n\n\nclass RequestedPage(object):\n def __init__(self, qualified_page):\n self.qualified_page = qualified_page\n self.req_path = None\n self.page_num = 1\n self.not_found_errors = []\n\n\ndef find_routes(routes, uri):\n res = []\n tax_res = []\n for route in routes:\n metadata = route.matchUri(uri)\n if metadata is not None:\n if route.is_taxonomy_route:\n tax_res.append((route, metadata))\n else:\n res.append((route, metadata))\n return res + tax_res\n\n\ndef get_requested_page(app, req_path):\n # Try to find what matches the requested URL.\n req_path, page_num = split_sub_uri(app, req_path)\n\n routes = find_routes(app.routes, req_path)\n if len(routes) == 0:\n raise RouteNotFoundError(\"Can't find route for: %s\" % req_path)\n\n qp = None\n not_found_errors = []\n for route, route_metadata in routes:\n try:\n qp = _get_requested_page_for_route(\n app, route, route_metadata, req_path)\n if qp is not None:\n break\n except PageNotFoundError as nfe:\n not_found_errors.append(nfe)\n\n req_page = RequestedPage(qp)\n req_page.req_path = req_path\n req_page.page_num = page_num\n req_page.not_found_errors = not_found_errors\n return req_page\n\n\ndef _get_requested_page_for_route(app, route, route_metadata, req_path):\n taxonomy = None\n source = app.getSource(route.source_name)\n if route.taxonomy_name is None:\n factory = source.findPageFactory(route_metadata, MODE_PARSING)\n if factory is None:\n raise PageNotFoundError(\"No path found for '%s' in source '%s'.\" %\n (req_path, source.name))\n else:\n taxonomy = app.getTaxonomy(route.taxonomy_name)\n\n # This will raise `PageNotFoundError` naturally if not found.\n tax_page_ref = taxonomy.getPageRef(source)\n factory = tax_page_ref.getFactory()\n\n # Build the page.\n page = factory.buildPage()\n qp = QualifiedPage(page, route, route_metadata)\n return qp\n\n\ndef load_mimetype_map():\n mimetype_map = {}\n sep_re = re.compile(r'\\s+')\n path = os.path.join(os.path.dirname(__file__), 'mime.types')\n with open(path, 'r') as f:\n for line in f:\n tokens = sep_re.split(line)\n if len(tokens) > 1:\n for t in tokens[1:]:\n mimetype_map[t] = tokens[0]\n return mimetype_map\n\n\ndef make_wrapped_file_response(environ, request, path):\n logger.debug(\"Serving %s\" % path)\n\n # Check if we can return a 304 status code.\n mtime = os.path.getmtime(path)\n etag_str = '%s$$%s' % (path, mtime)\n etag = hashlib.md5(etag_str.encode('utf8')).hexdigest()\n if etag in request.if_none_match:\n response = Response()\n response.status_code = 304\n return response\n\n wrapper = wrap_file(environ, open(path, 'rb'))\n response = Response(wrapper)\n _, ext = os.path.splitext(path)\n response.set_etag(etag)\n response.last_modified = datetime.datetime.fromtimestamp(mtime)\n response.mimetype = mimetype_map.get(\n ext.lstrip('.'), 'text/plain')\n response.direct_passthrough = True\n return response\n\n\nmimetype_map = load_mimetype_map()\ncontent_type_map = {\n 'html': 'text/html',\n 'xml': 'text/xml',\n 'txt': 'text/plain',\n 'text': 'text/plain',\n 'css': 'text/css',\n 'xhtml': 'application/xhtml+xml',\n 'atom': 'application/atom+xml', # or 'text/xml'?\n 'rss': 'application/rss+xml', # or 'text/xml'?\n 'json': 'application/json'}\n\n","sub_path":"piecrust/serving/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":4433,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"281349045","text":"import tkinter as tk\r\n\r\nwindow = tk.Tk()\r\n\r\nwindow.title(\"My App\")\r\n\r\nwindow.geometry(\"400x400\")\r\n\r\n# Create label\r\n\r\ntitle = tk.Label(text=\"Hello World. Welcome to CS50 and to the GUI created.\")\r\ntitle.grid(column=0, row=0)\r\n\r\n# Create Button One\r\n\r\nbutton1 = tk.Button(text=\"Click Me!\")\r\nbutton1.grid(column=0, row=1)\r\n\r\n# Create Entry Field\r\nentry_field1 = tk.Entry()\r\nentry_field1.grid(column=0, row=2)\r\n\r\nwindow.mainloop()\r\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"232711730","text":"from __future__ import print_function\n\nfrom os import environ, getcwd\nfrom pdb import set_trace\nfrom random import uniform, randint\nimport sys\n\n# Update PYTHONPATH\nHOME = environ['HOME']\naxe = HOME + '/git/axe/axe/' # AXE\npystat = HOME + '/git/pystats/' # PySTAT\ncwd = getcwd() # Current Directory\nsys.path.extend([axe, pystat, cwd])\n\nfrom sklearn.ensemble import AdaBoostClassifier\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.tree import DecisionTreeClassifier\n\nfrom Prediction import *\nfrom _imports import *\nfrom abcd import _Abcd\nfrom cliffsDelta import *\nfrom contrastset import *\nfrom dectree import *\nfrom hist import *\nfrom smote import *\nimport makeAmodel as mam\nfrom methods1 import *\nimport numpy as np\nimport pandas as pd\nimport sk\n\n\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# PLANNING PHASE: 1. Decision Trees, 2. Contrast Sets\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n\ndef treatments(train = None, test = None, verbose = True, smoteit = False):\n\n def remember(node):\n key = node.f.name\n Val = node.val\n contrastSet.update({key: Val})\n # print contrastSet\n\n def forget(key):\n del contrastSet[key]\n\n def objectiveScores(lst):\n obj = ([k.cells[-2] for k in lst.rows])\n return np.mean([k for k in obj]), [k for k in obj]\n\n def compare(node, test):\n leaves = [n for n in test.kids] if len(test.kids) else [test]\n# set_trace()\n for k in leaves:\n return objectiveScores(k) < objectiveScores(node), [objectiveScores(k),\n objectiveScores(node)]\n def getKey():\n keys = {}\n for i in xrange(len(test_DF.headers)):\n keys.update({test_DF.headers[i].name[1:]:i})\n return keys\n\n #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n # New Methods - 02/03/2015\n #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n def leaves(node):\n L = []\n if node.kids:\n for l in node.kids:\n L.extend(leaves(node.kids))\n return L\n else:\n return L.extend(node)\n return L\n\n def score(node):\n pass\n\n #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n # Main\n #~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n # Training data\n train_DF = createTbl(train)\n print('Done')\n # if smoteit: train_DF = SMOTE(data = train_DF, atleast = 50, atmost = 100)\n # Testing data\n test_DF = createTbl(test)\n print('Done')\n# set_trace()\n # Decision Tree\n\n t = discreteNums(train_DF, map(lambda x: x.cells, train_DF._rows))\n print('Done')\n myTree = tdiv(t)\n print('Done')\n if verbose: showTdiv(myTree)\n\n # Testing data\n testCase = test_DF._rows\n\n keys = getKey();\n newTab = []\n for tC in testCase:\n newRow = tC;\n loc = drop(tC, myTree)\n newNode = loc;\n set_trace()\n if newNode.lvl > 0:\n # Go up one Level\n _up = newNode.up\n # look at the kids\n _kids = _up.kids\n _leaves = [leaves(_k) for _k in _kids]\n set_trace()\n branches = [];\n while newNode.lvl > 0:\n newNode = newNode.up;\n branches.append(newNode);\n # A dict of contrast sets\n contrastSet = {};\n # print loc.f.name, loc.lvl+1, loc.val\n for nn in branches:\n toScan = nn.kids\n # set_trace()\n for testing in toScan:\n isBetter, obj = compare(loc, testing)\n if isBetter:\n remember(testing)\n continue # As soon as the first better node is found, exit..\n\n # Pick a random value in the range suggested by the contrast set and\n # assign it to the row.\n for k in contrastSet:\n min, max = contrastSet[k]\n if isinstance(min, int) and isinstance(max, int):\n val = randint(min, max)\n else: val = uniform(min, max)\n newRow.cells[keys[k]] = val\n\n newTab.append(newRow.cells)\n\n updatedTab = clone(test_DF, rows = newTab, discrete = True)\n return updatedTab\n# saveImg(bugs(test_df), num_bins = 50, fname = 'bugsBefore', ext = '.jpg')\n# set_trace()\n\ndef planningTest():\n # Test contrast sets\n n = 1\n dir = '../Data'\n one, two = explore(dir)\n # Training data\n train_DF = createTbl(one[n])\n # Test data\n test_df = createTbl(two[n])\n newTab = treatments(train = [one[n][0]],\n test = [one[n][1]],\n verbose = False,\n smoteit = False)\n\n\nif __name__ == '__main__':\n planningTest()\n","sub_path":"ExtractFeatures/Data/rahul/Planning.py","file_name":"Planning.py","file_ext":"py","file_size_in_byte":4543,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"450316606","text":"# -*- coding: utf-8 -*-\n'''\nWrap libsodium routines\n'''\n# pylint: disable=C0103\n# Import python libs\nimport ctypes, ctypes.util\nimport sys\nimport os\n\n__SONAMES = (23, 18, 17, 13, 10, 5, 4)\n\n\ndef _get_nacl():\n '''\n Locate the nacl c libs to use\n '''\n # Import libsodium\n l_path = ctypes.util.find_library('sodium')\n if l_path is not None:\n return ctypes.cdll.LoadLibrary(l_path)\n\n if sys.platform.startswith('win'):\n try:\n return ctypes.cdll.LoadLibrary('libsodium')\n except OSError:\n pass\n for soname_ver in __SONAMES:\n try:\n return ctypes.cdll.LoadLibrary(\n 'libsodium-{0}'.format(soname_ver)\n )\n except OSError:\n pass\n msg = 'Could not locate nacl lib, searched for libsodium'\n raise OSError(msg)\n elif sys.platform.startswith('darwin'):\n try:\n return ctypes.cdll.LoadLibrary('libsodium.dylib')\n except OSError:\n pass\n try:\n libidx = __file__.find('lib')\n if libidx > 0:\n libpath = __file__[0:libidx+3] + '/libsodium.dylib'\n return ctypes.cdll.LoadLibrary(libpath)\n except OSError:\n msg = 'Could not locate nacl lib, searched for libsodium'\n raise OSError(msg)\n else:\n try:\n return ctypes.cdll.LoadLibrary('libsodium.so')\n except OSError:\n pass\n try:\n return ctypes.cdll.LoadLibrary('/usr/local/lib/libsodium.so')\n except OSError:\n pass\n try:\n libidx = __file__.find('lib')\n if libidx > 0:\n libpath = __file__[0:libidx+3] + '/libsodium.so'\n return ctypes.cdll.LoadLibrary(libpath)\n except OSError:\n pass\n\n for soname_ver in __SONAMES:\n try:\n return ctypes.cdll.LoadLibrary(\n 'libsodium.so.{0}'.format(soname_ver)\n )\n except OSError:\n pass\n try:\n # fall back to shipped libsodium, trust os version first\n libpath = os.path.join(os.path.dirname(__file__), 'libsodium.so')\n return ctypes.cdll.LoadLibrary(libpath)\n except OSError:\n pass\n msg = 'Could not locate nacl lib, searched for libsodium.so, '\n for soname_ver in __SONAMES:\n msg += 'libsodium.so.{0}, '.format(soname_ver)\n raise OSError(msg)\n\n# Don't load libnacl if we are in sphinx\nif not 'sphinx' in sys.argv[0]:\n nacl = _get_nacl()\n DOC_RUN = False\nelse:\n nacl = None\n DOC_RUN = True\n\n\n# Define exceptions\nclass CryptError(Exception):\n \"\"\"\n Base Exception for cryptographic errors\n \"\"\"\n\nif not DOC_RUN:\n sodium_init = nacl.sodium_init\n sodium_init.res_type = ctypes.c_int\n if sodium_init() < 0:\n raise RuntimeError('sodium_init() call failed!')\n\n # Define constants\n try:\n crypto_box_SEALBYTES = nacl.crypto_box_sealbytes()\n HAS_SEAL = True\n except AttributeError:\n HAS_SEAL = False\n try:\n crypto_aead_aes256gcm_KEYBYTES = nacl.crypto_aead_aes256gcm_keybytes()\n crypto_aead_aes256gcm_NPUBBYTES = nacl.crypto_aead_aes256gcm_npubbytes()\n crypto_aead_aes256gcm_ABYTES = nacl.crypto_aead_aes256gcm_abytes()\n HAS_AEAD_AES256GCM = bool(nacl.crypto_aead_aes256gcm_is_available())\n crypto_aead_chacha20poly1305_ietf_KEYBYTES = nacl.crypto_aead_chacha20poly1305_ietf_keybytes()\n crypto_aead_chacha20poly1305_ietf_NPUBBYTES = nacl.crypto_aead_chacha20poly1305_ietf_npubbytes()\n crypto_aead_chacha20poly1305_ietf_ABYTES = nacl.crypto_aead_chacha20poly1305_ietf_abytes()\n crypto_aead_xchacha20poly1305_ietf_KEYBYTES = nacl.crypto_aead_xchacha20poly1305_ietf_keybytes()\n crypto_aead_xchacha20poly1305_ietf_NPUBBYTES = nacl.crypto_aead_xchacha20poly1305_ietf_npubbytes()\n crypto_aead_xchacha20poly1305_ietf_ABYTES = nacl.crypto_aead_xchacha20poly1305_ietf_abytes()\n HAS_AEAD_CHACHA20POLY1305_IETF = True\n HAS_AEAD_XCHACHA20POLY1305_IETF = True\n HAS_AEAD = True\n except AttributeError:\n HAS_AEAD_AES256GCM = False\n HAS_AEAD_CHACHA20POLY1305_IETF = False\n HAS_AEAD_XCHACHA20POLY1305_IETF = False\n HAS_AEAD = False\n\n crypto_box_SECRETKEYBYTES = nacl.crypto_box_secretkeybytes()\n crypto_box_SEEDBYTES = nacl.crypto_box_seedbytes()\n crypto_box_PUBLICKEYBYTES = nacl.crypto_box_publickeybytes()\n crypto_box_NONCEBYTES = nacl.crypto_box_noncebytes()\n crypto_box_ZEROBYTES = nacl.crypto_box_zerobytes()\n crypto_box_BOXZEROBYTES = nacl.crypto_box_boxzerobytes()\n crypto_box_BEFORENMBYTES = nacl.crypto_box_beforenmbytes()\n crypto_scalarmult_BYTES = nacl.crypto_scalarmult_bytes()\n crypto_scalarmult_SCALARBYTES = nacl.crypto_scalarmult_scalarbytes()\n crypto_sign_BYTES = nacl.crypto_sign_bytes()\n crypto_sign_SEEDBYTES = nacl.crypto_sign_secretkeybytes() // 2\n crypto_sign_PUBLICKEYBYTES = nacl.crypto_sign_publickeybytes()\n crypto_sign_SECRETKEYBYTES = nacl.crypto_sign_secretkeybytes()\n crypto_sign_ed25519_PUBLICKEYBYTES = nacl.crypto_sign_ed25519_publickeybytes()\n crypto_sign_ed25519_SECRETKEYBYTES = nacl.crypto_sign_ed25519_secretkeybytes()\n crypto_box_MACBYTES = crypto_box_ZEROBYTES - crypto_box_BOXZEROBYTES\n crypto_secretbox_KEYBYTES = nacl.crypto_secretbox_keybytes()\n crypto_secretbox_NONCEBYTES = nacl.crypto_secretbox_noncebytes()\n crypto_secretbox_ZEROBYTES = nacl.crypto_secretbox_zerobytes()\n crypto_secretbox_BOXZEROBYTES = nacl.crypto_secretbox_boxzerobytes()\n crypto_secretbox_MACBYTES = crypto_secretbox_ZEROBYTES - crypto_secretbox_BOXZEROBYTES\n crypto_stream_KEYBYTES = nacl.crypto_stream_keybytes()\n crypto_stream_NONCEBYTES = nacl.crypto_stream_noncebytes()\n crypto_auth_BYTES = nacl.crypto_auth_bytes()\n crypto_auth_KEYBYTES = nacl.crypto_auth_keybytes()\n crypto_onetimeauth_BYTES = nacl.crypto_onetimeauth_bytes()\n crypto_onetimeauth_KEYBYTES = nacl.crypto_onetimeauth_keybytes()\n crypto_generichash_BYTES = nacl.crypto_generichash_bytes()\n crypto_generichash_BYTES_MIN = nacl.crypto_generichash_bytes_min()\n crypto_generichash_BYTES_MAX = nacl.crypto_generichash_bytes_max()\n crypto_generichash_KEYBYTES = nacl.crypto_generichash_keybytes()\n crypto_generichash_KEYBYTES_MIN = nacl.crypto_generichash_keybytes_min()\n crypto_generichash_KEYBYTES_MAX = nacl.crypto_generichash_keybytes_max()\n crypto_scalarmult_curve25519_BYTES = nacl.crypto_scalarmult_curve25519_bytes()\n crypto_hash_BYTES = nacl.crypto_hash_sha512_bytes()\n crypto_hash_sha256_BYTES = nacl.crypto_hash_sha256_bytes()\n crypto_hash_sha512_BYTES = nacl.crypto_hash_sha512_bytes()\n crypto_verify_16_BYTES = nacl.crypto_verify_16_bytes()\n crypto_verify_32_BYTES = nacl.crypto_verify_32_bytes()\n crypto_verify_64_BYTES = nacl.crypto_verify_64_bytes()\n\n try:\n randombytes_SEEDBYTES = nacl.randombytes_seedbytes()\n HAS_RAND_SEED = True\n except AttributeError:\n HAS_RAND_SEED = False\n\n try:\n crypto_kdf_PRIMITIVE = nacl.crypto_kdf_primitive()\n crypto_kdf_BYTES_MIN = nacl.crypto_kdf_bytes_min()\n crypto_kdf_BYTES_MAX = nacl.crypto_kdf_bytes_max()\n crypto_kdf_CONTEXTBYTES = nacl.crypto_kdf_contextbytes()\n crypto_kdf_KEYBYTES = nacl.crypto_kdf_keybytes()\n HAS_CRYPT_KDF = True\n except AttributeError:\n HAS_CRYPT_KDF = False\n\n try:\n crypto_kx_PUBLICKEYBYTES = nacl.crypto_kx_publickeybytes()\n crypto_kx_SECRETKEYBYTES = nacl.crypto_kx_secretkeybytes()\n crypto_kx_SEEDBYTES = nacl.crypto_kx_seedbytes()\n crypto_kx_SESSIONKEYBYTES = nacl.crypto_kx_sessionkeybytes()\n crypto_kx_PRIMITIVE = nacl.crypto_kx_primitive()\n HAS_CRYPT_KX = True\n except AttributeError:\n HAS_CRYPT_KX = False\n\n # pylint: enable=C0103\n\n# Pubkey defs\n\n\ndef crypto_box_keypair():\n '''\n Generate and return a new keypair\n\n pk, sk = nacl.crypto_box_keypair()\n '''\n pk = ctypes.create_string_buffer(crypto_box_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_box_SECRETKEYBYTES)\n nacl.crypto_box_keypair(pk, sk)\n return pk.raw, sk.raw\n\n\ndef crypto_box_seed_keypair(seed):\n '''\n Generate and return a keypair from a key seed \n '''\n if len(seed) != crypto_box_SEEDBYTES:\n raise ValueError('Invalid key seed')\n pk = ctypes.create_string_buffer(crypto_box_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_box_SECRETKEYBYTES)\n nacl.crypto_box_seed_keypair(pk, sk, seed)\n return pk.raw, sk.raw\n\n\ndef crypto_scalarmult_base(sk):\n '''\n Compute and return the scalar product of a standard group element and the given integer.\n\n This can be used to derive a Curve25519 public key from a Curve25519 secret key,\n such as for usage with crypto_box and crypto_box_seal.\n '''\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n pk = ctypes.create_string_buffer(crypto_box_PUBLICKEYBYTES)\n if nacl.crypto_scalarmult_base(pk, sk):\n raise CryptError('Failed to compute scalar product')\n return pk.raw\n\n\ndef crypto_box(msg, nonce, pk, sk):\n '''\n Using a public key and a secret key encrypt the given message. A nonce\n must also be passed in, never reuse the nonce\n\n enc_msg = nacl.crypto_box('secret message', , , )\n '''\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n pad = b'\\x00' * crypto_box_ZEROBYTES + msg\n c = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_box(c, pad, ctypes.c_ulonglong(len(pad)), nonce, pk, sk)\n if ret:\n raise CryptError('Unable to encrypt message')\n return c.raw[crypto_box_BOXZEROBYTES:]\n\n\ndef crypto_box_open(ctxt, nonce, pk, sk):\n '''\n Decrypts a message given the receiver's private key, and sender's public key\n '''\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n pad = b'\\x00' * crypto_box_BOXZEROBYTES + ctxt\n msg = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_box_open(\n msg,\n pad,\n ctypes.c_ulonglong(len(pad)),\n nonce,\n pk,\n sk)\n if ret:\n raise CryptError('Unable to decrypt ciphertext')\n return msg.raw[crypto_box_ZEROBYTES:]\n\n\ndef crypto_box_easy(msg, nonce, pk, sk):\n '''\n Using a public key and a secret key encrypt the given message. A nonce\n must also be passed in, never reuse the nonce\n\n enc_msg = nacl.crypto_box_easy('secret message', , , )\n '''\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n c = ctypes.create_string_buffer(len(msg) + crypto_box_MACBYTES)\n ret = nacl.crypto_box(c, msg, ctypes.c_ulonglong(len(msg)), nonce, pk, sk)\n if ret:\n raise CryptError('Unable to encrypt message')\n return c.raw\n\n\ndef crypto_box_open_easy(ctxt, nonce, pk, sk):\n '''\n Decrypts a message given the receiver's private key, and sender's public key\n '''\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n msg = ctypes.create_string_buffer(len(ctxt) - crypto_box_MACBYTES)\n ret = nacl.crypto_box_open(\n msg,\n ctxt,\n ctypes.c_ulonglong(len(ctxt)),\n nonce,\n pk,\n sk)\n if ret:\n raise CryptError('Unable to decrypt ciphertext')\n return msg.raw[crypto_box_ZEROBYTES:]\n\n\ndef crypto_box_beforenm(pk, sk):\n '''\n Partially performs the computation required for both encryption and decryption of data\n '''\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n k = ctypes.create_string_buffer(crypto_box_BEFORENMBYTES)\n ret = nacl.crypto_box_beforenm(k, pk, sk)\n if ret:\n raise CryptError('Unable to compute shared key')\n return k.raw\n\n\ndef crypto_box_afternm(msg, nonce, k):\n '''\n Encrypts a given a message, using partial computed data\n '''\n if len(k) != crypto_box_BEFORENMBYTES:\n raise ValueError('Invalid shared key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n pad = b'\\x00' * crypto_box_ZEROBYTES + msg\n ctxt = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_box_afternm(ctxt, pad, ctypes.c_ulonglong(len(pad)), nonce, k)\n if ret:\n raise CryptError('Unable to encrypt messsage')\n return ctxt.raw[crypto_box_BOXZEROBYTES:]\n\n\ndef crypto_box_open_afternm(ctxt, nonce, k):\n '''\n Decrypts a ciphertext ctxt given k\n '''\n if len(k) != crypto_box_BEFORENMBYTES:\n raise ValueError('Invalid shared key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n pad = b'\\x00' * crypto_box_BOXZEROBYTES + ctxt\n msg = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_box_open_afternm(\n msg,\n pad,\n ctypes.c_ulonglong(len(pad)),\n nonce,\n k)\n if ret:\n raise CryptError('unable to decrypt message')\n return msg.raw[crypto_box_ZEROBYTES:]\n\n\ndef crypto_box_easy_afternm(msg, nonce, k):\n '''\n Using a precalculated shared key, encrypt the given message. A nonce\n must also be passed in, never reuse the nonce\n\n enc_msg = nacl.crypto_box_easy_afternm('secret message', , )\n '''\n if len(k) != crypto_box_BEFORENMBYTES:\n raise ValueError('Invalid shared key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n ctxt = ctypes.create_string_buffer(len(msg) + crypto_box_MACBYTES)\n ret = nacl.crypto_box_easy_afternm(ctxt, msg, ctypes.c_ulonglong(len(msg)), nonce, k)\n if ret:\n raise CryptError('Unable to encrypt messsage')\n return ctxt.raw\n\n\ndef crypto_box_open_easy_afternm(ctxt, nonce, k):\n '''\n Decrypts a ciphertext ctxt given k\n '''\n if len(k) != crypto_box_BEFORENMBYTES:\n raise ValueError('Invalid shared key')\n if len(nonce) != crypto_box_NONCEBYTES:\n raise ValueError('Invalid nonce')\n msg = ctypes.create_string_buffer(len(ctxt) - crypto_box_MACBYTES)\n ret = nacl.crypto_box_open_easy_afternm(\n msg,\n ctxt,\n ctypes.c_ulonglong(len(ctxt)),\n nonce,\n k)\n if ret:\n raise CryptError('unable to decrypt message')\n return msg.raw\n\n\ndef crypto_box_seal(msg, pk):\n '''\n Using a public key to encrypt the given message. The identity of the sender cannot be verified.\n\n enc_msg = nacl.crypto_box_seal('secret message', )\n '''\n if not HAS_SEAL:\n raise ValueError('Underlying Sodium library does not support sealed boxes')\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if not isinstance(msg, bytes):\n raise TypeError('Message must be bytes')\n\n c = ctypes.create_string_buffer(len(msg) + crypto_box_SEALBYTES)\n ret = nacl.crypto_box_seal(c, msg, ctypes.c_ulonglong(len(msg)), pk)\n if ret:\n raise CryptError('Unable to encrypt message')\n return c.raw\n\n\ndef crypto_box_seal_open(ctxt, pk, sk):\n '''\n Decrypts a message given the receiver's public and private key.\n '''\n if not HAS_SEAL:\n raise ValueError('Underlying Sodium library does not support sealed boxes')\n if len(pk) != crypto_box_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n if len(sk) != crypto_box_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n if not isinstance(ctxt, bytes):\n raise TypeError('Message must be bytes')\n\n c = ctypes.create_string_buffer(len(ctxt) - crypto_box_SEALBYTES)\n ret = nacl.crypto_box_seal_open(c, ctxt, ctypes.c_ulonglong(len(ctxt)), pk, sk)\n if ret:\n raise CryptError('Unable to decrypt message')\n return c.raw\n\n# Signing functions\n\n\ndef crypto_sign_keypair():\n '''\n Generates a signing/verification key pair\n '''\n vk = ctypes.create_string_buffer(crypto_sign_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_sign_SECRETKEYBYTES)\n ret = nacl.crypto_sign_keypair(vk, sk)\n if ret:\n raise ValueError('Failed to generate keypair')\n return vk.raw, sk.raw\n\n\ndef crypto_sign_ed25519_keypair():\n '''\n Generates a signing/verification Ed25519 key pair\n '''\n vk = ctypes.create_string_buffer(crypto_sign_ed25519_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_sign_ed25519_SECRETKEYBYTES)\n ret = nacl.crypto_sign_ed25519_keypair(vk, sk)\n if ret:\n raise ValueError('Failed to generate keypair')\n return vk.raw, sk.raw\n\n\ndef crypto_sign_ed25519_sk_to_pk(sk):\n '''\n Extract the public key from the secret key\n '''\n if len(sk) != crypto_sign_ed25519_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n\n pk = ctypes.create_string_buffer(crypto_sign_PUBLICKEYBYTES)\n ret = nacl.crypto_sign_ed25519_sk_to_pk(pk, sk)\n if ret:\n raise ValueError('Failed to generate public key')\n return pk.raw\n\n\ndef crypto_sign_ed25519_sk_to_seed(sk):\n '''\n Extract the seed from the secret key \n '''\n if len(sk) != crypto_sign_ed25519_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n\n seed = ctypes.create_string_buffer(crypto_sign_SEEDBYTES)\n ret = nacl.crypto_sign_ed25519_sk_to_seed(seed, sk)\n if ret:\n raise ValueError('Failed to generate seed')\n return seed.raw\n\n\ndef crypto_sign(msg, sk):\n '''\n Sign the given message with the given signing key\n '''\n if len(sk) != crypto_sign_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n\n sig = ctypes.create_string_buffer(len(msg) + crypto_sign_BYTES)\n slen = ctypes.pointer(ctypes.c_ulonglong())\n ret = nacl.crypto_sign(\n sig,\n slen,\n msg,\n ctypes.c_ulonglong(len(msg)),\n sk)\n if ret:\n raise ValueError('Failed to sign message')\n return sig.raw\n\n\ndef crypto_sign_detached(msg, sk):\n '''\n Return signature for the given message with the given signing key\n '''\n if len(sk) != crypto_sign_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n\n sig = ctypes.create_string_buffer(crypto_sign_BYTES)\n slen = ctypes.pointer(ctypes.c_ulonglong())\n ret = nacl.crypto_sign_detached(\n sig,\n slen,\n msg,\n ctypes.c_ulonglong(len(msg)),\n sk)\n if ret:\n raise ValueError('Failed to sign message')\n return sig.raw[:slen.contents.value]\n\n\ndef crypto_sign_seed_keypair(seed):\n '''\n Computes and returns the secret and verify keys from the given seed\n '''\n if len(seed) != crypto_sign_SEEDBYTES:\n raise ValueError('Invalid Seed')\n\n sk = ctypes.create_string_buffer(crypto_sign_SECRETKEYBYTES)\n vk = ctypes.create_string_buffer(crypto_sign_PUBLICKEYBYTES)\n\n ret = nacl.crypto_sign_seed_keypair(vk, sk, seed)\n if ret:\n raise CryptError('Failed to generate keypair from seed')\n return (vk.raw, sk.raw)\n\n\ndef crypto_sign_open(sig, vk):\n '''\n Verifies the signed message sig using the signer's verification key\n '''\n if len(vk) != crypto_sign_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n\n msg = ctypes.create_string_buffer(len(sig))\n msglen = ctypes.c_ulonglong()\n msglenp = ctypes.pointer(msglen)\n ret = nacl.crypto_sign_open(\n msg,\n msglenp,\n sig,\n ctypes.c_ulonglong(len(sig)),\n vk)\n if ret:\n raise ValueError('Failed to validate message')\n return msg.raw[:msglen.value] # pylint: disable=invalid-slice-index\n\n\ndef crypto_sign_verify_detached(sig, msg, vk):\n '''\n Verifies that sig is a valid signature for the message msg using the signer's verification key\n '''\n if len(sig) != crypto_sign_BYTES:\n raise ValueError('Invalid signature')\n if len(vk) != crypto_sign_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n\n ret = nacl.crypto_sign_verify_detached(\n sig,\n msg,\n ctypes.c_ulonglong(len(msg)),\n vk)\n if ret:\n raise ValueError('Failed to validate message')\n return msg\n\n\n# Authenticated Symmetric Encryption\n\n\ndef crypto_secretbox(message, nonce, key):\n \"\"\"Encrypts and authenticates a message using the given secret key, and nonce\n\n Args:\n message (bytes): a message to encrypt\n nonce (bytes): nonce, does not have to be confidential must be\n `crypto_secretbox_NONCEBYTES` in length\n key (bytes): secret key, must be `crypto_secretbox_KEYBYTES` in\n length\n\n Returns:\n bytes: the ciphertext\n\n Raises:\n ValueError: if arguments' length is wrong or the operation has failed.\n \"\"\"\n if len(key) != crypto_secretbox_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_secretbox_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n pad = b'\\x00' * crypto_secretbox_ZEROBYTES + message\n ctxt = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_secretbox(\n ctxt, pad, ctypes.c_ulonglong(len(pad)), nonce, key)\n if ret:\n raise ValueError('Failed to encrypt message')\n return ctxt.raw[crypto_secretbox_BOXZEROBYTES:]\n\n\ndef crypto_secretbox_open(ctxt, nonce, key):\n \"\"\"\n Decrypts a ciphertext ctxt given the receivers private key, and senders\n public key\n \"\"\"\n if len(key) != crypto_secretbox_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_secretbox_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n pad = b'\\x00' * crypto_secretbox_BOXZEROBYTES + ctxt\n msg = ctypes.create_string_buffer(len(pad))\n ret = nacl.crypto_secretbox_open(\n msg,\n pad,\n ctypes.c_ulonglong(len(pad)),\n nonce,\n key)\n if ret:\n raise ValueError('Failed to decrypt message')\n return msg.raw[crypto_secretbox_ZEROBYTES:]\n\n# Authenticated Symmetric Encryption improved version\n\n\ndef crypto_secretbox_easy(cmessage, nonce, key):\n if len(key) != crypto_secretbox_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_secretbox_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n \n ctxt = ctypes.create_string_buffer(crypto_secretbox_MACBYTES + len(cmessage))\n ret = nacl.crypto_secretbox_easy(ctxt, cmessage, ctypes.c_ulonglong(len(cmessage)), nonce, key)\n if ret:\n raise ValueError('Failed to encrypt message')\n return ctxt.raw[0:]\n\ndef crypto_secretbox_open_easy(ctxt, nonce, key):\n\n if len(key) != crypto_secretbox_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_secretbox_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n msg = ctypes.create_string_buffer(len(ctxt))\n ret = nacl.crypto_secretbox_open_easy(msg, ctxt, ctypes.c_ulonglong(len(ctxt)), nonce, key)\n if ret:\n raise ValueError('Failed to decrypt message')\n return msg.raw[0:len(ctxt) - crypto_secretbox_MACBYTES] \n\n# Authenticated Symmetric Encryption with Additional Data\n\n\ndef crypto_aead_aes256gcm_encrypt(message, aad, nonce, key):\n \"\"\"Encrypts and authenticates a message with public additional data using the given secret key, and nonce\n\n Args:\n message (bytes): a message to encrypt\n aad (bytes): additional public data to authenticate\n nonce (bytes): nonce, does not have to be confidential must be\n `crypto_aead_aes256gcm_NPUBBYTES` in length\n key (bytes): secret key, must be `crypto_aead_aes256gcm_KEYBYTES` in\n length\n\n Returns:\n bytes: the ciphertext\n\n Raises:\n ValueError: if arguments' length is wrong or the operation has failed.\n \"\"\"\n if not HAS_AEAD_AES256GCM:\n raise ValueError('Underlying Sodium library does not support AES256-GCM AEAD')\n\n if len(key) != crypto_aead_aes256gcm_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_aes256gcm_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(message) + crypto_aead_aes256gcm_ABYTES\n clen = ctypes.c_ulonglong()\n c = ctypes.create_string_buffer(length)\n ret = nacl.crypto_aead_aes256gcm_encrypt(\n c, ctypes.pointer(clen),\n message, ctypes.c_ulonglong(len(message)),\n aad, ctypes.c_ulonglong(len(aad)),\n None,\n nonce, key)\n if ret:\n raise ValueError('Failed to encrypt message')\n return c.raw\n\n\ndef crypto_aead_chacha20poly1305_ietf_encrypt(message, aad, nonce, key):\n \"\"\"Encrypts and authenticates a message with public additional data using the given secret key, and nonce\n\n Args:\n message (bytes): a message to encrypt\n aad (bytes): additional public data to authenticate\n nonce (bytes): nonce, does not have to be confidential must be\n `crypto_aead_chacha20poly1305_ietf_NPUBBYTES` in length\n key (bytes): secret key, must be `crypto_aead_chacha20poly1305_ietf_KEYBYTES` in\n length\n\n Returns:\n bytes: the ciphertext\n\n Raises:\n ValueError: if arguments' length is wrong or the operation has failed.\n \"\"\"\n if not HAS_AEAD_CHACHA20POLY1305_IETF:\n raise ValueError('Underlying Sodium library does not support IETF variant of ChaCha20Poly1305 AEAD')\n\n if len(key) != crypto_aead_chacha20poly1305_ietf_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_chacha20poly1305_ietf_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(message) + crypto_aead_chacha20poly1305_ietf_ABYTES\n clen = ctypes.c_ulonglong()\n c = ctypes.create_string_buffer(length)\n ret = nacl.crypto_aead_chacha20poly1305_ietf_encrypt(\n c, ctypes.pointer(clen),\n message, ctypes.c_ulonglong(len(message)),\n aad, ctypes.c_ulonglong(len(aad)),\n None,\n nonce, key)\n if ret:\n raise ValueError('Failed to encrypt message')\n return c.raw\n\n\ndef crypto_aead_xchacha20poly1305_ietf_encrypt(message, aad, nonce, key):\n \"\"\"Encrypts and authenticates a message with public additional data using the given secret key, and nonce\n\n Args:\n message (bytes): a message to encrypt\n aad (bytes): additional public data to authenticate\n nonce (bytes): nonce, does not have to be confidential must be\n `crypto_aead_xchacha20poly1305_ietf_NPUBBYTES` in length\n key (bytes): secret key, must be `crypto_aead_chacha20poly1305_ietf_KEYBYTES` in\n length\n\n Returns:\n bytes: the ciphertext\n\n Raises:\n ValueError: if arguments' length is wrong or the operation has failed.\n \"\"\"\n if not HAS_AEAD_XCHACHA20POLY1305_IETF:\n raise ValueError('Underlying Sodium library does not support IETF variant of ChaCha20Poly1305 AEAD')\n\n if len(key) != crypto_aead_xchacha20poly1305_ietf_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_xchacha20poly1305_ietf_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(message) + crypto_aead_xchacha20poly1305_ietf_ABYTES\n clen = ctypes.c_ulonglong()\n c = ctypes.create_string_buffer(length)\n ret = nacl.crypto_aead_xchacha20poly1305_ietf_encrypt(\n c, ctypes.pointer(clen),\n message, ctypes.c_ulonglong(len(message)),\n aad, ctypes.c_ulonglong(len(aad)),\n None,\n nonce, key)\n if ret:\n raise ValueError('Failed to encrypt message')\n return c.raw\n\n\ndef crypto_aead_aes256gcm_decrypt(ctxt, aad, nonce, key):\n \"\"\"\n Decrypts a ciphertext ctxt given the key, nonce, and aad. If the aad\n or ciphertext were altered then the decryption will fail.\n \"\"\"\n if not HAS_AEAD_AES256GCM:\n raise ValueError('Underlying Sodium library does not support AES256-GCM AEAD')\n\n if len(key) != crypto_aead_aes256gcm_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_aes256gcm_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(ctxt)-crypto_aead_aes256gcm_ABYTES\n mlen = ctypes.c_ulonglong()\n m = ctypes.create_string_buffer(length)\n\n ret = nacl.crypto_aead_aes256gcm_decrypt(\n m, ctypes.byref(mlen),\n None,\n ctxt, ctypes.c_ulonglong(len(ctxt)),\n aad, ctypes.c_ulonglong(len(aad)),\n nonce, key)\n if ret:\n raise ValueError('Failed to decrypt message')\n return m.raw\n\n\ndef crypto_aead_chacha20poly1305_ietf_decrypt(ctxt, aad, nonce, key):\n \"\"\"\n Decrypts a ciphertext ctxt given the key, nonce, and aad. If the aad\n or ciphertext were altered then the decryption will fail.\n \"\"\"\n if not HAS_AEAD_CHACHA20POLY1305_IETF:\n raise ValueError('Underlying Sodium library does not support IETF variant of ChaCha20Poly1305 AEAD')\n\n if len(key) != crypto_aead_chacha20poly1305_ietf_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_chacha20poly1305_ietf_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(ctxt)-crypto_aead_chacha20poly1305_ietf_ABYTES\n mlen = ctypes.c_ulonglong()\n m = ctypes.create_string_buffer(length)\n\n ret = nacl.crypto_aead_chacha20poly1305_ietf_decrypt(\n m, ctypes.byref(mlen),\n None,\n ctxt, ctypes.c_ulonglong(len(ctxt)),\n aad, ctypes.c_ulonglong(len(aad)),\n nonce, key)\n if ret:\n raise ValueError('Failed to decrypt message')\n return m.raw\n\n\ndef crypto_aead_xchacha20poly1305_ietf_decrypt(ctxt, aad, nonce, key):\n \"\"\"\n Decrypts a ciphertext ctxt given the key, nonce, and aad. If the aad\n or ciphertext were altered then the decryption will fail.\n \"\"\"\n if not HAS_AEAD_CHACHA20POLY1305_IETF:\n raise ValueError('Underlying Sodium library does not support IETF variant of ChaCha20Poly1305 AEAD')\n\n if len(key) != crypto_aead_xchacha20poly1305_ietf_KEYBYTES:\n raise ValueError('Invalid key')\n\n if len(nonce) != crypto_aead_xchacha20poly1305_ietf_NPUBBYTES:\n raise ValueError('Invalid nonce')\n\n length = len(ctxt)-crypto_aead_xchacha20poly1305_ietf_ABYTES\n mlen = ctypes.c_ulonglong()\n m = ctypes.create_string_buffer(length)\n\n ret = nacl.crypto_aead_xchacha20poly1305_ietf_decrypt(\n m, ctypes.byref(mlen),\n None,\n ctxt, ctypes.c_ulonglong(len(ctxt)),\n aad, ctypes.c_ulonglong(len(aad)),\n nonce, key)\n if ret:\n raise ValueError('Failed to decrypt message')\n return m.raw\n\n# Symmetric Encryption\n\n\ndef crypto_stream(slen, nonce, key):\n '''\n Generates a stream using the given secret key and nonce\n '''\n if len(key) != crypto_stream_KEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_stream_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n stream = ctypes.create_string_buffer(slen)\n ret = nacl.crypto_stream(stream, ctypes.c_ulonglong(slen), nonce, key)\n if ret:\n raise ValueError('Failed to init stream')\n return stream.raw\n\n\ndef crypto_stream_xor(msg, nonce, key):\n '''\n Encrypts the given message using the given secret key and nonce\n\n The crypto_stream_xor function guarantees that the ciphertext is the\n plaintext (xor) the output of crypto_stream. Consequently\n crypto_stream_xor can also be used to decrypt\n '''\n if len(key) != crypto_stream_KEYBYTES:\n raise ValueError('Invalid secret key')\n if len(nonce) != crypto_stream_NONCEBYTES:\n raise ValueError('Invalid nonce')\n\n stream = ctypes.create_string_buffer(len(msg))\n ret = nacl.crypto_stream_xor(\n stream,\n msg,\n ctypes.c_ulonglong(len(msg)),\n nonce,\n key)\n if ret:\n raise ValueError('Failed to init stream')\n return stream.raw\n\n\n# Authentication\n\n\ndef crypto_auth(msg, key):\n '''\n Constructs a one time authentication token for the given message msg\n using a given secret key\n '''\n if len(key) != crypto_auth_KEYBYTES:\n raise ValueError('Invalid secret key')\n\n tok = ctypes.create_string_buffer(crypto_auth_BYTES)\n ret = nacl.crypto_auth(tok, msg, ctypes.c_ulonglong(len(msg)), key)\n if ret:\n raise ValueError('Failed to auth msg')\n return tok.raw[:crypto_auth_BYTES]\n\n\ndef crypto_auth_verify(tok, msg, key):\n '''\n Verifies that the given authentication token is correct for the given\n message and key\n '''\n if len(key) != crypto_auth_KEYBYTES:\n raise ValueError('Invalid secret key')\n if len(tok) != crypto_auth_BYTES:\n raise ValueError('Invalid authenticator')\n\n ret = nacl.crypto_auth_verify(tok, msg, ctypes.c_ulonglong(len(msg)), key)\n if ret:\n raise ValueError('Failed to auth msg')\n return msg\n\n# One time authentication\n\n\ndef crypto_onetimeauth_primitive():\n \"\"\"\n Return the onetimeauth underlying primitive\n\n Returns:\n str: always ``poly1305``\n \"\"\"\n func = nacl.crypto_onetimeauth_primitive\n func.restype = ctypes.c_char_p\n return func().decode()\n\n\ndef crypto_onetimeauth(message, key):\n \"\"\"\n Constructs a one time authentication token for the given message using\n a given secret key\n\n Args:\n message (bytes): message to authenticate.\n key (bytes): secret key - must be of crypto_onetimeauth_KEYBYTES length.\n\n Returns:\n bytes: an authenticator, of crypto_onetimeauth_BYTES length.\n\n Raises:\n ValueError: if arguments' length is wrong.\n \"\"\"\n if len(key) != crypto_onetimeauth_KEYBYTES:\n raise ValueError('Invalid secret key')\n\n tok = ctypes.create_string_buffer(crypto_onetimeauth_BYTES)\n # cannot fail\n _ = nacl.crypto_onetimeauth(\n tok, message, ctypes.c_ulonglong(len(message)), key)\n\n return tok.raw[:crypto_onetimeauth_BYTES]\n\n\ndef crypto_onetimeauth_verify(token, message, key):\n \"\"\"\n Verifies, in constant time, that ``token`` is a correct authenticator for\n the message using the secret key.\n\n Args:\n token (bytes): an authenticator of crypto_onetimeauth_BYTES length.\n message (bytes): The message to authenticate.\n key: key (bytes): secret key - must be of crypto_onetimeauth_KEYBYTES\n length.\n\n Returns:\n bytes: secret key - must be of crypto_onetimeauth_KEYBYTES length.\n\n Raises:\n ValueError: if arguments' length is wrong or verification has failed.\n \"\"\"\n if len(key) != crypto_onetimeauth_KEYBYTES:\n raise ValueError('Invalid secret key')\n if len(token) != crypto_onetimeauth_BYTES:\n raise ValueError('Invalid authenticator')\n\n ret = nacl.crypto_onetimeauth_verify(\n token, message, ctypes.c_ulonglong(len(message)), key)\n if ret:\n raise ValueError('Failed to auth message')\n return message\n\n# Hashing\n\n\ndef crypto_hash(msg):\n '''\n Compute a hash of the given message\n '''\n hbuf = ctypes.create_string_buffer(crypto_hash_BYTES)\n nacl.crypto_hash(hbuf, msg, ctypes.c_ulonglong(len(msg)))\n return hbuf.raw\n\n\ndef crypto_hash_sha256(msg):\n '''\n Compute the sha256 hash of the given message\n '''\n hbuf = ctypes.create_string_buffer(crypto_hash_sha256_BYTES)\n nacl.crypto_hash_sha256(hbuf, msg, ctypes.c_ulonglong(len(msg)))\n return hbuf.raw\n\n\ndef crypto_hash_sha512(msg):\n '''\n Compute the sha512 hash of the given message\n '''\n hbuf = ctypes.create_string_buffer(crypto_hash_sha512_BYTES)\n nacl.crypto_hash_sha512(hbuf, msg, ctypes.c_ulonglong(len(msg)))\n return hbuf.raw\n\n# Generic Hash\n\n\ndef crypto_generichash(msg, key=None):\n '''\n Compute the blake2 hash of the given message with a given key\n '''\n hbuf = ctypes.create_string_buffer(crypto_generichash_BYTES)\n if key:\n key_len = len(key)\n else:\n key_len = 0\n nacl.crypto_generichash(\n hbuf,\n ctypes.c_size_t(len(hbuf)),\n msg,\n ctypes.c_ulonglong(len(msg)),\n key,\n ctypes.c_size_t(key_len))\n return hbuf.raw\n\n\n# String cmp\n\n\ndef crypto_verify_16(string1, string2):\n '''\n Compares the first crypto_verify_16_BYTES of the given strings\n\n The time taken by the function is independent of the contents of string1\n and string2. In contrast, the standard C comparison function\n memcmp(string1,string2,16) takes time that is dependent on the longest\n matching prefix of string1 and string2. This often allows for easy\n timing attacks.\n '''\n a, b, c = (len(string1) >= 16), (len(string2) >= 16), (not nacl.crypto_verify_16(string1, string2))\n return a & b & c\n\n\ndef crypto_verify_32(string1, string2):\n '''\n Compares the first crypto_verify_32_BYTES of the given strings\n\n The time taken by the function is independent of the contents of string1\n and string2. In contrast, the standard C comparison function\n memcmp(string1,string2,32) takes time that is dependent on the longest\n matching prefix of string1 and string2. This often allows for easy\n timing attacks.\n '''\n a, b, c = (len(string1) >= 32), (len(string2) >= 32), (not nacl.crypto_verify_32(string1, string2))\n return a & b & c\n\n\ndef crypto_verify_64(string1, string2):\n '''\n Compares the first crypto_verify_64_BYTES of the given strings\n\n The time taken by the function is independent of the contents of string1\n and string2. In contrast, the standard C comparison function\n memcmp(string1,string2,64) takes time that is dependent on the longest\n matching prefix of string1 and string2. This often allows for easy\n timing attacks.\n '''\n a, b, c = (len(string1) >= 64), (len(string2) >= 64), (not nacl.crypto_verify_64(string1, string2))\n return a & b & c\n\n\ndef bytes_eq(a, b):\n '''\n Compares two byte instances with one another. If `a` and `b` have\n different lengths, return `False` immediately. Otherwise `a` and `b`\n will be compared in constant time.\n\n Return `True` in case `a` and `b` are equal. Otherwise `False`.\n\n Raises :exc:`TypeError` in case `a` and `b` are not both of the type\n :class:`bytes`.\n '''\n if not isinstance(a, bytes) or not isinstance(b, bytes):\n raise TypeError('Both arguments must be bytes.')\n\n len_a = len(a)\n len_b = len(b)\n if len_a != len_b:\n return False\n\n return nacl.sodium_memcmp(a, b, len_a) == 0\n\n# Random byte generation\n\n\ndef randombytes(size):\n '''\n Return a string of random bytes of the given size\n '''\n buf = ctypes.create_string_buffer(size)\n nacl.randombytes(buf, ctypes.c_ulonglong(size))\n return buf.raw\n\n\ndef randombytes_buf(size):\n '''\n Return a string of random bytes of the given size\n '''\n size = int(size)\n buf = ctypes.create_string_buffer(size)\n nacl.randombytes_buf(buf, size)\n return buf.raw\n\ndef randombytes_buf_deterministic(size, seed):\n '''\n Returns a string of random byles of the given size for a given seed. \n For a given seed, this function will always output the same sequence. \n Size can be up to 2^70 (256 GB).\n '''\n\n if not HAS_RAND_SEED:\n raise ValueError('Underlying Sodium library does not support randombytes_seedbytes')\n if len(seed) != randombytes_SEEDBYTES:\n raise ValueError('Invalid key seed')\n\n size = int(size)\n buf = ctypes.create_string_buffer(size)\n nacl.randombytes_buf_deterministic(buf, size, seed)\n return buf.raw \n\ndef randombytes_close():\n '''\n Close the file descriptor or the handle for the cryptographic service\n provider\n '''\n nacl.randombytes_close()\n\n\ndef randombytes_random():\n '''\n Return a random 32-bit unsigned value\n '''\n return nacl.randombytes_random()\n\n\ndef randombytes_stir():\n '''\n Generate a new key for the pseudorandom number generator\n\n The file descriptor for the entropy source is kept open, so that the\n generator can be reseeded even in a chroot() jail.\n '''\n nacl.randombytes_stir()\n\n\ndef randombytes_uniform(upper_bound):\n '''\n Return a value between 0 and upper_bound using a uniform distribution\n '''\n return nacl.randombytes_uniform(upper_bound)\n\n# Key derivation API \n\ndef crypto_kdf_keygen():\n '''\n Returns a string of random bytes to generate a master key\n '''\n if not HAS_CRYPT_KDF:\n raise ValueError('Underlying Sodium library does not support crypto_kdf_keybytes')\n size = crypto_kdf_KEYBYTES\n buf = ctypes.create_string_buffer(size)\n nacl.crypto_kdf_keygen(buf)\n return buf.raw \n\ndef crypto_kdf_derive_from_key(subkey_size, subkey_id, context, master_key):\n '''\n Returns a subkey generated from a master key for a given subkey_id. \n For a given subkey_id, the subkey will always be the same string.\n '''\n size = int(subkey_size)\n buf = ctypes.create_string_buffer(size)\n nacl.crypto_kdf_derive_from_key(buf, subkey_size, ctypes.c_ulonglong(subkey_id), context, master_key)\n return buf.raw\n\n# Key Exchange API\n\ndef crypto_kx_keypair():\n '''\n Generate and return a new keypair\n '''\n if not HAS_CRYPT_KX:\n raise ValueError('Underlying Sodium library does not support crypto_kx')\n pk = ctypes.create_string_buffer(crypto_kx_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_kx_SECRETKEYBYTES)\n nacl.crypto_kx_keypair(pk, sk)\n return pk.raw, sk.raw\n\ndef crypto_kx_seed_keypair(seed):\n '''\n Generate and return a keypair from a key seed\n '''\n if not HAS_CRYPT_KX:\n raise ValueError('Underlying Sodium library does not support crypto_kx')\n\n if len(seed) != crypto_kx_SEEDBYTES:\n raise ValueError('Invalid key seed')\n pk = ctypes.create_string_buffer(crypto_kx_PUBLICKEYBYTES)\n sk = ctypes.create_string_buffer(crypto_kx_SECRETKEYBYTES)\n nacl.crypto_kx_seed_keypair(pk, sk, seed)\n return pk.raw, sk.raw\n\ndef crypto_kx_client_session_keys(client_pk, client_sk, server_pk):\n '''\n Computes a pair of shared keys (rx and tx) using the client's public key client_pk, \n the client's secret key client_sk and the server's public key server_pk.\n Status returns 0 on success, or -1 if the server's public key is not acceptable.\n '''\n if not HAS_CRYPT_KX:\n raise ValueError('Underlying Sodium library does not support crypto_kx')\n\n rx = ctypes.create_string_buffer(crypto_kx_SESSIONKEYBYTES)\n tx = ctypes.create_string_buffer(crypto_kx_SESSIONKEYBYTES)\n status = nacl.crypto_kx_client_session_keys(rx, tx, client_pk, client_sk, server_pk)\n return rx.raw, tx.raw, status\n\ndef crypto_kx_server_session_keys(server_pk, server_sk, client_pk):\n '''\n Computes a pair of shared keys (rx and tx) using the server's public key server_pk, \n the server's secret key server_sk and the client's public key client_pk.\n Status returns 0 on success, or -1 if the client's public key is not acceptable.\n '''\n if not HAS_CRYPT_KX:\n raise ValueError('Underlying Sodium library does not support crypto_kx')\n\n rx = ctypes.create_string_buffer(crypto_kx_SESSIONKEYBYTES)\n tx = ctypes.create_string_buffer(crypto_kx_SESSIONKEYBYTES)\n status = nacl.crypto_kx_server_session_keys(rx, tx, server_pk, server_sk, client_pk)\n return rx.raw, tx.raw, status\n\n\n\n# Utility functions\n\ndef sodium_library_version_major():\n '''\n Return the major version number\n '''\n return nacl.sodium_library_version_major()\n\n\ndef sodium_library_version_minor():\n '''\n Return the minor version number\n '''\n return nacl.sodium_library_version_minor()\n\n\ndef sodium_version_string():\n '''\n Return the version string\n '''\n func = nacl.sodium_version_string\n func.restype = ctypes.c_char_p\n return func()\n\n\ndef crypto_sign_ed25519_pk_to_curve25519(ed25519_pk):\n '''\n Convert an Ed25519 public key to a Curve25519 public key\n '''\n if len(ed25519_pk) != crypto_sign_ed25519_PUBLICKEYBYTES:\n raise ValueError('Invalid public key')\n\n curve25519_pk = ctypes.create_string_buffer(crypto_scalarmult_curve25519_BYTES)\n ret = nacl.crypto_sign_ed25519_pk_to_curve25519(curve25519_pk, ed25519_pk)\n if ret:\n raise CryptError('Failed to generate Curve25519 public key')\n return curve25519_pk.raw\n\n\ndef crypto_sign_ed25519_sk_to_curve25519(ed25519_sk):\n '''\n Convert an Ed25519 secret key to a Curve25519 secret key\n '''\n if len(ed25519_sk) != crypto_sign_ed25519_SECRETKEYBYTES:\n raise ValueError('Invalid secret key')\n\n curve25519_sk = ctypes.create_string_buffer(crypto_scalarmult_curve25519_BYTES)\n ret = nacl.crypto_sign_ed25519_sk_to_curve25519(curve25519_sk, ed25519_sk)\n if ret:\n raise CryptError('Failed to generate Curve25519 secret key')\n return curve25519_sk.raw\n\n\n","sub_path":"libnacl/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":45854,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"48363167","text":"\"\"\"Apipu\"\"\"\ndef main(text):\n \"\"\"Main function\"\"\"\n element = [\"Normal\", \"Fire\", \"Water\", \"Grass\", \"Electric\", \"Ice\", \"Fighting\", \"Poison\", \\\n \"Ground\", \"Flying\", \"Psychic\", \"Bug\", \"Rock\", \"Ghost\", \"Dragon\", \"Dark\", \"Steel\", \"Fairy\"]\n dic = {\"Normal\":\"6\", \"Fire\":\"2,8,12\", \"Water\":\"3,4\", \"Grass\":\"1,5,7,9,11\", \"Electric\":\"8\", \\\n \"Ice\":\"1,6,12,16\", \"Fighting\":\"9,10,17\", \"Poison\":\"8,10\", \"Ground\":\"2,3,5\", \"Flying\":\"4,5,12\",\\\n \"Psychic\":\"11,13,15\", \"Bug\":\"1,9,12\", \"Rock\":\"2,3,6,8,16\", \"Ghost\":\"13,15\", \\\n \"Dragon\":\"5,14,17\", \"Dark\":\"6,11,17\", \"Steel\":\"1,6,8\", \"Fairy\":\"7,16\"}\n if text not in element:\n print(\"Not Found\")\n else:\n weak = dic[text].split(\",\")\n weak = [int(i) for i in weak]\n for i in weak:\n print(element[i])\nmain(input())\n","sub_path":"Python9/Elements Battle V1.py","file_name":"Elements Battle V1.py","file_ext":"py","file_size_in_byte":803,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"208414098","text":"# -*- coding: utf-8 -*-\nimport paho.mqtt.client as mqtt\n\ndef on_connect(self, client, userdata, rc):\n print(\"Connected with result code \"+str(rc))\n self.subscribe(\"Topico/Lampada\", 0)\n\ndef on_message(client, userdata, msg):\n print(\"Topic: \"+ msg.topic + \"Message: \"+str(msg.payload))\n if msg.payload == \"Ligar\":\n print(\"Lampada ligada!\")\n\n\n#def on_connect(self, mosq, obj, rc):\n# self.subscribe(\"hello/world\", 0)\n\n\nclient = mqtt.Client()\n\nclient.on_connect = on_connect\nclient.on_message = on_message\n\nclient.username_pw_set(\"rcjwrmig\",\"luNqlsda3Ma_\")\nclient.connect(\"m13.cloudmqtt.com\", 12389, 60) #test.mosquitto.org\n\nclient.loop_forever()\n","sub_path":"subscribe-cloudmqtt.py","file_name":"subscribe-cloudmqtt.py","file_ext":"py","file_size_in_byte":668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"11377552","text":"'''\nThis script is written 4 anyalyzing KBNUCH's Inpatients' Data based on 'DRGNO by the order of payment'.\n\nWritten Date: 2019.12.17.\nWritten By: Peter JH Park\n\n'''\n\n### Import modules in needs\n\nimport os, sys, csv\nimport pandas as pd\nfrom pandas import DataFrame\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport datetime, time\nimport re\nfrom itertools import product\nimport math\n\nprint(\"\\n Current Working Directory is: \", os.getcwd())\n\n### READ Files & Check\n\nKBNUCHIn = pd.read_csv(\"./KBNUCH/KBNUCHInP_R4A.csv\", encoding=\"utf-8\")\nDiagCN = pd.read_csv(\"./master_Dcode&Dname.csv\", encoding=\"utf-8\")\nSurCN = pd.read_csv(\"./master_Scode&Sname.csv\", encoding=\"utf-8\")\nDRGCNmain5 = pd.read_csv(\"./SevCri_5digit(DRGname).csv\", encoding=\"utf-8\")\nDRGCNsub4 = pd.read_csv(\"./SevCri_4digit(DRGname).csv\", encoding=\"utf-8\")\n\n\nDiagCN.drop(['Dname(ENG)'], axis=1, inplace=True)\nDiagCN.rename(columns={'Dcode' : 'D_Code', 'Dname(KOR)' :'D_Name'}, inplace=True)\nDiagCN=DiagCN.groupby(by='D_Code', as_index=False).agg({'D_Name': lambda a: a.value_counts().index[0]})\n\nSurCN = SurCN.rename(columns={'Scode':'Sur_Code', 'Sname':'Sur_Name'})\nSurCN=SurCN.groupby(by='Sur_Code', as_index=False).agg({'Sur_Name': lambda a: a.value_counts().index[0]})\n\nDRGCNmain5=DRGCNmain5.groupby(by='DRGNO', as_index=False).agg({'DRGname': lambda a: a.value_counts().index[0]})\nDRGCNsub4=DRGCNsub4.groupby(by='DRGNO', as_index=False).agg({'DRGname': lambda a: a.value_counts().index[0]})\n\nprint(KBNUCHIn.info())\nprint(KBNUCHIn.columns)\n\n\n############################# Analyzing #############################\n\n\n## RAW DATA based on DRGNO by the order of payment: append rows\nKBNUCHIn4SevPbase = KBNUCHIn.copy()\n#KBNUCHIn4SevPbase.DRGNO = KBNUCHIn4SevPbase.DRGNO.str.split('/')\n#KBNUCHIn4SevPbase.Severity = KBNUCHIn4SevPbase.Severity.str.split('/')\n#KBNUCHIn4SevPbase = KBNUCHIn4SevPbase.apply(pd.Series.explode).reset_index(drop=True)\nprint(KBNUCHIn4SevPbase.info())\nprint(KBNUCHIn4SevPbase.columns)\n'''\nKBNUCHIn4SevPbase['Severity'] = KBNUCHIn4SevPbase['Severity'].map({'Severe' : 4, 'Normal' : 3, 'Simple' : 2, 'SortError' : 1})\n\nKBNUCHIn4SevPbasesub = KBNUCHIn4SevPbase.copy()\nKBNUCHIn4SevPbase = KBNUCHIn4SevPbase.groupby(['PT_No'], as_index= False)['Severity'].agg(lambda x: x.max())\nKBNUCHIn4SevPbase= KBNUCHIn4SevPbase.merge(KBNUCHIn4SevPbasesub, on=['PT_No', 'Severity'], how='left')\nKBNUCHIn4SevPbase.drop_duplicates(subset =['PT_No', 'Severity'], inplace = True)\nKBNUCHIn4SevPbase.reset_index(drop=True, inplace=True)\n\nKBNUCHIn4SevPbase['Severity'] = KBNUCHIn4SevPbase['Severity'].map({4:'Severe', 3:'Normal', 2:'Simple', 1:'SortError'})\n'''\n\nDRGCNmain5.rename(columns={DRGCNmain5.columns[0] : 'DRGNO01', DRGCNmain5.columns[1] : 'DRGName01'}, inplace=True)\nKBNUCHIn4SevPbase['DRGNO01'] = KBNUCHIn4SevPbase['DRGNO'].astype(str).str[0:5]\nKBNUCHIn4SevPbase['DRGNO01'].fillna('NoDRG', inplace=True)\n\nDRGCNsub4.rename(columns={DRGCNsub4.columns[0] : 'DRGNO02', DRGCNsub4.columns[1] : 'DRGName02'}, inplace=True)\nKBNUCHIn4SevPbase['DRGNO02'] = KBNUCHIn4SevPbase['DRGNO'].astype(str).str[0:4]\nKBNUCHIn4SevPbase['DRGNO02'].fillna('NoDRG', inplace=True)\n\nKBNUCHIn4SevPbase = KBNUCHIn4SevPbase.merge(DRGCNmain5, on='DRGNO01', how='left').merge(DRGCNsub4, on='DRGNO02', how='left')\nKBNUCHIn4SevPbase.reset_index(drop=True, inplace=True)\n\nKBNUCHIn4SevPbase.loc[(KBNUCHIn4SevPbase['DRGName01'].isna() & KBNUCHIn4SevPbase['DRGName02'].notna()), 'DRGName01'] = KBNUCHIn4SevPbase['DRGName02']\nKBNUCHIn4SevPbase.rename(columns={'DRGName01' : 'DRGName'}, inplace=True)\ndel KBNUCHIn4SevPbase['DRGNO01']; del KBNUCHIn4SevPbase['DRGNO02']; del KBNUCHIn4SevPbase['DRGName02']\nKBNUCHIn4SevPbase.DRGName.fillna('NoDRG', inplace=True)\n\nKBNUCHIn4SevPbase.loc[(KBNUCHIn4SevPbase['DRGName'] == 'NoDRG'), 'DRGNO'] = 'NoDRG'\nKBNUCHIn4SevPbase = KBNUCHIn4SevPbase[KBNUCHIn4SevPbase.D_Code != 'NoDiag']\nprint(KBNUCHIn4SevPbase.info())\n\nKBNUCHIn4SevPbase['TotalPay'] = KBNUCHIn4SevPbase['Pay_InsSelf'] + KBNUCHIn4SevPbase['Pay_InsCorp'] + KBNUCHIn4SevPbase['Pay_NoIns'] + KBNUCHIn4SevPbase['Pay_Sel']\n\nbpay=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'TotalPay': pd.Series.mean})\nbpay = bpay[bpay.DRGNO != 'NoDRG']\nbpay.sort_values(by='TotalPay', ascending=False, inplace=True)\nbpay.reset_index(drop=True, inplace=True)\nbpay.index += 1\nbpay = bpay.rename_axis('Rank').reset_index()\nbpay['TotalPay'] = bpay['TotalPay'].astype(int)\nprint (bpay)\n\nbdname=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'DRGName': lambda a: a.value_counts().index[0]})\nbdname = bdname[bdname.DRGNO != 'NoDRG']\nbdname.reset_index(drop=True, inplace=True)\n#print (bdname)\n\n\nbsev=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'Severity': pd.Series.unique})\nbsev.reset_index(drop=True, inplace=True)\nprint (bsev)\n\n\n## 01. BASIC INFORMATION by Severity Payment\npnum=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'PT_No': pd.Series.nunique})\npnum.rename(columns={'PT_No':'P_Num'}, inplace=True)\npnum['P_Num'] = pnum['P_Num'].astype(int)\npnum.reset_index(drop=True, inplace=True)\nprint (pnum)\n\navage=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'Age': pd.Series.mean})\navage.rename(columns={'Age':'AvAge'}, inplace=True)\navage.AvAge = avage.AvAge.round(1)\navage.reset_index(drop=True, inplace=True)\nprint (avage)\n\navinprd=KBNUCHIn4SevPbase.groupby(by='DRGNO', as_index=False).agg({'In_Prd': pd.Series.mean})\navinprd.rename(columns={'In_Prd':'AvInPrd'}, inplace=True)\navinprd.AvInPrd = avinprd.AvInPrd.round(1)\navinprd.reset_index(drop=True, inplace=True)\nprint (avinprd)\n\nKBNUCHIn4SevPbase_Ins = KBNUCHIn4SevPbase.copy()\nKBNUCHIn4SevPbase_Ins['IndPaidExp'] = KBNUCHIn4SevPbase_Ins['Pay_InsSelf'] + KBNUCHIn4SevPbase_Ins['Pay_NoIns'] + KBNUCHIn4SevPbase_Ins['Pay_Sel']\ninspaid = KBNUCHIn4SevPbase_Ins.groupby(by='DRGNO', as_index=False).agg({'IndPaidExp' : pd.Series.mean, 'Pay_InsCorp' : pd.Series.mean})\ninspaid.rename(columns={'Pay_InsCorp':'CorpPaidExp'}, inplace=True)\ninspaid['IndPaidExp'] = inspaid['IndPaidExp'].astype(int)\ninspaid['CorpPaidExp'] = inspaid['CorpPaidExp'].astype(int)\ninspaid['IndPaidExp'] = inspaid.apply(lambda x: \"{:,}\".format(x['IndPaidExp']), axis=1)\ninspaid['CorpPaidExp'] = inspaid.apply(lambda x: \"{:,}\".format(x['CorpPaidExp']), axis=1)\ninspaid.reset_index(drop=True, inplace=True)\nprint (inspaid)\n\n\nKBNUCHInSevP_base = bpay.merge(bdname,on='DRGNO',how='left').merge(bsev,on='DRGNO',how='left').merge(pnum,on='DRGNO',how='left').merge(avage,on='DRGNO',how='left').merge(avinprd,on='DRGNO',how='left').merge(inspaid,on='DRGNO',how='left')\nKBNUCHInSevP_base = KBNUCHInSevP_base[['Rank', 'DRGNO', 'DRGName', 'Severity', 'TotalPay', 'IndPaidExp', 'CorpPaidExp', 'P_Num', 'AvAge', 'AvInPrd']]\nKBNUCHInSevP_base.reset_index(drop=True, inplace=True)\nprint(KBNUCHInSevP_base)\nprint(KBNUCHInSevP_base.columns)\n\n\nKBNUCHInSevP_base.Severity = KBNUCHInSevP_base.Severity.map({'Severe' : '전문', 'Normal' : '일반', 'Simple' : '단순', 'SortError' : '분류오류'})\n\n\nKBNUCHInSevP_baseSev = KBNUCHInSevP_base[KBNUCHInSevP_base.Severity == '전문']\nKBNUCHInSevP_baseSev = KBNUCHInSevP_baseSev.sort_values(by ='TotalPay', ascending = 0)\nKBNUCHInSevP_baseSev.drop(['Rank'], axis=1, inplace=True)\nKBNUCHInSevP_baseSev.reset_index(drop=True, inplace=True)\nKBNUCHInSevP_baseSev.index += 1\nKBNUCHInSevP_baseSev['TotalPay'] = KBNUCHInSevP_baseSev.apply(lambda x: \"{:,}\".format(x['TotalPay']), axis=1)\nKBNUCHInSevP_baseSev = KBNUCHInSevP_baseSev.rename_axis('Rank').reset_index()\n\n\nKBNUCHInSevP_baseNorm = KBNUCHInSevP_base[KBNUCHInSevP_base.Severity == '일반']\nKBNUCHInSevP_baseNorm = KBNUCHInSevP_baseNorm.sort_values(by ='TotalPay', ascending = 0)\nKBNUCHInSevP_baseNorm.drop(['Rank'], axis=1, inplace=True)\nKBNUCHInSevP_baseNorm.reset_index(drop=True, inplace=True)\nKBNUCHInSevP_baseNorm.index += 1\nKBNUCHInSevP_baseNorm['TotalPay'] = KBNUCHInSevP_baseNorm.apply(lambda x: \"{:,}\".format(x['TotalPay']), axis=1)\nKBNUCHInSevP_baseNorm = KBNUCHInSevP_baseNorm.rename_axis('Rank').reset_index()\n\n\nKBNUCHInSevP_baseSimp = KBNUCHInSevP_base[KBNUCHInSevP_base.Severity == '단순']\nKBNUCHInSevP_baseSimp = KBNUCHInSevP_baseSimp.sort_values(by ='TotalPay', ascending = 0)\nKBNUCHInSevP_baseSimp.drop(['Rank'], axis=1, inplace=True)\nKBNUCHInSevP_baseSimp.reset_index(drop=True, inplace=True)\nKBNUCHInSevP_baseSimp.index += 1\nKBNUCHInSevP_baseSimp['TotalPay'] = KBNUCHInSevP_baseSimp.apply(lambda x: \"{:,}\".format(x['TotalPay']), axis=1)\nKBNUCHInSevP_baseSimp = KBNUCHInSevP_baseSimp.rename_axis('Rank').reset_index()\n\n\nKBNUCHInSevP_base['TotalPay'] = KBNUCHInSevP_base.apply(lambda x: \"{:,}\".format(x['TotalPay']), axis=1)\n\n\nKBNUCHInSevP_base = KBNUCHInSevP_base.rename(columns={'Rank':'순위', 'DRGNO':'DRG번호', 'DRGName':'DRG명', 'Severity':'중증도', 'TotalPay':'총부담금(원)',\n 'IndPaidExp':'본인부담금(원)', 'CorpPaidExp':'공단부담금(원)', 'P_Num':'환자수(명)', 'AvAge':'평균연령(세)', 'AvInPrd':'평균재원기간(일)'})\nKBNUCHInSevP_baseSev = KBNUCHInSevP_baseSev.rename(columns={'Rank':'순위', 'DRGNO':'DRG번호', 'DRGName':'DRG명', 'Severity':'중증도', 'TotalPay':'총부담금(원)',\n 'IndPaidExp':'본인부담금(원)', 'CorpPaidExp':'공단부담금(원)', 'P_Num':'환자수(명)', 'AvAge':'평균연령(세)', 'AvInPrd':'평균재원기간(일)'})\nKBNUCHInSevP_baseNorm = KBNUCHInSevP_baseNorm.rename(columns={'Rank':'순위', 'DRGNO':'DRG번호', 'DRGName':'DRG명', 'Severity':'중증도', 'TotalPay':'총부담금(원)',\n 'IndPaidExp':'본인부담금(원)', 'CorpPaidExp':'공단부담금(원)', 'P_Num':'환자수(명)', 'AvAge':'평균연령(세)', 'AvInPrd':'평균재원기간(일)'})\nKBNUCHInSevP_baseSimp = KBNUCHInSevP_baseSimp.rename(columns={'Rank':'순위', 'DRGNO':'DRG번호', 'DRGName':'DRG명', 'Severity':'중증도', 'TotalPay':'총부담금(원)',\n 'IndPaidExp':'본인부담금(원)', 'CorpPaidExp':'공단부담금(원)', 'P_Num':'환자수(명)', 'AvAge':'평균연령(세)', 'AvInPrd':'평균재원기간(일)'})\n\nprint(KBNUCHInSevP_base.columns)\n\nKBNUCHInSevP_base50 = KBNUCHInSevP_base.loc[0:49, :]\nKBNUCHInSevP_base50Sev = KBNUCHInSevP_baseSev.loc[0:49, :]\nKBNUCHInSevP_base50Norm = KBNUCHInSevP_baseNorm.loc[0:49, :]\nKBNUCHInSevP_base50Simp = KBNUCHInSevP_baseSimp.loc[0:49, :]\n\n\n\n'''\n# No Executuon\n## 02. DIAGNOSIS & SURGERY INFORMATION by Severity Payment\nKBNUCHIn4SevPbase_Diag = KBNUCHIn4SevPbase.copy()\nKBNUCHIn4SevPbase_Diag.D_Code = KBNUCHIn4SevPbase_Diag.D_Code.str.split('/')\nKBNUCHIn4SevPbase_Diag.D_Name = KBNUCHIn4SevPbase_Diag.D_Name.str.split('/')\nKBNUCHIn4SevPbase_Diag = KBNUCHIn4SevPbase_Diag.apply(pd.Series.explode).reset_index(drop=True)\nKBNUCHIn4SevPbase_Diag = KBNUCHIn4SevPbase_Diag.drop_duplicates(['PT_No', 'In_Date', 'DRGNO', 'Severity', 'D_Code', 'D_Name'], keep='first') # Need D_Date after for proper drop\nKBNUCHIn4SevPbase_Diag = KBNUCHIn4SevPbase_Diag[KBNUCHIn4SevPbase_Diag.D_Code != 'NoDiag']\n\n##### another way to selcet most frequent value using groupby(Not Accurate in Certain Condition)#####\n#dcodename = KBNUCHIn4SevPbase_Diag.groupby(by='DRGNO', as_index=False).agg({'D_Code': lambda a: ' / '.join(pd.Series.mode(a)), 'D_Name': lambda b: ' / '.join(pd.Series.mode(b))})\n\n\ndcode = KBNUCHIn4SevPbase_Diag.groupby(by='DRGNO', as_index=False)['D_Code'].agg(lambda x:x.value_counts().index[0])\ndname = KBNUCHIn4SevPbase_Diag.groupby(by='D_Code', as_index=False).agg({'D_Name': lambda a: '/'.join(pd.unique(a))})\ndcode.reset_index(drop=True, inplace=True)\ndname.reset_index(drop=True, inplace=True)\ndcodename = dcode.merge(dname, on='D_Code', how='left')\ndcodename.reset_index(drop=True, inplace=True)\ndfreq = KBNUCHIn4SevPbase_Diag.groupby(by='DRGNO', as_index=False)['D_Code'].agg(lambda x: x.value_counts().head(1))\ndfreq.rename(columns={'D_Code':'Dfreq'}, inplace=True)\ndfreq['Dfreq'] = dfreq['Dfreq'].astype('int64')\ndfreq.reset_index(drop=True, inplace=True)\ndtot = KBNUCHIn4SevPbase_Diag.groupby(by='DRGNO', as_index=False)['D_Code'].agg(lambda x: x.count())\ndtot.rename(columns={'D_Code':'Dtot'}, inplace=True)\ndtot['Dtot'] = dtot['Dtot'].astype('int64')\ndtot.reset_index(drop=True, inplace=True)\ndcodenamefreq = dcodename.merge(dfreq,on='DRGNO',how='left').merge(dtot,on='DRGNO',how='left')\ndcodenamefreq.rename(columns={'D_Code':'Dcode', 'D_Name':'Dname'}, inplace=True)\ndcodenamefreq.reset_index(drop=True, inplace=True)\ndcodenamefreq['Dratio'] = (dcodenamefreq.Dfreq / dcodenamefreq.Dtot) * 100\ndcodenamefreq.Dratio = dcodenamefreq.Dratio.round(2)\ndcodenamefreq['Dratio'] = dcodenamefreq['Dratio'].astype(str) + ' %'\nwith pd.option_context('display.max_columns', None):\n print (dcodenamefreq)\n\n\nKBNUCHIn4SevPbase_Sur = KBNUCHIn4SevPbase.copy()\nKBNUCHIn4SevPbase_Sur.Sur_Date = KBNUCHIn4SevPbase_Sur.Sur_Date.str.split('/')\nKBNUCHIn4SevPbase_Sur.Sur_Code = KBNUCHIn4SevPbase_Sur.Sur_Code.str.split('/')\nKBNUCHIn4SevPbase_Sur.Sur_Name = KBNUCHIn4SevPbase_Sur.Sur_Name.str.split('/')\nKBNUCHIn4SevPbase_Sur = KBNUCHIn4SevPbase_Sur.apply(pd.Series.explode).reset_index(drop=True)\nKBNUCHIn4SevPbase_Sur = KBNUCHIn4SevPbase_Sur.drop_duplicates(['PT_No', 'In_Date', 'DRGNO', 'Severity', 'Sur_Date', 'Sur_Code', 'Sur_Name'], keep='first')\nKBNUCHIn4SevPbase_OnlySur = KBNUCHIn4SevPbase_Sur[KBNUCHIn4SevPbase_Sur.Sur_Code != 'NoSur']\nKBNUCHIn4SevPbase_OnlySur = KBNUCHIn4SevPbase_OnlySur[KBNUCHIn4SevPbase_OnlySur.Sur_Code != 'GroupPay_KBNUCH']\n\n##### another way to selcet most frequent value using groupby(Not Accurate in Certain Condition)#####\n#scodename = KBNUCHIn4SevPbase_Sur.groupby(by='DRGNO', as_index=False).agg({'Sur_Code': lambda a: ' / '.join(pd.Series.mode(a)), 'Sur_Name' : lambda b: ' / '.join(pd.Series.mode(b))})\n\n\nscode = KBNUCHIn4SevPbase_OnlySur.groupby(by='DRGNO', as_index=False)['Sur_Code'].agg(lambda x : x.value_counts().index[0])\nsname = KBNUCHIn4SevPbase_OnlySur.groupby(by='Sur_Code', as_index=False)['Sur_Name'].agg(lambda x : x.value_counts().index[0])\nscode.reset_index(drop=True, inplace=True)\nsname.reset_index(drop=True, inplace=True)\nscodename = scode.merge(sname, on='Sur_Code', how='left')\nscodename.reset_index(drop=True, inplace=True)\nsfreq = KBNUCHIn4SevPbase_OnlySur.groupby(by='DRGNO', as_index=False)['Sur_Code'].agg(lambda x : x.value_counts().head(1))\nsfreq.rename(columns={'Sur_Code':'Sfreq'}, inplace=True)\nsfreq['Sfreq'] = sfreq['Sfreq'].astype('int64')\nsfreq.reset_index(drop=True, inplace=True)\nstot = KBNUCHIn4SevPbase_OnlySur.groupby(by='DRGNO', as_index=False)['Sur_Code'].agg(lambda x : x.count())\nstot.rename(columns={'Sur_Code':'Stot'}, inplace=True)\nstot['Stot'] = stot['Stot'].astype('int64')\nstot.reset_index(drop=True, inplace=True)\nscodenamefreq = scodename.merge(sfreq,on='DRGNO',how='left').merge(stot,on='DRGNO',how='left')\nscodenamefreq.rename(columns={'Sur_Code':'Scode', 'Sur_Name':'Sname'}, inplace=True)\nscodenamefreq.reset_index(drop=True, inplace=True)\nscodenamefreq['Sratio'] = (scodenamefreq.Sfreq / scodenamefreq.Stot) * 100\nscodenamefreq.Sratio = scodenamefreq.Sratio.round(2)\nscodenamefreq['Sratio'] = scodenamefreq['Sratio'].astype(str) + ' %'\nwith pd.option_context('display.max_columns', None):\n print (scodenamefreq)\n\n\nKBNUCHInSevP_diagsur = bpay.merge(bsev,on='DRGNO',how='left').merge(dcodenamefreq,on='DRGNO',how='left').merge(scodenamefreq,on='DRGNO',how='left')\nKBNUCHInSevP_diagsur.drop(['TotalPay' ,'Dtot', 'Stot'], axis=1, inplace=True)\nKBNUCHInSevP_diagsur.reset_index(drop=True, inplace=True)\nKBNUCHInSevP_diagsur['Dcode'].fillna('NoDiag', inplace=True)\nKBNUCHInSevP_diagsur['Dname'].fillna('NoDiag', inplace=True)\nKBNUCHInSevP_diagsur['Dfreq'].fillna(0, inplace=True)\nKBNUCHInSevP_diagsur['Dratio'].fillna('0.0 %', inplace=True)\nKBNUCHInSevP_diagsur['Scode'].fillna('NoSur', inplace=True)\nKBNUCHInSevP_diagsur['Sname'].fillna('NoSur', inplace=True)\nKBNUCHInSevP_diagsur['Sfreq'].fillna(0, inplace=True)\nKBNUCHInSevP_diagsur['Sratio'].fillna('0.0 %', inplace=True)\n\nwith pd.option_context('display.max_columns', None):\n print (KBNUCHInSevP_diagsur)\n\nprint(KBNUCHInSevP_diagsur.columns)\n\nKBNUCHInSevP_diagsur50 = KBNUCHInSevP_diagsur.loc[0:49, :]\n'''\n\n\nKBNUCHInSevP_base50.to_csv('./KBNUCH/KBNUCH 입원환자 중증도 총부담금별 순위.csv', encoding='cp949', index=False)\nKBNUCHInSevP_base50Sev.to_csv('./KBNUCH/KBNUCH 입원환자 중증도 총부담금별 순위(전문상위50).csv', encoding='cp949', index=False)\nKBNUCHInSevP_base50Norm.to_csv('./KBNUCH/KBNUCH 입원환자 중증도 총부담금별 순위(일반상위50).csv', encoding='cp949', index=False)\nKBNUCHInSevP_base50Simp.to_csv('./KBNUCH/KBNUCH 입원환자 중증도 총부담금별 순위(단순상위50).csv', encoding='cp949', index=False)\n#KBNUCHInSevP_diagsur50.to_csv('./KBNUCH/KBNUCHInSevP_diagsur50.csv', index=False)\n\nKBNUCHInSevP_base.to_csv('./KBNUCH/(원)KBNUCH 입원환자 중증도 총부담금별 순위.csv', encoding='cp949', index=False)\nKBNUCHInSevP_baseSev.to_csv('./KBNUCH/(원)KBNUCH 입원환자 중증도 총부담금별 순위(전문상위50).csv', encoding='cp949', index=False)\nKBNUCHInSevP_baseNorm.to_csv('./KBNUCH/(원)KBNUCH 입원환자 중증도 총부담금별 순위(일반상위50).csv', encoding='cp949', index=False)\nKBNUCHInSevP_baseSimp.to_csv('./KBNUCH/(원)KBNUCH 입원환자 중증도 총부담금별 순위(단순상위50).csv', encoding='cp949', index=False)\n\n#KBNUCHInSevP_inst = pd.DataFrame(columns=['Rank' ,'DRGNO', 'Severity', 'A_Ratio', 'B_Ratio', 'C_Ratio', 'D_Ratio', 'E_Ratio', 'F_Ratio', 'G_Ratio', 'H_Ratio'])\n#KBNUCHInSevP_regn = pd.DataFrame(columns=['Rank' ,'DRGNO', 'Severity', 'fSeoul', 'fBusan', 'fDaegu', 'fGwangju', 'fDaejeon', 'fIncheon', 'fJeju', 'fSejong', 'fJeonnam', 'fJeonbuk',\n# 'fGyeongnam', 'fGyeongbuk', 'fChungnam', 'fChungbuk', 'fGangwon', 'fGyeonggi'])\n\n\n\n","sub_path":"08-5. InpAnlyzSevPaybase_KBNUCH.py","file_name":"08-5. InpAnlyzSevPaybase_KBNUCH.py","file_ext":"py","file_size_in_byte":18000,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"604861833","text":"##+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\n## Created by: Yaoyao Liu\n## Tianjin University\n## Email: liuyaoyao@tju.edu.cn\n## Copyright (c) 2019\n##\n## This source code is licensed under the MIT-style license found in the\n## LICENSE file in the root directory of this source tree\n##+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++\n\n\"\"\" The base model class. \"\"\"\nimport numpy as np\nimport sys\nimport tensorflow as tf\nfrom tensorflow.python.platform import flags\nfrom utils.misc import mse, softmaxloss, xent, resnet_conv_block, resnet_nob_conv_block, normalize\n\nFLAGS = flags.FLAGS\n\nclass Models:\n \"\"\"The class that contains the code for the basic resnet models and SS weights\"\"\"\n def __init__(self):\n # Set the dimension number for the input feature maps\n self.dim_input = FLAGS.img_size * FLAGS.img_size * 3\n # Set the dimension number for the outputs\n self.dim_output = FLAGS.way_num\n # Load base learning rates from FLAGS\n self.update_lr = FLAGS.base_lr\n # Load the pre-train phase class number from FLAGS\n self.pretrain_class_num = FLAGS.pretrain_class_num\n # Set the initial meta learning rate\n self.meta_lr = tf.placeholder_with_default(FLAGS.meta_lr, ())\n # Set the initial pre-train learning rate\n self.pretrain_lr = tf.placeholder_with_default(FLAGS.pre_lr, ())\n\n # Set the default objective functions for meta-train and pre-train\n self.loss_func = xent\n self.pretrain_loss_func = softmaxloss\n\n # Set the default channel number to 3\n self.channels = 3\n # Load the image size from FLAGS\n self.img_size = FLAGS.img_size\n\n def process_ss_weights(self, weights, ss_weights, label): \n \"\"\"The function to process the scaling operation\n Args:\n weights: the weights for the resnet.\n ss_weights: the weights for scaling and shifting operation.\n label: the label to indicate which layer we are operating.\n Return:\n The processed weights for the new resnet.\n \"\"\" \n [dim0, dim1] = weights[label].get_shape().as_list()[0:2]\n this_ss_weights = tf.tile(ss_weights[label], multiples=[dim0, dim1, 1, 1])\n return tf.multiply(weights[label], this_ss_weights)\n\n def forward_pretrain_resnet(self, inp, weights, reuse=False, scope=''):\n \"\"\"The function to forward the resnet during pre-train phase\n Args:\n inp: input feature maps.\n weights: input resnet weights.\n reuse: reuse the batch norm weights or not.\n scope: the label to indicate which layer we are processing.\n Return:\n The processed feature maps.\n \"\"\" \n inp = tf.reshape(inp, [-1, self.img_size, self.img_size, self.channels])\n net = self.pretrain_block_forward(inp, weights, 'block1', reuse, scope)\n net = self.pretrain_block_forward(net, weights, 'block2', reuse, scope)\n net = self.pretrain_block_forward(net, weights, 'block3', reuse, scope)\n net = self.pretrain_block_forward(net, weights, 'block4', reuse, scope)\n net = tf.nn.avg_pool(net, [1,5,5,1], [1,5,5,1], 'VALID')\n net = tf.reshape(net, [-1, np.prod([int(dim) for dim in net.get_shape()[1:]])])\n return net\n\n def forward_resnet(self, inp, weights, ss_weights, reuse=False, scope=''):\n \"\"\"The function to forward the resnet during meta-train phase\n Args:\n inp: input feature maps.\n weights: input resnet weights.\n ss_weights: input scaling weights.\n reuse: reuse the batch norm weights or not.\n scope: the label to indicate which layer we are processing.\n Return:\n The processed feature maps.\n \"\"\" \n inp = tf.reshape(inp, [-1, self.img_size, self.img_size, self.channels])\n net = self.block_forward(inp, weights, ss_weights, 'block1', reuse, scope)\n net = self.block_forward(net, weights, ss_weights, 'block2', reuse, scope)\n net = self.block_forward(net, weights, ss_weights, 'block3', reuse, scope)\n net = self.block_forward(net, weights, ss_weights, 'block4', reuse, scope)\n net = tf.nn.avg_pool(net, [1,5,5,1], [1,5,5,1], 'VALID')\n net = tf.reshape(net, [-1, np.prod([int(dim) for dim in net.get_shape()[1:]])])\n return net\n\n def forward_fc(self, inp, fc_weights):\n \"\"\"The function to forward the fc layer\n Args:\n inp: input feature maps.\n fc_weights: input fc weights.\n Return:\n The processed feature maps.\n \"\"\" \n net = tf.matmul(inp, fc_weights['w5']) + fc_weights['b5']\n return net\n\n def pretrain_block_forward(self, inp, weights, block, reuse, scope):\n \"\"\"The function to forward a resnet block during pre-train phase\n Args:\n inp: input feature maps.\n weights: input resnet weights.\n block: the string to indicate which block we are processing.\n reuse: reuse the batch norm weights or not.\n scope: the label to indicate which layer we are processing.\n Return:\n The processed feature maps.\n \"\"\" \n net = resnet_conv_block(inp, weights[block + '_conv1'], weights[block + '_bias1'], reuse, scope+block+'0')\n net = resnet_conv_block(net, weights[block + '_conv2'], weights[block + '_bias2'], reuse, scope+block+'1')\n net = resnet_conv_block(net, weights[block + '_conv3'], weights[block + '_bias3'], reuse, scope+block+'2')\n res = resnet_nob_conv_block(inp, weights[block + '_conv_res'], reuse, scope+block+'res')\n net = net + res\n net = tf.nn.max_pool(net, [1,2,2,1], [1,2,2,1], 'VALID')\n net = tf.nn.dropout(net, keep_prob=FLAGS.pretrain_dropout_keep)\n return net\n\n def block_forward(self, inp, weights, ss_weights, block, reuse, scope):\n \"\"\"The function to forward a resnet block during meta-train phase\n Args:\n inp: input feature maps.\n weights: input resnet weights.\n ss_weights: input scaling weights.\n block: the string to indicate which block we are processing.\n reuse: reuse the batch norm weights or not.\n scope: the label to indicate which layer we are processing.\n Return:\n The processed feature maps.\n \"\"\" \n net = resnet_conv_block(inp, self.process_ss_weights(weights, ss_weights, block + '_conv1'), \\\n ss_weights[block + '_bias1'], reuse, scope+block+'0')\n net = resnet_conv_block(net, self.process_ss_weights(weights, ss_weights, block + '_conv2'), \\\n ss_weights[block + '_bias2'], reuse, scope+block+'1')\n net = resnet_conv_block(net, self.process_ss_weights(weights, ss_weights, block + '_conv3'), \\\n ss_weights[block + '_bias3'], reuse, scope+block+'2')\n res = resnet_nob_conv_block(inp, weights[block + '_conv_res'], reuse, scope+block+'res')\n net = net + res\n net = tf.nn.max_pool(net, [1,2,2,1], [1,2,2,1], 'VALID')\n net = tf.nn.dropout(net, keep_prob=1)\n return net\n\n def construct_fc_weights(self):\n \"\"\"The function to construct fc weights.\n Return:\n The fc weights.\n \"\"\" \n dtype = tf.float32 \n fc_weights = {}\n fc_initializer = tf.contrib.layers.xavier_initializer(dtype=dtype)\n if FLAGS.phase=='pre':\n fc_weights['w5'] = tf.get_variable('fc_w5', [512, FLAGS.pretrain_class_num], initializer=fc_initializer)\n fc_weights['b5'] = tf.Variable(tf.zeros([FLAGS.pretrain_class_num]), name='fc_b5')\n else:\n fc_weights['w5'] = tf.get_variable('fc_w5', [512, self.dim_output], initializer=fc_initializer)\n fc_weights['b5'] = tf.Variable(tf.zeros([self.dim_output]), name='fc_b5')\n return fc_weights\n\n def construct_resnet_weights(self):\n \"\"\"The function to construct resnet weights.\n Return:\n The resnet weights.\n \"\"\" \n weights = {}\n dtype = tf.float32\n conv_initializer = tf.contrib.layers.xavier_initializer_conv2d(dtype=dtype)\n fc_initializer = tf.contrib.layers.xavier_initializer(dtype=dtype)\n weights = self.construct_residual_block_weights(weights, 3, 3, 64, conv_initializer, dtype, 'block1')\n weights = self.construct_residual_block_weights(weights, 3, 64, 128, conv_initializer, dtype, 'block2')\n weights = self.construct_residual_block_weights(weights, 3, 128, 256, conv_initializer, dtype, 'block3')\n weights = self.construct_residual_block_weights(weights, 3, 256, 512, conv_initializer, dtype, 'block4')\n weights['w5'] = tf.get_variable('w5', [512, FLAGS.pretrain_class_num], initializer=fc_initializer)\n weights['b5'] = tf.Variable(tf.zeros([FLAGS.pretrain_class_num]), name='b5')\n return weights\n\n def construct_residual_block_weights(self, weights, k, last_dim_hidden, dim_hidden, conv_initializer, dtype, scope='block0'):\n \"\"\"The function to construct one block of the resnet weights.\n Args:\n weights: the resnet weight list.\n k: the dimension number of the convolution kernel.\n last_dim_hidden: the hidden dimension number of last block.\n dim_hidden: the hidden dimension number of the block.\n conv_initializer: the convolution initializer.\n dtype: the dtype for numpy.\n scope: the label to indicate which block we are processing.\n Return:\n The resnet block weights.\n \"\"\" \n weights[scope + '_conv1'] = tf.get_variable(scope + '_conv1', [k, k, last_dim_hidden, dim_hidden], \\\n initializer=conv_initializer, dtype=dtype)\n weights[scope + '_bias1'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias1')\n weights[scope + '_conv2'] = tf.get_variable(scope + '_conv2', [k, k, dim_hidden, dim_hidden], \\\n initializer=conv_initializer, dtype=dtype)\n weights[scope + '_bias2'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias2')\n weights[scope + '_conv3'] = tf.get_variable(scope + '_conv3', [k, k, dim_hidden, dim_hidden], \\\n initializer=conv_initializer, dtype=dtype)\n weights[scope + '_bias3'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias3')\n weights[scope + '_conv_res'] = tf.get_variable(scope + '_conv_res', [1, 1, last_dim_hidden, dim_hidden], \\\n initializer=conv_initializer, dtype=dtype)\n return weights\n\n def construct_resnet_ss_weights(self):\n \"\"\"The function to construct ss weights.\n Return:\n The ss weights.\n \"\"\" \n ss_weights = {}\n ss_weights = self.construct_residual_block_ss_weights(ss_weights, 3, 64, 'block1')\n ss_weights = self.construct_residual_block_ss_weights(ss_weights, 64, 128, 'block2')\n ss_weights = self.construct_residual_block_ss_weights(ss_weights, 128, 256, 'block3')\n ss_weights = self.construct_residual_block_ss_weights(ss_weights, 256, 512, 'block4')\n return ss_weights\n\n def construct_residual_block_ss_weights(self, ss_weights, last_dim_hidden, dim_hidden, scope='block0'):\n \"\"\"The function to construct one block ss weights.\n Args:\n ss_weights: the ss weight list.\n last_dim_hidden: the hidden dimension number of last block.\n dim_hidden: the hidden dimension number of the block.\n scope: the label to indicate which block we are processing.\n Return:\n The ss block weights.\n \"\"\" \n ss_weights[scope + '_conv1'] = tf.Variable(tf.ones([1, 1, last_dim_hidden, dim_hidden]), name=scope + '_conv1')\n ss_weights[scope + '_bias1'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias1')\n ss_weights[scope + '_conv2'] = tf.Variable(tf.ones([1, 1, dim_hidden, dim_hidden]), name=scope + '_conv2')\n ss_weights[scope + '_bias2'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias2')\n ss_weights[scope + '_conv3'] = tf.Variable(tf.ones([1, 1, dim_hidden, dim_hidden]), name=scope + '_conv3')\n ss_weights[scope + '_bias3'] = tf.Variable(tf.zeros([dim_hidden]), name=scope + '_bias3')\n return ss_weights\n\n\n\n\n","sub_path":"models/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":12317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"307766714","text":"import random\n\nfrom django.shortcuts import render, redirect, get_object_or_404\n\nfrom models import Recommendation, User\nfrom forms import UserForm, RecommendationForm\n\ndef home(request):\n if request.method == 'POST':\n form = UserForm(request.POST)\n if form.is_valid():\n username = form.cleaned_data['username']\n User.objects.get_or_create(username=username)\n return redirect(user, username)\n else:\n form = UserForm()\n\n return render(request, 'home.html', {'form': form})\n\ndef users(request):\n users = User.objects.all()\n return render(request, 'users.html', {'users': users})\n\ndef user(request, username):\n user = get_object_or_404(User, username=username)\n\n if request.method == 'POST':\n form = RecommendationForm(request.POST)\n if form.is_valid():\n\n url = form.cleaned_data['url']\n recommender = user\n recommendees_pks = form.cleaned_data['recommendees']\n recommendees = User.objects.filter(pk__in=recommendees_pks)\n\n recommendation = Recommendation.create(url=url, recommender=recommender)\n recommendation.save()\n recommendation.recommendees.add(*recommendees)\n recommendation.save()\n else:\n form = RecommendationForm()\n\n recommendations_received = user.recommendations_received.all()\n recommendations_made = user.recommendations_made.all()\n\n users = User.objects.all()\n\n recommendation_examples = [\n \"http://www.amazon.co.uk/Thinking-Fast-Slow-Daniel-Kahneman/dp/0141033576\",\n \"http://www.justiceharvard.org/about/justice-book/\",\n \"http://theleanstartup.com/\",\n \"http://paulgraham.com/love.html\",\n \"http://web.mit.edu/newsoffice/2013/commencement-address-houston-0607.html\"\n ]\n\n return render(request, 'user.html', {\n 'user': user,\n 'form': form,\n 'recommendations_received': recommendations_received,\n 'recommendations_made': recommendations_made,\n 'users': users,\n 'recommendation_example': random.choice(recommendation_examples)\n })\n\ndef recommendations(request):\n recommendations = Recommendation.objects.all()\n return render(request, 'recommendations.html', {'recommendations': recommendations})\n\ndef out(request, recommendation_id):\n recommendation = get_object_or_404(Recommendation, pk=recommendation_id)\n return redirect(recommendation.url)","sub_path":"recommend/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2451,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"35212273","text":"# encoding=utf8\nimport sys\nreload(sys)\nsys.setdefaultencoding('utf8')\n\nfrom math import sqrt\nfrom pymongo import MongoClient\nimport datetime\n\n#Calcula a similaridade de 2 itens com formato Padrao\ndef similaridade(lista,item1,item2):\n si = 0\n \n for item in lista[item1] :\n if item in lista[item2] : si =1\n\n if(si == 0) : return 0\n\n soma = sum([pow(lista[item1][item] + lista[item2][item],2)\n for item in lista[item1] if item in lista[item2] ])\n return 1 / (1+ sqrt(soma))\n\ndef GetSimilares(lista,itemOriginal):\n si = [(similaridade(lista,itemOriginal,item),item) \n for item in lista if item != itemOriginal]\n si.sort()\n si.reverse()\n return si","sub_path":"UsuarioRecomendacao/FuncoesGerais.py","file_name":"FuncoesGerais.py","file_ext":"py","file_size_in_byte":705,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"312670197","text":"def py_cc_library(name, deps, srcs, visibility=None):\n native.cc_binary(\n name = name + \".dll\",\n deps = deps,\n srcs = srcs,\n linkshared = True,\n linkstatic = True,\n )\n\n native.genrule(\n name = name + \"_pyd\",\n srcs = [name + \".dll\"],\n outs = [name + \".pyd\"],\n cmd = \"\"\"cp $< $@\"\"\",\n )\n\n native.py_library(\n name = name,\n data = [\":\" + name + \".pyd\"],\n imports = [\".\"],\n visibility = visibility,\n )\n","sub_path":"rodeo/formats/python/build_defs.bzl","file_name":"build_defs.bzl","file_ext":"bzl","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"604173137","text":"\"\"\"\n\nObject oriented programming\nWrite a class MyClass it contains 2 members\nOne number and one string\nWrite default constructor, parametrized constructor\nOverload +,*,- operator\n+ - will do the addition of the number and\nascii of 1 st character of the string\n‘- - it will chk if num>ascii of 1 st character of the string\nThen num-ascii of 1 st character of the string\nElse ascii of 1 st character of the string – num\n‘* will do the multiplication num and\nascii of last character of the string\n\"\"\"\n\n\nclass Myclass:\n\n def __init__(self, str, number):\n self.str = str\n self.number = number\n print(self.str)\n print(self.number)\n\n def __add__(self, other):\n print(self.number + ord(self.str[0]))\n return self.number + ord(self.str[0])\n\n def __sub__(self, other):\n # return self.str[0]\n if self.number > ord(self.str[0]):\n print(self.number - ord(self.str[0]))\n return self.number - ord(self.str[0])\n\n else:\n print(ord(self.str[0]) - self.number)\n return ord(self.str[0]) - self.number\n\n def __mul__(self, other):\n print(self.number * ord(self.str[-1]))\n return self.number * ord(self.str[-1])\n\ns1=Myclass('ABC',5)\ns2='+'\ns4='-'\ns6='*'\n\n\n\n#Add method\ns3=s1+s2\nprint(s3)\n\n\n#Subtract method\n# s5=s1-s4\n# print(s5)\n\n#Multiply method\n# s7=s1*s6\n# print(s7)\n\n\n# print(ord('A')+5)\n# print(ord('A')-5)\n# print(5*ord('C'))\n","sub_path":"python_assignments/Lab Assignments/operator_overloading.py","file_name":"operator_overloading.py","file_ext":"py","file_size_in_byte":1449,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"300335445","text":"from django.urls import re_path\nfrom django.views.decorators.cache import cache_page\n\nfrom violation import views\n\nurlpatterns = [\n re_path(r'csv/', views.violation_csv, name='violation_csv'),\n re_path(r'^view/(?P[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12})/$',\n cache_page(60 * 60 * 24)(views.ViolationDetail.as_view()),\n name=\"view-violation\"),\n re_path(r'^edit/(?P[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12})/$',\n views.ViolationEditBasicsView.as_view(),\n name=\"edit-violation\"),\n re_path(r'^edit/locations/(?P[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12})/$',\n views.ViolationEditLocationsView.as_view(),\n name=\"edit-violation-locations\"),\n re_path(r'^delete/(?P[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12})/$',\n views.ViolationDeleteView.as_view(),\n name=\"delete-violation\"),\n re_path(r'^create/$',\n views.ViolationCreateBasicsView.as_view(),\n name=\"create-violation\"),\n re_path(r'type/autocomplete/$', views.violation_type_autocomplete, name='violation-type-autocomplete'),\n re_path(r'classification/autocomplete/$',\n views.violation_perpetrator_classification_autocomplete,\n name='perpetrator-classification-autocomplete'),\n]\n","sub_path":"violation/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"382602652","text":"from .tokenizer import SimpleTokenizer\nfrom .reader import ReaderFactory\n\nimport collections\nimport numpy as np\nimport warnings\nimport pdb\n\ndef get_vocab_size(corpus):\n \"\"\" words are {0, 1, ..., n_words - 1}\"\"\"\n vocabulary_size = 1\n for idx, center_word_id in enumerate(corpus):\n if center_word_id + 1> vocabulary_size:\n vocabulary_size = center_word_id + 1\n print(\"vocabulary_size={}\".format(vocabulary_size))\n return vocabulary_size\n\ndef build_cooccurance_dict(data, skip_window, vocabulary_size):\n cooccurance_count = collections.defaultdict(collections.Counter)\n for idx, center_word_id in enumerate(data):\n if center_word_id > vocabulary_size:\n vocabulary_size = center_word_id\n for i in range(max(idx - skip_window - 1, 0), min(idx + skip_window + 1, len(data))):\n cooccurance_count[center_word_id][data[i]] += 1\n cooccurance_count[center_word_id][center_word_id] -= 1\n return cooccurance_count\n\nfrom scipy.sparse import coo_matrix, csr_matrix\nfrom tqdm import tqdm\ndef worker(pij_dok,pi,pj,k,pair):\n i,j=pair\n x = np.log(pij_dok) - np.log(pi*pj) - np.log(k)\n x = np.array(x)[0][0]\n if np.isinf(x) or np.isnan(x):\n x = 0\n if x>0:\n return [i,j,x]\n else:\n return []\n\ndef construct_coo_matrix(data, k, skip_window):\n vocabulary_size = get_vocab_size(data)\n cooccur = build_cooccurance_dict(data, skip_window, vocabulary_size)\n\n rows = []\n cols = []\n data = []\n print(\"Constructing Nij...\")\n for i in tqdm(range(vocabulary_size)):\n for j in range(vocabulary_size):\n if cooccur[i][j] != 0:\n rows.append(i)\n cols.append(j)\n data.append(cooccur[i][j])\n Nij = coo_matrix((np.array(data), (np.array(rows), np.array(cols))),shape=(vocabulary_size,vocabulary_size))\n Ni = np.sum(Nij, axis=1)\n tot = np.sum(Nij)\n with warnings.catch_warnings():\n \"\"\"log(0) is going to throw warnings, but we will deal with it.\"\"\"\n warnings.filterwarnings(\"ignore\")\n\n Pij = Nij / tot\n Pi = Ni / np.sum(Ni)\n # c.f.Neural Word Embedding as Implicit Matrix Factorization, Levy & Goldberg, 2014\n from concurrent.futures import ThreadPoolExecutor\n from concurrent.futures import as_completed\n import multiprocessing as mp\n #idx_pairs = []\n #for i in range(vocabulary_size):\n # for j in range(vocabulary_size):\n # idx_pairs.append((i,j))\n\n rows = []\n cols = []\n data = []\n results = []\n #Pij_dok = Pij.todok()\n Pij_csr = Pij.tocsr()\n print(\"Constructing Pij...\")\n def worker_(pair):\n i,j=pair\n x = np.log(Pij_dok[i,j]) - np.log(Pi[i]*Pi[j]) - np.log(k)\n x = np.array(x)[0][0]\n if np.isinf(x) or np.isnan(x):\n return []\n if x>0:\n return [i,j,x]\n else:\n return []\n \"\"\"\n import time\n stime = time.time()\n for i in tqdm(range(vocabulary_size)):\n #pool = mp.Pool(processes=4)\n #results = [pool.apply(worker, args=(Pij_dok[i,j], Pi[i], Pi[j], k,(i,j),)) for j in range(vocabulary_size)]\n #results = [i for i in results if len(i) >0]\n\n with ThreadPoolExecutor(max_workers = 8) as executor:\n future_workers = [executor.submit(worker, (i,j)) for j in range(vocabulary_size)]\n #results = executor.map(worker, idx_pairs)\n for future in tqdm(as_completed(future_workers)):\n try:\n result = future.result()\n except:\n pass\n else:\n if len(result)>0:\n results.append(result)\n\n print(\"Running time: \", time.time()-stime)\n rows, cols, data = zip(*results)\n \"\"\"\n PMI = csr_matrix((vocabulary_size, vocabulary_size))\n for i in tqdm(range(vocabulary_size)):\n tmp = np.log(Pij_csr.getrow(i)/(Pi[i]*Pi.T*k))\n tmp[np.isnan(tmp)] =0\n tmp[np.isinf(tmp)]=0\n tmp[tmp<0] =0\n tmp = coo_matrix(tmp)\n #import pdb;pdb.set_trace()\n data.append(tmp.data.astype(np.float16))\n cols.append(tmp.col.astype(np.int16))\n rows.append(np.array([i]*tmp.col.shape[0]).astype(np.int16))\n\n #import pdb;pdb.set_trace()\n continue\n for j in range(vocabulary_size):\n #import time\n #stime = time.time()\n #a= Pij_dok[i,j]\n #b = Pi[i]\n #c = Pi[j]\n #print(\"Reading time \", time.time()-stime)\n #stime = time.time()\n #x= np.log(a)-np.log(b*c) - np.log(k)\n #print(\"CPU time \", time.time()-stime)\n x = np.log(Pij_dok[i,j]) - np.log(Pi[i]*Pi[j]) - np.log(k)\n x = np.array(x)[0][0]\n if np.isinf(x) or np.isnan(x):\n x = 0\n if x>00:\n rows.append(i)\n cols.append(j)\n data.append(x)\n #\"\"\"\n data = np.concatenate(data)\n cols = np.concatenate(cols)\n rows = np.concatenate(rows)\n PMI = coo_matrix((data, (rows, cols)),shape=(vocabulary_size,vocabulary_size))\n #PMI = coo_matrix((np.array(data), (np.array(rows), np.array(cols))),shape=(vocabulary_size,vocabulary_size))\n #import time\n #stime = time.time()\n #PMI = coo_matrix(PMI)\n #print(\"Convert time \", time.time()-stime)\n return PMI, Nij\n\n\ndef construct_matrix(data, k, skip_window):\n vocabulary_size = get_vocab_size(data)\n cooccur = build_cooccurance_dict(data, skip_window, vocabulary_size)\n\n Nij = np.zeros([vocabulary_size, vocabulary_size])\n for i in range(vocabulary_size):\n for j in range(vocabulary_size):\n Nij[i,j] += cooccur[i][j]\n Ni = np.sum(Nij, axis=1)\n tot = np.sum(Nij)\n with warnings.catch_warnings():\n \"\"\"log(0) is going to throw warnings, but we will deal with it.\"\"\"\n warnings.filterwarnings(\"ignore\")\n Pij = Nij / tot\n Pi = Ni / np.sum(Ni)\n # c.f.Neural Word Embedding as Implicit Matrix Factorization, Levy & Goldberg, 2014\n PMI = np.log(Pij) - np.log(np.outer(Pi, Pi)) - np.log(k)\n PMI[np.isinf(PMI)] = 0\n PMI[np.isnan(PMI)] = 0\n return PMI, Nij\n\ndef generate_matrix(afile, vocab_size, min_count, neg_samples, skip_window):\n reader = ReaderFactory.produce(afile[-3:])\n data = reader.read_data(afile)\n tokenizer = SimpleTokenizer()\n indexed_corpus = tokenizer.do_index_data(data,\n n_words=vocab_size,\n min_count=min_count)\n pmi, nij = construct_matrix(indexed_corpus, neg_samples, skip_window)\n return pmi, tokenizer.dictionary, tokenizer.reversed_dictionary, nij\n\nif __name__ == \"__main__\":\n vocabulary_size = 10000\n min_count = 100\n neg_samples = 1\n skip_window = 5\n pmi, dd, rd, nij = generate_matrix(\"../data/text8.zip\", vocabulary_size, min_count, neg_samples, skip_window)\n import pickle\n pickle.dump([pmi, dd, rd, nij], open('../text8_{}.pkl'.format(vocabulary_size),'wb'), protocol=2)\n","sub_path":"matrix/signal_matrix.py","file_name":"signal_matrix.py","file_ext":"py","file_size_in_byte":7391,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"177892894","text":"#!/usr/bin/env python3\r\n\r\nimport docker\r\nimport argparse\r\nimport sys\r\nfrom datetime import datetime\r\n\r\n\r\ndef logando(mensagem, e, logfile=\"docker-cli.log\"):\r\n data_atual = datetime.now().strftime('%d/%m/%Y %H:%M:%S')\r\n with open('docker-cli.log', 'a') as log:\r\n texto = \"%s \\t %s \\t %s \\n\" % (data_atual, mensagem, str(e))\r\n log.write(texto)\r\n\r\n#def criar_container(imagem, comando):\r\n# client = docker.from_env()\r\n# executando = client.containers.run(imagem, comando)\r\n# print(executando)\r\n\r\ndef criar_container_args(args):\r\n try:\r\n client = docker.from_env()\r\n executando = client.containers.run(args.imagem, args.comando)\r\n print(executando)\r\n return (executando)\r\n except docker.errors.ImageNotFound as e:\r\n logando(\"Erro: Esta imagem não existe!\", e)\r\n except docker.errors.NotFound as e:\r\n logando(\"Erro: Este comando não existe!\", e)\r\n except Exception as e:\r\n logando(\"Erro! Falha não prevista!\", e)\r\n finally:\r\n print(\"Execuntando...!\")\r\n\r\ndef listar_containers():\r\n try:\r\n client = docker.from_env()\r\n get_all = client.containers.list(all)\r\n for cada_container in get_all:\r\n conectando = client.containers.get(cada_container.id)\r\n print (\"O container %s utiliza a imagem %s rodando o comando %s\" %(conectando.short_id, conectando.attrs['Config']['Image'], conectando.attrs['Config']['Cmd']))\r\n except Exception as e:\r\n logando(\"Erro: Contatar seu administrator\", e)\r\n\r\ndef procurar_container(imagem):\r\n try:\r\n client = docker.from_env()\r\n get_all = client.containers.list(all)\r\n for cada_container in get_all:\r\n conectando = client.containers.get(cada_container.id)\r\n imagem_container = conectando.attrs['Config']['Image']\r\n if str(imagem).lower() in str(imagem_container).lower():\r\n print (\"Achei o container %s que contem a palavra %s no nome da sua imagem %s\" %(conectando.short_id, imagem, imagem_container))\r\n except Exception as e:\r\n logando(\"Erro: Contatar seu administrator\", e) \r\n\r\ndef remover_container():\r\n try:\r\n client = docker.from_env()\r\n get_all = client.containers.list(all)\r\n for cada_container in get_all:\r\n conectando = client.containers.get(cada_container.id)\r\n portas = conectando.attrs['HostConfig']['PortBindings']\r\n if isinstance(portas,dict):\r\n for porta, porta1 in portas.items():\r\n porta1 = str(porta1)\r\n porta2 = ''.join(filter(str.isdigit, porta1))\r\n if int(porta2) <= 1024:\r\n print(\"Removendo o container %s que esta escutando na porta %s e bindando no host na porta %s\" % (cada_container.short_id, porta, porta2))\r\n removendo = cada_container.remove(force=True)\r\n except Exception as e:\r\n logando(\"Erro: Contatar seu administrator\", e)\r\n\r\n\r\nparser = argparse.ArgumentParser(description=\"docker-cli criado na aula de python\")\r\nsubparser = parser.add_subparsers()\r\n\r\ncriar_opt = subparser.add_parser('criar')\r\ncriar_opt.add_argument('--imagem', required=True)\r\ncriar_opt.add_argument('--comando', required=True)\r\ncriar_opt.set_defaults(func=criar_container_args)\r\n\r\ncmd = parser.parse_args()\r\ncmd.func(cmd)\r\n\r\n\r\n#listar_containers()\r\n#criar_container(\"alpine\", \"echo vAII\")\r\n#procurar_container(\"nginx\")\r\n#remover_container()","sub_path":"docker-cli.py","file_name":"docker-cli.py","file_ext":"py","file_size_in_byte":3494,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"324313365","text":"import numpy as np\nfrom scipy.spatial.distance import cdist\n\nimport matplotlib as mpl\nmpl.use(\"Agg\")\nimport matplotlib.pyplot as plt\nimport matplotlib.collections as mc\nimport matplotlib.cm as cm\n\nfrom VISolver.Domain import Domain\n\nfrom IPython import embed\nclass MLN(Domain):\n\n def __init__(self,Network):\n self.Network = Network\n self.UnpackNetwork(Network)\n self.dim = self.CalculateNetworkSize()\n\n def F(self,Data):\n return self._F(Data)\n\n # Functions Used to Animate and Save Network Run to Movie File\n\n def plotboundary(self,ax,B,b,c='w',boundary=None,mask=False):\n if B[0] != 0 and B[1] != 0:\n start = [self.xlims[0],-(B[0]*self.xlims[0]+b)/B[1]]\n if start[1] < self.ylims[0]:\n start = [-(B[1]*self.ylims[0]+b)/B[0],self.ylims[0]]\n elif start[1] > self.ylims[1]:\n start = [-(B[1]*self.ylims[1]+b)/B[0],self.ylims[1]]\n\n end = [self.xlims[1],-(B[0]*self.xlims[1]+b)/B[1]]\n if end[1] < self.ylims[0]:\n end = [-(B[1]*self.ylims[0]+b)/B[0],self.ylims[0]]\n elif end[1] > self.ylims[1]:\n end = [-(B[1]*self.ylims[1]+b)/B[0],self.ylims[1]]\n elif B[0] == 0 and B[1] != 0:\n start = [self.xlims[0],-b/B[1]]\n end = [self.xlims[1],-b/B[1]]\n elif B[1] != 0 and B[1] == 0:\n start = [-b/B[0],self.ylims[0]]\n end = [-b/B[0],self.ylims[1]]\n else:\n start = [self.xlims[0]-1,self.ylims[0]-1]\n end = [self.xlims[0]-1,self.ylims[0]-1]\n\n xs = [start[0],end[0]]\n ys = [start[1],end[1]]\n\n if mask:\n xs = np.ma.array(xs,mask=True)\n ys = np.ma.array(ys,mask=True)\n\n if boundary is None:\n boundary, = ax.plot(xs,ys,c=c)\n else:\n boundary.set_xdata(xs)\n boundary.set_ydata(ys)\n\n return boundary\n\n def InitVisual(self):\n\n assert self.D == 2\n\n assert self.I == 4\n colors = ['c','m','b','y']\n\n ax = plt.gca()\n fig = plt.gcf()\n\n ax.set_xlim(self.xlims)\n ax.set_ylim(self.ylims)\n\n pos = ax.plot(self.pos[0],self.pos[1],'g+',mew=5,ms=10)\n neg = ax.plot(self.neg[0],self.neg[1],'rx',mew=5,ms=10)\n\n datacenters = ax.scatter(self.xjd[:,0],self.xjd[:,1],c='k',s=50)\n boundary = self.plotboundary(ax,self.B,self.b,c='k')\n\n maskedI = np.ma.array(np.zeros(self.I),mask=True)\n self.datastreams = ax.scatter(maskedI,maskedI,c='w',s=50)\n self.datastreams.set_facecolors(colors)\n\n self.boundaries = [self.plotboundary(ax,np.zeros(self.D),0,c=colors[i],mask=True) for i in range(self.I)]\n\n return datacenters,boundary,pos,neg\n\n def UpdateVisual(self,num,ax,Frames):\n\n Data = Frames[num]\n\n qij, xid, Bid, bi = self.UnpackData(Data)\n\n one = np.ones(xid.shape[0])\n pyx = self.Py_x(one,xid,Bid,bi)\n pred_one = pyx > 0.5\n colors = ['g' if po else 'r' for po in pred_one]\n\n self.datastreams.set_offsets(xid)\n self.datastreams.set_edgecolors(colors)\n boundaries = [self.plotboundary(ax,B,b,boundary=self.boundaries[i]) for i,B,b in zip(range(self.I),Bid,bi)]\n\n return self.datastreams,boundaries\n\n # Functions used to Initialize the Machine Learning Network and Calculate F\n\n def UnpackNetwork(self,Network):\n I,J,D,pAj,pbj,ri,Ci,ci,lamj,xjd,Gj,gami,duration,Ki,alphai,B,b,los,his,pos,neg = Network\n self.I = I\n self.J = J\n self.D = D\n self.pAj = pAj\n self.pbj = pbj\n self.ri = ri\n self.Ci = Ci\n self.ci = ci\n self.lamj = lamj\n self.clam = np.outer(self.ci,self.lamj)\n self.xjd = xjd\n self.Gj = Gj\n self.gami = gami\n self.duration = duration\n self.counter = 0\n self.nt = None\n self.Ki = Ki\n self.alphai = alphai\n self.B = B\n self.b = b\n self.los = los\n self.his = his\n self.xlims = [self.los[I*J],self.his[I*J]]\n self.ylims = [self.los[I*J+1],self.his[I*J+1]]\n self.pos = pos\n self.neg = neg\n\n def CalculateNetworkSize(self):\n I,J,D = self.I, self.J, self.D\n return I*J + 2*I*D + I\n\n def UnpackData(self,Data):\n I,J,D = self.I, self.J, self.D\n ptr = 0\n qij = Data[ptr:ptr+I*J].reshape((I,J))\n ptr += I*J\n xid = Data[ptr:ptr+I*D].reshape((I,D))\n ptr += I*D\n Bid = Data[ptr:ptr+I*D].reshape((I,D))\n ptr += I*D\n bi = Data[ptr:ptr+I]\n return qij,xid,Bid,bi\n\n def Price(self,qij):\n return self.pAj - self.pbj*np.sum(qij,axis=0)\n\n def Distances(self,xid):\n return np.hstack([cdist(xid,self.xjd[j,None],'mahalanobis',VI=self.Gj[j]) for j in range(self.J)])\n\n def NewsTip(self,xid):\n if self.nt is None or self.counter % self.duration == 0:\n los = self.los[self.I*self.J:self.I*(self.J+self.D)]\n his = self.his[self.I*self.J:self.I*(self.J+self.D)]\n closest = np.round((xid.flatten() - los)/(his-los))\n farthest = closest*los + (1-closest)*his\n self.nt = farthest.reshape(xid.shape)\n self.counter = 1\n else:\n self.counter += 1\n return self.nt\n\n def y(self,xid):\n one = np.ones(xid.shape[0])\n return (self.Py_x(one,xid,self.B,self.b) > 0.5).astype(float)*2-1\n\n def Py_x(self,yi,xid,Bid,bi):\n return 1/(1+np.exp(-yi*(np.sum(Bid*xid,axis=1)+bi)))\n\n def _F(self,Data):\n qij, xid, Bid, bi = self.UnpackData(Data)\n pj = self.Price(qij)\n dij = self.Distances(xid)\n xtd = self.NewsTip(xid)\n yi = self.y(xid)\n pyx = self.Py_x(yi,xid,Bid,bi)\n\n dfi_dqij = pj - self.pbj*qij + 2*(self.ri*qij.T).T - self.clam + (self.ci*dij.T).T\n temp_dij = np.swapaxes([np.dot((xid-self.xjd[j]),self.Gj[j]) for j in range(self.J)],0,2)\n dfi_dxid = 2*(self.gami*(xid-xtd).T).T + 2*(self.ci*np.sum(qij*temp_dij,axis=-1)).T + 2*(self.alphai*(((Bid*xid).T-bi).T*Bid).T).T\n dLRi_dBid = (self.Ki*Bid.T).T - (yi*(1-pyx)*xid.T).T\n dLRi_dbi = self.Ki*bi - yi*(1-pyx)\n\n return np.hstack([delta.flatten() for delta in [dfi_dqij,dfi_dxid,dLRi_dBid,dLRi_dbi]])\n\ndef CreateRandomNetwork(I=4,J=3,D=2,seed=None):\n if seed is not None:\n np.random.seed(seed)\n\n pAj = np.random.rand(J)*.005 + 0.02\n pbj = np.random.rand(J)*0.2\n ri = np.ones(I)*.02\n Ci = np.random.rand(I)\n ci = np.random.rand(I)*.02\n lamj = np.random.rand(J)*5\n xjd = np.random.rand(J,D)\n Gj = np.asarray([np.eye(D) for j in range(J)])\n gami = np.ones(I)\n duration = 100\n Ki = np.ones(I)*0.\n alphai = np.random.rand(I)*.02\n B = np.array([1,-1])*0.5\n b = 0\n\n los = np.array([0]*I*J + [0]*I*D + [-1]*I*D + [-1]*I)\n his = np.array([1]*I*J + [1]*I*D + [1]*I*D + [1]*I)\n\n pos = np.array([0.75,0.25])\n neg = np.array([0.25,0.75])\n\n return [I,J,D,pAj,pbj,ri,Ci,ci,lamj,xjd,Gj,gami,duration,Ki,alphai,B,b,los,his,pos,neg]\n\n\n\n","sub_path":"VISolver/Domains/MLN.py","file_name":"MLN.py","file_ext":"py","file_size_in_byte":7114,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"61866036","text":"\n\n#calss header\nclass _TONIGHT():\n\tdef __init__(self,): \n\t\tself.name = \"TONIGHT\"\n\t\tself.definitions = [u'(during) the night of the present day: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'adverbs'\n\n\n\tdef run(self, obj1, obj2):\n\t\tself.jsondata[obj2] = {}\n\t\tself.jsondata[obj2]['properties'] = self.name.lower()\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/adverbs/_tonight.py","file_name":"_tonight.py","file_ext":"py","file_size_in_byte":396,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"612027698","text":"# Copyright 2011 The fast-python-pb Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom collections import defaultdict\n\ntry:\n from collections import OrderedDict\nexcept ImportError:\n # just ignore the order with Python <2.7 for now\n OrderedDict = dict\n\ndef order_dependencies(dependencies):\n depends_on = OrderedDict()\n dependent_off = defaultdict(set)\n \n for item, deps in dependencies:\n depends_on[item] = set()\n for dep in deps:\n dependent_off[dep].add(item)\n depends_on[item].add(dep)\n \n result = []\n while depends_on:\n # collect items wihout any dependencies\n for item, dependens in depends_on.iteritems():\n if not dependens: # item without dependencies\n result.append(item)\n del depends_on[item]\n if item in dependent_off:\n # item resolved -> remove from dependents\n for dependents in dependent_off[item]:\n depends_on[dependents].remove(item)\n break # start again to keep general order of dependencies\n else:\n assert False, 'recursive dependency detected'\n\n return result\n\nclass TestOrderDependencies(object):\n def test_tree(self):\n tree = [\n ('a', ('b', 'c')),\n ('b', ('c', )),\n ('c', ()),\n ]\n assert order_dependencies(tree) == ['c', 'b', 'a']\n \n def test_flat(self):\n tree = [\n ('a', ()),\n ('b', ()),\n ('c', ()),\n ]\n assert order_dependencies(tree) == ['a', 'b', 'c']\n \n def test_diamond(self):\n tree = [\n ('a', ('b', 'c')),\n ('b', ('d',)),\n ('c', ('d',)),\n ('d', ()),\n ]\n assert order_dependencies(tree) == ['d', 'b', 'c', 'a']\n\n def test_loop(self):\n tree = [\n ('a', ('b')),\n ('b', ('c',)),\n ('c', ('a',)),\n ]\n try:\n order_dependencies(tree)\n except AssertionError:\n pass\n else:\n assert False, 'assertion expected'\n\n","sub_path":"src/fastpb/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":2661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"480410245","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\nimport logging\nimport os\nimport subprocess\nimport sys\nfrom configparser import ConfigParser, NoOptionError\nfrom zipfile import ZipFile\n\nfrom P4 import P4, P4Exception\n\nimport const\n\n\nclass ComponentException(Exception):\n pass\n\n\ndef customized(func):\n def exec(*args, **kwargs):\n self = args[0]\n self.customized_action.append(func.__name__)\n func(*args, **kwargs)\n\n return exec\n\n\nclass DefaultComponent:\n __logger = logging.getLogger(__name__)\n\n def __init__(self, name, config):\n self.__config = config\n self.__name = name\n self.__customized_action = []\n self.__tokens = {}\n self.__init_tokens()\n\n def __init_tokens(self):\n if self.__get_component_config_value(\"token.file\") is None:\n return\n token_resource_dir_path = os.path.join(self.resources_dir_path, self.__config.get(\"core\", \"resources.token\"))\n token_resource_file_path = os.path.join(token_resource_dir_path,\n self.__get_component_config_value(\"token.file\"))\n token_file_config_parser = ConfigParser()\n token_file_config_parser.read(token_resource_file_path)\n for k, v in token_file_config_parser.items(\"DEFAULT\"):\n self.__tokens[k] = v\n\n def p4_fetch(self):\n p4 = P4()\n p4.exception_level = 1\n p4.user = self.p4user\n p4.password = self.p4password\n p4.port = self.p4port\n p4.client = self.p4client\n # TODO if not exist create\n try:\n DefaultComponent.__logger.info(\"Begin to sync component [%s] from p4 client [%s]\" % (self.name, p4.client))\n p4.connect()\n p4.run_sync()\n except P4Exception:\n for e in p4.errors:\n DefaultComponent.__logger.error(\"Fail to sync component [%s] code.\" % self.name, e)\n\n def info(self):\n pass\n\n def build(self):\n DefaultComponent.__logger.info(\"Begin to build component [%s].\" % self.name)\n build_cmd = \"%s %s\" % (self.mvn_cmd_path, self.build_cmd)\n build_result = subprocess.run(build_cmd, stdout=sys.stdout, cwd=self.build_dir_path)\n if build_result and build_result.returncode == 0:\n DefaultComponent.__logger.info(\"Success to build component [%s].\" % self.name)\n return\n DefaultComponent.__logger.info(\"Fail to build component [%s].\" % self.name)\n\n def deploy_apache(self):\n DefaultComponent.__extract_zip(self.apache_zip_path, self.apache_deploy_path)\n\n def deploy_tomcat(self):\n DefaultComponent.__extract_zip(self.tomcat_zip_path, self.tomcat_deploy_path)\n\n @staticmethod\n def __extract_zip(resource_path, target_path):\n logging.info(\"Begin to extract apache to [%s]\" % str(target_path))\n resource_zip_file = ZipFile(resource_path)\n resource_zip_file.extractall(target_path)\n logging.info(\"Success to extract apache to [%s]\" % str(target_path))\n return\n\n def deploy(self):\n for one_result_path in self.build_result_paths:\n if not os.path.exists(one_result_path):\n continue\n self.__extract_zip(one_result_path, self.deploy_target_path)\n self._replace_token()\n\n def config(self):\n pass\n\n def start(self):\n pass\n\n def stop(self):\n pass\n\n def _replace_token(self):\n for dir_root_path, sub_dir_names, file_names in os.walk(self.deploy_target_path):\n if not file_names:\n continue\n for file_name in file_names:\n if file_name.endswith(\".tmpl\"):\n output_file_name = file_name[0: file_name.find(\".tmpl\")]\n tmpl_file_absolute_path = os.path.join(dir_root_path, file_name)\n output_file_absolute_path = os.path.join(dir_root_path, output_file_name)\n with open(tmpl_file_absolute_path, mode=\"r\", encoding=\"utf-8\") as tmpl_file:\n with open(output_file_absolute_path, mode=\"w\", encoding=\"utf-8\") as output_file:\n for line in tmpl_file:\n replaced_line = line\n for k, v in self.tokens.items():\n replaced_line = replaced_line.replace(\"@\" + k.upper() + \"@\", v)\n output_file.write(replaced_line)\n\n def __get_component_config_value(self, key):\n section_name = \"%s.%s\" % (const.COMPONENT_MODULE_PACKAGE_NAME, self.name)\n if not self.__config.has_section(section_name):\n return None\n try:\n return self.__config.get(section_name, key)\n except NoOptionError:\n return None\n\n @property\n def p4user(self):\n return self.__config.get(\"p4\", \"user\")\n\n @property\n def p4password(self):\n return self.__config.get(\"p4\", \"password\")\n\n @property\n def p4port(self):\n return \"%s:%s\" % (\n self.__config.get(\"p4\", \"host\"), self.__config.get(\"p4\", \"port\"))\n\n @property\n def p4client(self):\n return self.__get_component_config_value(\"p4.client.name\")\n\n @property\n def name(self):\n return self.__name\n\n @property\n def code_base_dir_path(self):\n return os.path.join(os.path.abspath(self.__config.get(\"p4\", \"workspace.root.path\")),\n self.__get_component_config_value(\"p4.client.name\"))\n\n @property\n def resources_dir_path(self):\n return os.path.abspath(\"resources\")\n\n @property\n def tomcat_zip_path(self):\n tomcat_zip_file_path = os.path.join(self.resources_dir_path, self.__config.get(\"core\", \"resources.tomcat\"))\n return tomcat_zip_file_path\n\n @property\n def tomcat_deploy_path(self):\n tomcat_deploy_path = os.path.join(self.__config.get(\"core\", \"deploy.target.dir.root\"),\n self.__config.get(\"core\", \"deploy.target.dir.relative.tomcat\"))\n return tomcat_deploy_path\n\n @property\n def apache_zip_path(self):\n apache_zip_file_path = os.path.join(self.resources_dir_path, self.__config.get(\"core\", \"resources.apache\"))\n return apache_zip_file_path\n\n @property\n def apache_deploy_path(self):\n apache_deploy_path = os.path.join(self.__config.get(\"core\", \"deploy.target.dir.root\"),\n self.__config.get(\"core\", \"deploy.target.dir.relative.apache\"))\n return apache_deploy_path\n\n @property\n def build_dir_path(self):\n return os.path.join(self.code_base_dir_path, self.__get_component_config_value(\"build.dir\"))\n\n @property\n def build_result_paths(self):\n build_results = self.__get_component_config_value(\"build.result\")\n if build_results is None:\n return []\n build_result_list = build_results.split(\",\")\n return [os.path.join(self.code_base_dir_path, one_result) for one_result in build_result_list]\n\n @property\n def deploy_target_root_dir_path(self):\n return os.path.abspath(self.__config.get(\"core\", \"deploy.target.dir.root\"))\n\n @property\n def deploy_target_root_components_dir_path(self):\n return os.path.join(self.deploy_target_root_dir_path,\n self.__config.get(\"core\", \"deploy.target.dir.relative.components\"))\n\n @property\n def mvn_cmd_path(self):\n return os.path.abspath(self.__config.get(\"core\", \"maven.path\"))\n\n @property\n def build_cmd(self):\n return self.__get_component_config_value(\"build.cmd\")\n\n @property\n def deploy_target_path(self):\n deploy_relative_path = self.__get_component_config_value(\n \"deploy.target.dir\")\n if deploy_relative_path is None:\n deploy_relative_path = self.name\n deploy_path = os.path.join(self.deploy_target_root_components_dir_path,\n deploy_relative_path)\n return deploy_path\n\n @property\n def customized_action(self):\n return self.__customized_action\n\n @property\n def tokens(self):\n return self.__tokens\n\n\n__all__ = [ComponentException, DefaultComponent]\n","sub_path":"src/component/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":8224,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"328807429","text":"# Evaluate net accuracy on test set\n\n\nimport numpy as np\nimport os\nimport LabPicsVesselInstanceReader as ChemScapeInstanceReader\nimport torch\n\n############################################################################################################\n#########################################################################################################################\nclass Evaluator:\n def __init__(self, AnnDir,OutFile): # Make reader for test set, and open file for eevaluation results\n self.AnnDir = AnnDir\n self.OutFile=OutFile\n if not os.path.exists(OutFile):\n f=open(OutFile,\"w\")\n f.close()\n print(\"-------------------------------------Creating test evaluator------------------------------------------------------\")\n self.Reader = ChemScapeInstanceReader.Reader(MainDir=self.AnnDir, TrainingMode=False)\n#####################################################################################333\n def Eval(self,Net,itr):\n print(\"Evaluating\")\n Finished=False\n\n IOUSum = 0\n InterSum = 0\n UnionSum = 0\n ImSum=0\n IOUDif = 0\n CatAccuracy = 0\n\n # IOUSumCat = np.zeros([20])\n # InterSumCat = np.zeros([20])\n # UnionSumCat = np.zeros([20])\n # ImSumCat = np.zeros([20])\n #\n\n while (not Finished):\n Imgs, AnnMapGt, BG,ROI,PointerMap, Ignore, Cats, Finished=self.Reader.LoadSingle()\n # --------------------------------------\n # Imgs[:, :, 0] *= 1 - AnnMapGt.astype(np.uint8)\n # Imgs[:, :, 1] *= 1 - Ignore.astype(np.uint8)\n # print(Cats)\n # misc.imshow(Imgs)\n # misc.imshow((ROI + AnnMapGt * 2 + PointerMap * 3).astype(np.uint8) * 40)\n # print(ROI.shape)\n # ----------------------------------------------\n Imgs=np.expand_dims(Imgs,axis=0)\n PointerMap = np.expand_dims(PointerMap,axis=0)\n ROI = np.expand_dims(ROI, axis=0)\n with torch.autograd.no_grad():\n Prob, LbPred, PredIOU, PredIsVessel = Net.forward(Images=Imgs, Pointer=PointerMap,ROI=ROI) # Run net inference and get prediction\n\n PredIOU = np.squeeze(PredIOU.data.cpu().numpy())\n Pred= LbPred.data.cpu().numpy()[0]*(1-Ignore)\n GT=AnnMapGt*(1-Ignore)\n Inter=(Pred*GT).sum()\n Union=(Pred).sum()+(GT).sum()-Inter\n if Union.sum()>0: #Union>0:\n IOUSum += Inter/Union\n InterSum += Inter\n UnionSum += Union\n ImSum += 1\n IOUDif += np.abs(Inter / Union - PredIOU)\n CatAccuracy += (PredIsVessel.data.cpu().numpy()>0).mean()\n # for k in Cats:\n # IOUSumCat[k] += Inter / Union\n # InterSumCat[k] += Inter\n # UnionSumCat[k]+= Union\n # ImSumCat[k] += 1\n #\n # # if GT.sum()>0:\n # # print(k)\n # # Im=Imgs[0].copy()\n # # print( Inter / Union)\n # # Im[:, :, 0] *= 1 - GT.astype(np.uint8)\n # # Im[:, :, 2] *= (1-Ignore).astype(np.uint8)\n # # Im[:, :, 1] *= 1 - Pred.astype(np.uint8)\n # # misc.imshow(Im)\n # # break\n #\n #\n f = open(self.OutFile, \"a\")\n txt=\"\\n=====================\"+str(itr)+\"==============================================\\n\"\n # txt+=str(itr)+\"\\n\"\n # PerPixelPerCat = []\n # PerImagePerCat = []\n # for nm in range(IOUSumCat.shape[0]):\n # if UnionSumCat[nm]>0:\n # txt += str(nm) + \"\\t\" +CatName[nm]+\"\\t\"\n # txt += \"IOU Average Per Pixel=\\t\"+str(InterSumCat[nm]/UnionSumCat[nm])+\"\\t\"\n # txt += \"IOU Average Per Image=\\t\" + str(IOUSumCat[nm]/ImSumCat[nm])+\"\\tNum Examples\\t\"+str(ImSumCat[nm])+\"\\n\"\n # PerPixelPerCat.append(InterSumCat[nm]/UnionSumCat[nm])\n # PerImagePerCat.append(IOUSumCat[nm]/ImSumCat[nm])\n\n\n txt += \"\\n\\n Total IOU Average Per Pixel=\\t\" + str(InterSum / UnionSum) + \"\\t\"\n txt += \"Total IOU Average Per Image=\\t\" + str(IOUSum / ImSum) + \"\\n\"\n txt +=\"\\n------EVAL------\\n\"\n txt += \"\\n\\n Dif Pred IOU=\\t\" + str(IOUDif / ImSum) + \"\\t\"\n txt += \"\\n------Category------\\n\"\n txt += \"Accuracy Rate Cat=\\t\" + str(CatAccuracy/ImSum) + \"\\n\"\n # txt += \"\\n\\n Cat Total IOU Average Per Pixel=\\t\" + str(np.mean(PerPixelPerCat)) + \"\\t\"\n # txt += \"Cat Total IOU Average Per Image=\\t\" + str(np.mean(PerImagePerCat)) + \"\\n\"\n f.write(txt)\n f.close()\n print(txt)\n\n\n\n\n\n\n","sub_path":"InstanceVesselWithCOCO/Evaluator.py","file_name":"Evaluator.py","file_ext":"py","file_size_in_byte":5146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"464514673","text":"import mailchimp_marketing as MailchimpMarketing\nfrom mailchimp_marketing.api_client import ApiClientError\n\nAPI_KEY = \"YOUR_API_KEY\"\nSERVER_PREFIX = \"YOUR_SERVER_PREFIX\"\n\nLIST_ID = \"YOUR_LIST_ID\"\n\nclient = MailchimpMarketing.Client()\nclient.setConfig(API_KEY, SERVER_PREFIX)\ntry:\n response = client.lists.create_segment(LIST_ID, {\n \"name\": \"MegaInfluencer\",\n \"static_segment\": [\n \"dolly.parton@example.com\",\n \"rihanna@example.com\",\n ]\n }, LIST_ID, SUBSCRIBER_HASH)\n print(\"client.ping.get() response: {}\".format(response))\nexcept ApiClientError as error:\n print(\"An exception occurred: {}\".format(error.text))","sub_path":"snippets/python/new_bulk_tags.py","file_name":"new_bulk_tags.py","file_ext":"py","file_size_in_byte":664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"232318219","text":"class Settings(object):\n \"\"\"\n A settings object controls a variety of parameters that are used in\n falsification. There is a single default settings object that all other\n Settings will use as its values s defaults.\n\n Not all settings parameters are guaranteed to be stable. However the\n following are:\n\n max_examples: Once this many examples have been considered without finding\n any counter-example, falsify will terminate\n timeout: Once this amount of time has passed, falsify will terminate even\n if it has not found many examples. This is a soft rather than a hard\n limit - Hypothesis won't e.g. interrupt execution of the called\n function to stop it.\n derandomize: If this is True then hypothesis will run in deterministic mode\n where each falsification uses a random number generator that is seeded\n based on the hypothesis to falsify, which will be consistent across\n multiple runs. This has the advantage that it will eliminate any\n randomness from your tests, which may be preferable for some situations\n . It does have the disadvantage of making your tests less likely to\n find novel breakages.\n\n \"\"\"\n def __init__(\n self,\n min_satisfying_examples=None,\n max_examples=None,\n max_skipped_examples=None,\n timeout=None,\n derandomize=None,\n ):\n self.min_satisfying_examples = (\n min_satisfying_examples or default.min_satisfying_examples)\n self.max_examples = max_examples or default.max_examples\n self.timeout = timeout or default.timeout\n self.max_skipped_examples = (\n max_skipped_examples or default.max_skipped_examples)\n if derandomize is None:\n self.derandomize = default.derandomize\n else:\n self.derandomize = derandomize\n\n\ndefault = Settings(\n min_satisfying_examples=5,\n max_examples=200,\n timeout=60,\n max_skipped_examples=50,\n derandomize=False,\n)\n","sub_path":"src/hypothesis/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":2016,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"453814123","text":"# program must have 0 in user input boxes\n# handle if 0 is passed for Division\n# Input boxes should be lined up nicely\n# output of calultor should be in another text box\n# color the font on buttons in blue\n# add 1,2,3,4,5,6,7,8,9 and dot and equal button keys for calculator\n\n\nimport tkinter as tk # Python 3 import\nfrom tkinter import messagebox\n\n# create canvas to draw boxed and buttons\nroot = tk.Tk()\n# define size of canvas\nroot.geometry(\"500x500\")\n\n# create the add function\n\n\ndef add():\n # get the first number input by the user\n number1Input = number1.get()\n # get the second number input by the user\n number2Input = number2.get()\n # add them and alert user the sum\n msg = messagebox.showinfo(\"Answer!\", int(number1Input) + int(number2Input))\n\n\ndef subtract():\n number1Input = number1.get()\n number2Input = number2.get()\n msg = messagebox.showinfo(\"Answer!\", int(number1Input) - int(number2Input))\n\n\ndef multiply():\n number1Input = number1.get()\n number2Input = number2.get()\n msg = messagebox.showinfo(\"Answer!\", int(number1Input) * int(number2Input))\n\n\ndef division():\n number1Input = number1.get()\n number2Input = number2.get()\n msg = messagebox.showinfo(\"Answer!\", int(number1Input) / int(number2Input))\n\n\n# Create the label to show on canvas\nmy_label = tk.Label(root, text=\"number1\")\n# define where to position or show it on canvas\nmy_label.grid(row=0, column=0)\nnumber1 = tk.Entry(root)\nnumber1.grid(row=0, column=1)\n\nmy_label2 = tk.Label(root, text=\"number2\")\nmy_label2.grid(row=10, column=10)\nnumber2 = tk.Entry(root)\nnumber2.grid(row=20, column=2)\n\nmy_button = tk.Button(root, text=\"ADD\", command=add)\nmy_button.grid(row=1, column=1)\nmy_button = tk.Button(root, text=\"SUBTRACT\", command=subtract)\nmy_button.grid(row=2, column=1)\nmy_button = tk.Button(root, text=\"Division\", command=division)\nmy_button.grid(row=3, column=1)\nmy_button = tk.Button(root, text=\"MULTIPLY\", command=multiply)\nmy_button.grid(row=4, column=1)\n\nroot.mainloop()\n","sub_path":"calculator.py","file_name":"calculator.py","file_ext":"py","file_size_in_byte":1999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"357502305","text":"import torch, pdb\nfrom torch.autograd import Function\nfrom utils import decode, nms\n\n\nclass Detect(Function):\n \"\"\"At test time, Detect is the final layer of SSD. Decode location preds,\n apply non-maximum suppression to location predictions based on conf\n scores and threshold to a top_k number of output predictions for both\n confidence score and locations.\n \"\"\"\n def __init__(self, num_classes, top_k, conf_thresh, nms_thresh, anchors):\n self.num_classes = num_classes\n self.top_k = top_k\n # Parameters used in nms.\n self.nms_thresh = nms_thresh\n if nms_thresh <= 0:\n raise ValueError('nms_threshold must be non negative.')\n self.conf_thresh = conf_thresh\n self.anchors = anchors\n\n def forward(self, loc_data, conf_data):\n \"\"\"\n Args:\n loc_data: (tensor) Loc preds from loc layers\n Shape: [batch,num_priors*4]\n conf_data: (tensor) Shape: Conf preds from conf layers\n Shape: [batch*num_priors,num_classes]\n \"\"\"\n\n batch_size, num_priors, _ = loc_data.shape\n output = torch.zeros(batch_size, self.num_classes, self.top_k, 5)\n conf_preds = conf_data.transpose(2, 1)\n\n # Decode predictions into bboxes.\n for i in range(batch_size):\n decoded_boxes = self.anchors.decode(loc_data[i])\n\n # For each class, perform nms\n conf_scores = conf_preds[i].clone()\n\n for cl in range(1, self.num_classes):\n c_mask = conf_scores[cl].gt(self.conf_thresh)\n scores = conf_scores[cl][c_mask]\n if scores.dim() == 0:\n continue\n l_mask = c_mask.unsqueeze(1).expand_as(decoded_boxes)\n boxes = decoded_boxes[l_mask].view(-1, 4)\n # idx of highest scoring and non-overlapping boxes per class\n ids, count = nms(boxes, scores, self.nms_thresh, self.top_k)\n\n output[i, cl, :count] = torch.cat((scores[ids[:count]].unsqueeze(1), boxes[ids[:count]]), 1)\n flt = output.contiguous().view(batch_size, -1, 5)\n _, idx = flt[:, :, 0].sort(1, descending=True)\n _, rank = idx.sort(1)\n flt[(rank < self.top_k).unsqueeze(-1).expand_as(flt)].fill_(0)\n return output\n","sub_path":"detection.py","file_name":"detection.py","file_ext":"py","file_size_in_byte":2326,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"253497517","text":"__author__ = 'vinodkumar545'\n\nfrom IPython import embed\n\n\"\"\"\nValidate whether the arthemtic parathensis expression is valid or not\n\nreturn True or False\n\n\"\"\"\n\nclass Solution:\n\n\tdef validateExpression(self, value):\n\t\tused_index, i = [], 0\n\t\topened, closed, match = 0, 0, 0\n\t\t\n\t\twhile(i < len(value)):\n\t\t\t\n\t\t\tif value[i] == \")\":\n\t\t\t\tclosed += 1\n\t\t\t\n\t\t\tif value[i] == '(':\n\t\t\t\topened += 1\n\t\t\t\tincreament = 0\n\t\t\t\t\n\t\t\t\tfor j in range(i + 1, len(value)):\n\t\t\t\t\tif value[j] == '(':\n\t\t\t\t\t\tincreament += 1\n\t\t\t\t\t\n\t\t\t\t\tif value[j] == ')' and increament != 0:\n\t\t\t\t\t\tincreament -= 1\n\t\t\t\t\t\n\t\t\t\t\tif value[j] == ')' and increament == 0 and j not in used_index:\n\t\t\t\t\t\tmatch += 2\n\t\t\t\t\t\tused_index.append(j)\n\t\t\t\t\t\tprint(\"one closed expression found\")\n\t\t\t\t\t\tbreak\n\t\t\t\n\t\t\tif value[i] == \")\" and i not in used_index:\n\t\t\t\t\n\t\t\t\tif opened == 0:\n\t\t\t\t\tclosed += 1\n\t\t\t\t\tprint(\"closed parathensis found first\")\n\t\t\t\n\t\t\t# print(opened, closed)\n\t\t\ti += 1\n\t\t\n\t\tprint(opened, closed)\n\t\t\n\t\tif match == (opened + closed) and opened == closed:\n\t\t\tprint('True')\n\t\t\treturn True\n\t\t\n\t\telse:\n\t\t\tprint('False')\n\t\t\treturn False\n\nsolution = Solution()\n\nvalue = \"(a + b) * (c + d)\"\n# value = \"((a + b) * ab) + (a + b)\"\n# value = '3) + (5j'\n# value = \"(((a + b )\"\n# value = \"((((()\"\n# value = \"))((\"\n# value = \"((a + b)\"\n\nsolution.validateExpression(value)\n\n","sub_path":"expression.py","file_name":"expression.py","file_ext":"py","file_size_in_byte":1310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"572503749","text":"# Definition for a binary tree node.\r\nclass TreeNode:\r\n def __init__(self, x):\r\n self.val = x\r\n self.left = None\r\n self.right = None\r\n\r\nclass Solution:\r\n def buildTree(self, inorder, postorder) -> TreeNode:\r\n # edge case\r\n if len(postorder) == 0:\r\n return None\r\n n = len(postorder)\r\n root = TreeNode(postorder[n - 1])\r\n index = -1\r\n for i in range(len(inorder)):\r\n if inorder[i] == root.val:\r\n index = i\r\n break\r\n postLeft = postorder[0:index]\r\n postRight = postorder[index:n - 1]\r\n inLeft = inorder[0:index]\r\n inRight = inorder[index + 1:len(inorder)]\r\n\r\n root.left = self.buildTree(inLeft, postLeft)\r\n root.right = self.buildTree(inRight, postRight)\r\n\r\n return root","sub_path":"45_ConstructTreeUsingInorderPostOrder.py","file_name":"45_ConstructTreeUsingInorderPostOrder.py","file_ext":"py","file_size_in_byte":837,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"14658452","text":"from malcolm.compat import OrderedDict\nfrom malcolm.core.loggable import Loggable\n\n\nclass StateMachine(Loggable):\n\n RESETTING = \"Resetting\"\n DISABLED = \"Disabled\"\n DISABLING = \"Disabling\"\n FAULT = \"Fault\"\n\n # Subclasses must override this\n AFTER_RESETTING = None\n\n def __init__(self):\n self.set_logger_name(type(self).__name__)\n self.allowed_transitions = OrderedDict()\n self.busy_states = []\n assert self.AFTER_RESETTING is not None, \\\n \"No AFTER_RESETTING state given\"\n self.set_allowed(self.RESETTING, self.AFTER_RESETTING)\n self.set_busy(self.RESETTING)\n self.create_states()\n custom_states = list(self.allowed_transitions) + [self.AFTER_RESETTING]\n\n # Set transitions for standard states\n for state in custom_states:\n self.set_allowed(state, self.FAULT)\n self.set_allowed(state, self.DISABLING)\n self.set_allowed(self.FAULT, [self.RESETTING, self.DISABLING])\n self.set_allowed(self.DISABLING, [self.FAULT, self.DISABLED])\n self.set_allowed(self.DISABLED, self.RESETTING)\n\n # These are all the states we can possibly be in\n self.possible_states = list(self.allowed_transitions)\n\n def create_states(self):\n raise NotImplementedError()\n\n def is_allowed(self, initial_state, target_state):\n \"\"\"\n Check if a transition between two states is allowed\n\n Args:\n initial_state(str): Initial state\n target_state(str): Target state\n\n Returns:\n bool: True if allowed, False if not\n \"\"\"\n assert initial_state in self.allowed_transitions, \\\n \"%s is not in %s\" % (initial_state, list(self.allowed_transitions))\n return target_state in self.allowed_transitions[initial_state]\n\n def set_allowed(self, initial_state, allowed_states):\n \"\"\"\n Add an allowed transition state\n\n Args:\n initial_state(str): Initial state\n allowed_states(list(str) / str): States that initial_state can\n transition to\n \"\"\"\n\n if not isinstance(allowed_states, list):\n allowed_states = [allowed_states]\n\n self.allowed_transitions.setdefault(initial_state, set()).update(\n allowed_states)\n\n def set_busy(self, state, busy=True):\n \"\"\"\n Set the busy-ness of a state; i.e. whether the block is considered\n to be busy in a certain state\n\n Args:\n state(str): State to update\n busy(bool): True or False for whether state is a busy state\n \"\"\"\n\n if not busy and state in self.busy_states:\n self.busy_states.remove(state)\n\n elif busy and state not in self.busy_states:\n self.busy_states.append(state)\n\n def is_busy(self, state):\n \"\"\"\n Check if a state is a busy state\n\n Args:\n state(str): State to check busy-ness for\n\n Returns:\n bool: True if state is a busy state, False if not\n \"\"\"\n return state in self.busy_states\n\n\nclass DefaultStateMachine(StateMachine):\n\n READY = \"Ready\"\n\n AFTER_RESETTING = READY\n\n def create_states(self):\n pass\n\n\nclass ManagerStateMachine(DefaultStateMachine):\n\n EDITABLE = \"Editable\"\n SAVING = \"Saving\"\n REVERTING = \"Reverting\"\n\n def create_states(self):\n super(ManagerStateMachine, self).create_states()\n self.set_allowed(self.AFTER_RESETTING, self.EDITABLE)\n self.set_allowed(self.EDITABLE, self.SAVING)\n self.set_allowed(self.EDITABLE, self.REVERTING)\n self.set_allowed(self.SAVING, self.AFTER_RESETTING)\n self.set_allowed(self.REVERTING, self.AFTER_RESETTING)\n\n\nclass RunnableStateMachine(ManagerStateMachine):\n\n IDLE = \"Idle\"\n CONFIGURING = \"Configuring\"\n READY = \"Ready\"\n RUNNING = \"Running\"\n POSTRUN = \"PostRun\"\n PAUSED = \"Paused\"\n SEEKING = \"Seeking\"\n ABORTING = \"Aborting\"\n ABORTED = \"Aborted\"\n\n AFTER_RESETTING = IDLE\n\n def create_states(self):\n super(RunnableStateMachine, self).create_states()\n # Set transitions for normal states\n self.set_allowed(self.IDLE, self.CONFIGURING)\n self.set_allowed(\n self.READY, [self.RUNNING, self.SEEKING, self.RESETTING])\n self.set_allowed(self.CONFIGURING, self.READY)\n self.set_allowed(self.RUNNING, [self.POSTRUN, self.SEEKING])\n self.set_allowed(self.POSTRUN, [self.IDLE, self.READY])\n self.set_allowed(self.PAUSED, [self.SEEKING, self.RUNNING])\n self.set_allowed(self.SEEKING, [self.READY, self.PAUSED])\n\n # Add Abort to all normal states\n normal_states = [\n self.IDLE, self.READY, self.CONFIGURING, self.RUNNING, self.POSTRUN,\n self.PAUSED, self.SEEKING]\n for state in normal_states:\n self.set_allowed(state, self.ABORTING)\n\n # Set transitions for other states\n self.set_allowed(self.ABORTING, self.ABORTED)\n self.set_allowed(self.ABORTED, self.RESETTING)\n","sub_path":"malcolm/core/statemachine.py","file_name":"statemachine.py","file_ext":"py","file_size_in_byte":5067,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"435428253","text":"# -*- coding: UTF-8 -*-\n# from search import Search\n# import sys\n\n# reload(sys)\n# sys.setdefaultencoding('utf8')\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.chrome.options import Options\nimport time\n\n\n# 默认访问一下搜索界面,这样数据接口才能正常访问\nclass Chrome(object):\n def run(self):\n delay = 3\n chrome_options = Options()\n chrome_options.add_argument('--headless')\n # chrome_options.add_argument(('--proxy-server=http://' + '132.58.168.12'))\n driver = webdriver.Chrome(options=chrome_options, executable_path=\"./chromedriver.exe\")\n\n driver.get(\"https://www.12306.cn/index/\")\n \n time.sleep(delay)\n start = driver.find_element_by_id('fromStationText')\n start.clear()\n start.send_keys('重庆')\n start.send_keys(Keys.ENTER)\n start.send_keys(Keys.TAB)\n\n end = driver.find_element_by_id('toStationText')\n end.clear()\n end.send_keys('成都')\n end.send_keys(Keys.ENTER)\n\n query = driver.find_element_by_id('search_one')\n query.click()\n\n time.sleep(delay)\n","sub_path":"ticket/sence/chrome.py","file_name":"chrome.py","file_ext":"py","file_size_in_byte":1179,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"486205189","text":"\n# coding: utf-8\n\n# In[1]:\n\n# importing function\n\nimport os, struct\nfrom array import array as pyarray\nfrom pylab import *\nfrom numpy import *\nimport numpy as np\n\ndef load_mnist(dataset=\"training\", digits=np.arange(10), path=\".\"):\n \"\"\"\n Loads MNIST files into 3D numpy arrays\n\n Adapted from: http://abel.ee.ucla.edu/cvxopt/_downloads/mnist.py\n \"\"\"\n\n if dataset == \"training\":\n fname_img = os.path.join(path, 'train-images-idx3-ubyte')\n fname_lbl = os.path.join(path, 'train-labels-idx1-ubyte')\n elif dataset == \"testing\":\n fname_img = os.path.join(path, 't10k-images-idx3-ubyte')\n fname_lbl = os.path.join(path, 't10k-labels-idx1-ubyte')\n else:\n raise ValueError(\"dataset must be 'testing' or 'training'\")\n\n flbl = open(fname_lbl, 'rb')\n magic_nr, size = struct.unpack(\">II\", flbl.read(8))\n lbl = pyarray(\"b\", flbl.read())\n flbl.close()\n\n fimg = open(fname_img, 'rb')\n magic_nr, size, rows, cols = struct.unpack(\">IIII\", fimg.read(16))\n img = pyarray(\"B\", fimg.read())\n fimg.close()\n\n ind = [ k for k in range(size) if lbl[k] in digits ]\n N = len(ind)\n\n images = zeros((N, rows, cols))\n labels = zeros((N, 1))\n for i in range(len(ind)):\n images[i] = array(img[ ind[i]*rows*cols : (ind[i]+1)*rows*cols ]).reshape((rows, cols))\n labels[i] = lbl[ind[i]]\n \n # image vectors\n imvs = zeros((N, rows * cols))\n for i in range(len(ind)):\n imvs[i] = array(img[ ind[i]*rows*cols : (ind[i]+1)*rows*cols ])\n for j in range(len(imvs[0])):\n imvs[i][j] = imvs[i][j] * (1./255.)\n\n return images, labels, imvs\n\n# Gradient Calculation\ndef sigmoid(x):\n \"Numerically-stable sigmoid function.\"\n if x >= 0:\n z = exp(-x)\n return 1 / (1 + z)\n else:\n z = exp(x)\n return z / (1 + z)\n\ndef gradient(W, X, n, lbl, target):\n T = 0\n if lbl[n] == target:\n T = 1\n \n W_T = np.transpose(W)\n \n val = -1 * np.dot(W_T, X[n])\n\n y_n = 1/(1 + exp(val))\n if y_n == 1:\n y_n = 0.999999\n elif y_n == 0:\n y_n = 0.000001\n \n y_n = sigmoid(-val)\n \n return (y_n - T) * X[n]\n\n# Gradient Descent Calculation\n\ndef gradient_desc(m, nu, T, N, X, lbl, target):\n # W_0\n W = zeros((len(X[0]), 1))\n \n for j in range(m):\n S = zeros((len(X[0]), 1))\n for i in range(N):\n S = S + gradient(W, X, i, lbl, target)\n \n W = W - (nu/(1 + j/T)) * S\n return W\n\n\n# In[2]:\n\n# Load images and labels\n\nimages, labels, imvs = load_mnist('training') \n\n# Take from 20,000, images with 2's and 8's\n\nL = []\nL1 = []\nfor i in range(20000):\n if labels[i] == 2 or labels[i] == 8:\n L.append(labels[i])\n L1.append(imvs[i])\n\n# Append '1' to image vectors\n\nX = zeros((len(L1), len(L1[0]) + 1, 1))\n\nfor i in range(len(L1)):\n X[i] = (np.append([1], L1[i])).reshape((len(L1[0]) + 1, 1))\n\n\n# In[3]:\n\n# Calculate gradient descent\n\nN = len(X)\nm = 100\nnu = 0.01#0.6#0.8#0.001\nT = 3#2#1\ntarget = 2\n\nW = zeros((len(X[0]), 1))\nalpha = gradient(W, X, 0, L, target)\n\n#print(alpha)\nw = gradient_desc(m, nu, T, N, X, L, target)\n\n# Show the weight image\n\nsth = w[1:785]\n#print(size(sth))\na = np.reshape(sth, (28, 28))\nimshow(a)\nshow()\n\n\n# In[4]:\n\nprint(N)\n\n\n# In[5]:\n\nfor i in range(len(w)):\n if w[i] > 100:\n print(i, \" \", w[i])\n\n\n# In[6]:\n\n# Load images and labels\n\ntimages, tlabels, timvs = load_mnist('testing') \n\n# Take from 2,000, images with 2's and 3's\ntL = []\ntL1 = []\nfor i in range(2000):\n if tlabels[i] == 2 or tlabels[i] == 3:\n tL.append(tlabels[i])\n tL1.append(timvs[i])\n\n# number of test images\nnum_test = len(tL)\n\n# Append '1' to image vectors\nY = zeros((len(tL1), len(tL1[0]) + 1, 1))\n\nfor i in range(len(tL1)):\n Y[i] = (np.append([1], tL1[i])).reshape((len(tL1[0]) + 1, 1))\n\n\n# In[7]:\n\n# Accuracy percentage function\n\ndef accuracy_perc(W, N, X, lbl, target):\n W_T = np.transpose(W)\n counter = 0\n for i in range(N):\n val = np.dot(W_T, X[i])\n #if val <= -1000:\n # val = -1000\n \n p = sigmoid(val)#1/(1 + exp(val))\n \n if p >= 0.5 and lbl[i] == target:\n #print(lbl[i])\n counter = counter + 1\n elif p < 0.5 and lbl[i] != target:\n #print(lbl[i])\n counter = counter + 1\n return 100 * counter/N\n\n\n# In[8]:\n\n# Calculate accuracy percentage on test data\n\nacc = accuracy_perc(w, num_test, Y, tL, 2)\nprint(acc)\n\n\n# In[9]:\n\ndef avg_err2(W, X, minv, maxv, lbl, target):\n \n W_T = np.transpose(W)\n S = 0\n counter = 0\n y_n = 0\n t_n = 0\n \n for n in range(minv, maxv):\n \n val = np.dot(W_T, X[n])\n #if val <= -1000:\n # val = -1000\n \n y_n = sigmoid(val)#1/(1 + exp(val))\n \n counter = counter + 1\n \n if lbl[n] == target:\n t_n = 1\n if y_n == 0:\n y_n = 0.000000001\n err_n = log(y_n)\n\n else:\n t_n = 0\n if y_n == 1:\n y_n = 0.999999999\n err_n = log(1 - y_n)\n \n #err_n = t_n * log(y_n) + (1 - t_n) * log(1 - y_n)\n\n S = S + err_n\n #print(counter)\n #print(maxv-minv)\n \n #print(counter)\n return -1 * S / counter\n\n\n# In[10]:\n\n# Gradient Descent Calculation\n\ndef gradient_desc1(m, nu, T, N, X, lbl, target):\n # W_0\n W = zeros((len(X[0]), 1))\n E = zeros((m, 1))\n \n for j in range(m):\n S = zeros((len(X[0]), 1))\n for i in range(N):\n S = np.add(S, gradient(W, X, i, lbl, target))\n \n W = W - (nu/(1 + j/T)) * S\n \n E[j] = accuracy_perc(W, N, X, lbl, target)#avg_err2(W, X, 0, N, lbl, target)\n \n return W, E\n\n\n# In[11]:\n\nTr = X[:3500]\nTrl = L[:3500]\n\nHo = X[3500:]\nHol = L[3500:]\n\nTe = Y\nTel = tL\n\nprint(len(Tr))\nprint(len(Ho))\nprint(len(Te))\n\n\n# In[12]:\n\nw100, err100 = gradient_desc1(500, 0.01, 2, len(Tr), Tr, Trl, target)\n\n\n# In[13]:\n\nimport matplotlib.pyplot as plt\n\nindex = 500\nxrange = [i for i in range(index)]\n\n#for x in xrange:\ntr = err100[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('error rate')\nplt.title('Training')\nplt.plot(range(index), tr, 'g-')\nplt.axis([0, 100, 0, 100])\nplt.show()\n\n\n# In[14]:\n\nw100, err100 = gradient_desc1(500, 0.01, 2, len(Ho), Ho, Hol, target)\n\n\n# In[15]:\n\nimport matplotlib.pyplot as plt\n\nindex = 500\nxrange = [i for i in range(index)]\n\n#for x in xrange:\nho = err100[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('error rate')\nplt.title('Hold-out')\nplt.plot(range(index), ho, 'b-')\nplt.axis([0, 100, 0, 100])\nplt.show()\n\n\n# In[16]:\n\nw100, err100 = gradient_desc1(500, 0.01, 2, len(Te), Te, Tel, target)\n\n\n# In[ ]:\n\n\n\n\n# In[30]:\n\nimport matplotlib.pyplot as plt\n\nindex = 500\nxrange = [i for i in range(index)]\n\n#for x in xrange:\nte = err100[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('error rate')\nplt.title('Test')\nplt.plot(range(index), te, 'r-')\nplt.axis([1, 100, 0, 20])\nplt.show()\n\n\n# In[19]:\n\nplt.xlabel('m iterations')\nplt.ylabel('error rate')\nplt.title('Training, Hold-out, and Test')\nplt.plot(range(index), tr, 'g-')\nplt.plot(range(index), ho, 'b-')\nplt.plot(range(index), te, 'r-')\nplt.axis([1, 100, 80, 100])\nplt.show()\n\n\n# In[20]:\n\n# Gradient Descent Calculation\n\ndef gradient_desc2(m, nu, T, tr, trl, ho, hol, te, tel, target):#N, X, lbl, target):\n # W_0\n W = zeros((len(X[0]), 1))\n E = zeros((m, 1))\n E1 = zeros((m, 1))\n E2 = zeros((m, 1))\n \n for j in range(m):\n S = zeros((len(X[0]), 1))\n for i in range(len(tr)):\n S = np.add(S, gradient(W, tr, i, trl, target))\n \n W = W - (nu/(1 + j/T)) * S\n \n E[j] = avg_err2(W, tr, 0, len(tr), trl, target)\n E1[j] = avg_err2(W, ho, 0, len(ho), hol, target)\n E2[j] = avg_err2(W, te, 0, len(te), tel, target)\n \n return W, E, E1, E2\n\n\n# In[21]:\n\nw, tre, hoe, tee = gradient_desc2(500, 0.01, 2, Tr, Trl, Ho, Hol, Te, Tel, target)\n\n\n# In[22]:\n\nimport matplotlib.pyplot as plt\n\nindex = 500\nxrange = [i for i in range(index)]\n\n#for x in xrange:\ntr1 = tre[:index]\nho1 = hoe[:index]\nte1 = tee[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('loss')\nplt.title('Training, Hold-out, and Test')\nplt.plot(range(index), tr1, 'g-')\nplt.plot(range(index), ho1, 'b-')\nplt.plot(range(index), te1, 'r-')\nplt.axis([0, index, 0, 10])\nplt.show()\n\n\n# In[23]:\n\n# Gradient Descent Calculation\n\ndef gradient_desc3(m, nu, T, tr, trl, ho, hol, te, tel, target):#N, X, lbl, target):\n # W_0\n W = zeros((len(X[0]), 1))\n E = zeros((m, 1))\n E1 = zeros((m, 1))\n E2 = zeros((m, 1))\n counter = 0\n \n for j in range(m):\n S = zeros((len(X[0]), 1))\n for i in range(len(tr)):\n S = np.add(S, gradient(W, tr, i, trl, target))\n \n W = W - (nu/(1 + j/T)) * S\n \n E[j] = avg_err2(W, tr, 0, len(tr), trl, target)\n E1[j] = avg_err2(W, ho, 0, len(ho), hol, target)\n E2[j] = avg_err2(W, te, 0, len(te), tel, target)\n \n # Early stopping\n if j >= 10 and E1[j - 1] <= E[j]:\n counter = counter + 1\n \n if counter >= 3:\n return W, E, E1, E2, j\n \n return W, E, E1, E2, j\n\n\n# In[24]:\n\nw, tre, hoe, tee, ind = gradient_desc3(500, 0.01, 2, Tr, Trl, Ho, Hol, Te, Tel, target)\n\n\n# In[25]:\n\nprint(ind)\n\n\n# In[26]:\n\nimport matplotlib.pyplot as plt\n\nindex = ind\nxrange = [i for i in range(index)]\n\n#for x in xrange:\ntr1 = tre[:index]\nho1 = hoe[:index]\nte1 = tee[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('loss')\nplt.title('Training, Hold-out, and Test')\nplt.plot(range(index), tr1, 'g-')\nplt.plot(range(index), ho1, 'b-')\nplt.plot(range(index), te1, 'r-')\nplt.axis([0, index, 0, 40])\nplt.show()\n\nprint(ind)\n\n\n# In[27]:\n\n# minibatch\n\ndef minibatch(m, nu, T, tr, trl, ho, hol, te, tel, target):\n # W_0\n W = zeros((len(X[0]), 1))\n E = zeros((m, 1))\n E1 = zeros((m, 1))\n E2 = zeros((m, 1))\n counter = 0\n \n for j in range(m):\n S = zeros((len(X[0]), 1))\n for i in range(len(tr)):\n S = np.add(S, gradient(W, tr, i, trl, target))\n if i % (int)(1/10 * len(tr)) == 0:\n W = W - (nu/(1 + j/T)) * S\n S = zeros((len(X[0]), 1))\n \n W = W - (nu/(1 + j/T)) * S\n \n E[j] = avg_err2(W, tr, 0, len(tr), trl, target)\n E1[j] = avg_err2(W, ho, 0, len(ho), hol, target)\n E2[j] = avg_err2(W, te, 0, len(te), tel, target)\n \n # Early stopping\n if j >= 10 and E1[j - 1] <= E[j]:\n counter = counter + 1\n \n if counter >= 3:\n return W, E, E1, E2, j\n \n return W, E, E1, E2, j \n\n\n# In[28]:\n\nw, tre, hoe, tee, ind = minibatch(500, 0.01, 2, Tr, Trl, Ho, Hol, Te, Tel, target)\n\n\n# In[29]:\n\nimport matplotlib.pyplot as plt\n\nindex = ind\nxrange = [i for i in range(index)]\n\n#for x in xrange:\ntr1 = tre[:index]\nho1 = hoe[:index]\nte1 = tee[:index]\n\nplt.xlabel('m iterations')\nplt.ylabel('loss')\nplt.title('Training, Hold-out, and Test')\nplt.plot(range(index), tr1, 'g-')\nplt.plot(range(index), ho1, 'b-')\nplt.plot(range(index), te1, 'r-')\nplt.axis([0, index, 0, 2])\nplt.show()\n\n\n# In[ ]:\n\n\n\n","sub_path":"cse253hw1_2vs8_jean.py","file_name":"cse253hw1_2vs8_jean.py","file_ext":"py","file_size_in_byte":11206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"484279276","text":"import asyncio\nimport playground\nfrom playground.network.packet import PacketType\nfrom PacketConnect import RequestStringCompare,StringCompareQuestion,StringCompareAnswer,StringCompareResult\n\nclass ServerProtocol(asyncio.Protocol):\n\tdef __init__(self):\n\t\tself.transport=None\n\t\tself._Deserializer = PacketType.Deserializer()\n\t\tprint('server')\n\n\tdef connection_made(self,transport):\n\t\tprint('server')\n\t\tself.transport=transport\n\t\tself._Deserializer=PacketType.Deserializer()\n\t\tpeername = transport.get_extra_info('peername')\n\t\tprint('server(prepare)-->client(prepare):Connection from {}'.format(peername))\n\n\tdef data_received(self, data):\n\t\tself._Deserializer=PacketType.Deserializer()\n\t\tself._Deserializer.update(data)\n\t\tfor pkt in self._Deserializer.nextPackets():\n\t\t\tself.stringclass(pkt)\n\n\tdef stringclass(self,pkt):\n\t\tif isinstance(pkt,RequestStringCompare):\n\t\t\tpacket2=StringCompareQuestion()\n\t\t\tprint(\"Server(giving question)-->client(result): Question of asking compared string received.\")\n\t\t\tpacket2.id=100\n\t\t\tpacket2.STR1=\"YANG\"\n\t\t\tpacket2.STR2=\"LI\"\n\t\t\tprint(\"Server(giving question)-->client(result): Question of compared string sent.\")\n\t\t\tself.transport.write(packet2.__serialize__())\n\t\tif isinstance(pkt,StringCompareAnswer):\n\t\t\tpacket4=StringCompareResult()\n\t\t\tprint('Server(judge)-->client(final): Answer received and is working on the result...')\n\t\t\tpacket4.id=100\n\t\t\tprint(pkt.Answer)\n\t\t\tif pkt.Answer==\"STR1\":\n\t\t\t\tpacket4.output=\"TRUE\"\n\t\t\t\tprint('Server(judge)-->client(FINAL) : The output is {}.'.format(packet4.output))\n\t\t\tif pkt.Answer==\"STR2\":\n\t\t\t\tpacket4.output=\"FALSE\"\n\t\t\t\tprint('Server(judge)-->client(FINAL) : The output is {}.'.format(packet4.output))\n\t\t\tself.transport.write(packet4.__serialize__())\n\t\t\tself.transport.close()\n\t\t\tprint('Server(FINAL) :Close the client socket.')\n\t\t\t\n\nif __name__ == \"__main__\":\n loop = asyncio.get_event_loop()\n coro = playground.getConnector().create_playground_server(ServerProtocol,888)\n server = loop.run_until_complete(coro)\n loop.run_forever()\n loop.close()\n\n","sub_path":"netsec_fall2017/lab_1d/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":2038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"267513703","text":"import re\nimport string\n\ndef extract_tags(tags):\n whitespace = re.compile('\\s')\n nowhite = whitespace.sub(\"\", tags)\n tags_array = nowhite.split(',')\n\n cleaned = []\n for tag in tags_array:\n if tag not in cleaned and tag != \"\":\n cleaned.append(tag)\n\n return cleaned\n\ndef clean_tags(tags):\n return tags.replace(\"[\", \"\").replace(\"]\", \"\")\n","sub_path":"app/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"481133091","text":"# -*- coding: utf-8 -*-\n#! /usr/bin/env python\n\nimport numpy as np\n\n\n# 共分散行列を求める\ndef Cov(X):\n\tret = np.array([x - np.mean(x) for x in X.T]).T\n\tret = np.dot(ret.T, ret) / ret.shape[0]\n\treturn ret\n\n# 固有値問題を解いて固有値の大きい順に固有値と固有ベクトルをソート\ndef sortedEigen(X):\n\t(val, vec) = np.linalg.eig(X)\n\tvec = vec.T\n\tret_vec = []\n\tfor (key, value) in enumerate(sorted(val, reverse=True)):\n\t\tret_vec.append(vec[key])\n\tret_vec = np.array(ret_vec)\n\tret_val = np.array(sorted(val, reverse=True))\n\treturn (ret_val, ret_vec)\n\n# 線形写像\ndef linMap(X, eigvec, dim):\n\tcomponents = eigvec[:dim]\n\tret = np.dot(X, components.T)\n\treturn ret\n\n# 主成分分析\ndef PCA(X,dim):\n\tcovmat = Cov(X)\n\t(eigval,eigvec) = sortedEigen(covmat)\n\tret = linMap(X,eigvec,dim)\n\treturn ret\n\n","sub_path":"pattern/pca.py","file_name":"pca.py","file_ext":"py","file_size_in_byte":826,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"403497517","text":"#George Emanuel\n#emanuega0@gmail.com\n\nimport multiprocessing\nfrom functools import reduce\n\nimport numpy as np\nfrom scipy.optimize import least_squares\nfrom scipy.optimize import minimize_scalar\n\nfrom core.analysis import analysis\nfrom core.data import sequence\n\nclass BarcodeAggregator(object):\n\n def __init__(self, matchedBarcodes, bcToAATable):\n self.matchedBarcodes = matchedBarcodes\n self.bcToAATable = bcToAATable\n\n uniqueSequencedBarcodes = set(\n self.bcToAATable.get_unique_barcodes_as_ints())\n matchingUnique = {k: b for k,b in self.matchedBarcodes.items() \\\n if uniqueSequencedBarcodes.__contains__(b)}\n\n uniqueBarcodes, uniqueInverse = np.unique(\n list(matchingUnique.values()), return_inverse = True)\n keys = np.array(list(matchingUnique.keys()))\n\n self.groupedCells = {uniqueBarcodes[i]: \\\n [j for j in keys[uniqueInverse==i]]\\\n for i in range(len(uniqueBarcodes))}\n\n def get_cells_by_barcode(self):\n return self.groupedCells \n\n def reverse(self, x):\n return x[::-1]\n\n def get_barcodes_abundance_sorted(self, abundanceThreshold = 0):\n groupedCells = self.get_cells_by_barcode()\n\n cellCounts = {i: len(x) for i,x in groupedCells.items()}\n sortedCells = [(key, value) for key, value \\\n in sorted(cellCounts.items(), key=self.reverse, reverse=True)]\n return [k for k,v in sortedCells if v > abundanceThreshold]\n\n def partition_phenotypes(self, phenotypeDictionary, abundanceThreshold=0,\n index = None):\n groupedCells = self.get_cells_by_barcode()\n\n partitionedPhenotypes = None\n if index is None:\n partitionedPhenotypes = {k: {j: phenotypeDictionary[j] \\\n for j in v} for k,v in groupedCells.items() \\\n if len(v) > abundanceThreshold}\n else:\n partitionedPhenotypes = {k: {j: phenotypeDictionary[j][index] \\\n for j in v} for k,v in groupedCells.items() \\\n if len(v) > abundanceThreshold}\n\n return partitionedPhenotypes\n\n def median_partition_phenotypes(\n self, phenotypeDictionary, abundanceThreshold=0, index=None):\n\n partitionedPhenotypes = self.partition_phenotypes(\n phenotypeDictionary, abundanceThreshold, index)\n return {k: np.median([x for x in v.values()], axis=0) \\\n for k,v in partitionedPhenotypes.items() \\\n if len(v) > abundanceThreshold}\n\n def std_partition_phenotypes(\n self, phenotypeDictionary, abundanceThreshold=0, index=None):\n\n partitionedPhenotypes = self.partition_phenotypes(\n phenotypeDictionary, abundanceThreshold, index)\n return {k: np.std([x for x in v.values()], axis=0) \\\n for k,v in partitionedPhenotypes.items() \\\n if len(v) > abundanceThreshold}\n\n def partition_by_mutant(\n self, phenotypeDictionary, abundanceThreshold=0, index = None):\n\n partitionedByBC = self.partition_phenotypes(\n phenotypeDictionary, abundanceThreshold = abundanceThreshold,\n index = index)\n mediansByBC = self.median_partition_phenotypes(\n phenotypeDictionary, abundanceThreshold = abundanceThreshold,\n index = index)\n\n byMutant = self._sort_measurements_by_mutant(partitionedByBC,\n {k: len(v) for k,v in partitionedByBC.items()})\n\n return byMutant\n\n def median_partition_by_mutant(\n self, phenotypeDictionary, abundanceThreshold=0, index = None):\n\n byMutant = self.partition_by_mutant(phenotypeDictionary,\n abundanceThreshold = 0, index = index)\n\n\n valuesByMutant = [reduce((lambda x,y: x+y),\n [list(x[0].values()) for x in currentMutant]) \\\n for currentMutant in byMutant]\n\n mediansByMutant = [(np.median(v), [x[3] for x in y]) \\\n for y,v in zip(byMutant, valuesByMutant) \\\n if len(v)>abundanceThreshold]\n\n return mediansByMutant\n\n\n def _sort_measurements_by_mutant(self, measurements, counts):\n\n barcodes = self.bcToAATable.get_barcode_list_as_ints()\n sequences = self.bcToAATable.get_sequence_list()\n uniqueSequences, uniqueI, uniqueCounts = np.unique(\n sequences, return_inverse = True, return_counts = True)\n\n medians = [[(measurements[barcodes[j]], counts[barcodes[j]], j, \\\n self.bcToAATable.bcToAA[j])\\\n for j in np.where(uniqueI==i)[0]\\\n if barcodes[j] in measurements.keys()]\\\n for i in range(len(uniqueSequences))]\n \n medians = [x for x in medians if len(x) > 0]\n return medians\n\n\nclass PhenotypeExtractor(analysis.AbstractAnalysisTask):\n\n def __init__(self, cellIntensities, phenotypeParameters):\n self.cellIntensities = cellIntensities\n self.phenotypeParameters = phenotypeParameters\n\n self.complete = False\n self.coreCount = 1\n\n def run(self):\n if self.coreCount == 1:\n self.phenotypes = {id: self._cell_intensities_to_phenotypes(\n intensities) \\\n for id, intensities in self.cellIntensities.items()}\n else:\n pool = multiprocessing.Pool(processes = self.coreCount)\n self.phenotypes = dict(pool.map(\n self._intensities_to_id_phenotypes, \n self.cellIntensities.items()))\n\n self.complete = True\n\n def set_cpu_count(self, cpuCount):\n self.coreCount = cpuCount\n\n def _cell_intensities_to_phenotypes(self, intensities):\n return [self._extract_phenotype(intensities, p) for p \\\n in self.phenotypeParameters]\n\n def _intensities_to_id_phenotypes(self, intensityId):\n return (intensityId[0], self._cell_intensities_to_phenotypes(\n intensityId[1]))\n\n def _extract_phenotype(self, intensities, phenotypeProperties):\n if phenotypeProperties['type'] == 'Intensity':\n measuredIntensity = self._intensity_at_index(\n intensities, phenotypeProperties['index'])\n if 'normalization' in phenotypeProperties:\n normalization = self._intensity_at_index(\n intensities, phenotypeProperties['normalization'])\n measuredIntensity /= normalization\n return measuredIntensity\n\n if phenotypeProperties['type'] == 'Intensity Difference':\n intensity1 = self._intensity_at_index(\n intensities, phenotypeProperties['intensity_1'])\n intensity2 = self._intensity_at_index(\n intensities, phenotypeProperties['intensity_2'])\n difference = intensity1 - intensity2\n if 'normalization' in phenotypeProperties:\n normalization = self._intensity_at_index(\n intensities, phenotypeProperties['normalization'])\n difference /= normalization\n return difference\n\n elif phenotypeProperties['type'] == 'Recovery':\n initialIntensity = self._intensity_at_index(\n intensities, phenotypeProperties['initial'])\n finalIntensity = self._intensity_at_index(\n intensities, phenotypeProperties['final'])\n recoveredIntensity = self._intensity_at_index(\n intensities, phenotypeProperties['recovered'])\n return (recoveredIntensity-finalIntensity)/ \\\n (initialIntensity-finalIntensity)\n\n return np.NaN\n\n def _intensity_at_index(self, intensities, index):\n return intensities[index[0]][index[1]]\n\n def result(self):\n if self.complete:\n return self.phenotypes\n \n def to_string(self):\n pass\n\n\n\n\n","sub_path":"core/analysis/phenotype.py","file_name":"phenotype.py","file_ext":"py","file_size_in_byte":7974,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"383393774","text":"from html.parser import HTMLParser\nfrom urllib.request import urlopen\n\n\ndef get(url):\n r = urlopen(url)\n return r.read().decode(r.info().get_content_charset())\n\n\nclass P0(HTMLParser):\n def __init__(self):\n super().__init__()\n self.__gate = False\n self.success_builds = []\n\n def handle_starttag(self, tag, attrs):\n if attrs == [(\"class\", \"status-Success\")]:\n self.__gate = True\n elif self.__gate and tag == \"a\":\n self.success_builds.append(\"https://ci.chromium.org\" + attrs[0][1])\n self.__gate = False\n\n\nclass P1(HTMLParser):\n def __init__(self):\n super().__init__()\n self.download_link = None\n\n def handle_starttag(self, tag, attrs):\n if len(attrs) >= 2 and attrs[1][1] == \"step link for download\":\n self.download_link = attrs[0][1]\n\n\nif __name__ == \"__main__\":\n p0 = P0()\n p0.feed(get(\"https://ci.chromium.org/p/v8/builders/luci.v8.ci/V8%20Linux64%20-%20node.js%20integration\"))\n p1 = P1()\n p1.feed(get(p0.success_builds[0]))\n print(p1.download_link, end=\"\")\n","sub_path":"utils/v8-node.py","file_name":"v8-node.py","file_ext":"py","file_size_in_byte":1093,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"340086348","text":"from project.people.models import Person\nfrom rest_framework import serializers, viewsets, status\nfrom rest_framework.decorators import action\nfrom rest_framework.response import Response\nfrom rest_framework import filters\nfrom django.core.exceptions import ValidationError\n\n# Serializers\nclass PersonSerializer(serializers.ModelSerializer):\n \"\"\" Serializer for the Person model. \"\"\"\n class Meta:\n model = Person\n fields = (\n 'id',\n 'first_name',\n 'last_name',\n 'email',\n 'birthday',\n 'zip_code'\n )\n\n# Views\nclass PersonViewSet(viewsets.ModelViewSet):\n \"\"\" This viewset provides API routes for the Person model.\n Provides get, put, patch and delete method handlers. \"\"\"\n\n queryset = Person.objects.all()\n serializer_class = PersonSerializer\n filter_backends = (filters.OrderingFilter,)\n ordering_fields = '__all__'\n ordering = ('first_name',)\n\n\n # DELETE /api/people//\n @action(detail=True, methods=['delete'])\n def delete_person(self, request, pk=None):\n \"\"\" Delete a single person object. \"\"\"\n person = self.get_object(pk)\n person.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n\n # POST /api/people/\n def create(self, request):\n \"\"\" Create a person object. \"\"\"\n person_data = request.data\n new_obj = Person.objects.create(**person_data)\n try:\n new_obj.save()\n serializer = self.serializer_class(new_obj, context={\"request\": request})\n return Response(serializer.data, status=status.HTTP_200_OK)\n except ValidationError as error:\n Response(error, status=status.HTTP_400_BAD_REQUEST)\n","sub_path":"project/api/api_people.py","file_name":"api_people.py","file_ext":"py","file_size_in_byte":1742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"57981615","text":"\nimport numpy as np\n\n# Load the data from the file and return a list of lists, with each line as a list\ndef load_data(filename):\n file = open(filename, 'r')\n lines = file.readlines()\n dataList = []\n \n for line in lines:\n lineList = line.split()\n fLineList = [float(i) for i in lineList]\n dataList.append(fLineList)\n return dataList\n\n# Returns a set of unique values for a column in the data\ndef unique(rows, col):\n return set([row[col] for row in rows])\n\n# Class that holds the column number and value of a feature_split\nclass Feature_Split:\n\n def __init__(self, column, value):\n self.column = column\n self.value = value\n\n # Compares this feature split to an example\n def compare(self, example):\n val = example[self.column]\n return val >= self.value\n \n def __repr__(self):\n return \"Is %s %s %s\" % (str(self.column), \">=\", str(self.value))\n\ndef partition(rows, split):\n true_rows, false_rows = [], []\n for row in rows:\n if split.compare(row):\n true_rows.append(row)\n else:\n false_rows.append(row)\n return true_rows, false_rows\n\n# Counts the number of labels for each label, and the total number of labels\ndef num_labels(rows):\n counts = {}\n for row in rows:\n label = row[-1]\n if label not in counts:\n counts[label] = 0\n counts[label] += 1\n return counts, len(rows)\n\ndef entropy(true_rows, false_rows):\n \n # Calculate conditional entropy on true\n true_probs = {}\n true_counts, true_total = num_labels(true_rows)\n\n # Convert the counts into probabilities\n for label in true_counts:\n true_probs[label] = float(true_counts[label]) / true_total\n \n true_entropy = 0.0\n for label in true_probs:\n p = true_probs[label]\n if p != 0.0:\n true_entropy -= p * np.log(p)\n\n # Calculate conditional entropy on false\n false_probs = {}\n false_counts, false_total = num_labels(false_rows)\n\n # Convert the counts into probabilities\n for label in false_counts:\n false_probs[label] = float(false_counts[label]) / false_total\n \n false_entropy = 0.0\n for label in false_probs:\n p = false_probs[label]\n if p != 0.0:\n false_entropy -= p * np.log(p)\n\n total = true_total + false_total\n\n split_entropy = (true_total / total) * true_entropy + (false_total / total) * false_entropy\n\n return split_entropy\n\n# Find the best split\ndef find_best_split(rows):\n lowest_entropy = 1\n best_split = None\n num_cols = len(rows[0]) - 1 # Subtract one to account for label\n\n for col in range(num_cols):\n values = unique(rows, col)\n \n for val in values:\n split = Feature_Split(col, val)\n \n true_rows, false_rows = partition(rows, split)\n\n # If the split doesn't divide any of the dataset, skip it\n if len(true_rows) == 0 or len(false_rows) == 0:\n continue\n\n # Calculate entropy of the split\n ent = entropy(true_rows, false_rows)\n\n if ent < lowest_entropy:\n lowest_entropy = ent\n best_split = split\n return lowest_entropy, best_split\n","sub_path":"pa2decisiontree.py","file_name":"pa2decisiontree.py","file_ext":"py","file_size_in_byte":3258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"602123204","text":"\r\n \r\n\r\ndef tableau(tab):\r\n for k in range(len(tab)):\r\n if k%10==9:\r\n print(tab[k])\r\n else:\r\n print(tab[k],\"\\t\",end=\" \")\r\n\r\ndef Fusion(t1,t2):\r\n res=[ ]\r\n m1=n1=0\r\n while m1 1 :\r\n if len(T) == 2 :\r\n if T[0]>T[1] : T[0],T[1] = T[1],T[0]\r\n else :\r\n m = len(T)/2\r\n T = Fusion(TriFusion(T[:m]),TriFusion(T[m:]))\r\n return T\r\n\r\nfrom random import*\r\nn=int(input (\"entrez le nombre n d'éléments à trier \"))\r\ntab1=[]\r\nfor k in range(n):\r\n tab1=tab1+[randrange(1,100)]\r\n if k%10==9:\r\n print(tab1[k])\r\n else:\r\n print (tab1[k],\"\\t\",end=\" \")\r\n\r\n\r\nn=int(input (\"entrez le nombre n d'éléments à trier \"))\r\ntab2=[]\r\nfor k in range(n):\r\n tab2=tab2+[randrange(1,100)]\r\n if k%10==9:\r\n print(tab2[k])\r\n else:\r\n print (tab2[k],\"\\t\",end=\" \")\r\n\r\ntableau(tab1)\r\ntableau(tab2)\r\nTriFusion(tab1)\r\n\r\n","sub_path":"insertiondouble.py","file_name":"insertiondouble.py","file_ext":"py","file_size_in_byte":1259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"406959177","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May 12 15:04:33 2016\n\n@author: snoran\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\nfname = \"../data/sample/RSSI.csv\"\nrssi = pd.read_csv(fname,sep=',').values\n\nplt.plot(rssi[:,1])\nplt.show()\n\nfreqInMHz = 2462\n#levelInDb = -83\nSNR = -87\n\nresult = (27.55 - (20 * np.log10(freqInMHz)) + np.fabs(rssi)) / 20.0\nmeters = np.power(10, result)\n\nfeet = meters * 3.2808","sub_path":"analysis/read_data.py","file_name":"read_data.py","file_ext":"py","file_size_in_byte":433,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"78533476","text":"\nimport random\nimport sys\nimport os\n\nimport dumbTemplate\n\ndef page_initialise(template_filename):\n\n ROOT_DIR = \"/www/no_ssl/\"\n TEMPLATE_DIR = \"/www/no_ssl/templates/\"\n\n #First need to print this\n print(\"Content-Type: text/html\")\n print(\"\")\n\n #Open page template\n template = dumbTemplate.dumbTemplate(TEMPLATE_DIR + template_filename)\n if not template:\n print(\"Couldn't open webpage??!1\")\n return\n\n #Open side_bar.template and put it into webpage\n side_bar_file = open(TEMPLATE_DIR + \"side_bar.template\", 'r', -1, 'utf-8', 'replace')\n side_bar_html = side_bar_file.read()\n side_bar_file.close()\n template.substitute(\"SIDE_BAR\", side_bar_html)\n\n #pick a random image of billd and put into webpage(sidebar)\n try:\n billd_file = random.choice(os.listdir(ROOT_DIR + \"webpage-files/billd/\"))\n template.substitute(\"BILLD\", \"/webpage-files/billd/\" + billd_file)\n except IndexError:\n billd_file = \"\"\n\n #randomise position of billd at the bottom\n dice = random.randint(0, 2)\n if(dice == 0):\n billd2 = \"\"\"
\n
\"\"\".format(random.randint(0, 130))\n elif(dice == 1):\n billd2 = \"\"\"
\n
\"\"\".format(random.randint(0, 50))\n elif(dice == 2):\n billd2 = \"\"\"
\n \"\"\".format(random.randint(0, 200))\n\n template.substitute(\"BILLD2\", billd2)\n\n return(template)\n\n\n\nif __name__ == \"__main__\":\n print(\"wtf\")\n\n","sub_path":"lib/webpageCommon.py","file_name":"webpageCommon.py","file_ext":"py","file_size_in_byte":2118,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"561316431","text":"class DVDScreenSaver:\n def __init__(self,\n size: tuple,\n clip_size: tuple,\n start_position: int = 0,\n speed: int = 5):\n self._W, self._H = size\n self._w, self._h = clip_size\n self._start_position = start_position\n self._xV = speed\n self._yV = speed\n self._x = start_position\n self._y = start_position\n\n def calculate_position(self, t: int) -> tuple:\n \"\"\"Returns x, y position given the time\n\n :type t: int\n :param t: current time\n \"\"\"\n self._x += self._xV\n self._y += self._yV\n self.detect_collision()\n return self._x, self._y\n\n def detect_collision(self):\n \"\"\"Detects collision with the outer walls and prevents it. \"\"\"\n if (self._y + self._h) == self._H:\n self._yV = -self._yV\n if (self._x + self._w) == self._W:\n self._xV = -self._xV\n if self._x == 0:\n self._xV = -self._xV\n if self._y == 0:\n self._yV = -self._yV\n","sub_path":"animations/dvd_screen_saver.py","file_name":"dvd_screen_saver.py","file_ext":"py","file_size_in_byte":1074,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"69732516","text":"#!/usr/bin/env python\r\n\r\ndict_yyz = dict()\r\nlist_names = ['bob','joe','harry','bob','bob','harry']\r\nfor name in list_names :\r\n dict_yyz[name] = dict_yyz.get(name, 0) + 1\r\nprint(dict_yyz)\r\n\r\n\r\nf=open('myfile.txt')\r\ndict_a=dict()\r\nfor line in f:\r\n line = line.rstrip()\r\n print(line)\r\n words = line.split()\r\n for word in words :\r\n dict_a[word] = dict_a.get(word, 0) + 1\r\n\r\nprint(dict_a)\r\n\r\nbiggestword = None\r\nbiggestcount = 0\r\nfor key,val in dict_a.items() :\r\n if biggestword is None or val > biggestcount :\r\n biggestword = key\r\n biggestcount = val\r\n\r\nprint(f'The biggest word is { biggestword } with count of: { biggestcount }')\r\n","sub_path":"dict.py","file_name":"dict.py","file_ext":"py","file_size_in_byte":668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"552339155","text":"import pandas as pd\nimport numpy as np\n\n_VALID_ANOM_TYPES = ['median', 'mean', 'frac']\n\n\ndef load_temp_precip_data(crop: str, season: str, country, regions: list, month_indexes, anom_type='mean'):\n anom_type = anom_type.lower()\n if anom_type not in _VALID_ANOM_TYPES:\n raise ValueError(f\"{anom_type} is not a valid anomaly type.\"\n f\"Must be one of {anom_type}\")\n\n crop_season_country = [crop, season, country] if season != '' \\\n else [crop, country]\n crop_season_country = '_'.join(crop_season_country)\n\n # Read in climate temperatures\n clim_temp_crop = pd.read_csv(f'./Crop_data_files/clim_file/temp_climatology_{crop}.csv', sep='\\t')\n clim_temp_crop.rename(columns={'Unnamed: 0': 'Crop_season_location'}, inplace=True)\n # Read in climate precipitation\n clim_precip_crop = pd.read_csv(f'./Crop_data_files/clim_file/precip_climatology_{crop}.csv', sep='\\t')\n clim_precip_crop.rename(columns={'Unnamed: 0': 'Crop_season_location'}, inplace=True)\n # Read in Yields\n yields = pd.read_csv(f'./Crop_data_files/{crop}_{anom_type}_yield_anoms.csv', sep='\\t')\n years = None\n # Read in and add back mean temperature to get real temperature values\n temp_regions = []\n for i, region in enumerate(regions):\n maize_temp = pd.read_csv(f'./Crop_data_files/{crop}_met_anoms/{crop_season_country}'\n f'_{region}_temp_anom_real.csv', sep='\\t')\n maize_temp.rename(columns={'Unnamed: 0': 'Year'}, inplace=True)\n if years is None: # we need to know which yield years we have climatology data for\n years = maize_temp['Year'].apply(str).values\n means = clim_temp_crop[clim_temp_crop['Crop_season_location']\n == f'{crop_season_country}_{regions[i]}'].iloc[0, 1:, ]\n tmp = maize_temp.iloc[:, 1:].add(means)\n temp_regions.append(tmp)\n temp_regions = pd.concat(temp_regions, keys=regions)\n\n # Read in and add back mean precipitation to get real precipitation values\n pecip_regions = []\n for i, region in enumerate(regions):\n maize_precip = pd.read_csv(\n f'./Crop_data_files/{crop}_met_anoms/{crop_season_country}_{region}_precip_anom_real.csv', sep='\\t')\n maize_precip.rename(columns={'Unnamed: 0': 'Year'}, inplace=True)\n means = clim_precip_crop[clim_precip_crop['Crop_season_location']\n == f'{crop_season_country}_{regions[i]}'].iloc[0, 1:, ]\n tmp = maize_precip.iloc[:, 1:].add(means)\n pecip_regions.append(tmp)\n pecip_regions = pd.concat(pecip_regions, keys=regions)\n\n n_years = yields[years].shape[1]\n n_months = len(month_indexes)\n n_regions = len(regions)\n\n d_yields = yields[yields[\"Region\"].isin(\n [f'{crop_season_country}_{region}' for region in regions])][years]\n missing_year_inds = d_yields.isna().any().to_numpy().nonzero()[0]\n n_years -= len(missing_year_inds)\n d_yields.dropna(axis=1, inplace=True)\n\n # Drop years with missing yield values\n d_temp = np.array(temp_regions.iloc[:, month_indexes]).reshape(\n n_regions, -1, n_months).astype(float)\n d_precip = np.array(pecip_regions.iloc[:, month_indexes]).reshape(\n n_regions, -1, n_months).astype(float)\n d_temp = np.delete(d_temp, missing_year_inds, axis=1)\n d_precip = np.delete(d_precip, missing_year_inds, axis=1)\n\n offset = 1 if anom_type == 'frac' else 9.75\n data = {\n 'n_regions': n_regions,\n 'n_years': n_years,\n 'n_months': n_months,\n 'd_temp': d_temp,\n 'd_precip': d_precip,\n 'd_yields': np.array(d_yields).astype(float) + offset,\n # adjust for current values (this adjustment factor only applies to USA Maize)\n 'n_gf': 40,\n 'temp': np.arange(0, 40, 1),\n 'precip': np.arange(0, 200, 5),\n }\n\n return data\n\n\ndef batch_data(data):\n x_cat = np.concatenate((data['d_temp'], data['d_precip']), -1)\n x_annual = x_cat.reshape(data['n_years'] * data['n_regions'], -1)\n\n batched_data = {\n 'X': x_annual,\n 'y': np.array(data['d_yields']).flatten()\n }\n\n return batched_data\n\n\ndef extract_data_by_year_index(data, indexes):\n data['n_years'] = len(indexes)\n temporal_data = ['d_temp', 'd_precip', 'd_yields']\n new_data = {key: value[:, indexes] if key in temporal_data else value for key, value in data.items()}\n return new_data\n","sub_path":"utils/data_loading.py","file_name":"data_loading.py","file_ext":"py","file_size_in_byte":4438,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"62769451","text":"import pygame\nimport os\n\nfrom ..globals import ASSETS_DIR\n\n\nclass Entity:\n \"\"\"Class that represents entities in Treasure Keeper.\"\"\"\n def __init__(self, pos, board, sprite_fname, name):\n self.pos = pos\n self.board = board\n self.set_sprite(sprite_fname)\n self.rect = self.sprite.get_rect()\n self.name = name\n\n def set_sprite(self, sprite_fname):\n \"\"\"Change/set the entity's sprite.\"\"\"\n sprite = pygame.image.load(os.path.join(ASSETS_DIR, sprite_fname))\n self.sprite = sprite\n","sub_path":"treasurekeeper/entity/entity.py","file_name":"entity.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"342068929","text":"import time\nimport math\nimport random\nimport subprocess\nimport digitalio\nimport board\nfrom PIL import Image, ImageDraw, ImageFont\nimport adafruit_rgb_display.st7789 as st7789\nfrom time import strftime, sleep\nfrom datetime import datetime as dt\nfrom datetime import date, timedelta\n\n# Configuration for CS and DC pins (these are FeatherWing defaults on M0/M4):\ncs_pin = digitalio.DigitalInOut(board.CE0)\ndc_pin = digitalio.DigitalInOut(board.D25)\nreset_pin = None\n\n# Config for display baudrate (default max is 24mhz):\nBAUDRATE = 64000000\n\n# Setup SPI bus using hardware SPI:\nspi = board.SPI()\n\n# Create the ST7789 display:\ndisp = st7789.ST7789(\n spi,\n cs=cs_pin,\n dc=dc_pin,\n rst=reset_pin,\n baudrate=BAUDRATE,\n width=135,\n height=240,\n x_offset=53,\n y_offset=40,\n)\n\nbacklight = digitalio.DigitalInOut(board.D22)\nbacklight.switch_to_output()\nbacklight.value = True\nbuttonA = digitalio.DigitalInOut(board.D23)\nbuttonB = digitalio.DigitalInOut(board.D24)\nbuttonA.switch_to_input()\nbuttonB.switch_to_input()\n\n# Create blank image for drawing.\n# Make sure to create image with mode 'RGB' for full color.\nheight = disp.width # we swap height/width to rotate it to landscape!\nwidth = disp.height\nimage = Image.new(\"RGB\", (width, height))\nrotation = 90\n\n# Get drawing object to draw on image.\ndraw = ImageDraw.Draw(image)\n\n# Draw a black filled box to clear the image.\ndraw.rectangle((0, 0, width, height), outline=0, fill=(0, 0, 0))\ndisp.image(image, rotation)\n# Draw some shapes.\n# First define some constants to allow easy resizing of shapes.\npadding = -2\ntop = padding\nbottom = height - padding\n# Move left to right keeping track of the current x position for drawing shapes.\nx = 0\n\n# Alternatively load a TTF font. Make sure the .ttf font file is in the\n# same directory as the python script!\n# Some other nice fonts to try: http://www.dafont.com/bitmap.php\nfont = ImageFont.truetype(\"/usr/share/fonts/truetype/dejavu/DejaVuSans.ttf\", 20)\nfont1 = ImageFont.truetype(\"zawijasy.otf\", 30)\nfont2 = ImageFont.truetype(\"BEECH___.TTF\", 80)\nfont3 = ImageFont.truetype(\"CheerfulYellow.ttf\", 30)\n#font4 = ImageFont.truetype(\"Kind and Rich - Personal Use.otf\", 30)\n\n# Turn on the backlight\nbacklight = digitalio.DigitalInOut(board.D22)\nbacklight.switch_to_output()\nbacklight.value = True\n\nwhile True:\n if buttonA.value and buttonB.value:\n d = dt.now(tz=None)\n fri = 4 - d.weekday()\n time_change = timedelta(days=fri)\n now = dt.now()\n seconds_since_midnight = (now - now.replace(hour=0, minute=0, second=0, microsecond=0)).total_seconds()\n time_change_2 = timedelta(seconds=seconds_since_midnight)\n next_friday = time_change + d - time_change_2\n # Draw a black filled box to clear the image.\n draw.rectangle((0, 0, width, height), outline=0, fill=0)\n\n #TODO: fill in here. You should be able to look in cli_clock.py and stats.py\n DATE = strftime(\"%m/%d/%Y %H:%M:%S\")\n s0 = str(fri) + \" days\"\n s1 = str(math.trunc((next_friday - d).total_seconds()/60)) + \" m\"\n s2 = str(math.trunc((next_friday - d).total_seconds())) + \" s\"\n \n y = top\n draw.text((x, y), DATE, font=font, fill=\"#fab300\")\n\n y += font.getsize(DATE)[1]\n y += font.getsize(DATE)[1]\n draw.text((x, y), \"Days to Friday: \", font=font, fill=\"#fc7200\")\n\n y += font.getsize(DATE)[1]\n draw.text((x, y), s0, font=font, fill=\"#eb4034\")\n\n # y += font.getsize(DATE)[1]\n # draw.text((x, y), \"Minutes to Friday: \", font=font, fill=\"#fc7200\")\n \n # y += font.getsize(s0)[1]\n # draw.text((x, y), s1, font=font, fill=\"#fc7200\")\n y += font.getsize(DATE)[1]\n draw.text((x, y), \"Seconds to Friday: \", font=font, fill=\"#fc7200\")\n \n y += font.getsize(s1)[1]\n draw.text((x, y), s2, font=font, fill=\"#eb4034\")\n\n \n # Display image.\n disp.image(image, rotation)\n time.sleep(1)\n\n else:\n if buttonB.value and not buttonA.value:\n tasks = [\"Cook Some Food!\", \"Go Hiking!\", \"Go Shopping!\", \"Watch a Movie!\", \"Watch YouTube!\", \"Read a book!\", \"Take a Nap!\"]\n colors = [\"#eb4034\", \"#ff7700\", \"#fffb00\", \"#bbff00\", \"#00ff84\", \"#00f7ff\", \"#ff00e6\"]\n ri = random.randint(0,6)\n ri_2 = random.randint(0,6)\n\n draw.rectangle((0, 0, width, height), outline=0, fill=0)\n y = top\n draw.text((x+15, y), \"Random Tasks!\", font=font1, fill=colors[random.randint(0,6)])\n y = top + 50\n draw.text((x+15, y), tasks[ri], font=font3, fill=colors[random.randint(0,6)])\n disp.image(image, rotation)\n time.sleep(1)\n if buttonA.value and not buttonB.value:\n ri_2 = random.randint(0,6)\n colors = [\"#eb4034\", \"#ff7700\", \"#fffb00\", \"#bbff00\", \"#00ff84\", \"#00f7ff\", \"#ff00e6\"]\n ri_3 = random.randint(1,6)\n\n draw.rectangle((0, 0, width, height), outline=0, fill=0)\n\n y = top\n draw.text((x, y), \"Roll Dice!\", font=font3, fill=colors[random.randint(0,6)])\n y = top + 50\n draw.text((x+100, y), str(ri_3), font=font2, fill=colors[random.randint(0,6)])\n disp.image(image, rotation)\n time.sleep(1)\n\n \n","sub_path":"Lab 2/screen_clock.py","file_name":"screen_clock.py","file_ext":"py","file_size_in_byte":5306,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"54205879","text":"from zoundry.appframework.resources.resourceutils import ZMappedImageList\r\nfrom zoundry.appframework.ui.events.listevents import ZEVT_POPUP_LIST_SELECTION\r\nfrom zoundry.appframework.ui.util.fontutil import getTextDimensions\r\nfrom zoundry.appframework.ui.widgets.controls.listex import IZListViewExContentProvider\r\nfrom zoundry.appframework.ui.widgets.controls.listex import ZListViewEx\r\nfrom zoundry.appframework.ui.widgets.controls.listex import ZPopupListView\r\nimport wx\r\n\r\n# ------------------------------------------------------------------------------\r\n# Content provider for listing all of the tabs in the drop-down list.\r\n# ------------------------------------------------------------------------------\r\nclass ZTabSelectionListContentProvider(IZListViewExContentProvider):\r\n \r\n def __init__(self, tabs):\r\n self.tabs = tabs\r\n \r\n self.imageList = self._createImageList()\r\n # end __init__()\r\n\r\n def _createImageList(self):\r\n imageList = ZMappedImageList()\r\n for tab in self.tabs:\r\n bitmap = tab.getTabInfo().getBitmap()\r\n if bitmap is not None:\r\n imageList.addImage(unicode(tab.getTabId()), bitmap)\r\n return imageList\r\n # end _createImageList()\r\n\r\n def getImageList(self):\r\n return self.imageList\r\n # end getImageList()\r\n\r\n def getNumColumns(self):\r\n return 1\r\n # end getNumColumns()\r\n\r\n def getNumRows(self):\r\n return len(self.tabs)\r\n # end getNumRows()\r\n\r\n def getColumnInfo(self, columnIndex): #@UnusedVariable\r\n return (u\"\", None, None, ZListViewEx.COLUMN_RELATIVE, 100) #$NON-NLS-1$\r\n # end getColumnInfo()\r\n\r\n def getRowText(self, rowIndex, columnIndex): #@UnusedVariable\r\n return self.tabs[rowIndex].getTabInfo().getTitle()\r\n # end getRowText()\r\n\r\n def getRowImage(self, rowIndex, columnIndex): #@UnusedVariable\r\n tab = self.tabs[rowIndex]\r\n return self.imageList[unicode(tab.getTabId())]\r\n # end getRowImage()\r\n\r\n# end ZTabSelectionListContentProvider\r\n\r\n\r\n# ------------------------------------------------------------------------------\r\n# This class implements a popup window that is used to select a tab from a \r\n# list of tabs. This is useful when there are more tabs than can be shown\r\n# on the tab bar.\r\n# ------------------------------------------------------------------------------\r\nclass ZTabSelectionPopupWindow(wx.PopupTransientWindow):\r\n\r\n def __init__(self, parent, tabs):\r\n self.parent = parent\r\n self.tabs = tabs\r\n \r\n wx.PopupTransientWindow.__init__(self, parent, style = wx.SIMPLE_BORDER)\r\n\r\n self.SetBackgroundColour(u\"#FFFFE6\") #$NON-NLS-1$\r\n \r\n self._createWidgets()\r\n self._layoutWidgets()\r\n self._bindWidgetEvents()\r\n \r\n self.SetSize(self.GetBestSize())\r\n self.Layout()\r\n # end __init__()\r\n \r\n def _createWidgets(self):\r\n # FIXME (EPW) add a filter/nav bar to the popup to narrow down the search\r\n self.provider = ZTabSelectionListContentProvider(self.tabs)\r\n self.tabList = ZPopupListView(self.provider, self, style = wx.LC_VIRTUAL | wx.LC_REPORT | wx.LC_SINGLE_SEL | wx.LC_NO_HEADER | wx.NO_BORDER)\r\n self.tabList.SetBackgroundColour(u\"#FFFFE6\") #$NON-NLS-1$\r\n # end _createWidgets()\r\n \r\n def _layoutWidgets(self):\r\n self.tabList.SetSizeHints(self._getPreferredWidth(), self._getPreferredHeight())\r\n\r\n self.sizer = wx.BoxSizer(wx.VERTICAL)\r\n self.sizer.Add(self.tabList, 1, wx.EXPAND | wx.ALL, 5)\r\n\r\n self.SetSizer(self.sizer)\r\n self.SetAutoLayout(True)\r\n self.Layout()\r\n # end _layoutWidgets()\r\n \r\n def _bindWidgetEvents(self):\r\n self.Bind(ZEVT_POPUP_LIST_SELECTION, self.onTabSelected, self.tabList)\r\n # end _bindWidgetEvents()\r\n\r\n def onTabSelected(self, event):\r\n idx = event.getListId()\r\n tab = self.tabs[idx]\r\n self.parent._fireSelectedEvent(tab)\r\n self.Dismiss()\r\n event.Skip()\r\n # end onTabSelected()\r\n \r\n def _getPreferredWidth(self):\r\n largestW = 50\r\n largestBmpW = 0\r\n for tab in self.tabs:\r\n title = tab.getTabInfo().getTitle()\r\n (w, h) = getTextDimensions(title, self.tabList) #@UnusedVariable\r\n largestW = max(w + 10, largestW)\r\n bitmap = tab.getTabInfo().getBitmap()\r\n if bitmap is not None:\r\n largestBmpW = max(largestBmpW, bitmap.GetWidth())\r\n return largestW + largestBmpW\r\n # end _getPreferredWidth()\r\n \r\n def _getPreferredHeight(self):\r\n # FIXME (EPW) this logic is not guaranteed to work well for all operating systems\r\n totalH = 0\r\n imageH = 0\r\n imageList = self.provider.getImageList()\r\n if imageList is not None:\r\n (ilW, imageH) = imageList.GetSize(0) #@UnusedVariable\r\n for tab in self.tabs:\r\n (tabW, tabH) = getTextDimensions(tab.getTabInfo().getTitle(), self.tabList) #@UnusedVariable\r\n tabH = max(tabH, imageH)\r\n totalH += tabH + 2\r\n return totalH\r\n # end _getPreferredHeight()\r\n\r\n# end ZTabSelectionPopupWindow\r\n\r\n","sub_path":"src/python/zoundry/appframework/ui/widgets/controls/advanced/support/tabselpopup.py","file_name":"tabselpopup.py","file_ext":"py","file_size_in_byte":5204,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"172054392","text":"import json\nimport time\nimport pickle\nimport gensim\nfrom textblob.classifiers import NaiveBayesClassifier\ntraining_data=[]\nmodel=gensim.models.Word2Vec.load('model.txt')\nwith open(\"para0.json\", encoding='utf8') as f:\n\tdata=json.load(f)\n\tfor i in data:\n\t\ttry:\n\t\t\tx=model[i]\n\t\t\ttmp=(i,str(0))\n\t\t\ttraining_data.append(tmp)\n\t\texcept:\n\t\t\tcontinue\nwith open(\"para3.json\", encoding='utf8') as f:\n\tdata=json.load(f)\n\tfor i in data:\n\t\ttry:\n\t\t\tx=model[i]\n\t\t\ttmp=(i,str(0.3))\n\t\t\ttraining_data.append(tmp)\n\t\texcept:\n\t\t\tcontinue\nwith open(\"para6.json\", encoding='utf8') as f:\n\tdata=json.load(f)\n\tfor i in data:\n\t\ttry:\n\t\t\tx=model[i]\n\t\t\ttmp=(i,str(0.6))\n\t\t\ttraining_data.append(tmp)\n\t\texcept:\n\t\t\tcontinue\nwith open(\"para10.json\", encoding='utf8') as f:\n\tdata=json.load(f)\n\tfor i in data:\n\t\ttry:\n\t\t\tx=model[i]\n\t\t\ttmp=(i,str(1))\n\t\t\ttraining_data.append(tmp)\n\t\texcept:\n\t\t\tcontinue\nx=[]\nfor i in range(0,500):\n x.append(training_data[i])\n x.append(training_data[i+16000])\n x.append(training_data[i+20000])\n x.append(training_data[i+22000])\n\nprint(time.asctime( time.localtime(time.time()) ))\nmodel = NaiveBayesClassifier(x)\nprint(time.asctime( time.localtime(time.time()) ))\n'''\nsave_classifier = open(\"naivebayes.pickle\",\"wb\")\npickle.dump(model, save_classifier)\nsave_classifier.close()\n'''","sub_path":"baye_text.py","file_name":"baye_text.py","file_ext":"py","file_size_in_byte":1284,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"166989282","text":"from datetime import datetime\nimport warnings\n\nimport numpy as np\nimport pandas as pd\nfrom pandas import DataFrame\nimport matplotlib.pyplot as plt\nfrom statsmodels.stats.diagnostic import acorr_ljungbox\nfrom statsmodels.tsa.stattools import adfuller as ADF\nfrom statsmodels.graphics.tsaplots import plot_acf, plot_pacf\nimport talib\n\nfrom vnpy.trader.constant import Exchange, Interval\nfrom vnpy.trader.database import database_manager\nwarnings.filterwarnings(\"ignore\")\n\n\nclass DataAnalysis:\n\n def __init__(self):\n \"\"\"\"\"\"\n self.symbol = \"\"\n self.exchange = None\n self.interval = None\n self.start = None\n self.end = None\n self.rate = 0.0\n\n self.window_volatility = 20\n self.window_index = 20\n\n self.orignal = pd.DataFrame()\n\n self.index_1to1 = []\n self.index_2to2 = []\n self.index_3to1 = []\n self.index_2to1 = []\n self.index_4to1 = []\n self.intervals = []\n\n self.results = {}\n\n def load_history(\n self,\n symbol: str,\n exchange: Exchange,\n interval: Interval,\n start: datetime,\n end: datetime,\n rate: float = 0.0,\n index_1to1: list = None,\n index_2to2: list = None,\n index_3to1: list = None,\n index_2to1: list = None,\n index_4to1: list = None,\n window_index: int = 20,\n window_volatility: int = 20,\n\n ):\n \"\"\"\"\"\"\n output(\"开始加载历史数据\")\n\n self.window_volatility = window_volatility\n self.window_index = window_index\n self.rate = rate\n self.index_1to1 = index_1to1\n self.index_2to2 = index_2to2\n self.index_3to1 = index_3to1\n self.index_2to1 = index_2to1\n self.index_4to1 = index_4to1\n\n # Load history data from database\n bars = database_manager.load_bar_data(\n symbol=symbol,\n exchange=exchange,\n interval=interval,\n start=start,\n end=end,\n\n )\n\n output(f\"历史数据加载完成,数据量:{len(bars)}\")\n\n # Generate history data in DataFrame\n t = []\n o = []\n h = []\n l = [] # noqa\n c = []\n v = []\n r = []\n\n for i in range(1,len(bars)):\n time = bars[i].datetime\n open_price = bars[i].open_price\n high_price = bars[i].high_price\n low_price = bars[i].low_price\n close_price = bars[i].close_price\n volume = bars[i].volume\n ret = np.log(bars[i]/bars[i-1])\n\n t.append(time)\n o.append(open_price)\n h.append(high_price)\n l.append(low_price)\n c.append(close_price)\n v.append(volume)\n r.append(ret)\n\n self.orignal[\"open\"] = o\n self.orignal[\"high\"] = h\n self.orignal[\"low\"] = l\n self.orignal[\"close\"] = c\n self.orignal[\"volume\"] = v\n self.orignal[\"time\"] = t\n self.orignal[\"return\"] = r\n\n def base_analysis(self, df: DataFrame = None):\n \"\"\"\"\"\"\n if df is None:\n df = self.orignal\n\n if df is None:\n output(\"数据为空,请输入数据\")\n\n output(\"检验空值\")\n nan_num = df.isnull().sum()\n output(f\"总体空值为 {nan_num}\")\n\n close_price = df[\"close\"]\n retrun_series = df[\"return\"]\n\n output(\"画出收盘价行情图,检查数据断点\")\n\n #close_price.plot(figsize=(20, 8), title=\"close_price\")\n plt.figure(figsize=(20,8))\n plt.plot(range(0,len(close_price)), close_price, title=\"close_price\")\n plt.show()\n\n output(\"画出收益率图,检查数据断点\")\n plt.figure(figsize=(20,8))\n plt.plot(range(0,len(retrun_series)), retrun_series, title=\"retrun_series\")\n plt.show()\n\n\n random_test(retrun_series)\n stability_test(retrun_series)\n autocorrelation_test(retrun_series)\n self.relative_volatility_analysis(retrun_series)\n #self.growth_analysis(df)\n #self.trend_analysis(df)\n\n #self.calculate_index(df)\n\n return df\n\n def relative_volatility_analysis(self, df: DataFrame = None):\n \"\"\"\n 相对波动率\n \"\"\"\n output(\"第五步:相对波动率分析\")\n df[\"volatility\"] = talib.ATR(\n np.array(df[\"high\"]),\n np.array(df[\"low\"]),\n np.array(df[\"close\"]),\n self.window_volatility\n )\n\n df[\"fixed_cost\"] = df[\"close\"] * self.rate\n df[\"relative_vol\"] = df[\"volatility\"] - df[\"fixed_cost\"]\n\n #df[\"relative_vol\"].plot(figsize=(20, 6), title=\"relative volatility\")\n plt.figure(figsize=(20,6))\n plt.plot(range(0,len(df[\"relative_vol\"])),df[\"relative_vol\"])\n plt.show()\n\n df[\"relative_vol\"].hist(bins=200, figsize=(20, 6), grid=False)\n plt.show()\n\n statitstic_info(df[\"relative_vol\"])\n\n def growth_analysis(self, df: DataFrame = None):\n \"\"\"\n 百分比K线变化率\n \"\"\"\n output(\"第六步:变化率分析\")\n df[\"pre_close\"] = df[\"close\"].shift(1).fillna(0)\n df[\"g%\"] = 100 * (df[\"close\"] - df[\"pre_close\"]) / df[\"close\"]\n\n #df[\"g%\"].plot(figsize=(20, 6), title=\"growth\", ylim=(-5, 5))\n plt.figure(figsize=(20,6))\n plt.ylim(-5,5)\n plt.plot(range(0,len(df[\"g%\"])),df[\"g%\"])\n plt.show()\n\n df[\"g%\"].hist(bins=1000, figsize=(20, 6), grid=False)\n plt.xlim(-10,10)\n \n plt.show()\n\n statitstic_info(df[\"g%\"])\n\n def calculate_index(self, df: DataFrame = None):\n \"\"\"\"\"\"\n output(\"第七步:计算相关技术指标,返回DataFrame\\n\")\n\n if self.index_1to1:\n for i in self.index_1to1:\n func = getattr(talib, i)\n df[i] = func(\n np.array(df[\"close\"]),\n self.window_index\n )\n\n if self.index_3to1:\n for i in self.index_3to1:\n func = getattr(talib, i)\n df[i] = func(\n np.array(df[\"high\"]),\n np.array(df[\"low\"]),\n np.array(df[\"close\"]),\n self.window_index\n )\n\n if self.index_2to2:\n for i in self.index_2to2:\n func = getattr(talib, i)\n result_down, result_up = func(\n np.array(df[\"high\"]),\n np.array(df[\"low\"]),\n self.window_index\n )\n up = i + \"_UP\"\n down = i + \"_DOWN\"\n df[up] = result_up\n df[down] = result_down\n\n if self.index_2to1:\n for i in self.index_2to1:\n func = getattr(talib, i)\n df[i] = func(\n np.array(df[\"high\"]),\n np.array(df[\"low\"]),\n self.window_index\n )\n\n if self.index_4to1:\n for i in self.index_4to1:\n func = getattr(talib, i)\n df[i] = func(\n np.array(df[\"open\"]),\n np.array(df[\"high\"]),\n np.array(df[\"low\"]),\n np.array(df[\"close\"]),\n )\n\n return df\n\n def multi_time_frame_analysis(self, intervals: list = None, df: DataFrame = None):\n \"\"\"\"\"\"\n if not intervals:\n output(\"请输入K线合成周期\")\n return\n\n if df is None:\n df = self.orignal\n\n if df is None:\n output(\"请先加载数据\")\n return\n\n for interval in intervals:\n output(\"------------------------------------------------\")\n output(f\"合成{interval}周期K先并开始数据分析\")\n\n data = pd.DataFrame()\n data[\"open\"] = df[\"open\"].resample(interval, label='right', how=\"first\")\n data[\"high\"] = df[\"high\"].resample(interval, label='right',how=\"max\")\n data[\"low\"] = df[\"low\"].resample(interval, label='right',how=\"min\")\n data[\"close\"] = df[\"close\"].resample(interval, label='right',how=\"last\")\n data[\"volume\"] = df[\"volume\"].resample(interval, label='right',how=\"sum\")\n data = data[data.close.notnull()]\n data.reset_index(inplace=True)\n\n result = self.base_analysis(data)\n self.results[interval] = result\n\n def show_chart(self, data, boll_wide):\n \"\"\"\"\"\"\n data[\"boll_up\"] = data[\"SMA\"] + data[\"STDDEV\"] * boll_wide\n data[\"boll_down\"] = data[\"SMA\"] - data[\"STDDEV\"] * boll_wide\n\n up_signal = []\n down_signal = []\n len_data = len(data[\"close\"])\n for i in range(1, len_data):\n if data.iloc[i][\"close\"] > data.iloc[i][\"boll_up\"]and data.iloc[i - 1][\"close\"] < data.iloc[i - 1][\"boll_up\"]:\n up_signal.append(i)\n\n elif data.iloc[i][\"close\"] < data.iloc[i][\"boll_down\"] and data.iloc[i - 1][\"close\"] > data.iloc[i - 1][\"boll_down\"]:\n down_signal.append(i)\n\n plt.figure(figsize=(20, 8))\n close = data[\"close\"]\n plt.plot(range(0,len(close)),close, lw=1)\n plt.plot(range(0,len(close)),close, '^', markersize=5, color='r',\n label='UP signal', markevery=up_signal)\n plt.plot(range(0,len(close)),close, 'v', markersize=5, color='g',\n label='DOWN signal', markevery=down_signal)\n plt.plot(range(0,len(data[\"boll_up\"])),data[\"boll_up\"], lw=0.5, color=\"r\")\n plt.plot(range(0,len(data[\"boll_down\"])),data[\"boll_down\"], lw=0.5, color=\"g\")\n plt.legend()\n plt.show()\n\n def trend_analysis(self, df: DataFrame = None):\n \"\"\"\n ER位移路程比\n \"\"\"\n output(\"ER位移路程比\")\n df[\"pre_close\"] = df[\"close\"].shift(1).fillna(0)\n df[\"x\"] = abs(df[\"close\"] - df[\"close\"].shift(self.window_index).fillna(0))\n df[\"m1\"] = abs(df[\"close\"] - df[\"pre_close\"])\n\n df[\"cumsum\"] = np.cumsum(df[\"m1\"])\n df[\"pre_cumsum\"] = df[\"cumsum\"].shift(self.window_index).fillna(0)\n df[\"s\"] = df[\"cumsum\"]-df[\"pre_cumsum\"]\n\n #print(df[\"s\"].head(10))\n df[\"ER\"] = df[\"x\"]/df[\"s\"]\n #print(df[\"ER\"].tail(10))\n\n plt.figure(figsize=(20,6))\n plt.plot(range(0,len(df[\"ER\"])),df[\"ER\"])\n plt.show()\n\n statitstic_info(df[\"ER\"])\n\n def show_index(self, data, index_list: list = None):\n \"\"\"\n plot \n \"\"\"\n for i in index_list:\n value = data[i]\n plt.figure(figsize=(20,8))\n plt.plot(range(0,len(value)),value)\n plt.show()\n\ndef random_test(retrun_series):\n \"\"\"\"\"\"\n acorr_result = acorr_ljungbox(retrun_series, lags=1)\n p_value = acorr_result[1]\n if p_value < 0.05:\n output(\"随机性检验:非纯随机性\")\n else:\n output(\"随机性检验:纯随机性\")\n output(f\"白噪声检验结果:{acorr_result}\\n\")\n\n\ndef stability_test(retrun_series):\n \"\"\"\"\"\"\n statitstic = ADF(retrun_series)\n t_s = statitstic[1]\n t_c = statitstic[4][\"5%\"]\n\n if t_s > t_c:\n output(\"平稳性检验:存在单位根,时间序列不平稳\")\n else:\n output(\"平稳性检验:不存在单位根,时间序列平稳\")\n\n output(f\"ADF检验结果:{statitstic}\\n\")\n\n\ndef autocorrelation_test(retrun_series):\n \"\"\"\"\"\"\n output(\"画出自相关性图,观察自相关特性\")\n\n plot_acf(retrun_series, lags=60)\n plt.show()\n\n plot_pacf(retrun_series, lags=60).show()\n plt.show()\n\n\ndef statitstic_info(df):\n \"\"\"\"\"\"\n mean = round(df.mean(), 4)\n median = round(df.median(), 4)\n output(f\"样本平均数:{mean}, 中位数: {median}\")\n\n skew = round(df.skew(), 4)\n kurt = round(df.kurt(), 4)\n\n if skew == 0:\n skew_attribute = \"对称分布\"\n elif skew > 0:\n skew_attribute = \"分布偏左\"\n else:\n skew_attribute = \"分布偏右\"\n\n if kurt == 0:\n kurt_attribute = \"正态分布\"\n elif kurt > 0:\n kurt_attribute = \"分布陡峭\"\n else:\n kurt_attribute = \"分布平缓\"\n\n output(f\"偏度为:{skew},属于{skew_attribute};峰度为:{kurt},属于{kurt_attribute}\\n\")\n\n\ndef output(msg):\n \"\"\"\n Output message of backtesting engine.\n \"\"\"\n print(f\"{datetime.now()}\\t{msg}\")\n","sub_path":"examples/data_analysis/data_analysis.py","file_name":"data_analysis.py","file_ext":"py","file_size_in_byte":12463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"295383945","text":"from numbertheory import totientList\n\nt = totientList(10**7)\nnext(t)\nnext(t)\nm = 5\nfor n in range(2, 10**7):\n k = next(t)\n sk = str(k)\n if len(sk) < len(str(n)):\n sk += '0'\n if n/k < m and sorted(sk) == sorted(str(n)):\n m = n/k\n print(n, k, n / k)\nprint(m)\n","sub_path":"p070.py","file_name":"p070.py","file_ext":"py","file_size_in_byte":290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"527842338","text":"import os\nfrom setuptools import setup\n\n\ndef find_packages(dir_):\n packages = []\n for _dir, subdirectories, files in os.walk(os.path.join(dir_, 'pyflipdot')):\n if '__init__.py' in files:\n lib, fragment = _dir.split(os.sep, 1)\n packages.append(fragment.replace(os.sep, '.'))\n return packages\n\nsetup(\n name='pyflipdot',\n version='0.1',\n long_description=__doc__,\n packages=find_packages('lib'),\n package_dir={'': 'lib'},\n include_package_data=True,\n zip_safe=False,\n entry_points={'console_scripts': ['pyflipdot = pyflipdot.app:main']},\n install_requires=['pytz', 'pyusb', 'pyaml', 'pillow', 'numpy', 'setproctitle', 'python-dateutil', 'requests', 'PyAudio']\n)\n","sub_path":"pybrose/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":724,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"225199634","text":"import json\nimport urllib.request\n\nclass ApiManager:\n \"\"\"docstring fs ApiManager.\"\"\"\n\n def __init__(self):\n self.list_product = []\n\n def load_json(self, category):\n url = \"https://fr.openfoodfacts.org/category/\"+ category + \".json\"\n response = urllib.request.urlopen(url)\n data = json.loads(response.read())\n\n return data\n\n def get_data(self, category):\n data = self.load_json(category)\n\n for dictionary in data[\"products\"]:\n product = dictionary.get(\"product_name_fr\",\"\")\n stores = dictionary.get(\"stores\", \"\")\n description = dictionary.get(\"ingredients_text_fr\", \"\")\n score = dictionary.get(\"nutrition_grades\", \"\")\n brand = dictionary.get(\"brands\", \"\")\n link = dictionary.get(\"url\", \"\")\n\n tuple_product = (category, product, description, score, stores, brand, link)\n\n self.list_product.append(tuple_product)\n return self.list_product\n","sub_path":"myapp/models/api_manager.py","file_name":"api_manager.py","file_ext":"py","file_size_in_byte":992,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"605570770","text":"\"\"\"Test our caching zonal_stats\"\"\"\n\nimport numpy as np\nfrom geopandas import GeoSeries\nfrom shapely.geometry import Polygon\nfrom affine import Affine\nfrom pyiem.grid import zs\n\n\ndef test_gen_stats():\n \"\"\"Run a test\"\"\"\n affine = Affine(10., 0., 0., 0., -10, 100)\n grid = np.reshape(np.arange(100), (10, 10))\n sq1 = Polygon([(50, 50), (50, 60), (60, 60), (60, 50)])\n sq2 = Polygon([(60, 60), (60, 70), (70, 70), (70, 60)])\n geometries = GeoSeries([sq1, sq2])\n czs = zs.CachingZonalStats(affine)\n res = czs.gen_stats(np.flipud(grid), geometries)\n assert len(res) == 2\n assert abs(res[0] - 55.0) < 0.1\n assert abs(res[1] - 66.0) < 0.1\n\n czs = zs.CachingZonalStats(affine)\n res = czs.gen_stats(np.flipud(grid))\n assert not res\n","sub_path":"pyiem/grid/tests/test_zs.py","file_name":"test_zs.py","file_ext":"py","file_size_in_byte":764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"607244573","text":"from scipy.io import loadmat\n\ndef get_atlas_info(parcellation='Sch240'):\n if parcellation == 'Sch240':\n file = '/Users/luke/Documents/Projects/StrokeNet/Docs/Atlas/Schaefer200/240COG.mat'\n MNIcoords = loadmat(file)['COG']\n\n\n if parcellation == 'Sch240':\n file = '/Users/luke/Documents/Projects/StrokeNet/Docs/Atlas/Schaefer200/240parcellation_Yeo8Index.mat'\n networks = loadmat(file)['Yeo8Index']\n network_labels = ['Vis', 'SomMat', 'DorstAttn', 'SalVentAttn', 'Limbic', 'Control', 'Default', 'SC','Cerebellum']\n \n return MNIcoords, networks, network_labels","sub_path":"Scripts/functions/.ipynb_checkpoints/get_atlas_info-checkpoint.py","file_name":"get_atlas_info-checkpoint.py","file_ext":"py","file_size_in_byte":592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"74688037","text":"# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport paddle\nimport os\nimport paddle.nn as nn\nimport mmoe_net as net\nimport time\nimport logging\n\nfrom utils import load_yaml, get_abs_model, save_model, load_model\nfrom census_reader_dygraph import CensusDataset\nfrom paddle.io import DistributedBatchSampler, DataLoader\nimport argparse\n\nlogging.basicConfig(\n format='%(asctime)s - %(levelname)s - %(message)s', level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\n\ndef parse_args():\n parser = argparse.ArgumentParser(description='paddle-rec run')\n parser.add_argument(\"-m\", \"--config_yaml\", type=str)\n args = parser.parse_args()\n args.config_yaml = get_abs_model(args.config_yaml)\n return args\n\n\ndef create_feeds(batch, feature_size):\n input_data = paddle.to_tensor(batch[0].numpy().astype('float32').reshape(\n -1, feature_size))\n label_income = paddle.to_tensor(batch[1].numpy().astype('float32').reshape(\n -1, 1))\n label_marital = paddle.to_tensor(batch[2].numpy().astype('float32')\n .reshape(-1, 1))\n return input_data, label_income, label_marital\n\n\ndef create_loss(pred_income, pred_marital, label_income, label_marital):\n pred_income_1d = paddle.slice(pred_income, axes=[1], starts=[1], ends=[2])\n pred_marital_1d = paddle.slice(\n pred_marital, axes=[1], starts=[1], ends=[2])\n cost_income = paddle.nn.functional.log_loss(\n input=pred_income_1d, label=label_income)\n cost_marital = paddle.nn.functional.log_loss(\n input=pred_marital_1d, label=label_marital)\n\n avg_cost_income = paddle.mean(x=cost_income)\n avg_cost_marital = paddle.mean(x=cost_marital)\n\n cost = avg_cost_income + avg_cost_marital\n return cost\n\n\ndef create_model(config):\n feature_size = config.get('hyper_parameters.feature_size', None)\n expert_num = config.get('hyper_parameters.expert_num', None)\n expert_size = config.get('hyper_parameters.expert_size', None)\n tower_size = config.get('hyper_parameters.tower_size', None)\n gate_num = config.get('hyper_parameters.gate_num', None)\n\n MMoE = net.MMoELayer(feature_size, expert_num, expert_size, tower_size,\n gate_num)\n\n return MMoE\n\n\ndef create_data_loader(dataset, mode, place, config):\n batch_size = config.get('dygraph.batch_size', None)\n is_train = mode == 'train'\n batch_sampler = DistributedBatchSampler(\n dataset, batch_size=batch_size, shuffle=is_train)\n loader = DataLoader(dataset, batch_sampler=batch_sampler, places=place)\n return loader\n\n\ndef main(args):\n paddle.seed(12345)\n config = load_yaml(args.config_yaml)\n use_gpu = config.get(\"dygraph.use_gpu\", True)\n train_data_dir = config.get(\"dygraph.train_data_dir\", None)\n epochs = config.get(\"dygraph.epochs\", None)\n feature_size = config.get('hyper_parameters.feature_size', None)\n print_interval = config.get(\"dygraph.print_interval\", None)\n model_save_path = config.get(\"dygraph.model_save_path\", \"model_output\")\n\n print(\"***********************************\")\n logger.info(\n \"use_gpu: {}, train_data_dir: {}, epochs: {}, print_interval: {}, model_save_path: {}\".\n format(use_gpu, train_data_dir, epochs, print_interval,\n model_save_path))\n print(\"***********************************\")\n\n place = paddle.set_device('gpu' if use_gpu else 'cpu')\n\n mmoe_model = create_model(config)\n model_init_path = config.get(\"dygraph.model_init_path\", None)\n if model_init_path is not None:\n load_model(model_init_path, mmoe_model)\n\n # to do : add optimizer function\n optimizer = paddle.optimizer.Adam(parameters=mmoe_model.parameters())\n\n # to do init model\n file_list = [\n os.path.join(train_data_dir, x) for x in os.listdir(train_data_dir)\n ]\n print(\"read data\")\n dataset = CensusDataset(file_list)\n train_dataloader = create_data_loader(\n dataset, mode='test', place=place, config=config)\n\n last_epoch_id = config.get(\"last_epoch\", -1)\n\n for epoch_id in range(last_epoch_id + 1, epochs):\n # set train mode\n mmoe_model.train()\n auc_metric_marital = paddle.metric.Auc(\"ROC\")\n auc_metric_income = paddle.metric.Auc(\"ROC\")\n epoch_begin = time.time()\n interval_begin = time.time()\n train_reader_cost = 0.0\n train_run_cost = 0.0\n total_samples = 0\n reader_start = time.time()\n\n for batch_id, batch in enumerate(train_dataloader()):\n train_reader_cost += time.time() - reader_start\n optimizer.clear_grad()\n train_start = time.time()\n batch_size = len(batch[0])\n\n input_data, label_income, label_marital = create_feeds(\n batch, feature_size)\n\n pred_income, pred_marital = mmoe_model(input_data)\n loss = create_loss(pred_income, pred_marital, label_income,\n label_marital)\n\n loss.backward()\n optimizer.step()\n train_run_cost += time.time() - train_start\n total_samples += batch_size\n # for auc\n auc_metric_income.update(\n preds=pred_income.numpy(), labels=label_income.numpy())\n auc_metric_marital.update(\n preds=pred_marital.numpy(), labels=label_marital.numpy())\n\n if batch_id % print_interval == 1:\n logger.info(\n \"epoch: {}, batch_id: {}, auc_income: {:.6f}, auc_marital: {:.6f}, avg_reader_cost: {:.5f} sec, avg_batch_cost: {:.5f} sec, avg_samples: {:.5f}, ips: {:.5f} images/sec\".\n format(epoch_id, batch_id,\n auc_metric_income.accumulate(),\n auc_metric_marital.accumulate(), train_reader_cost /\n print_interval, (train_reader_cost + train_run_cost\n ) / print_interval, total_samples /\n print_interval, total_samples / (train_reader_cost +\n train_run_cost)))\n train_reader_cost = 0.0\n train_run_cost = 0.0\n total_samples = 0\n reader_start = time.time()\n\n logger.info(\n \"epoch: {} done, auc_income: {:.6f}, auc_marital: {:.6f}, : epoch time{:.2f} s\".\n format(epoch_id,\n auc_metric_income.accumulate(),\n auc_metric_marital.accumulate(), time.time() - epoch_begin))\n\n save_model(\n mmoe_model, optimizer, model_save_path, epoch_id, prefix='rec')\n\n\nif __name__ == '__main__':\n args = parse_args()\n main(args)\n","sub_path":"models/multitask/mmoe/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":7280,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"52940330","text":"import requests\r\nfrom bs4 import BeautifulSoup\r\n\r\nheaders = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36\"}\r\nlink = \"https://beijing.anjuke.com/sale/p\"\r\n\r\nfor i in range(1,11):\r\n r = requests.get(link + str(i),headers=headers)\r\n\r\n soup = BeautifulSoup(r.text, 'lxml')\r\n house_list = soup.find_all(\"li\", class_='list-item')\r\n\r\n for house in house_list:\r\n name = house.find('div', class_='house-title').a.text.strip()\r\n price = house.find('span', class_='price-det').text.strip()\r\n price_area = house.find('span', class_='unit-price').text.strip()\r\n\r\n selector = house.find(\"div\",class_='details-item')\r\n no_room = selector.span.text\r\n area = selector.contents[3].text\r\n floor = selector.contents[5].text\r\n year = selector.contents[7].text\r\n\r\n broker = house.find(\"span\", class_=\"brokername\").text\r\n broker = broker[1:]\r\n address = house.find(\"span\", class_='comm-address')[\"title\"]\r\n tag_list = house.find_all('span', class_='item-tags tag-others')\r\n tags = [i.text for i in tag_list]\r\n print(name, price, price_area, no_room, area, floor, year, broker, address, tags)\r\n\r\n filename = \"AnjukeBJ\" + \".txt\"\r\n \r\n with open(filename, 'a+', encoding='utf-8') as f:\r\n f.write(\"%s \\t %s \\t %s \\t %s \\t %s \\t %s \\t %s \\t %s \\t %s \\t %s \\t \\r\\n\"\r\n % (name, price, price_area, no_room, area, floor, year, broker, address, tags))\r\n","sub_path":"Anjuke/AnjukeBJ.py","file_name":"AnjukeBJ.py","file_ext":"py","file_size_in_byte":1560,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"225352915","text":"#Enumerating k-mers Lexicogrphically\n#Calvin D. Cox\n#12/10/2017\n\n#Given: A collection of at most 10 symbols defining an ordered alphabet, and a positive integer nn (n≤10n≤10).\n#Return: All strings of length nn that can be formed from the alphabet, ordered lexicographically (use the standard order of symbols in the English alphabet).\n\nfrom itertools import product as prod\nimport re\n\n#Loading in the Sequence Files\nwith open('exampletext.txt', 'r') as et:\n data_set = et.readline()\n\n#Create chart with all possible permutations \ni = 1\nini_mirror = []\nwhile i <= int(data_set):\n ini_mirror.append(int(i))\n ini_mirror.append(-int(i))\n i += 1\n \n\n#Store permutation in an acceptable format to be sorted\nperm_store = []\nnum_perm = prod(ini_mirror, repeat=int(data_set))\nfor i in list(num_perm):\n perm_store.append(re.sub(\"[^-\\d+$]\", \"\", str(i)))\n \nwith open('resulttext.txt', 'w') as rt:\n for i in perm_store:\n print(i)\n rt.write(i + '\\n')\n \n","sub_path":"Bioinformatics Stronk/Enumerating_Oriented_Gene_Orderings.py","file_name":"Enumerating_Oriented_Gene_Orderings.py","file_ext":"py","file_size_in_byte":981,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"219087545","text":"import sys\nfrom PyQt5.QtWidgets import (QWidget,QTabWidget,QSizePolicy,QListWidget,QVBoxLayout,\n QHBoxLayout,QPushButton,QListWidgetItem)\nfrom PyQt5.QtCore import Qt,QSize\n\nfrom PyQt5.QtWidgets import QLabel\n\n'''\nclass MarkList(QListWidget):\n def __init__(self):\n super().__init__()\n'''\n\nclass MarkItem(QListWidgetItem):\n def __init__(self, list, name, pose):\n print(\"[MarkItem {0} - {1}] >>>\".format(sys._getframe().f_code.co_name, sys._getframe().f_lineno))\n super().__init__()\n\n widget = QWidget()\n\n layout = QHBoxLayout()\n\n self.label1 = QLabel(name)\n self.label2 = QLabel(pose)\n\n layout.addWidget(self.label1)\n layout.addWidget(self.label2)\n\n widget.setLayout(layout)\n\n self.setSizeHint(QSize(10,30))\n\n list.addItem(self)\n list.setItemWidget(self, widget)\n\n def getMarkName(self):\n return self.label1.text()\n\n def getMarkValue(self):\n return self.label2.text()\n\nclass TabWidget(QTabWidget):\n def __init__(self):\n super().__init__()\n\n self.markListTab = QWidget()\n self.pathTab = QWidget()\n\n self.addTab(self.markListTab, u\"标签列表\")\n self.addTab(self.pathTab, u\"轨迹列表\")\n\n # 设置水平宽度固定,垂直方向随窗口伸展\n self.setSizePolicy(QSizePolicy.Fixed, QSizePolicy.Preferred)\n\n self.markListTabUi()\n self.pathTabUi()\n\n def registerCB(self, callback):\n if callback is not None:\n self.cb = callback\n\n def markListTabUi(self):\n layout = QVBoxLayout()\n\n self.markList = QListWidget()\n\n btn_remove = QPushButton('&Remove')\n btn_remove.clicked.connect(self.removeMark)\n\n btn_save = QPushButton('&Save')\n btn_save.clicked.connect(self.saveMark)\n\n layout.addWidget(self.markList)\n layout.addWidget(btn_remove, 0, Qt.AlignRight | Qt.AlignBottom)\n layout.addWidget(btn_save, 0, Qt.AlignLeft | Qt.AlignBottom)\n\n self.markListTab.setLayout(layout)\n\n #设置标签头\n MarkItem(self.markList, '标签名', '坐标值')\n\n #MarkItem(self.markList, '厨房', '11.0 x 22.0')\n\n def pathTabUi(self):\n pass\n\n # 保存标签点到文件中\n def saveMark(self):\n fp = open('labels.yaml', 'w')\n\n for mark_id in range(1,self.markList.count()):\n #print('888> ',self.markList.item(mark_id).getMarkName(), self.markList.item(mark_id).getMarkValue())\n line = '{0}: {1}'.format(self.markList.item(mark_id).getMarkName(), self.markList.item(mark_id).getMarkValue())\n fp.writelines(line + '\\n')\n\n fp.close()\n\n def removeMark(self):\n # 判断选中的哪一个标签\n mark = self.markList.currentItem()\n if mark:\n print('*******> markName:', mark, mark.getMarkName(), self.markList.row(mark))\n else:\n print('Get current Item failed')\n return\n\n if self.cb is not None:\n self.cb(mark.getMarkName())\n\n # pixmap中删除标签点后,还需删除标签列表中的对应信息\n #self.markList.removeItemWidget(mark) # 方法一\n self.markList.takeItem(self.markList.row(mark)) # 方法二\n\n def addMark(self, markName, markPose):\n if markName is not None and markPose is not None:\n MarkItem(self.markList, markName, markPose)\n","sub_path":"ROSClientGui/ui/TabWidget.py","file_name":"TabWidget.py","file_ext":"py","file_size_in_byte":3456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"600584447","text":"from flask import Flask,request,jsonify\r\nimport cv2\r\nimport os\r\nimport json\r\n\r\nfrom .Translator.translate import Translator\r\nfrom .Sudoku_Solver.Grid_Detection.Optimised import SudokuSolver\r\nfrom .Math_Equation_Solver.math_equation_solver import MathEquationSolver\r\nfrom .Barcode_Product.barcode_to_product_details import BarcodeToProductDetails\r\n\r\napp=Flask(__name__)\r\n\r\n@app.route('/Sudoku',methods=['POST','GET'])\r\ndef Sudoku():\r\n\r\n upload=request.files['picture']\r\n\r\n if(upload.filename != ''):\r\n upload.save(upload.filename)\r\n\r\n img=cv2.imread(upload.filename)\r\n os.remove(upload.filename)\r\n\r\n solver=SudokuSolver(img)\r\n solvedSudoku=solver.solveSudoku()\r\n\r\n print(solvedSudoku)\r\n d={'data':solvedSudoku}\r\n\r\n return jsonify(d)\r\n\r\n@app.route('/MathEquation',methods=['POST','GET'])\r\ndef MathEquation():\r\n\r\n upload=request.files['picture']\r\n\r\n if(upload.filename != ''):\r\n upload.save(upload.filename)\r\n\r\n img=cv2.imread(upload.filename)\r\n os.remove(upload.filename)\r\n\r\n solver=MathEquationSolver()\r\n equation=solver.solveEquation(img)\r\n\r\n print(equation)\r\n d={'data':equation}\r\n\r\n return jsonify(d)\r\n\r\n@app.route('/Barcode',methods=['POST','GET'])\r\ndef Barcode():\r\n\r\n upload=request.files['picture']\r\n\r\n if(upload.filename != ''):\r\n upload.save(upload.filename)\r\n\r\n img=cv2.imread(upload.filename)\r\n os.remove(upload.filename)\r\n\r\n solver=BarcodeToProductDetails(img)\r\n details=solver.getProductInformation()\r\n\r\n print(details)\r\n d={'data':details}\r\n\r\n return jsonify(d)\r\n\r\n@app.route('/Translate',methods=['POST'])\r\ndef Translate():\r\n # takes body as simple string\r\n a=request.get_data(as_text=True)\r\n\r\n # deserializing string to json\r\n js=json.loads(a)\r\n\r\n trans=Translator()\r\n data=trans.translate(str(js['from']),str(js['to']),str(js['text']))\r\n d={'data':data}\r\n\r\n return jsonify(d)\r\n\r\n\r\nif __name__==\"main\":\r\n app.run(debug=True)\r\n\r\n# app.run(port=8000)","sub_path":"Flask-Server/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2075,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"420588374","text":"from greyatomlib.python_getting_started.q01_read_data.build import read_data\ndata1 = read_data()\n\n# solution\n\n#lst = []\ndef deliveries_count(data):\n lst = data['innings'][0]['1st innings']['deliveries']\n #print balls[0]['batsman']\n count = 0\n for d in lst:\n balls = d.values()\n if balls[0]['batsman'] == 'RT Ponting':\n count+=1\n return count\n\ndeliveries_count(data1)\n","sub_path":"q04_count/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"586897886","text":"from flask import Flask, request, redirect, url_for, render_template\nfrom werkzeug.utils import secure_filename\nimport pandas as pd\nimport numpy as np\nimport os\nimport pickle\nimport math\nimport matplotlib as mpl\nmpl.use('TkAgg')\nfrom wordclouds import wordclouds\nfrom KeyWords import Rake\nfrom collections import Counter\nfrom TFIDF import digram_model\nfrom CaseTrack import case_track\n\n\nAPP_ROOT = os.path.dirname(os.path.abspath(__file__))\nUPLOAD_FOLDER = os.path.join(APP_ROOT, './static')\nALLOWED_EXTENSIONS = set(['txt', 'pdf', 'png', 'jpg', 'jpeg', 'gif','csv'])\n\napp = Flask(__name__)\napp.config['DEBUG'] = True\napp.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER\napp.config['SEND_FILE_MAX_AGE_DEFAULT'] = 0\n\ndef allowed_file(filename):\n return '.' in filename and \\\n filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS\n\n@app.route('/', methods=['GET', 'POST'])\ndef cover():\n if request.method == 'POST':\n if request.form['name'] == 'Medallia':\n return redirect(url_for('Medallia'))\n elif request.form['name'] == 'Salesforce':\n return redirect(url_for('Salesforce'))\n else:\n pass\n return render_template('index.html')\n\n@app.route('/Medallia',methods=['GET', 'POST'])\ndef Medallia():\n # upload file\n if request.method == 'POST':\n # check if the post request has the file part\n if 'file' not in request.files:\n return redirect(url_for('cover'))\n file = request.files['file']\n # if user does not select file, browser also\n # submit a empty part without filename\n if file.filename == '':\n return redirect(url_for('cover'))\n if file and allowed_file(file.filename):\n file.filename = \"Medallia.csv\"\n filename = secure_filename(file.filename)\n file.save(os.path.join(app.config['UPLOAD_FOLDER'], filename))\n return redirect(url_for('select'))\n return render_template(\"Medallia.html\")\n@app.route('/Medallia/select',methods=['GET','POST'])\ndef select():\n SITE_ROOT = os.path.realpath(os.path.dirname(__file__))\n url_ = os.path.join(SITE_ROOT, 'static', 'Medallia.csv')\n if os.path.isfile(url_):\n dat2 = pd.read_csv(url_, sep=\",\", encoding='latin1')\n dat2['Responsedate'] = pd.to_datetime(dat2['Responsedate'], infer_datetime_format=True)\n dat2['year'] = dat2['Responsedate'].dt.year\n dat2['month'] = dat2['Responsedate'].dt.month\n year = sorted(list(dat2['year'].unique()))\n month = sorted(list(dat2['month'].unique()))\n # return \"YES\"\n if request.method == 'POST':\n # return \"YES\"\n selected_year = request.form.getlist('Year')\n selected_month = request.form.getlist('Month')\n dat1 = dat2[dat2['year'].isin([int(i) for i in selected_year if i != ''])]\n dat1 = dat1[dat1['month'].isin([int(i) for i in selected_month if i != ''])]\n\n # Likelihood to Recommend //bar chart 1\n scores = list(set(dat1['Likelihood to Recommend'].dropna()))\n Score_dic = {}\n for i in range(len(scores)):\n Score_dic[int(scores[i])] = np.sum(dat1['Likelihood to Recommend'] == scores[i])\n Score_dic['NaN'] = np.sum(dat1['Likelihood to Recommend'] != dat1['Likelihood to Recommend'])\n labelsLR = list(Score_dic.keys())\n valuesLR = list(Score_dic.values())\n\n # Visiting Purposes //pie chart1\n reasons = list(set(dat1['Reason for Visit'].dropna()))\n reasons_dic = {}\n for i in range(len(reasons)):\n reasons_dic[reasons[i]] = round(np.sum(dat1['Reason for Visit'] == reasons[i]) / len(dat1),2)\n reasons_dic['NaN'] = round(np.sum(dat1['Reason for Visit'] != dat1['Reason for Visit']) / len(dat1),2)\n labelsP = list(reasons_dic.keys())\n valuesP = list(reasons_dic.values())\n colorsP = [\"#F7464A\", \"#46BFBD\", \"#FDB45C\", \"#FEDCBA\", \"#ABCDEF\", \"#DDDDDD\", \"#ABCABC\"]\n\n # Purpose Completion //bar chart 2\n complete = list(set(dat1['Complete Purpose'].dropna()))\n complete_dic = {}\n for i in range(len(complete)):\n complete_dic[complete[i]] = np.sum(dat1['Complete Purpose'] == complete[i])\n complete_dic['NaN'] = np.sum(dat1['Complete Purpose'] != dat1['Complete Purpose'])\n labelsCP = list(complete_dic.keys())\n valuesCP = list(complete_dic.values())\n return render_template('Medallia.html', set=zip(valuesP, labelsP, colorsP),\n valuesLR=valuesLR, labelsLR=labelsLR,\n valuesCP=valuesCP, labelsCP=labelsCP,\n year=year, month=month)\n\n return render_template(\"Medallia.html\",year = year, month = month)\n\n\n@app.route('/Medallia/wordcloud', methods=['GET','POST'])\ndef mdwordcloud():\n SITE_ROOT = os.path.realpath(os.path.dirname(__file__))\n url_ = os.path.join(SITE_ROOT, 'static', 'Medallia.csv')\n if os.path.isfile(url_):\n wordcloud = pd.read_csv(url_, sep=\",\", encoding='latin1')\n wordcloud['Responsedate'] = pd.to_datetime(wordcloud['Responsedate'], infer_datetime_format=True)\n wordcloud['year'] = wordcloud['Responsedate'].dt.year\n wordcloud['month'] = wordcloud['Responsedate'].dt.month\n year = sorted(list(wordcloud['year'].unique()))\n month = sorted(list(wordcloud['month'].unique()))\n score = sorted([int(i) for i in list(wordcloud['Likelihood to Recommend'].unique()) if not math.isnan(i)])\n if request.method == 'POST':\n selected_year = request.form.getlist('Year')\n selected_month = request.form.getlist('Month')\n selected_score = request.form.getlist('Score')\n wordcloud = wordcloud[wordcloud['year'].isin([int(i) for i in selected_year if i != ''])]\n wordcloud = wordcloud[wordcloud['month'].isin([int(i) for i in selected_month if i != ''])]\n wordcloud = wordcloud[wordcloud['Likelihood to Recommend'].isin([int(i) for i in selected_score if i != ''])]\n wordclouds(wordcloud, SITE_ROOT)\n return render_template('Wordcloud.html', year=year, month=month, score=score)\n\n\n@app.route('/Salesforce',methods=['GET', 'POST'])\ndef Salesforce():\n # upload file\n if request.method == 'POST':\n # check if the post request has the file part\n if 'file' not in request.files:\n return redirect(url_for('cover'))\n file = request.files['file']\n # if user does not select file, browser also\n # submit a empty part without filename\n if file.filename == '':\n return redirect(url_for('cover'))\n if file and allowed_file(file.filename):\n file.filename = \"Salesforce.csv\"\n filename = secure_filename(file.filename)\n file.save(os.path.join(app.config['UPLOAD_FOLDER'], filename))\n SITE_ROOT = os.path.realpath(os.path.dirname(__file__))\n url_ = os.path.join(SITE_ROOT, 'static', 'Salesforce.csv')\n return redirect(url_for('sf_select'))\n return render_template('Salesforce.html')\n\n@app.route('/Salesforce/select',methods=['GET','POST'])\ndef sf_select():\n SITE_ROOT = os.path.realpath(os.path.dirname(__file__))\n url_ = os.path.join(SITE_ROOT, 'static', 'Salesforce.csv')\n if os.path.isfile(url_):\n Salesforce = pd.read_csv(url_, sep=\",\", encoding='latin1')\n Salesforce = Salesforce.dropna(subset=['Case Record Type'])\n Salesforce['Date/Time Opened'] = pd.to_datetime(Salesforce['Date/Time Opened'], infer_datetime_format=True)\n Salesforce['year'] = Salesforce['Date/Time Opened'].dt.year\n Salesforce['month'] = Salesforce['Date/Time Opened'].dt.month\n year = sorted(list(Salesforce['year'].unique()))\n month = sorted(list(Salesforce['month'].unique()))\n Category = list(Salesforce['Case Record Type'].unique())\n\n if request.method == 'POST':\n selected_year = request.form.getlist('Year')\n selected_month = request.form.getlist('Month')\n selected_category = request.form.getlist('Category')\n\n Salesforce = Salesforce[Salesforce['year'].isin([int(i) for i in selected_year if i != ''])]\n Salesforce = Salesforce[Salesforce['month'].isin([int(i) for i in selected_month if i != ''])]\n Salesforce = Salesforce[Salesforce['Case Record Type'].isin(selected_category)]\n\n #bar chart\n labels = Salesforce[\"Case Record Type\"].unique()\n dic1 = {}\n for category in labels:\n value = len(Salesforce[Salesforce[\"Case Record Type\"] == category])\n dic1[category] = value\n values = dic1.values()\n labels = dic1.keys()\n\n #sentiment pie chart\n Salesforce['combine'] = Salesforce['Description'].fillna('') + ' ' + Salesforce['Case Comments'].fillna('')\n tmodel = digram_model()\n transformed = tmodel.transform(Salesforce['combine'])\n cmodel = pickle.load(open(os.path.join(SITE_ROOT,'static','MultinomialNB.pkl'), 'rb'))\n predict = cmodel.predict(transformed)\n sentiments, counts = np.unique(predict, return_counts=True)\n dic = {1:'Positive',0:'Neutral',-1:'Negative'}\n sentiment_labels = []\n for i in list(sentiments):\n sentiment_labels.append(dic[i])\n #sentiment_labels = sentiment_labels#list(sentiments)\n sentiment_counts = list(counts)\n colorsP = [\"#F7464A\", \"#46BFBD\", \"#FDB45C\"]\n\n #case track charts\n #Salesforce['result'] = predict\n trace = case_track(Salesforce)\n ctvalues = trace.values()\n ctlabels = trace.keys()\n ctcolors = [\"#F7464A\", \"#FFB6C1\", \"#46BFBD\", \"#FDB45C\", \"#FEDCBA\", \"#ABCDEF\", \"#DDDDDD\", \"#AFEEEE\", \"#CD6889\", \"#BEE554\", \"#EEA9B8\"]\n set2=zip(ctvalues, ctlabels, ctcolors)\n return render_template('Salesforce.html', values=values, labels=labels, year=year, month=month,\n category=Category, set=zip(sentiment_counts, sentiment_labels, colorsP),\n set2=set2, ctvalues=ctvalues, ctlabels=ctlabels)\n\n return render_template('Salesforce.html',year = year, month = month, category = Category)\n\n\n@app.route('/Salesforce/keyword', methods=['GET', 'POST'])\ndef sfkeyword():\n SITE_ROOT = os.path.realpath(os.path.dirname(__file__))\n url_ = os.path.join(SITE_ROOT, 'static', 'Salesforce.csv')\n if os.path.isfile(url_):\n Salesforce = pd.read_csv(url_, sep=\",\", encoding='latin1')\n Salesforce = Salesforce.dropna(subset=['Case Record Type'])\n Salesforce['Description'] = Salesforce['Description'].apply(lambda x: x if isinstance(x, str) else '')\n Salesforce['Case Comments'] = Salesforce['Case Comments'].apply(lambda x: x if isinstance(x, str) else '')\n Salesforce['Combine'] = Salesforce.apply(lambda x: x['Case Comments'] if x['Case Comments'] != '' else (x['Description'] if x['Description'] != '' else ''), axis=1)\n Salesforce['Date/Time Opened'] = pd.to_datetime(Salesforce['Date/Time Opened'], infer_datetime_format=True)\n Salesforce['year'] = Salesforce['Date/Time Opened'].dt.year\n Salesforce['month'] = Salesforce['Date/Time Opened'].dt.month\n year = sorted(list(Salesforce['year'].unique()))\n month = sorted(list(Salesforce['month'].unique()))\n category = list(Salesforce['Case Record Type'].unique())\n keywords=[]\n\n if request.method == 'POST':\n # return \"YES\"\n selected_year = request.form.getlist('Year')\n selected_month = request.form.getlist('Month')\n selected_category = request.form.getlist('Category')\n\n Salesforce = Salesforce[Salesforce['year'].isin([int(i) for i in selected_year if i != ''])]\n Salesforce = Salesforce[Salesforce['month'].isin([int(i) for i in selected_month if i != ''])]\n Salesforce = Salesforce[Salesforce['Case Record Type'].isin(selected_category)]\n Stoplist_url = os.path.join(SITE_ROOT, 'static', 'SmartStoplist.txt')\n rakes = Rake(Salesforce['Combine'], Stoplist_url, 3, 5)\n phrase_s = rakes.execute()\n keywords = Counter(phrase_s).most_common()[:30]\n return render_template('Keywords.html', year=year, month=month, category=category, keywords=keywords)\n\n return render_template('Keywords.html', year=year, month=month, category=category, keywords=keywords)\n\n\n# If 'Address has been used', do following in terminal\n# ps aux | grep python\n# sudo kill -9 PID\n\nif __name__ == '__main__':\n app.debug = True\n app.run()\n","sub_path":"Zoetis_app/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":12518,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"554321318","text":"# encoding:UTF-8\nimport re\nimport urllib.request\nfrom urllib import request,parse\nfrom bs4 import BeautifulSoup\nimport os\nimport ssl\nfrom selenium import webdriver\n\nssl._create_default_https_context = ssl._create_unverified_context\n\n\nurl = 'https://www.817cf.com/xiaoshuo/39095.html'\nheader = {'User-Agent':'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:23.0) Gecko/20100101 Firefox/23.0'}\nres = request.Request(url = url, headers = header)\nrep = request.urlopen(res)\nimre = rep.read().decode('utf-8')\n# print(imre)\n# print(res)\n# print(rep)\n\nsoup = BeautifulSoup(imre,'lxml')\n# print(soup.p.string) #获取p标签中的文字\n# print(soup.find_all('a')\n\nget_text_title = []\nget_xstext = []\n\nfor text in soup.find_all('a'):\n # print(text.get_text())\n get_text_title.append(text.get_text())\n\nxstitle = get_text_title[-3]\n\nfor xstext in soup.find_all('p'):\n # print(xstext.get_text())\n get_xstext.append(xstext.get_text())\n\n# print(get_xstext)\n# print(text_list[1])\n#test\n# os.mkdir('xiaoshuo_text')\n\ntxt_name = xstitle + '.txt'\n\n\n\npage_text = open(txt_name,'w',encoding='utf-8')\nfor webtext in get_xstext[4:]: #取消前几行无关信息\n page_text.write(str(webtext))\n page_text.write('\\n')\npage_text.close()\n","sub_path":"sometests/pagetest.py","file_name":"pagetest.py","file_ext":"py","file_size_in_byte":1221,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"44542457","text":"#### import the simple module from the paraview\nfrom paraview.simple import *\n#### disable automatic camera reset on 'Show'\nparaview.simple._DisableFirstRenderCameraReset()\n\nlist_nu = [1, 1.5, 2]\nlist_lc = [0.01, 0.1, 0.2]\n\nni = len(list_nu)\nnj = len(list_lc)\n\n# create a new 'Legacy VTK Reader'\nfor i in range(ni):\n for j in range(nj):\n nu = list_nu[i]\n lc = list_lc[j]\n file_name = '/Users/eroubin/git/test.python/clients/matern2D_lc{lc}_nu{nu}.vtk'.format(nu=nu, lc=lc)\n rf_vtk = LegacyVTKReader(FileNames=[file_name])\n # get active view\n render = GetActiveViewOrCreate('RenderView')\n render.OrientationAxesVisibility = 0\n # show rf in view\n rf_vtkDisplay = Show(rf_vtk, render)\n Hide(rf_vtk, render)\n\n # create a new 'Clip'\n clip11 = Clip(Input=rf_vtk)\n clip11.ClipType = 'Scalar'\n clip11.Value = 0.5\n # create a new 'Transform'\n tran11 = Transform(Input=clip11)\n tran11.Transform = 'Transform'\n tran11.Transform.Translate = [i*1.1, j*1.1, 0.0]\n\n clip12 = Clip(Input=rf_vtk)\n clip12.ClipType = 'Scalar'\n clip12.InsideOut = 1\n clip12.Value = -0.5\n tran12 = Transform(Input=clip12)\n tran12.Transform = 'Transform'\n tran12.Transform.Translate = [i*1.1, j*1.1, 0.0]\n # show dat in view\n tran11Display = Show(tran11, render)\n tran12Display = Show(tran12, render)\n # turn off scalar coloring\n # ColorBy(tran11Display, None)\n\n # save screenshot\n xc = 0.5*ni + 0.05*(ni-1)\n yc = 0.5*nj + 0.05*(nj-1)\n zc = 0.0\n magnification = 5\n height = 1.0*ni+0.1*(ni+1)\n width = 1.0*nj+0.1*(nj+1)\n\n ratio = height / width\n height_p = 1155 * magnification\n width_p = int(height_p * 1.0 / ratio / magnification)\n\n render.ViewSize = [width_p , height_p]\n render.InteractionMode = '2D'\n render.CameraPosition = [xc, yc, 10000.0 + zc]\n render.CameraFocalPoint = [xc, yc, zc]\n render.CameraParallelScale = (height / 2.0)\n Render()\n\nSaveScreenshot('/Users/eroubin/git/test.python/clients/coucou.png', magnification=2, quality=100, view=render)\n","sub_path":"clients/2DImagesMultiple.py","file_name":"2DImagesMultiple.py","file_ext":"py","file_size_in_byte":2246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"576071531","text":"import socket\n\n\n# define constants\nbufferSize = 1024\n\n\ndef client(n_channels, ch_base_port, cl_base_port, n_pkts):\n ch_addr = [None]*n_channels\n cl_addr = [None]*n_channels\n cl_skt = [None]*n_channels\n for i in range(n_channels):\n # create a list of pairs of channel address and port\n ch_addr[i] = (\"127.0.0.1\", ch_base_port+i)\n \n # create and bind client sockets\n cl_addr[i] = (\"127.0.0.1\", cl_base_port+i)\n cl_skt[i] = socket.socket(family=socket.AF_INET, type=socket.SOCK_DGRAM)\n cl_skt[i].bind(cl_addr[i])\n print(\"CL{0:d}: The client is up and listening.\".format(i))\n \n # send a request message to the server through channels\n rqst = \"Request\"\n for i in range(n_channels):\n cl_skt[i].sendto(rqst.encode(), ch_addr[i])\n print(\"CL{0:d}: Sent {1} to {2}\".format(i, rqst, ch_addr[i]))\n\n sn = -1\n while True:\n for i in range(n_channels):\n rx_data, rx_addr = cl_skt[i].recvfrom(bufferSize)\n rx_msg = rx_data.decode(\"utf-8\")\n print(\"CL{0:d}: Received {1} from {2}\".format(i, rx_msg, rx_addr))\n\n # if there is at least one \"Data\" from the channles, SN is updated\n if \"Data\" in rx_msg:\n # update the data sequence number\n tmp = int((rx_msg.split(' ')[1]).split('=')[1]) # 'tmp' is an integer\n if tmp > sn:\n sn = tmp\n\n ack_msg = \"ACK: SN=\" + str(sn)\n ack_data = ack_msg.encode()\n for i in range(n_channels):\n cl_skt[i].sendto(ack_data, ch_addr[i])\n print(\"CL{0:d}: Sent {1} to {2}\".format(i, ack_msg, ch_addr[i]))\n\n if sn == n_pkts-1:\n return 0 # to end the function\n\n\nif __name__ == '__main__':\n import argparse\n parser = argparse.ArgumentParser()\n parser.add_argument(\n \"-C\",\n \"--n_channels\",\n help=\"number of channels; default is 2\",\n default=2,\n type=int)\n parser.add_argument(\n \"--ch_base_port\",\n help=\"base port number for channels; default is 31100\",\n default=31100,\n type=int)\n parser.add_argument(\n \"--cl_base_port\",\n help=\"base port number for client; default is 31200\",\n default=31200,\n type=int)\n parser.add_argument(\n \"-P\",\n \"--n_pkts\",\n help=\"number of packets to receive from the server; default is 10\",\n default=10,\n type=int)\n args = parser.parse_args()\n n_channels = args.n_channels\n ch_base_port = args.ch_base_port\n cl_base_port = args.cl_base_port\n n_pkts = args.n_pkts\n\n\n # run the client\n client(n_channels=n_channels,\n ch_base_port=ch_base_port,\n cl_base_port=cl_base_port,\n n_pkts=n_pkts)\n","sub_path":"mptcp/components/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":2803,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"323180270","text":"import datetime\n\nfrom bs4 import BeautifulSoup\n\nfrom extraction.value_extractor import extract_money, extract_percent, parse_german_float\n\n\ndef translate_row(row_name):\n values = {\n 'Umsatz': 'revenue',\n 'Umsatz je Aktie': 'revenuePerStock',\n 'Cashflow je Aktie': 'cashFlowPerStock',\n 'Eigenkapitalrendite': 'returnOnEquity',\n 'Umsatzrendite': 'returnOnRevenue',\n 'Gesamtrendite': 'totalReturn',\n 'Return on Investment': 'returnOnInvest',\n 'Anlageintensität': 'shareOfLongTermAssets',\n 'Arbeitsintensität': 'shareOfCurrentAssets',\n 'Eigenkapitalquote': 'shareOfEquity',\n 'Verschuldungsgrad': 'debtToEquityRatio',\n 'Liquidität 1. Grades': 'firstGradeLiquidity',\n 'Liquidität 2. Grades': 'secondGradeLiquidity',\n 'Liquidität 3. Grades': 'thirdGradeLiquidity',\n 'Deckungsgrad 1': 'firstCoverageRatio',\n 'Deckungsgrad 2': 'secondCoverageRatio',\n 'Deckungsgrad 3': 'thirdCoverageRatio',\n 'Fremdkapitalquote': 'gearingRatio',\n 'Dividende': 'dividend',\n 'Dividende je Aktie': 'dividendPerShare',\n 'Ergebnis je Aktie verwässert': 'earningsPerShare',\n 'Operatives Ergebnis': 'operatingProfit',\n 'Jahresüberschuss': 'earnings',\n 'Forschungs- und Entwicklungskosten': 'researchAndDevelopment',\n 'Ausstehende Aktien in Mio. (verwässert)': 'amountOfSharesOutstanding',\n 'Cash flow': 'cashFlow',\n 'Cashflow aus der Investitionstätigkeit': 'cashFlowFromInvestments',\n 'Cashflow aus der Finanzierungstätigkeit': 'cashFlowFromFinance',\n 'Veränderung der Finanzmittel': 'cashReserveDelta',\n 'Finanzmittel am Ende der Periode': 'cashReserveEndOfPeriod',\n 'Buchwert je Aktie': 'bookValuePerShare',\n 'Anzahl der Mitarbeiter': 'amountOfEmployees',\n 'Personalkosten': 'laborCost',\n 'Umsatz pro Mitarbeiter': 'revenuePerEmployee'\n }\n if row_name in values:\n return values[row_name]\n else:\n return None\n\n\ndef extract_years_header(table, title):\n years = []\n for header_column in table.thead.tr(\"th\"):\n if header_column.contents[0].string != title:\n years.append(int(header_column.contents[0].string))\n return years\n\n\ndef has_header(table, title):\n if len(table.tr(\"th\")) == 0:\n return False\n else:\n table_title = table.tr(\"th\")[0].contents[0].string\n return table_title == title\n\n\nclass Extractor:\n\n def __init__(self, file_content, reference):\n\n self.soup = BeautifulSoup(file_content, 'html.parser')\n content_section = self.soup.find(id='vwd_content')\n self.content_section = content_section\n self.reference = reference\n\n @staticmethod\n def has_isin_content(span):\n return 'ISIN' in span.string\n\n def stock_info(self):\n return {'head': self.head(), }\n\n def head(self):\n\n stock_head_data = {}\n\n price_info_section = self.content_section.find_all(\"div\", {\"class\": \"einzelkurs_header\"}, limit=2, recursive=False)\n stock_head_data['title'] = price_info_section[0].h1.string\n price_text = price_info_section[1].find(title=\"aktueller Wert\").string.replace(u'\\xa0', ' ')\n\n stock_head_data['current_price'] = extract_money(price_text)\n isin_and_wkn = price_info_section[1](string=self.has_isin_content)[0].string\n stock_head_data['isin'] = isin_and_wkn.split('|')[0].replace('ISIN', '').strip()\n stock_head_data['wkn'] = isin_and_wkn.split('|')[1].replace('WKN', '').strip()\n stock_head_data['timestamp'] = datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n\n stock_head_data['url'] = self.reference\n\n return stock_head_data\n\n def content(self):\n\n print('Starting to extract content of ', self.head())\n\n tables = self.content_section(\"table\")\n owner_structure_table = self.find_table(tables, 'Aktionär')\n balance_sheet_table = self.find_table(tables, 'Bilanz')\n income_statement_table = self.find_table(tables, 'Gewinn- und Verlustrechnung')\n cash_flow_table = self.find_table(tables, 'Jahrescashflow')\n stock_data_table = self.find_table(tables, 'Wertpapierdaten')\n kpi_data_table = self.find_table(tables, 'Bewertungszahlen')\n employee_table = self.find_table(tables, 'Mitarbeiter')\n\n if None not in [owner_structure_table, balance_sheet_table, income_statement_table, cash_flow_table, stock_data_table, kpi_data_table, employee_table]:\n owner_structure = self.extract_owners(owner_structure_table)\n balance_sheet = self.extract_balance_sheet(balance_sheet_table)\n income_statement = self.extract_kpi(income_statement_table, 'Gewinn- und Verlustrechnung')\n cash_flow = self.extract_kpi(cash_flow_table, 'Jahrescashflow')\n stock_data = self.extract_kpi(stock_data_table, 'Wertpapierdaten')\n kpi_data = self.extract_kpi(kpi_data_table, 'Bewertungszahlen')\n employees = self.extract_kpi(employee_table, 'Mitarbeiter')\n all_kpi = {**income_statement, **cash_flow, **stock_data, **kpi_data, **employees}\n\n if self.kpi_are_valid(all_kpi):\n return {'balance_sheet': balance_sheet, 'head': self.head(), 'owner_structure': owner_structure, 'kpi': all_kpi}\n print('!! FOUND FOULTY STUFF HERE ...', self.head())\n return None\n\n def kpi_are_valid(self, all_kpi):\n return any('dividendPerShare' == key for key in all_kpi.keys()) \\\n and any('earningsPerShare' == key for key in all_kpi.keys()) \\\n and any('bookValuePerShare' == key for key in all_kpi.keys())\n\n def find_table(self, tables, title):\n corresponding = [table for table in tables if has_header(table, title)][:1]\n if len(corresponding) > 0:\n return corresponding[0]\n else:\n return None\n\n def extract_owners(self, owner_table):\n\n owner_data = []\n all_rows = owner_table(\"tr\")\n for row in all_rows:\n cells = row(\"td\")\n if cells and len(cells) > 1:\n owner_data.append({\n 'owner': cells[0].string,\n 'percent': extract_percent(cells[1].string)\n })\n return owner_data\n\n def extract_kpi(self, kpi_table, header_name):\n\n years = extract_years_header(kpi_table, header_name)\n\n kpi_data = {\n 'years': years\n }\n\n for value_row in kpi_table.tbody(\"tr\"):\n columns = value_row(\"td\")\n row_name = columns[0].string\n\n kpi_field = translate_row(row_name)\n if kpi_field is not None:\n row_values = []\n for column_index in range(1, len(columns)):\n value = parse_german_float(columns[column_index].string)\n year = years[column_index - 1]\n row_values.append({'year': year, 'value': value})\n kpi_data[kpi_field] = row_values\n\n return kpi_data\n\n def extract_balance_sheet(self, balance_sheet_table):\n\n years = extract_years_header(balance_sheet_table, 'Bilanz')\n\n balance_sheet = {\n 'years': years,\n 'assets': {},\n 'liabilities_and_equity': {}\n }\n\n active_element = {}\n\n for value_row in balance_sheet_table.tbody(\"tr\"):\n columns = value_row(\"td\")\n row_name = columns[0].string\n\n if row_name == 'Aktiva':\n active_element = balance_sheet['assets']\n continue\n\n if row_name == 'Passiva':\n active_element = balance_sheet['liabilities_and_equity']\n continue\n\n if row_name.strip() == '':\n continue\n\n row_values = []\n for column_index in range(1, len(columns)):\n value = parse_german_float(columns[column_index].string)\n year = years[column_index - 1]\n row_values.append({'year': year, 'value': value})\n active_element[row_name] = row_values\n\n return balance_sheet\n","sub_path":"extraction/extractor.py","file_name":"extractor.py","file_ext":"py","file_size_in_byte":7463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"539339084","text":"from pylab import * \n\nlabels = ['skin','HEPMASS-high level','HIGGS-high level','synthetic','HEPMASS','HIGGS'] # do HIGGS wide\nx = 1 + arange(len(labels))\n\n# y1 = [0.33,6.83,21.69,122.21] # base\n# y2 = [0.07,0.32,0.28,0.23] # radon\n# y3 = [0.07,0.18,0.20,0.23] # PCA\n\n\n\ny1 = [0.33,6.83,21.69,122.21,19.04,74.72] # base\ny2 = [0.07,0.32,0.28 ,0.23 ,0.48 ,0.52] # radon\ny3 = [0.07,0.18,0.20 ,0.23 ,0.31 ,0.49] # PCA\n\n# print(y1)\n# print(y2)\n# print(y3)\n# exit()\ny1 = array(y1)\ny2 = array(y2)\ny3 = array(y3)\n\n# y1 /= y3 \n# y2 /= y3 \n# y3 /= y3\n\nfigure(dpi=300)\ntitle('execution time across various datasets')\nbar(x-0.2,y1,width=0.2,color='r', log=True, align='center')\nbar(x,y2,width=0.2,color='b', log=True, align='center')\nbar(x+0.2,y3,width=0.2,color='g', log=True, align='center')\nxticks(x, labels)\nylabel('seconds')\nlegend(('base learner','radon machine','PCA'))\n\nshow()","sub_path":"Plotting/bargrapher-template.py","file_name":"bargrapher-template.py","file_ext":"py","file_size_in_byte":872,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"641217960","text":"from PyQt5.QtWidgets import (QAction, qApp, QMessageBox, QWidget, QInputDialog)\r\nfrom PyQt5.QtGui import QIcon\r\nimport webbrowser\r\nimport os\r\nfrom brotherData import ImportExportDB\r\n\r\nclass BrotherLogic(QWidget): \r\n \r\n def createActions(self):\r\n \r\n #Menubar granola goes here, it's cunchy and suite.\r\n self.exitAction = QAction(QIcon('icon/exit.png'), '&Exit', self)\r\n self.exitAction.setShortcut('Ctrl+Q')\r\n self.exitAction.setStatusTip('Exit Application')\r\n self.exitAction.triggered.connect(qApp.instance().exit)\r\n \r\n self.queueAction = QAction(QIcon('icon/btn-queue1.png'), '&View Next Queue', self)\r\n self.queueAction.setShortcut('F1')\r\n self.queueAction.setStatusTip('View Next Queue')\r\n self.queueAction.triggered.connect(BrotherLogic.btnNextQueue_Click)\r\n \r\n self.loadAction = QAction(QIcon('icon/load.png'), '&Load Queue', self)\r\n self.loadAction.setShortcut('F2')\r\n self.loadAction.setStatusTip('Load Next queue into database.')\r\n self.loadAction.triggered.connect(self.btnLoadQueue_Click)\r\n \r\n self.printAction = QAction(QIcon('icon/print-shirt.png'), '&Print', self)\r\n self.printAction.setShortcut('F3')\r\n self.printAction.setStatusTip('Print from queue server.')\r\n self.printAction.triggered.connect(self.btnPrint_Click)\r\n \r\n self.sourceAction = QAction(QIcon('icon/print-source.png'), 'Print &Source', self)\r\n self.sourceAction.setShortcut('F4')\r\n self.sourceAction.setStatusTip('Print from art server source.')\r\n self.sourceAction.triggered.connect(self.btnPrintSource_Click)\r\n \r\n self.finishAction = QAction(QIcon('icon/finish-queue.png'), 'Add queue to finished', self)\r\n self.finishAction.setShortcut('F5')\r\n self.finishAction.setStatusTip('Add queue to database as printed.')\r\n self.finishAction.triggered.connect(self.btnFinishQueue_Click)\r\n \r\n self.killAction = QAction(QIcon('icon/kill-arp.png'), '&Kill ARP', self)\r\n self.killAction.setShortcut('F6')\r\n self.killAction.setStatusTip('Kill artwork for recreation.')\r\n self.killAction.triggered.connect(self.btnKillArp_Click)\r\n \r\n self.blankAction = QAction(QIcon(\"icon/out-of-stock.png\"), \"&Out of stock\", self)\r\n self.blankAction.setShortcut('F7')\r\n self.blankAction.setStatusTip(\"Add item to blank out of stock\")\r\n self.blankAction.triggered.connect(self.btnBlankStock_Click)\r\n \r\n def btnNextQueue_Click(self):\r\n \r\n ie = webbrowser.get(webbrowser.iexplore)\r\n ie.open('http://sqlrptserver/ReportServer_SQLREPORTS/Pages/ReportViewer.aspx?%2fInkPixi+Reports%' +\\\r\n '2frptIPGetOldestOpenQueueNotLoaded&rs:Command=Render&rc:Parameters=Collapsed&rc:Toolbar=False')\r\n \r\n def loginDialog(self):\r\n\r\n prnName, ok = QInputDialog.getText(self, 'Change User', 'Please enter user name:')\r\n\r\n if ok:\r\n if ImportExportDB.printerCount(self, prnName) > 0:\r\n self.lblUser.setText(prnName)\r\n ImportExportDB.insertPrinter(self, prnName)\r\n else:\r\n msg = \"Please enter your correct user name, which is your first name and last initial. Example: \\\"brodiew\\\"\"\r\n QMessageBox.warning(self, 'User not found.',\r\n msg,\r\n QMessageBox.Ok)\r\n BrotherLogic.loginDialog(self)\r\n else:\r\n BrotherLogic.loginDialog(self) \r\n \r\n def loadQueue(self, queue):\r\n \r\n ImportExportDB.insertQueue(self, queue)\r\n \r\n #Force web browser to use IE for opening report, because it is currently the only browser\r\n #That can digest the AD user so no pop-up occurs.\r\n ie = webbrowser.get(webbrowser.iexplore)\r\n ie.open('http://sqlrptserver/ReportServer_SQLREPORTS/Pages/ReportViewer.aspx?%' + \\\r\n '2fInkPixi+Reports%2frptQueueExceptions&queue='+ queue + '&rs:' + \\\r\n 'Command=Render&rc:Parameters=Collapsed&rc:Toolbar=False') \r\n \r\n def printBarcode(self):\r\n\r\n barcodeNumber, ok = QInputDialog.getText(self, 'Print from QServer', 'Please scan barcode:')\r\n\r\n if ok:\r\n if barcodeNumber == '':\r\n self.btnPrint_Click()\r\n elif barcodeNumber == 'Send,':\r\n self.btnPrint_Click()\r\n else: \r\n try:\r\n arp = \"\\\\\\\\qserver\\\\Brother Artwork Storage\\\\\" + barcodeNumber + \".arp\"\r\n os.startfile(arp)\r\n ImportExportDB.printARP(self, barcodeNumber)\r\n except BaseException as e:\r\n QMessageBox.warning(self, 'Artwork not found!', \"Artwork not found! \\n\" + str(e), QMessageBox.Ok)\r\n #self.btnPrint_Click() \r\n finally:\r\n self.btnPrint_Click()\r\n \r\n def printBarcodeSource(self):\r\n\r\n barcodeNumber, ok = QInputDialog.getText(self, 'Print from Source.', 'Please scan barcode:')\r\n \r\n if ok:\r\n if barcodeNumber == '':\r\n self.btnPrintSource_Click()\r\n elif barcodeNumber == 'Send,':\r\n self.btnPrintSource_Click()\r\n else:\r\n try:\r\n arp = ImportExportDB.printSource(self, barcodeNumber)\r\n os.startfile(arp)\r\n except BaseException as e:\r\n QMessageBox.warning(self, 'Artwork not found!', \"Artwork not found! \\n\" + str(e), QMessageBox.Ok)\r\n finally:\r\n self.btnPrintSource_Click()\r\n \r\n def finishQueue(self):\r\n queueNumber, ok = QInputDialog.getText(self, 'Enter Queue', 'Please scan or enter queue:')\r\n\r\n if ok:\r\n ImportExportDB.insertQueuePrinted(self, queueNumber)\r\n QMessageBox.information(self, 'Queue finished', 'You\\'re queue as been marked complete.',\r\n QMessageBox.Close) \r\n \r\n def killARP(self, barcodeNumber):\r\n try:\r\n ImportExportDB.killARP(self, barcodeNumber)\r\n QMessageBox.information(self, 'He\\'s dead Jim.', 'ARP has been terminated with extreme prejudice.',\r\n QMessageBox.Ok)\r\n except:\r\n QMessageBox.information(self, 'She\\'s still there.', 'Your ARP is still kicking',\r\n QMessageBox.Ok) \r\n def blankStock(self):\r\n printTypes = (\"Brother\", \"Breeze\", \"Embroidery\", \"Kornit\")\r\n barcode, ok = QInputDialog.getText(self, \"Enter barcode\", \"Please scan or enter barcode:\")\r\n if ok:\r\n bc = barcode\r\n printType, ok = QInputDialog.getItem(self, \"Select Print Type\", \"Print Type\", printTypes, 0, True)\r\n if ok and printType:\r\n pt = printType\r\n quantity, ok = QInputDialog.getInt(self, \"Quantity\", \"Please enter quantity:\", 1, 0, 25, 1)\r\n if ok:\r\n qty = quantity\r\n ImportExportDB.insBlankStock(self, bc, pt, qty)\r\n \r\n \r\n \r\n ","sub_path":"brother_print/brotherLogic.py","file_name":"brotherLogic.py","file_ext":"py","file_size_in_byte":7391,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"9957622","text":"\"\"\"The WaveBlocks Project\n\nThis file contains a the block factory.\n\n@author: R. Bourquin\n@copyright: Copyright (C) 2012 R. Bourquin\n@license: Modified BSD License\n\"\"\"\n\n__all__ = [\"BlockFactory\"]\n\n\nclass BlockFactory(object):\n \"\"\"A factory to create instances of various classes\n based on a simple description ``dict``.\n \"\"\"\n\n def __init__(self):\n\n # Load different factory methods\n import GridFactory\n self.__dict__[\"create_grid\"] = GridFactory.create_grid\n\n import PotentialFactory\n self.__dict__[\"create_potential\"] = PotentialFactory.create_potential\n\n import MatrixExponentialFactory\n self.__dict__[\"create_matrixexponential\"] = MatrixExponentialFactory.create_matrixexponential\n\n\n # TODO: Consider \"local\" vs \"global\" description dicts\n # TODO: Consider putting defaults into \"GlobalDefaults\"\n\n\n def create_basis_shape(self, description):\n try:\n bs_type = description[\"type\"]\n except:\n # Default setting\n bs_type = \"HyperCubicShape\"\n\n if bs_type == \"HyperCubicShape\":\n from HyperCubicShape import HyperCubicShape\n limits = description[\"limits\"]\n BS = HyperCubicShape(limits)\n\n elif bs_type == \"HyperbolicCutShape\":\n from HyperbolicCutShape import HyperbolicCutShape\n K = description[\"K\"]\n D = description[\"dimension\"]\n BS = HyperbolicCutShape(D, K)\n\n else:\n raise ValueError(\"Unknown basis shape type \"+str(bs_type))\n\n return BS\n\n\n def create_wavepacket(self, description):\n\n wp_type = description[\"type\"]\n\n if wp_type == \"HagedornWavepacket\" or wp_type == \"HagedornWavepacketCpp\":\n if wp_type == \"HagedornWavepacket\":\n from HagedornWavepacket import HagedornWavepacket\n else:\n from HagedornWavepacketCpp import HagedornWavepacketCpp as HagedornWavepacket\n\n # Initialize a packet\n WP = HagedornWavepacket(description[\"dimension\"],\n description[\"ncomponents\"],\n description[\"eps\"])\n\n # Set parameters\n if description.has_key(\"Pi\"):\n Pi = description[\"Pi\"]\n WP.set_parameters(Pi)\n\n # Configure basis shapes\n if description.has_key(\"basis_shapes\"):\n for component, shapedescr in enumerate(description[\"basis_shapes\"]):\n BS = self.create_basis_shape(shapedescr)\n WP.set_basis_shape(BS, component=component)\n\n # Set coefficients\n if description.has_key(\"coefficients\"):\n for component, data in enumerate(description[\"coefficients\"]):\n for index, value in data:\n WP.set_coefficient(component, index, value)\n\n # And the quadrature\n if description.has_key(\"quadrature\"):\n QE = self.create_quadrature(description[\"quadrature\"])\n WP.set_quadrature(QE)\n else:\n print(\"Warning: no quadrature specified!\")\n\n elif wp_type == \"HagedornWavepacketInhomogeneous\":\n from HagedornWavepacketInhomogeneous import HagedornWavepacketInhomogeneous\n\n # Initialize a packet\n WP = HagedornWavepacketInhomogeneous(description[\"dimension\"],\n description[\"ncomponents\"],\n description[\"eps\"])\n\n # Set parameters\n if description.has_key(\"Pi\"):\n Pi = description[\"Pi\"]\n WP.set_parameters(Pi)\n\n # Configure basis shapes\n if description.has_key(\"basis_shapes\"):\n for component, shapedescr in enumerate(description[\"basis_shapes\"]):\n BS = self.create_basis_shape(shapedescr)\n WP.set_basis_shape(BS, component=component)\n\n # Set coefficients\n if description.has_key(\"coefficients\"):\n for component, data in enumerate(description[\"coefficients\"]):\n for index, value in data:\n WP.set_coefficient(component, index, value)\n\n # And the quadrature\n if description.has_key(\"quadrature\"):\n QE = self.create_quadrature(description[\"quadrature\"])\n WP.set_quadrature(QE)\n else:\n print(\"Warning: no quadrature specified!\")\n\n else:\n raise ValueError(\"Unknown wavepacket type \"+str(wp_type))\n\n return WP\n\n\n def create_quadrature(self, description):\n try:\n qe_type = description[\"type\"]\n except:\n # Default setting\n qe_type = \"InhomogeneousQuadrature\"\n\n # TODO: Maybe denest QR initialization?\n if qe_type == \"HomogeneousQuadrature\":\n from HomogeneousQuadrature import HomogeneousQuadrature\n QR = self.create_quadrature_rule(description[\"qr\"])\n QE = HomogeneousQuadrature(QR)\n\n elif qe_type == \"InhomogeneousQuadrature\":\n from InhomogeneousQuadrature import InhomogeneousQuadrature\n QR = self.create_quadrature_rule(description[\"qr\"])\n QE = InhomogeneousQuadrature(QR)\n\n else:\n raise ValueError(\"Unknown basis shape type \"+str(qe_type))\n\n return QE\n\n\n def create_quadrature_rule(self, description):\n qr_type = description[\"type\"]\n\n if qr_type == \"GaussHermiteQR\":\n from GaussHermiteQR import GaussHermiteQR\n order = description[\"order\"]\n assert type(order) == int\n QR = GaussHermiteQR(order)\n\n elif qr_type == \"TensorProductQR\":\n from TensorProductQR import TensorProductQR\n # Iteratively create all quadrature rules necessary\n qrs = [ self.create_quadrature_rule(desc) for desc in description[\"qr_rules\"] ]\n QR = TensorProductQR(qrs)\n\n return QR\n","sub_path":"src/WaveBlocksND/BlockFactory.py","file_name":"BlockFactory.py","file_ext":"py","file_size_in_byte":6077,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"603875786","text":"nome = input('Qual é o seu nome?')\nprint ('Muito prazer {:10}!'.format(nome))\n\nnome = input('Qual é o seu nome?')\nprint ('Muito prazer {:>10}!'.format(nome))\n\nnome = input('Qual é o seu nome?')\nprint ('Muito prazer {:<10}!'.format(nome))\n\nnome = input('Qual é o seu nome?')\nprint ('Muito prazer {:^10}!'.format(nome))\n\nnome = input('Qual é o seu nome?')\nprint('Muito prazer {}!'.format(nome))\n\n##\n\nn1 = int(input('Um valor: '))\nn2 = int(input('Outro valor: '))\ns = n1 + n2\nm = n1 * n2\nd = n1 / n2\ndi = n1 // n2\ne = n1 ** n2\nr = n1 % n2\n\nprint ('A soma é {}, o produto {}, e a divisão {:.3f}'.format(s, m, d), end='')\n# utilize end='' para não quebrar linha\nprint (' Divisão inteira é {}, a exponenciação {}, e o \\n módulo {}'.format(di,e,r))\n# utilize \\n para quebrar linha\n\n\n\n","sub_path":"Aula7/Aula07a.py","file_name":"Aula07a.py","file_ext":"py","file_size_in_byte":818,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"155324418","text":"import os\nfrom flask import Flask\nimport twitter\nimport markovify\n\napp = Flask(__name__)\n\nif os.environ.get('IS_HEROKU') == None:\n from dotenv import load_dotenv\n load_dotenv('.env')\n\napi = twitter.Api(consumer_key=os.environ.get(\"TWITTER_CONSUMER_KEY\"),\n consumer_secret=os.environ.get(\"TWITTER_CONSUMER_SECRET\"),\n access_token_key=os.environ.get(\"TWITTER_ACCESS_TOKEN\"),\n access_token_secret=os.environ.get(\"TWITTER_TOKEN_SECRET\"))\n\n@app.route(\"/\")\ndef hello():\n return \"Hello World!\"\n\n@app.route('/u/')\ndef get_tweets(user):\n statuses = api.GetUserTimeline(screen_name=user)\n text_model = markovify.Text(' '.join([t.text for t in statuses]))\n return text_model.make_sentence()\n\nif __name__ == \"__main__\":\n app.run()\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":809,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"328626598","text":"from __future__ import unicode_literals\nfrom django.db import models\nfrom tenant_schemas.models import TenantMixin\n\nclass Client(TenantMixin):\n name = models.CharField(max_length=100)\n email = models.EmailField()\n paid_until = models.DateField()\n on_trial = models.BooleanField()\n created_on = models.DateField(auto_now_add=True)\n\n # default true, schema will be automatically created and synced when it is saved\n auto_create_schema = True\n\n def __unicode__(self):\n return u'%s - %s' % (self.domain_url, self.schema_name)","sub_path":"gym_server/customers/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"92242571","text":"from tkinter import *\r\nimport webbrowser\r\nwindow = Tk()\r\nroot = Tk()\r\nframe = Frame(root)\r\nframe.pack()\r\ndef OpenUrl():\r\n webbrowser.open_new(url)\r\nurl = 'webcam.html'\r\nwindow.title(\"ASH VLOGS\")\r\nlbl = Label(window, text=\"click.exe prompt\", font=(\"Arial Bold\", 50))\r\nlbl.grid(column=0, row=0)\r\nwindow.wm_iconbitmap('drip.ico')\r\ndef command():\r\n Toplevel(root)\r\nroot.title(\"Look Behind You\")\r\nbutton = Button(root, text=\"I'm Watching You\", command= OpenUrl, font=(\"Arial Bold\", 20))\r\nroot.wm_iconbitmap('drip.ico')\r\nbutton.pack()\r\nwindow.mainloop()","sub_path":"ASH_VLOGS.py","file_name":"ASH_VLOGS.py","file_ext":"py","file_size_in_byte":553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"546884514","text":"# -*- coding: utf-8 -*-\n\nimport xml.etree.cElementTree as ec\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport os\n\npath = os.getcwd()\npath1 = path + '/wiki_data/FA/'\npath2 = path + '/wiki_data/GA/'\nFAfilenames = os.listdir(path1)\nGAfilenames = os.listdir(path2)\n\nfeaturedArticleList = []\ngoodArticleList = []\nfor filename in FAfilenames:\n if '.xml' in filename:\n filename = path1 + filename\n featuredArticleList.append(filename)\n \nfor filename in GAfilenames:\n if '.xml' in filename:\n filename = path2 + filename\n goodArticleList.append(filename)\n\n# featuredArticleList = ['Arabian_Sea.xml']\n\ndef CurrentText(articleName):\n\ttry:\n\t\ttree = ec.parse(articleName) \n\t\troot = tree.getroot()\n\n\t\tpageElement = root[1]\n\n\t\tlatestRevisionText = ''\n\t\tfor child in pageElement:\n\t\t\tif 'revision' in child.tag:\n\t\t\t\tfor each in child:\n\t\t\t\t\tif 'text' in each.tag:\n\t\t\t\t\t\tlatestRevisionText = each.text\n\n\t\treturn latestRevisionText\n\texcept:\n\t\tprint('\\n'+'Error! in parsing '+articleName+'\\n')\n\t\treturn -1\n\ndef NumberOfImages(currentText, articleName):\n\timageFormates = ['.jpg','.jpeg','.svg','.gif','.png','.bmp','.tiff']\n\ttry:\n\t\tcount = 0\n\t\tfor image in imageFormates:\n\t\t\tcount += currentText.count(image)\t\n\t\tprint(articleName + ' ' + str(count))\t\n\t\treturn count\n\n\texcept:\n\t\tprint('\\n'+'Something went wrong!'+articleName+'\\n')\n\t\treturn -1\n\ndef main():\n listOfImagesFA = []\n listOfImagesGA = []\n xAxis = []\n sc = []\n count=1\n for articleName in featuredArticleList: \n currentText = CurrentText(articleName)\n if currentText != -1:\n listOfImagesFA.append(NumberOfImages(currentText, articleName))\n xAxis.append(count)\n sc.append(5)\n count+=1\n\n listOfImagesFA.sort()\n for articleName in goodArticleList: \n currentText = CurrentText(articleName)\n if currentText != -1:\n listOfImagesGA.append(NumberOfImages(currentText, articleName))\n\n listOfImagesGA.sort()\n\n plt.xlabel('Articles', fontsize=12)\n plt.ylabel('Number of Images', fontsize=12)\n plt.scatter(xAxis,listOfImagesFA, c='b',s=sc, marker='o',label='Number Of Images in FA')\n plt.scatter(xAxis,listOfImagesGA, c='r',s=sc, marker='^',label='Number Of Images in GA')\n plt.legend()\n plt.savefig('numberOfImages.png',dpi=800)\n plt.show()\t\n #npArray = np.array(listOfImages)\n #print('Mean number of Images ' + str(np.mean(npArray)))\n #print('Standard Deviation ' + str(np.std(npArray)) + '\\n')\n\nmain()","sub_path":"wiki_codes/NumberOfImages.py","file_name":"NumberOfImages.py","file_ext":"py","file_size_in_byte":2521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"98950489","text":"# Created by: Dr. David John & Kenneth Meza.\n# Created at: March, 2016.\n# Updated at: April, 2016.\n\n# LIBRARIES\nfrom business_logic.chromosome import Chromosome\nfrom business_logic.general_functions import *\nfrom copy import deepcopy\nfrom random import randint\n\n\n# ===================\n# SELECTION FUNCTIONS\n# ===================\n\n# FUNCTION: fitness_calculator\ndef fitness_calculator(ordered_population):\n \"\"\"\n Calculates the 'fitness' that will be used for doing a rank based selection.\n\n Args:\n ordered_population : LIST[Chromosome]\n A list filled with 'Chromosome' objects sorted by the likelihood result\n \"\"\"\n size = len(ordered_population)\n for i in range(0, size):\n ordered_population[i].fitness = (2*(size+1-(i+1)))/(size*(size+1))\n\n\n# FUNCTION: selection_function\ndef selection_function(population, threshold):\n \"\"\"\n Creates a new generation of chromosomes, based on the CHC Algorithm.\n\n Args:\n population : LIST[Chromosome]\n A list filled with 'Chromosome' objects\n threshold : INT\n The limit used as a filter in the algorithm\n\n Returns:\n LIST[Chromosome]\n A list filled with 'Chromosome' objects, containing the new population\n \"\"\"\n population_visited = [i for i in range(0, len(population))]\n children_population = []\n for i in range(1, len(population)//2):\n random1 = randint(0, len(population_visited)-1)\n random2 = randint(0, len(population_visited)-1)\n while random1 == random2:\n random2 = randint(0, len(population_visited)-1)\n\n index_a = population_visited[random1]\n index_b = population_visited[random2]\n\n parent_a = population[index_a]\n parent_b = population[index_b]\n\n hamming_distance = calculate_hamming_distance(parent_a.get_genes(), parent_b.get_genes())\n if hamming_distance > threshold:\n children = HUX_crossover_function(parent_a, parent_b)\n children_population.append(children[0])\n children_population.append(children[1])\n population_visited.remove(index_a)\n population_visited.remove(index_b)\n return children_population\n\n\n# FUNCTION: HUX_crossover_function\ndef HUX_crossover_function(parent_a, parent_b):\n \"\"\"\n The HUX crossover (Half Uniform Crossover) consists in finding all the bits that differ between two parents and\n swapping the half of this bits, creating two new offsprings.\n\n Args:\n parent_a : Chromosome\n An object 'Chromosome' that will be used for mating\n parent_b : Chromosome\n An object 'Chromosome' that will be used for mating\n\n Returns:\n TUPLE(Chromosome, Chromosome)\n A tuple formed by two offsprings, represented by a 'Chromosome' object\n \"\"\"\n genes_a = deepcopy(parent_a.get_genes())\n genes_b = deepcopy(parent_b.get_genes())\n\n positions = get_positions(genes_a, genes_b)\n for i in range(0, len(positions)//2):\n index_pos = randint(0, len(positions)-1)\n x = positions[index_pos][0]\n y = positions[index_pos][1]\n genes_a[x][y], genes_b[x][y] = genes_b[x][y], genes_a[x][y]\n positions.remove(positions[index_pos])\n return Chromosome(genes_a), Chromosome(genes_b)\n","sub_path":"business_logic/selection_functions.py","file_name":"selection_functions.py","file_ext":"py","file_size_in_byte":3288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"365342730","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2017/10/3 PM4:16\n# @Author : Shiloh Leung\n# @Site : \n# @File : ntucker_demo.py\n# @Software: PyCharm Community Edition\n\nimport tensorflow as tf\nfrom tensorD.factorization.env import Environment\nfrom tensorD.dataproc.provider import Provider\nfrom tensorD.factorization.ntucker import NTUCKER_BCU\nfrom tensorD.demo.DataGenerator import *\n\nif __name__ == '__main__':\n print('=========Train=========')\n X = synthetic_data_tucker([30, 30, 30], [10, 10, 10])\n data_provider = Provider()\n data_provider.full_tensor = lambda: X\n env = Environment(data_provider, summary_path='/tmp/ntucker_demo_' + '30')\n ntucker = NTUCKER_BCU(env)\n args = NTUCKER_BCU.NTUCKER_Args(ranks=[10, 10, 10], validation_internal=1)\n ntucker.build_model(args)\n ntucker.train(1000)\n factor_matrices = ntucker.factors\n core_tensor = ntucker.core\n print('Train ends.\\n\\n\\n')\n","sub_path":"tensorD/demo/ntucker_demo.py","file_name":"ntucker_demo.py","file_ext":"py","file_size_in_byte":940,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"35486632","text":"import requests\nfrom bs4 import BeautifulSoup\n\n\nurl = \"https://www.imdb.com/chart/top/?ref_=nv_mv_250\"\n\nhtml = requests.get(url).content\nsoup = BeautifulSoup(html,\"html.parser\")\n\nlist = soup.find('tbody',{\"class\":\"lister-list\"}).find_all(\"tr\",limit=10)\nprint(\"film adı ve yılı\")\nfor tr in list:\n title = tr.find(\"td\",{\"class\":\"titleColumn\"}).find(\"a\").text\n year = tr.find(\"td\", {\"class\": \"titleColumn\"}).find(\"span\").text\n rating = tr.find(\"td\",{\"class\": \"ratingColumn\"}).find(\"strong\").text\n print(\"film adı:\"+title)\n print(\"yapım yılı: \"+year)\n print(\"rating: \"+rating)\n\n","sub_path":"imdb.py","file_name":"imdb.py","file_ext":"py","file_size_in_byte":599,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"636706068","text":"from portality.lib import dataobj\nfrom portality import models\n\nfrom portality.api.v1.data_objects.common_journal_application import OutgoingCommonJournalApplication\n\n# we only have outgoing journals for the moment\nJOURNAL_STRUCT = {\n \"objects\": [\"bibjson\", \"admin\"],\n \"fields\": {\n \"id\": {\"coerce\": \"unicode\"},\n \"created_date\": {\"coerce\": \"utcdatetime\"},\n \"last_updated\": {\"coerce\": \"utcdatetime\"}\n },\n \"structs\": {\n \"admin\": {\n \"fields\": {\n \"in_doaj\": {\"coerce\": \"bool\", \"get__default\": False},\n \"ticked\": {\"coerce\": \"bool\", \"get__default\": False},\n \"seal\": {\"coerce\": \"bool\", \"get__default\": False} }\n },\n \"bibjson\": {\n \"fields\": {\n \"allows_fulltext_indexing\": {\"coerce\": \"bool\"},\n \"alternative_title\": {\"coerce\": \"unicode\"},\n \"apc_url\": {\"coerce\": \"unicode\"},\n \"country\": {\"coerce\": \"country_code\"},\n \"institution\": {\"coerce\": \"unicode\"},\n \"provider\": {\"coerce\": \"unicode\"},\n \"publication_time\": {\"coerce\": \"integer\"},\n \"publisher\": {\"coerce\": \"unicode\"},\n \"submission_charges_url\": {\"coerce\": \"unicode\"},\n \"title\": {\"coerce\": \"unicode\"},\n },\n \"lists\": {\n \"deposit_policy\": {\"coerce\": \"deposit_policy\", \"contains\": \"field\"},\n \"format\": {\"coerce\": \"format\", \"contains\": \"field\"},\n \"identifier\": {\"contains\": \"object\"},\n \"keywords\": {\"coerce\": \"unicode\", \"contains\": \"field\"},\n \"language\": {\"coerce\": \"isolang_2letter\", \"contains\": \"field\"},\n \"license\": {\"contains\": \"object\"},\n \"link\": {\"contains\": \"object\"},\n \"persistent_identifier_scheme\": {\"coerce\": \"persistent_identifier_scheme\", \"contains\": \"field\"},\n \"subject\": {\"contains\": \"object\"}\n },\n \"objects\": [\n \"apc\",\n \"archiving_policy\",\n \"article_statistics\",\n \"author_copyright\",\n \"author_publishing_rights\",\n \"editorial_review\",\n \"oa_start\",\n \"oa_end\",\n \"plagiarism_detection\",\n \"submission_charges\",\n ],\n\n \"structs\": {\n \"apc\": {\n \"fields\": {\n \"currency\": {\"coerce\": \"currency_code\"},\n \"average_price\": {\"coerce\": \"integer\"}\n }\n },\n\n \"archiving_policy\": { # NOTE: this is not the same as the storage model, so beware when working with this\n \"fields\": {\n \"url\": {\"coerce\": \"unicode\"},\n },\n \"lists\": {\n \"policy\": {\"coerce\": \"unicode\", \"contains\": \"object\"},\n },\n\n \"structs\" : {\n \"policy\" : {\n \"fields\" : {\n \"name\" : {\"coerce\": \"unicode\"},\n \"domain\" : {\"coerce\" : \"unicode\"}\n }\n }\n }\n },\n\n \"article_statistics\": {\n \"fields\": {\n \"statistics\": {\"coerce\": \"bool\"},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"author_copyright\": {\n \"fields\": {\n \"copyright\": {\"coerce\": \"unicode\"},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"author_publishing_rights\": {\n \"fields\": {\n \"publishing_rights\": {\"coerce\": \"unicode\"},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"editorial_review\": {\n \"fields\": {\n \"process\": {\"coerce\": \"unicode\", \"allowed_values\" : [\"Editorial review\", \"Peer review\", \"Blind peer review\", \"Double blind peer review\", \"Open peer review\", \"None\"]},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"identifier\": {\n \"fields\": {\n \"type\": {\"coerce\": \"unicode\"},\n \"id\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"license\": {\n \"fields\": {\n \"title\": {\"coerce\": \"license\"},\n \"type\": {\"coerce\": \"license\"},\n \"url\": {\"coerce\": \"unicode\"},\n \"version\": {\"coerce\": \"unicode\"},\n \"open_access\": {\"coerce\": \"bool\"},\n \"BY\": {\"coerce\": \"bool\"},\n \"NC\": {\"coerce\": \"bool\"},\n \"ND\": {\"coerce\": \"bool\"},\n \"SA\": {\"coerce\": \"bool\"},\n \"embedded\": {\"coerce\": \"bool\"},\n \"embedded_example_url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"link\": {\n \"fields\": {\n \"type\": {\"coerce\": \"unicode\"},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"oa_start\": {\n \"fields\": {\n \"year\": {\"coerce\": \"integer\"},\n \"volume\": {\"coerce\": \"unicode\"},\n \"number\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"oa_end\": {\n \"fields\": {\n \"year\": {\"coerce\": \"integer\"},\n \"volume\": {\"coerce\": \"unicode\"},\n \"number\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"plagiarism_detection\": {\n \"fields\": {\n \"detection\": {\"coerce\": \"bool\"},\n \"url\": {\"coerce\": \"unicode\"},\n }\n },\n\n \"submission_charges\": {\n \"fields\": {\n \"currency\": {\"coerce\": \"currency_code\"},\n \"average_price\": {\"coerce\": \"integer\"}\n }\n },\n \"subject\": {\n \"fields\": {\n \"scheme\": {\"coerce\": \"unicode\"},\n \"term\": {\"coerce\": \"unicode\"},\n \"code\": {\"coerce\": \"unicode\"},\n }\n }\n }\n }\n }\n}\n\n\nclass OutgoingJournal(OutgoingCommonJournalApplication):\n\n def __init__(self, raw=None):\n super(OutgoingJournal, self).__init__(raw, struct=JOURNAL_STRUCT, construct_silent_prune=True, expose_data=True)\n\n @classmethod\n def from_model(cls, jm):\n assert isinstance(jm, models.Journal)\n return super(OutgoingJournal, cls).from_model(jm)\n\n @classmethod\n def from_model_by_id(cls, id_):\n j = models.Journal.pull(id_)\n return cls.from_model(j)","sub_path":"portality/api/v1/data_objects/journal.py","file_name":"journal.py","file_ext":"py","file_size_in_byte":7343,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"89430611","text":"# Copyright 2016 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport netaddr\n\nfrom collections import defaultdict\n\nfrom datetime import datetime\nfrom pepper import libpepper\nfrom tcp_tests.helpers import utils\nfrom tcp_tests import logger\nfrom tcp_tests import settings\nfrom tcp_tests.managers.execute_commands import ExecuteCommandsMixin\n\nLOG = logger.logger\n\n\nclass SaltManager(ExecuteCommandsMixin):\n \"\"\"docstring for SaltManager\"\"\"\n\n __config = None\n __underlay = None\n _map = {\n 'enforceState': 'enforce_state',\n 'enforceStates': 'enforce_states',\n 'runState': 'run_state',\n 'runStates': 'run_states',\n }\n\n def __init__(self, config, underlay, host=None, port='6969',\n username=None, password=None):\n self.__config = config\n self.__underlay = underlay\n self.__port = port\n self.__host = host\n self.__api = None\n self.__user = username or settings.SALT_USER\n self.__password = password or settings.SALT_PASSWORD\n self._salt = self\n\n super(SaltManager, self).__init__(config=config, underlay=underlay)\n\n def install(self, commands):\n # if self.__config.salt.salt_master_host == '0.0.0.0':\n # # Temporary workaround. Underlay should be extended with roles\n # salt_nodes = self.__underlay.node_names()\n # self.__config.salt.salt_master_host = \\\n # self.__underlay.host_by_node_name(salt_nodes[0])\n\n self.execute_commands(commands=commands,\n label=\"Install and configure salt\")\n\n def change_creds(self, username, password):\n self.__user = username\n self.__password = password\n\n @property\n def port(self):\n return self.__port\n\n @property\n def host(self):\n if self.__host:\n return self.__host\n else:\n # TODO(ddmitriev): consider to add a check and raise\n # exception if 'salt_master_host' is not initialized.\n return self.__config.salt.salt_master_host\n\n @property\n def api(self):\n def login():\n LOG.info(\"Authentication in Salt API\")\n self.__api.login(\n username=self.__user,\n password=self.__password,\n eauth='pam')\n return datetime.now()\n\n if self.__api:\n if (datetime.now() - self.__session_start).seconds < 5 * 60:\n return self.__api\n else:\n # FIXXME: Change to debug\n LOG.info(\"Session's expired\")\n self.__session_start = login()\n return self.__api\n\n url = \"http://{host}:{port}\".format(\n host=self.host, port=self.port)\n LOG.info(\"Connecting to Salt API {0}\".format(url))\n self.__api = libpepper.Pepper(url)\n self.__session_start = login()\n return self.__api\n\n def local(self, tgt, fun, args=None, kwargs=None):\n return self.api.local(tgt, fun, args, kwargs, expr_form='compound')\n\n def local_async(self, tgt, fun, args=None, kwargs=None):\n return self.api.local_async(tgt, fun, args, kwargs)\n\n def lookup_result(self, jid):\n return self.api.lookup_jid(jid)\n\n def check_result(self, r):\n if len(r.get('return', [])) == 0:\n raise LookupError(\"Result is empty or absent\")\n\n result = r['return'][0]\n if len(result) == 0:\n raise LookupError(\"Result is empty or absent\")\n LOG.info(\"Job has result for %s nodes\", result.keys())\n fails = defaultdict(list)\n for h in result:\n host_result = result[h]\n LOG.info(\"On %s executed:\", h)\n if isinstance(host_result, list):\n fails[h].append(host_result)\n continue\n for t in host_result:\n task = host_result[t]\n if task['result'] is False:\n fails[h].append(task)\n LOG.error(\"%s - %s\", t, task['result'])\n else:\n LOG.info(\"%s - %s\", t, task['result'])\n\n return fails if fails else None\n\n def enforce_state(self, tgt, state, args=None, kwargs=None):\n r = self.local(tgt=tgt, fun='state.sls', args=state)\n f = self.check_result(r)\n return r, f\n\n def enforce_states(self, tgt, state, args=None, kwargs=None):\n rets = []\n for s in state:\n r = self.enforce_state(tgt=tgt, state=s)\n rets.append(r)\n return rets\n\n def run_state(self, tgt, state, args=None, kwargs=None):\n return self.local(tgt=tgt, fun=state, args=args, kwargs=kwargs), None\n\n def run_states(self, tgt, state, args=None, kwargs=None):\n rets = []\n for s in state:\n r = self.run_state(tgt=tgt, state=s, args=args, kwargs=kwargs)\n rets.append(r)\n return rets\n\n def get_pillar(self, tgt, pillar):\n result = self.local(tgt=tgt, fun='pillar.get', args=pillar)\n return result['return']\n\n def get_grains(self, tgt, grains):\n result = self.local(tgt=tgt, fun='grains.get', args=grains)\n return result['return']\n\n def get_ssh_data(self):\n \"\"\"Generate ssh config for Underlay\n\n :param roles: list of strings\n \"\"\"\n\n pool_name = self.__config.underlay.net_mgmt\n pool_net = netaddr.IPNetwork(self.__config.underlay.address_pools[\n self.__config.underlay.net_mgmt])\n hosts = self.local('*', 'grains.item', ['host', 'ipv4'])\n\n if len(hosts.get('return', [])) == 0:\n raise LookupError(\"Hosts is empty or absent\")\n hosts = hosts['return'][0]\n if len(hosts) == 0:\n raise LookupError(\"Hosts is empty or absent\")\n\n def host(node_name, ip):\n return {\n 'roles': ['salt_minion'],\n 'keys': [\n k['private'] for k in self.__config.underlay.ssh_keys\n ],\n 'node_name': node_name,\n 'host': ip,\n 'address_pool': pool_name,\n 'login': settings.SSH_NODE_CREDENTIALS['login'],\n 'password': settings.SSH_NODE_CREDENTIALS['password']\n }\n\n try:\n ret = [\n host(k, next(i for i in v['ipv4'] if i in pool_net))\n for k, v in hosts.items()\n if next(i for i in v['ipv4'] if i in pool_net)]\n LOG.debug(\"Fetched ssh data from salt grains - {}\".format(ret))\n return ret\n except StopIteration:\n msg = (\"Can't match nodes ip address with network cidr\\n\"\n \"Managment network - {net}\\n\"\n \"Host with address - {host_list}\".format(\n net=pool_net,\n host_list={k: v['ipv4'] for k, v in hosts.items()}))\n raise StopIteration(msg)\n\n def service_status(self, tgt, service):\n result = self.local(tgt=tgt, fun='service.status', args=service)\n return result['return']\n\n def service_restart(self, tgt, service):\n result = self.local(tgt=tgt, fun='service.restart', args=service)\n return result['return']\n\n def service_stop(self, tgt, service):\n result = self.local(tgt=tgt, fun='service.stop', args=service)\n return result['return']\n\n @utils.retry(3, exception=libpepper.PepperException)\n def sync_time(self, tgt='*'):\n LOG.info(\"NTP time sync on the salt minions '{0}'\".format(tgt))\n # Force authentication update on the next API access\n # because previous authentication most probably is not valid\n # before or after time sync.\n self.__api = None\n self.run_state(\n tgt,\n 'cmd.run', 'service ntp stop; if [ -x /usr/sbin/ntpdate ]; then ntpdate -s ntp.ubuntu.com; else ntpd -gq ; fi; service ntp start') # noqa\n new_time_res = self.run_state(tgt, 'cmd.run', 'date')\n for node_name, time in sorted(new_time_res[0]['return'][0].items()):\n LOG.info(\"{0}: {1}\".format(node_name, time))\n self.__api = None\n","sub_path":"tcp_tests/managers/saltmanager.py","file_name":"saltmanager.py","file_ext":"py","file_size_in_byte":8690,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"183976873","text":"__author__ = \"Qiyao_Qin\"\n\nclass TrieNode:\n def __init__(self):\n self.has_word = False\n self.children = {}\n\n\nclass Trie:\n def __init__(self):\n self.root = TrieNode()\n\n def insert(self, word):\n node = self.root\n for letter in word:\n child = node.children.get(letter)\n if child is None:\n child = TrieNode()\n node.children[letter] = child\n node = child\n child.has_word = True\n\n\nclass Solution:\n # @param board, a list of lists of 1 length string\n # @param words: A list of string\n # @return: A list of string\n def wordSearchII(self, board, words):\n # write your code here\n if not board or not words:\n return []\n\n trie = Trie()\n for word in words:\n trie.insert(word)\n\n root = trie.root\n self.result = []\n for row in range(len(board)):\n for col in range(len(board[0])):\n self.search(board, root, row, col, '')\n return self.result\n\n def search(self, board, trie, row, col, res):\n if row < 0 or row >= len(board) or col < 0 or col >= len(board[0]):\n return\n if board[row][col] == '#':\n return\n if board[row][col] not in trie.children:\n return\n trie = trie.children.get(board[row][col])\n res += board[row][col]\n if trie.has_word:\n if res not in self.result:\n trie.has_word = False\n self.result.append(res)\n temp = board[row][col]\n board[row][col] = '#'\n self.search(board, trie, row + 1, col, res)\n self.search(board, trie, row - 1, col, res)\n self.search(board, trie, row, col + 1, res)\n self.search(board, trie, row, col - 1, res)\n board[row][col] = temp\n\n","sub_path":"Trie/word_search_ii.py","file_name":"word_search_ii.py","file_ext":"py","file_size_in_byte":1847,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"527779632","text":"#!/usr/bin/env python3\nimport itertools\nfrom collections import OrderedDict\nfrom keyword import kwlist\nfrom argparse import *\nkwset=set(kwlist)\n\ndef varsNamesGen(alph, l=255):\n\ti=0\n\tj=1\n\twhile True:\n\t\tfor name in map(lambda x: \"\".join(x), itertools.combinations(alph, j)):\n\t\t\tif name not in kwset:\n\t\t\t\tyield name\n\t\t\t\ti+=1\n\t\t\t\tif i>l:\n\t\t\t\t\treturn\n\t\tj+=1\n\ndef genVarNames(len=255):\n\ta=ord(\"a\")\n\talphabetLen=26\n\talphabet=[chr(i) for i in range(a, a+alphabetLen)]\n\ta=ord(\"A\")\n\talphabet.extend([chr(i) for i in range(a, a+alphabetLen)])\n\treturn list(varsNamesGen(alphabet, len))\n\ndef genFunc(name, varNames, indent=0):\n\treturn \"\\n\".join((\n\t\t\"\\t\"*indent+\"def \"+name+\"():\",\n\t\t\"\\t\"*(indent+1)+\"return \"+\"+\".join(varNames),\n\t\t\"\"\n\t))\n\ndef genVarsInitializer(varsDict, indent=0):\n\treturn \"\\n\".join((indent*\"\\t\"+\"=\".join(i) for i in varsDict.items()))+\"\\n\"\n\ndef genMeasuredFunc(name, varsDict, indent=0):\n\treturn \"\".join((\n\t\tgenFunc(name, varsDict.keys(), indent),\n\t\tgenVarsInitializer(varsDict, indent),\n\t\t\"\\t\"*indent+name+\"_inlined=bind.bind(\"+name+\", locals())\"\n\t))\n\ndef genClosureMeasuredFunc(name, varsDict, innerName=None, indent=0):\n\tif innerName is None:\n\t\tinnerName=\"inner_\"+name\n\treturn \"\\n\".join((\n\t\t\"\\t\"*indent+\"def \"+name+\"_gen():\",\n\t\tgenMeasuredFunc(innerName, varsDict, indent+1),\n\t\t\"\\t\"*(indent+1)+\"return (\"+innerName+\", \"+innerName+\"_inlined)\\n\",\n\t\tindent*\"\\t\"+\"(\"+name+\", \"+name+ \"_inlined)=\"+name+\"_gen()\",\n\t\t\"\"\n\t))\n\ndef genMeasurement(name):\n\treturn \"\\n\".join((\n\t\tmeasResultVarName+\"['\"+name+\"']=OrderedDict((\",\n\t\t\"\\t('orig'\t, timeit.timeit(\"+name+\"\t\t\t )),\",\n\t\t\"\\t('inlined', timeit.timeit(\"+name+\"_inlined)),\",\n\t\t\"))\"\n\t))\n\ndef genDis(name):\n\treturn \"dis.dis(\"+name+\")\"\n\nmeasResultVarName=\"measResults\"\npreamble=r\"\"\"\nimport dis, json, timeit, bind\nfrom collections import OrderedDict\n\ndef computeDelta(res):\n\tfor i in res:\n\t\tres[i][\"% faster\"]=(res[i][\"orig\"]-res[i][\"inlined\"])/res[i][\"orig\"]*100\n\n\"\"\"+ measResultVarName+\"={}\"\n\nif __name__ == \"__main__\":\n\targp=ArgumentParser(\n\t\tdescription=\"Generates code for benchmarking the library\"\n\t)\n\targp.add_argument(\"--count\", type=int, help=\"Count of variables per function to be inlined\", default=0xff)\n\targs=argp.parse_args()\n\n\tvarNames=genVarNames(args.count)\n\tvarsDict=dict(zip(varNames, map(str, range(len(varNames)))))\n\n\ttext=\"\\n\".join((\n\t\tpreamble,\n\t\tgenClosureMeasuredFunc(\"load_deref\", varsDict),\n\t\tgenMeasuredFunc(\"load_global\", varsDict),\n\t\tgenMeasurement(\"load_global\"),\n\t\tgenMeasurement(\"load_deref\"),\n\t\t\"computeDelta(\"+measResultVarName+\")\",\n\t\t\"print(json.dumps(\"+measResultVarName+\", indent='\\t'))\"\n\t))\n\tprint(text)\n","sub_path":"benchmarkGen.py","file_name":"benchmarkGen.py","file_ext":"py","file_size_in_byte":2586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"303854784","text":"import random\nimport os\nimport time\n\n\nx = 780\nreal = x/9\nx2 = 350\ncount = 0\nreal2 = real\nwhile (real2 <= x2):\n real2 += real\n count += 1\n print(\"Distância entre o objeto e o início do campo: %.2fm\" % count)\nprint(\"A cada %.2f pixels, um metro em X é caminhado.\" % real)\n","sub_path":"projSIM2111-master/CONTADOR_DISTANCIA.py","file_name":"CONTADOR_DISTANCIA.py","file_ext":"py","file_size_in_byte":282,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"452010524","text":"#should have just used collections.defaultdict\nclass Solution(object):\n def canFinish(self, numCourses, prerequisites):\n \"\"\"\n :type numCourses: int\n :type prerequisites: List[List[int]]\n :rtype: bool\n \"\"\"\n if not prerequisites:\n return True\n g={}\n degrees={}\n zeros=[]\n res=set()\n \n for x in prerequisites:\n degrees.setdefault(x[0],0)\n degrees.setdefault(x[1],0)\n degrees[x[0]]+=1\n \n g.setdefault(x[1],set())\n g.setdefault(x[0],set())\n g[x[1]].add(x[0])\n \n for x in range(numCourses):\n degrees.setdefault(x,0)\n if degrees[x]==0:\n res.add(x)\n if x in g:\n zeros.append(x)\n \n print(zeros) \n while zeros:\n new=zeros.pop(0)\n res.add(new)\n #print(new)\n for x in g[new]:\n degrees[x]-=1\n if degrees[x]==0:\n zeros.append(x)\n \n print(res)\n return numCourses==len(res)\n \n","sub_path":"Graph_TopologicalSort_CircularPermutation/Course_Schedule/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":1172,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"401949092","text":"from appium import webdriver\nimport time\nimport base64\n\nfrom appium.webdriver.common.touch_action import TouchAction\n\ndesired_caps = {}\ndesired_caps['platformName'] = 'Android'\ndesired_caps['platformVersion'] = '7.1.2'\ndesired_caps['deviceName'] = '127.0.0.1:62001'\ndesired_caps['appPackage'] = 'com.android.settings'\ndesired_caps['appActivity'] = '.Settings'\ndesired_caps['unicodeKeyboadr'] = True\ndesired_caps['resetKeyboard'] = True\ndriver = webdriver.Remote(\"http://127.0.0.1:4723/wd/hub\", desired_caps)\n\n# 发送文件到手机目录下\n# with open(\"./python.txt\", \"r\") as f:\n# data = str(base64.b64encode(f.read().encode('utf-8')), 'utf-8')\n# driver.push_file('/storage/emulated/0', data)\n# 拉取手机文件到电脑\n# data = driver.pull_file(\"storage/emulated/0/push.txt\")\n# with open(\"C:/Users/ZXY/Desktop/phone.txt\", \"a\")as f:\n# f.write(str(base64.b64decode(data),\"utf-8\"))\n\n# 屏幕滑动\n# driver.close_app()\n#\n# w_h = driver.get_window_size()\n# w = w_h.get(\"width\")\n# h = w_h.get(\"height\")\n# driver.swipe(w*0.8, h*0.5, w*0.2, h*0.5, 1000)\n\n# 手指轻敲操作\n# 通过元素定位\n# WL_T = driver.find_element_by_xp ath(\"//*[contains(@text, 'WLA')]\")\n# TouchAction(driver).tap(WL_T).perform()\n\n# 通过位置定位\nWL_T = driver.find_element_by_xpath(\"//*[contains(@text, 'WLA')]\")\nTouchAction(driver).tap(x=WL_T.location.get(\"x\"), y=WL_T.location.get(\"y\")).perform()","sub_path":"App-Api/作业练习.py","file_name":"作业练习.py","file_ext":"py","file_size_in_byte":1394,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"319629308","text":"import os\nimport json\nfrom pprint import pprint\nfrom parser import webtotrain \n\n# Cost of correction, in seconds, per word\ncost = {\n \"not_gt_uncorrectable\": 16,\n \"gt_correctable\": 5,\n \"gt_uncorrectable\": 5,\n \"not_gt_corerctable\": 5\n}\n\n# Cost of human per day\nhuman_cost = 350 # Per working day\nworking_hours = 6 # Hours working a day\nhuman_cost_per_hour = human_cost/working_hours\n\ndef analyze(save):\n new = {\n \"not_gt_uncorrectable\": save[\"uncorrectable\"] - \\\n save[\"ocr_equals_gt\"][\"uncorrectable\"] ,\n \"gt_correctable\" : save[\"ocr_equals_gt\"][\"correctable\"],\n \"not_gt_correctable\": save[\"correctable\"] - save[\"ocr_equals_gt\"][\"correctable\"],\n \"gt_uncorrectable\": save[\"ocr_equals_gt\"][\"uncorrectable\"]\n }\n\n total_cost = 0\n for key in cost.keys():\n total_cost += cost[key] * new[key]\n return total_cost\n\n\ndef final_error(save):\n return save[\"real_word_error\"]/save[\"total\"]\n\ndef tohours(seconds):\n return seconds/3600;\n\ndef to_money_units(hours):\n return human_cost_per_hour * hours\n\n\nfmap = {}\nfmap['hi'] = dict(map(lambda x: x.strip().split('_'), open(\"hi.fmap\")))\nfmap['ml'] = dict(map(lambda x: x.strip().split('_'), open(\"ml.fmap\")))\n\ndef extract_bcode(save):\n bname = save[\"book_dir\"].split('/')[-2]\n return bname\n\ndef get_total_words(save):\n td = save[\"0.0\"][\"unseen\"]\n return td[\"total\"]\n\nprint('Lang,Book,Cost,Error,Word')\nfor lang in ['hi', 'ml']:\n saves = []\n for dr, drs, fls in os.walk('output/%s'%(lang)):\n for fn in fls:\n fn_with_path = dr + '/' + fn\n saves.append(json.load(open(fn_with_path)))\n\n words = list(map(get_total_words, saves))\n\n pages = list(map(lambda x: len(webtotrain.read_book(x[\"book_dir\"])), saves))\n bcodes = list(map(extract_bcode, saves))\n bnames = list(map(lambda bc: fmap[lang][bc], bcodes))\n costs_in_seconds = map(lambda x: analyze(x[\"0.0\"][\"unseen\"]), saves)\n errors = map(lambda x: final_error(x[\"0.0\"][\"unseen\"]), saves)\n #avg_error = sum(errors)/len(list(errors)\n costs_in_hours = map(tohours, costs_in_seconds)\n costs_in_cash = map(to_money_units, costs_in_hours)\n pretty_print = lambda name, cost, error, word, page: '%s,%s,%.2lf,%.2lf,%d,%d'%(lang, name, cost, error, word, page)\n str_ls = map(pretty_print, bnames, costs_in_cash, errors, words, pages)\n print('\\n'.join(str_ls))\n\n\n","sub_path":"src/experiments/cost_analysis.py","file_name":"cost_analysis.py","file_ext":"py","file_size_in_byte":2420,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"116473262","text":"from pylab import *\r\nfigure(figsize=(8,8))\r\nax=subplot(aspect='equal')\r\n\r\n#plot one circle (the biggest one on bottom-right)\r\n#circles(1, 0, 0.5, 'r', alpha=0.2, lw=5, edgecolor='b', transform=ax.transAxes)\r\n\r\n#plot a set of circles (circles in diagonal)\r\n#a=arange(11)\r\n#out = circles(a, a,a*0.2, c=a, edgecolor='k')\r\n#colorbar(out)\r\n\r\na=arange(4)\r\n\r\ncircles(5, 5,1/1, c='r', edgecolor='k')\r\ncircles(5, 5,1/2, c='g', edgecolor='k')\r\ncircles(5, 5,1/3, c='b', edgecolor='k')\r\nxlim(0,10)\r\nylim(0,10)","sub_path":"Mapa/prueba.py","file_name":"prueba.py","file_ext":"py","file_size_in_byte":497,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"621491027","text":"\r\ndef make_heap(arr, length): # 创建堆结构函数 length为需要计算的列表个数, (length-2)//2 为节点个数\r\n for node in range((length-2)//2, -1, -1): # node 为节点下标\r\n arr = max_heap(arr, node, length)\r\n return arr\r\n\r\ndef max_heap(arr,node,length): # 寻找基础节点的最大值函数 # length为需要计算的列表个数\r\n left = 2 * node + 1 # left 左子节点\r\n right = left + 1 # right 右子节点,(右子节点可能不存在,当right=length时,右子节点不存在,)\r\n if arr[left] > arr[node]:\r\n arr[left], arr[node] = arr[node], arr[left]\r\n if right < length and arr[right] > arr[node]:\r\n arr[right], arr[node] = arr[node], arr[right]\r\n return arr\r\n\r\n\r\ndef getresult(arr):\r\n i = len(arr)\r\n while True:\r\n arr = make_heap(arr, i)\r\n arr[0], arr[i-1] = arr[i-1], arr[0]\r\n i -= 1\r\n if i == 1:\r\n break\r\n return arr\r\n\r\n\r\narr = [32, 45, 65, 74, 48, 49, 12, 21, 32, 14, 16, 88, 88]\r\n\r\nprint(arr)\r\nprint(getresult(arr))","sub_path":"sort/demo/heapsort.py","file_name":"heapsort.py","file_ext":"py","file_size_in_byte":1212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"551547098","text":"__author__ = 'kot'\nfrom django.core.management.base import NoArgsCommand\n\n\nclass Command(NoArgsCommand):\n help = 'Update corporation members list'\n\n def handle_noargs(self, **options):\n from corp.models import CorporationMember, Corporation\n from corp.models import CorporationAccount\n from eve.views import get_corp_member_tracking\n\n for corporation in Corporation.objects.all():\n validate = []\n try:\n c_account = corporation.corporation_account.latest('id')\n except CorporationAccount.DoesNotExist:\n continue\n members_tracking = get_corp_member_tracking(\n c_account.key_id,\n c_account.v_code\n )\n for member in members_tracking.members:\n CorporationMember.objects.get_or_create(\n corporation=corporation,\n member_id=member.get('characterID'),\n member_name=member.get('name')\n )\n validate.append(member.get('characterID'))\n invalid = CorporationMember.objects.filter(\n corporation=corporation\n ).exclude(\n member_id__in=validate\n )\n for member in invalid: member.delete()","sub_path":"corp/management/commands/updatemembers.py","file_name":"updatemembers.py","file_ext":"py","file_size_in_byte":1307,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"4192792","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport json\n\nfrom alipay.aop.api.constant.ParamConstants import *\n\n\nclass Matcher(object):\n\n def __init__(self):\n self._identity_card = None\n self._mobile_no = None\n self._open_id = None\n self._user_id = None\n\n @property\n def identity_card(self):\n return self._identity_card\n\n @identity_card.setter\n def identity_card(self, value):\n self._identity_card = value\n @property\n def mobile_no(self):\n return self._mobile_no\n\n @mobile_no.setter\n def mobile_no(self, value):\n self._mobile_no = value\n @property\n def open_id(self):\n return self._open_id\n\n @open_id.setter\n def open_id(self, value):\n self._open_id = value\n @property\n def user_id(self):\n return self._user_id\n\n @user_id.setter\n def user_id(self, value):\n self._user_id = value\n\n\n def to_alipay_dict(self):\n params = dict()\n if self.identity_card:\n if hasattr(self.identity_card, 'to_alipay_dict'):\n params['identity_card'] = self.identity_card.to_alipay_dict()\n else:\n params['identity_card'] = self.identity_card\n if self.mobile_no:\n if hasattr(self.mobile_no, 'to_alipay_dict'):\n params['mobile_no'] = self.mobile_no.to_alipay_dict()\n else:\n params['mobile_no'] = self.mobile_no\n if self.open_id:\n if hasattr(self.open_id, 'to_alipay_dict'):\n params['open_id'] = self.open_id.to_alipay_dict()\n else:\n params['open_id'] = self.open_id\n if self.user_id:\n if hasattr(self.user_id, 'to_alipay_dict'):\n params['user_id'] = self.user_id.to_alipay_dict()\n else:\n params['user_id'] = self.user_id\n return params\n\n @staticmethod\n def from_alipay_dict(d):\n if not d:\n return None\n o = Matcher()\n if 'identity_card' in d:\n o.identity_card = d['identity_card']\n if 'mobile_no' in d:\n o.mobile_no = d['mobile_no']\n if 'open_id' in d:\n o.open_id = d['open_id']\n if 'user_id' in d:\n o.user_id = d['user_id']\n return o\n\n\n","sub_path":"alipay/aop/api/domain/Matcher.py","file_name":"Matcher.py","file_ext":"py","file_size_in_byte":2308,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"507392987","text":"print(\"HANGMAN lETTER GAME\")\nimport random\nwords=['orange','lime','lemon','melon','grape','mango','apple']\nwhile True:\n start=input(\"Press enter/return to start, or enter Q to quit\")\n if start.lower()=='q':\n break\n secret_word=random.choice(words)\n bad_guesses=[]\n good_guesses=[]\n while len(bad_guesses)<7 and len(good_guesses)!= len(list(secret_word)):\n for letter in secret_word:\n if letter in good_guesses:\n print(letter,end=' ')\n else:\n print('_',end=' ')\n print('\\nstrike:{}/7'.format(len(bad_guesses)))\n print('') \n guess=input(\"Guess a letter:\").lower()\n if len(guess)!=1:\n print(\"You can only guess a single letter\")\n continue\n elif guess in bad_guesses or guess in good_guesses:\n print(\"You have already guess that letter\")\n continue\n elif not guess.isalpha():\n print(\"You can only guess letters!\")\n continue\n if guess in secret_word:\n good_guesses.append(guess)\n if len(good_guesses)==len((secret_word)):\n print(\"You win : The word was {}\".format(secret_word))\n break\n else:\n bad_guesses.append(guess)\nprint(\"You didn't guess it : My secret word was {}\".format(secret_word))\n \n \n \n\n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n \n \n \n ","sub_path":"Hangman_Game.py","file_name":"Hangman_Game.py","file_ext":"py","file_size_in_byte":1722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"357909815","text":"# coding: utf-8\nimport argparse\nimport torch\nimport torch.nn as nn\nfrom torch.autograd import Variable\nfrom collections import Counter\nimport torchtext\nfrom torchtext.vocab import Vocab\nfrom diagnostic_classifier import DiagnoseLM as DiagnosticClassifier\nfrom lstm import forward\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--train', type=str, help='location of the training data')\nparser.add_argument('--test', type=str, help='location of the test data')\nparser.add_argument('--model', type=str, help='path to load the model')\nparser.add_argument('--vocab', type=str, help='path to load the model dict')\nparser.add_argument('--save', type=str, help='path to save the model', default='dc.pt')\nparser.add_argument('--bptt', type=int, default=60, help='max sequence length')\nparser.add_argument('--batch_size', type=int, default=24, help='batch size for training')\nparser.add_argument('--eval_batch_size', type=int, default=64, help='eval batch size')\nparser.add_argument('--cuda', action='store_true')\n\n######################################################################\n# Prepare data\n\ntry:\n raw_input\nexcept:\n raw_input = input\n\nargs = parser.parse_args()\nmax_len = 20\ndevice = None if torch.cuda.is_available() else -1\n\ndef len_filter(example):\n return len(example.sentences) <= max_len and len(example.targets) <= max_len\n\ndef tokeniser(text):\n return text.split()\n\ndef tokeniser_targets(targets):\n return [float(target) for target in targets.split()]\n\ndef preprocessing(seq):\n return seq\n\ndef get_vocab(vocab_file):\n vocab = Vocab(Counter())\n words = open(vocab_file, 'rb').read().splitlines()\n pad = vocab.itos[0]\n vocab.itos = []\n for word in words:\n vocab.itos.append(word)\n vocab.itos.append(pad)\n vocab.stoi.update({tok: i for i, tok in enumerate(vocab.itos)})\n return vocab\n\ndef repackage_hidden(h):\n \"\"\"Wraps hidden states in new Variables, to detach them from their history.\"\"\"\n if type(h) == Variable:\n return Variable(h.data)\n else:\n return tuple(repackage_hidden(v) for v in h)\n\nif __name__ == '__main__':\n\n args = parser.parse_args()\n\n # load model\n with open(args.model, 'rb') as f:\n model = torch.load(f, map_location=lambda storage, loc: storage)\n\n # generate datafields\n print(\"Generate training data and vocabulary\")\n sentences = torchtext.data.Field(sequential=True, tokenize=tokeniser, preprocessing=preprocessing, include_lengths=True, use_vocab=True)\n targets = torchtext.data.Field(sequential=True, tokenize=tokeniser_targets, use_vocab=False, include_lengths=True, tensor_type=torch.FloatTensor, pad_token=-1.)\n\n\n # generate vocab and attach to data field\n vocab = get_vocab(args.vocab)\n sentences.vocab = vocab\n pad = vocab.stoi['']\n\n # generate train and test data and vocabulary\n train_data = torchtext.data.TabularDataset(\n path=args.train, format='tsv',\n fields=[('sentences', sentences), ('targets', targets)],\n filter_pred=len_filter\n )\n\n test_data = torchtext.data.TabularDataset(\n path=args.test, format='tsv',\n fields=[('sentences', sentences), ('targets', targets)],\n filter_pred=len_filter\n )\n\n print(\"Create diagnostic model\")\n # create diagnostic classifier\n dc = DiagnosticClassifier(model, n_layer=-1)\n\n dc.add_linear()\n \n print(\"Run training\")\n loss = dc.diagnose(train_data, test_data, n_epochs=30, batch_size=args.batch_size, print_every=600)\n\n dc.save(args.save)\n\n print(loss)\n\n # test_loss = evaluate(model, dictionary, test_data, criterion, args.eval_batch_size)\n # print('=' * 89)\n # print('| End of training | test loss {:5.2f} | test ppl {:8.2f}'.format(\n # test_loss, math.exp(test_loss)))\n # print('=' * 89)\n","sub_path":"diagnose.py","file_name":"diagnose.py","file_ext":"py","file_size_in_byte":3798,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"13863930","text":"from flask import current_app\nfrom flask_login import (login_required, current_user)\nfrom flask import render_template, request\n\nfrom ..models import Note\n\n@current_app.route('/')\n@login_required\ndef home():\n notes = (Note.query\n .filter_by(user=current_user)\n .order_by(_get_order_by(request.args.get('order')))\n .all())\n return render_template('notes/list.html', notes=notes)\n\ndef _get_order_by(param='-updated_at'):\n ''' get model order param by string description '''\n return {\n 'name': Note.name.asc(),\n '-name': Note.name.desc(),\n 'updated_at': Note.updated_at.asc(),\n '-updated_at': Note.updated_at.desc(),\n }.get(param, Note.updated_at.desc())\n","sub_path":"app/views/home.py","file_name":"home.py","file_ext":"py","file_size_in_byte":743,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"355793109","text":"\n\nfrom xai.brain.wordbase.nouns._freshman import _FRESHMAN\n\n#calss header\nclass _FRESHMEN(_FRESHMAN, ):\n\tdef __init__(self,): \n\t\t_FRESHMAN.__init__(self)\n\t\tself.name = \"FRESHMEN\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"freshman\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_freshmen.py","file_name":"_freshmen.py","file_ext":"py","file_size_in_byte":250,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"498619413","text":"from flask import Flask, json\nfrom flask import request\nimport os\n\napi = Flask(__name__)\n\n@api.route('/faucet', methods=['GET'])\ndef get_faucet():\n address = request.args.get('address')\n commandline = 'bitcoin-cli -named sendtoaddress address=\"' + address + '\" amount=1 fee_rate=1'\n os.system(commandline)\n return \"Success\"\n\nif __name__ == '__main__':\n api.run() \n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":371,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"275652199","text":"# Copyright 2015: Mirantis Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport mock\n\nfrom rally.benchmark.scenarios.mistral import utils\nfrom tests.unit import test\n\nMISTRAL_UTILS = \"rally.benchmark.scenarios.mistral.utils\"\n\n\nclass MistralScenarioTestCase(test.TestCase):\n\n def _test_atomic_action_timer(self, atomic_actions, name):\n action_duration = atomic_actions.get(name)\n self.assertIsNotNone(action_duration)\n self.assertIsInstance(action_duration, float)\n\n @mock.patch(MISTRAL_UTILS + \".MistralScenario.clients\")\n def test_list_workbooks(self, mock_clients):\n wbs_list = []\n mock_clients(\"mistral\").workbooks.list.return_value = wbs_list\n scenario = utils.MistralScenario()\n return_wbs_list = scenario._list_workbooks()\n self.assertEqual(wbs_list, return_wbs_list)\n self._test_atomic_action_timer(scenario.atomic_actions(),\n \"mistral.list_workbooks\")\n\n @mock.patch(MISTRAL_UTILS + \".MistralScenario.clients\")\n def test_create_workbook(self, mock_clients):\n definition = \"version: \\\"2.0\\\"\\nname: wb\"\n mock_clients(\"mistral\").workbooks.create.return_value = \"wb\"\n scenario = utils.MistralScenario()\n wb = scenario._create_workbook(definition)\n self.assertEqual(\"wb\", wb)\n self._test_atomic_action_timer(scenario.atomic_actions(),\n \"mistral.create_workbook\")\n\n @mock.patch(MISTRAL_UTILS + \".MistralScenario.clients\")\n def test_delete_workbook(self, mock_clients):\n wb = mock.Mock()\n wb.name = \"wb\"\n mock_clients(\"mistral\").workbooks.delete.return_value = \"ok\"\n scenario = utils.MistralScenario()\n scenario._delete_workbook(wb.name)\n mock_clients(\"mistral\").workbooks.delete.assert_called_once_with(\n wb.name\n )\n self._test_atomic_action_timer(scenario.atomic_actions(),\n \"mistral.delete_workbook\")\n","sub_path":"tests/unit/benchmark/scenarios/mistral/test_utils.py","file_name":"test_utils.py","file_ext":"py","file_size_in_byte":2568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"136830033","text":"\"\"\"\r\nmiddle 2022-03-03 双指针\r\n思路:从中心向���边扩散\r\n[中心扩展法](https://leetcode-cn.com/problems/longest-palindromic-substring/solution/zui-chang-hui-wen-zi-chuan-by-leetcode-solution/)\r\n[动态规划](https://leetcode-cn.com/problems/longest-palindromic-substring/solution/zhong-xin-kuo-san-dong-tai-gui-hua-by-liweiwei1419/)\r\n\"\"\"\r\nclass Solution:\r\n def longestPalindrome(self, s: str) -> str:\r\n if not s or len(s)==1:return s\r\n start,end=0,0\r\n for i in range(len(s)):\r\n # 由于存在奇数的字符串和偶数的字符串,所以我们需要从一个字符开始扩展,或者从两个字符之间开始扩展,所以总共有 n+n-1 个中心。\r\n left1,right1 = self.expand(s,i,i) # a\r\n left2,right2 = self.expand(s,i,i+1) # aa\r\n # print(left1,right1,left2,right2)\r\n # right1-left1+1 是以i为中心点的最长回文子串\r\n if right1-left1>end-start:\r\n start,end=left1,right1\r\n if right2-left2>end-start:\r\n start,end=left2,right2\r\n return s[start:end+1]\r\n\r\n def expand(self, s, left, right):\r\n while left>=0 and rightmaxlen:\r\n maxlen = j-i+1\r\n start = i\r\n return s[start:start+maxlen]\r\n\r\n\r\nif __name__ == '__main__':\r\n # s = \"babab\"\r\n # 输出:\"bab\"\r\n # 解释:\"aba\" 同样是符合题意的答案。\r\n s = \"cbbd\"\r\n print(Solution().longestPalindrome(s))\r\n print(Solution().dp(s))","sub_path":"02_双指针/5-最长回文子串.py","file_name":"5-最长回文子串.py","file_ext":"py","file_size_in_byte":3389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"350320919","text":"import pandas as pd\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.cluster import KMeans\r\nfrom sklearn.cluster import AgglomerativeClustering\r\nfrom sklearn.preprocessing import StandardScaler\r\nfrom sklearn.decomposition import PCA\r\nfrom sklearn import metrics\r\n\r\n\r\ndef clustering(X, method=\"kmeans\", filt=\"\"):\r\n max_nmi = 0\r\n k_optimal_nmi = 0\r\n x_nmi = []\r\n y_nmi = []\r\n for i in range(2, 20):\r\n if method==\"kmeans\":\r\n labels = KMeans(n_clusters=i, random_state=0).fit(X).labels_\r\n color_plot = \"red\"\r\n elif method==\"agglomerative_complete\":\r\n labels = AgglomerativeClustering(n_clusters=i, linkage=\"complete\").fit(X).labels_\r\n color_plot = \"blue\"\r\n elif method==\"agglomerative_average\":\r\n labels = AgglomerativeClustering(n_clusters=i, linkage=\"average\").fit(X).labels_\r\n color_plot = \"cyan\"\r\n elif method==\"agglomerative_single\":\r\n labels = AgglomerativeClustering(n_clusters=i, linkage=\"single\").fit(X).labels_\r\n color_plot = \"green\"\r\n elif method==\"agglomerative_ward\":\r\n labels = AgglomerativeClustering(n_clusters=i, linkage=\"ward\").fit(X).labels_\r\n color_plot = \"magenta\"\r\n\r\n nmi = metrics.normalized_mutual_info_score(class_labels, labels, average_method=\"geometric\")\r\n x_nmi.append(i)\r\n y_nmi.append(nmi)\r\n if nmi > max_nmi:\r\n max_nmi = nmi\r\n k_optimal_nmi = i\r\n labels_optimal = labels\r\n plt.clf()\r\n plt.title(method)\r\n plt.xlabel(\"num clusters\")\r\n plt.ylabel(\"nmi\")\r\n plt.plot(x_nmi, y_nmi, color_plot)\r\n plt.savefig(\"../plots/\" + filt + method + \"_nmi_clusters_msdata.png\")\r\n #plt.show()\r\n\r\n return(k_optimal_nmi, max_nmi, labels_optimal)\r\n\r\ndef compute_PCA(data_set, num_components=2):\r\n x = data_set.values\r\n x = StandardScaler().fit_transform(x)\r\n pca = PCA(n_components=num_components)\r\n principalComponents = pca.fit_transform(x)\r\n\r\n return(pd.DataFrame(principalComponents))\r\n\r\n\r\n\r\n# Data contains the initial dataset, with classes and features\r\ndata = pd.read_csv(\"../data/msdata.csv\")\r\n\r\n\r\n# Extract class labels\r\nclass_labels = data['class']\r\nclass_labels = np.array(class_labels)\r\ndata = data.drop(columns = ['id', 'class'])\r\n\r\n\r\n\r\nX = data\r\nfilt = \"\"\r\n#######################################################################\r\n# Comment the following lines to run the program without performing PCA\r\nX = compute_PCA(X)\r\nfilt = \"filtered_\"\r\n#######################################################################\r\nX = np.array(X)\r\n\r\n\r\n\r\n#CLUSTERIZATION\r\n#k-means\r\nnum_clusters, nmi_kmeans, labels_kmeans = clustering(X, \"kmeans\", filt)\r\nprint(\"NMI kmeans:\", nmi_kmeans)\r\nprint(\"Num_clusters:\", num_clusters)\r\nprint()\r\n\r\n#Agglomerative complete\r\nnum_clusters, nmi_complete, labels_complete = clustering(X, \"agglomerative_complete\", filt)\r\nprint(\"NMI agg complete:\", nmi_complete)\r\nprint(\"Num_clusters:\", num_clusters)\r\nprint()\r\n\r\n#Agglomerative average\r\nnum_clusters, nmi_avg, labels_avg = clustering(X, \"agglomerative_average\", filt)\r\nprint(\"NMI agg avg:\", nmi_avg)\r\nprint(\"Num_clusters:\", num_clusters)\r\nprint()\r\n\r\n#Agglomerative single\r\nnum_clusters, nmi_single, labels_single = clustering(X, \"agglomerative_single\", filt)\r\nprint(\"NMI agg single:\", nmi_single)\r\nprint(\"Num_clusters:\", num_clusters)\r\nprint()\r\n\r\n#Agglomerative ward\r\nnum_clusters, nmi_ward, labels_ward = clustering(X, \"agglomerative_ward\", filt)\r\nprint(\"NMI agg ward:\", nmi_ward)\r\nprint(\"Num_clusters:\", num_clusters)\r\n\r\n#Create and save txt with the id and the new labels\r\nnmi_array = np.array([nmi_kmeans, nmi_complete, nmi_avg, nmi_single, nmi_ward])\r\nmax_idx = np.argmax(nmi_array)\r\n\r\nif max_idx==0:\r\n labels_optimal = labels_kmeans\r\nelif max_idx==1:\r\n labels_optimal = labels_complete\r\nelif max_idx==2:\r\n labels_optimal = labels_avg\r\nelif max_idx==3:\r\n labels_optimal = labels_single\r\nelif max_idx==4:\r\n labels_optimal = labels_ward\r\n\r\nlabels_optimal = labels_optimal.astype(int)\r\nnp.savetxt(\"../results/output_ms_data.txt\", labels_optimal, fmt=\"%d\", delimiter=\",\")\r\n","sub_path":"src/clustering_msdata.py","file_name":"clustering_msdata.py","file_ext":"py","file_size_in_byte":4156,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"589650011","text":"import requests\ndef translate(str):\n url = 'https://fanyi.baidu.com/#en/zh/'+ str\n try:\n response = requests.get(url)\n response.raise_for_status()\n response.encoding = response.apparent_encoding\n html = response.text\n print(html)\n except:\n print(\"连接错误\")\ndef weiboImg():\n\n print( requests.get('//img.t.sinajs.cn/t4/appstyle/expression/ext/normal/a1/2018new_doge02_org.png').content)\n\ndef new():\n res = requests.get('http://dl.sina.com.cn/news/2019-03-06/detail-ihrfqzkc1551167.shtml')\n res.encoding = res.apparent_encoding\n html = res.text\n print(html)\nnew()","sub_path":"百度翻译.py","file_name":"百度翻译.py","file_ext":"py","file_size_in_byte":631,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"607791982","text":"# -*- coding: utf-8 -*-\n# Part of Odoo. See LICENSE file for full copyright and licensing details.\n\nfrom datetime import datetime\n\nfrom odoo import models, fields, api, exceptions, _\nfrom odoo.tools import DEFAULT_SERVER_DATETIME_FORMAT\n\n\nclass EmpBreakTime(models.Model):\n _name = \"emp.breaktime\"\n _description = \"Breaktime\"\n _order = \"check_in desc\"\n\n def _default_employee(self):\n return self.env['emp.employee'].search([('user_id', '=', self.env.uid)], limit=1)\n\n employee_id = fields.Many2one('hr.employee', string=\"Employee\", default=_default_employee, required=True, ondelete='cascade', index=True)\n department_id = fields.Many2one('hr.department', string=\"Department\", related=\"employee_id.department_id\")\n check_in = fields.Datetime(string=\"Check In\", default=fields.Datetime.now, required=True)\n check_out = fields.Datetime(string=\"Check Out\")\n worked_hours = fields.Float(string='Break Hours', compute='_compute_worked_hours', store=True, readonly=True)\n\n @api.multi\n def name_get(self):\n result = []\n for breaktime in self:\n if not breaktime.check_out:\n result.append((breaktime.id, _(\"%(empl_name)s from %(check_in)s\") % {\n 'empl_name': breaktime.employee_id.name_related,\n 'check_in': fields.Datetime.to_string(fields.Datetime.context_timestamp(breaktime, fields.Datetime.from_string(breaktime.check_in))),\n }))\n else:\n result.append((breaktime.id, _(\"%(empl_name)s from %(check_in)s to %(check_out)s\") % {\n 'empl_name': breaktime.employee_id.name_related,\n 'check_in': fields.Datetime.to_string(fields.Datetime.context_timestamp(breaktime, fields.Datetime.from_string(breaktime.check_in))),\n 'check_out': fields.Datetime.to_string(fields.Datetime.context_timestamp(breaktime, fields.Datetime.from_string(breaktime.check_out))),\n }))\n return result\n\n @api.depends('check_in', 'check_out')\n def _compute_worked_hours(self):\n for breaktime in self:\n if breaktime.check_out:\n delta = datetime.strptime(breaktime.check_out, DEFAULT_SERVER_DATETIME_FORMAT) - datetime.strptime(\n breaktime.check_in, DEFAULT_SERVER_DATETIME_FORMAT)\n breaktime.worked_hours = delta.total_seconds() / 3600.0\n\n @api.constrains('check_in', 'check_out')\n def _check_validity_check_in_check_out(self):\n \"\"\" verifies if check_in is earlier than check_out. \"\"\"\n for breaktime in self:\n if breaktime.check_in and breaktime.check_out:\n if breaktime.check_out < breaktime.check_in:\n raise exceptions.ValidationError(_('\"Check Out\" time cannot be earlier than \"Check In\" time.'))\n\n @api.constrains('check_in', 'check_out', 'employee_id')\n def _check_validity(self):\n \"\"\" Verifies the validity of the breaktime record compared to the others from the same employee.\n For the same employee we must have :\n * maximum 1 \"open\" breaktime record (without check_out)\n * no overlapping time slices with previous employee records\n \"\"\"\n for breaktime in self:\n # we take the latest breaktime before our check_in time and check it doesn't overlap with ours\n last_breaktime_before_check_in = self.env['emp.breaktime'].search([\n ('employee_id', '=', breaktime.employee_id.id),\n ('check_in', '<=', breaktime.check_in),\n ('id', '!=', breaktime.id),\n ], order='check_in desc', limit=1)\n if last_breaktime_before_check_in and last_breaktime_before_check_in.check_out and last_breaktime_before_check_in.check_out > breaktime.check_in:\n raise exceptions.ValidationError(_(\"Cannot create new breaktime record for %(empl_name)s, the employee was already checked in on %(datetime)s\") % {\n 'empl_name': breaktime.employee_id.name_related,\n 'datetime': fields.Datetime.to_string(fields.Datetime.context_timestamp(self, fields.Datetime.from_string(breaktime.check_in))),\n })\n\n if not breaktime.check_out:\n # if our breaktime is \"open\" (no check_out), we verify there is no other \"open\" breaktime\n no_check_out_breaktimes = self.env['emp.breaktime'].search([\n ('employee_id', '=', breaktime.employee_id.id),\n ('check_out', '=', False),\n ('id', '!=', breaktime.id),\n ])\n if no_check_out_breaktimes:\n raise exceptions.ValidationError(_(\"Cannot create new breaktime record for %(empl_name)s, the employee hasn't checked out since %(datetime)s\") % {\n 'empl_name': breaktime.employee_id.name_related,\n 'datetime': fields.Datetime.to_string(fields.Datetime.context_timestamp(self, fields.Datetime.from_string(no_check_out_breaktimes.check_in))),\n })\n else:\n # we verify that the latest breaktime with check_in time before our check_out time\n # is the same as the one before our check_in time computed before, otherwise it overlaps\n last_breaktime_before_check_out = self.env['emp.breaktime'].search([\n ('employee_id', '=', breaktime.employee_id.id),\n ('check_in', '<', breaktime.check_out),\n ('id', '!=', breaktime.id),\n ], order='check_in desc', limit=1)\n if last_breaktime_before_check_out and last_breaktime_before_check_in != last_breaktime_before_check_out:\n raise exceptions.ValidationError(_(\"Cannot create new breaktime record for %(empl_name)s, the employee was already checked in on %(datetime)s\") % {\n 'empl_name': breaktime.employee_id.name_related,\n 'datetime': fields.Datetime.to_string(fields.Datetime.context_timestamp(self, fields.Datetime.from_string(last_breaktime_before_check_out.check_in))),\n })\n\n @api.multi\n def copy(self):\n raise exceptions.UserError(_('You cannot duplicate any Breaktime.'))\n","sub_path":"beta-dev1/employee_breaktime/models/employee_breaktime.py","file_name":"employee_breaktime.py","file_ext":"py","file_size_in_byte":6317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"77238395","text":"from game import Game\nfrom corrupted_chess_file_error import *\nfrom player import Player\nfrom piece import Piece\nfrom board import Board\n\n\n\nclass HumanWriteableIO(object):\n\n def load_game(self, input):\n\n '''\n This is the game object this method will fill with data. The object\n is returned when the file ends and everything is ok.\n '''\n\n self.game = Game()\n board = Board()\n \n self.game.set_board(board)\n \n info_read = False\n white_read = False\n black_read = False\n\n \n #Use this variable for reading all the section headers.\n current_line = ''\n\n try:\n\n # Read the file header and the save date\n\n current_line = input.readline()\n header_parts = current_line.split(\" \")\n\n # Process the data we just read.\n # NOTE: To test the line below you must test the class once with a\n # broken header\n\n if header_parts[0] != \"SHAKKI\":\n raise CorruptedChessFileError(\"Unknown file type\")\n\n if header_parts[2].strip().lower() != 'tallennustiedosto':\n raise CorruptedChessFileError(\"Unknown file type\")\n \n\n # The version information and the date are not used in this\n # exercise\n\n # *************************************************************\n current_line = input.readline()\n block = []\n \n while input:\n block = []\n \n #break the while-loop when the file is read to the end\n if current_line == '':\n break\n \n #if current line is includes start of information block indicator #, save block name and read next line\n elif current_line[0] == '#':\n block_name = current_line[1:].strip().lower()\n current_line = input.readline()\n \n #save all lines until end of file or until end of block into a list\n while current_line != '' and current_line[0] != '#':\n block.append(current_line)\n current_line = input.readline()\n \n #implement player data from \"pelin tiedot\" block\n if block_name == 'pelin tiedot':\n players = self.read_info(block)\n for i in range(len(players)):\n self.game.add_player(players[i])\n info_read = True\n \n #list each player's chess pieces and set on board. Available only after \"pelin tiedot\" block has been read. \n elif block_name == 'musta' or block_name == 'valkoinen' and info_read == True:\n #creates pieces for each player\n if block_name == 'musta':\n pieces = self.read_pieces(block, self.game.black)\n else:\n pieces = self.read_pieces(block, self.game.white)\n \n #sets pieces on board\n for i in range(len(pieces)):\n try:\n self.game.board.set_piece(pieces[i][1], self.game.board.column_char_to_integer(pieces[i][0][0]), self.game.board.row_char_to_integer(pieces[i][0][1])) \n except:\n raise CorruptedChessFileError(\"Invalid coordinates for a piece\")\n \n if block_name =='musta':\n black_read = True\n else:\n white_read = True \n \n del block \n \n else:\n current_line = input.readline()\n \n \n if info_read != True or black_read != True or white_read != True:\n raise CorruptedChessFileError(\"Black, White or Info-block missing\")\n\n self.piece_types = {'Kuningas':Piece.KING, 'Kuningatar':Piece.QUEEN, 'Torni':Piece.ROOK, 'Lahetti':Piece.BISHOP, 'Ratsu':Piece.KNIGHT, 'Sotilas':Piece.PAWN}\n\n \n \n # If we reach this point the Game-object should now have the proper\n # players and\n # a fully set up chess board. Therefore we might as well return it.\n \n return self.game\n \n except IOError:\n\n # To test this part the stream would have to cause an\n # IOError. That's a bit complicated to test. Therefore we have\n # given you a \"secret tool\", class BrokenReader, which will throw\n # an IOError at a requested position in the stream.\n # Throw the exception inside any chunk, but not in the chunk\n # header.\n\n raise CorruptedChessFileError(\"Reading the chess data failed.\")\n \n def read_info(self, block):\n players = []\n \n for i in range(len(block)):\n line_parts = block[i].split(':')\n if line_parts[0].strip().lower() == 'valkoinen':\n name = line_parts[1].strip()\n players.append(Player(name, Player.WHITE))\n elif line_parts[0].strip().lower() == 'musta':\n name = line_parts[1].strip()\n players.append(Player(name, Player.BLACK))\n \n if len(players) != 2:\n raise CorruptedChessFileError('Odd number of players')\n \n return players\n \n \n \n def read_pieces(self, block, player):\n pieces_dic = {'kuningas':Piece.KING, 'kuningatar':Piece.QUEEN, 'torni':Piece.ROOK, 'lahetti':Piece.BISHOP, 'ratsu':Piece.KNIGHT, 'sotilas':Piece.PAWN} \n \n pieces = [] #list of all pieces\n \n for i in range(len(block)):\n piece = [] #information of 1)coordinates and 2)type\n \n line_parts = block[i].split(':')\n if len(line_parts) > 1:\n try:\n piece.append(line_parts[1].strip()) #coordinates\n piece.append(Piece(player, pieces_dic[line_parts[0].strip().lower()])) #type\n pieces.append(piece)\n except:\n raise CorruptedChessFileError(\"Invalid \")\n \n del line_parts \n return pieces\n\n","sub_path":"human_writeable_IO.py","file_name":"human_writeable_IO.py","file_ext":"py","file_size_in_byte":6647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"78502256","text":"import tensorflow as tf\nfrom imputation.utils import xavier_init\n\n\ndef set_pre_encoder(theta_pre_Enc, init_weights, trainable=False):\n for idx in range(0, 12, 2):\n theta_pre_Enc.append([tf.Variable(initial_value=init_weights[idx], name='Encoder_W' + str(idx//2), trainable=trainable),\n tf.Variable(initial_value=init_weights[idx + 1], name='Encoder_B' + str(idx//2), trainable=trainable)])\n return None\n\ndef set_encoder(theta_Enc, init_weights, trainable=True):\n for idx in range(0, 12, 2):\n theta_Enc.append([tf.Variable(initial_value=init_weights[idx], name='Encoder_W' + str(idx//2), trainable=trainable),\n tf.Variable(initial_value=init_weights[idx + 1], name='Encoder_B' + str(idx//2), trainable=trainable)])\n return None\n\n\ndef set_dec(theta_Dec, init_weights, numeric_dim, categorical_cols, trainable):\n cat_start_idx = 0\n # decoder weights for numeric data (non-trainable)\n if numeric_dim > 0:\n for idx in range(0, 16, 2):\n theta_Dec.append([tf.Variable(initial_value=init_weights[idx], name='Decoder_Num_W' + str(idx//2), trainable=trainable),\n tf.Variable(initial_value=init_weights[idx+1], name='Decoder_Num_B' + str(idx//2), trainable=trainable)])\n cat_start_idx = 16\n # decoder weights for categorical data (non-trainable)\n for idx in range(len(categorical_cols)):\n for in_idx in range(0, 16, 2):\n theta_Dec.append([tf.Variable(initial_value=init_weights[cat_start_idx + in_idx + 16 * idx],\n name='Decoder_' + categorical_cols[idx] + '_W' + str(in_idx//2), trainable=trainable),\n tf.Variable(initial_value=init_weights[cat_start_idx + (in_idx + 1) + 16 * idx],\n name='Decoder_' + categorical_cols[idx] + '_B' + str(in_idx//2), trainable=trainable)])\n return None\n\ndef set_disc(theta_Disc, dim):\n # Discriminator weights\n l_dim = int(dim * 0.7)\n l1_dim = int(l_dim * 0.7)\n l2_dim = int(l_dim * 0.3)\n\n theta_Disc.append([tf.Variable(xavier_init([l_dim, l1_dim]), name='Discriminator_W0'),\n tf.Variable(tf.zeros([l1_dim]), name='Discriminator_b0')])\n theta_Disc.append([tf.Variable(xavier_init([l1_dim, l1_dim]), name='Discriminator_W1'),\n tf.Variable(tf.zeros([l1_dim]), name='Discriminator_b1')])\n theta_Disc.append([tf.Variable(xavier_init([l1_dim, l2_dim]), name='Discriminator_W2'),\n tf.Variable(tf.zeros([l2_dim]), name='Discriminator_b2')])\n theta_Disc.append([tf.Variable(xavier_init([l2_dim, l2_dim]), name='Discriminator_W3'),\n tf.Variable(tf.zeros([l2_dim]), name='Discriminator_b3')])\n theta_Disc.append([tf.Variable(xavier_init([l2_dim, 1]), name='Discriminator_W4'),\n tf.Variable(tf.zeros([1]), name='Discriminator_b4')])\n return None\n\ndef pre_encoder(data, theta_pre_Enc):\n # Auto-encoder (Generator) structure\n tmp_latent0 = tf.nn.tanh(tf.matmul(data, theta_pre_Enc[0][0]) + theta_pre_Enc[0][1])\n tmp_latent1 = tf.nn.tanh(tf.matmul(tmp_latent0, theta_pre_Enc[1][0]) + theta_pre_Enc[1][1])\n tmp_latent2 = tf.nn.tanh(tf.matmul(tmp_latent1, theta_pre_Enc[2][0]) + theta_pre_Enc[2][1])\n tmp_latent3 = tf.nn.tanh(tf.matmul(tmp_latent2, theta_pre_Enc[3][0]) + theta_pre_Enc[3][1])\n tmp_latent4 = tf.nn.tanh(tf.matmul(tmp_latent3, theta_pre_Enc[4][0]) + theta_pre_Enc[4][1])\n latent = tf.matmul(tmp_latent4, theta_pre_Enc[5][0]) + theta_pre_Enc[5][1]\n return latent\n\ndef encoder(data, theta_Enc):\n # Auto-encoder (Generator) structure\n tmp_latent0 = tf.nn.tanh(tf.matmul(data, theta_Enc[0][0]) + theta_Enc[0][1])\n tmp_latent1 = tf.nn.tanh(tf.matmul(tmp_latent0, theta_Enc[1][0]) + theta_Enc[1][1])\n tmp_latent2 = tf.nn.tanh(tf.matmul(tmp_latent1, theta_Enc[2][0]) + theta_Enc[2][1])\n tmp_latent3 = tf.nn.tanh(tf.matmul(tmp_latent2, theta_Enc[3][0]) + theta_Enc[3][1])\n tmp_latent4 = tf.nn.tanh(tf.matmul(tmp_latent3, theta_Enc[4][0]) + theta_Enc[4][1])\n latent = tf.matmul(tmp_latent4, theta_Enc[5][0]) + theta_Enc[5][1]\n return latent\n\ndef generator(data, theta_Enc, theta_Dec, numeric_dim):\n # Auto-encoder (Generator) structure\n tmp_latent0 = tf.nn.tanh(tf.matmul(data, theta_Enc[0][0]) + theta_Enc[0][1])\n tmp_latent1 = tf.nn.tanh(tf.matmul(tmp_latent0, theta_Enc[1][0]) + theta_Enc[1][1])\n tmp_latent2 = tf.nn.tanh(tf.matmul(tmp_latent1, theta_Enc[2][0]) + theta_Enc[2][1])\n tmp_latent3 = tf.nn.tanh(tf.matmul(tmp_latent2, theta_Enc[3][0]) + theta_Enc[3][1])\n tmp_latent4 = tf.nn.tanh(tf.matmul(tmp_latent3, theta_Enc[4][0]) + theta_Enc[4][1])\n latent = tf.matmul(tmp_latent4, theta_Enc[5][0]) + theta_Enc[5][1]\n # Decoder structure\n dec = []\n cat_start_idx = 0\n if numeric_dim > 0:\n # add layer for numerics\n numeric_latent0 = tf.nn.tanh(tf.matmul(latent, theta_Dec[0][0]) + theta_Dec[0][1])\n numeric_latent1 = tf.nn.tanh(tf.matmul(numeric_latent0, theta_Dec[1][0]) + theta_Dec[1][1])\n numeric_latent2 = tf.nn.tanh(tf.matmul(numeric_latent1, theta_Dec[2][0]) + theta_Dec[2][1])\n numeric_latent3 = tf.nn.tanh(tf.matmul(numeric_latent2, theta_Dec[3][0]) + theta_Dec[3][1])\n numeric_latent4 = tf.nn.tanh(tf.matmul(numeric_latent3, theta_Dec[4][0]) + theta_Dec[4][1])\n numeric_latent5 = tf.nn.tanh(tf.matmul(numeric_latent4, theta_Dec[5][0]) + theta_Dec[5][1])\n numeric_latent6 = tf.nn.tanh(tf.matmul(numeric_latent5, theta_Dec[6][0]) + theta_Dec[6][1])\n dec.append(tf.matmul(numeric_latent6, theta_Dec[7][0]) + theta_Dec[7][1])\n cat_start_idx = 8\n # add layer for categorical\n for i in range(cat_start_idx, len(theta_Dec), 8):\n dec_latent0 = tf.nn.tanh(tf.matmul(latent, theta_Dec[i][0]) + theta_Dec[i][1])\n dec_latent1 = tf.nn.tanh(tf.matmul(dec_latent0, theta_Dec[i + 1][0]) + theta_Dec[i + 1][1])\n dec_latent2 = tf.nn.tanh(tf.matmul(dec_latent1, theta_Dec[i + 2][0]) + theta_Dec[i + 2][1])\n dec_latent3 = tf.nn.tanh(tf.matmul(dec_latent2, theta_Dec[i + 3][0]) + theta_Dec[i + 3][1])\n dec_latent4 = tf.nn.tanh(tf.matmul(dec_latent3, theta_Dec[i + 4][0]) + theta_Dec[i + 4][1])\n dec_latent5 = tf.nn.tanh(tf.matmul(dec_latent4, theta_Dec[i + 5][0]) + theta_Dec[i + 5][1])\n dec_latent = tf.nn.tanh(tf.matmul(dec_latent5, theta_Dec[i + 6][0]) + theta_Dec[i + 6][1])\n if theta_Dec[i + 7][0].shape[1] > 1:\n dec.append(tf.nn.softmax(tf.matmul(dec_latent, theta_Dec[i + 7][0]) + theta_Dec[i + 7][1]))\n else:\n dec.append(tf.nn.sigmoid(tf.matmul(dec_latent, theta_Dec[i + 7][0]) + theta_Dec[i + 7][1]))\n # Output\n outputs = tf.concat(values=dec, axis=1)\n return outputs, dec\n\n\ndef discriminator(data, theta_Disc):\n # imputed_data: outputs (of generator function)\n # Discriminator structure\n Disc_h0 = tf.nn.tanh(tf.matmul(data, theta_Disc[0][0]) + theta_Disc[0][1])\n Disc_h1 = tf.nn.tanh(tf.matmul(Disc_h0, theta_Disc[1][0]) + theta_Disc[1][1])\n Disc_h2 = tf.nn.tanh(tf.matmul(Disc_h1, theta_Disc[2][0]) + theta_Disc[2][1])\n Disc_h3 = tf.nn.tanh(tf.matmul(Disc_h2, theta_Disc[3][0]) + theta_Disc[3][1])\n Disc_prob = tf.nn.sigmoid(tf.matmul(Disc_h3, theta_Disc[4][0]) + theta_Disc[4][1])\n return Disc_prob","sub_path":"imputation/utils_imp.py","file_name":"utils_imp.py","file_ext":"py","file_size_in_byte":7413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"41881416","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n\nclass Solution:\n def findTilt(self, root):\n \"\"\"\n :type root: TreeNode\n :rtype: int\n \"\"\"\n self.result = 0\n def helper(root): #used to calculate the summation and update the self.result\n if not root:\n return 0\n leftSum = helper(root.left)\n rightSum = helper(root.right)\n self.result += abs(leftSum - rightSum)\n return leftSum + rightSum + root.val\n helper(root)\n return self.result\n","sub_path":"TangYuCreated/leetcode/Python/简单题目/563.py","file_name":"563.py","file_ext":"py","file_size_in_byte":673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"404341445","text":"a=open('digits-large-in.txt','r')\r\nb=a.readlines()\r\nfor i in range(len(b)):\r\n\tb[i]=b[i].rstrip('\\n')\r\na.close()\r\n\r\nc=open('digits-large-out.txt','w')\r\n\r\ndef checkNumTimes(key,tempDict):\r\n\tif key in tempDict.keys():\r\n\t\treturn tempDict[key]\r\n\treturn 0\r\n\r\n\r\nfor i in range(1,len(b)):\r\n\tnumList=[]\r\n\tfreq={}\r\n\tfor letter in b[i]:\r\n\t\tif letter in freq.keys():\r\n\t\t\tfreq[letter]+=1\r\n\t\telse:\r\n\t\t\tfreq[letter]=1\r\n\tfor j in range(checkNumTimes(\"Z\",freq)): #ZERO\r\n\t\tnumList.append(0)\r\n\t\tfreq[\"E\"]-=1\r\n\t\tfreq[\"R\"]-=1\r\n\t\tfreq[\"O\"]-=1\r\n\tfor j in range(checkNumTimes(\"W\",freq)): #TWO\r\n\t\tnumList.append(2)\r\n\t\tfreq['T']-=1\r\n\t\tfreq['O']-=1\r\n\tfor j in range(checkNumTimes(\"U\",freq)): #FOUR\r\n\t\tnumList.append(4)\r\n\t\tfreq['F']-=1\r\n\t\tfreq['O']-=1\r\n\t\tfreq['R']-=1\r\n\tfor j in range(checkNumTimes(\"F\",freq)): #FIVE\r\n\t\tnumList.append(5)\r\n\t\tfreq[\"I\"]-=1\r\n\t\tfreq[\"V\"]-=1\r\n\t\tfreq[\"E\"]-=1\r\n\tfor j in range(checkNumTimes(\"X\",freq)): #SIX\r\n\t\tnumList.append(6)\r\n\t\tfreq['S']-=1\r\n\t\tfreq['I']-=1\r\n\tfor j in range(checkNumTimes(\"V\",freq)): #SEVEN\r\n\t\tnumList.append(7)\r\n\t\tfreq['S']-=1\r\n\t\tfreq['E']-=2\r\n\t\tfreq['N']-=1\r\n\tfor j in range(checkNumTimes(\"G\",freq)): #EIGHT\r\n\t\tnumList.append(8)\r\n\t\tfreq['E']-=1\r\n\t\tfreq['I']-=1\r\n\t\tfreq['H']-=1\r\n\t\tfreq['T']-=1\r\n\tfor j in range(checkNumTimes(\"I\",freq)): #NINE\r\n\t\tnumList.append(9)\r\n\t\tfreq['N']-=2\r\n\t\tfreq['E']-=1\r\n\tfor j in range(checkNumTimes(\"H\",freq)): #THREE\r\n\t\tnumList.append(3)\r\n\t\tfreq['T']-=1\r\n\t\tfreq['R']-=1\r\n\t\tfreq['E']-=2\r\n\tfor j in range(checkNumTimes(\"O\",freq)): #ONE\r\n\t\tnumList.append(1)\r\n\t\tfreq['N']-=1\r\n\t\tfreq['E']-=1\r\n\tnumList.sort()\r\n\tresult=\"\"\r\n\tfor num in numList:\r\n\t\tresult+=str(num)\r\n\tc.write(\"Case #\"+str(i)+\": \"+result+\"\\n\")\r\nc.close()\r\n","sub_path":"codes/BuildLinks1.10/test_input/CJ_16_2/16_2_1_plin_digits.py","file_name":"16_2_1_plin_digits.py","file_ext":"py","file_size_in_byte":1662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"552184639","text":"#%% Classifications using Premade Estimators\r\nimport pandas as pd\r\nimport numpy as np\r\nimport tensorflow as tf\r\nimport os\r\nfrom sklearn.model_selection import train_test_split\r\nfrom pandas_ml import ConfusionMatrix\r\nimport matplotlib.pyplot as plt\r\nimport itertools\r\nimport gc; gc.enable()\r\n\r\nos.chdir(\"D:\\\\trainings\\\\tensorflow\")\r\nexec(open(os.path.abspath('tf_CommonUtils.py')).read())\r\ntf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.INFO)\r\n\r\n# fix random seed for reproducibility\r\nseed = 123; np.random.seed(seed); tf.compat.v1.set_random_seed(seed)\r\n\r\n# identifications of features and response\r\nFEATURES = [\"SepalLength\",\"SepalWidth\",\"PetalLength\",\"PetalWidth\"]\r\nLABEL = \"Species\"\r\nbatch_size = 8\r\n\r\ndata = pd.read_csv(\"./data/iris.csv\")\r\ndata.shape\r\ndata.dtypes\r\ndata.head(2)\r\ndata.info()\r\nprint(data.describe()) # .unstack()\r\n#print(data.describe(include = [np.number])) # for number only\r\n\r\n# Labels dtype should be integer\r\ndata[LABEL].unique() #.astype(int)\r\nnum_mapping = {\"setosa\":0 ,\"versicolor\" :1,\"virginica\":2}\r\ndata[LABEL] = data[LABEL].replace(num_mapping)\r\n\r\n# now convert the types\r\ndata[LABEL] = pd.to_numeric(data[LABEL], errors='coerce')\r\ndata.dtypes\r\n\r\n#Segragate 85% and 15%\r\ntraining_set ,test_set = train_test_split(data,test_size=0.15)\r\ndel(data)\r\n\r\n#An input function returns a tf.data.Dataset object which contains features (dictionary -\r\n#with key (feature name) and value (feature's values)\r\n#label - An array containing the values of the label for every row.\r\n\r\n#Building the input_fn: regressor accepts Tensors and custom function to convert pandas\r\n#Dataframe and return feature column and label values as Tensors:\r\ndef input_fn(features, labels = None, custom_batch_size = batch_size, caller_source = 'train'):\r\n # Convert the inputs to a Dataset.\r\n dataset = tf.data.Dataset.from_tensor_slices(dict(features))\r\n \r\n if caller_source != 'test':\r\n dataset = tf.data.Dataset.from_tensor_slices((dict(features), labels))\r\n \r\n if caller_source == 'train': \r\n dataset = dataset.shuffle(len(features)).repeat()\r\n \r\n dataset = dataset.batch(custom_batch_size)\r\n\r\n return dataset\r\n\r\n#Defining FeatureColumns and Creating the classifier\r\n#All features in data set contain continuous values, hence create their FeatureColumns\r\nfeature_cols = [tf.feature_column.numeric_column(k) for k in FEATURES]\r\n\r\n# Build 3 layer DNN with 10, 20, 10 units respectively.\r\nclassifier = tf.estimator.DNNClassifier(feature_columns=feature_cols, hidden_units=[30,10], n_classes=3)\r\n\r\n# Fit model. Note: If error comes then clean folder 'model_dir' and restart Spyder\r\nclassifier.train(input_fn=lambda: input_fn(training_set[FEATURES], training_set[LABEL],custom_batch_size = batch_size), steps=5000)\r\n\r\n#Evaluating the Model. Note: If error comes then clean folder 'model_dir' and restart Spyder\r\nev = classifier.evaluate(input_fn=lambda: input_fn(test_set[FEATURES], test_set[LABEL],custom_batch_size = batch_size, caller_source = 'eval'))\r\nprint(\"\\nTest Accuracy: {0:f}\\n\".format(ev[\"accuracy\"])) # 78%\r\n\r\n#Making Predictions\r\npredictions = classifier.predict(input_fn=lambda: input_fn(test_set[FEATURES], None, custom_batch_size = batch_size, caller_source = 'test'))\r\n\r\n# .predict() returns an iterator; convert to a list and print predictions\r\npredictions = list(pred_tensor[\"class_ids\"][0] for pred_tensor in itertools.islice(predictions, test_set.shape[0]))\r\nprint (\"Predictions: {}\".format(str(predictions)))\r\n\r\n# Compute confusion matrix\r\nconfusion_matrix = ConfusionMatrix(test_set[LABEL].values, predictions)\r\nconfusion_matrix\r\n\r\n# normalized confusion matrix\r\nconfusion_matrix.plot(normalized=True)\r\nplt.show()\r\n\r\n#Statistics are also available as follows\r\nconfusion_matrix.print_stats()\r\ncms = confusion_matrix.stats()\r\nprint(\"Overall Accuracy is \", round(cms['overall']['Accuracy'], 2),\", Kappa is \", round(cms['overall']['Kappa'], 2))\r\n#Overall Accuracy is 0.78 , Kappa is 0.66\r\n\r\ndf = cms['class'].reset_index()\r\ndf[df['index'].str.contains('Precision')]\r\ndf[df['index'].str.contains('Sensitivity')]\r\ndf[df['index'].str.contains('Specificity')]\r\n\r\n# Cleaning\r\ndel(training_set, test_set, predictions, feature_cols, df, cms, ev, num_mapping, batch_size); gc.collect()\r\n\r\n#%% Multiclass classification using Tensorflow Multi level\r\nimport pandas as pd\r\nimport numpy as np\r\nimport tensorflow as tf\r\nimport os\r\nfrom pandas_ml import ConfusionMatrix\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.model_selection import train_test_split\r\nimport gc; gc.enable()\r\n\r\nos.chdir(\"D:\\\\trainings\\\\tensorflow\")\r\ntf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.INFO)\r\n\r\n# fix random seed for reproducibility\r\nseed = 123; np.random.seed(seed); tf.compat.v1.set_random_seed(seed)\r\n\r\n# identifications of features and response\r\nFEATURES = [\"SepalLength\",\"SepalWidth\",\"PetalLength\",\"PetalWidth\"]\r\nLABEL = \"Species\"\r\n\r\n# Read data\r\ndata = pd.read_csv(\"./data/iris.csv\")\r\ndata.shape\r\ndata.dtypes\r\ndata.head(2)\r\ndata.info()\r\nprint(data.describe()) # .unstack()\r\n#print(data.describe(include = [np.number])) # for number only\r\n\r\n# Labels dtype should be integer\r\ndata[LABEL].unique() #.astype(int)\r\nnum_mapping = {\"setosa\":0 ,\"versicolor\" :1,\"virginica\":2}\r\ndata[LABEL] = data[LABEL].replace(num_mapping)\r\n\r\n# now convert the types\r\ndata[LABEL] = pd.to_numeric(data[LABEL], errors='coerce')\r\ndata.dtypes\r\n\r\n#Segragate 85% and 15%\r\ntraining_set ,test_set = train_test_split(data,test_size=0.15)\r\ndel(data)\r\n\r\n# Build the model\r\nmodel = tf.keras.models.Sequential() # same as tf.keras.Sequential()\r\nmodel.add(tf.keras.layers.Dense(4, input_shape=(4,), activation=tf.nn.relu)) # , kernel_initializer = tf.random_normal_initializer\r\nmodel.add(tf.keras.layers.Dense(3, activation=tf.nn.softmax)) # , kernel_initializer = tf.random_normal_initializer\r\n\r\nmodel.summary()\r\n\r\n#Compile\r\nmodel.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy'])\r\n\r\n# Train it\r\nmodel.fit(training_set[FEATURES].values, training_set[LABEL].values, epochs=1000,batch_size=8) # 6 min\r\n\r\n# Evaluate on test data\r\nmodel.evaluate(test_set[FEATURES].values, test_set[LABEL].values)\r\n# loss value & metrics values: [0.13, 0.95]\r\n\r\n#Making Predictions\r\npredictions = model.predict(x=test_set[FEATURES].values, verbose=1)\r\n\r\n# Extracting max probability\r\npredictions_number = np.array([])\r\nfor row_num in range(predictions.shape[0]): # row_num = 0\r\n predictions_number = np.append(predictions_number, np.argmax(predictions[row_num]))\r\n\r\n# Compute confusion matrix\r\nconfusion_matrix = ConfusionMatrix(test_set[LABEL].values, predictions_number)\r\nconfusion_matrix\r\n\r\n# normalized confusion matrix\r\nconfusion_matrix.plot(normalized=True)\r\nplt.show()\r\n\r\n#Statistics are also available as follows\r\nconfusion_matrix.print_stats()\r\ncms = confusion_matrix.stats()\r\nprint(\"Overall Accuracy is \", round(cms['overall']['Accuracy'], 2),\", Kappa is \", round(cms['overall']['Kappa'], 2))\r\n# 1000: Overall Accuracy is 0.96 , Kappa is 0.93\r\n\r\ndf = cms['class'].reset_index()\r\ndf[df['index'].str.contains('Precision')]\r\ndf[df['index'].str.contains('Sensitivity')]\r\ndf[df['index'].str.contains('Specificity')]\r\n\r\ndel(training_set, test_set, predictions, df, cms, num_mapping, predictions_number); gc.collect()\r\n","sub_path":"tensorflowShivOmkar/1.3.dl_tf_basic_classifications.py","file_name":"1.3.dl_tf_basic_classifications.py","file_ext":"py","file_size_in_byte":7260,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"293826826","text":"import pglet\nfrom pglet import Textbox\n\ndef test_textbox_add():\n tb = pglet.Textbox(id=\"txt1\", label=\"Your name:\")\n assert isinstance(tb, pglet.Control)\n assert isinstance(tb, pglet.Textbox)\n assert tb.get_cmd_str(indent=' ') == ' textbox id=\"txt1\" label=\"Your name:\"', \"Test failed\"\n\ndef test_textbox_update():\n tb = Textbox(id=\"txt1\", error_message=\"Enter the value\")\n assert tb.get_cmd_str(update=True) == '\"txt1\" errorMessage=\"Enter the value\"', \"Test failed\"\n\ndef test_add_textbox():\n # open page\n p = pglet.page('test_textbox', no_window=True)\n\n tb_value = \"Line1\\nLine2\\nLine3\"\n\n # add textbox\n txt1 = Textbox(value=tb_value, multiline=True)\n tb = p.add(txt1)\n assert tb.id.startswith('_'), \"Test failed\"\n assert tb.id == txt1.id, \"Test failed\"\n\n # get textbox value\n ret_tb_value = p.get_value(tb)\n assert ret_tb_value == tb_value, \"Test failed\"","sub_path":"tests/test_textbox.py","file_name":"test_textbox.py","file_ext":"py","file_size_in_byte":906,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"156594096","text":"# +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ #\n# Import data provided by Towne et al. and perform Spectral-POD #\n# #\n# do it in Python and train yourself #\n# #\n# +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ #\n\n# import libraries -------------\n#import sys\nimport h5py\nimport time\nimport numpy as np\nimport matplotlib\nmatplotlib.use('TkAgg') # do this before importing pylab\nimport matplotlib.pyplot as plt\n\nfrom pod import pod_time\nfrom spod import spod\n\n# beginning of the program -----\nprint('\\n Start of the program.\\n')\n\n# read data --------------------\nfilen = '/home/davide/PROJECTS/SpectralPOD/spod-towne/spod_matlab-master/jet_data/jetLES.mat'\nf = h5py.File(filen, 'r') # h5py.File acts like a Python dictionary\n\nprint('list(f.keys()):', list(f.keys()))\nprint('f:', f)\n\nfor i1 in range(0,len(f)):\n print(' field', i1,'. key: ',list(f.keys())[i1], \\\n ' shape: ',f[list(f.keys())[i1]].shape)\n\n# plot mean pressure field -----\ncmap = plt.get_cmap('PiYG')\nfig, ax = plt.subplots()\nax.axis('equal')\nax.autoscale(enable=True, axis='y', tight=True)\ncs = ax.contourf(f[\"x\"],f[\"r\"],f[\"p_mean\"])\n\n\n# Proper Orthogoanl Decomposition of p\npres = f[\"p\"]\ndt = f['dt'][0,0]\nprint('dt:', dt)\n# only on the first 100 snapshots\n# pres = pres[:,:,0:100]\nuPOD, s, vh = pod_time(pres[:,:,0:300],dt)\n\n# SPOD -------------------------\nnSnap1 = 256 ; nOverlap = 128 ; nSnap = 5000 # pres.shape[len(pres.shape)-1]\nPPsi, LLambda = spod(pres[:,:,0:nSnap],nSnap1,nOverlap,dt)\n\nT = nSnap1 * dt\ndOm = 2*np.pi / T\nom = np.arange(0,nSnap1) * dOm\nprint(' dOmega: ', dOm)\n\nprint('LLambda.shape: ', LLambda.shape)\n\nplt.figure(101)\nplt.semilogy(om,LLambda[:,:])\n# plt.plot(om,LLambda[:,:])\n\n# plot sum of lambda\nplt.figure(102)\nplt.semilogy(om,np.sum(LLambda,axis=1))\n\n# plot mode contributions at peak frequency\nisort = np.argsort(LLambda[0:int(np.ceil(nSnap1/2)),0]) ; isort = isort[::-1]\n \n# print(' omega: ', dOm*indmax )\n# plt.figure(103)\n# plt.semilogy(np.arange(0,LLambda.shape[1]),LLambda[indmax,:],'o')\n\nnplot = 2\nfor ip in range(0,nplot):\n indMode = isort[ip]\n # Plot mode contribution at the considered frequency\n plt.figure(51+ip)\n plt.semilogy(np.arange(0,LLambda.shape[1]),LLambda[indMode,:],'o')\n plt.title(' Mode contributions at $\\omega$ = %8.3f' % (dOm*indMode))\n\n # Plot optimal mode\n # reshape back to bi-dimensional field\n OptMode = np.reshape(PPsi[:,indMode,1], pres.shape[0:2])\n plt.figure(111+ip, figsize=(9, 4.5))\n plt.subplot(211)\n plt.contourf(f[\"x\"],f[\"r\"],np.real(OptMode))\n plt.axis('scaled')\n ax.autoscale(enable=True, axis='y', tight=True)\n plt.title(' Optimal mode: $\\omega$ = %8.3f. Real part' % (dOm*indMode))\n plt.ylabel('r')\n \n plt.subplot(212)\n plt.contourf(f[\"x\"],f[\"r\"],np.imag(OptMode))\n plt.axis('scaled')\n ax.autoscale(enable=True, axis='y', tight=True)\n plt.title(' Optimal mode: $\\omega$ = %8.3f. Imag part' % (dOm*indMode))\n plt.xlabel('x')\n plt.ylabel('r')\n \n\nplt.figure(201, figsize=(5, 10))\nnPlot = 9\nstepPlot = 2\nfor ip in range(1,nPlot):\n Mode = np.reshape(PPsi[:,ip*stepPlot,1], pres.shape[0:2])\n ax = plt.subplot(str(nPlot-1)+str(1)+str(ip))\n plt.contourf(f[\"x\"],f[\"r\"],np.imag(Mode))\n plt.axis('scaled')\n ax.autoscale(enable=True, axis='y', tight=True)\n plt.title('$\\omega$ = %8.3f , $\\lambda$ = %8.3f' \\\n % (dOm*ip*stepPlot,LLambda[ip*stepPlot,0]))\n ax.title.set_fontsize(10)\n\nplt.show()\n\n# # low-dimensional example ----------------------\n# data = np.array( [[[0, 1],\n# [2, 3],\n# [4, 5]],\n# [[6, 7],\n# [8, 9], \n# [10, 11]]] )\n# datav_C = np.ravel(data,order='C')\n# datav_F = np.ravel(data,order='F')\n# print('data.shape: ', data.shape)\n# print('datav_C.shape: ', datav_C.shape)\n# print('datav_C : ', datav_C )\n# print('datav_F.shape: ', datav_F.shape)\n# print('datav_F : ', datav_F )\n# pod_time(data)\n# # low-dimensional example ----------------------\n\n# # plt.show()\n# \n# # plot movie pressure time evolution ----\n# dt = f['dt'][0][0]\n# nt = f['p'].shape[2] # f['nt'][0][0]\n# print('nt: ',f['p'].shape[2],', dt: ',dt)\n# pmin = 4.33963 # np.amin(f['p'])\n# pmax = 4.51078 # np.amax(f['p'])\n# print('min(p):', pmin)\n# print('max(p):', pmax)\n# \n# fig_mov = plt.figure()\n# nclevs = 30\n# #plt.ion()\n# def animate():\n# for i in range(0,int(nt),20):\n# print('i: ',i,'. t = ',i*f['dt'][0,0])\n# # im=plt.imshow(f[\"p\"][:,:,i])\n# im=plt.contourf(f[\"x\"],f[\"r\"],f[\"p\"][:,:,i], nclevs, \\\n# vmin=pmin, vmax=pmax , \\\n# cmap=plt.cm.bone )\n# fig_mov.canvas.draw()\n# # time.sleep(1.0)\n# \n# win = fig_mov.canvas.manager.window\n# fig_mov.canvas.manager.window.after(1000, animate)\n# #plt.axis('equal')\n# plt.axis([0, 20, 0, 2.9],'equal')\n# plt.gca().set_aspect('equal', adjustable='box')\n# plt.show()\n# \n# \n# \n# print('\\n End of the program. Bye!\\n')\n# # end of the program -----\n","sub_path":"projects/SpectralPOD/spod-python/old_py/towne_data3.py","file_name":"towne_data3.py","file_ext":"py","file_size_in_byte":5206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"579768759","text":"import openpyxl\n\n# Location of the file\npath = 'C:/Users/melikyan.tatevik/Desktop\\megafeed_test_app/Soccer.xlsx'\n\n# To open the workbook\nwb = openpyxl.load_workbook(path)\n\n# Get workbook active sheet\nsheet = wb.active\n\n# Rows to lists\nrow_list = [row for row in sheet.values]\n# print(row_list)\n\nlist_of_lists = [list(elem) for elem in row_list]\nprint(list_of_lists)\n\nfor item in list_of_lists:\n if item[3] == None:\n item[3] = ''\n # elif item[3] == [1,2]:\n\nfor item in list_of_lists:\n if item[4] == None:\n item[4] = ''\n","sub_path":"project/apitest/Sheet1.py","file_name":"Sheet1.py","file_ext":"py","file_size_in_byte":541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"340198634","text":"# -*- coding: utf-8 -*-\nimport tornado.web\nimport tornado.ioloop\nimport tornado.websocket\nimport tornado.httpserver\nfrom sockjs.tornado import SockJSConnection\nimport tornadoredis\nimport redis\n\nfrom importlib import import_module\nimport json\n\nfrom django.conf import settings\nfrom django.contrib.auth import get_user as guser\nfrom django.db import connection\n\nfrom chat.models import Thread\n\nservice_queue = redis.StrictRedis().publish\n\n_engine = import_module(settings.SESSION_ENGINE)\n\n\ndef get_session(session_key):\n return _engine.SessionStore(session_key)\n\n\ndef get_user(session):\n\n class Dummy(object):\n pass\n\n django_request = Dummy()\n django_request.session = session\n return guser(django_request)\n\n\nclass WebSocketsConnection(SockJSConnection):\n channels = []\n thread_id = None\n trip_id = None\n\n @tornado.gen.engine\n def listen_redis(self):\n self.redis_client = tornadoredis.Client()\n self.redis_client.connect()\n\n yield tornado.gen.Task(self.redis_client.subscribe, self.channels)\n self.redis_client.listen(self.on_redis_queue)\n\n def send(self, message, func):\n msg_type = message.channel\n message = json.loads(message.body)\n return super(WebSocketsConnection, self).send({\n 'type': msg_type,\n 'data': message,\n 'func': func\n })\n\n def append_channels(self, channel):\n if type(channel) == list:\n self.channels = self.channels + channel\n else:\n self.channels.append(channel)\n self.channels = list(set(self.channels))\n self.listen_redis()\n\n def on_open(self, info):\n self.django_session = get_session(info.get_cookie(settings.SESSION_COOKIE_NAME).value)\n self.user = get_user(self.django_session)\n connection.close()\n\n self.append_channels(['unread_chat_messages_count_%s' % self.user.pk,\n 'unread_trip_notifications_count_%s' % self.user.pk])\n\n def on_message(self, message):\n pack, data = message.split('::', 1)\n\n # открыли чат с кем-либо\n if pack == 'open_chat':\n try:\n Thread.objects.get(pk=data, participants=self.user)\n self.thread_id = data\n except Thread.DoesNotExist:\n return\n self.append_channels('chat_%s_%s' % (data, self.user.pk))\n\n if pack == 'open_dialogs':\n self.append_channels('dialogs_%s' % self.user.pk)\n\n if pack == 'open_comments':\n self.trip_id = data\n self.append_channels('comments_%s' % data)\n\n if pack == 'open_notifications':\n self.trip_id = data\n self.append_channels('notifications_%s_%s' % (data, self.user.pk))\n\n def on_close(self):\n self.redis_client.unsubscribe(self.channels)\n self.redis_client.disconnect()\n\n def on_redis_queue(self, message):\n if message.kind == 'message':\n # новые сообщения в чате\n if message.channel == 'chat_%s_%s' % (self.thread_id, self.user.pk):\n self.send(message, 'new_message')\n\n # непрочитанные сообщения\n if message.channel == 'unread_chat_messages_count_%s' % self.user.pk:\n self.send(message, 'unread_chat_messages_count')\n\n # диалоги\n if message.channel == 'dialogs_%s' % self.user.pk:\n self.send(message, 'dialogs')\n\n # комментарии в путешествии\n if message.channel == 'comments_%s' % self.trip_id:\n self.send(message, 'comments')\n\n # непрочитанные уведомления\n if message.channel == 'unread_trip_notifications_count_%s' % self.user.pk:\n self.send(message, 'unread_trip_notifications_count')\n\n # уведомления в путешествии\n if message.channel == 'notifications_%s_%s' % (self.trip_id, self.user.pk):\n self.send(message, 'notifications')\n","sub_path":"apps/websockets/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":4117,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"610815689","text":"import tensorflow as tf\nimport numpy as np\nimport random\nimport sys\n\n\nclass StackedLstmRNN(object):\n def __init__(self, num_classes, state_size, num_layers, learning_rate, model_name='stacked_lstm_rnn_model',\n ckpt_path='./ckpt/stacked_lstm/'):\n self.num_classes = num_classes\n self.state_size = state_size\n self. num_layers = num_layers\n self.learning_rate = learning_rate\n self.model_name = model_name\n self.ckpt_path = ckpt_path\n # build graph\n sys.stdout.write('\\nBuilding Graph...')\n tf.reset_default_graph()\n # inputs\n self.xs_ = tf.placeholder(shape=[None, None], dtype=tf.int32)\n self.ys_ = tf.placeholder(shape=[None], dtype=tf.int32)\n # embeddings\n embs = tf.get_variable('emb', [self.num_classes, self.state_size])\n rnn_inputs = tf.nn.embedding_lookup(embs, self.xs_)\n # initial hidden state\n self.init_state = tf.placeholder(shape=[2, self.num_layers, None, self.state_size], dtype=tf.float32,\n name='initial_state')\n # initializer and params\n xav_init = tf.contrib.layers.xavier_initializer\n w = tf.get_variable('W', shape=[num_layers, 4, self.state_size, self.state_size], initializer=xav_init())\n u = tf.get_variable('U', shape=[num_layers, 4, self.state_size, self.state_size], initializer=xav_init())\n b = tf.get_variable('b', shape=[num_layers, 2, self.state_size], initializer=tf.constant_initializer(0.0))\n\n # lstm step\n def __step__(prev, x):\n st_1, ct_1 = tf.unstack(prev)\n # iterate through layers\n st, ct = [], []\n inp = x\n for i in range(self.num_layers):\n # input gate\n ig = tf.sigmoid(tf.matmul(inp, u[i][0]) + tf.matmul(st_1[i], w[i][0]))\n # forget gate\n fg = tf.sigmoid(tf.matmul(inp, u[i][1]) + tf.matmul(st_1[i], w[i][1]))\n # output gate\n og = tf.sigmoid(tf.matmul(inp, u[i][2]) + tf.matmul(st_1[i], w[i][2]))\n # gate weights\n g = tf.tanh(tf.matmul(inp, u[i][3]) + tf.matmul(st_1[i], w[i][3]))\n # new internal cell state\n ct_i = ct_1[i] * fg + g * ig + b[i][0]\n # output state\n st_i = tf.tanh(ct_i) * og + b[i][1]\n inp = st_i\n st.append(st_i)\n ct.append(ct_i)\n return tf.stack([st, ct])\n\n states = tf.scan(__step__, tf.transpose(rnn_inputs, [1, 0, 2]), initializer=self.init_state)\n # predictions\n v = tf.get_variable('V', shape=[self.state_size, self.num_classes], initializer=xav_init())\n bo = tf.get_variable('bo', shape=[self.num_classes], initializer=tf.constant_initializer(0.0))\n # get last state before reshape/transpose\n self.last_state = states[-1]\n # transpose/slice -> pick st from [ct, st] -> pick st[-1] from st\n # and flatten states to 2d matrix for matmult with v\n states = tf.reshape(tf.transpose(states, [1, 2, 3, 0, 4])[0][-1], [-1, self.state_size])\n logits = tf.add(tf.matmul(states, v), bo)\n # predictions\n self.predictions = tf.nn.softmax(logits)\n # optimization\n self.loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(logits=logits, labels=self.ys_))\n self.train_op = tf.train.AdagradOptimizer(learning_rate=self.learning_rate).minimize(self.loss)\n sys.stdout.write(' Done...\\n')\n\n def train(self, train_set, epochs=50, steps_per_epoch=100):\n with tf.Session() as sess:\n sess.run(tf.global_variables_initializer())\n train_loss = 0\n epoch = 0\n try:\n for epoch in range(epochs):\n for step in range(steps_per_epoch):\n xs, ys = train_set.__next__()\n batch_size = xs.shape[0]\n feed_dict = {self.xs_: xs, self.ys_: ys.flatten(),\n self.init_state: np.zeros([2, self.num_layers, batch_size, self.state_size])}\n _, train_loss_ = sess.run([self.train_op, self.loss], feed_dict=feed_dict)\n train_loss += train_loss_\n print('Epoch [{}] loss : {}'.format(epoch, train_loss / steps_per_epoch))\n train_loss = 0\n except KeyboardInterrupt:\n print('Interrupted by user at ' + str(epoch))\n saver = tf.train.Saver()\n saver.save(sess, self.ckpt_path + self.model_name, global_step=epoch)\n\n def generate(self, idx2w, w2idx, num_words=100, separator=' '):\n random_init_word = random.choice(idx2w)\n current_word = w2idx[random_init_word]\n with tf.Session() as sess:\n sess.run(tf.global_variables_initializer()) # init session\n # restore session\n ckpt = tf.train.get_checkpoint_state(self.ckpt_path)\n saver = tf.train.Saver()\n if ckpt and ckpt.model_checkpoint_path:\n saver.restore(sess, ckpt.model_checkpoint_path)\n words = [current_word] # generate operation\n state = None\n state_ = None\n # enter the loop\n for i in range(num_words):\n if state:\n feed_dict = {self.xs_: np.array([current_word]).reshape([1, 1]), self.init_state: state_}\n else:\n feed_dict = {self.xs_: np.array([current_word]).reshape([1, 1]),\n self.init_state: np.zeros([2, self.num_layers, 1, self.state_size])}\n # forward propagation\n preds, state_ = sess.run([self.predictions, self.last_state], feed_dict=feed_dict)\n state = True # set flag to true\n # set new word\n current_word = np.random.choice(preds.shape[-1], 1, p=np.squeeze(preds))[0]\n words.append(current_word) # add to list of words\n return separator.join([idx2w[w] for w in words])\n","sub_path":"rnn_from_scratch/rnn_units/stacked_lstm_rnn.py","file_name":"stacked_lstm_rnn.py","file_ext":"py","file_size_in_byte":6173,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"438135029","text":"\"\"\"\n\n8kyu\nRemoving Elements\n\n\nTake an array and remove every second element out of that array.\nAlways keep the first element and start removing with the next element.\n\nExample:\n\nmy_list = ['Keep', 'Remove', 'Keep', 'Remove', 'Keep', ...]\nNone of the arrays will be empty, so you don't have to worry about that!\n\n\n\"\"\"\n\ndef remove_every_other(my_list):\n counter = 1\n new_list = []\n for i, elm in enumerate(my_list):\n if counter % 2 != 0:\n new_list.append(elm)\n counter += 1\n return new_list\n\n\"\"\"\nOne linear\n\"\"\"\ndef remove_every_other2(my_list):\n return my_list[::2]\n\n\n\nprint(remove_every_other(['Hello', 'Goodbye', 'Hello Again']))\nprint(remove_every_other([1, 2, 3, 4, 5, 6, 7, 8, 9, 10]))\nprint(remove_every_other([['Goodbye'], {'Great': 'Job'}]))\n","sub_path":"src/codewars/python/8kyu/removing_element.py","file_name":"removing_element.py","file_ext":"py","file_size_in_byte":788,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"158931807","text":"import numpy as np\n\n\ndef to_chemcraft_struct(charges, x, comment=''):\n lines = [str(len(charges)), comment]\n for i in range(len(charges)):\n lines.append('{}\\t{:.11f}\\t{:.11f}\\t{:.11f}'.format(charges[i], *x[i * 3: i * 3 + 3]))\n lines.append('')\n return '\\n'.join(lines)\n\n\ndef from_chemcraft_struct(s):\n lines = s.split('\\n')\n\n cnt = int(lines[0])\n\n charges, struct = [], []\n for i in range(cnt):\n splitted = lines[2 + i].split()\n charges.append(splitted[0])\n struct.append(np.array(list(map(float, splitted[1:]))))\n\n return charges, np.concatenate(struct)\n\n\nif __name__ == '__main__':\n import numpy as np\n\n X = np.array([0.000000000, -0.859799324, 0.835503236,\n 0.000000000, -0.100462324, 1.431546236,\n 0.000000000, -1.619136324, 1.431546236])\n\n print(to_chemcraft_struct([1, 1, 8], X, 'chemcraft format test'))\n","sub_path":"chemistry/utils/io.py","file_name":"io.py","file_ext":"py","file_size_in_byte":910,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"388519042","text":"#!/usr/bin/env python2\n#encoding: UTF-8\n\n# Post an anonymous gist and get the URL back\n\nclass Gist:\n \n def post_gist(self,file_content):\n files = {\n 'spam.txt' : {\n 'content': 'What... is the air-speed velocity of an unladen swallow?'\n }\n }\n gist = create_gist('Answer this to cross the bridge', files)\n comments = [c for c in gist.iter_comments()]\n # []\n comment = gist.create_comment('Bogus. This will not work.')\n # Which of course it didn't, because you're not logged in\n # comment == None\n print(gist.html_url)","sub_path":"classes/GitHubController.py","file_name":"GitHubController.py","file_ext":"py","file_size_in_byte":629,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"234320402","text":"from flask import request\n\nfrom forms import AuthForm\nfrom models import App\nfrom utils import send_json_response\nfrom views import views\n\n\n@views.before_request\ndef exchange_format():\n if not request.is_json:\n return send_json_response(\n message={\n 'message': 'Не валидный формат. Должен быть JSON.'\n },\n status_code=400\n )\n\n\n@views.before_request\ndef auth():\n form = AuthForm()\n\n if not form.validate_on_submit():\n return send_json_response(\n message=form.errors,\n status_code=400\n )\n\n if not is_valid_api_key(api_key=form.api_key.data):\n return send_json_response(\n message={'message': 'Ошибка аутентификации'},\n status_code=401\n )\n\n\ndef is_valid_api_key(api_key):\n if App.query.filter_by(key=api_key).scalar():\n return True\n return False\n\n","sub_path":"middleware.py","file_name":"middleware.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"432988020","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('form', '0003_auto_20151209_2352'),\n ]\n\n operations = [\n migrations.RenameField(\n model_name='checkin',\n old_name='full_name',\n new_name='first_name',\n ),\n migrations.AddField(\n model_name='checkin',\n name='last_name',\n field=models.CharField(null=True, max_length=120),\n ),\n ]\n","sub_path":"customer/form/migrations/0004_auto_20151210_0415.py","file_name":"0004_auto_20151210_0415.py","file_ext":"py","file_size_in_byte":559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"250572247","text":"import sys\nimport pytest\nimport os\n\nimport pandas as pd\nimport numpy as np\nimport math\nimport scipy\nfrom scipy import stats\nfrom sklearn import metrics, linear_model\n\nfrom gpmodel import gpkernel\nfrom gpmodel import gpmodel\nfrom gpmodel import gpmean\nfrom gpmodel import chimera_tools\n\n\nseqs = pd.DataFrame([['R', 'Y', 'M', 'A'], ['R', 'T', 'H', 'A'],\n ['R', 'T', 'M', 'A']],\n index=['A', 'B', 'C'], columns=[0, 1, 2, 3])\nseqs = seqs.append(seqs.iloc[0])\nseqs.index = ['A', 'B', 'C', 'A']\n\nspace = [('R', 'T', 'C'), ('Y', 'T', 'B'), ('M', 'H', 'I'), ('A', 'A', 'B')]\ncontacts = [(0, 1), (2, 3)]\n\nclass_Ys = pd.Series([-1, 1, 1, -1], index=seqs.index)\nreg_Ys = pd.Series([-1, 1, 0.5, -.4], index=seqs.index)\nvariances = pd.Series([0.11, 0.18, 0.13, 0.14], index=seqs.index)\nstruct = gpkernel.StructureKernel(contacts)\nSE_kern = gpkernel.StructureSEKernel(contacts)\nalpha = 0.1\nfunc = gpmean.StructureSequenceMean(space, contacts, linear_model.Lasso,\n alpha=alpha)\n\ntest_seqs = pd.DataFrame([['R', 'Y', 'M', 'A'],\n ['R', 'T', 'H', 'A']],\n index=['A', 'D'])\n\n\ndef test_creation():\n model = gpmodel.GPRegressor(struct, mean_func=func,\n objective='LOO_log_p', guesses=(1.0,))\n assert model.mean_func == func\n assert model.objective == model._LOO_log_p\n pytest.raises(AttributeError, 'model.fit(seqs, reg_Ys)')\n model._set_params(objective=model._log_ML)\n assert model.objective == model._log_ML\n\n model_2 = gpmodel.GPRegressor(struct, objective='log_ML')\n assert model != model_2\n\n # create a model\n model = gpmodel.GPRegressor(struct, guesses=(2, 2), objective='LOO_log_p')\n # pickle the model\n model.dump('test_creation.pkl')\n # reload the model\n model = gpmodel.GPRegressor.load('test_creation.pkl')\n # delete the pickle\n os.remove('test_creation.pkl')\n # test the model\n assert model.objective == model._LOO_log_p\n assert model.guesses == (2, 2)\n\n # fit the model\n model.fit(seqs, reg_Ys)\n ML = model.ML\n lp = model.log_p\n K = model._K\n Ky = model._Ky\n alpha = model._alpha\n L = model._L\n hypers = model.hypers\n # pickle the model\n model.dump('test_creation.pkl')\n # reload the model\n model = gpmodel.GPRegressor.load('test_creation.pkl')\n # delete the pickle\n os.remove('test_creation.pkl')\n # test the model\n assert model.ML == ML\n assert model.log_p == lp\n assert np.isclose(model.hypers.var_n, hypers.var_n)\n assert np.isclose(model.hypers.var_p, hypers.var_p)\n assert np.isclose(model._K, K).all()\n assert np.isclose(model._Ky, Ky).all()\n assert np.isclose(model._alpha, alpha).all()\n assert np.isclose(model._L, L).all()\n\n model = gpmodel.GPClassifer(struct)\n assert model.objective == model._log_ML\n model.dump('test_creation.pkl')\n model = gpmodel.GPClassifer.load('test_creation.pkl')\n assert model.objective == model._log_ML\n os.remove('test_creation.pkl')\n\n\ndef test_score():\n model = gpmodel.GPRegressor(struct)\n model.fit(seqs, reg_Ys)\n preds = model.predict(seqs)\n pred_Y = [p[0] for p in preds]\n r1 = stats.rankdata(reg_Ys)\n r2 = stats.rankdata(pred_Y)\n kendall = stats.kendalltau(r1, r2).correlation\n R = np.corrcoef(reg_Ys, pred_Y)[0, 1]\n u = sum((y1-y2)**2 for y1, y2 in zip(reg_Ys, pred_Y))\n m = sum(reg_Ys)/len(reg_Ys)\n v = sum((y1-m)**2 for y1 in reg_Ys)\n R2 = 1 - u/v\n scores = model.score(seqs, reg_Ys, 'R', 'kendalltau', 'R2')\n assert R == scores['R']\n assert kendall == scores['kendalltau']\n assert R2 == scores['R2']\n\n scores = model.score(seqs, reg_Ys)\n assert kendall == scores\n\n scores = model.score(seqs, reg_Ys, 'R')\n assert R == scores\n\n scores = model.score(seqs, reg_Ys, 'R', 'R2')\n assert R == scores['R']\n assert R2 == scores['R2']\n\n pytest.raises(ValueError, 'model.score(seqs, reg_Ys, \"R3\")')\n\n model.fit(seqs, class_Ys)\n preds = [p[0] for p in model.predict(seqs)]\n score = model.score(seqs, class_Ys)\n fpr, tpr, _ = metrics.roc_curve(class_Ys, preds)\n AUC = metrics.auc(fpr, tpr)\n assert AUC == score\n\n\ndef test_regression():\n model = gpmodel.GPRegressor(struct)\n model.fit(seqs, reg_Ys)\n assert np.isclose(model.hypers.var_p, 0.63016924576335664),\\\n 'Regression model.hypers.var_p is incorrect'\n assert np.isclose(model.hypers.var_n, 0.18044635639161319),\\\n 'Regression model.hypers.var_n is incorrect'\n assert np.isclose(model.ML, 4.59859002013),\\\n 'Regression model.ML is incorrect'\n assert np.isclose(model.log_p, 3.79099390643),\\\n 'Regression model.log_p is incorrect'\n assert np.isclose(model._K,\n struct.make_K(seqs,\n [model.hypers.var_p],\n normalize=True)).all()\n assert model._ell == len(seqs.index)\n\n m = reg_Ys.mean()\n s = reg_Ys.std()\n assert model.mean == m\n assert model.std == s\n\n normed_Ys = (reg_Ys - m) / s\n assert (y1 == y2 for y1, y2 in zip(normed_Ys, model.normed_Y)), \\\n 'Model does not normalize Y-values correctly.'\n\n Y_mat = np.matrix(normed_Ys)\n\n # test marginal likelihood\n vp = 1.0\n vn = model.hypers.var_n\n K_mat = np.matrix(struct.make_K(seqs, normalize=True))\n Ky = vp*K_mat + np.eye(len(reg_Ys))*vn\n first = 0.5*Y_mat*np.linalg.inv(Ky)*Y_mat.T\n second = math.log(np.linalg.det(Ky))*0.5\n third = len(reg_Ys)*math.log(2*math.pi)*.5\n ML = first + second + third\n\n assert np.isclose(model._log_ML((vn, vp)), ML.item()), \\\n 'log_ML fails: ' + ' '.join([str(first), str(second), str(third)])\n\n K_inv = np.linalg.inv(Ky)\n mus = np.diag(Y_mat.T - K_inv*Y_mat.T/K_inv)\n vs = np.diag(1/K_inv)\n res2 = pd.DataFrame(list(zip(mus, vs)), index=normed_Ys.index,\n columns=['mu', 'v'])\n log_p_1 = model._LOO_log_p((vn, vp))\n log_p_2 = 0.5*np.sum(np.log(res2['v']) +\n np.power(normed_Ys-res2['mu'], 2)/res2['v'] +\n np.log(2 * np.pi))\n assert np.isclose(log_p_1, log_p_2), \\\n ('Regression model does not correctly '\n 'calculate LOO log predictive probability')\n # test predictions\n kA = np.matrix([model.kern.calc_kernel(test_seqs.loc['A'],\n seq1, [model.hypers.var_p],\n normalize=True) for seq1\n in [seqs.iloc[i] for i in range(len(seqs.index))]])\n kD = np.matrix([model.kern.calc_kernel(test_seqs.loc['D'],\n seq1, [model.hypers.var_p],\n normalize=True) for seq1\n in [seqs.iloc[i] for i in range(len(seqs.index))]])\n EA = (kA*np.linalg.inv(model._Ky)*Y_mat.T) * s + m\n ED = (kD*np.linalg.inv(model._Ky)*Y_mat.T) * s + m\n k_star_A = model.kern.calc_kernel(test_seqs.loc['A'],\n test_seqs.loc['A'],\n normalize=True)*model.hypers.var_p\n k_star_D = model.kern.calc_kernel(test_seqs.loc['D'],\n test_seqs.loc['D'],\n normalize=True)*model.hypers.var_p\n\n var_A = (k_star_A - kA*np.linalg.inv(model._Ky)*kA.T) * s**2\n var_D = (k_star_D - kD*np.linalg.inv(model._Ky)*kD.T) * s**2\n predictions = model.predict(test_seqs, delete=False)\n assert np.isclose(EA, predictions[0][0])\n assert np.isclose(ED, predictions[1][0])\n assert np.isclose(var_A, predictions[0][1])\n assert np.isclose(var_D, predictions[1][1])\n\n [(E, v)] = model.predict(test_seqs.loc[['D']])\n assert np.isclose(E, ED)\n assert np.isclose(v, var_D)\n\n # test regression with StructureSEKernel\n model = gpmodel.GPRegressor(SE_kern)\n model.fit(seqs, reg_Ys)\n kA = np.matrix([model.kern.calc_kernel(test_seqs.loc['A'],\n seq1, [model.hypers.sigma_f,\n model.hypers.ell]) for seq1\n in [seqs.iloc[i] for i in range(len(seqs.index))]])\n kD = np.matrix([model.kern.calc_kernel(test_seqs.loc['D'],\n seq1, [model.hypers.sigma_f,\n model.hypers.ell]) for seq1\n in [seqs.iloc[i] for i in range(len(seqs.index))]])\n EA = (kA*np.linalg.inv(model._Ky)*Y_mat.T) * s + m\n ED = (kD*np.linalg.inv(model._Ky)*Y_mat.T) * s + m\n k_star_A = model.kern.calc_kernel(test_seqs.loc['A'],\n test_seqs.loc['A'],\n [model.hypers.sigma_f,\n model.hypers.ell])\n k_star_D = model.kern.calc_kernel(test_seqs.loc['D'],\n test_seqs.loc['D'],\n [model.hypers.sigma_f,\n model.hypers.ell])\n var_A = (k_star_A - kA*np.linalg.inv(model._Ky)*kA.T) * s**2\n var_D = (k_star_D - kD*np.linalg.inv(model._Ky)*kD.T) * s**2\n predictions = model.predict(test_seqs, delete=False)\n assert np.isclose(EA, predictions[0][0])\n assert np.isclose(ED, predictions[1][0])\n assert np.isclose(var_A, predictions[0][1])\n assert np.isclose(var_D, predictions[1][1])\n\n [(E, v)] = model.predict(test_seqs.loc[['D']])\n assert np.isclose(E, ED)\n assert np.isclose(v, var_D)\n\n h = model.hypers\n ML = model.ML\n model._set_params(hypers=(1, 1, 10))\n assert np.isclose([1, 1, 10], model.hypers).all()\n\n model._set_params(hypers=h)\n assert np.isclose(h, model.hypers).all()\n assert model.ML == ML\n\n # test predictions with mean function\n model = gpmodel.GPRegressor(struct, mean_func=func)\n model.fit(seqs, reg_Ys)\n X, terms = chimera_tools.make_X([''.join(row) for _, row\n in seqs.iterrows()],\n space, contacts, collapse=False)\n clf = linear_model.Lasso(alpha=alpha)\n clf.fit(X, reg_Ys)\n preds = clf.predict(X)\n assert np.array_equal(model.mean_func.means, preds)\n assert np.array_equal(model.Y, reg_Ys - model.mean_func.means)\n # test accuracy of predictions\n Y_mat = np.matrix(model.normed_Y.values.T)\n s = model.std\n m = model.mean\n kA = np.matrix([model.kern.calc_kernel(test_seqs.loc['A'],\n seq1, [model.hypers.var_p])\n for seq1 in [seqs.iloc[i]\n for i in range(len(seqs.index))]])\n kD = np.matrix([model.kern.calc_kernel(test_seqs.loc['D'],\n seq1, [model.hypers.var_p])\n for seq1 in [seqs.iloc[i]\n for i in range(len(seqs.index))]])\n EA = (kA*np.linalg.inv(model._Ky)*Y_mat.T) * s + m + \\\n model.mean_func.mean(test_seqs.loc[['A']])\n ED = (kD*np.linalg.inv(model._Ky)*Y_mat.T) * s + m + \\\n model.mean_func.mean(test_seqs.loc[['D']])\n k_star_A = model.kern.calc_kernel(test_seqs.loc['A'],\n test_seqs.loc['A'],\n [model.hypers.var_p])\n k_star_D = model.kern.calc_kernel(test_seqs.loc['D'],\n test_seqs.loc['D'],\n [model.hypers.var_p])\n var_A = (k_star_A - kA*np.linalg.inv(model._Ky)*kA.T) * s**2\n var_D = (k_star_D - kD*np.linalg.inv(model._Ky)*kD.T) * s**2\n predictions = model.predict(test_seqs, delete=False)\n assert np.isclose(EA, predictions[0][0])\n assert np.isclose(ED, predictions[1][0])\n assert np.isclose(var_A, predictions[0][1])\n assert np.isclose(var_D, predictions[1][1])\n K_inv = np.linalg.inv(Ky)\n mus = np.diag(Y_mat.T - K_inv*Y_mat.T/K_inv)\n vs = np.diag(1/K_inv)\n\n res1 = model.LOO_res((vn, vp))\n res2 = pd.DataFrame(list(zip(mus, vs)), index=normed_Ys.index,\n columns=['mu', 'v'])\n assert res1.equals(res2), \\\n 'Regression model does not correctly predict LOO values'\n res2['mu'] = model.unnormalize(res2['mu'])\n res2['v'] *= model.std**2\n res = model.LOO_res((vn, vp), unnorm=True)\n assert res.equals(res2), \\\n 'Regression model does not correctly predict LOO values'\n res2['mu'] += model.mean_func.means\n res = model.LOO_res((vn, vp), add_mean=True)\n assert res.equals(res2), \\\n 'Regression model does not correctly predict LOO values'\n\n model = gpmodel.GPRegressor(struct, guesses=(100.0,))\n model.fit(seqs, reg_Ys, variances=variances)\n assert np.array_equal(model._K + np.diag(variances) / model.std**2,\n model._Ky)\n assert np.isclose(model.hypers.var_p, 0.6657325327454634),\\\n 'Regression model.hypers.var_p is incorrect'\n assert np.isclose(model.ML, 4.69092081175),\\\n 'Regression model.ML is incorrect'\n assert np.isclose(model.log_p, 3.96114185627),\\\n 'Regression model.log_p is incorrect'\n variances.index = ['C', 'D', 'E', 'F']\n pytest.raises(AttributeError, \"model.fit(seqs, reg_Ys, variances)\")\n\n\ndef test_classification():\n model = gpmodel.GPClassifer(struct)\n model.fit(seqs, class_Ys)\n test_F = pd.Series([-.5, .5, .6, .1])\n true_vps = np.array([43.7865018929, 43.763806573,\n 43.7993636657, 43.7893659609])\n assert np.isclose(model.hypers.var_p, true_vps).any(),\\\n 'Classification model.hypers.var_p is incorrect'\n assert np.isclose(model.ML, 2.45520196), \\\n 'Classification model.ML is incorrect'\n assert model._ell == len(seqs.index)\n assert model._logistic_likelihood(1, 0) == 0.5\n assert model._logistic_likelihood(-1, 0) == 0.5\n pytest.raises(RuntimeError, \"model._logistic_likelihood(-2,1)\")\n\n assert model._log_logistic_likelihood(class_Ys,\n pd.Series([0, 0, 0, 0])) \\\n == math.log(0.125/2)\n pytest.raises(RuntimeError,\n (\"model._log_logistic_likelihood\"\n \"(class_Ys,pd.Series([0,0,0]))\"))\n # Test the gradient and hessian functions\n glll = model._grad_log_logistic_likelihood(class_Ys, test_F)\n for i in range(model._ell):\n for j in range(model._ell):\n if i != j:\n assert glll[i, j] == 0, 'Non-zero non-diagonal element of glll'\n else:\n assert np.isclose(glll[i, j], (class_Ys.iloc[i]+1)/2. -\n model._logistic_likelihood(1,\n test_F.iloc[i]))\n\n hess = model._hess(test_F)\n for i in range(model._ell):\n for j in range(model._ell):\n if i != j:\n assert hess[i, j] == 0, 'Non-zero non-diagonal element of W'\n else:\n pi_i = model._logistic_likelihood(1, test_F.iloc[i])\n assert np.isclose(hess[i, j], pi_i*(1-pi_i))\n\n # Test find_F (Algorithm 3.1) by seeing if\n # the result satisfies Eq 3.17 from RW\n f_hat = model._find_F(hypers=(1,))\n K = model.kern.make_K(seqs, hypers=(1,), normalize=True)\n K_mat = np.matrix(K)\n glll = np.matrix(np.diag(model._grad_log_logistic_likelihood\n (class_Ys, f_hat))).T\n f_check = K_mat*glll\n for fh, fc in zip(f_hat, f_check):\n assert np.isclose(fh, fc), 'find_F fails for var_p = 1.'\n\n vp = 0.1\n f_hat = model._find_F(hypers=(vp,))\n K_mat = K_mat*vp\n glll = np.matrix(np.diag(model._grad_log_logistic_likelihood\n (class_Ys, f_hat))).T\n f_check = K_mat*glll\n for fh, fc in zip(f_hat, f_check):\n assert np.isclose(fh, fc), 'find_F fails for var_p ~= 1.'\n\n # Test the functions that calculate marginal likelihood\n logq = model._logq(f_hat, hypers=(vp,))\n W = model._hess(f_hat)\n W_root = scipy.linalg.sqrtm(W)\n F_mat = np.matrix(f_hat)\n l = len(f_hat)\n L = np.linalg.cholesky(np.matrix(np.eye(l))+W_root*K_mat*W_root)\n b = W*F_mat.T + glll\n a = b - W_root*np.linalg.lstsq(L.T,\n np.linalg.lstsq(L, W_root*K_mat*b)[0])[0]\n check_q = 0.5*a.T*F_mat.T - model._log_logistic_likelihood(class_Ys, f_hat)\n check_q += sum(np.log(np.diag(L)))\n assert np.isclose(check_q, logq)\n assert np.isclose(model._log_ML([vp]), check_q)\n\n # Test the function that is integrated\n v = 0.4\n m = 0.5\n for z in [0.5, 0, 10000, -10000]:\n val = 1./(1+np.exp(-z))/math.sqrt(2*math.pi*v)*math.exp(-(z-m)**2/2/v)\n assert np.isclose(val, model._p_integral(z, m, v))\n for z in [np.inf, -np.inf]:\n assert np.isclose(0., model._p_integral(z, m, v))\n\n # test predictions\n preds = model.predict(test_seqs)\n for p1, p2 in zip(preds, [0.19135285438781463, 0.779243365755702]):\n assert np.isclose(p1[0], p2), 'Predictions failed.'\n\n\nif __name__ == \"__main__\":\n test_creation()\n test_regression()\n test_classification()\n test_score\n","sub_path":"gpmodel/test/model_test.py","file_name":"model_test.py","file_ext":"py","file_size_in_byte":17218,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"89498602","text":"from flask import Blueprint, render_template, request\nfrom app.database import db\nfrom app.mod_results.models import SentimentResult\n\nmod_results = Blueprint('results', __name__, url_prefix='/results')\n\n\n@mod_results.route('/', methods=['GET'])\ndef result():\n keyword = request.args.get('keyword', default='HWAT', type=str)\n sentiment_results = db.query(SentimentResult).filter_by(\n keyword=keyword)\n avrg_pol = 0\n standing = 'Neutral'\n av = False\n\n if sentiment_results.count() > 0:\n av = True\n for result in sentiment_results:\n avrg_pol += result.polarity\n\n avrg_pol = avrg_pol/sentiment_results.count()\n\n\n if avrg_pol < 0:\n standing = 'Bad'\n\n if avrg_pol > 0:\n standing = 'Good'\n\n return render_template(\"results/results.html\", sentiment_results=sentiment_results, standing=standing, av=av)\n","sub_path":"web/app/mod_results/controllers.py","file_name":"controllers.py","file_ext":"py","file_size_in_byte":894,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"474116630","text":"# reports/urls.py\nfrom django.conf.urls import url\n\nfrom reports import views\n\napp_name = 'reports'\nurlpatterns = [\n url(r'^reports/$', views.reports, name='reports'),\n url(r'^reports/(?P\\w+)/(?P\\d+)/(?P\\d+)/$', views.reports, name='by_month_id'),\n url(r'^reports/category/(?P\\d+)/$', views.reports, name='by_category_id'),\n]\n","sub_path":"reports/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":367,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"149908578","text":"import nltk\nimport webbrowser\nfrom cPickle import dump,load\n\n\n##1. CC Coordinating conjunction\n##2. CD Cardinal number\n##3. DT Determiner\n##4. EX Existential there\n##5. FW Foreign word\n##6. IN Preposition or subordinating conjunction\n##7. JJ Adjective\n##8. JJR Adjective, comparative\n##9. JJS Adjective, superlative\n##10. LS List item marker\n##11. MD Modal\n##12. NN Noun, singular or mass\n##13. NNS Noun, plural\n##14. NNP Proper noun, singular\n##15. NNPS Proper noun, plural\n##16. PDT Predeterminer\n##17. POS Possessive ending\n##18. PRP Personal pronoun\n##19. PRP$ Possessive pronoun\n##20. RB Adverb\n##21. RBR Adverb, comparative\n##22. RBS Adverb, superlative\n##23. RP Particle\n##24. SYM Symbol\n##25. TO to\n##26. UH Interjection\n##27. VB Verb, base form\n##28. VBD Verb, past tense\n##29. VBG Verb, gerund or present participle\n##30. VBN Verb, past participle\n##31. VBP Verb, non-3rd person singular present\n##32. VBZ Verb, 3rd person singular present\n##33. WDT Wh-determiner\n##34. WP Wh-pronoun\n##35. WP$ Possessive wh-pronoun\n##36. WRB Wh-adverb \n\n\ndef read_file(file_name):\n fileHandle = open ( file_name ) \n string= fileHandle.read() \n fileHandle.close()\n return string\ndef parse_text(text):\n \n strlist = text.split(',')\n #return strlist[1:]\n return strlist\ndef tag_text(command):\n \"\"\"tagger_pkl=open('t2.pkl','rb')\n tagger=load(tagger_pkl)\n tagger_pkl.close()\"\"\"\n\n document=open('t3.pkl','rb')\n tagger=load(document)\n document.close()\n after_tagged=[]\n for c in command:\n tokens=c.split()\n after_tagged.append(tagger.tag(tokens))\n\n #print after_tagged\n return after_tagged\n \ndef entity_recognition(tagged):\n\n \n grammar=r\"\"\"\nCommand: {+??+?}\n {++}\nGoogle: {<.*>*}\n {<.*>*}\n \n\"\"\"\n\n \"\"\"I want to turn off the light, open the door,\n unlock window. What is the capital of China?\"\"\"\n cp=nltk.RegexpParser(grammar)\n trees=[]\n for t in tagged:\n tree=cp.parse(t)\n trees.append(tree)\n #tree.draw()\n\n lis1=[]\n lis2=[]\n for tree in trees:\n for subtree in tree.subtrees():\n if subtree.node=='Command':\n #print subtree,type(subtree)\n lis1.append(subtree.pos())\n if subtree.node=='Google':\n lis2.append(subtree.pos())\n return lis1,lis2\n\ndef command_transformation(commands):\n lis=[]\n for command in commands:\n #print command\n lis.append([command[i][0] for i in range(len(command))])\n #for single_word in commands:\n #print single_word\n return lis\ndef google_transformation(commands):\n lis=[]\n for command in commands:\n lis.append([command[i][0][0] for i in range(len(command))])\n return lis\n##def understand():\n## s=read_file(\"received_parameter.txt\")\n## \n## #s=\"what is the capital of China, what is the capital of China. What is the capital of China ?\"\n## command=parse_text(s)\n## tagged=tag_text(command)\n## \n## command1,command2=entity_recognition(tagged)\n##\n## command1=command_transformation(command1)\n## command2=google_transformation(command2)\n## #print command1\n## #print command2\n## return command1,command2\n\nif __name__ == '__main__':\n #understand()\n c1,c2=understand();\n if c2:\n webbrowser.open(\"www.google.com/?#newwindow=1&q=\"+ ' '.join(c2[0]))\n \n","sub_path":"understand.py","file_name":"understand.py","file_ext":"py","file_size_in_byte":3721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"57830204","text":"#!/usr/bin/env python3 \n#-*- coding: utf-8 -*- \n\n\"\"\"\nCreated on Wed Dec 25 12:47:44 2019\n@author: Amal Htait\n\"\"\"\n\nfrom elasticsearch import Elasticsearch, helpers \nimport os, uuid, json\nimport common as c\nimport sentiment as s\n\ndirpath = os.getcwd()\n\n# Generator to push bulk data from a JSON file into an Elasticsearch index / that function is changed according to files content\ndef bulkJsonData(json_file, _index,whatStuff):\n\tjson_list = c.getDataFromFile(json_file)\n\tfor doc in json_list:\n\n\t\tjson_doc = json.loads(doc)\n\n\t\tsentiment=[0,0,0]\n\n\t\t# use a 'yield' generator so that the data isn't loaded into memory\n\t\tif '{\"index\"' not in doc:\n\n\t\t\t# clean the text in comments and title from special character and emojies after json conversion\n\t\t\tif 'data' in json_doc:\n\t\t\t\tfor dt in json_doc['data']:\n\t\t\t\t\tif 'post' in dt:\n\t\t\t\t\t\tmy_text = dt[\"post\"]\n\n\t\t\t\t\t\t#get sentiment\n\t\t\t\t\t\tsentiment = s.getSentiment(my_text)\n\n\t\t\t\t\t\tclean_my_text = c.cleanText(my_text)\n\t\t\t\t\t\tdt.update([ (\"post\", clean_my_text) ])\n\t\t\t\t\t\tjson_doc.update([ (\"all_text\", clean_my_text) ])\t\n\n\t\t\tif 'attachments' in json_doc:\n\t\t\t\tfor att in json_doc['attachments']:\n\t\t\t\t\tif 'data' in att:\n\t\t\t\t\t\tfor dt in att['data']:\n\t\t\t\t\t\t\tif 'external_context' in dt:\n\t\t\t\t\t\t\t\tif 'name' in dt[\"external_context\"]:\n\t\t\t\t\t\t\t\t\tmy_text2 = dt[\"external_context\"][\"name\"]\n\t\t\t\t\t\t\t\t\tclean_my_text2 = c.cleanText(my_text2)\n\t\t\t\t\t\t\t\t\tdt[\"external_context\"].update([ (\"name\", clean_my_text2) ])\t \n\n\t\t\t\t\t\t\tif 'media' in dt:\n\t\t\t\t\t\t\t\tmy_title2 = dt['media']['title']\n\t\t\t\t\t\t\t\tclean_my_title2 = c.cleanText(my_title2)\n\t\t\t\t\t\t\t\tdt['media'].update([ (\"title\", clean_my_title2) ])\t \n\n\t\t\t\t\t\t\t\tif 'description' in dt['media']:\n\t\t\t\t\t\t\t\t\tmy_description = dt['media'][\"description\"]\n\t\t\t\t\t\t\t\t\tclean_my_description = c.cleanText(my_description)\n\t\t\t\t\t\t\t\t\tdt['media'].update([ (\"description\", clean_my_description) ])\n\n\t\t\t\t\t\t\tif 'place' in dt:\n\t\t\t\t\t\t\t\tmy_loc = dt[\"place\"][\"coordinate\"]\n\t\t\t\t\t\t\t\tmy_lat = my_loc[\"latitude\"]\n\t\t\t\t\t\t\t\tmy_lon = my_loc[\"longitude\"]\n\t\t\t\t\t\t\t\tnew_my_loc = [my_lon,my_lat]\n\t\t\t\t\t\t\t\tdt[\"place\"].update([ (\"location\", new_my_loc) ])\t \t \n\t \t\t\n\n\t\t\tif 'title' in json_doc:\n\t\t\t\tmy_title = json_doc[\"title\"]\n\t\t\t\tclean_my_title = c.cleanText(my_title)\n\t\t\t\tjson_doc.update([ (\"title\", clean_my_title) ])\t\n\n\t\t\t# add sentiment\n\t\t\tjson_doc.update([ (\"mySentiment\", sentiment[0]) ]) \n\t\t\tjson_doc.update([ (\"sentPositive\", sentiment[1]) ]) \n\t\t\tjson_doc.update([ (\"sentNegative\", sentiment[2]) ]) \n\n\t\t\t# add load_type, used later for filter\n\t\t\tjson_doc.update([ (\"load_type\", whatStuff) ]) \n\t\t\tjson_doc.update([ (\"source_type\", \"facebook\") ])\n\t\t\tnew_doc = str(json_doc).replace(\"'\", '\"')\n\n\t\t\tyield {\n\t\t\t\t\"_index\": _index,\n\t\t\t\t\"_id\": uuid.uuid4(),\n\t\t\t\t \"_source\": new_doc\n\t\t\t}\n\ndef fct():\n\telastic = Elasticsearch(hosts=[{'host':'localhost','port':9200}])\n\n\t# the Schema, used to force specific types and to add alias/ it is changed according to files content\n\tschema = { \n\n\t \"settings\": {\n\t \"analysis\": {\n\t \"analyzer\": {\n\t \"my_english_analyzer\": {\"type\": \"standard\",\"stopwords\": \"_english_\"}\n\t }\n\t }\n\t },\n\t\t \n\t \"mappings\":{\n\t \"properties\":{ \n\t\t \"timestamp\": { \"type\":\"date\", \"format\":\"EEE MMM dd HH:mm:ss ZZ yyyy||date_optional_time||epoch_second\"},\n\t\t \"created_at\": { \"type\": \"alias\", \"path\": \"timestamp\" },\n\t\t \"all_text\": { \"type\": \"text\", \"analyzer\": \"my_english_analyzer\", \"fields\": {\"keyword\": { \"type\": \"keyword\"}}, \"fielddata\": True},\n \"mySentiment\": { \"type\":\"float\"},\n \"sentPositive\": { \"type\":\"float\"},\n \"sentNegative\": { \"type\":\"float\"},\n\t\t \n\t\t \"data\": {\n\t \t\"properties\": {\n\t\t \t\t\"post\": { \"type\": \"text\", \"analyzer\": \"my_english_analyzer\", \"fields\": {\"keyword\": { \"type\": \"keyword\"}}, \"fielddata\": True}\n\t\t \t}\n\t\t },\n\t\t \"attachments\": {\n\t\t \t\"properties\": {\n\t\t\t \"data\": {\n\t\t\t \t\"properties\": {\n\t\t\t\t \"place\": {\n\t\t\t\t \t\"properties\": {\n\t\t\t\t \t\t\"location\": {\"type\": \"geo_point\"}\n\t\t\t\t \t}\t\t \t\n\t\t\t \t}\n\t\t\t }\n\t\t\t }\n\t\t\t \n\t\t \t}\n\t\t } \n\t \t}\n\t \t}\n\n\t}\n\n\t# Create index with a schema\n\tc.createIndex('dfp_text_fb_groups', schema, elastic)\n\n\n\tinputFolder = dirpath+\"/script/dataSource/json-facebook_data/groups\"\n\tfor loadType in [\"your_posts_and_comments_in_groups_fixed\",\"your_group_membership_activity\"]:\n\t\twhatFile = os.path.join(inputFolder, loadType+'.json')\n\t\t\n\t\ttry:\n\t\t\tresponse = helpers.bulk(elastic, bulkJsonData(whatFile, \"dfp_text_fb_groups\",loadType))\n\t\t\tprint (\"Insert Facebook Groups\")\n\t\texcept:\n\t\t\tprint (\"Error in Facebook posts in Group :\"+ whatFile)\n\t\t\tpass\n\n\n\t\n\n","sub_path":"script/transformer/facebook/FacebookBulkInsert_postGrp.py","file_name":"FacebookBulkInsert_postGrp.py","file_ext":"py","file_size_in_byte":4666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"385967661","text":"import glob, os\nimport nltk\nimport json\n\n\nrootdir='C:/Documents and Settings/cukier_j.OECDMAIN/My Documents/vastChallenge/mc3/MC_3_Materials_4_4_2011'\ndef count_words(string, key):\n return float(len(string) - len(string.replace(key, ''))) / float(len(key))\n\ndef contains(theString, theQueryValue):\n return theString.find(theQueryValue) > -1\n\nplacesW=(\n #\"Vastopolis\",\n \"Cornertown\", \"Westside\", \"Villa\", \"Smogtown\", \"Plainville\", \"Southville\", \"Downtown\", \"Riverside\", \"Suburbia\", \"Eastside\", \"Uptown\", \"Lakeside\",\n \"Vastopolis Armed Forces\",\"Vastopolis Dome\", \"Courthouse\", \"Capital Building\", \"Convention Center\", \"Vastopolis Airport\",\"Westside Stadium\", \n \"Vastopolis City Hospital\",\"St. Georges Hospital\",\"Corner Hospital\", \"Westside Hospital\", \"Northside Hospital\", \"River Hospital\", \"Villa Hospital\", \"Smogtown Hospital\", \n \"Interstate 610\", \"Interstate 67\", \"Interstate 435\", \"Interstate 494\", \"Interstate 269\", \"Interstate 270\",\"Interstate 124\", \"Interstate 905\",\n \"Vast River\")\n\nthreatW=(\"threat\", \"menace\", \"terror\", \"hack\", \"security\")\n\nbioW=(\"bioterrorism\", \"microbe\")\nairW=(\"hijack\", \"crash\", \"flight 256\")\nfireW=(\"arson\", \"bomb\", \"explosive\", \"explosion\", \"arson\", \"fire\")\nhackW=(\"hack\", \"virus\")\n\n\nout=open('C:/Documents and Settings/cukier_j.OECDMAIN/My Documents/vastChallenge/largeTaggedFile.js', 'w')\nscores=[]\n\n\n\nfor n in range(1,4475):\n file=\"0\"*(5-len(str(n)))+str(n)+\".txt\"\n f=open(rootdir + '/' + file, 'r')\n lines=f.readlines()\n f.close()\n t=''\n for l in lines:\n l=l.decode('iso-8859-1')\n t=t+l\n thisScore={\"name\":file[:5], \"values\":[0,0], \"size\":len(t)}\n type='none'\n typeMax=0\n for word in placesW:\n thisScore[\"values\"][0]=thisScore[\"values\"][0]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n for word in threatW:\n thisScore[\"values\"][1]=thisScore[\"values\"][1]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n if thisScore[\"values\"][1]>typeMax:\n type='threat'\n typeMax=thisScore[\"values\"][1]\n for word in bioW:\n thisScore[\"values\"][2]=thisScore[\"values\"][1]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n if thisScore[\"values\"][2]>typeMax:\n type='bio'\n typeMax=thisScore[\"values\"][2]\n for word in airW:\n thisScore[\"values\"][3]=thisScore[\"values\"][1]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n if thisScore[\"values\"][3]>typeMax:\n type='air'\n typeMax=thisScore[\"values\"][3]\n for word in fireW:\n thisScore[\"values\"][4]=thisScore[\"values\"][1]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n if thisScore[\"values\"][4]>typeMax:\n type='fire'\n typeMax=thisScore[\"values\"][4]\n for word in hackW:\n thisScore[\"values\"][5]=thisScore[\"values\"][1]+count_words(t,word)\n lines[2]=lines[2].replace(word, ''+word+'')\n if thisScore[\"values\"][5]>typeMax:\n type='hack'\n \n thisScore[\"title\"]=lines[0][:-1].decode('iso-8859-1')\n thisScore[\"text\"]=lines[2][:-1].decode('iso-8859-1')\n date=lines[1][:-1].replace(',','').split(' ')\n if(date[0]=='May'):\n thisScore[\"day\"]=date[1]\n else:\n thisScore[\"day\"]=0\n thisScore[\"date\"]=lines[1][:-1]\n thisScore[\"type\"]=type\n scores.append(thisScore)\n\nout.write(json.dumps(scores))\nout.close()","sub_path":"VAST challenge/MC3-createLargeFile.py","file_name":"MC3-createLargeFile.py","file_ext":"py","file_size_in_byte":3629,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"44388764","text":"from pathlib import Path\nfrom urllib.request import urlretrieve\nimport zipfile\n\n\ndef get_wikitext(level):\n \"\"\"\n Download Wikitext-2 https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/\n \"\"\"\n assert level in [\"2_raw\", \"2_word\", \"103_raw\", \"103_word\"]\n\n level_url = {\n \"2_raw\": \"https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-2-raw-v1.zip\",\n \"2_word\": \"https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-2-v1.zip\",\n \"103_raw\": \"https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-103-raw-v1.zip\",\n \"103_word\": \"https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-103-v1.zip\",\n }\n level_filename = {\n \"2_raw\": \"wikitext-2-raw.zip\",\n \"2_word\": \"wikitext-2.zip\",\n \"103_raw\": \"wikitext-103-raw.zip\",\n \"103_word\": \"wikitext-103.zip\",\n }\n url = level_url[level]\n filename = level_filename[level]\n filepath, _ = urlretrieve(url, filename)\n print(\"Download url:\", url)\n print(\"Downloaded file path:\", filepath)\n\n path_to_unzip = Path(filepath).parent\n print(\"Unarchived and placed under\", path_to_unzip)\n with zipfile.ZipFile(filepath) as zip_:\n zip_.extractall(path_to_unzip)\n\n\nif __name__ == \"__main__\":\n import sys\n\n level = sys.argv[1] if len(sys.argv) >= 2 else \"raw\"\n get_wikitext(level)\n","sub_path":"examples/benchmark-wikitext/get_wikitext.py","file_name":"get_wikitext.py","file_ext":"py","file_size_in_byte":1394,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"533185491","text":"from conf import settings\nimport importlib\nfrom lib.response import BaseResponse\n# def pack(host):\n# package = {}\n# for plugin, v in settings.plugins.items():\n# tmp = {'status':True,'data':None}\n# try:\n# path, cls = v.rsplit(\".\", 1)\n# p = importlib.import_module(path)\n# tmp['data']=getattr(p,cls)().execute(host)\n# except Exception as e:\n# print(e)\n# tmp['status'] = False\n# package[plugin] = tmp\n# print(tmp)\n# return package\n\ndef pack(host):\n package = {}\n for plugin, v in settings.plugins.items():\n try:\n path, cls = v.rsplit(\".\", 1)\n p = importlib.import_module(path)\n data = getattr(p,cls)().execute(host)\n package[plugin]=data\n except Exception as e:\n resp = BaseResponse()\n resp.status = False\n resp.error = e.__str__()\n package[plugin]=resp\n return package","sub_path":"src/plugins/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":989,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"556996366","text":"from ch03_quicksort_v30 import quickSort\nfrom ch04_graphutils_v10 import * \nfrom random import randint\n\ndef prim_MST(g):\n # 입력: 가중치 그래프 G=(V, E), n개의 점과 m개의 선분\n # 출력: 최소 신장 트리\n n = len(g.vertexes())\n m = len(g.edges())\n\n #1. 그래프 G에서 임의의 점 p를 시작점으로 선택하고, D[p]=0으로 놓는다.\n\n # // 배열 D[v]는 점 v에 대하여,\n # // T에 속한 점과 연결되어 있는 선분의 최소 가중치를 저장해두는 배열이다.\n D = {} # 파이선 구현에서는 배열 구조 대신 딕션너리 구조로 D를 표현한다. \n v_list_remained = list(g.vertexes())\n \n #p = v_list_remained.pop( randint(0, len(v_list_remained)-1) )\n p = v_list_remained.pop(0)\n \n D[p] = 0\n\n #2. for (점 p가 아닌 각 점 v에 대하여) { // 배열 D의 초기화 과정(2~6)\n for v in v_list_remained:\n \n #3. if ( 선분 (p,v)가 그래프에 있으면 )\n w_pv = g.edgeweight ( (p, v) )\n print(w_pv)\n if w_pv : \n\n #4. D[v] = 선분 (p,v)의 가중치\n D[v] = w_pv \n\n #5. else \n #6. D[v]=∞\n else:\n D[v] = float('inf') \n # } // end of for \n\n #T= {p} // 초기에 트리 T는 점 p만을 가진다.\n t = Tree_Weighted([p], dict({}))\n\n #8. while (T에 있는 점의 수 < n) {\n updated = True\n while (updated and len(t.vertexes()) < n ) : \n\n #9. T에 속하지 않은 각 점 v에 대하여,\n # D[v]가 최소인 점 vmin과 연결된 선분 (u,vmin)을 T에 추가한다.\n # 단, u는 T에 속한 점이고, 점 vmin은 T에 새로 추가된다.\n d_min=float('inf') \n updated = False\n for v in t.vertexes():\n for e in g.edges_started_from(v):\n if e[1] in v_list_remained :\n if g.edgeweight(e) < d_min :\n d_min = g.edgeweight(e)\n u = v\n v_min = e[1]\n updated = True\n if ( updated ) : \n t.addedge( (u, v_min), d_min )\n v_list_remained.remove( v_min )\n \n #10. for (T에 속하지 않은 각 점 w에 대해서) { // D[w]를 갱신 과정\n #11. if (선분 (vmin,w)의 가중치 < D[w]) // 새로 추가된 점 vmin과 비교하여\n # D[w] = 선분 (vmin,w)의 가중치 // 기존 값보다 작으면 갱신한다.\n for e in g.edges_started_from(v_min):\n if e[1] in v_list_remained:\n D[e[1]] = g.edgeweight( e )\n # }\n # }\n return t\n\n\nif __name__ == \"__main__\" :\n\n V_tuple = ('a', 'b', 'c', 'd', 'e', 'f')\n EW_dict = {('a','b'):3, ('a','d'):2, ('a','e'):4,\n ('b','c'):1, ('b','d'):4, ('b','f'):2,\n ('c','f'):1,\n ('d','e'):5, ('d','f'):7,\n ('e','f'):9}\n \n## V_tuple = ('a', 'b', 'c', 'd', 'e', 'f')\n## EW_dict = {('a','b'):8, ('a','d'):2, ('a','e'):4,\n## ('b','c'):1, ('b','d'):4, ('b','f'):2,\n## ('c','f'):1,\n## ('d','e'):3, ('d','f'):7,\n## ('e','f'):9}\n\n g = Graph_Weighted(V_tuple, EW_dict, undirected=True)\n g.print()\n \n p_tree = prim_MST(g)\n p_tree.print()\n","sub_path":"정보처리알고리즘/실습과제/ch04_primmst_v00.py","file_name":"ch04_primmst_v00.py","file_ext":"py","file_size_in_byte":3390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"601739995","text":"# -*- coding: utf-8 -*-\n\n# The MIT License (MIT)\n#\n# Copyright (c) 2015 Alex Headley \n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nimport zlib\n\nclass ZLIB(object):\n def __init__(self, chunk_size=4 * 1024):\n self._chunk_size = chunk_size\n\n def compress_stream(self, input_buffer, output_buffer):\n output_buffer_pos_start = output_buffer.tell()\n worker = zlib.compressobj()\n chunk = input_buffer.read(self._chunk_size)\n while len(chunk) != 0:\n output_buffer.write(worker.compress(chunk))\n chunk = input_buffer.read(self._chunk_size)\n output_buffer.write(worker.flush())\n return output_buffer.tell() - output_buffer_pos_start\n\n def compress(self, input_data):\n return zlib.compress(input_data)\n\n def decompress_stream(self, input_buffer, output_buffer):\n output_buffer_pos_start = output_buffer.tell()\n worker = zlib.decompressobj()\n chunk = input_buffer.read(self._chunk_size)\n while len(chunk) != 0:\n output_buffer.write(worker.decompress(worker.unconsumed_tail + chunk))\n chunk = input_buffer.read(self._chunk_size)\n output_buffer.write(worker.flush())\n return output_buffer.tell() - output_buffer_pos_start\n\n def decompress(self, input_data):\n return zlib.decompress(input_data)\n\ndecompress = zlib.decompress\ncompress = zlib.compress\n","sub_path":"naabal/util/zlib_wrapper.py","file_name":"zlib_wrapper.py","file_ext":"py","file_size_in_byte":2441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"67691239","text":"# coding: utf-8\nfrom video_downloader import helper\nimport re\n\n\ndef check_and_prepare_url(url):\n\n if 'video?z=' in url:\n url = url.replace('video?z=','')\n end_sign = url.find('%2')\n url = url[0:int(end_sign)]\n return url\n else:\n return url\n\ndef parse_vk_video_urls(url):\n\n soup = helper.get_content_soup(check_and_prepare_url(url))\n scripts = soup.find_all('script')\n current_tag = scripts[-1].string\n\n current_urls = []\n\n for m in re.finditer(r\"(?<=src\\=\\\\\\\")(https:\\\\\\/\\\\\\/c[\\s\\S]*?mp4)\", current_tag):\n current_urls.append(('%s' % (m.group(0))))\n\n\n current_urls = [sl.replace('/','') for sl in current_urls]\n current_urls = [sl.replace('\\\\','/') for sl in current_urls]\n\n print(current_urls)\n\n\nparse_vk_video_urls('https://vk.com/video?z=video-16395362_456239091%2F83b7a2b89db60dacbb%2Fpl_cat_featured')","sub_path":"video_downloader/vk.py","file_name":"vk.py","file_ext":"py","file_size_in_byte":945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"406895218","text":"from turtle import *\r\nimport turtle\r\nimport random\r\ncolors=['red','blue','purple','yellow']\r\n\r\n\r\nr1=random.randrange(450,500)\r\nr2=random.randrange(450,500)\r\nr3=random.randrange(450,500)\r\nr4=random.randrange(450,500)\r\n\r\nt1=Turtle()\r\nt1.hideturtle()\r\nt1.shape(\"turtle\")\r\nt1.color(random.choice(colors))\r\nt1.penup()\r\nt1.setpos(-250,-200)\r\nt1.pendown()\r\nt1.showturtle()\r\nt1.setheading(90)\r\nt1.speed(random.randrange(3,6))\r\n\r\nt2=Turtle()\r\nt2.hideturtle()\r\nt2.shape(\"turtle\")\r\nt2.color(random.choice(colors))\r\nt2.penup()\r\nt2.setpos(-200,-200)\r\nt2.pendown()\r\nt2.showturtle()\r\nt2.setheading(90)\r\nt2.speed(random.randrange(3,6))\r\n\r\n\r\nt3=Turtle()\r\nt3.hideturtle()\r\nt3.shape(\"turtle\")\r\nt3.color(random.choice(colors))\r\nt3.penup()\r\nt3.setpos(-150,-200)\r\nt3.pendown()\r\nt3.showturtle()\r\nt3.setheading(90)\r\nt3.speed(random.randrange(3,6))\r\n\r\nt4=Turtle()\r\nt4.hideturtle()\r\nt4.shape(\"turtle\")\r\nt4.color(random.choice(colors))\r\nt4.penup()\r\nt4.setpos(-100,-200)\r\nt4.pendown()\r\nt4.showturtle()\r\nt4.setheading(90)\r\nt4.speed(random.randrange(3,6))\r\n\r\n\r\n\r\nt1.fd(r1)\r\nt2.fd(r2)\r\nt3.fd(r3)\r\nt4.fd(r4)\r\n\r\n\r\nturtle.mainloop()\r\n","sub_path":"Turtle_Race.py","file_name":"Turtle_Race.py","file_ext":"py","file_size_in_byte":1100,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"466934432","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2019. Mike Herbert\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA\nimport copy\nimport logging\nimport math\nimport re\n\nfrom geofinder import Loc, GeoKeys\n\n\nclass CSVEntry:\n PLACE_ID = 0\n TITLE = 1\n NAME = 2\n LAT = 3\n LON = 4\n FEAT = 5\n ADMIN1_ID = 6\n ADMIN2_ID = 7\n ISO = 8\n ENCLOSED_BY = 9\n TYPE = 10\n\n\nclass GrampsCsv:\n def __init__(self, in_path: str, geodata):\n self.logger = logging.getLogger(__name__)\n self.csv_path = in_path + '.' + 'csv'\n\n self.admin_table = [{}, {}, {}, {}, {}]\n self.csvfile = None\n self.geodata = geodata\n\n @staticmethod\n def get_dict_id(place):\n if place.place_type == Loc.PlaceType.COUNTRY:\n dict_idx = 0\n elif place.place_type == Loc.PlaceType.ADMIN1:\n dict_idx = 1\n elif place.place_type == Loc.PlaceType.ADMIN2:\n dict_idx = 2\n elif place.place_type == Loc.PlaceType.CITY:\n dict_idx = 3\n elif place.place_type == Loc.PlaceType.PREFIX:\n dict_idx = 4\n else:\n msg = f'Dictionary ID - Unknown place type for {place.name}. Type={place.place_type}'\n raise Exception(msg)\n return dict_idx\n\n def write_asis(self, entry:str):\n pass\n\n def get_csv_key(self, place):\n # Fill in admin2\n if place.admin2_id == '':\n self.geodata.geo_files.geodb.get_admin2_id(place)\n if place.admin2_id == '' and len(place.admin2_name.strip(' ')) > 0:\n place.admin2_id = ' '\n if place.place_type == Loc.PlaceType.COUNTRY:\n key = f'{place.country_iso}'\n elif place.place_type == Loc.PlaceType.ADMIN1:\n key = f'{place.admin1_id}_{place.country_iso}'\n elif place.place_type == Loc.PlaceType.ADMIN2:\n key = f'{place.admin2_id}_{place.admin1_id}_{place.country_iso}'\n elif place.place_type == Loc.PlaceType.CITY:\n key = f'{place.city1.strip(\" \")}_{place.admin2_id}_{place.admin1_id}_{place.country_iso}'\n elif place.place_type == Loc.PlaceType.PREFIX:\n key = f'{place.prefix.strip(\" \")}_{place.city1.strip(\" \")}_{place.admin2_id}_{place.admin1_id}_{place.country_iso}'\n else:\n msg = f'Get key - Unknown place type for {place.name}. Type={place.place_type}'\n raise Exception(msg)\n\n key = key.strip('_')\n key = key.strip('_')\n key = key.strip(' ')\n # self.logger.debug(f'key={key.upper().strip(\"_\")} type={place.place_type}')\n return key.upper()\n\n @staticmethod\n def set_CSV_place_type(place: Loc.Loc):\n place.set_place_type()\n if len(place.prefix) > 0:\n place.place_type = Loc.PlaceType.PREFIX\n\n @staticmethod\n def get_csv_name(place):\n if place.place_type == Loc.PlaceType.COUNTRY:\n place.name = place.country_name\n nm = place.country_name\n elif place.place_type == Loc.PlaceType.ADMIN1:\n nm = place.admin1_name\n elif place.place_type == Loc.PlaceType.ADMIN2:\n nm = place.admin2_name\n elif place.place_type == Loc.PlaceType.CITY:\n nm = place.city1\n elif place.place_type == Loc.PlaceType.PREFIX:\n nm = place.prefix\n else:\n msg = f'Get name - Unknown place type for {place.name}. Type={place.place_type}'\n raise Exception(msg)\n return nm\n\n def create_csv_node(self, place: Loc.Loc):\n \"\"\"\n Create CSV row in Dictionary: Place (ID), Title, Name, Type, latitude, longitude,enclosed_by\n :param place:\n :return: None\n \"\"\"\n if place.original_entry == '':\n return\n\n row = [''] * 11\n self.set_CSV_place_type(place)\n\n if place.id == '':\n self.set_CSV_place_type(place)\n place.id = self.get_csv_key(place)\n\n row[CSVEntry.PLACE_ID] = place.id\n row[CSVEntry.ENCLOSED_BY] = place.enclosed_by\n place.id = row[CSVEntry.PLACE_ID]\n\n row[CSVEntry.TITLE] = place.prefix + place.prefix_commas + place.original_entry\n row[CSVEntry.FEAT] = place.feature\n row[CSVEntry.LAT] = f'{float(place.lat):.4f}'\n\n row[CSVEntry.LAT] = f'{float(place.lat):.4f}'\n row[CSVEntry.LON] = f'{float(place.lon):.4f}'\n row[CSVEntry.ADMIN2_ID] = place.admin2_id\n row[CSVEntry.ADMIN1_ID] = place.admin1_id\n row[CSVEntry.ISO] = place.country_iso\n\n place.set_place_type_text()\n row[CSVEntry.NAME] = self.get_csv_name(place)\n row[CSVEntry.TYPE] = place.result_type_text\n key = self.get_csv_key(place)\n dict_idx = self.get_dict_id(place)\n\n if dict_idx == 0:\n place.enclosed_by = ''\n row[CSVEntry.ENCLOSED_BY] = ''\n\n if place.enclosed_by != '':\n if key.count('_') <= row[CSVEntry.ENCLOSED_BY].count('_') and key.count('_') > 0:\n msg = f'Incorrect Enclosure for [{place.original_entry}]. Key= [{key}] Enclosure= [{row[CSVEntry.ENCLOSED_BY]}]'\n self.logger.warning(msg)\n elif key.count('_') < row[CSVEntry.ENCLOSED_BY].count('_') and key.count('_') == 0:\n msg = f'Incorrect Enclosure for [{place.original_entry}]. Key= [{key}] Enclosure= [{row[CSVEntry.ENCLOSED_BY]}]'\n self.logger.warning(msg)\n\n if re.match(r'P\\d\\d\\d\\d', place.id):\n # our item has an ID with P9999, add this row\n self.admin_table[dict_idx][key.upper()] = row\n else:\n res = self.admin_table[dict_idx].get(key.upper())\n if res is None:\n # Nothing there, add this row\n self.admin_table[dict_idx][key.upper()] = row\n else:\n # A node is already there and we don't have a P, so do nothing\n place.id = res[CSVEntry.PLACE_ID]\n\n #self.logger.debug(f'\\nCREATE CSV NODE {key.upper()} idx={dict_idx}: {row}\\n{place.name}')\n\n def move_up_level(self, enclosure_place, idx) -> bool:\n enclosure_place.lat = 99.9\n enclosure_place.lon = 99.9\n\n # Switch place type to next level higher\n if enclosure_place.place_type == Loc.PlaceType.COUNTRY:\n # Already at top\n enclosure_place.feature = 'ADM0'\n return False\n elif enclosure_place.place_type == Loc.PlaceType.ADMIN1:\n enclosure_place.feature = 'ADM1'\n enclosure_place.place_type = Loc.PlaceType.COUNTRY\n elif enclosure_place.place_type == Loc.PlaceType.ADMIN2:\n enclosure_place.feature = 'ADM2'\n enclosure_place.place_type = Loc.PlaceType.ADMIN1\n elif enclosure_place.place_type == Loc.PlaceType.CITY:\n enclosure_place.place_type = Loc.PlaceType.ADMIN2\n elif enclosure_place.place_type == Loc.PlaceType.PREFIX:\n enclosure_place.place_type = Loc.PlaceType.CITY\n else:\n msg = f'Move Up - Unknown place type for {enclosure_place.name}. Type={enclosure_place.place_type}'\n raise Exception(msg)\n\n enclosure_place.remove_old_fields()\n enclosure_place.name = enclosure_place.format_full_nm(None)\n enclosure_place.set_place_type_text()\n # enclosure_place.city1 = tkns[1]\n save_type = enclosure_place.place_type\n self.geodata.find_first_match(enclosure_place.name, enclosure_place)\n enclosure_place.place_type = save_type\n enclosure_place.remove_old_fields()\n enclosure_place.name = enclosure_place.format_full_nm(None)\n enclosure_place.set_place_type_text()\n enclosure_place.id = self.get_csv_key(enclosure_place)\n\n # place.name = place.format_full_nm(None)\n self.logger.debug(f'\\nMOVED UP TO {enclosure_place.name}')\n\n self.create_csv_node(enclosure_place)\n\n new_idx = self.get_dict_id(enclosure_place)\n if new_idx < idx:\n return True\n else:\n msg = f'Move Up - Index error {enclosure_place.name}. Type={enclosure_place.place_type} idx={idx} new_idx={new_idx}'\n raise Exception(msg)\n\n def create_enclosed_by(self, place: Loc.Loc):\n \"\"\"\n Create EnclosedBy elements in Dictionary for CSV file\n :return: None\n \"\"\"\n self.logger.debug(f'\\nCREATE ENCLOSURE FOR {place.original_entry}')\n enclosure_place = copy.copy(place)\n enclosure_place.id = ''\n\n # Move up to enclosure level\n success = self.move_up_level(enclosure_place, idx=self.get_dict_id(enclosure_place))\n if success:\n place.enclosed_by = enclosure_place.id\n self.update_enclosure_id(place)\n return\n\n def update_enclosure_id(self, place):\n key = self.get_csv_key(place)\n dict_idx = self.get_dict_id(place)\n row = self.admin_table[dict_idx].get(key.upper())\n if row:\n if not re.match(r'P\\d\\d\\d\\d', row[CSVEntry.ENCLOSED_BY]):\n row[CSVEntry.ENCLOSED_BY] = place.enclosed_by\n self.admin_table[dict_idx][key.upper()] = row\n self.logger.debug(f'UPDATE ENC for {dict_idx}:{key} New Enclosure=[{place.enclosed_by}]')\n else:\n pass\n else:\n self.logger.warning(f'@@@@@@@@ Cant find row {key}')\n\n def complete_csv(self):\n # Add location enclosures. Create if not there already. Then add as reference.\n self.logger.debug('\\n\\n******** DONE - CREATE CSV ENCLOSURES *********')\n place = Loc.Loc()\n\n # There are separate dictionaries for each hierarchy (prefix, city, county, country).\n # We need to go through prefix table, then city, etc (e.g. reversed order)\n # Create Enclosure records\n for idx, tbl in reversed(list(enumerate(self.admin_table))):\n self.logger.debug(f'===TABLE {idx}===')\n for key in tbl:\n self.retrieve_csv_place(self.admin_table, self.geodata, place, key, idx)\n self.logger.debug(f'** CSV {key} {place.original_entry}')\n\n # Create enclosure for each node at this level\n self.create_enclosed_by(place)\n\n if self.csv_path is not None:\n self.csvfile = open(self.csv_path, \"w\", encoding='utf-8')\n self.logger.debug(f'CSV file {self.csv_path}')\n self.csvfile.write('Place,Title,Name,Type,latitude,longitude,enclosed_by\\n')\n\n # List CSV\n self.logger.debug('*** OUTPUT TABLE ***')\n for idx, tbl in enumerate(self.admin_table):\n for key in tbl:\n # TODO\n row = tbl[key]\n #self.logger.debug(f'IDX={idx} {key} : {row}')\n self.output_row(row)\n\n if self.csv_path is not None:\n self.csvfile.close()\n\n def output_row(self, row):\n if len(row[CSVEntry.ENCLOSED_BY]) > 0:\n enc = f'[{row[CSVEntry.ENCLOSED_BY]}]'\n else:\n enc = ''\n if self.csv_path is not None:\n # 0Place (ID), 1Title, 2Name, 3Type, 4latitude, 5longitude, 6enclosed_by\n title = GeoKeys.capwords(row[CSVEntry.TITLE])\n name = GeoKeys.capwords(row[CSVEntry.NAME])\n\n if math.isnan(float(row[CSVEntry.LAT])) or math.isnan(float(row[CSVEntry.LAT])) :\n self.csvfile.write(f'[{row[CSVEntry.PLACE_ID]}],\"{title}\",\"{name}\",{row[CSVEntry.TYPE]},'\n f' , ,{enc},\\n')\n else:\n self.csvfile.write(f'[{row[CSVEntry.PLACE_ID]}],\"{title}\",\"{name}\",{row[CSVEntry.TYPE]},'\n f'{row[CSVEntry.LAT]},{row[CSVEntry.LON]},{enc},\\n')\n\n @staticmethod\n def retrieve_csv_place(admin_table, geodata, place: Loc.Loc, key, idx):\n # 0Place (ID), 1Title, 2Name, 3Type, 4latitude, 5longitude, 6enclosed_by\n row = admin_table[idx].get(key)\n key_tokens = key.split(\"_\")\n place.place_type = len(key_tokens) - 1\n # self.logger.debug(f'{row}')\n place.feature = row[CSVEntry.FEAT]\n\n place.original_entry = row[CSVEntry.TITLE]\n place.country_iso = row[CSVEntry.ISO]\n place.country_name = geodata.geo_files.geodb.get_country_name(place.country_iso)\n place.enclosed_by = row[CSVEntry.ENCLOSED_BY]\n\n place.lat: float = float(row[CSVEntry.LAT])\n place.lon: float = float(row[CSVEntry.LON])\n\n place.admin2_id = row[CSVEntry.ADMIN2_ID]\n place.admin1_id = row[CSVEntry.ADMIN1_ID]\n place.admin1_name = str(geodata.geo_files.geodb.get_admin1_name(place))\n place.admin2_name = str(geodata.geo_files.geodb.get_admin2_name(place))\n if place.admin2_name is None:\n place.admin2_name = ''\n if place.admin1_name is None:\n place.admin1_name = ''\n\n tokens = place.original_entry.split(',')\n if len(tokens) > 3:\n place.city1 = tokens[-4]\n if len(tokens) > 4:\n place.prefix = tokens[-5]\n\n place.id = row[CSVEntry.PLACE_ID]\n","sub_path":"geofinder/GrampsCsv.py","file_name":"GrampsCsv.py","file_ext":"py","file_size_in_byte":13758,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"227716859","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Aug 1 00:01:33 2020\n\n@author: Puneet Singh\n\"\"\"\n\n# here we are implementing transfer learning using Resnet v2 model..\n# you known about transfer learning.\n# this is tranfer learning for object recognize.\nfrom keras.layers import Input, Lambda, Dense, Flatten # here we are implementing keras layers\n#we are importing this beacuse , we will create last layer for our output, resnet50 model developement 10000 images classifier, so we dont need of 1000\n# classifier nueron on the last layer, so we are modify this last layer according to our problem statement using flatten,dense layers\nfrom keras.models import Model\nfrom keras.applications.resnet50 import ResNet50 # Resnet50 its preimplement and predefine model. we will use this model. this model preimplement in keras liberary\n# for our problem object recongize.\n# use preimplemented model for another similer type problem statements , that concept is called tranfer learning. \nfrom keras.applications.vgg16 import preprocess_input # it for preprocess our input..\nfrom keras.preprocessing import image\nfrom keras.preprocessing.image import ImageDataGenerator # its liberary use for image arugmentaion. its help us to generate new images by zooming in zooming out, horiental flip , vertical flip etc.\nfrom keras.models import Sequential #implement deep learning model.\nimport numpy as np\nfrom glob import glob\nimport matplotlib.pyplot as plt\n\n# re-size all the images to this\nIMAGE_SIZE = [224, 224] # this regnet50 model take image size is 224,224.\n\ntrain_path = 'Datasets/Train' # its path of train dataset..\nvalid_path = 'Datasets/Test' # its path of test dataset..\n\n# add preprocessing layer to the front of VGG\nresnet = ResNet50(input_shape=IMAGE_SIZE + [3], weights='imagenet', include_top=False)\n# here input_shape=IMAGE_size +[3] mean , our image size 224,224 *3 , its basically mean our image is RBG image.\n# weights=imagenet , it already present in keras vgg16 model, so we are not modifiying it.\n#include_top=false - its very important parameter , using this we are tellling last layer we are adding in regnet model, once last layer remove , when we will use this model for transfer leraning.\n\n# here we dont have train our regnet50 layers, because its all layers are already trained , weights are fixed \nfor layer in resnet.layers: # here we are iterating all the layer\n layer.trainable = False # here tell them we are not train , so we give false\n \n # if you dont do this , your model will start train for your small data and give bad accuracy..\n # becoz it regnet trained with many many images.it trained on 10 m of images.\n# dont need to train this model. becoz if you train again for good accuaracy , you required large dataset and high computation power.\n \n\n \n # useful for getting number of classes\nfolders = glob('Datasets/Train/*') # using this we check , how many catergorize basically we have. \n#this check how many folder we have of catergorize. , \n \n\n# our layers - you can add more if you want\nx = Flatten()(resnet.output) # here we are flating last layer of regnet. for add last layer of my problem statement in the regnet\n# x = Dense(1000, activation='relu')(x)\nprediction = Dense(len(folders), activation='softmax')(x) # this statement add categorize in last layer of x. using activation function softmax. using this statement , our mannual create output layer append in regnet model layer\n\n# create a model object\nmodel = Model(inputs=resnet.input, outputs=prediction) # give input is regnet put and outputs is prediction to our vgg16 model\n\n# view the structure of the model\nmodel.summary() # here we check summary of our model\n\n# so we used the regnet model.\n\n# tell the model what cost and optimization method to use\nmodel.compile(\n loss='categorical_crossentropy',\n optimizer='adam',\n metrics=['accuracy']\n)\n# compile\n\nfrom keras.preprocessing.image import ImageDataGenerator\n\n# -------this statements we use for image argumentation-------------\n\n\ntrain_datagen = ImageDataGenerator(rescale = 1./255,\n shear_range = 0.2,\n zoom_range = 0.2,\n horizontal_flip = True)\n\ntest_datagen = ImageDataGenerator(rescale = 1./255)\n\ntraining_set = train_datagen.flow_from_directory('Datasets/Train',\n target_size = (224, 224),\n batch_size = 32,\n class_mode = 'categorical')\n\ntest_set = test_datagen.flow_from_directory('Datasets/Test',\n target_size = (224, 224),\n batch_size = 32,\n class_mode = 'categorical')\n\n'''r=model.fit_generator(training_set,\n samples_per_epoch = 8000,\n nb_epoch = 5,\n validation_data = test_set,\n nb_val_samples = 2000)'''\n#--------------------------------------------------------------------\n# here fit the model to our data\nr = model.fit_generator(\n training_set,\n validation_data=test_set,\n epochs=5,\n steps_per_epoch=len(training_set),\n validation_steps=len(test_set)\n)\n\n# here plot loss\nplt.plot(r.history['loss'], label='train loss')\nplt.plot(r.history['val_loss'], label='val loss')\nplt.legend()\nplt.show()\nplt.savefig('LossVal_loss')\n\n# here plot accuracies\nplt.plot(r.history['acc'], label='train acc')\nplt.plot(r.history['val_acc'], label='val acc')\nplt.legend()\nplt.show()\nplt.savefig('AccVal_acc')\n\nimport tensorflow as tf\n\nfrom keras.models import load_model\n\nmodel.save('facefeatures_new_model.h5') # so here we are saving this model into our locsl folder...\n\n","sub_path":"face_Recognition using regnet moel.py","file_name":"face_Recognition using regnet moel.py","file_ext":"py","file_size_in_byte":5805,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"374973845","text":"# stand-up bot\n\nfrom os import getenv\nfrom re import search\nimport json\nimport discord\nimport datetime\nfrom datetime import timedelta\nfrom dotenv import load_dotenv\n\nload_dotenv()\ntoken = getenv('DISCORD_TOKEN')\nchannels = [648006246380601355,0,613152346050002972]\n\nclient = discord.Client()\n\nwith open('userList.dat','r') as f:\n userList = json.loads(f.read())\n\ndef secondsUntilPostAllowed(userId):\n if userId in userList:\n diff = datetime.datetime.strptime(str(datetime.datetime.utcnow()),'%Y-%m-%d %H:%M:%S.%f')\\\n - datetime.datetime.strptime(userList.get(userId),'%Y-%m-%d %H:%M:%S.%f')\n if diff.days < 1:\n return diff.seconds\n return 0\n\n@client.event\nasync def on_message(message):\n x = search('(Yesterday I:).*\\n(Today I will:).*\\n(Potential hard problems).*(:).*',message.content)\n if message.channel.id in channels:\n if message.author == client.user:\n return\n elif x is None:\n await message.delete()\n await message.author.send('Not the correct format, please read the channel description and try again ' + message.author.name + '.')\n return \n elif secondsUntilPostAllowed(str(message.author.id)) > 0:\n await message.delete()\n await message.author.send('You\\'ve already posted today, try again in ' + str(round((86400-secondsUntilPostAllowed(str(message.author.id)))/3600,2)) + ' hours.')\n return\n \n await message.channel.send('Thank you for the stand-up submission ' + message.author.name + ', good luck!')\n userList.update({str(message.author.id):str(datetime.datetime.utcnow())})\n with open('userlist.dat', \"r+\") as f:\n f.seek(0)\n f.write(json.dumps(userList))\n f.truncate()\n\nclient.run(token)","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":1824,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"544939323","text":"\n#!/usr/bin/python3\n\n###############\n### Modules ###\n###############\n\nimport os\nimport json\nimport wget\nimport time\nimport requests\nimport datetime\nimport numpy as np\nimport pandas as pd\nfrom lxml import html\nfrom datetime import date\nfrom moviepy.editor import *\n\n#################\n### Functions ###\n#################\n\ndef checkDuplicates():\n #load data from txt file\n with open('dataVideo.txt') as f:\n dataVideo = json.load(f)\n\n #load data from txt file\n with open('dataVideo.txt') as f:\n dataVideo2 = json.load(f)\n\n for index in dataVideo:\n for index2 in dataVideo2:\n if index[\"id\"] == index2[\"id\"]:\n print(\"duplicate found\", index2[\"id\"])\n\ndef importTrendingDataToDB():\n \"\"\"\n Update the DB with new trending video\n \"\"\"\n\n def getTrendingUrl():\n \"\"\"\n function to generate the url with signature to retrieve the trending videos data. Trending page is opened in pyppeteer and all the requests url are captured\n INPUT: /\n OUTPUT: the urls are saved in the 2 global variable trendingUrl1 and trendingUrl2 and can be used to retrieve the trending data\n \"\"\"\n print(\"getting trending url\")\n #importing everything for the python version of Pupetteer\n import asyncio\n from pyppeteer import launch\n from pyppeteer_stealth import stealth\n import re\n\n def checkUrl(url):\n \"\"\"function that receive all the request urls and filter on the url to retrieve the trending video data with the signature\n INPUT: url from all the requests being made by the tiktok trending page\n OUTPUT: the 2 url that are used to retrieve trending video data are saved in 2 global variables\n \"\"\"\n #regex for the 2 types of url we are looking for (maxCursor is changing)\n pattern = re.compile(\"https://m.tiktok.com/api/item_list/\\?count=30&id=1&type=5&secUid=&maxCursor=0&minCursor=0.*\")\n pattern2 = re.compile(\"https://m.tiktok.com/api/item_list/\\?count=30&id=1&type=5&secUid=&maxCursor=1&minCursor=0.*\")\n if pattern.match(url):\n #print(url)\n global trendingUrl1\n trendingUrl1 = url\n #print('found trending url 1')\n elif pattern2.match(url):\n global trendingUrl2\n trendingUrl2 = url\n #print('found trending url 2')\n else:\n pass\n #print('not found')\n\n async def main():\n \"\"\"function to launch the browser and capture all the request that are being made by the tiktok page to tget the url with signature\n \"\"\"\n #launching the browser in headless mode\n browser = await launch({'headless': True})\n page = await browser.newPage()\n #removing the timeout\n page.setDefaultNavigationTimeout(0)\n #adding the stealth mode to be undetected\n await stealth(page)\n #capture the url of every request and save the ones we want\n page.on('request', lambda request: checkUrl(request.url))\n await page.goto('https://www.tiktok.com/trending/?lang=en')\n await page.waitFor(2000)\n #scroll down to trigger the second request to get trending video data\n await page.evaluate(\"\"\"{window.scrollBy(0, document.body.scrollHeight);}\"\"\")\n await page.waitFor(2000)\n await browser.close()\n\n try:\n asyncio.get_event_loop().run_until_complete(main())\n except:\n print(\"error to go on the trending page. Retrying...\")\n time.sleep(10)\n asyncio.get_event_loop().run_until_complete(main())\n return 1\n\n def processDataRequest(requestData):\n \"\"\"function to process the data from the trending request\n INPUT: response from trending request\n OUTPUT: list of dictionnary with processed video data\n \"\"\"\n listOfVideoDic = []\n data = requestData.json()\n if 'items' in data:\n for video in data['items']:\n #extracting the info we want to save\n dic = {}\n dic['id'] = video['id']\n dic['timeCreated'] = video['createTime']\n dic['likeCount'] = video['stats']['diggCount']\n dic['shareCount'] = video['stats']['shareCount']\n dic['playCount'] = video['stats']['playCount']\n dic['commentCount'] = video['stats']['commentCount']\n dic['videoUsed'] = False\n dic['videoUsedDate'] = ''\n listOfVideoDic.append(dic)\n return listOfVideoDic\n\n def getTrendingVideoData():\n \"\"\"function that send request to retrieve trending video data\n INPUT: /\n OUTPUT: DF with the video data\n \"\"\"\n print(\"Getting trending video data\")\n listOfVideoDic = []\n #setting the headers where the User-Agent have to be the SAME as the one used by pupeteer\n headers = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3494.0 Safari/537.36\",\n \"Accept-Encoding\": \"gzip, deflate, br\"}\n #store all the cookies\n session = requests.Session()\n #make the request type 1 for trending data\n requestData = session.get(url = trendingUrl1, headers=headers)\n #process data request and return in list of dictionnary\n listOfVideoDic = processDataRequest(requestData) \n\n #make the request type 2 x times\n for _ in range(100):\n #print('request')\n time.sleep(1) #time between each request\n requestData = session.get(url = trendingUrl2, headers=headers)\n #merge result with list of dictionnary\n listOfVideoDic.extend(processDataRequest(requestData))\n #transforming list of dic into df\n newDataDF = pd.DataFrame(listOfVideoDic)\n #dropping the duplicates (appeared in API update why ?)\n newDataDF.drop_duplicates(subset='id',inplace=True,keep='last') \n #setting the index with the id\n newDataDF.set_index('id', inplace=True)\n return newDataDF\n\n def updateInsertDB(newData):\n print(\"merging data into DB\")\n #Loading data DB from txt file\n with open('dataVideo.txt','r') as f:\n videos_dict = json.load(f)\n #loading the dic into a DF\n DB = pd.DataFrame.from_dict(videos_dict)\n #Using the ID of the video as DF index\n DB.set_index('id', inplace=True)\n #number of records before adding new data\n numOldRecord = len(DB)\n\n #adding all the data that are not in DB = insert\n DB = pd.concat([DB, newData[~newData.index.isin(DB.index)]])\n #removing the columns that don't have to be updated\n newData.drop(['videoUsed', 'videoUsedDate'], axis=1, inplace=True)\n #updating the data = updating only the numbers\n DB.update(newData)\n #calulating the number of new records added\n numNewRecord = len(DB)\n numRecordAdded = numNewRecord - numOldRecord\n print(\"Number of records added in DB:\", numRecordAdded)\n print(\"Total number of records:\", numNewRecord)\n return DB\n\n #getting the trending url in global variable\n getTrendingUrl()\n #getting the new data into a DF\n newDataDF = getTrendingVideoData()\n #merging new data in DB\n DB = updateInsertDB(newDataDF)\n #putting back the index as a column to have it in the export\n DB['id'] = DB.index\n #saving DF as json into file\n DB.to_json(r'dataVideo.txt',orient=\"records\")\n\ndef importChallengeDataToDB():\n\n #importing everything for the python version of Pupetteer\n import asyncio\n from pyppeteer import launch\n from pyppeteer_stealth import stealth\n import re\n\n def getDiscoverUrl():\n \"\"\"function to get the signed discover url that will allow to get the list of challenges\n IN: /\n OUT: discover url is saved in global variable\n \"\"\"\n print(\"getting the discover url...\")\n def checkUrlDiscover(url):\n \"\"\"function that receive all the request urls and filter on the discover url with the signature\n INPUT: url from all the requests being made by the tiktok trending page\n OUTPUT: discover url in global variable\n \"\"\"\n #print(url)\n pattern = re.compile(\"https://m.tiktok.com/node/share/discover?.*\")\n if pattern.match(url):\n global discoverUrl\n discoverUrl = url\n else:\n pass\n\n async def main():\n \"\"\"function to launch the browser and capture all the request that are being made by the tiktok page to get the url with signature\n \"\"\"\n #launching the browser in headless mode\n browser = await launch({'headless': True})\n page = await browser.newPage()\n #removing the timeout\n page.setDefaultNavigationTimeout(0)\n #adding the stealth mode to be undetected\n await stealth(page)\n #capture the url of every request and save the ones we want\n page.on('request', lambda request: checkUrlDiscover(request.url))\n await page.goto('https://www.tiktok.com/trending/?lang=en')\n await page.waitFor(3000)\n await browser.close()\n\n try:\n asyncio.get_event_loop().run_until_complete(main())\n except:\n print(\"error to go on the trending page. Retrying...\")\n time.sleep(10)\n asyncio.get_event_loop().run_until_complete(main())\n return 1\n\n def getChallengesList():\n \"\"\"function to retrieve the list of current challenges url\n INPUT:\n OUTPUT: list of all th current challenges link\n \"\"\"\n print(\"Getting the list of challenge...\")\n #setting the headers where the User-Agent have to be the same as the one used by pupeteer\n headers = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3494.0 Safari/537.36\",\n \"Accept-Encoding\": \"gzip, deflate, br\"}\n #store all the cookies\n session = requests.Session()\n try:\n #make the request to get the list of challenge\n r = session.get(url = discoverUrl, headers=headers)\n except:\n print(\"error to get list of challenge. Retrying...\")\n time.sleep(60)\n getChallengesList()\n\n data = r.json()\n listOfLinks = []\n listOfChallengeData = []\n if 'body' in data:\n for challenge in data['body'][2]['exploreList']:\n if challenge['cardItem']['type'] == 1: #check that it is a music challenge type\n listOfLinks.append('https://www.tiktok.com'+challenge['cardItem']['link'])\n dic = {}\n dic['link'] = 'https://www.tiktok.com'+challenge['cardItem']['link']\n dic['musicId'] = challenge['cardItem']['extraInfo']['musicId']\n dic['numberOfVideos'] = challenge['cardItem']['extraInfo']['posts']\n dic['challengeUsed'] = False\n dic['challengeUsedDate'] = ''\n listOfChallengeData.append(dic)\n else:\n print(\"wrong type in discover\")\n else:\n print(\"no body in discover data\")\n saveListOfChallenge(listOfChallengeData)\n return listOfLinks\n\n def saveListOfChallenge(listOfChallengeData):\n print(\"Saving the list of challenge...\")\n #putting list of challenge into a DF\n DFChallengeData = pd.DataFrame.from_dict(listOfChallengeData)\n DFChallengeData.set_index('musicId', inplace=True)\n #loading list of challenge from txt\n with open('listChallenge.txt','r') as f:\n videos_dict = json.load(f)\n challengeDB = pd.DataFrame.from_dict(videos_dict)\n #Using the music ID as DF index\n challengeDB.set_index('musicId', inplace=True)\n #adding all the data that are not in DB = insert\n challengeDB = pd.concat([challengeDB, DFChallengeData[~DFChallengeData.index.isin(challengeDB.index)]])\n #removing the columns that don't have to be updated\n DFChallengeData.drop(['challengeUsed', 'challengeUsedDate'], axis=1, inplace=True)\n #updating the data = updating only the numbers\n challengeDB.update(DFChallengeData)\n #putting back the index as a column to have it in the export\n challengeDB['musicId'] = challengeDB.index\n #saving DF as json into file\n challengeDB.to_json(r'listChallenge.txt',orient=\"records\")\n\n def getChallengeUrl(urlChallenge):\n \"\"\"function to retrieve the data urls for each challenge using pyppeteer\n INPUT: challenge urls\n OUTPUT: challenge datas urls\n \"\"\"\n print(\"Getting the challenge data url...\")\n urlList = []\n\n def checkUrlChallenge(url):\n pattern = re.compile(\"https://m.tiktok.com/share/item/list\\?secUid.*\")\n if pattern.match(url):\n urlList.append(url)\n else:\n pass\n #print('not found')\n\n async def main():\n \"\"\"function to launch the browser and capture all the request that are being made by the tiktok page to tget the url with signature\n \"\"\"\n #launching the browser in headless mode\n browser = await launch({'headless': True})\n page = await browser.newPage()\n #removing the timeout\n page.setDefaultNavigationTimeout(20000)\n #adding the stealth mode to be undetected\n await stealth(page)\n #capture the url of every request and save the ones we want\n page.on('request', lambda request: checkUrlChallenge(request.url))\n await page.goto(urlChallenge)\n await page.waitFor(1000)\n #scroll down to trigger the requests to get video data\n for _ in range(1):\n await page.evaluate(\"\"\"{window.scrollBy(0, document.body.scrollHeight);}\"\"\")\n await page.waitFor(1000)\n await page.waitFor(3000)\n await browser.close()\n\n try:\n asyncio.get_event_loop().run_until_complete(main())\n return urlList\n except:\n print(\"Error to get the challenge url data\")\n return urlList\n\n def processDataRequest(requestData):\n \"\"\"function to process the data from the trending request\n INPUT: response from trending request\n OUTPUT: list of dictionnary with processed video data\n \"\"\"\n listOfVideoDic = []\n data = requestData.json()\n if 'body' in data:\n for video in data['body']['itemListData']:\n #extracting the info we want to save\n dic = {}\n dic['id'] = video['itemInfos']['id']\n dic['musicId'] = video['itemInfos']['musicId']\n dic['timeCreated'] = video['itemInfos']['createTime']\n dic['likeCount'] = video['itemInfos']['diggCount']\n dic['shareCount'] = video['itemInfos']['shareCount']\n dic['playCount'] = video['itemInfos']['playCount']\n dic['commentCount'] = video['itemInfos']['commentCount']\n dic['videoUsed'] = False\n dic['videoUsedDate'] = ''\n listOfVideoDic.append(dic)\n return listOfVideoDic\n\n def getChallengeVideoData(challengeUrlDic):\n \"\"\"function to make the request to retrieve video data for all the challenge and call the function to process it\n INPUT: dic containing all the challenge data url where the challenges are the key\n OUTPUT: sending the response to function to process data\n \"\"\"\n print(\"Getting the challenge video data...\")\n listOfVideoDic = []\n #looping through each challenge and data url\n for challenge in challengeUrlDic:\n for url in challengeUrlDic[challenge]:\n time.sleep(1)\n #setting the headers where the User-Agent have to be the same as the one used by pupeteer\n headers = {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3494.0 Safari/537.36\",\n \"Accept-Encoding\": \"gzip, deflate, br\"}\n #store all the cookies\n session = requests.Session()\n try:\n #make the request type 1 for trending data\n requestData = session.get(url = url, headers=headers)\n listOfVideoDic.extend(processDataRequest(requestData))\n except:\n print(\"Error to get data for challenge\")\n\n #transforming list of dic into df\n newDataDF = pd.DataFrame(listOfVideoDic)\n #dropping the duplicates (appeared in API update why ?)\n newDataDF.drop_duplicates(subset='id',inplace=True,keep='last') \n #setting the index with the id\n newDataDF.set_index('id', inplace=True)\n return newDataDF\n\n def updateInsertDB(newData):\n \n #loading video challenge data into DF\n with open('dataVideoChallenge.txt','r') as f:\n videos_dict = json.load(f)\n DB = pd.DataFrame.from_dict(videos_dict)\n numOldRecord = len(DB)\n #Using the ID of the video as DF index\n DB.set_index('id', inplace=True)\n #adding all the data that are not in DB = insert\n DB = pd.concat([DB, newData[~newData.index.isin(DB.index)]])\n #removing the columns that don't have to be updated\n newData.drop(['videoUsed', 'videoUsedDate'], axis=1, inplace=True)\n #updating the data = updating only the numbers\n DB.update(newData)\n\n numNewRecord = len(DB)\n numRecordAdded = numNewRecord - numOldRecord\n print(\"Number of records added in DB:\", numRecordAdded)\n print(\"Total number of records:\", numNewRecord)\n return DB\n\n challengeUrlDic = {}\n #save discover url in global variable\n getDiscoverUrl()\n #get the list of music challenges url\n challengeList = getChallengesList()\n #looping through each challenge and getting the data url for each challenge\n for challenge in challengeList:\n challengeUrlDic[challenge] = getChallengeUrl(challenge)\n #print(challengeUrlDic)\n newDataDF = getChallengeVideoData(challengeUrlDic)\n #merging new data in DB\n DB = updateInsertDB(newDataDF)\n #putting back the index as a column to have it in the export\n DB['id'] = DB.index\n #saving DF as json into file\n DB.to_json(r'dataVideoChallenge.txt',orient=\"records\")\n\ndef loadDbIntoDf(file):\n \"\"\"\n Function that load the json file with all the data and treat them to return\n original dataframe and a shorter one and likeCount,commentCount,...\n columns rescalled for the score calculation.\n INPUT: json file\n OUTPUT; dataframes and columns\n \"\"\"\n #Loading data\n with open(file,'r') as f:\n videos_dict = json.load(f)\n df = pd.DataFrame.from_dict(videos_dict)\n df_shorter = df[df.videoUsed == False] #take only videos no used before\n df_shorter = df_shorter[df_shorter.id != '6815763642621889797']\n df_shorter = df_shorter[df_shorter.id != '6810425865206304006']\n df_shorter = df_shorter.drop(columns=['timeCreated','videoUsed','videoUsedDate'])\n columns_name = ['id','commentCount','likeCount','playCount','shareCount']\n df_shorter = df_shorter.reindex(columns=columns_name)\n df_shorter = df_shorter.apply(lambda x: x/x.max() if x.name in columns_name[1:] else x)\n return df,df_shorter\n\ndef select(df_shorter,nbvideos):\n \"\"\"\n Function to select a range of best videos according to the value of its score\n defined as combinaton of likeCount, playCount, shareCount and commentCount\n INPUT: DataFrame\n OUTPUT: New DataFrame with only x top videos sorted by score\n\n \"\"\"\n score = (35/100 * df_shorter['likeCount'] + 20/100*df_shorter['playCount'] + 35/100* df_shorter['shareCount']\n + 10/100*df_shorter['commentCount'])*100\n df_shorter['score'] = score\n df_shorter = df_shorter.sort_values('score',ascending=False)\n df_shorter = df_shorter.head(nbvideos)\n #print(df_shorter)\n return df_shorter\n\ndef generateLinkFromId(videoId):\n \"\"\"\n function to generate a valid link to download a video from a video ID. Link is extracted from html trending page\n INPUT: video ID\n OUTPUT: valid video link\n\n \"\"\"\n page = requests.get('https://www.tiktok.com/embed/v2/'+videoId+'?lang=en')\n tree = html.fromstring(page.content)\n buyers = tree.xpath('//*[@id=\"main\"]/div/div/div[1]/div/div/div/div[2]/div[1]/video/@src')\n return buyers[0]\n\ndef download(df_shorter):\n \"\"\"\n Functions to download videos selected using urls.\n INPUT: DataFrame\n OUTPUT: list of videos dowloaded and stored on the folder\n \"\"\"\n path = os.getcwd()+'\\\\'\n df_shorter['urls'] = df_shorter['id'].apply(lambda x: generateLinkFromId(x))\n vid_dl = []\n i = 1\n for u in df_shorter['urls']:\n name = str(i)+'.mp4'\n vid_dl.append(wget.download(u,path+name))\n i = i+1\n return vid_dl\n\ndef merge(vidlist):\n \"\"\"\n Function to merge videos dowloaded in one video.\n INPUT: list of videos downloaded\n OUTPUT: One video (not stored as variable)\n \"\"\"\n today = date.today()\n d = today.strftime(\"%Y_%m_%d\")\n clips = []\n for vid in vidlist:\n if vid.endswith(\".mp4\"):\n clips.append(VideoFileClip(vid))\n m = max(c.h for c in clips)\n clips = [c.resize(height=m) for c in clips]\n #print(clips[0].size)\n finalrender = concatenate_videoclips(clips,method='compose')\n finalrender.write_videofile('TiktokCompile'+d+'.mp4',codec='libx264')\n\ndef update(df,df_shorter):\n \"\"\"\n Function to update videoUsed and videoUsedDate info in the original dataframe\n and save it as json.\n \"\"\"\n today = date.today()\n d = today.strftime(\"%Y_%m_%d\")\n for id in df_shorter['id']:\n df.loc[df['id'] == id,'videoUsed'] = True\n df.loc[df['id'] == id,'videoUsedDate'] = d\n #print(df)\n df.to_json(r'dataVideo.txt',orient=\"records\")\n\ndef importData():\n ### Import new challenge data in the DB ###\n importChallengeDataToDB()\n #importTrendingDataToDB()\n\ndef makeVideo():\n ### Import and manip dataVideo ###\n df,df_shorter = loadDbIntoDf('dataVideo.txt')\n print('Initialization is done...')\n print('')\n ##################\n ### Processing ###\n ##################\n\n print('##################')\n print('### Processing ###')\n print('##################')\n print('')\n\n ### Select x best videos and download them ###\n df_shorter = select(df_shorter,20)\n vid_dl = download(df_shorter)\n\n ### merge videos ###\n merge(vid_dl)\n\n ### Check ID of selected videos and updtate videoUsed status ###\n update(df,df_shorter)\n\n######################\n### Initialization ###\n######################\nstart_time = time.time()\nprint('######################')\nprint('### Initialization ###')\nprint('######################')\n\n### Global variable for the trending urls (should be avoided) ###\ntrendingUrl1 = ''\ntrendingUrl2 = ''\ndiscoverUrl = ''\n\nfor _ in range(100):\n importData()\n time.sleep(1) #time between each request\n #makeVideo()\n\n print('Processing is done... ')\n print(\"--- %s seconds ---\" % (time.time() - start_time))\n print('')\n#makeVideo()\nprint('############')\nprint('### DONE ###')\nprint('############')\n\n############\n### Publish on YT ###\n","sub_path":"videomaker.py","file_name":"videomaker.py","file_ext":"py","file_size_in_byte":24039,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"216141344","text":"import os\nfrom sklearn.metrics import accuracy_score\nfrom os.path import dirname\nimport numpy as np\n\ndef eval_condition(iepoch,print_result_every_x_epoch):\n if (iepoch + 1) % print_result_every_x_epoch == 0:\n return True\n else:\n return False\n\n\ndef eval_model(model, dataloader):\n predict_list = np.array([])\n label_list = np.array([])\n for sample in dataloader:\n y_predict = model(sample[0])\n y_predict = y_predict.detach().cpu().numpy()\n y_predict = np.argmax(y_predict, axis=1)\n predict_list = np.concatenate((predict_list, y_predict), axis=0)\n label_list = np.concatenate((label_list, sample[1].detach().cpu().numpy()), axis=0)\n acc = accuracy_score(predict_list, label_list)\n return acc\n\n\ndef save_to_log(sentence, Result_log_folder, dataset_name):\n father_path = Result_log_folder + dataset_name\n if not os.path.exists(father_path):\n os.makedirs(father_path)\n path = father_path + '/' + dataset_name + '_.txt'\n print(path)\n with open(path, \"a\") as myfile:\n myfile.write(sentence + '\\n')","sub_path":"utils/log_manager.py","file_name":"log_manager.py","file_ext":"py","file_size_in_byte":1090,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"294037600","text":"import numpy as np\nimport pandas as pd\nfrom sklearn.datasets import make_regression\n\n\ndef make_sarsa_frame(\n n_samples: int = 10000,\n n_features: int = 40,\n num_actions: int = 2,\n state_name: str = \"state\",\n action_name: str = \"actions\",\n response_name: str = \"reward\"\n):\n state_x, reward_y = make_regression(\n n_samples=n_samples, n_features=n_features\n )\n xl = len(state_x[0, :])\n state_x, action_z = np.hsplit(state_x, [xl - num_actions])\n\n x_len = len(state_x)\n actions = [action_z[i] for i in range(x_len)]\n state = [state_x[i] for i in range(x_len)]\n reward = [reward_y[i] for i in range(x_len)]\n state_act: pd.DataFrame = pd.DataFrame({\n f\"{state_name}\": state,\n f\"{action_name}\": actions,\n f\"{response_name}\": reward\n })\n return state_act","sub_path":"continuum/data/generator.py","file_name":"generator.py","file_ext":"py","file_size_in_byte":825,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"171570671","text":"# -*- coding: utf-8 -*-\nimport context\nfrom my_lib.crawler import Crawler\nfrom my_lib.JournalDB import Util\nfrom my_lib.exception import AbstractNullError, INS_TO_DBexception, Unknown_HTML_ERROR\nimport new_elsevier_crawler\nimport ori_elsevier_crawler\nimport atexit\n\n# Constant\nJOURNAL = \"Journal of Accounting and Economics\"\nHOME_URL = \"http://www.sciencedirect.com/science/journal/01654101\"\nBASE_URL = \"http://www.sciencedirect.com\"\nURL_GROUP_FILE = \"urls.txt\"\nDOI_PREFIX = \"https://doi.org/\"\n\n# Global Var\nUrlGroup = []\ncrawler = Crawler(JOURNAL, HOME_URL, BASE_URL)\nisduplicate_cnt = 0\n\n# status Var for every loop through an article\nsoup = None\nroot = None\n\n\ndef main():\n setupUrlGroup()\n crawler.LOGGER.info(\"All need to parsed:{}\".format(len(UrlGroup)))\n init_two_crawlers()\n parse_all_url()\n\n\ndef setupUrlGroup():\n global UrlGroup\n f = open(URL_GROUP_FILE, \"r\")\n for line in f:\n line = line.strip()\n UrlGroup.append(line)\n f.close()\n return UrlGroup\n\n\ndef init_two_crawlers():\n ori_elsevier_crawler.init(crawler)\n new_elsevier_crawler.init(crawler)\n\n\ndef parse_all_url():\n global soup, root, isduplicate_cnt\n\n for url in UrlGroup[-1000:-500:10]:\n try:\n getSoup(url)\n if isOrigin(soup):\n crawler.LOGGER.debug(\"isOrigin \\n url:{}\".format(url))\n record = ori_elsevier_crawler.parse_article(url, soup, root)\n elif isNew(soup):\n crawler.LOGGER.debug(\"isNew \\n url:{}\".format(url))\n record = new_elsevier_crawler.parse_article(url, soup, root)\n else:\n raise Unknown_HTML_ERROR(url)\n except AbstractNullError as ex:\n crawler.abstract_null_list.append(url)\n crawler.abstract_null_cnt += 1\n continue\n else:\n try:\n if not isduplicate(record):\n insertToDB(record)\n else:\n crawler.LOGGER.debug(\"isduplicate!\")\n isduplicate_cnt += 1\n except INS_TO_DBexception as ex:\n crawler.LOGGER.error(\"url:{}\\nmsg:{}\".format(url, ex))\n crawler.db_ins_ex_list.append(url)\n crawler.db_ins_ex_cnt += 1\n\n\ndef getSoup(url):\n global soup\n\n html = crawler.scraper(url)\n soup = crawler.getSoup(html)\n assert(soup)\n\n\ndef isOrigin(soup):\n global root\n root = soup.find(\"div\", {'id': 'centerInner'})\n if root:\n return True\n\n return False\n\n\ndef isNew(soup):\n global root\n root = soup.find(\"div\", {'role': 'main'})\n if root:\n return True\n\n\ndef isduplicate(record):\n data = record[0]\n doi = data['doi']\n conn = crawler.CONN_HANDLER.conn\n cursor = crawler.CONN_HANDLER.cursor\n Util_handler = Util(conn, cursor)\n result = Util_handler.isduplicate('data', 'd_doi', doi)\n return result[0]\n\n\ndef insertToDB(record):\n data_record = record[0]\n author_record_group = record[1]\n revised_date = record[2]\n keyword_record_group = record[3]\n jel_record_group = record[4]\n db_ex_flag = False\n db_ex_table_group = []\n\n try:\n paper_id = insertToData(data_record)\n except Exception as ex:\n db_ex_flag = True\n db_ex_table_group.append('Data')\n crawler.LOGGER.error(\"msg:{}\".format(ex))\n crawler.LOGGER.error(ex, exc_info=True)\n else:\n try:\n insertToAuthor(author_record_group, paper_id)\n except Exception as ex:\n db_ex_flag = True\n db_ex_table_group.append('Author')\n crawler.LOGGER.error(\"msg:{}\".format(ex))\n crawler.LOGGER.error(ex, exc_info=True)\n\n if revised_date:\n try:\n insertToRevisedDate(revised_date, paper_id)\n except Exception as ex:\n db_ex_flag = True\n db_ex_table_group.append('Revised_date')\n crawler.LOGGER.error(\"msg:{}\".format(ex))\n crawler.LOGGER.error(ex, exc_info=True)\n\n if keyword_record_group:\n try:\n insertToKeyword(keyword_record_group, paper_id)\n except Exception as ex:\n db_ex_flag = True\n db_ex_table_group.append('Keyword')\n crawler.LOGGER.error(\"msg:{}\".format(ex))\n crawler.LOGGER.error(ex, exc_info=True)\n\n if jel_record_group:\n try:\n insertToJel(jel_record_group, paper_id)\n except Exception as ex:\n db_ex_flag = True\n db_ex_table_group.append('Classification')\n crawler.LOGGER.error(\"msg:{}\".format(ex))\n crawler.LOGGER.error(ex, exc_info=True)\n\n if db_ex_flag:\n raise INS_TO_DBexception(db_ex_table_group)\n\n\ndef insertToData(data_record):\n data_obj = crawler.setData(data_record)\n paper_id = crawler.insertToData(data_obj)\n return paper_id\n\n\ndef insertToAuthor(author_record_group, paper_id):\n for author_record in author_record_group:\n author_obj = crawler.setAuthor(author_record)\n author_obj.setPaperID(paper_id)\n crawler.insertToAuthor(author_obj)\n\n# TODO: for revised_date_record in revised_date_group...\n\n\ndef insertToRevisedDate(revised_date, paper_id):\n revised_date_obj = crawler.setRevised(revised_date)\n revised_date_obj.setPaperID(paper_id)\n crawler.insertToRevisedDate(revised_date_obj)\n\n\ndef insertToKeyword(keyword_record_group, paper_id):\n for keyword_record in keyword_record_group:\n keyword_obj = crawler.setKeyword(keyword_record)\n keyword_obj.setPaperID(paper_id)\n crawler.insertToKeyword(keyword_obj)\n\n\ndef insertToJel(jel_record_group, paper_id):\n for jel_record in jel_record_group:\n jel_obj = crawler.setJel(jel_record)\n jel_obj.setPaperID(paper_id)\n crawler.insertToJel(jel_obj)\n\n\ndef showStatus():\n crawler.LOGGER.warning(\"isduplicate_cnt:{}\".format(isduplicate_cnt))\n crawler.showStatus()\n ori_elsevier_crawler.showStatus()\n new_elsevier_crawler.showStatus()\n\n\nif __name__ == '__main__':\n atexit.register(showStatus)\n main()\n","sub_path":"elsevier/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":6151,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"53114584","text":"# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n\nimport unittest\n\ntry:\n # Attempt to load mock (works on Python 3.3 and above)\n from unittest.mock import patch\nexcept ImportError:\n # Attempt to load mock (works on Python version below 3.3)\n from mock import patch\n\nfrom azure.cli.testsdk import ScenarioTest\nfrom azure_devtools.scenario_tests import AllowLargeResponse\nfrom azext_devops.dev.team.credentials import credential_set\nfrom .utilities.helper import ( DEVOPS_CLI_TEST_ORGANIZATION , DEVOPS_CLI_TEST_PAT_TOKEN, disable_telemetry )\n\nclass PipelinesBuildTaskTests(ScenarioTest): \n @AllowLargeResponse(size_kb=3072)\n @disable_telemetry\n def test_build_task_listShow(self):\n\n with patch('azext_devops.dev.team.credentials._get_pat_token') as mock_pat_token: \n mock_pat_token.return_value = DEVOPS_CLI_TEST_PAT_TOKEN\n self.cmd('az devops login')\n self.cmd('az devops configure --defaults organization=' + DEVOPS_CLI_TEST_ORGANIZATION + ' project=buildtests')\n\n list_task_command = 'az pipelines build task list --detect off --output json'\n list_task_output = self.cmd(list_task_command).get_output_in_json()\n assert len(list_task_output) > 0\n for task in list_task_output:\n assert task[\"definitionType\"] == 'task'\n \n task_to_query = list_task_output[0]\n task_UUID_to_query = task_to_query[\"id\"]\n task_version_to_query = task_to_query[\"version\"]\n version_value = (str(task_version_to_query[\"major\"]) + '.' + str(task_version_to_query[\"minor\"]) + '.' \n + str(task_version_to_query[\"patch\"]))\n\n show_task_command = ('az pipelines build task show --task-id ' + task_UUID_to_query + ' --version ' + version_value + \n ' --detect off --output json')\n show_task_output = self.cmd(show_task_command).get_output_in_json()\n assert show_task_output[\"id\"] == task_UUID_to_query\n","sub_path":"tests/pipelinesBuildTaskTest.py","file_name":"pipelinesBuildTaskTest.py","file_ext":"py","file_size_in_byte":2245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"234297011","text":"import h5py\nimport numpy as np\nimport yt\nimport os\n\n\nID = 3\n\nfor idx in range(ID,ID+1):\n\n filename = \"Data_%06d\"%idx\n\n print('convert '+ filename + ' ...')\n\n fs = h5py.File(filename,\"r\")\n fd = h5py.File(filename+\"_single\",\"w\")\n\n Dens=fs['/GridData/Dens']\n MomX=fs['/GridData/MomX']\n MomY=fs['/GridData/MomY']\n MomZ=fs['/GridData/MomZ']\n Engy=fs['/GridData/Engy']\n Temp=fs['/GridData/Temp']\n\n\n Dens_single=np.empty(Dens.shape,dtype=np.float32)\n MomX_single=np.empty(Dens.shape,dtype=np.float32)\n MomY_single=np.empty(Dens.shape,dtype=np.float32)\n MomZ_single=np.empty(Dens.shape,dtype=np.float32)\n Engy_single=np.empty(Dens.shape,dtype=np.float32)\n Temp_single=np.empty(Dens.shape,dtype=np.float32)\n\n\n Dens.read_direct(Dens_single)\n MomX.read_direct(MomX_single)\n MomY.read_direct(MomY_single)\n MomZ.read_direct(MomZ_single)\n Engy.read_direct(Engy_single)\n Temp.read_direct(Temp_single)\n\n fd.create_dataset('GridData/Dens', data=Dens_single)\n fd.create_dataset('GridData/MomX', data=MomX_single)\n fd.create_dataset('GridData/MomY', data=MomY_single)\n fd.create_dataset('GridData/MomZ', data=MomZ_single)\n fd.create_dataset('GridData/Engy', data=Engy_single)\n fd.create_dataset('GridData/Temp', data=Temp_single)\n\n array=fs['Info/InputPara']\n fd.create_dataset('Info/InputPara', data=array)\n\n array=fs['Info/KeyInfo']\n fd.create_dataset('Info/KeyInfo', data=array)\n\n array=fs['Info/Makefile']\n fd.create_dataset('Info/Makefile', data=array)\n\n array=fs['Info/SymConst']\n fd.create_dataset('Info/SymConst', data=array)\n\n array=fs['Tree/Corner']\n fd.create_dataset('Tree/Corner', data=array)\n fd['Tree/Corner'].attrs['Cvt2Phy'] = fs['Tree/Corner'].attrs['Cvt2Phy']\n\n array=fs['Tree/Father']\n fd.create_dataset('Tree/Father', data=array)\n\n array=fs['Tree/LBIdx']\n fd.create_dataset('Tree/LBIdx', data=array)\n\n array=fs['Tree/Sibling']\n fd.create_dataset('Tree/Sibling', data=array)\n\n array=fs['Tree/Son']\n fd.create_dataset('Tree/Son', data=array)\n\n\n fd.flush()\n\n #After you are done\n fs.close()\n fd.close()\n","sub_path":"FloatConversion.py","file_name":"FloatConversion.py","file_ext":"py","file_size_in_byte":2104,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"135930461","text":"import os\nimport json\nimport sys\n\nimport tensorflow as tf\nfrom keras.models import load_model\n\nfrom models import *\nfrom GAN_training import train\nfrom config import CONFIG_ABSOLUTE_PATH, PROJECT_ABSOLUTE_PATH\nfrom image_generation import generate_and_save_images\nfrom load_data import load_data\n\nos.environ['CUDA_VISIBLE_DEVICES'] = '-1' # change or not depending on your machine\n\n\ndef gan_train():\n train_images = load_data()\n train_images = train_images.astype('float32')\n train_images = (train_images - 127.5) / 127.5 # Normalize from [0,255] to [-1,1]\n\n with open(os.path.join(CONFIG_ABSOLUTE_PATH, 'params.json'), 'r') as param_file:\n params = json.load(param_file)\n\n BUFFER_SIZE = params['BUFFER_SIZE']\n BATCH_SIZE = params['BATCH_SIZE']\n EPOCHS = params['EPOCHS']\n noise_dimension = params['noise_dimension']\n num_examples_to_generate = params['num_examples_to_generate_train']\n image_width = params['image_width']\n image_height = params['image_height']\n seed = tf.random.normal([num_examples_to_generate, noise_dimension]) # generates random normal distribution for seed\n\n # Batch and shuffle the data\n train_dataset = tf.data.Dataset.from_tensor_slices(train_images).shuffle(BUFFER_SIZE).batch(BATCH_SIZE)\n\n generator = build_generator(image_width, image_height, BATCH_SIZE)\n discriminator = build_discriminator((image_height, image_width, 3))\n\n train(train_dataset, seed, EPOCHS, BATCH_SIZE, noise_dimension, generator, discriminator, generator_loss,\n discriminator_loss, generator_optimizer, discriminator_optimizer)\n\n\n# create_gif('gif/test.gif')\n\ndef generate(n):\n load_path = os.path.join(PROJECT_ABSOLUTE_PATH, 'saved_generator')\n generator = load_model(load_path, compile=False)\n\n with open(os.path.join(CONFIG_ABSOLUTE_PATH, 'params.json'), 'r') as param_file:\n params = json.load(param_file)\n\n noise_dimension = params['noise_dimension']\n num_examples_to_generate = params['num_examples_to_generate_train']\n\n for i in range(n):\n seed = tf.random.normal([num_examples_to_generate, noise_dimension])\n generate_and_save_images(generator, seed, i, training=False)\n\n\nif __name__ == '__main__':\n train_or_generate = sys.argv[1]\n if train_or_generate == 'train':\n gan_train()\n elif train_or_generate == 'generate':\n number_to_generate = int(sys.argv[2])\n generate(number_to_generate)\n else:\n raise ValueError('Parameters not recognized')\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2499,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"527841970","text":"__author__ = 'ebruma'\n\n#Benutzerschnittstelle gemäss Kapitel 3 aus 2_1_konzeptbericht_Olivier Mingard_Marc Brünisholz.docx\n\nimport sys\nimport os\n\n#Darin wird Konfiguration gespeichert als Key wird der Parameterbuchstabe verwendet.\nconfig = {\"d\":\"0\",\n \"i\":\"./training_solutions_rev1.csv\",\n \"b\":\"./images\",\n \"o\":\"./output\",\n \"p27\":\"C:\\Python27\\python.exe\"}\n\nhelpText = \"\"\"\nHelp--------------------------------------------------------------------------------------------------------------------\n./ meint das aktuelle Arbeitsverzeichnis.\n\nMögliche Parameter:\n -d <0 oder 1> Bestimmt, ob der Debugmodus angeschaltet sein soll. Im Debumodus werden Zwischenresultate\n der Bildbearbeitung im Outputordner abgespeichert und während der Abarbeitung in Fenstern\n angezeigt. Standard: {d}\n\n -i Pfad zur CSV-Datei, welche in Spalte 0 die ID und in den Spalten 16-18\n die Häufigkeit für die Klassifizierungen 7.1, 7.2 und 7.3 gemäss\n https://www.kaggle.com/c/galaxy-zoo-the-galaxy-challenge/details/the-galaxy-zoo-decision-tree\n enthält. Die erste Zeile wird ignoriert. Es werden nur die Galaxien berücksichtigt,\n die für die Klassifizierung 1.1 (Spalte 1, smooth) eine Wahrscheinlichkeit von mehr als\n 50% bzw. 0.5 aufweisen. Galaxien, die zu mehr als 1% der Klassifizierung 8.6 (Spalte 24, merger)\n zugeordnet wurden, werden ebenfalls ignoriert. Standardpfad: {i}\n\n -t Wenn dieser Parameter verwendet wird, findet kein Machine Learning statt. Stattdessen\n wird das Bild analysiert und das Resultat in der Form \"number, number, number\" ausgegeben, wobei\n number eine Fliesskommazahl ist. Die drei Zahlen sind Vorschläge für die Klassifizierungen 7.1,\n 7.2 und 7.3. Zu Beachten: Die Galaxy sollte etwa mittig ausgerichtet sein, gleichmässig und\n rundlich sein.\n\n -b Der Verzeichnispfad zu den Bildern, welche in der Form .jpg vorliegen. Standard: {b}\n\n -o Pfad zu Ordner, wo erzeugte Dateien gespeichert werden sollen: CSV-Datei mit ID-Kennzahl\n Zuordnung, Machine Learning Zustand und evt. Debugdateien. Standard: {o}\n\n -p27 Pfad zu python.exe der Version 2.7 (OpenCV ist zur Zeit der Erstellung dieses Skripts noch nicht\n für Python 3 veröffentlicht worden). Standardpfad: {p27}\n\n -h Gibt einen Hilfetext aus. Skript wird nicht ausgeführt.\n\"\"\".format(**config)\n\n\ndef runWithArguments(python27, imagesPath, pathCSVsolutions):\n import subprocess\n subprocess.call(\"\\\"\"+python27 + \"\\\"\"+ \" galaxyCutScale.py \\\"\" + imagesPath + \"\\\" \" + \"\\\"\" + pathCSVsolutions + \"\\\"\")\n print(\"Alle Bilder wurden zugeschnitten und im Ordner ./Zugeschnitten gespeichert. Es werden nun Kennzahlen \"+\n \"zur Eigenschaft Rundlichkeit erzeugt.\")\n\n subprocess.call(python27 + \" berechneKennzahlen.py \")\n\n import lernen\n\n #Bilder analysieren\n lernen.learnFromData(pathCSVsolutions)\n\ndef run():\n import subprocess\n print(\"\\\"\"+config[\"p27\"]+ \"\\\"\" + \" galaxyCutScale.py \\\"\" + config[\"b\"] + \"\\\"\")\n subprocess.call(\"\\\"\"+config[\"p27\"]+ \"\\\"\" + \" galaxyCutScale.py \\\"\" + config[\"b\"] + \"\\\"\")\n print(\"Alle Bilder wurden zugeschnitten und im Ordner ./Zugeschnitten gespeichert. Es werden nun Kennzahlen \"+\n \"zur Eigenschaft Rundlichkeit erzeugt.\")\n\n subprocess.call(\"\\\"\"+config[\"p27\"]+ \"\\\"\" + \" berechneKennzahlen.py \")\n\n import lernen\n\n #Bilder analysieren\n lernen.learnFromData(config[\"i\"])\n # import kennzahl\n # for bild in os.listdir(\"Zugeschnitten\"):\n # kennzahl, egal = kennzahl.kennzahl(id,True)\n\n # print(\"Berechnete Kennzahl von Bild \" + id + \" :\" + kennzahl)\n\ndef analyse():\n import subprocess\n print(\"Bild \" + config[\"t\"] + \" wird zugeschnitten\")\n subprocess.call(config[\"p27\"] + \" galaxyCutScale.py \" + config[\"t\"])\n import ntpath\n\n print(subprocess.call(config[\"p27\"] + \" berechneKennzahlen.py Zugeschnitten/\"+ntpath.basename(config[\"t\"])))\n\n\n\nclass UnknownParameter(Exception):\n def __init__(self, parameter):\n self.message = \"Parameter \" + parameter + \" unbekannt.\"\n\nclass FileNotExistent(Exception):\n def __init__(self, filePath):\n self.message = \"Fehler: Die Datei \" + filePath + \" existiert nicht.\"\n\nclass DirNotExistent(Exception):\n def __init__(self, dirPath):\n self.message = \"Fehler: Das Verzeichnis \" + dirPath + \" existiert nicht.\"\n\n#Die Werte dieser Parameter sind Verzeichnisse, die auf Existenz prüft werden sollen\npruefVerzeichnisExist = [\"b\"]\n\n#Die Werte dieser Parameter sind Dateien, die auf Existenz geprüft werden sollen\npruefDateienAufExistenz = [\"i\",\"t\"]\n\n#Warnung wenn Verzeichnis nicht existent, wird aber neu angelegt\nwarnungWennNichtExistent = [\"o\"]\n\nerlaubteParameter_name = {\"d\":\"Debugausgabe\",\n \"i\":\"Inputdatei\",\n \"t\":\"Bildklassifizierung\",\n \"b\":\"Bildverzeichnis\",\n \"o\":\"Outputverzeichnis\",\n \"h\":\"Hilfe\",\n \"p27\":\"Pfad zu Python2.7\"}\n\n#Zum überprüfen, ob ein Parameter mehrfach angegeben wurde\nverwendeteParameter = []\n\nif __name__ == \"__main__\":\n #Wird ausgeführt, wenn thegalaxychallenge direkt in der Kommandozeile ausgeführt wird.\n\n #Wenn keine Parameter angegeben wurden\n if(len(sys.argv) == 1):\n print(\"Es werden Standardeinstellungen benutzt. \\\"thegalaxychallenge -h\\\" gibt \"+\n \"Informationen zur Änderung der Einstellungen aus.\")\n run()\n\n #Wenn -h als Parameter enthalten ist, wird nur Hilfetext angezeigt\n if(\"-h\" in sys.argv):\n print(helpText)\n else:\n #Parameter einlesen\n try:\n for i in range(1, len(sys.argv),2):\n originalParameter = sys.argv[i]\n\n #Von Parameter Symbol \"-\" wegschneiden\n parameter = originalParameter[1:]\n value = sys.argv[i+1]\n\n #Parameter auf Gültigkeit prüfen\n if parameter in erlaubteParameter_name.keys():\n print(\"Parameter -\"+parameter+\" für \"+erlaubteParameter_name[parameter]+\" erkannt. Wert: \" + value)\n\n if parameter in pruefDateienAufExistenz:\n if not os.path.exists(value):\n raise FileNotExistent(value)\n\n if parameter in pruefVerzeichnisExist:\n if not os.path.isdir(value):\n raise DirNotExistent(value)\n\n if parameter in warnungWennNichtExistent:\n if not os.path.isdir(value):\n print(\"Warnung: Verzeichnis \" + value + \" existiert nicht, würde aber neu angelegt.\")\n\n if parameter in verwendeteParameter:\n print(\"Warnung: Parameter \" + originalParameter + \" schon verwendet. Wert wird überschrieben.\")\n\n #In Konfiguration übernehmen\n config[parameter] = value\n\n verwendeteParameter.append(parameter)\n else:\n raise UnknownParameter(originalParameter)\n\n #Falls ein Bild zur Analyse angegeben wurde\n if \"t\" in config:\n analyse()\n else:\n run()\n\n except (UnknownParameter, IndexError) as e:\n if type(e) == UnknownParameter:\n print(e.message)\n elif type(e) == IndexError:\n print(\"Fehler: Parameteranzahl ungültig.\")\n\n #Wenn es einen Fehler gibt, weil die Parameter falsch sind, soll Hilfetext angezeigt werden.\n print(helpText)\n except (FileNotExistent, DirNotExistent) as e:\n print(e.message)\n\n\n\n\n\n\n\n\n\n","sub_path":"thegalaxychallenge.py","file_name":"thegalaxychallenge.py","file_ext":"py","file_size_in_byte":8104,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"218002062","text":"import pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nsns.set(style=\"ticks\")\n\n\ndef scatterMatrix(data, labels, count=5):\n '''Use seaborn to produce a pairplot of columns\n\n data: numpy array of data\n labels: numpy array of labels\n count: number of columns to scatter (larger will result in slower)\n '''\n # convert to dataframe, limit number of columns shown for time reasons\n df = pd.DataFrame(data[:,:count])\n\n # use labels as class\n df[\"labels\"]=labels\n\n # pairplot\n sns.pairplot(df,hue='labels')\n\n # show plot\n plt.show()\n\ndef correlationHeatmap(data):\n '''Use seaborn to produce a heatmap of the columns' correlation\n\n data: numpy array of data\n '''\n # convert to dataframe\n df = pd.DataFrame(data)\n \n # heatmap of correlations\n sns.heatmap(df.corr())\n\n # show plot\n plt.show()\n","sub_path":"engi1006/advanced/sb.py","file_name":"sb.py","file_ext":"py","file_size_in_byte":870,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"81462752","text":"# -*- coding: utf-8 -*-\n# model.py\n# author: Antoine Passemiers\n\nfrom archmm.ann.layers import *\n\nimport numpy as np\nfrom operator import mul\nfrom functools import reduce\n\n\nclass Optimizer:\n\n def __init__(self, learning_rate=.1, momentum=.2):\n self.learning_rate = learning_rate\n self.momentum = momentum\n self.previous_grad = None\n self.gradient_fragments = list()\n\n def update(self):\n gradient = list()\n for _, _, fragments in self.gradient_fragments:\n for fragment in fragments:\n gradient.append(fragment.flatten(order='C'))\n gradient = np.concatenate(gradient)\n\n delta = self.learning_rate * gradient\n if self.momentum > 0:\n if self.previous_grad is not None:\n delta += self.momentum * self.previous_grad\n self.previous_grad = delta\n\n self.update_layers(delta)\n self.gradient_fragments = list()\n\n def update_layers(self, delta):\n cursor = 0\n for src_layer, layer_param_shapes, _ in self.gradient_fragments:\n layer_fragments = list()\n for fragment_shape in layer_param_shapes:\n n_elements = reduce(mul, fragment_shape)\n fragment = delta[cursor:cursor+n_elements]\n layer_fragments.append(fragment.reshape(fragment_shape, order='C'))\n cursor += n_elements\n src_layer.update_parameters(tuple(layer_fragments))\n\n def add_gradient_fragments(self, src_layer, fragments):\n if not isinstance(fragments, (tuple, list)):\n fragments = [fragments]\n layer_param_shapes = list()\n for fragment in fragments:\n layer_param_shapes.append(fragment.shape)\n self.gradient_fragments.append((src_layer, layer_param_shapes, fragments))\n\n\nclass CrossEntropy:\n\n def __init__(self, epsilon=1e-15):\n self.epsilon = epsilon\n\n def eval(self, y, y_hat):\n indices = np.argmax(y, axis=1).astype(np.int)\n predictions = y_hat[np.arange(len(y_hat)), indices]\n log_predictions = np.log(np.maximum(predictions, self.epsilon))\n return -np.mean(log_predictions)\n\n def grad(self, y, y_hat):\n return y_hat - y\n\n\nclass NeuralStackClassifier:\n\n def __init__(self, optimizer=Optimizer(1)):\n self.layers = list()\n self.optimizer = optimizer\n self.cost = CrossEntropy()\n\n def add(self, layer):\n self.layers.append(layer)\n\n def eval(self, X):\n out = X\n for layer in self.layers:\n out = layer.forward(out)\n return out\n \n def fit(self, X, y, max_n_iter=100):\n print()\n for k in range(max_n_iter):\n batch_X, batch_y = X, y # TODO: batching\n y_hat = self.eval(X)\n\n print(self.cost.eval(batch_y, y_hat))\n\n signal = self.cost.grad(batch_y, y_hat)\n\n for layer in reversed(self.layers):\n signal, gradient = layer.backward(signal)\n if gradient is not None:\n self.optimizer.add_gradient_fragments(layer, gradient)\n self.optimizer.update()\n","sub_path":"archmm/ann/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":3148,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"315659894","text":"import numpy as np\nimport matplotlib\nmatplotlib.use('TkAgg')\nimport matplotlib.pyplot as plt\n# from IPython import embed\n\ndef draw(tmp, model_name):\n print(model_name)\n print(np.mean(tmp))\n mean_batch = 100\n train_x, reward = zip(*tmp)\n train_x = list(map(int, train_x))\n print(np.mean(reward))\n reward = [np.mean(reward[i:i+mean_batch]) for i in range(len(reward))]\n plt.plot(train_x[:-mean_batch], reward[:-mean_batch], '-', label=model_name)\n\n\n\ndef draw_2(tmp, model_name):\n mean_batch = 100\n print(model_name)\n plt.figure(figsize=(20,10))\n train_x, reward = zip(*tmp)\n train_x = list(map(int, train_x))\n print(np.mean(reward))\n reward = [np.mean(reward[i:i+mean_batch]) for i in range(len(reward))]\n plt.plot(train_x[:-mean_batch], reward[:-mean_batch], '-', label=model_name)\n\n\n tmp2 = np.load(\"dqn_0.99.npy\")\n train_x, reward = zip(*tmp2)\n print(len(train_x))\n train_x = list(map(int, train_x))\n print(np.mean(reward))\n reward = [np.mean(reward[i:i+mean_batch]) for i in range(len(reward))]\n plt.plot(train_x[:-mean_batch], reward[:-mean_batch], '-', label=\"dqn\")\n\n plt.title(\"Learning Curves of DQN\")\n plt.xlabel(\"Steps\")\n plt.ylabel(\"Reward\")\n\n plt.legend(loc=4)\n plt.savefig(\"./pic/{}.png\".format(model_name))\n plt.clf()\n\ndef draw_4():\n mean_batch = 100\n plt.figure(figsize=(20,10))\n model_name = \"dqn\"\n tmp2 = np.load(\"dqn_0.99.npy\")\n train_x, reward = zip(*tmp2)\n print(len(train_x))\n train_x = list(map(int, train_x))\n print(np.mean(reward))\n reward = [np.mean(reward[i:i+mean_batch]) for i in range(len(reward))]\n plt.plot(train_x[:-mean_batch], reward[:-mean_batch], '-', label=\"dqn\")\n\n plt.title(\"Learning Curves of DQN\")\n plt.xlabel(\"Steps\")\n plt.ylabel(\"Reward\")\n\n plt.legend(loc=4)\n plt.savefig(\"./pic/{}.png\".format(model_name))\n plt.clf()\n\ndef draw_3(tmp, model_name):\n print(model_name)\n print(np.max(tmp))\n mean_batch = 100\n reward = [np.mean(tmp[i:i+mean_batch]) for i in range(len(tmp)-mean_batch)]\n train_x = [i for i in range(1, len(reward)+1)]\n\n plt.figure(figsize=(20,10))\n plt.plot(train_x, reward, '-', label=model_name)\n\n plt.title(\"Learning Curves of PG\")\n plt.xlabel(\"Episodes\")\n plt.ylabel(\"Reward\")\n\n plt.legend()\n plt.savefig(\"./pic/{}.png\".format(model_name))\n plt.clf()\n\n\ndraw_3(np.load(\"pg.npy\"), \"pg\")\n\ndraw_4()\n\ndraw_2(np.load(\"duel_dqn_0.99.npy\"), \"duel dqn\")\ndraw_2(np.load(\"double_dqn_0.99.npy\"), \"double dqn\")\ndraw_2(np.load(\"noisy_dqn_0.99.npy\"), \"noisy dqn\")\ndraw_2(np.load(\"prioritized_dqn_0.99.npy\"), \"prioritized dqn\")\n\n\n\nplt.figure(figsize=(20,10))\ndraw(np.load(\"dqn_0.99.npy\"), \"dqn with gamma = 0.99\")\ndraw(np.load(\"dqn_0.95.npy\"), \"dqn with gamma = 0.95\")\ndraw(np.load(\"dqn_0.9.npy\"), \"dqn with gamma = 0.9\")\ndraw(np.load(\"dqn_0.85.npy\"), \"dqn with gamma = 0.85\")\nplt.title(\"Learning Curves of DQN\")\nplt.xlabel(\"Steps\")\nplt.ylabel(\"Reward\")\n\nplt.legend(loc=4)\nplt.savefig(\"./pic/dqn_gamma_compare.png\")\nplt.clf()","sub_path":"hw3/b05902002/improvement/draw.py","file_name":"draw.py","file_ext":"py","file_size_in_byte":3040,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"195579762","text":"import json\nimport chess\nimport chess.engine\nfrom models.drawTacticsModel import DrawTacticsModel\n\nengine = chess.engine.SimpleEngine.popen_uci(\"./stockfish_20090216_x64\")\n\n\ndef reevaluate_tactics(file_path):\n new_tactics = []\n with open(file_path) as json_file:\n json_tactics = json.load(json_file)\n for tactic in json_tactics:\n print(len(new_tactics)\n , \" done.\")\n board = chess.Board(tactic[\"fen\"])\n score = engine.analyse(board, chess.engine.Limit(time=1))['score']\n if score.is_mate() is False and score.relative.cp == 0:\n new_tactics.append(DrawTacticsModel(board.fen()))\n return new_tactics\n","sub_path":"tactics/drawTacticsFinder/util/reevaluateTactics.py","file_name":"reevaluateTactics.py","file_ext":"py","file_size_in_byte":697,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"555939552","text":"from PyQt5.QtCore import Qt\nfrom PyQt5.QtWidgets import QMessageBox\n\nfrom Business.SketchActions import create_all_areas, find_all_areas, create_area, create_composite_area\nfrom GUI.init import plugin_initializers\n\nfrom GUI.Ribbon.RibbonButton import RibbonButton\n\n\nclass SketchGenerateAreas():\n\tdef __init__(self, main_window):\n\t\tself._main_window = main_window\n\t\tself._generate_areas_action = None\n\t\tself._create_area_action = None\n\t\tself._create_composite_area_action = None\n\t\tself._states = main_window.states\n\t\tself._sketch_editor_view = main_window.sketch_editor_view\n\t\tself._sketch_editor_view.add_mouse_press_event_handler(self.on_mouse_press)\n\t\t# self._sketch_editor_view.add_mouse_move_event_handler(self.on_mouse_move)\n\t\tself._sketch_editor_view.add_escape_event_handler(self.on_escape)\n\t\tself._states.generate_areas = False\n\t\tself._states.create_area = False\n\t\tself._states.create_composite_area = False\n\t\tself._base_area = None\n\t\tself._subtracted_areas = []\n\t\tself.init_ribbon()\n\n\tdef init_ribbon(self):\n\t\tself._generate_areas_action = self._main_window.add_action(\"Create\\nAreas\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t \"createareas\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t \"Create areas from all edges\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t True,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t self.on_generate_areas)\n\t\tself._create_area_action = self._main_window.add_action(\"Create\\narea\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\"createarea\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\"Create area from selected edges\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tTrue,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tself.on_create_area,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tcheckable=True)\n\t\tself._create_composite_area_action = self._main_window.add_action(\"Create\\nComp. Area\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\"addcompositearea\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\"Create composite area from existing areas\",\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tTrue,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tself.on_create_composite_area,\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\tcheckable=True)\n\t\tribbon = self._main_window.ribbon\n\t\tsketch_tab = ribbon.get_ribbon_tab(\"Sketch\")\n\t\tinsert_pane = sketch_tab.get_ribbon_pane(\"Insert\")\n\t\tinsert_pane.add_ribbon_widget(RibbonButton(insert_pane, self._generate_areas_action, True))\n\t\tinsert_pane.add_ribbon_widget(RibbonButton(insert_pane, self._create_area_action, True))\n\t\tinsert_pane.add_ribbon_widget(RibbonButton(insert_pane, self._create_composite_area_action, True))\n\n\tdef on_generate_areas(self):\n\t\tself._sketch_editor_view.on_escape()\n\t\tsketch = self._sketch_editor_view.sketch\n\t\tview = self._sketch_editor_view\n\t\tif sketch is None:\n\t\t\treturn\n\n\t\tgo = False\n\t\tif len(sketch.get_areas()) > 0:\n\t\t\ttxt = \"This will replace all existing areas with new areas generated from the edges.\"\n\t\t\ttxt += \"Are you sure you want to do this?\"\n\t\t\tret = QMessageBox.warning(self._main_window, \"Create areas?\", txt, QMessageBox.Yes | QMessageBox.Cancel)\n\t\t\tif ret == QMessageBox.Yes:\n\t\t\t\tgo = True\n\t\telse:\n\t\t\tgo = True\n\t\tif go:\n\t\t\tcreate_all_areas(self._main_window.document, sketch)\n\t\t\tview.update()\n\n\tdef on_create_area(self):\n\t\tself._sketch_editor_view.on_escape()\n\t\tif self._sketch_editor_view.sketch is None:\n\t\t\treturn\n\t\tself._sketch_editor_view.setCursor(Qt.CrossCursor)\n\t\tself._states.select_edges = True\n\t\tself._states.create_area = True\n\t\tself._main_window.update_ribbon_state()\n\t\tdoc = self._main_window.document\n\t\tdoc.set_status(\"Click on edges to create area. Hold CTRL to multi select.\", 0, True)\n\n\tdef on_create_composite_area(self):\n\t\tself._sketch_editor_view.on_escape()\n\t\tif self._sketch_editor_view.sketch is None:\n\t\t\treturn\n\t\tself._sketch_editor_view.setCursor(Qt.CrossCursor)\n\t\tself._states.select_area = True\n\t\tself._states.create_composite_area = True\n\t\tself._base_area = None\n\t\tself._subtracted_areas = []\n\t\tself._main_window.document.set_status(\"Select base area for new composite area\", 0, True)\n\t\tself._main_window.update_ribbon_state()\n\n\tdef check_edge_loop(self):\n\t\tview = self._sketch_editor_view\n\t\tdoc = self._main_window.document\n\t\tsketch = view.sketch\n\t\tbranches = find_all_areas(view.selected_edges)\n\t\tfor branch in branches:\n\t\t\tif branch['enclosed']:\n\t\t\t\tcreate_area(sketch, branch)\n\t\t\t\tview.on_escape()\n\t\t\t\tview.update()\n\t\t\t\tbreak\n\n\tdef on_mouse_move(self, scale, x, y):\n\t\tpass\n\n\tdef on_mouse_press(self, scale, x, y):\n\t\tif self._states.create_area:\n\t\t\tself.check_edge_loop()\n\n\t\tif self._states.create_composite_area:\n\t\t\tview = self._sketch_editor_view\n\t\t\tdoc = self._main_window.document\n\t\t\tsketch = view.sketch\n\t\t\tif view.area_hover is not None and self._base_area is None:\n\t\t\t\tself._base_area = view.area_hover\n\t\t\t\tdoc.set_status(\"Select areas to subtract. Hold CTRL to select multiple areas.\", 33, True)\n\t\t\telif view.area_hover is not None and self._base_area is not None:\n\t\t\t\tif view.area_hover != self._base_area:\n\t\t\t\t\tself._subtracted_areas.append(view.area_hover)\n\t\t\t\t\tif self._states.multi_select:\n\t\t\t\t\t\tpass\n\t\t\t\t\telse:\n\t\t\t\t\t\tcreate_composite_area(sketch, self._base_area, self._subtracted_areas)\n\t\t\t\t\t\tview.update()\n\t\t\t\t\t\tview.on_escape()\n\n\tdef on_escape(self):\n\t\tself._base_area = None\n\t\tself._subtracted_areas = []\n\t\tself._states.create_composite_area = False\n\t\tself._states.create_area = False\n\n\tdef update_ribbon_state(self):\n\t\tself._create_area_action.setChecked(self._states.create_area)\n\t\tself._create_composite_area_action.setChecked(self._states.create_composite_area)\n\n\t@staticmethod\n\tdef initializer(main_window):\n\t\treturn SketchGenerateAreas(main_window)\n\n\nplugin_initializers.append(SketchGenerateAreas.initializer)\n","sub_path":"GUI/Plugins/SketchAreas.py","file_name":"SketchAreas.py","file_ext":"py","file_size_in_byte":5509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"409438909","text":"#!/usr/bin/python\n# -*- coding: utf8 -*-\n#\n\nimport shutil\nfrom collections import namedtuple\nfrom ansible import constants\nfrom ansible.parsing.dataloader import DataLoader\nfrom ansible.playbook.play import Play\nfrom ansible.executor.task_queue_manager import TaskQueueManager\nfrom ansible.executor.playbook_executor import PlaybookExecutor\nfrom ansible.plugins.callback import CallbackBase\nfrom ansible.inventory.manager import InventoryManager\nfrom ansible.vars.manager import VariableManager\nimport ansible.constants as C\nimport os\nimport json\n\nBASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\nclass AdHocResultsCollector(CallbackBase):\n def __init__(self, *args, **kwargs):\n super(AdHocResultsCollector, self).__init__(*args, **kwargs)\n self.host_ok = {}\n self.host_unreachable = {}\n self.host_failed = {}\n\n def v2_runner_on_unreachable(self, result):\n self.host_unreachable[result._host.get_name()] = result\n\n def v2_runner_on_ok(self, result, *args, **kwargs):\n self.host_ok[result._host.get_name()] = result\n\n def v2_runner_on_failed(self, result, *args, **kwargs):\n self.host_failed[result._host.get_name()] = result\n\nclass PlaybookResultsCollector(CallbackBase):\n CALLBACK_VERSION = 2.0\n def __init__(self, *args, **kwargs):\n super(PlaybookResultsCollector, self).__init__(*args, **kwargs)\n self.task_ok = {}\n self.task_skipped = {}\n self.task_failed = {}\n self.task_status = {}\n self.task_unreachable = {}\n\n def v2_runner_on_ok(self, result, *args, **kwargs):\n self.task_ok[result._host.get_name()] = result\n\n def v2_runner_failed(self, result, *args, **kwargs):\n self.task_failed[result_host.get_name()] = result\n\n def v2_runner_on_unreachable(self, result):\n self.task_unreachable[result._host.get_name()] = result\n\n def v2_runner_on_skipped(self, result):\n self.task_ok[result._host.get_name()] = result\n\n def v2_playbook_on_stats(self, stats):\n hosts = sorted(stats.processed.keys())\n for h in hosts:\n t = stats.summarize(h)\n self.task_status[h] = {\n \"ok\": t['ok'],\n \"changed\": t['changed'],\n \"unreachable\": t['unreachable'],\n \"skipped\": t['skipped'],\n \"failed\": t['failures']\n }\n\nclass AnsibleRunner:\n def __init__(self, redisKey=None, logId=None, *args, **kwargs):\n self.sources = \"{0}/utils/hostlist\".format(BASE_DIR)\n self.loader = DataLoader()\n self.inventory = InventoryManager(loader=self.loader, sources=self.sources)\n self.variable_manager = VariableManager(loader=self.loader, inventory=self.inventory)\n self.passwords = dict(sshpass=None, becomepass=None)\n self.callback = None\n self.__initializeData()\n self.results_raw = {}\n self.redisKey = redisKey\n self.logId = logId\n\n def __initializeData(self):\n Options = namedtuple('Options', ['listtags', 'listtasks', 'listhosts', 'syntax',\n 'connection','module_path', 'forks', 'remote_user',\n 'private_key_file', 'timeout','ssh_common_args',\n 'ssh_extra_args', 'sftp_extra_args','scp_extra_args',\n 'become', 'become_method', 'become_user','verbosity',\n 'check', 'extra_vars', 'playbook_path', 'passwords',\n 'diff', 'gathering', 'remote_tmp',])\n\n self.options = Options(listtags=False,listtasks=False,listhosts=False,syntax=False,timeout=60,\n connection='smart',module_path=None ,forks=10,remote_user='root',private_key_file=None,\n ssh_common_args=\"\",ssh_extra_args=\"\",sftp_extra_args=\"\",scp_extra_args=\"\",\n become=None,become_method=None,become_user=None,verbosity=None,extra_vars=[],\n check=False,playbook_path=None,passwords=None,diff=False,\n gathering='implicit',remote_tmp='/tmp/.ansible')\n \n def AdHocRunner(self, host_list, module_name, module_args):\n play_source = dict(\n name = \"Ansible Play\",\n hosts = host_list,\n gather_facts = 'no',\n tasks = [dict(action=dict(module=module_name, args=module_args))]\n )\n play = Play().load(play_source, variable_manager=self.variable_manager, loader=self.loader)\n tmq = None\n self.callback = AdHocResultsCollector()\n import traceback\n try:\n tqm = TaskQueueManager(\n inventory = self.inventory,\n variable_manager = self.variable_manager,\n loader = self.loader,\n options = self.options,\n passwords = self.passwords,\n stdout_callback = \"minimal\",\n )\n tqm._stdout_callback = self.callback\n constants.HOST_KEY_CHECKING = False\n tqm.run(play)\n except Exception as err:\n print(traceback.print_exc())\n finally:\n if tqm is not None:\n tqm.cleanup()\n\n shutil.rmtree(C.DEFAULT_LOCAL_TMP, True)\n\n def PlaybookRunner(self, playbook_path, extra_vars=None):\n try:\n self.callback = PlaybookResultsCollector()\n if extra_vars:\n self.variable_manager.extra_vars = extra_vars\n executor = PlaybookExecutor(\n playbooks=[playbook_path], inventory=self.inventory,\n variable_manager=self.variable_manager, loader=self.loader,\n options=self.options, passwords=self.passwords,\n )\n executor._tqm._stdout_callback = self.callback\n constants.HOST_KEY_CHECKING = False\n executor.run()\n except Exception as err:\n return False\n\n def get_AdHoc_result(self):\n self.results_raw = {'success': {}, 'failed': {}, 'unreachable': {}}\n for host, result in self.callback.host_ok.items():\n hostvisiable = host.replace('.', '_')\n self.results_raw['success'][hostvisiable] = result._result\n\n for host, result in self.callback.host_failed.items():\n hostvisiable = host.replace('.', '_')\n self.results_raw['failed'][hostvisiable] = result._result\n\n for host, result in self.callback.host_unreachable.items():\n hostvisiable = host.replace('.', '_')\n self.results_raw['unreachable'][hostvisiable] = result._result\n\n return self.results_raw\n\n def get_playbook_result(self):\n self.results_raw = {'skipped': {}, 'failed': {},'ok':{}, 'status': {}, 'unreachable': {}, 'changed': {}}\n for host, result in self.callback.task_ok.items():\n self.results_raw['ok'][host] = result\n\n for host, result in self.callback.task_failed.items():\n self.results_raw['failed'][host] = result\n\n for host, result in self.callback.task_status.items():\n self.results_raw['status'][host] = result\n\n for host, result in self.callback.task_skipped.items():\n self.results_raw['skipped'][host] = result\n\n for host, result in self.callback.task_unreachable.items():\n self.results_raw['unreachable'][host] = result\n\n return self.results_raw\n\n#if __name__ == '__main__':\n #rbt = AnsibleRunner()\n #Ansible Adhoc\n #rbt.AdHocRunner(host_list=['win'],module_name='win_ping',module_args=\"\")\n #data = json.dumps(rbt.get_AdHoc_result())\n #Ansible playbook\n #rbt.PlaybookRunner(playbook_path='{0}/utils/admin.yml'.format(BASE_DIR))\n #data = rbt.get_playbook_result()\n #print(data['ok'])\n","sub_path":"release/ansible_api.py","file_name":"ansible_api.py","file_ext":"py","file_size_in_byte":8068,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"426605093","text":"from collections import defaultdict\nimport math\n\n\nclass Solution:\n def minAreaFreeRect(self, points):\n \"\"\"\n :type points: List[List[int]]\n :rtype: float\n \"\"\"\n N = len(points)\n relations = [defaultdict(set) for i in range(N)]\n ijMap = dict()\n for i, (x0, y0) in enumerate(points):\n for j, (x1, y1) in enumerate(points[i + 1:], i + 1):\n incline = self.getIncline(x0, y0, x1, y1)\n relations[i][incline].add(j)\n relations[j][incline].add(i)\n ijMap[(i, j)] = ijMap[(j, i)] = incline\n minArea = math.inf\n for i0 in range(N):\n for i1 in range(i0 + 1, N):\n inc01 = ijMap[(i0, i1)]\n inc02 = self.getRevIncline(*inc01)\n i3Set = relations[i1][inc02]\n for i2 in relations[i0][inc02]:\n if not i3Set.isdisjoint(relations[i2][inc01]):\n minArea = min(\n minArea, self.calArea(points, i0, i1, i2))\n return minArea if minArea != math.inf else 0\n\n def calArea(self, points, i0, i1, i2):\n x0, y0 = points[i0]\n x1, y1 = points[i1]\n x2, y2 = points[i2]\n return math.sqrt(((x1 - x0)*(x1 - x0) + (y1 - y0) * (y1 - y0)) * ((x2 - x0) * (x2 - x0) + (y2 - y0) * (y2 - y0)))\n\n def getRevIncline(self, dy, dx):\n if dy == math.inf:\n return 0, 1\n if dy == 0:\n return math.inf, 1\n return (-dx, dy) if dy > 0 else (dx, -dy)\n\n def getIncline(self, x0, y0, x1, y1):\n if x0 == x1:\n return math.inf, 1\n if y0 == y1:\n return 0, 1\n dx, dy = x1 - x0, y1 - y0\n cf = self.getCommonFactor(dx, dy)\n dx, dy = dx // cf, dy // cf\n return (dy, dx) if dx > 0 else (-dy, -dx)\n\n def getCommonFactor(self, a, b):\n a, b = abs(a), abs(b)\n a = a % b\n while a != 0:\n b = b % a\n a, b = b, a\n return b\n\n\nsol = Solution()\n# ret = sol.minAreaFreeRect([[1, 2], [2, 1], [1, 0], [0, 1]])\nret = sol.minAreaFreeRect([[0, 3], [1, 2], [3, 1], [1, 3], [2, 1]])\nprint(ret)\n","sub_path":"src/minimum-area-rectangle-ii.py","file_name":"minimum-area-rectangle-ii.py","file_ext":"py","file_size_in_byte":2190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"297124316","text":"# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: https://doc.scrapy.org/en/latest/topics/item-pipeline.html\nimport logging\nfrom scrapy.pipelines.media import *\nfrom scrapy.pipelines.images import *\nfrom scrapy.pipelines.files import *\nfrom zzhlyx.settings import *\n\n\nclass ZzhlyxPipeline(FilesPipeline):\n\n def item_completed(self, results, item, info):\n # logger.info('results:', results)\n # logging.info(msg='pipelines------------------')\n # logging.info(msg=item)\n title = item['title']\n path = ''\n if title != '':\n path = FILES_STORE + title + '/'\n if not os.path.exists(path):\n os.mkdir(path)\n\n hash_path = results[0][1]['path']\n os.rename(FILES_STORE + hash_path, path + item['index'] + '.ts')\n item['path'] = path + item['index'] + '.ts'\n return item\n","sub_path":"Language/Python/spider/zzhlyx/zzhlyx/pipelines.py","file_name":"pipelines.py","file_ext":"py","file_size_in_byte":952,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"419261178","text":"# Definition for a binary tree node.\n# class TreeNode:\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n\nclass Solution:\n def trimBST(self, root: TreeNode, L: int, R: int) -> TreeNode:\n while root and not(L <= root.val <= R):\n if root.val < L:\n root = root.right \n if root.val > R:\n root = root.left \n stack = [root]\n while stack:\n node = stack[-1]\n if not node:\n stack.pop()\n continue\n update = 0\n if node.left and node.left.val < L:\n node.left = node.left.right\n update += 1\n if node.right and node.right.val > R:\n node.right = node.right.left\n update += 1\n if not update:\n stack.pop()\n stack.append(node.left)\n stack.append(node.right)\n return root\n ","sub_path":"Tree/Easy/669. Trim a Binary Search Tree - Iterative.py","file_name":"669. Trim a Binary Search Tree - Iterative.py","file_ext":"py","file_size_in_byte":1002,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"553614284","text":"import numpy as np\nimport h5py\nimport os\n\ndef load_dataset():\n dir_path = os.path.dirname(os.path.abspath(__file__))\n train_dataset = h5py.File(dir_path + '/datasets/train_catvnoncat.h5', \"r\")\n # 保存的是训练集里面的图像数据(本训练集有209张64x64的图像)。\n train_set_x_orig = np.array(train_dataset[\"train_set_x\"][:])\n # 保存的是训练集的图像对应的分类值(【0 | 1】,0表示不是猫,1表示是猫)。\n train_set_y_orig = np.array(train_dataset[\"train_set_y\"][:])\n\n test_dataset = h5py.File(dir_path + '/datasets/test_catvnoncat.h5', \"r\")\n # 保存的是测试集里面的图像数据(本训练集有50张64x64的图像)。\n test_set_x_orig = np.array(test_dataset[\"test_set_x\"][:])\n # 保存的是测试集的图像对应的分类值(【0 | 1】,0表示不是猫,1表示是猫)。\n test_set_y_orig = np.array(test_dataset[\"test_set_y\"][:])\n\n # 保存的是以bytes类型保存的两个字符串数据,数据为:[b’non-cat’ b’cat’]\n classes = np.array(test_dataset[\"list_classes\"][:])\n\n train_set_y_orig = train_set_y_orig.reshape((1, train_set_y_orig.shape[0]))\n test_set_y_orig = test_set_y_orig.reshape((1, test_set_y_orig.shape[0]))\n\n return train_set_x_orig, train_set_y_orig, test_set_x_orig, test_set_y_orig, classes","sub_path":"code/code1/lr_utils.py","file_name":"lr_utils.py","file_ext":"py","file_size_in_byte":1345,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"17620430","text":"import json\r\n\r\n\r\ndef write_json(data, filename):\r\n with open(filename, 'w') as addBook:\r\n json.dump(data, addBook, indent=4)\r\n\r\n\r\ndef elementExists(element): # Return True if passed Element is Present else return False\r\n with open('book.json') as addressBook:\r\n dataOnFile = json.load(addressBook)\r\n for datas in dataOnFile[\"personalDetail\"]:\r\n if datas.get(\"firstname\") == element:\r\n return True\r\n return False\r\n\r\n\r\ndef add(): # To add New Entry into the JSON File\r\n\r\n firstName = input(\"Enter Your First_Name: \")\r\n lastName = input(\"Enter Your Last_Name: \")\r\n address = input(\"Enter Your Address: \")\r\n city = input(\"Enter Your CITY: \")\r\n state = input(\"Enter Your State: \")\r\n zip = str(input(\"Enter your ZIP Number: \"))\r\n phoneNumber = str(input(\"Enter your Phone Number: \"))\r\n try:\r\n addressDetail = {\r\n \"firstname\": firstName,\r\n \"lastName\": lastName,\r\n \"address\": address,\r\n \"city\": city,\r\n \"state\": state,\r\n \"zip\": zip,\r\n \"phoneNumber\": phoneNumber\r\n }\r\n with open('book.json') as addressBook:\r\n dataOnFile = json.load(addressBook)\r\n temp = dataOnFile[\"personalDetail\"]\r\n temp.append(addressDetail)\r\n write_json(dataOnFile, 'book.json')\r\n print(\"Data Saved !!!\")\r\n except:\r\n addressDetail = {\r\n \"personalDetail\": [\r\n {\r\n \"firstname\": firstName,\r\n \"lastName\": lastName,\r\n \"address\": address,\r\n \"city\": city,\r\n \"state\": state,\r\n \"zip\": zip,\r\n \"phoneNumber\": phoneNumber}\r\n ]}\r\n write_json(addressDetail, 'book.json')\r\n print(\"Data Saved !!!\")\r\n\r\n\r\ndef search(element): # Search for Data Based on FirstName or MobileNumber or LastName of the Entry\r\n\r\n with open('book.json') as addressBook:\r\n dataOnFile = json.load(addressBook)\r\n for datas in dataOnFile[\"personalDetail\"]:\r\n if element == datas.get(\"firstname\") or element == datas.get(\"phoneNumber\") or element == datas.get(\"lastName\"):\r\n print(datas)\r\n\r\n\r\ndef delete(element): # Delete a Object from JSON file based on FirstName\r\n\r\n if elementExists(element) == False:\r\n print(\"Data Not Present\")\r\n return None\r\n with open('book.json') as addressBook:\r\n dataOnFile = json.load(addressBook)\r\n temp = []\r\n for datas in dataOnFile[\"personalDetail\"]:\r\n if element == datas.get(\"firstname\"):\r\n pass\r\n else:\r\n temp.append(datas)\r\n dictionary = {\"personalDetail\": temp}\r\n write_json(dictionary, 'book.json')\r\n print(\"Data Deleted\")\r\n\r\n\r\nsearch(\"Vishwas\")\r\nadd()\r\n","sub_path":"dataInventory.py","file_name":"dataInventory.py","file_ext":"py","file_size_in_byte":2843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"545683450","text":"def partition(arr, idx, low, high):\n pivot = arr[high]\n\n i = low - 1\n\n for j in range(low, high):\n if arr[j] < pivot:\n i += 1\n arr[i], arr[j] = arr[j], arr[i]\n idx[i], idx[j] = idx[j], idx[i]\n\n arr[i + 1], arr[high] = arr[high], arr[i + 1]\n idx[i + 1], idx[high] = idx[high], idx[i + 1]\n return i + 1\n\ndef quick_sort(arr, idx, low, high):\n if low < high:\n pi = partition(arr, idx, low, high)\n\n quick_sort(arr, idx, low, pi - 1)\n quick_sort(arr, idx, pi + 1, high)\n\n","sub_path":"quick_sort.py","file_name":"quick_sort.py","file_ext":"py","file_size_in_byte":547,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"755710","text":"import sys as s\n# Write your code here\n\"\"\"import math\ndef countDivisors(n):\n cnt = 0\n for i in range(1, (int)(math.sqrt(n)) + 1):\n if (n % i == 0):\n\n # If divisors are equal,\n # count only one\n if (n / i == i):\n cnt = cnt + 1\n else: # Otherwise count both\n cnt = cnt + 2\n\n print(cnt)\n\n\n\n\ndef query(l ,r ,list1):\n prod=1\n for i in range(list1.index(l),(list1.index(r)+1)):\n if 1<= list1[i] <= 10:\n prod*=list1[i]\n countDivisors(prod%(pow(10,9)+7))\n\n\n\n\nN, M = map(int, input().split())\nif 1 <= N <= (5 * pow(10, 5)) and 1 <= N <= (4 * pow(10, 5)):\n list1 = list(map(int, input().split()))\n for i in range(M):\n l, r = map(int, input().split())\n if 1 <= l <= r <= N:\n query(l, r, list1)\"\"\"\n\"\"\"def calculate(list1, n, k):\n list1.sort()\n if list1[0]>=k:\n print(0)\n else:\n print(k-list1[i])\"\"\"\n\n\n#t=int(input())\n\"\"\"if 1<=t<=5:\n for i in range(t):\n n,k=map(int,input().split())\n if 1<=n<=pow(10,5):\n list1=list(map(int,input().split()))\n calculate(list1,n,k)\"\"\"\n\n'''File CLassifier'''\nextension=['exe','pdf','docx','pptx','mdj','json']\ndef classifybasedonextension(directory):\n print('FILE TYPE', '\\t', 'FILE NAME', '\\t', 'FILE SIZE')\n for i in range(len(directory)):\n for j in range(len(extension)):\n if directory[i].split('.')[1]==extension[j]:\n print(directory[i].split('.')[1],'\\t\\t',directory[i].split('.')[0],'\\t\\t\\t',s.getsizeof(directory[i]))\n\n\n\nn=int(input())\ndirectory=[]\nfor i in range(n):\n directory.append(input())\n\nclassifybasedonextension(directory)\n\n\n\n\n\n\n\n\n\n","sub_path":"roughwork.py","file_name":"roughwork.py","file_ext":"py","file_size_in_byte":1759,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"234357354","text":"# Copyright (c) 2014-2015, NVIDIA CORPORATION. All rights reserved.\n\nimport os.path\nimport pickle\nimport numpy as np\nfrom collections import OrderedDict\n\nfrom bokeh.embed import components\nfrom bokeh.plotting import figure, hplot\nfrom bokeh.charts import Bar, HeatMap\nfrom bokeh.models import HoverTool, ColumnDataSource, TapTool, OpenURL\nfrom bokeh.models import Range1d\n\nfrom sklearn import metrics\nfrom sklearn.externals import joblib\nfrom scipy.stats import mode\n\nfrom digits.trial import tasks\nfrom digits import utils\nfrom digits.utils import subclass, override\nfrom ..job import ImageTrialJob\n\n# NOTE: Increment this everytime the pickled object changes\nPICKLE_VERSION = 1\n\n@subclass\nclass ImageClassificationTrialJob(ImageTrialJob):\n \"\"\"\n A Job that creates an image trial for a classification network\n \"\"\"\n\n def __init__(self, **kwargs):\n super(ImageClassificationTrialJob, self).__init__(**kwargs)\n self.pickver_job_trial_image_classification = PICKLE_VERSION\n \n # graph data\n self.confusion_matrix = None\n self.scores = None\n self.track_confusion_matrix = None\n self.track_scores = None\n self.class_labels = None\n self.hit_miss_results = None\n\n @override\n def job_type(self):\n return 'Image Classification Trial'\n \n \n # Plotting Functions \n def generate_hit_miss_results(self, class_labels):\n \n if self.learning_method == 'dnn':\n prob = joblib.load(self.experiment.path('ftrs_test/prob.jbl'))\n ypred = np.argmax(prob, axis=1) \n else:\n prob = joblib.load(self.path('prob.jbl'))\n ypred = joblib.load(self.path('ypred.jbl'))\n \n ytrue = joblib.load(self.path('test_label.jbl')) \n prob = np.max(prob, axis=1)\n \n self.hit_miss_results = {class_label: {'hit': [], 'miss': []} for class_label in class_labels}\n for class_label_index, class_label in enumerate(class_labels):\n \n bool_index = ytrue == class_label_index\n index = np.where(bool_index)[0]\n \n ytrue1 = ytrue[bool_index]\n ypred1 = ypred[bool_index]\n prob1 = prob[bool_index]\n \n hit_ind = np.logical_and(ytrue1 == ypred1, ytrue1 == class_label_index)\n hit_ind1 = np.where(hit_ind)[0]\n miss_ind = np.logical_and(ytrue1 != ypred1, ytrue1 == class_label_index)\n miss_ind1 = np.where(miss_ind)[0]\n \n prob_hit = prob1[hit_ind]\n hit = index[hit_ind]\n \n prob_hit_ind_sorted = (-prob_hit).argsort() \n hit = hit[prob_hit_ind_sorted]\n prob_hit = prob_hit[prob_hit_ind_sorted]\n hit_ind1 = hit_ind1[prob_hit_ind_sorted]\n \n prob_miss = prob1[miss_ind]\n miss = index[miss_ind]\n ypred_miss = ypred1[miss_ind]\n \n prob_miss_ind_sorted = (-prob_miss).argsort() \n miss = miss[prob_miss_ind_sorted]\n ypred_miss = ypred_miss[prob_miss_ind_sorted]\n ypred_miss = ypred_miss.astype(np.int)\n prob_miss = prob_miss[prob_miss_ind_sorted]\n miss_ind1 = miss_ind1[prob_miss_ind_sorted]\n \n for miss1, ypred_miss1, prob_miss1, index1 in zip(miss, ypred_miss, prob_miss, miss_ind1):\n key = '%08d' % (miss1,)\n predicted_label = class_labels[ypred_miss1]\n score = prob_miss1\n self.hit_miss_results[class_label]['miss'].append((predicted_label, score, key, index1))\n \n for hit1, prob_hit1, index1 in zip(hit, prob_hit, hit_ind1):\n key = '%08d' % (hit1,)\n score = prob_hit1\n self.hit_miss_results[class_label]['hit'].append((score, key, index1))\n \n \n def generate_data(self, y_pred_file, y_true_file, class_labels):\n \n # Generate Standard Confusion Matrix Results\n if not (os.path.exists(y_pred_file) and os.path.exists(y_true_file)):\n return False\n \n self.class_labels = class_labels\n\n y_pred = joblib.load(y_pred_file)\n y_pred = np.argmax(y_pred, axis=1)\n y_true = joblib.load(y_true_file)\n y_true = y_true.astype(int)\n self.confusion_matrix, self.scores = self.generate_confusion_matrix(y_pred, y_true)\n\n # Generate Results incorporating temporal information\n per_track_results = self.generate_per_track_results_mode(y_pred, y_true)\n if per_track_results is not None:\n y_pred, y_true = per_track_results\n self.track_confusion_matrix, self.track_scores = self.generate_confusion_matrix(y_pred, y_true)\n \n return True\n\n def generate_per_track_results_mode(self, y_pred, y_true):\n info_file = self.experiment.dataset.path('test_info.txt')\n index_file = self.experiment.dataset.path('test_indices.txt')\n \n if not (os.path.exists(info_file) and os.path.exists(index_file)):\n return None\n\n f = open(index_file)\n lines = f.readlines()\n lines = [l.strip() for l in lines]\n indices = np.array([int(l) for l in lines])\n f.close()\n\n f = open(info_file)\n lines = f.readlines()\n lines = [l.strip().split(' ') for l in lines]\n video_ids = np.array([int(l[0]) for l in lines])\n video_ids = video_ids[indices]\n frames = np.array([int(l[1]) for l in lines])\n frames = frames[indices]\n track_ids = np.array([int(l[2])for l in lines])\n track_ids = track_ids[indices]\n bb_ids = np.array([int(l[3]) for l in lines])\n bb_ids = bb_ids[indices]\n f.close()\n\n y_true_track = []\n y_pred_track = []\n unique_video_ids = np.unique(video_ids)\n for vid_id in unique_video_ids:\n unique_track_ids = np.unique(track_ids[video_ids==vid_id])\n for track_id in unique_track_ids:\n y_true1 = y_true[np.logical_and(video_ids == vid_id, track_ids == track_id)]\n y_true_track.append(mode(y_true1)[0] * np.ones(y_true1.shape[0]))\n y_pred1 = y_pred[np.logical_and(video_ids == vid_id, track_ids == track_id)]\n y_pred_track.append(mode(y_pred1)[0] * np.ones(y_pred1.shape[0]))\n \n y_true_track = np.hstack(y_true_track)\n y_pred_track = np.hstack(y_pred_track)\n return y_pred_track, y_true_track\n \n\n def generate_confusion_matrix(self, y_pred, y_true):\n confusion_matrix = metrics.confusion_matrix(y_true, y_pred)\n scores = metrics.precision_recall_fscore_support(y_true, y_pred)\n return confusion_matrix, scores\n \n \n def plot_confusion_matrix(self, confusion_matrix):\n counts = confusion_matrix \n if counts is None:\n return None\n \n counts_normalised = counts.astype('float') / counts.sum(axis=1)[:, np.newaxis]\n names = self.class_labels\n \n colormap = [\n \"#444444\", \"#a6cee3\", \"#1f78b4\", \"#b2df8a\", \"#33a02c\", \"#fb9a99\",\n \"#e31a1c\", \"#fdbf6f\", \"#ff7f00\", \"#cab2d6\", \"#6a3d9a\"\n ]\n\n xname = []\n yname = []\n color = []\n alpha = []\n for i, n1 in enumerate(names):\n for j, n2 in enumerate(names):\n xname.append(n1)\n yname.append(n2)\n \n a = np.log10(counts[j, i]+1)/np.log10(np.max(counts)+1)\n a = counts_normalised[j,i] * .9 + 0.1\n #color.append(\"#%02x%02x%02x\" % (255, 255 - counts_normalised[j,i] * 255.0, 255 - (count / max_count) * 255.0))\n alpha.append(a)\n color.append(colormap[8])\n\n\n source = ColumnDataSource(\n data=dict(\n xname=xname,\n yname=yname,\n alphas=alpha,\n colors=color,\n count=counts.T.flatten(),\n count_normalised=counts_normalised.T.flatten() * 100.\n )\n )\n\n p = figure(title=\"Confusion Matrix\",\n x_axis_location=\"above\", tools=\"pan,resize,box_zoom,hover,save,tap,reset\",\n x_range=names, y_range=list(reversed(names)))\n p.plot_width = 500\n p.plot_height = 500\n\n p.rect('xname', 'yname', 0.9, 0.9, source=source,\n color='colors', alpha='alphas', line_color=None)\n\n p.grid.grid_line_color = None\n p.axis.axis_line_color = None\n p.axis.major_tick_line_color = None\n p.axis.major_label_text_font_size = \"15pt\"\n p.axis.major_label_standoff = 0\n p.xaxis.major_label_orientation = np.pi/3\n p.xaxis.axis_label = \"Predicted Class\"\n p.yaxis.axis_label = \"Actual Class\"\n\n hover = p.select(dict(type=HoverTool))\n hover.tooltips = OrderedDict([\n ('classes', '@yname, @xname'),\n ('count', '@count'),\n ('percentage', '@count_normalised'),\n ])\n \n url = \"/trials/images/classification/\" + self.id() + \"/results/@yname\"\n taptool = p.select(type=TapTool)\n taptool.action=OpenURL(url=url)\n\n script, div = components(p)\n return (script, div)\n\n def plot_scores2(self):\n \n if self.scores is None:\n return None\n \n scores = np.vstack(self.scores)\n data = OrderedDict()\n for i in xrange(len(self.class_labels)):\n data[self.class_labels[i]] = scores[:2, i]\n s1 = Bar(data, cat=['precision', 'recall'], title=\"Per Class Scores\",\n width=500, height=500, legend=True, tools=\"pan,resize,box_zoom,hover,save,reset\") \n \n data = OrderedDict()\n for i in xrange(len(self.class_labels)):\n data[self.class_labels[i]] = scores[3:4, i]\n s2 = Bar(data, cat=['support'], title=\"Support\",\n width=500, height=500, legend=True, tools=\"pan,resize,box_zoom,hover,save,reset\") \n \n p = hplot(s1,s2)\n \n script, div = components(p)\n return (script, div) \n \n \n def plot_scores(self, all_scores):\n \n if all_scores is None:\n return None\n \n scores = all_scores[0][:]\n scores = np.hstack((scores,np.mean(scores)))\n class_labels = self.class_labels[:]\n class_labels.append('average')\n data = {\"precision\": scores}\n s1 = Bar(data, cat=class_labels, title=\"Per Class Precision\",\n xlabel='categories', ylabel='precision', width=500, height=500,\n tools=\"pan,resize,box_zoom,hover,save,reset\", stacked=True, palette=[\"#b2df8a\"]) \n \n hover = s1.select(dict(type=HoverTool))\n hover.tooltips = OrderedDict([\n ('precision', '@precision'),\n ])\n \n scores = all_scores[1][:]\n scores = np.hstack((scores,np.mean(scores)))\n class_labels = self.class_labels[:]\n class_labels.append('average')\n data = {\"recall\": scores}\n s2 = Bar(data, cat=class_labels, title=\"Per Class Recall\",\n xlabel='categories', ylabel='recall', width=500, height=500,\n tools=\"pan,resize,box_zoom,hover,save,reset\", stacked=True, palette=[\"#a6cee3\"]) \n \n hover = s2.select(dict(type=HoverTool))\n hover.tooltips = OrderedDict([\n ('recall', '@recall'),\n ])\n \n data = {\"support\": all_scores[3]}\n s3 = Bar(data, cat=self.class_labels, title=\"Per Class Support\",\n xlabel='categories', ylabel='support', width=500, height=500,\n tools=\"pan,resize,box_zoom,hover,save,reset\", stacked=True, palette=[\"#6a3d9a\"]) \n \n hover = s3.select(dict(type=HoverTool))\n hover.tooltips = OrderedDict([\n ('support', '@support'),\n ])\n \n p = hplot(s1, s2, s3)\n \n \n \n script, div = components(p)\n return (script, div)\n \n def calculate_average_accuracy(self, confusion_matrix):\n if confusion_matrix is None:\n return None\n\n scores = []\n n = np.shape(confusion_matrix)[0]\n mean_acc = 0\n for index, r in enumerate(confusion_matrix):\n ss = sum(r)\n if ss != 0:\n scores.append(float(r[index]) / ss)\n \n\n scores = np.hstack((scores,np.mean(scores)))\n class_labels = self.class_labels[:]\n class_labels.append('average')\n data = {\"accuracy\": scores}\n s = Bar(data, cat=class_labels, title=\"Per Class Accuracy\",\n xlabel='categories', ylabel='accuracy', width=500, height=500,\n tools=\"pan,resize,box_zoom,hover,save,reset\", stacked=True, palette=[\"#ec5d5e\"]) \n \n hover = s.select(dict(type=HoverTool))\n hover.tooltips = OrderedDict([\n ('accuracy', '@accuracy'),\n ])\n\n p = hplot(s)\n script, div = components(p)\n return (script, div)\n\n\n","sub_path":"digits/trial/images/classification/job.py","file_name":"job.py","file_ext":"py","file_size_in_byte":13200,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"586035557","text":"from django import forms\nfrom .models import Cursos, Programaciones, Horarios\n\nclass CursosForm(forms.ModelForm):\n\tclass Meta:\n\t\tmodel = Cursos\n\t\tfields = ('programcurso','namecurso',)\n\t\tlabels = {\n 'programcurso': 'Programa',\n 'namecurso':'Curso',\n }\n\nclass ProgramacionesForm(forms.ModelForm):\n\tclass Meta:\n\t\tmodel = Programaciones\n\t\tfields = ('idprogramacion',)\n\nclass HorariosForm(forms.ModelForm):\n\tclass Meta:\n\t\tmodel = Horarios\n\t\tfields = ('dayhorario','timestarthorario','timeendhorario',)\n\t\tlabels = {\n\t\t\t'dayhorario':'Dia',\n\t\t\t'timestarthorario':'Hora Inicio',\n\t\t\t'timeendhorario':'Hora Termino',\n\t\t}","sub_path":"horarios/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"451140907","text":"'''\nGiven a string and a string dictionary, find the longest string in the dictionary that can be formed by deleting some characters of the given string. If there are more than one possible results, return the longest word with the smallest lexicographical order. If there is no possible result, return the empty string.\n\nExample 1:\nInput:\ns = \"abpcplea\", d = [\"ale\",\"apple\",\"monkey\",\"plea\"]\n\nOutput: \n\"apple\"\nExample 2:\nInput:\ns = \"abpcplea\", d = [\"a\",\"b\",\"c\"]\n\nOutput: \n\"a\"\nNote:\nAll the strings in the input will only contain lower-case letters.\nThe size of the dictionary won't exceed 1,000.\nThe length of all the strings in the input won't exceed 1,000.\n'''\n\nclass Solution:\n def findLongestWord(self, s: str, d: List[str]) -> str:\n longest_word = \"\"\n for word in d:\n if self.isSubsequence(word, s):\n if len(word) > len(longest_word) or len(word) == len(longest_word) and word < longest_word:\n longest_word = word\n \n return longest_word\n \n def isSubsequence(self, word, s):\n i = 0\n j = 0\n while i < len(word) and j < len(s):\n if word[i] == s[j]:\n i += 1\n j += 1 \n return i == len(word)\n \n","sub_path":"problems/524_longest_word_in_dictionary_through_deleting.py","file_name":"524_longest_word_in_dictionary_through_deleting.py","file_ext":"py","file_size_in_byte":1247,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"186169881","text":"#Predicting Systolic Blood Pressure\n\n#data is from:\n#http://college.cengage.com/mathematics/brase/understandable_statistics/7e/students/datasets/mlr/frames/mlr02.html\n\n# The data (X1, X2, X3)/patient.\n# X1 => systolic blood pressure\n# X2 => age (years)\n# X3 => weight (lbs)\n\n#Requres Pandas to read the Excel Sheet => 'pip install xlrd'\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\n\ndataFrame = pd.read_excel('mlr02.xls')\nX = dataFrame.as_matrix()\n\n#Age vs Blood Pressure\nplt.scatter(X[:, 1], X[:, 0])\nplt.show()\n\n#Weight vs Blood Pressure\nplt.scatter(X[:, 2], X[:, 0])\nplt.show()\n\n#Partitioning\ndataFrame['ones'] = 1\nY = dataFrame['X1']\nX = dataFrame[['X2', 'X3', 'ones']]\nx2_Only = dataFrame[['X2', 'ones']]\nx3_Only = dataFrame[['X3', 'ones']]\n\n#R^2 Calculate\ndef get_r2(X, Y):\n weight = np.linalg.solve(X.T.dot(X), X.T.dot(Y))\n Yhat = X.dot(weight)\n\n d1 = Y - Yhat\n d2 = Y - Y.mean()\n r2 = 1 - d1.dot(d1) / d2.dot(d2)\n return r2\n\nprint(\"R2 for Age Only is : \", get_r2(x2_Only, Y))\nprint(\"R2 for Weight Only is : \", get_r2(x3_Only, Y))\nprint(\"R2 for both Age & Weight is : \", get_r2(X, Y))\n","sub_path":"linearRegression/systolicRegression.py","file_name":"systolicRegression.py","file_ext":"py","file_size_in_byte":1137,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"307614475","text":"import torch\nimport torch.nn.functional as F\nimport matplotlib.pyplot as plt\n\n\n''' creat data '''\ntorch.manual_seed(1)\n\nx = torch.unsqueeze(torch.linspace(-1, 1, 100), dim=1) #in torch, only 2-dimentions can be processed. torch.unsqueeze() add another dim, dim= means which dim be added\ny = x.pow(2) + 0.2*torch.rand(x.size())\n\n\nplt.figure()\nplt.scatter(x.data.numpy(), y.data.numpy())\nplt.grid(True)\nplt.draw()\n\n''' build NN model structure '''\nclass Net(torch.nn.Module): # define neural network that inherits the module (torch.nn.Module) form torch\n def __init__(self, n_features, n_hidden, n_output): # required information for building layers\n super(Net, self).__init__()\n self.hidden = torch.nn.Linear(n_features, n_hidden)\n self.predict = torch.nn.Linear(n_hidden, n_output)\n\n def forward(self, x): # forward process of NN\n x = F.relu(self.hidden(x))\n x = self.predict(x)\n return x\n\n\n''' define the number of neurons in each layers '''\nnet = Net(1, 10, 1)\nprint(net)\n\n\n\n\n''' optimize function '''\noptimizer = torch.optim.SGD(net.parameters(), lr=0.2)\nloss_func = torch.nn.MSELoss() # MSELoss: Mean Square Error for regression problem\n\n\nnet2 = torch.load('net.pkl')\n\nfor t in range(200):\n prediction = net2(x)\n if t % 10 == 0: # plot and show learning process\n plt.cla()\n plt.scatter(x.data.numpy(), y.data.numpy())\n plt.plot(x.data.numpy(), prediction.data.numpy(), 'r-', lw=5) # lw: line width\n plt.pause(0.1)\n\n\nplt.ioff() # Turn the interactive mode off.\nplt.show()\n\n\n","sub_path":"morvan_pytorch/morvan_Regression_load the trained net.py","file_name":"morvan_Regression_load the trained net.py","file_ext":"py","file_size_in_byte":1569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"201917262","text":"from firebase.firebase import FirebaseApplication, FirebaseAuthentication\n# import ijson\nimport pandas as pd\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\n# import time\nfrom datetime import datetime\nfrom flask import Flask, render_template, request\nimport io\nimport base64\n# import re\n\napp = Flask(__name__)\n\n\n@app.route('/', methods=['POST', 'GET'])\ndef main():\n if request.method == 'POST':\n result = fetch_data() #fetch from Firebasse\n data, columns = organize_data(result)\n load_choice = []\n box = request.form.get('all')\n if box == '1':\n load_choice = 'all'\n # print(\"checkbox\", box)\n else:\n for item in columns:\n # if item != 'timestamp':\n curr_box = request.form.get(item)\n if curr_box == '1':\n load_choice.append(item)\n\n # load_choice = request.form['load_choice']\n station_name = request.form['station_name']\n\n # load_choice = ['Grid','Tablet1']\n data_slice, tail_data = process_data(data, columns, load_choice=load_choice)\n image_url = plot_data(data_slice) # get graph image url to display\n\n # cut the tail_data into last 5 measurements\n max_index = max(tail_data, key=int)\n indexes = {max_index, max_index-1, max_index-2, max_index-3, max_index-4}\n cut_data = dict([(key, tail_data[key]) for key in indexes if key in tail_data])\n # print(cut_data)\n return render_template('graph/data_layout.html', station_name=station_name, image_url=image_url, tail_data=cut_data)\n # time.sleep(5)\n elif request.method == 'GET':\n result = fetch_data() # fetch data from Firebase\n data, columns = organize_data(result)\n load_keys = list(columns)\n i = -1\n # remove 'timestamp' from the dropdown list\n for item in load_keys:\n i += 1\n if item == 'timestamp':\n del load_keys[i]\n return render_template('inheritance/child_template_form.html', loads=load_keys)\n\n\ndef fetch_data():\n print(\"START\")\n SECRET = '2woGLBK6IxfOtDCM553cZshFZJAZxSHVd3mAcGRY'\n DSN = 'https://jakartasmartpark.firebaseio.com/'\n EMAIL = 'tamanhijaujakarta'\n authentication = FirebaseAuthentication(SECRET,EMAIL, True, True)\n firebase = FirebaseApplication(DSN, authentication)\n result = firebase.get('/measurements', None)\n return result\n# print(result)\n\n\ndef organize_data(result):\n columns = result.keys()\n # print(columns)\n print(\"DATA FETCHED\")\n status = []\n timestamp = []\n data = []\n for f in columns:\n raw_data = []\n raw_time = result[f]['timestamp']\n real_timestamp = datetime.strptime(raw_time,'%Y-%m-%dT%H:%M:%S.%f') #convert str into datetime object\n #real_timestamp = datetime.strftime(real_timestamp,'%Y-%m-%d %H:%M:%S')\n if real_timestamp.year > 2017:\n keys = list(result[f].keys())\n for key in keys:\n if key == 'timestamp':\n format_timestamp = datetime.strftime(real_timestamp, '%Y-%m-%d %H:%M:%S')\n real_timestamp = datetime.strptime(format_timestamp, '%Y-%m-%d %H:%M:%S')\n raw_data.append(real_timestamp)\n else:\n if key in result[f]:\n raw_data.append(result[f][key])\n else:\n raw_data.append('NaN')\n\n data.append(raw_data) # create a new dictionary for the data\n # print(data[0])\n # columns = result[f].keys()\n return data, keys\n\n\ndef process_data(data, columns, load_choice):\n status = pd.DataFrame(data, columns=columns)\n print(status.tail())\n tail_dict = status.to_dict('index')\n status = status.set_index('timestamp')\n if load_choice == 'all':\n pass\n else:\n status = status.loc[:, load_choice]\n # print(status['timestamp'].dtype)\n # status = status.loc[:,['Grid','Inverter','Load']]\n status = status.astype(float)\n\n return status, tail_dict\n\n\ndef plot_data(status):\n\n img = io.BytesIO()\n plt.gcf().clear()\n # status = status.cumsum()\n status.plot()\n plt.legend(loc='best')\n plt.tight_layout()\n # plt.show()\n plt.savefig(img, format='png')\n img.seek(0)\n plot_url = base64.b64encode(img.getvalue()).decode()\n # print(plot_url)\n return 'data:image/png;base64,{}'.format(plot_url)\n","sub_path":"library/get_load_data.py","file_name":"get_load_data.py","file_ext":"py","file_size_in_byte":4477,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"338755849","text":"#!/usr/bin/env python\n\nimport glob\n\nimport multiprocessing as mp\nfrom multiprocessing import shared_memory\nimport argparse\nimport numpy as np\nimport xarray as xr\nimport os\nimport time\n\npool = None\ndata_vars = None\n\n\ndef load_file(file_name,var_mems):\n '''Load a netcdf dataset into memory'''\n d = xr.open_dataset(file_name)\n subset(d,var_mems)\n return 'loaded '+file_name\n\n\ndef get_dims(dataset, section=\"d\"):\n '''Get the global attributes defining the domain, memory, or tile space'''\n results = []\n for axis in [\"i\",\"j\",\"k\"]:\n for position in [\"s\",\"e\"]:\n results.append(int(dataset.attrs[axis + section + position]))\n return results\n\ndef get_dim_offset(dims):\n '''Return x_offset, y_offset\n For the staggered dims, offset=1, otherwise offset=0'''\n x_off = 0\n if 'lon_u' in dims: x_off = 1\n\n y_off = 0\n if 'lat_v' in dims: y_off = 1\n\n return x_off, y_off\n\ndef set_up_data_vars(d,var_names):\n ids, ide, jds, jde, kds, kde = get_dims(d, section='d')\n nx = ide - ids + 1\n ny = jde - jds + 1\n nz = kde - kds + 1\n\n data_vars = dict()\n shms = dict()\n\n if var_names is None: var_names = d.variables\n\n for v in var_names:\n coords = [c for c in d[v].coords]\n dims = d[v].dims\n name = d[v].name\n attrs = d[v].attrs\n\n x_off, y_off = get_dim_offset(dims)\n\n if len(dims) == 1:\n nt = d.dims[dims[0]]\n data = np.zeros((nt))\n if len(dims) == 2:\n data = np.zeros((ny + y_off, nx + x_off))\n if len(dims) == 3:\n data = np.zeros((d.dims[dims[0]], ny + y_off, nx + x_off))\n if len(dims) == 4:\n nt = d.dims[dims[0]]\n nz = d.dims[dims[1]]\n data = np.zeros((nt, nz, ny + y_off, nx + x_off))\n \n #make data into shared memory\n shm = mp.shared_memory.SharedMemory(create=True, size=data.nbytes)\n shms[v] = shm\n shm_data = np.ndarray(data.shape, dtype=data.dtype, buffer=shm.buf)\n # print(name, data.shape, dims, attrs)\n data_vars[v] = xr.DataArray(shm_data, dims=dims, name=name, attrs=attrs)#, coords=coords)\n \n return data_vars, shms\n\ndef set_up_dataset(d,data_vars):\n '''Create a dataset to cover the entire domain with the variables present in d\n\n d : an input dataset covering part of the domain\n d must have global attributes ids, ide, jds, jde, kds, kde that define the full domain\n\n A new dataset is created with all the variables+attributes in d covering the full domain\n '''\n ds = xr.Dataset(data_vars, attrs=d.attrs)\n ds.encoding = d.encoding\n ds[\"time\"] = d[\"time\"]\n cords = []\n \n for v in d.variables:\n if v in data_vars:\n for c in d[v].coords:\n if not(c in cords) and (c in data_vars):\n cords.append(c)\n return ds.set_coords(cords)\n\n\ndef agg_file(first_file,var_names,verbose=True):\n '''Aggregated all files that come from the same time step as first_file\n\n first_file should have _001_ in the filename somewhere. This will be replaced\n with * to search for all matching files from this date. Once files are found, a\n dataset containing the entire domain is created and the data from each file are\n added to the master dataset.\n\n Result: aggregated dataset is written to a netcdf file'''\n \n if verbose:print(first_file)\n date_search = first_file.replace(\"_000001_\",\"*\")\n outputfile = first_file.replace(\"000001_\",\"_\").replace(\"__\",\"_\")\n if os.path.isfile(outputfile):\n return\n this_date_files = glob.glob(date_search)\n this_date_files.sort()\n \n template = xr.open_dataset(this_date_files[0])\n data_vars, shms = set_up_data_vars(template,var_names)\n\n args = [(d,shms) for d in this_date_files]\n results = pool.starmap_async(load_file, args)\n\n #Just use get to wait for result\n message = results.get()\n\n data_set = set_up_dataset(template,data_vars)\n data_set.load().to_netcdf(outputfile)\n for key in shms:\n shms[key].close()\n shms[key].unlink()\n\ndef subset(d,var_mems):\n ids, ide, jds, jde, kds, kde = get_dims(d, section='d')\n ims, ime, jms, jme, kms, kme = get_dims(d, section='m')\n its, ite, jts, jte, kts, kte = get_dims(d, section='t')\n\n xts, xte = its - ims, ite - ims + 1\n yts, yte = jts - jms, jte - jms + 1\n zts, zte = kts - kms, kte - kms + 1\n\n xs, xe = its - ids, ite - ids + 1\n ys, ye = jts - jds, jte - jds + 1\n zs, ze = kts - kds, kte - kds + 1\n\n nx = ide - ids + 1\n ny = jde - jds + 1\n nz = kde - kds + 1\n\n if ims==ids:\n its = ids\n if ime==ide:\n ite = ide\n\n if jms==jds:\n jts = jds\n if jme==jde:\n jte = jde\n\n for v in var_mems:\n dims = d[v].dims\n existing_mem = mp.shared_memory.SharedMemory(name=var_mems[v].name)\n x_off, y_off = get_dim_offset(dims)\n\n if len(dims) == 2:\n data = np.ndarray((ny + y_off, nx + x_off),buffer=existing_mem.buf)\n data[ys:ye, xs:xe] = d[v].values[yts:yte, xts:xte]\n if len(dims) == 3:\n data = np.ndarray((d.dims[dims[0]], ny + y_off, nx + x_off),buffer=existing_mem.buf)\n if dims[0] == \"time\":\n data[:, ys:ye+y_off, xs:xe+x_off] = d[v].values[:, yts:yte+y_off, xts:xte+x_off]\n else:\n data[zs:ze, ys:ye+y_off, xs:xe+x_off] = d[v].values[zts:zte, yts:yte+y_off, xts:xte+x_off]\n if len(dims) == 4:\n nt = d.dims[dims[0]]\n nz = d.dims[dims[1]]\n data = np.ndarray((nt, nz, ny + y_off, nx + x_off),buffer=existing_mem.buf)\n data[:,zs:ze, ys:ye+y_off, xs:xe+x_off] = d[v].values[:,zts:zte, yts:yte+y_off, xts:xte+x_off]\n existing_mem.close()\n return\n\ndef main(file_search,cpus,var_names):\n first_files = glob.glob(file_search.format(ens=\"000001\"))\n first_files.sort()\n\n for f in first_files:\n agg_file(f,var_names,verbose=True)\n \ndef continuous(file_search,cpus,var_names):\n print(\"Running continuous aggregation, Ctrl-C to stop\")\n while True:\n first_files = glob.glob(file_search.format(ens=\"000001\"))\n first_files.sort()\n\n # skip the last file in the list as ICAR might still be running\n for f in first_files[:-1]:\n agg_file(f, var_names,verbose=False)\n \n time.sleep(10)\n\nif __name__ == '__main__':\n\n # This should be an input, this is the search string that is assumed to match\n # the output files to be aggregated.\n file_search = \"icar_out_{ens}_*\"\n \n parser = argparse.ArgumentParser()\n parser.add_argument(\"-n\", \"--n_cpus\", type=int,\n help=\"Number of cpus to use\")\n parser.add_argument(\"--continuous\", action=\"store_true\",\n help=\"Use continuous aggregation of files\")\n parser.add_argument(\"-v\", \"--vars\", type=str,\n help=\"File containing var names to save (comma-delimited; .csv)\")\n parser.add_argument(\"-s\", \"--search_string\", type=str,\n help=\"Format of search string\")\n \n args = parser.parse_args()\n\n cpus = mp.cpu_count() \n if ( args.n_cpus and args.n_cpus > 0 and args.n_cpus < cpus): cpus = args.n_cpus\n \n if ( args.vars): var_names = np.loadtxt(args.vars,dtype=str,delimiter=',')\n else: var_names = None\n \n if (args.search_string): file_search = args.search_string\n \n mp.set_start_method('spawn')\n pool = mp.Pool(cpus)\n \n if args.continuous:\n try:\n continuous(file_search,cpus,var_names)\n except KeyboardInterrupt:\n pass\n else:\n main(file_search,cpus,var_names)\n pool.close()\n","sub_path":"helpers/aggregate_parallel_files_par.py","file_name":"aggregate_parallel_files_par.py","file_ext":"py","file_size_in_byte":7876,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"424289101","text":"import unittest\n\nfrom brew.parsers import JSONDataLoader\n\n\nclass TestJSONDataLoader(unittest.TestCase):\n\n def setUp(self):\n self.parser = JSONDataLoader('./')\n\n def test_format_name(self):\n name_list = [('pale malt 2-row us', 'pale_malt_2_row_us'),\n ('caramel crystal malt 20l', 'caramel_crystal_malt_20l'),\n ('centennial', 'centennial'),\n ('cascade us', 'cascade_us'),\n ('Wyeast 1056', 'wyeast_1056'),\n ]\n for name, expected in name_list:\n out = self.parser.format_name(name)\n self.assertEquals(out, expected)\n","sub_path":"tests/test_parsers.py","file_name":"test_parsers.py","file_ext":"py","file_size_in_byte":660,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"400948451","text":"from tags import Tag\n\nclass Interaction(Tag):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\n\t\tself.class_type = \"interaction\"\n\t\tself.need_prefix = [\"i\",]\n\t\tself.parents = False\n\nclass ITarget(Tag):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\n\t\tself.class_type = \"i_target\"\n\t\tself.need_prefix = [\"it\",]\n\t\tself.parents = [\"interaction\",]\n\t\tself.allow_duplicates = True\n\nclass IEffect(Tag):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\n\t\tself.class_type = \"i_effect\"\n\t\tself.need_prefix = [\"ie\", \"syndrome\"]\n\t\tself.parents = [\"interaction\",]","sub_path":"class_types/interaction_classes.py","file_name":"interaction_classes.py","file_ext":"py","file_size_in_byte":627,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"167469469","text":"class Settings:\n def __init__(self):\n self.screen_width = 1200\n self.screen_height = 800\n #ciekawe czy uda się zdefiniowac taki event żeby obraz rosnął po naciśnięciu konkretnego klawisza?\n #chyba za duzo zmiennyych\n self.bg_color = (50, 50, 125)\n self.ship_speed = 2\n self.Full_screen = 0\n\n self.bullet_speed = 2.0\n self.bullet_width = 6\n self.bullet_height = 6\n self.bullet_color = (255, 0, 0)\n self.bullets_allowed = 3","sub_path":"venv/lib/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":514,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"280288163","text":"#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport math\nimport numpy as np\nimport pandas as pd\n# from scipy.spatial.distance import cosine\n\n\ndef get_vec(word, embeddings_df):\n \"\"\"Find word on the list of word_embeddings and return its vector\"\"\"\n return embeddings_df.loc[embeddings_df[0] == word].values[:, 1:]\n\n\ndef vector_len(v):\n return math.sqrt(sum([x*x for x in v]))\n\n\ndef dot_product(v1, v2):\n assert len(v1) == len(v2)\n return sum([x * y for (x, y) in zip(v1, v2)])\n\n\ndef cosine_similarity(v1, v2):\n \"\"\"\n Returns the cosine of the angle between the two vectors.\n Results range from -1 (very different) to 1 (very similar).\n \"\"\"\n return dot_product(v1, v2) / (vector_len(v1) * vector_len(v2))\n\n\ndef similarity_calculator(vec):\n return lambda x: cosine_similarity(vec, x)\n\n\ndef find_word(vec, embeddings_df, skip_first=False):\n vecs = embeddings_df.values[:, 1:]\n words = embeddings_df.values[:, 0]\n similarity_to_vec = similarity_calculator(vec)\n similarities = np.apply_along_axis(similarity_to_vec, 1, vecs)\n order = np.argsort(-similarities)\n words_sorted = words[order]\n if skip_first:\n return words_sorted[1:6]\n return words_sorted[:5]\n\n\ndef main():\n \"\"\"Main function\"\"\"\n # embeddings = pd.read_csv('kgr_mwe/v100/cbow_v100m8.w2v.txt')\n print('Reading embeddings from file... ', end='', flush=True)\n embeddings = pd.read_csv('truncated.w2v.txt', sep=' ', skiprows=[0], header=None)\n print('DONE')\n while True:\n print('> ', end='')\n words = input().split()\n if len(words) == 1:\n a = words[0]\n avs = get_vec(a, embeddings)\n if not avs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(a))\n continue\n for av in avs:\n similar = find_word(av, embeddings, skip_first=True)\n print(av)\n print(similar)\n elif len(words) == 2:\n a, b = words\n avs = get_vec(a, embeddings)\n if not avs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(a))\n continue\n bvs = get_vec(b, embeddings)\n if not bvs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(b))\n continue\n for av in avs:\n for bv in bvs:\n print(cosine_similarity(av, bv))\n elif len(words) == 3:\n a, b, c = words\n avs = get_vec(a, embeddings)\n if not avs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(a))\n continue\n bvs = get_vec(b, embeddings)\n if not bvs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(b))\n continue\n cvs = get_vec(c, embeddings)\n if not cvs.any():\n print('Nie znaleziono wektora dla \"{}\"'.format(c))\n continue\n for av in avs:\n for bv in bvs:\n for cv in cvs:\n dv = cv + bv - av\n d = find_word(dv, embeddings, skip_first=True)\n print(d)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"pl_word_embeddings.py","file_name":"pl_word_embeddings.py","file_ext":"py","file_size_in_byte":3255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"209668943","text":"'''\nCreated on Nov 2, 2015\n@author: tudorstanila\n'''\nfrom InMemoryRepository import *\nfrom collections import OrderedDict\n\nclass AssignmentRepository(InMemoryRepository):\n def __init__(self,aNewStudentRepository):\n InMemoryRepository.__init__(self)\n self._aStudentRepository=aNewStudentRepository\n \n \n def getPosition(self,id):\n for i in range(0,len(self._listRepr)):\n if self._listRepr[i].getID()==id:\n return i\n \n \n def getByAssign(self,assign):\n rez=[]\n for i in range (0, len(self._listRepr)):\n s=self.getElemenAtPos(i)\n if self.getAssign(s)==assign:\n rez.append(self.getElemenAtPos(i))\n \n return rez \n def getAssign(self,a):\n return a._AssignDescr\n \n def getID(self,a):\n return a._AssignID\n \n def getGradeByAssign(self,assign):\n list=[]\n for i in range (0, len(self._listRepr)):\n s=self.getElemenAtPos(i)\n nameS=self.getStudentName(s.getID())\n if self.getAssign(s)==assign:\n list.append(StudentGrade(nameS,self.getGrade(s)))\n return list \n\n def getStudentGrade(self):\n list=[]\n for i in range (0, len(self._listRepr)):\n s=self.getElemenAtPos(i)\n nameS=self.getStudentName(s.getID())\n list.append(StudentGrade(nameS,self.getGrade(s)))\n return list \n def getGrade(self,assign):\n return assign._AssignGrade\n \n def getStudentName(self,id):\n s=self._aStudentRepository.findByID(id)\n if s!=None:\n return s.getsName()\n else:\n return ''\n \n def sortAlphabetically(self,list):\n return sorted(list, key=lambda StudentGrade: StudentGrade._nameS)\n \n def sortByGrade(self,list):\n return sorted(list, key=lambda StudentGrade: StudentGrade._gradeS)\n def getGradeAverage(self):\n list=self.getStudentGrade()\n rez=[] \n names=[]\n newlist=[]\n for i in range (0,len(list)):\n names.append(list[i].getName())\n \n for i in range(0,len(names)-1):\n if names[i]!=names[i+1]:\n newlist.append(names[i])\n for i in range(0,len(newlist)):\n avg=0\n n=0\n for j in range(0,len(list)):\n if newlist[i]==list[j].getName():\n avg+=list[j].getGrade()\n n=n+1\n avg=avg/n\n rez.append(StudentGrade(newlist[i],avg))\n return rez\n \n @staticmethod\n def test_getAssgin():\n s=Assignment(1,\"Mate\",\"23.05\",10)\n assert self.getAssign(s)==\"Mate\"\n \n @staticmethod\n def test_getByAssign():\n ls=[]\n a1=Assignment(1,\"Mate\",\"23.05\",10)\n ls.append(a1)\n a2=Assignment(2,\"Engl\",\"10.11\",10)\n ls.append(a2)\n s=a1\n assert self.getByAssign(\"Mate\")==s\n \n @staticmethod\n def test_getGradeByAssign():\n ls=[]\n a1=Assignment(1,\"Mate\",\"23.05\",10)\n ls.append(a1)\n a2=Assignment(2,\"Engl\",\"10.11\",5)\n ls.append(a2)\n assert self.getByAssign(\"Mate\")==10\n \n\"\"\" \nclass StudentGrade:\n def __init__(self,nameS,gradeS):\n self._nameS=nameS\n self._gradeS=gradeS\n \n def getName(self):\n return self._nameS\n def getGrade(self):\n return self._gradeS\n \n def __str__(self):\n elem=''\n elem+=str(self._nameS)+' '\n elem+=str(self._gradeS)+' '\n elem+='\\n'\n return elem\n \n def strName(self):\n elem=str(self._nameS)\n return elem\n \n \n \"\"\"","sub_path":"Python/Lab05-07/AssignmentRepository.py","file_name":"AssignmentRepository.py","file_ext":"py","file_size_in_byte":3730,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"52639174","text":"\n\n#calss header\nclass _MORALE():\n\tdef __init__(self,): \n\t\tself.name = \"MORALE\"\n\t\tself.definitions = [u'the amount of confidence felt by a person or group of people, especially when in a dangerous or difficult situation: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_morale.py","file_name":"_morale.py","file_ext":"py","file_size_in_byte":397,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"270798664","text":"# -*- coding:utf-8 -*-\nfrom .api_base import JsonHandler\nfrom utils import get_tags, get_tags_v2, get_tags_parents, get_tags_v3\n# get_tags_v2_by_name\nfrom db import Tag, Share, User\nimport tornado\nimport time\nimport copy\nd_tags = get_tags()\nd_tags_v2 = get_tags_v2()\nd_tags_v3 = get_tags_v3()\n\nd_tags_parents = get_tags_parents()\n\n\n# get_tags_v2_by_name\nclass TagsV2Handler(JsonHandler):\n\n def get(self):\n ver = self.get_argument(\"ver\", 3)\n name = self.get_argument(\"name\", '')\n sid = self.get_argument(\"id\", 0)\n ver = int(ver)\n sid = int(sid)\n # parents [0]\n if name or sid:\n # 具体某个标签\n if not name and sid:\n tag = Tag.by_sid(sid)\n name = tag['name']\n self.res = d_tags_v3.get(name, {})\n\n if self.res:\n parents = d_tags_parents.get(self.res['name'], {})\n node = d_tags_v3.get(parents, {})\n self.res['parents'] = copy.deepcopy(node)\n self.res['parents'].pop('subs')\n\n brothers = []\n for sub in copy.deepcopy(node)['subs']:\n sub.pop('subs')\n brothers.append(sub)\n self.res['brothers'] = brothers\n if parents:\n parents_p = d_tags_parents.get(parents, {})\n if parents_p:\n node = d_tags_v3.get(parents_p, {})\n\n self.res['parents']['parents'] = copy.deepcopy(node)\n self.res['parents']['parents'].pop('subs')\n\n parents_pp = d_tags_parents.get(parents_p, {})\n if parents_pp:\n node = d_tags_v3.get(parents_pp, {})\n self.res['parents']['parents']['parents'] = copy.deepcopy(node)\n self.res['parents']['parents']['parents'].pop('subs')\n\n self.res['articleNumber'] = Share.count_by_tag(self.res['name'])\n self.res['followerNumber'] = User.find({'user_tags': {'$in': [name]}}).count()\n self.res['isFollowing'] = False\n user_id = self.current_user[\"user_id\"] if self.current_user else None\n if user_id:\n user = User.by_sid(user_id)\n if user:\n # model1\n self.res['isFollowing'] = name in user['user_tags']\n # model2 查看like 表\n # self.res['isFollowing'] = name in user['user_tags']\n\n print(\"self.res['followerNumber']\", self.res['followerNumber'])\n tag = Tag.by_name(self.res['name'])\n\n if tag:\n self.res['id'] = tag['id']\n else:\n self.res['id'] = -1\n else:\n # 从根节点开始\n if ver == 3:\n self.res = d_tags_v2\n self.res['parents'] = {} # root\n self.res['articleNumber'] = Share.count_by_tag(self.res['name'])\n tag = Tag.by_name(self.res['name']) # 7491\n self.res['id'] = -1\n if tag:\n self.res['id'] = tag['id']\n elif ver == 2:\n self.res = d_tags_v2\n if self.current_user and 'user_id' in self.current_user:\n user = User.by_sid(self.current_user['user_id'])\n self.res['watched_tags'] = user['user_tags']\n else:\n self.res = d_tags\n self.write_json()\n\n @tornado.web.authenticated\n def post(self):\n share_id = self.get_argument(\"share_id\", None)\n tags = self.get_argument(\"tags\", '')\n tags = tags.strip()\n if share_id:\n share = Share.by_sid(share_id)\n if share and tags not in share.tags:\n tags = share.tags + ' ' + tags\n res = {\n 'tags': tags,\n 'updated': time.time(),\n }\n share.update(res)\n share.save()\n tags = tags.split(' ')\n tags = list(set(tags))\n for i in tags:\n doc = {\n 'name': i,\n 'share_ids': share.id\n }\n Tag.new(doc)\n\n\nclass TagsHandler(JsonHandler):\n\n def get(self):\n ver = self.get_argument(\"ver\", 2)\n ver = int(ver)\n if ver == 2:\n self.res = d_tags_v2\n if self.current_user and 'user_id' in self.current_user:\n user = User.by_sid(self.current_user['user_id'])\n self.res['watched_tags'] = user['user_tags']\n else:\n self.res = d_tags\n self.write_json()\n\n\n# print(self.current_user)\n# print(hasattr(self.current_user, 'user_id'))\n# if hasattr(self.current_user, 'user_id'):\n# user_id = self.current_user[\"user_id\"]\n","sub_path":"anwen/api_tag.py","file_name":"api_tag.py","file_ext":"py","file_size_in_byte":5064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"193695634","text":"'''\nBottom-up and Top-down Designs of Privacy Compensation\n@Application: Gaussian Distribution Fitting (Sum and Sum of Squares)\n@Author: Chaoyue Niu\n@Email: rvincency@gmail.com\n@Reference: C. Niu, Z. Zheng, F. Wu, S. Tang, X. Gao, and G. Chen,\n \"Unlocking the Value of Privacy: Trading Aggregate Statistics over Private Correlated Data\", in KDD, 2018\n@Settings: ERATO based;\n Total privacy compensations B = 100000.0;\n Number of data owners = 10000;\n i: data owner; j: dependent data owners; k: type of energy consumption.\n'''\n\n\nfrom random import uniform\nimport math\nimport numpy as np\n\n\ndef xBottomUpCompensate(x, k):\n v = 100.0\n return math.tanh(x/math.sqrt(v/2))\n\n\ndef x2BottomUpCompensate(x):\n v = 100.0\n return math.tanh(x/math.sqrt(v/2))\n\n\nif __name__==\"__main__\":\n \"\"\"\n Some Prior Information about MIN, MAX\n \"\"\"\n #For Sum: E(x)\n lowBound = [0, 0, 0, 0, 0]\n upBound = [5000, 3000, 3000, 400, 5000]\n\n\n \"\"\"\n Read Dependent Sensitivity\n \"\"\"\n #For Sum: E(x)\n xDSf = np.zeros((10000,5),float)\n f1 = open(\"xDSf\")\n for i in range(10000):\n line = f1.readline()\n if not line:\n break\n linetmp = line.split()\n for k in range(5):\n xDSf[i][k] = float(linetmp[k])\n print(xDSf)\n f1.close()\n\n # For Sum of Squares: E(x^2)\n x2DSf = np.zeros((10000, 5), float)\n f2 = open(\"x2DSf\")\n for i in range(10000):\n line = f2.readline()\n if not line:\n break\n linetmp = line.split()\n for k in range(5):\n x2DSf[i][k] = float(linetmp[k])\n print(x2DSf)\n f2.close()\n\n #Total Privacy Compensations B\n total_micro = 100000.0\n\n\n \"\"\"\n Case of Sum: E(x) \n \"\"\"\n #Bottom-up Design\n xbuPsi = np.zeros((10000, 5), float)\n xbusumPsi = np.zeros((5, 1), float)\n for k in range(5):\n for i in range(10000):\n xbuPsi[i][k] = xBottomUpCompensate(xDSf[i][k], k)\n xbusumPsi[k] += xbuPsi[i][k]\n\n xbuMicroPay = np.zeros((10000, 5), float)\n # xbuCnt: Count different interval values\n xbuCnt = np.zeros((6, 50), int)\n xbuMicoPayAvg = np.zeros((10000, 1), float)\n for k in range(5):\n for i in range(10000):\n xbuMicroPay[i][k] = (xbuPsi[i][k]/xbusumPsi[k]) * total_micro\n xbuMicoPayAvg[i] += xbuMicroPay[i][k]/5.0\n xbuCnt[k][int(xbuMicroPay[i][k])] += 1\n for i in range(10000):\n xbuCnt[5][int(xbuMicoPayAvg[i])] += 1\n np.savetxt('xbuMicroPay', xbuMicroPay, fmt='%.10f')\n np.savetxt('xbuMicroPay-Count', xbuCnt, fmt='%d')\n\n #Top-down Design\n xtdsumDSfi = np.zeros((5, 1), float)\n for k in range(5):\n for i in range(10000):\n xtdsumDSfi[k] += xDSf[i][k]\n\n xtdMicroPay = np.zeros((10000, 5), float)\n # xtdCnt: Count different interval values\n xtdCnt = np.zeros((6, 50), int)\n xtdMicoPayAvg = np.zeros((10000, 1), float)\n\n for k in range(5):\n for i in range(10000):\n xtdMicroPay[i][k] = (xDSf[i][k] / xtdsumDSfi[k]) * total_micro\n xtdMicoPayAvg[i] += xtdMicroPay[i][k] / 5.0\n xtdCnt[k][int(xtdMicroPay[i][k])] += 1\n for i in range(10000):\n xtdCnt[5][int(xtdMicoPayAvg[i])] += 1\n np.savetxt('xtdMicroPay', xtdMicroPay, fmt='%.10f')\n np.savetxt('xtdMicroPay-Count', xtdCnt, fmt='%d')\n\n \"\"\"\n Case of Sum of Squares: E(x^2)\n \"\"\"\n # Bottom-up Design\n x2buPsi = np.zeros((10000, 5), float)\n x2busumPsi = np.zeros((5, 1), float)\n for k in range(5):\n for i in range(10000):\n x2buPsi[i][k] = x2BottomUpCompensate(x2DSf[i][k])\n # print(buPsi[i][k])\n x2busumPsi[k] += x2buPsi[i][k]\n\n x2buMicroPay = np.zeros((10000, 5), float)\n # x2buCnt: Count different interval values\n x2buCnt = np.zeros((6, 100), int)\n x2buMicoPayAvg = np.zeros((10000, 1), float)\n for k in range(5):\n for i in range(10000):\n x2buMicroPay[i][k] = (x2buPsi[i][k] / x2busumPsi[k]) * total_micro\n x2buMicoPayAvg[i] += x2buMicroPay[i][k] / 5.0\n x2buCnt[k][int(x2buMicroPay[i][k])] += 1\n for i in range(10000):\n x2buCnt[5][int(x2buMicoPayAvg[i])] += 1\n np.savetxt('x2buMicroPay', x2buMicroPay, fmt='%.10f')\n np.savetxt('x2buMicroPay-Count', x2buCnt, fmt='%d')\n\n # Top-down Design\n x2tdsumDSfi = np.zeros((5, 1), float)\n for k in range(5):\n for i in range(10000):\n x2tdsumDSfi[k] += x2DSf[i][k]\n\n x2tdMicroPay = np.zeros((10000, 5), float)\n # x2tdCnt: Count different interval values\n x2tdCnt = np.zeros((6, 100), int)\n x2tdMicoPayAvg = np.zeros((10000, 1), float)\n\n for k in range(5):\n for i in range(10000):\n x2tdMicroPay[i][k] = (x2DSf[i][k]/ x2tdsumDSfi[k]) * total_micro\n x2tdMicoPayAvg[i] += x2tdMicroPay[i][k] / 5.0\n x2tdCnt[k][int(x2tdMicroPay[i][k])] += 1\n for i in range(10000):\n x2tdCnt[5][int(x2tdMicoPayAvg[i])] += 1\n np.savetxt('x2tdMicroPay', x2tdMicroPay, fmt='%.10f')\n np.savetxt('x2tdMicroPay-Count', x2tdCnt, fmt='%d')\n","sub_path":"Fine-Grained-Privacy-Compensations/Gaussian-Distribution-Fitting/ERATO-based.py","file_name":"ERATO-based.py","file_ext":"py","file_size_in_byte":5108,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"368434723","text":"from django.db import models\nfrom django.utils import timezone\nfrom ckeditor.fields import RichTextField\nfrom imagekit.models import ImageSpecField\nfrom imagekit.processors import ResizeToFill\n\n\nclass Video(models.Model):\n author = models.ForeignKey('auth.User', verbose_name='Автор')\n title = models.CharField(verbose_name='Название', max_length=100, default='Название для видео')\n video_link = models.CharField(verbose_name='Ссылка на видео', max_length=200, default='http://youtube.com/')\n img = models.FileField(verbose_name='Картинка', default='st_img.png')\n img_thumbnail = ImageSpecField(source='img', processors=[ResizeToFill(380, 350)], format='JPEG', options={'quality': 100})\n created_date = models.DateField(verbose_name='Дата', default=timezone.now)\n\n class Meta:\n verbose_name = 'Видео'\n verbose_name_plural = 'Видео'\n\n def __str__(self):\n return self.title\n\n\nclass About(models.Model):\n title = models.CharField(verbose_name='Заголовок', max_length=100, default='Заголовок')\n text = RichTextField(verbose_name='Текст', default='Описание')\n img = models.FileField(verbose_name='Картинка', default='st_img.png')\n img_thumbnail = ImageSpecField(source='img', processors=[ResizeToFill(1550, 380)], format='JPEG', options={'quality': 100})\n\n class Meta:\n verbose_name = 'О нас'\n verbose_name_plural = 'О нас'\n\n def __str__(self):\n return self.text\n\n\nclass Contact(models.Model):\n text = RichTextField(verbose_name='Текст', default='Описание')\n\n class Meta:\n verbose_name = 'Контакты'\n verbose_name_plural = 'Контакты'\n\n def __str__(self):\n return self.text\n","sub_path":"main/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1821,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"37099699","text":"def inverse(num):\n\tif num < 10:\n\t\treturn num\n\tauxNum = num\n\texponent = 0\n\twhile auxNum > 10:\n\t\tauxNum = int(auxNum / 10)\n\t\texponent = exponent + 1\n\tlast = nun % 10 * 10**exponent\n\tnum = int(num/ 10)\n\treturn last + inverse(num)\nprint(inverse(1245234534))","sub_path":"inverse.py","file_name":"inverse.py","file_ext":"py","file_size_in_byte":253,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"553415688","text":"\"\"\"this is main module for routes in flask app\n\"\"\"\nfrom flask import render_template\nfrom flask import flash\nfrom flask import redirect\nfrom flask import request\nfrom flask import url_for\nfrom flask_login import current_user\nfrom flask_login import login_user\nfrom flask_login import logout_user\nfrom flask_login import login_required\n\nfrom datetime import datetime\n\nfrom application.forms import LoginForm\nfrom application.forms import RegistrationForm\nfrom application.forms import EditProfileForm\n\nfrom application import microapp\nfrom application import db\nfrom application.models import User\n\n\n@microapp.before_request\ndef before_request():\n if current_user.is_authenticated:\n current_user.last_seen = datetime.utcnow()\n db.session.commit()\n\n@microapp.route('/')\n@microapp.route('/index')\n@login_required\ndef index():\n\n\n posts = [\n\n {\n 'author': {'username':'John'},\n 'body': 'Beautifull day in portland'\n\n },\n\n {\n 'author': {'username':'John'},\n 'body':'The Avengers movie was so cool'\n }\n\n\n ]\n\n return render_template('index.html', title='home',posts=posts)\n\n\n@microapp.route('/login', methods=['GET', 'POST'])\ndef login():\n if current_user.is_authenticated:\n return redirect(url_for('index'))\n\n form = LoginForm()\n\n if form.validate_on_submit():\n user = User.query.filter_by(username=form.username.data).first()\n print(user,'user')\n if user is None or not user.check_password(form.password.data):\n flash('Invalid creditinals or password')\n return redirect('/index')\n login_user(user, remember=form.remember_me.data)\n next_page = request.args.get('next')\n if not next_page or url_for('index'):\n next_page = url_for('index')\n return redirect(next_page)\n return render_template('login.html', form=form, title= 'Sign in')\n\n\n@microapp.route('/logout')\ndef logout():\n logout_user()\n return redirect(url_for('index'))\n\n\n@microapp.route('/register', methods=['GET', 'POST'])\ndef register():\n if current_user.is_authenticated:\n return redirect(url_for('index'))\n form = RegistrationForm()\n if form.validate_on_submit():\n user = User(username=form.username.data, email=form.email.data)\n user.set_password(form.password.data)\n db.session.add(user)\n db.session.commit()\n flash('Congratulations, you are now a registered user!')\n return redirect(url_for('login'))\n return render_template('register.html', title='Register', form=form)\n\n\n@microapp.route('/user/')\n@login_required\ndef user(username):\n user = User.query.filter_by(username=username).first_or_404()\n posts = [\n {'author': user, 'body': 'Test post #1'},\n {'author': user, 'body': 'Test post #1'}\n ]\n return render_template('user.html', user=user, posts=posts)\n\n\n@microapp.route('/edit_profile', methods=['GET', 'POST'])\n@login_required\ndef edit_profile():\n form = EditProfileForm(current_user.username)\n if form.validate_on_submit():\n current_user.username = form.username.data\n current_user.about_me = form.about_me.data\n db.session.commit()\n flash('Your changes have beensaved')\n return redirect(url_for('edit_profiler'))\n elif request.method == 'GET':\n form.username.data = current_user.username\n form.about_me.data = current_user.about_me\n return render_template('edit_profiler.html', title='Edit profile', form=form)\n\n\n@microapp.route('/follow/')\n@login_required\ndef follow(username):\n user = User.query.filter_by(username=username).first()\n if user is None:\n flash('User {} not found'.format(username))\n return redirect(url_for('index'))\n if user ==current_user:\n flash('You cannot follow yourself')\n return redirect(url_for('user', username=username))\n current_user.follow(user)\n db.session.commit()\n flash('You are following {}'.format(username))\n return redirect(url_for('user', username=username))\n\n\n@app.route('/unfollow/')\n@login_required\ndef unfollow(username):\n user = User.query.filter_by(username=username).first()\n if user is None:\n flash('User {} not found.'.format(username))\n return redirect(url_for('index'))\n if user == current_user:\n flash('You cannot unfollow yourself!')\n return redirect(url_for('user', username=username))\n current_user.unfollow(user)\n db.session.commit()\n flash('You are not following {}.'.format(username))\n return redirect(url_for('user', username=username))","sub_path":"application/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":4643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"130196831","text":"class Solution(object):\n def __init__(self):\n self.result = []\n\n def combine(self, n, k):\n \"\"\"\n :type n: int\n :type k: int\n :rtype: List[List[int]]\n \"\"\"\n if n < k:\n return []\n\n def dfs(left_element, cur, target):\n if target == 0:\n res = cur[:]\n self.result.append(res)\n return\n for idx, ele in enumerate(left_element):\n cur.append(ele)\n dfs(left_element[idx+1:], cur, target-1)\n cur.pop()\n return\n\n all_ele = [i + 1 for i in range(n)]\n dfs(all_ele, [], k)\n\n return self.result\n\n\nif __name__ == '__main__':\n s = Solution()\n print(s.combine(5, 3))\n","sub_path":"LeetCode(Python)/77. Combinations.py","file_name":"77. Combinations.py","file_ext":"py","file_size_in_byte":769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"356485702","text":"# Nelson Crockett\n# nwcrockett@alaska.edu\n# CS 684 HW4\n# 7OCT19\n# main source of programming help https://techwithtim.net/tutorials/\n# game-development-with-python/pygame-tutorial/scoring-health-bars/\n\nimport pygame\nimport chars\nimport random\n\nblack = (0,0,0)\nred = (255, 0, 0)\ngreen = (0, 200, 0)\nbright_green = (0, 255, 0)\n\npygame.init()\n\nscreen_width = 1800\nwin = pygame.display.set_mode((screen_width, screen_width))\npygame.display.set_caption(\"ldjam45 hw5\")\n\nstart = pygame.image.load(\"start.png\")\nstart = pygame.transform.scale(start, (screen_width, screen_width))\n\nbackground = pygame.image.load(\"field.png\")\nbackground = pygame.transform.scale(background, (screen_width, screen_width))\n\n# function for frame rate\nclock = pygame.time.Clock()\n\nrun = True\nmouse = chars.Player(200, 410, 64, 64)\n\ncats = []\ncat_counter = 3\ncheese = []\nfor cat in range(0, cat_counter):\n cat_x = random.randint(200, screen_width - 200)\n cat_y = random.randint(200, screen_width - 200)\n cats.append(chars.Enemy(cat_x, cat_y, 64, 64, screen_width - 100))\n cheese.append(chars.cheese(cat_x, cat_y, 64, 64))\n\n\ndef cycle_window():\n win.blit(background, (0, 0))\n mouse.draw(win)\n cats[0].draw(win, to_player=True, p_x=mouse.x, p_y=mouse.y)\n cheese[0].draw(win)\n\n x = 1650\n y = 10\n pygame.draw.rect(win, red, (x, y, 100, 50))\n screen_mousse = pygame.mouse.get_pos()\n click = pygame.mouse.get_pressed()\n if x + 100 > screen_mousse[0] > x and y + 50 > screen_mousse[1] > y:\n if click[0] == 1:\n pygame.quit()\n quit()\n pygame.display.update()\n\n\n# game is in the loop below\n# array that holds the bullets that the gun fires]\ncaught_cat = False\ncurrent_cat = 0\nintro = True\nwhile run:\n\n # intro loop, green button starts the game, red button quits\n while intro:\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n break\n win.blit(start, (0, 0))\n largeText = pygame.font.Font('freesansbold.ttf', 95)\n textSurface = largeText.render(\"Get the Cheese, Run from the Cats\", True, black)\n rect = textSurface.get_rect()\n rect.center = ((screen_width / 2), 100)\n win.blit(textSurface, rect)\n\n screen_mousse = pygame.mouse.get_pos()\n click = pygame.mouse.get_pressed()\n\n if 1200 + 100 > screen_mousse[0] > 1200 and 500 + 50 > screen_mousse[1] > 500:\n pygame.draw.rect(win, bright_green, (1200, 500, 100, 50))\n if click[0] == 1:\n intro = False\n break\n else:\n pygame.draw.rect(win, green, (1200, 500, 100, 50))\n pygame.draw.rect(win, red, (400, 500, 100, 50))\n\n if 400 + 100 > screen_mousse[0] > 400 and 500 + 50 > screen_mousse[1] > 500:\n if click[0] == 1:\n intro = False\n pygame.quit()\n quit()\n break\n\n pygame.display.update()\n clock.tick(15)\n # End intro loop\n\n # frame rate set to 27 frames per second\n if current_cat == cat_counter:\n break\n\n clock.tick(27)\n\n if mouse.x - 10 < cheese[0].x < mouse.x + 10 and mouse.y - 10 < cheese[0].y < mouse.y + 10:\n cats.pop(0)\n cheese.pop(0)\n mouse.go_faster()\n mouse.up_sprite_size()\n current_cat += 1\n caught_cat = False\n\n if mouse.x - 10 < cats[0].x < mouse.x + 10 and mouse.y - 10 < cats[0].y < mouse.y + 10:\n pygame.QUIT\n\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n run = False\n\n keys = pygame.key.get_pressed()\n\n if keys[pygame.K_LEFT] and mouse.x > mouse.vel:\n mouse.x -= mouse.vel\n mouse.left = True\n mouse.right = False\n mouse.standing = False\n\n elif keys[pygame.K_RIGHT] and mouse.x < screen_width - mouse.width:\n mouse.x += mouse.vel\n mouse.left = False\n mouse.right = True\n mouse.standing = False\n\n elif keys[pygame.K_UP] and mouse.y > mouse.vel:\n mouse.y -= mouse.vel\n\n elif keys[pygame.K_DOWN] and mouse.y < screen_width - mouse.height:\n mouse.y += mouse.vel\n\n else:\n mouse.standing = True\n mouse.walk_count = 0\n\n cycle_window()\n\npygame.QUIT\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4295,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"646473730","text":"#!/usr/local/bin/python3\n\nfrom cgitb import enable \nenable()\n\nfrom os import environ\n\nfrom cgi import FieldStorage, escape\nimport pymysql as db\n\n\n \nprint('Content-Type: text/plain')\nprint()\n\nname = environ['QUERY_STRING']\nname2 = name.split(\"=\")\nname3 = name2[0]\n\nform_data = FieldStorage()\nselection = form_data.getlist(name3)\n\ntry: \n connection = db.connect('cs1.ucc.ie','dr13','chujohqu','csdipact2017_dr13')\n cursor = connection.cursor(db.cursors.DictCursor)\n cursor.execute(\"\"\"SELECT *\n FROM questions WHERE ID = %s\"\"\" %(name3))\n try:\n if cursor.rowcount == 0:\n print('Undergoing scheduled mainenance. Please call back later')\n else:\n\n for row in cursor.fetchall():\n if selection[0] != row['CorrectAnswer']:\n print(\"Incorrect\")\n else:\n print('Correct' )\n except IndexError:\n print('Error!!')\n cursor.close() \n connection.close()\nexcept db.Error:\n print('Sorry. We are experiencing technical difficulties. Please Try Again Later')\n \n \n\n","sub_path":"compare.py","file_name":"compare.py","file_ext":"py","file_size_in_byte":1114,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"244046839","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n'''\n Author: kun.wang\n Create: 2013-01-21\n'''\nimport os, sys, string\nimport PyQt4.QtCore as QtCore\nimport PyQt4.QtGui as QtGui\n\nfrom pcore import *\nfrom pcloud import *\n\n\n\nclass ProjectItem(QtGui.QListWidgetItem):\n def __init__(self, name, project_id, parent = None):\n super(ProjectItem, self).__init__(parent)\n self.name = name\n self.project_id = project_id\n self.updateUI()\n\n def updateUI(self):\n self.setText(self.name)\n\n\nclass TaskItem(QtGui.QListWidgetItem):\n def __init__(self, task_name, task_id, asset_name, asset_id, parent = None):\n super(TaskItem, self).__init__(parent)\n self.task_name = task_name\n self.task_id = task_id\n self.asset_name = asset_name\n self.asset_id = asset_id\n\n self.updateUI()\n\n def updateUI(self):\n self.setText('Asset : <%s> * Task : <%s>' % (self.asset_name, self.task_name))\n\n\nclass Goto(QtGui.QWidget):\n def __init__(self, parent = None):\n super(Goto, self).__init__(parent)\n self.setWindowTitle('Pillars Goto')\n self.resize(350, 100)\n\n self.project_list = None\n\n self.createMainlayout()\n self.createMenubar()\n self.createLoginArea()\n self.createGotoTools()\n\n self.loadInformation()\n\n def createMainlayout(self):\n self.main_layout = QtGui.QVBoxLayout()\n self.setLayout(self.main_layout)\n \n def createMenubar(self):\n self.actHoudiniTool = QtGui.QAction('Houdini Tools', self)\n self.actAssetTool = QtGui.QAction('Asset Tools', self)\n self.connect(self.actHoudiniTool, QtCore.SIGNAL('triggered()'), self.showHoudiniTool)\n self.connect(self.actAssetTool, QtCore.SIGNAL('triggered()'), self.showAssetTool)\n self.tool_menu = QtGui.QMenu('Tools')\n self.tool_menu.addAction(self.actHoudiniTool)\n self.tool_menu.addAction(self.actAssetTool)\n\n \n self.help_menu = QtGui.QMenu('Help')\n\n self.pmenu = QtGui.QMenuBar()\n self.pmenu.addMenu(self.tool_menu)\n self.pmenu.addMenu(self.help_menu)\n self.main_layout.addWidget(self.pmenu)\n \n def createLoginArea(self):\n self.lServer = QtGui.QLabel(\"Server :\")\n self.tServer = QtGui.QLineEdit('')\n self.lPort = QtGui.QLabel(\"Port :\")\n self.tPort = QtGui.QLineEdit('')\n self.tPort.setMaximumWidth(50)\n \n self.lCompany = QtGui.QLabel(\"Company :\")\n self.lAuthor = QtGui.QLabel('User :')\n self.lPassword = QtGui.QLabel('Password :')\n self.tCompany = QtGui.QLineEdit('')\n self.tAuthor = QtGui.QLineEdit('')\n self.tPassword = QtGui.QLineEdit('')\n self.tPassword.setEchoMode(QtGui.QLineEdit.Password)\n \n self.loginButton = QtGui.QPushButton(' Login ')\n self.loginButton.setMinimumHeight(40)\n \n self.gridLayout = QtGui.QGridLayout()\n # self.gridLayout.setContentsMargins(50, 20, 50, 20)\n self.gridLayout.setColumnStretch(1, 1)\n self.gridLayout.addWidget(self.lServer, 0, 0, QtCore.Qt.AlignRight)\n self.gridLayout.addWidget(self.tServer, 0, 1)\n self.gridLayout.addWidget(self.lPort, 0, 2, QtCore.Qt.AlignRight)\n self.gridLayout.addWidget(self.tPort, 0, 3)\n self.gridLayout.addWidget(self.lCompany, 1, 0, QtCore.Qt.AlignRight)\n self.gridLayout.addWidget(self.tCompany, 1, 1, 1, 3)\n self.gridLayout.addWidget(self.lAuthor, 2, 0, QtCore.Qt.AlignRight)\n self.gridLayout.addWidget(self.tAuthor, 2, 1, 1, 3)\n self.gridLayout.addWidget(self.lPassword, 3, 0, QtCore.Qt.AlignRight)\n self.gridLayout.addWidget(self.tPassword, 3, 1, 1, 3)\n # self.gridLayout.addWidget(self.loginButton, 4, 0, 1, 4, QtCore.Qt.AlignCenter)\n \n self.connect(self.loginButton, QtCore.SIGNAL('clicked()'), self.login)\n self.main_layout.addLayout(self.gridLayout)\n self.main_layout.addWidget(self.loginButton)\n \n def createGotoTools(self):\n self.project_list = QtGui.QListWidget()\n self.project_list.setMinimumHeight(150)\n self.gotoProjButton = QtGui.QPushButton('Goto Project')\n self.gotoProjButton.setMinimumHeight(40)\n self.connect(self.gotoProjButton, QtCore.SIGNAL('clicked()'), self.gotoProject)\n self.isProjectListShow = False\n\n self.task_list = QtGui.QListWidget()\n self.task_list.setMinimumHeight(300)\n self.gotoTaskButton = QtGui.QPushButton('Goto Task')\n self.gotoTaskButton.setMinimumHeight(40)\n self.connect(self.gotoTaskButton, QtCore.SIGNAL('clicked()'), self.gotoTask)\n self.isTaskListShow = False\n \n\n def settingFile(self):\n return os.path.join(PILLARS_HOME, 'goto.txt')\n\n def loadInformation(self):\n if not os.path.isfile(self.settingFile()):\n return False\n\n try:\n f = open(self.settingFile(), 'r')\n setting = json.loads(f.read())\n f.close()\n self.tServer.setText(setting.get('server', ''))\n self.tPort.setText(setting.get('port', ''))\n self.tCompany.setText(setting.get('company', ''))\n self.tAuthor.setText(setting.get('user', ''))\n self.tPassword.setText(setting.get('password', ''))\n return True\n except:\n return False\n\n def saveInformation(self):\n info = {}\n info['server'] = unicode(self.tServer.text())\n info['port'] = unicode(self.tPort.text())\n info['company'] = unicode(self.tCompany.text())\n info['user'] = unicode(self.tAuthor.text())\n info['password'] = unicode(self.tPassword.text())\n\n f = open(self.settingFile(), 'w')\n f.write(json.dumps(info, indent = 4))\n f.close()\n return True\n\n def showProjectList(self):\n if not self.isProjectListShow:\n self.main_layout.addWidget(self.project_list)\n self.main_layout.addWidget(self.gotoProjButton)\n self.isProjectListShow = True\n\n def showTaskList(self):\n if not self.isTaskListShow:\n self.main_layout.addWidget(self.task_list)\n self.main_layout.addWidget(self.gotoTaskButton)\n self.isTaskListShow = True\n \n def login(self):\n server = str(self.tServer.text())\n port = int(str(self.tPort.text()))\n cookie = PSetting.cookie\n user = unicode(self.tAuthor.text())\n password = unicode(self.tPassword.text())\n\n os.environ['PILLARS_CLOUD'] = '%s:%d' % (server, port)\n os.environ['PILLARS_COOKIE'] = cookie\n os.environ['PILLARS_USER'] = str(user)\n\n PillarsCloud.setServer(server, port)\n PillarsCloud.setCookie(cookie)\n PillarsCloud.login(user, password)\n\n self.showProjectList()\n self.loadProject()\n\n self.saveInformation()\n\n def loadProject(self):\n self.project_list.clear()\n projs = PillarsCloud.getProjects()\n for proj in projs:\n item = ProjectItem(proj['name'], proj['id'])\n self.project_list.addItem(item)\n return True\n\n def loadTask(self):\n item = self.project_list.currentItem()\n if not item:\n QtGui.QMessageBox.about(self, u\"错误\", u\"请选择项目\")\n return False\n\n self.task_list.clear()\n \n p_id = item.project_id\n p_name = item.name\n os.environ['PILLARS_PROJECT_ID'] = p_id.replace('p', '')\n os.environ['PILLARS_PROJECT'] = p_name\n if os.name == \"nt\":\n os.environ['JOB'] = 'P:/Pub/Project/%s' % p_name\n elif os.name == \"posix\":\n os.environ['JOB'] = '/Pillars/Pub/Project/%s' % p_name\n\n tasks = PillarsCloud.myTask(str(self.tAuthor.text()), project_id = int(os.environ['PILLARS_PROJECT_ID']))\n if not tasks:\n return False\n for task in tasks:\n asset = PillarsCloud.getAsset(task_id = task['id'])\n t_item = TaskItem(task['name'], task['id'], asset['name'], asset['id'])\n self.task_list.addItem(t_item)\n return True\n\n def gotoProject(self):\n if self.loadTask():\n self.showTaskList()\n return True\n else:\n return False\n\n def gotoTask(self):\n task_item = self.task_list.currentItem()\n if not task_item:\n QtGui.QMessageBox.about(self, u\"错误\", u\"请选择任务\")\n return False\n\n os.environ['PILLARS_ASSET'] = task_item.asset_name\n os.environ['PILLARS_ASSET_ID'] = str(task_item.asset_id)\n os.environ['PILLARS_TASK'] = task_item.task_name\n os.environ['PILLARS_TASK_ID'] = str(task_item.task_id)\n\n os.environ['PILLARS_ASSET_PATH'] = '%s/asset/%s' % (os.environ['JOB'], os.environ['PILLARS_ASSET'])\n\n if os.name == \"nt\":\n os.system('start cmd')\n return True\n elif os.name == \"posix\":\n return True\n\n def showHoudiniTool(self):\n import htools\n self.ht = htools.HTools()\n self.ht.show()\n\n def showAssetTool(self):\n import asset\n self.asset_tool = asset.AssetEditor()\n self.asset_tool.resize(600, 500)\n self.asset_tool.show()\n self.asset_tool.loadProject()\n\n\ndef main():\n app = QtGui.QApplication(sys.argv)\n ui = Goto()\n ui.show()\n sys.exit(app.exec_())\n\nif __name__ == \"__main__\":\n main()","sub_path":"cgPipeline/old/Mango0.4/goto.py","file_name":"goto.py","file_ext":"py","file_size_in_byte":9466,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"612060375","text":"\r\nfrom pytube import YouTube, Playlist\r\nimport ffmpeg\r\nimport re\r\nimport os\r\nimport shutil\r\n\r\nclass YouTube_Downloader() :\r\n\r\n def __init__( self, save_path='' ) :\r\n self.save_path = save_path\r\n self.caption_title_adjust = '[\\\\\\/:*?\"<>|\\[\\]\\(\\)\\【\\】]'\r\n\r\n if os.path.exists( 'combine_temp_workplace' ) :\r\n shutil.rmtree( 'combine_temp_workplace' )\r\n\r\n def print_information( self, url, show_details=True, show_subtitle=True, show_stream=True,\r\n only_audio=True, only_video=True, show_progressive=True ) :\r\n try :\r\n yt = YouTube( url )\r\n\r\n if show_details :\r\n ## Showing details\r\n print( '------details--------------------' )\r\n print( \"Title: \", yt.title )\r\n print( \"Number of views: \", yt.views )\r\n print( \"Length of video: \", yt.length )\r\n print( \"Rating of video: \", yt.rating )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if show_subtitle :\r\n ## printing all Subtitle/Caption\r\n print( '------captions-------------------' )\r\n for caption in yt.captions.all() :\r\n print( caption ) ## name, code\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if show_stream :\r\n ## printing all the available streams\r\n print( '------streams--------------------' )\r\n for stream in yt.streams :\r\n print( stream )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if only_audio :\r\n ## filter out audio-only streams\r\n print( '------streams (only_audio)-------' )\r\n for stream in yt.streams.filter(only_audio=True) :\r\n print( stream )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if only_video :\r\n ## filter out video-only streams\r\n print( '------streams (only_video)-------' )\r\n for stream in yt.streams.filter(only_video=True) :\r\n print( stream )\r\n print( '---------------------------------', end='\\n\\n' )\r\n \r\n if show_progressive :\r\n ## filter out progressive streams\r\n print( '------streams (progressive)------' )\r\n for stream in yt.streams.filter(progressive=True) :\r\n print( stream )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n except Exception as e :\r\n raise e\r\n\r\n def single_download( self, url, itag=None, download=True, language_code=None, filter_dic=None ) :\r\n\r\n ## filter_dic\r\n ##\r\n ## type = None | \"video\" | \"audio\"\r\n ## subtype = None | \"mp4\" | ...\r\n ## res = None | \"1080p\" | \"720p\" | ...\r\n ## fps = None | \"30fps\" | ...\r\n ## abr = None | \"128kbps\" | \"50kbps\" | ... \r\n ## only_audio = None | True | False\r\n ## only_video = None | True | False\r\n ## progressive = None | True | False\r\n\r\n try :\r\n yt = YouTube( url )\r\n\r\n if download :\r\n \r\n ys = None\r\n\r\n if itag != None :\r\n ys = yt.streams.get_by_itag( itag )\r\n elif filter_dic != None :\r\n ys = yt.streams.filter( type = filter_dic['type'],\r\n subtype = filter_dic['subtype'],\r\n res = filter_dic['res'],\r\n fps = filter_dic['fps'],\r\n abr = filter_dic['abr'],\r\n only_audio = filter_dic['only_audio'],\r\n only_video = filter_dic['only_video'],\r\n progressive = filter_dic['progressive'] )\r\n ys = ys[0]\r\n else :\r\n ys = yt.streams.filter( progressive=True ).get_highest_resolution()\r\n \r\n print( '------target stream--------------' )\r\n print( ys )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n ys.download( self.save_path ) if self.save_path != None else ys.download()\r\n \r\n if language_code != None : \r\n ## download captions\r\n self.captions_download( url, language_code )\r\n\r\n print( 'Done' if download else 'Done ( no download )', end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n def captions_download( self, url, language_code ) :\r\n try :\r\n yt = YouTube( url )\r\n\r\n caption = yt.captions.get_by_language_code( language_code )\r\n caption = caption.generate_srt_captions() ## convert to the srt format\r\n\r\n title = yt.title \r\n\r\n if self.caption_title_adjust != None :\r\n title = re.sub( self.caption_title_adjust, '_', yt.title )\r\n\r\n caption_file = None\r\n\r\n if self.save_path == None :\r\n caption_file = open( ( title + \".srt\" ), \"w\", encoding=\"utf-8\" )\r\n else :\r\n caption_file = open( ( self.save_path + \"\\\\\" + title + \".srt\" ), \"w\", encoding=\"utf-8\" )\r\n\r\n caption_file.write( caption )\r\n caption_file.close()\r\n\r\n print( 'Successfully download captions :', title + \".srt\", end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n def create_ip_file( self, playlist_url, file_name='ip_file.txt' ) :\r\n try :\r\n ip_file = None\r\n\r\n if self.save_path == None :\r\n ip_file = open( file_name, 'w' )\r\n else :\r\n ip_file = open( ( self.save_path + '\\\\' + file_name ), 'w' )\r\n\r\n playlist = Playlist( playlist_url )\r\n\r\n ## this fixes the empty playlist.videos list\r\n playlist._video_regex = re.compile( r\"\\\"url\\\":\\\"(/watch\\?v=[\\w-]*)\" )\r\n\r\n total = len( playlist.video_urls )\r\n print( 'video num :', total, end='\\n\\n' )\r\n\r\n num = 1\r\n\r\n for url in playlist.video_urls:\r\n\r\n print( num, '/', total, ':', url, end='\\n\\n' ) \r\n ip_file.write( url )\r\n\r\n if num != total :\r\n ip_file.write( '\\n' )\r\n\r\n num += 1\r\n\r\n print( 'Create : ' + file_name, end='\\n\\n' )\r\n\r\n ip_file.close()\r\n except Exception as e :\r\n raise e\r\n\r\n def download_from_file( self, file_name, download=True, language_code=None, filter_dic=None ) :\r\n try :\r\n ip_file = open( file_name, 'r' )\r\n num = 1\r\n\r\n for url in ip_file.readlines() :\r\n print( num ) \r\n print( url.strip(), end='\\n\\n' ) ## delete the '\\n' at last\r\n self.single_download( url=url, \r\n download=download, \r\n language_code=language_code,\r\n filter_dic=filter_dic )\r\n num += 1\r\n\r\n ip_file.close()\r\n\r\n print( 'All Done', end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n def playlist_download( self, playlist_url, download=True, language_code=None, filter_dic=None,\r\n limit_num=None ) :\r\n try :\r\n playlist = Playlist( playlist_url )\r\n\r\n ## this fixes the empty playlist.videos list\r\n playlist._video_regex = re.compile( r\"\\\"url\\\":\\\"(/watch\\?v=[\\w-]*)\" )\r\n\r\n total = len( playlist.video_urls )\r\n print( 'video num :', total, end='\\n\\n' )\r\n\r\n num = 1\r\n\r\n for url in playlist.video_urls:\r\n print( num, '/', total, ':', url, end='\\n\\n' ) \r\n self.single_download( url=url, \r\n download=download, \r\n language_code=language_code,\r\n filter_dic=filter_dic )\r\n if ( limit_num != None ) and ( num == limit_num ) :\r\n print( 'Reach limit_num ! (', limit_num, ')', end='\\n\\n' )\r\n break\r\n num += 1\r\n\r\n print( 'All Done', end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n def download_single_video_audio_combine( self, url, audio_itag, video_itag, download=True, language_code=None ) :\r\n try :\r\n yt = YouTube( url )\r\n\r\n if download :\r\n\r\n audio_ys = yt.streams.get_by_itag( audio_itag )\r\n video_ys = yt.streams.get_by_itag( video_itag )\r\n \r\n print( '------combine target streams--------------' )\r\n print( audio_ys )\r\n print( video_ys )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if os.path.exists( 'combine_temp_workplace' ) :\r\n shutil.rmtree( 'combine_temp_workplace' )\r\n\r\n os.mkdir( 'combine_temp_workplace' )\r\n os.mkdir( 'combine_temp_workplace\\\\audio' )\r\n os.mkdir( 'combine_temp_workplace\\\\video' )\r\n\r\n audio_ys.download( 'combine_temp_workplace\\\\audio' ) \r\n video_ys.download( 'combine_temp_workplace\\\\video' ) \r\n\r\n audio_name = 'combine_temp_workplace\\\\audio\\\\' + os.listdir( 'combine_temp_workplace\\\\audio' )[0]\r\n video_name = 'combine_temp_workplace\\\\video\\\\' + os.listdir( 'combine_temp_workplace\\\\video' )[0]\r\n\r\n input_audio = ffmpeg.input( audio_name )\r\n input_video = ffmpeg.input( video_name )\r\n\r\n title = yt.title \r\n\r\n if self.caption_title_adjust != None :\r\n title = re.sub( self.caption_title_adjust, '_', yt.title )\r\n \r\n ffmpeg.concat( input_video, input_audio, v=1, a=1 ).output( self.save_path + '\\\\' + title + '.mp4' ).run( overwrite_output=True )\r\n # .run( overwrite_output=True, capture_stdout=True, capture_stderr=True )\r\n\r\n print()\r\n\r\n shutil.rmtree( 'combine_temp_workplace' )\r\n \r\n if language_code != None : \r\n ## download captions\r\n self.captions_download( url, language_code )\r\n\r\n print( 'Done' if download else 'Done ( no download )', end='\\n\\n' )\r\n\r\n # except ffmpeg.Error as e :\r\n # print( e.stdout, end='\\n\\n' )\r\n # print( e.stderr, end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n def multi_video_audio_combine( self, video_files_path, audio_files_path ) :\r\n try :\r\n video_files_list = os.listdir( video_files_path )\r\n audio_files_list = os.listdir( audio_files_path )\r\n\r\n print( '------video files----------------' )\r\n print( video_files_list )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n print( '------audio files----------------' )\r\n print( audio_files_list )\r\n print( '---------------------------------', end='\\n\\n' )\r\n\r\n if len( video_files_list ) == len( audio_files_list ) :\r\n\r\n if all( video_files_list[i].split( '.' )[0] == audio_files_list[i].split( '.' )[0]\r\n for i in range( len( video_files_list ) ) ) :\r\n\r\n total = len( video_files_list )\r\n num = 1\r\n\r\n for i in range( len( video_files_list ) ) :\r\n print( num, '/', total, end='\\n\\n' )\r\n\r\n input_video = ffmpeg.input( video_files_path + '\\\\' + video_files_list[i] )\r\n input_audio = ffmpeg.input( audio_files_path + '\\\\' + audio_files_list[i] )\r\n \r\n title = video_files_list[i]\r\n\r\n if self.caption_title_adjust != None :\r\n title = re.sub( self.caption_title_adjust, '_', title )\r\n\r\n ffmpeg.concat( input_video, input_audio, v=1, a=1 ).output( self.save_path + '\\\\' + title ).run( overwrite_output=True )\r\n # .run( overwrite_output=True, capture_stdout=True, capture_stderr=True )\r\n\r\n print()\r\n num += 1\r\n\r\n else :\r\n print( 'ERROR : Without own extension, some file names are not equal !', end='\\n\\n' )\r\n\r\n else :\r\n print( 'ERROR : File numbers are not equal !', end='\\n\\n' )\r\n\r\n print( 'All Done', end='\\n\\n' )\r\n\r\n # except ffmpeg.Error as e :\r\n # print( e.stdout, end='\\n\\n' )\r\n # print( e.stderr, end='\\n\\n' )\r\n except Exception as e :\r\n raise e\r\n\r\n\r\n","sub_path":"YouTube_Downloader_v4/data/tool/YouTube_DLer_v4.py","file_name":"YouTube_DLer_v4.py","file_ext":"py","file_size_in_byte":13240,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"253181434","text":"from dataclasses import dataclass\nfrom typing import Optional\n\nfrom thenewboston_node.core.utils.dataclass import cover_docstring, revert_docstring\nfrom thenewboston_node.core.utils.types import hexstr\n\nfrom ..node import Node\nfrom .base import SignedChangeRequestMessage\n\n\n@revert_docstring\n@dataclass\n@cover_docstring\nclass NodeDeclarationSignedChangeRequestMessage(SignedChangeRequestMessage):\n\n node: Node\n\n @classmethod\n def create(\n cls, identifier: hexstr, network_addresses: list[str], fee_amount: int, fee_account: Optional[hexstr] = None\n ):\n return cls(\n node=Node(\n identifier=identifier,\n network_addresses=network_addresses,\n fee_amount=fee_amount,\n fee_account=fee_account,\n )\n )\n\n def validate(self):\n self.node.validate()\n","sub_path":"thenewboston_node/business_logic/models/signed_change_request_message/node_declaration.py","file_name":"node_declaration.py","file_ext":"py","file_size_in_byte":864,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"338368275","text":"# -*- coding: utf-8 -*-\nfrom torch.nn import Module, Identity, Conv1d, MaxPool1d, BatchNorm1d, ReLU\nimport torch\n\nclass InceptionModule(Module):\n \n def canUseBottleneck(self, inputTensor):\n return self.use_bottleneck and int(inputTensor.shape[-1]) > 1;\n \n def __init__(self, in_channels, out_channels, stride, kernel_size, nb_filters, bottleneck_size, use_bottleneck):\n super().__init__()\n \n self.bottleneck_size = bottleneck_size;\n self.stride = stride;\n self.kernel_size = kernel_size;\n self.nb_filters = nb_filters;\n self.use_bottleneck = use_bottleneck;\n \n self.input = Identity()\n \n self.bottleneckInput = Conv1d(in_channels, self.bottleneck_size, kernel_size=1, padding=0, bias=False);\n \n # kernel_size_s = [3, 5, 8, 11, 17]\n kernel_size_s = [self.kernel_size // (2 ** i) for i in range(3)]\n #print(\"kernel size\");\n #print(kernel_size_s)\n\n self.conv_list = []\n self.temp_input = self.bottleneck_size\n for i in range(len(kernel_size_s)):\n setattr(self, \"cov_parallel_%d\" % kernel_size_s[i],Conv1d(self.temp_input, self.nb_filters, kernel_size=kernel_size_s[i], stride=self.stride, padding=0,bias=False))\n self.temp_input= self.nb_filters;\n self.conv_list.append(getattr(self,\"cov_parallel_%d\" % kernel_size_s[i])) \n self.max_pool = MaxPool1d(kernel_size=3, stride=self.stride, padding=0)\n \n self.conv = Conv1d(self.bottleneck_size, self.nb_filters, kernel_size=1, padding=0, bias=False)\n \n self.bn = BatchNorm1d(self.nb_filters)\n \n def forward(self, x):\n if self.canUseBottleneck(x):\n x = self.bottleneckInput(x);\n parallelLayers = [];\n \n for layer in self.conv_list:\n parallelLayers.append(layer(x));\n res = self.max_pool(x);\n res = self.conv(res);\n \n parallelLayers.append(res)\n \n x = torch.cat(parallelLayers, 2)\n \n x = self.bn(x);\n \n x = ReLU()(x);\n ##print(\"final output\");\n #print(x.shape) \n return x;\n \n \n ","sub_path":"dac-noise/networks/inceptionModule.py","file_name":"inceptionModule.py","file_ext":"py","file_size_in_byte":2448,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"340481914","text":"from __future__ import unicode_literals\nfrom django.shortcuts import render_to_response\nfrom django.http import HttpResponse\nfrom django.http import HttpResponseRedirect\nfrom django.utils import timezone\nfrom django.shortcuts import render\nfrom .models import Post\nfrom . import models\nimport datetime\n\n\n# Create your views here.\ndef index(request):\n posts = Post.objects.filter(published_date__lte=timezone.now()).order_by('-published_date')\n return render(request, 'blog/post_list.html', {'posts': posts})\n\ndef post_page(request):\n\treturn render_to_response('blog/post_new.html')\n\ndef edit_page(request):\n\ttry:\n\t\tpid=request.GET[\"id\"]\n\t\tposts = Post.objects.filter(p_id = pid)\n\t\treturn render(request, 'blog/post_edit.html', {'posts': posts})\n\texcept:\n\t\treturn HttpResponseRedirect('/blog/')\n\ndef view(request):\n\ttry:\n\t\tpid=request.GET[\"id\"]\n\t\tposts = Post.objects.filter(p_id = pid)\n\t\ttitle= \"\"\n\n\t\tfor post in posts:\n\t\t\ttitle = post.title\n\n\t\treturn render(request, 'blog/post_view.html', {'posts': posts, 'title': title})\n\texcept:\n\t\treturn HttpResponseRedirect('/blog/')\n\t\t\n\ndef save(request):\n user = request.POST.get(\"tb_user\")\n ti = request.POST.get(\"tb_title\")\n tx = request.POST.get(\"tb_text\")\n created = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n publish = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n\n n_post = models.Post(author = user, title = ti, text = tx, created_date = created, published_date = publish)\n n_post.save()\n\n return HttpResponseRedirect('/blog/')\n\ndef update(request):\n try:\n pi = request.POST.get(\"tb_pid\")\n thePost = Post.objects.get(p_id=pi)\n thePost.author = request.POST.get(\"tb_user\")\n thePost.title = request.POST.get(\"tb_title\")\n thePost.text = request.POST.get(\"tb_text\")\n thePost.published_date = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n thePost.save()\n return HttpResponseRedirect('/blog/')\n except Exception as e:\n return HttpResponseRedirect('/blog/')\n\n\ndef delete(request):\n un = request.GET[\"id\"]\n \n try:\n post = Post.objects.get(p_id=un)\n post.delete()\n return HttpResponseRedirect('/blog/')\n except:\n return HttpResponseRedirect('/blog/')","sub_path":"0430/0430-2/ntust/mysite/blog/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2261,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"570554137","text":"# This script allow to scrape www.whoscored.com and extract\n# players statistics during a season.\n# The Web Scraper is partly inspired by a solution available on this\n# link : https://github.com/cboutaud/whoscraped. The structure of\n# the website's pages has changed since then, we had to ajust some\n# features and increase the robustness by adding an exception management\n# mechanism for when pages are not proprely charged.\n#########################################################################\n\n# Imports\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom bs4 import BeautifulSoup\nimport time\nimport csv\nimport os\n\n# Base URL\nbaseURL = \"https://www.whoscored.com/Teams/\"\n\n# EPL Teams\neplTeams = [\n \"167\", # MCFC / Manchester City\n \"32\", # MUFC / Manchester United\n \"30\", # THFC / Tottenham Hotspur\n \"26\", # LFC / Liverpool FC\n \"15\", # CFC / Chelsea FC\n \"13\", # AFC / Arsenal FC\n \"184\", # BFC / Burnley FC\n \"31\", # EFC / Everton FC\n \"14\", # LC / Leicester City\n \"23\", # NUFC / Newcastle United\n \"162\", # CP / Crystal Palace\n \"183\", # BOU / AFC Bournemouth\n \"29\", # WHUFC / West Ham United\n \"27\", # WAT / Watford FC\n \"211\", # BHA / Brighton & Hove Albion\n \"166\", # HUD / Huddersfield Town\n \"18\", # SFC / Southampton FC\n \"161\", # WWFC / Wolverhampton Wanderers\n \"188\", # CCFC / Cardiff City\n \"170\", # FFC / Fulham FC\n \"175\", # WBAFC / West Bromwich Albion\n \"16\", # XXX / Sunderland\n \"168\", # XXX / Norwich\n \"259\", # XXX / Swansea\n \"214\", # XXX / Hull\n \"21\", # XXX / Middlesbrough\n \"96\", # XXX / Stoke\n \"163\", # SUFC / Sheffield United\n]\n\n# Mapping Team ID -> Team Name\nteams_ID_to_name = {\n \"167\": \"MCFC\",\n \"32\": \"MUFC\",\n \"30\": \"THFC\",\n \"26\": \"LFC\",\n \"15\": \"CFC\",\n \"13\": \"AFC\",\n \"184\": \"BFC\",\n \"31\": \"EFC\",\n \"14\": \"LC\",\n \"23\": \"NUFC\",\n \"162\": \"CP\",\n \"183\": \"BOU\",\n \"29\": \"WHUFC\",\n \"27\": \"WAT\",\n \"211\": \"BHA\",\n \"166\": \"HUD\",\n \"18\": \"SFC\",\n \"161\": \"WWFC\",\n \"188\": \"CCFC\",\n \"170\": \"FFC\",\n \"175\": \"WBAFC\",\n \"16\": \"XXX1\",\n \"168\": \"XXX2\",\n \"259\": \"XXX3\",\n \"214\": \"XXX4\",\n \"21\": \"XXX5\",\n \"96\": \"XXX6\",\n \"163\": \"SUFC\",\n}\n\n# archiveUrl = \"/Archive?stageId=17590\" # season 2019/2020\n# archiveUrl = \"/Archive?stageId=16368\" # season 2018/2019\n# archiveUrl = \"/Archive?stageId=15151\" # season 2017/2018\n# archiveUrl = \"/Archive?stageId=13796\" # season 2016/2017\n# archiveUrl = \"/Archive?stageId=12496\" # season 2015/2016\n\narchiveUrls = [\n \"/Archive?stageId=17590\", # season 2019/2020\n \"/Archive?stageId=16368\", # season 2018/2019\n \"/Archive?stageId=15151\", # season 2017/2018\n \"/Archive?stageId=13796\", # season 2016/2017\n \"/Archive?stageId=12496\", # season 2015/2016\n]\n\narchiveUrls_to_season = {\n \"/Archive?stageId=17590\": \"2019_2020\",\n \"/Archive?stageId=16368\": \"2018_2019\",\n \"/Archive?stageId=15151\": \"2017_2018\",\n \"/Archive?stageId=13796\": \"2016_2017\",\n \"/Archive?stageId=12496\": \"2015_2016\",\n}\n\nos.chdir(\"data/raw\")\noutputfile = open(\"players_stats4.csv\", \"w\")\ncsv_writer = csv.writer(outputfile)\ncsv_writer.writerow(\n [\n \"Name\",\n \"Team\",\n \"Nat\", # Nationality\n \"Age\",\n \"Pos\", # Position in field\n \"Height\",\n \"Weight\",\n \"Apps(Subs)\", # Appearances\n \"Mins\", # Mins played\n \"Goals\", # Total goals per season \n \"Assists\", # Total assists per season\n \"Yel\", # Yellow cards during the season\n \"Red\", # Red cards during the season\n \"SpG\", # Shots per game\n \"PS%\", # Pass success percentage per season\n \"AerWon\", # Aerial duels won per game\n \"MoM\", # Man of the match per season\n \"Tackles\", # Tackles per game\n \"Inter\", # Interceptions per game\n \"Fouls\", # Fouls per game\n \"OffW\", # Offsides won per game\n \"Clear\", # Clearances per game\n \"DrbPast\", # Dribbled past per game\n \"Blocks\", # Outfielder block per game\n \"OwnG\", # Own goals per season\n \"KeyP\", # Key passes per game\n \"Drb\", # Dribbles per game\n \"Fouled\", # Fouled per game\n \"Off\", # Offsides per game\n \"Disp\", # Dspossessed per game\n \"UnsT\", # Bad control per game\n \"AvgP\", # Passes per game\n \"Crosses\", # Crosses per game\n \"LongB\", # Long balls per game\n \"ThrB\", # Through balls per game\n \"Rat\", # Rating\n \"Season\", # Season\n ]\n)\nfor team in eplTeams:\n print(\"Currently getting {}'s players data ...\".format(teams_ID_to_name[team]))\n\n for archiveUrl in archiveUrls:\n\n print(\"Season : \" + archiveUrls_to_season[archiveUrl])\n\n ALLTHEDAMNPLAYERS = {}\n\n # URL\n finalURL = baseURL + team + archiveUrl\n\n # Connect webdriver\n browser = webdriver.Firefox()\n browser.set_window_size(1920, 1080)\n browser.get(finalURL)\n time.sleep(3)\n\n # Different tables\n tableNames = tableNames = [\"summary\", \"defensive\", \"offensive\", \"passing\"]\n\n # Get all tables\n tables = []\n\n for tableName in tableNames:\n while True:\n element = WebDriverWait(browser, 10).until(\n EC.presence_of_element_located(\n (\n By.CSS_SELECTOR,\n \"a[href*='#team-squad-archive-stats-\" + tableName + \"']\",\n )\n )\n )\n time.sleep(10)\n browser.execute_script(\"arguments[0].click();\", element)\n time.sleep(10)\n\n # Get content\n content = browser.page_source\n soup = BeautifulSoup(\"\".join(content), \"lxml\")\n table = soup.find(\"div\", {\"id\": \"statistics-table-\" + tableName}).find(\n \"tbody\", {\"id\": \"player-table-statistics-body\"}\n )\n try:\n test_NoneType = table.findAll(\"tr\")[0].findAll(\"td\")\n break\n except AttributeError:\n print(\n \"An error with the page has occured, refreshing and trying again ...\"\n )\n browser.refresh()\n time.sleep(10)\n continue\n tables.append(table)\n print(\"\\t\" + tableName + \" \" + u\"\\u2713\")\n\n browser.quit()\n\n j = 0\n\n for table in tables:\n\n # Get players\n players = table.findAll(\"tr\")\n\n # Get stats\n for i in range(len(players)):\n stats = players[i].findAll(\"td\")\n\n if j == 0:\n index = stats[0].get_text().split()[0][0]\n nation = stats[1].find(\"span\").get(\"class\")[2].rsplit(\"-\", 1)[1]\n name = stats[1].findChildren()[0].get_text().strip()\n players_team = teams_ID_to_name[team]\n age = stats[1].findChildren()[2].get_text().strip()\n pos = stats[1].findChildren()[3].get_text().split(\",\", 1)[1].strip()\n cm = stats[2].get_text().strip()\n kg = stats[3].get_text().strip()\n apps = stats[4].get_text().strip()\n mins = stats[5].get_text().strip()\n goals = stats[6].get_text().strip()\n assists = stats[7].get_text().strip()\n yellow = stats[8].get_text().strip()\n red = stats[9].get_text().strip()\n SpG = stats[10].get_text().strip()\n PassPer = stats[11].get_text().strip()\n AerialsWon = stats[12].get_text().strip()\n MotM = stats[13].get_text().strip()\n\n ALLTHEDAMNPLAYERS[name] = []\n ALLTHEDAMNPLAYERS[name].append(players_team)\n ALLTHEDAMNPLAYERS[name].append(nation)\n ALLTHEDAMNPLAYERS[name].append(age)\n ALLTHEDAMNPLAYERS[name].append(pos)\n ALLTHEDAMNPLAYERS[name].append(cm)\n ALLTHEDAMNPLAYERS[name].append(kg)\n ALLTHEDAMNPLAYERS[name].append(apps)\n ALLTHEDAMNPLAYERS[name].append(mins)\n ALLTHEDAMNPLAYERS[name].append(goals)\n ALLTHEDAMNPLAYERS[name].append(assists)\n ALLTHEDAMNPLAYERS[name].append(yellow)\n ALLTHEDAMNPLAYERS[name].append(red)\n ALLTHEDAMNPLAYERS[name].append(SpG)\n ALLTHEDAMNPLAYERS[name].append(PassPer)\n ALLTHEDAMNPLAYERS[name].append(AerialsWon)\n ALLTHEDAMNPLAYERS[name].append(MotM)\n\n if j == 1:\n name = stats[1].findChildren()[0].get_text().strip()\n tackles = stats[6].get_text().strip()\n inter = stats[7].get_text().strip()\n fouls = stats[8].get_text().strip()\n offW = stats[9].get_text().strip()\n clear = stats[10].get_text().strip()\n drbP = stats[11].get_text().strip()\n blocks = stats[12].get_text().strip()\n ownG = stats[13].get_text().strip()\n\n ALLTHEDAMNPLAYERS[name].append(tackles)\n ALLTHEDAMNPLAYERS[name].append(inter)\n ALLTHEDAMNPLAYERS[name].append(fouls)\n ALLTHEDAMNPLAYERS[name].append(offW)\n ALLTHEDAMNPLAYERS[name].append(clear)\n ALLTHEDAMNPLAYERS[name].append(drbP)\n ALLTHEDAMNPLAYERS[name].append(blocks)\n ALLTHEDAMNPLAYERS[name].append(ownG)\n\n if j == 2:\n name = stats[1].findChildren()[0].get_text().strip()\n KP = stats[9].get_text().strip()\n drb = stats[10].get_text().strip()\n fouled = stats[11].get_text().strip()\n cOff = stats[12].get_text().strip()\n disp = stats[13].get_text().strip()\n unsT = stats[14].get_text().strip()\n\n ALLTHEDAMNPLAYERS[name].append(KP)\n ALLTHEDAMNPLAYERS[name].append(drb)\n ALLTHEDAMNPLAYERS[name].append(fouled)\n ALLTHEDAMNPLAYERS[name].append(cOff)\n ALLTHEDAMNPLAYERS[name].append(disp)\n ALLTHEDAMNPLAYERS[name].append(unsT)\n\n if j == 3:\n name = stats[1].findChildren()[0].get_text().strip()\n avgP = stats[8].get_text().strip()\n crosses = stats[10].get_text().strip()\n longB = stats[11].get_text().strip()\n thrB = stats[12].get_text().strip()\n rat = stats[13].get_text().strip()\n season = archiveUrls_to_season[archiveUrl]\n\n ALLTHEDAMNPLAYERS[name].append(avgP)\n ALLTHEDAMNPLAYERS[name].append(crosses)\n ALLTHEDAMNPLAYERS[name].append(longB)\n ALLTHEDAMNPLAYERS[name].append(thrB)\n ALLTHEDAMNPLAYERS[name].append(rat)\n ALLTHEDAMNPLAYERS[name].append(season)\n\n j += 1\n\n for indiv in ALLTHEDAMNPLAYERS:\n csv_writer.writerow(\n [indiv] + [ALLTHEDAMNPLAYERS[indiv][i] for i in range(36)]\n )\n","sub_path":"src/data/scrappers/scraper_whoscoredv3.py","file_name":"scraper_whoscoredv3.py","file_ext":"py","file_size_in_byte":11765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"433311800","text":"\nimport pandas as pd # dataframes y utilidades\nfrom datetime import timedelta # diferencia entre datos tipo tiempo\nfrom oandapyV20 import API # conexion con broker OANDA\nimport oandapyV20.endpoints.instruments as instruments # informacion de precios historicos\n\n\n# -- --------------------------------------------------------- FUNCION: Descargar precios -- #\n# -- Descargar precios historicos con OANDA\n\ndef f_precios_masivos(p0_fini, p1_ffin, p2_gran, p3_inst, p4_oatk, p5_ginc):\n \"\"\"\n Parameters\n ----------\n p0_fini\n p1_ffin\n p2_gran\n p3_inst\n p4_oatk\n p5_ginc\n\n Returns\n -------\n dc_precios\n\n Debugging\n ---------\n\n \"\"\"\n\n def f_datetime_range_fx(p0_start, p1_end, p2_inc, p3_delta):\n \"\"\"\n\n Parameters\n ----------\n p0_start\n p1_end\n p2_inc\n p3_delta\n\n Returns\n -------\n ls_resultado\n\n Debugging\n ---------\n \"\"\"\n\n ls_result = []\n nxt = p0_start\n\n while nxt <= p1_end:\n ls_result.append(nxt)\n if p3_delta == 'minutes':\n nxt += timedelta(minutes=p2_inc)\n elif p3_delta == 'hours':\n nxt += timedelta(hours=p2_inc)\n elif p3_delta == 'days':\n nxt += timedelta(days=p2_inc)\n\n return ls_result\n\n # inicializar api de OANDA\n\n api = API(access_token=p4_oatk)\n\n gn = {'S30': 30, 'S10': 10, 'S5': 5, 'M1': 60, 'M5': 60 * 5, 'M15': 60 * 15,\n 'M30': 60 * 30, 'H1': 60 * 60, 'H4': 60 * 60 * 4, 'H8': 60 * 60 * 8,\n 'D': 60 * 60 * 24, 'W': 60 * 60 * 24 * 7, 'M': 60 * 60 * 24 * 7 * 4}\n\n # -- para el caso donde con 1 peticion se cubran las 2 fechas\n if int((p1_ffin - p0_fini).total_seconds() / gn[p2_gran]) < 4999:\n\n # Fecha inicial y fecha final\n f1 = p0_fini.strftime('%Y-%m-%dT%H:%M:%S')\n f2 = p1_ffin.strftime('%Y-%m-%dT%H:%M:%S')\n\n # Parametros pra la peticion de precios\n params = {\"granularity\": p2_gran, \"price\": \"M\", \"dailyAlignment\": 16, \"from\": f1,\n \"to\": f2}\n\n # Ejecutar la peticion de precios\n a1_req1 = instruments.InstrumentsCandles(instrument=p3_inst, params=params)\n a1_hist = api.request(a1_req1)\n\n # Para debuging\n # print(f1 + ' y ' + f2)\n lista = list()\n\n # Acomodar las llaves\n for i in range(len(a1_hist['candles']) - 1):\n lista.append({'TimeStamp': a1_hist['candles'][i]['time'],\n 'Open': a1_hist['candles'][i]['mid']['o'],\n 'High': a1_hist['candles'][i]['mid']['h'],\n 'Low': a1_hist['candles'][i]['mid']['l'],\n 'Close': a1_hist['candles'][i]['mid']['c']})\n\n # Acomodar en un data frame\n r_df_final = pd.DataFrame(lista)\n r_df_final = r_df_final[['TimeStamp', 'Open', 'High', 'Low', 'Close']]\n r_df_final['TimeStamp'] = pd.to_datetime(r_df_final['TimeStamp'])\n r_df_final['Open'] = pd.to_numeric(r_df_final['Open'], errors='coerce')\n r_df_final['High'] = pd.to_numeric(r_df_final['High'], errors='coerce')\n r_df_final['Low'] = pd.to_numeric(r_df_final['Low'], errors='coerce')\n r_df_final['Close'] = pd.to_numeric(r_df_final['Close'], errors='coerce')\n\n return r_df_final\n\n # -- para el caso donde se construyen fechas secuenciales\n else:\n\n # hacer series de fechas e iteraciones para pedir todos los precios\n fechas = f_datetime_range_fx(p0_start=p0_fini, p1_end=p1_ffin, p2_inc=p5_ginc,\n p3_delta='minutes')\n\n # Lista para ir guardando los data frames\n lista_df = list()\n\n for n_fecha in range(0, len(fechas) - 1):\n\n # Fecha inicial y fecha final\n f1 = fechas[n_fecha].strftime('%Y-%m-%dT%H:%M:%S')\n f2 = fechas[n_fecha + 1].strftime('%Y-%m-%dT%H:%M:%S')\n\n # Parametros pra la peticion de precios\n params = {\"granularity\": p2_gran, \"price\": \"M\", \"dailyAlignment\": 16, \"from\": f1,\n \"to\": f2}\n\n # Ejecutar la peticion de precios\n a1_req1 = instruments.InstrumentsCandles(instrument=p3_inst, params=params)\n a1_hist = api.request(a1_req1)\n\n # Para debuging\n print(f1 + ' y ' + f2)\n lista = list()\n\n # Acomodar las llaves\n for i in range(len(a1_hist['candles']) - 1):\n lista.append({'TimeStamp': a1_hist['candles'][i]['time'],\n 'Open': a1_hist['candles'][i]['mid']['o'],\n 'High': a1_hist['candles'][i]['mid']['h'],\n 'Low': a1_hist['candles'][i]['mid']['l'],\n 'Close': a1_hist['candles'][i]['mid']['c']})\n\n # Acomodar en un data frame\n pd_hist = pd.DataFrame(lista)\n pd_hist = pd_hist[['TimeStamp', 'Open', 'High', 'Low', 'Close']]\n pd_hist['TimeStamp'] = pd.to_datetime(pd_hist['TimeStamp'])\n\n # Ir guardando resultados en una lista\n lista_df.append(pd_hist)\n\n # Concatenar todas las listas\n r_df_final = pd.concat([lista_df[i] for i in range(0, len(lista_df))])\n\n # resetear index en dataframe resultante porque guarda los indices del dataframe pasado\n r_df_final = r_df_final.reset_index(drop=True)\n r_df_final['Open'] = pd.to_numeric(r_df_final['Open'], errors='coerce')\n r_df_final['High'] = pd.to_numeric(r_df_final['High'], errors='coerce')\n r_df_final['Low'] = pd.to_numeric(r_df_final['Low'], errors='coerce')\n r_df_final['Close'] = pd.to_numeric(r_df_final['Close'], errors='coerce')\n\n return r_df_final\n","sub_path":"Notas_Python/Notas_Herramientas/Conteo de velas/funciones.py","file_name":"funciones.py","file_ext":"py","file_size_in_byte":5893,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"442054160","text":"from django.contrib.auth import authenticate, login, logout\nfrom django.db import IntegrityError\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render\nfrom django.urls import reverse\nfrom django.db.models import Max\n\nfrom .models import *\n\nfrom datetime import datetime\n\n\ndef index(request):\n listings = Listing.objects.filter(status=\"active\")\n return render(request, \"auctions/index.html\", {\n \"listings\":reversed(listings)\n })\n\n\ndef login_view(request):\n if request.method == \"POST\":\n\n # Attempt to sign user in\n username = request.POST[\"username\"]\n password = request.POST[\"password\"]\n user = authenticate(request, username=username, password=password)\n\n # Check if authentication successful\n if user is not None:\n login(request, user)\n return HttpResponseRedirect(reverse(\"index\"))\n else:\n return render(request, \"auctions/login.html\", {\n \"message\": \"Invalid username and/or password.\"\n })\n else:\n return render(request, \"auctions/login.html\")\n\n\ndef logout_view(request):\n logout(request)\n return HttpResponseRedirect(reverse(\"index\"))\n\n\ndef register(request):\n if request.method == \"POST\":\n username = request.POST[\"username\"]\n email = request.POST[\"email\"]\n\n # Ensure password matches confirmation\n password = request.POST[\"password\"]\n confirmation = request.POST[\"confirmation\"]\n if password != confirmation:\n return render(request, \"auctions/register.html\", {\n \"message\": \"Passwords must match.\"\n })\n\n # Attempt to create new user\n try:\n user = User.objects.create_user(username, email, password)\n user.save()\n except IntegrityError:\n return render(request, \"auctions/register.html\", {\n \"message\": \"Username already taken.\"\n })\n login(request, user)\n return HttpResponseRedirect(reverse(\"index\"))\n else:\n return render(request, \"auctions/register.html\")\n\ndef create(request):\n if request.method == \"POST\":\n title = request.POST.get('title')\n description = request.POST.get('description')\n image = request.POST.get('image')\n category_code = request.POST.get('category')\n if not category_code:\n category_code = \"CNP\"\n category = Category.objects.get(code=category_code)\n s_bid = request.POST.get('s_bid')\n condition = request.POST.get('condition')\n c_time = datetime.now()\n try:\n listing_ = Listing.objects.create(title=title, description=description, image=image, category=category, starting_bid=s_bid, current_bid=s_bid, condition=condition, create_time=c_time)\n user = User.objects.get(username=request.user.username)\n user.listing.add(listing_)\n return HttpResponseRedirect(reverse(\"index\"))\n except Exception as e:\n return HttpResponse(f\"There was a error while creating the listing. : {e}\")\n if not request.user.is_authenticated:\n return HttpResponseRedirect(reverse(\"login\"))\n else:\n return render(request, \"auctions/create.html\")\n\ndef search(request):\n query = request.GET['query']\n listings = Listing.objects.all()\n results = []\n for each in listings:\n title = each.title.lower()\n if title.find(query.lower()) >= 0:\n results.append(each)\n return render(request, \"auctions/index.html\", {\n \"query\":query,\n \"result_title\":f\"Search result for '{query}'\",\n \"listings\":reversed(results)\n })\n\ndef listing(request, id, status=\"None\"):\n list_item = Listing.objects.get(id=id)\n watchlist = \"\"\n watchers_count = list_item.watchers.count()\n no_of_bids = UserBid.objects.filter(listing=list_item).count()\n bids = UserBid.objects.filter(listing=list_item).all().order_by('-id')[:5]\n comments = UserComment.objects.filter(listing=list_item).all().order_by('-id')\n creater = list_item.creater.get()\n outbid = \"\"\n if request.user.is_authenticated:\n if (no_of_bids > 0):\n userBids = UserBid.objects.filter(bidder=request.user, listing=list_item)\n if userBids:\n userBid = max([userBid.bid for userBid in userBids])\n if userBid == list_item.current_bid:\n outbid = \"no\"\n else:\n outbid = \"yes\"\n if list_item.status == \"sold\":\n if request.user == list_item.buyer.get():\n success = True\n message = \"Congratulations! you have won this auction.\"\n else:\n success = False\n message = \"\"\n else:\n success = False\n message = \"\"\n if request.user == creater:\n creater_view = True\n else:\n creater_view = False\n if request.user.is_authenticated:\n watchlist = request.user.watchlist.filter(id=id).first()\n if status == \"success\":\n success = True\n message = \"Congratulations! You have successfully placed your bid.\"\n\n elif status == \"error_code1\":\n success = False\n message = f\"Error: Bid amount must be greater than or equal to US ${list_item.starting_bid}\"\n elif status == \"error_code2\":\n success = False\n message = f\"Error: Bid amount must be greater than US ${list_item.current_bid}\"\n elif status == \"error_code3\":\n success = False\n message = \"There was an error while placing your bid.\"\n return render(request, \"auctions/listing.html\", {\n \"list\":list_item,\n \"creater\":creater,\n \"watchlist\":watchlist,\n \"watchers_count\":watchers_count,\n \"outbid\":outbid,\n \"no_of_bids\":no_of_bids,\n \"bids\":bids,\n \"comments\":comments,\n \"creater_view\":creater_view,\n \"success\":success,\n \"message\":message\n })\n\n\ndef watchlist(request):\n if request.user.is_authenticated:\n listings = request.user.watchlist.all()\n return render(request, \"auctions/index.html\", {\n \"result_title\":\"Watchlist\",\n \"listings\":reversed(listings),\n })\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef add_watchlist(request):\n if request.method == \"POST\":\n if request.user.is_authenticated:\n id = request.POST['id']\n origin = request.POST['origin']\n list_item = Listing.objects.get(id=id)\n try:\n request.user.watchlist.add(list_item)\n except Exception as e:\n return render(request, \"auctions/message.html\", {\n \"message\":f\"There was an error while adding to watchlist : {e}\"\n })\n return HttpResponseRedirect(origin)\n else:\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef remove_watchlist(request):\n if request.method == \"POST\":\n if request.user.is_authenticated:\n id = request.POST['id']\n origin = request.POST['origin']\n list_item = Listing.objects.get(id=id)\n try:\n watchlist = request.user.watchlist.get(id__contains=id)\n request.user.watchlist.remove(watchlist)\n except Exception as e:\n return render(request, \"auctions/message.html\", {\n \"message\":f\"There was an error while removing from watchlist : {e}\"\n })\n return HttpResponseRedirect(origin)\n else:\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef my_listing(request):\n if request.user.is_authenticated:\n listings = request.user.listing.all()\n return render(request, \"auctions/index.html\", {\n \"listings\":reversed(listings),\n \"result_title\":\"My Listings\",\n \"creater\":True\n })\n else:\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef place_bid(request, id):\n if request.method == \"POST\":\n if request.user.is_authenticated:\n bid_amount = float(request.POST['bid'])\n list_id = request.POST.get('id')\n list_item = Listing.objects.get(id=list_id)\n bid_time = datetime.now()\n try:\n if UserBid.objects.filter(listing=list_item).count() == 0:\n if bid_amount < list_item.starting_bid:\n return HttpResponseRedirect(reverse(\"bid_result\", kwargs={\n \"id\":list_id,\n \"status\":\"error_code1\"\n }))\n else:\n if bid_amount <= list_item.current_bid:\n return HttpResponseRedirect(reverse(\"bid_result\", kwargs={\n \"id\":list_id,\n \"status\":\"error_code2\"\n }))\n list_item.current_bid = bid_amount\n list_item.save()\n user_bid = UserBid.objects.create(bidder=request.user, bid=bid_amount, time=bid_time)\n user_bid.listing.add(list_item)\n user_bid.save()\n return HttpResponseRedirect(reverse(\"bid_result\", kwargs={\n \"id\":list_id,\n \"status\":\"success\"\n }))\n except expression as e:\n return HttpResponseRedirect(reverse(\"bid_result\", kwargs={\n \"id\":list_id,\n \"status\":\"error_code3\"\n }))\n else:\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef bids(request, id):\n list_item = Listing.objects.get(id=id)\n bids = UserBid.objects.filter(listing=list_item).all().order_by('-id')\n creater = list_item.creater.get()\n if request.user == creater:\n creater_view = True\n else:\n creater_view = False\n return render(request, \"auctions/bids.html\", {\n \"bids\":bids,\n \"list\":list_item,\n \"creater_view\":creater_view\n })\n\ndef auction_command(request):\n if request.method == \"POST\":\n list_id = int(request.POST['id'])\n list_item = Listing.objects.get(id=list_id)\n if list_item.status == \"active\":\n if list_item.bids.count() == 0:\n list_item.status = \"unsold\"\n list_item.save()\n else:\n list_item.status = \"sold\"\n bids = UserBid.objects.filter(listing=list_item)\n mbid = bids.aggregate(Max('bid'))\n highest_bid = bids.filter(bid=mbid['bid__max']).first()\n highest_bid.bidder.bought_items.add(list_item)\n highest_bid.save()\n list_item.save()\n elif list_item.status == \"unsold\":\n list_item.status = \"active\"\n list_item.save()\n return HttpResponseRedirect(reverse(\"listing\", args={list_id}))\n\ndef add_comment(request):\n if request.method == \"POST\":\n if request.user.is_authenticated:\n list_id = int(request.POST['id'])\n comment_data = request.POST['comment']\n list_item = Listing.objects.get(id=list_id)\n comment_time = datetime.now()\n user_comment = UserComment.objects.create(commenter=request.user, comment=comment_data, time=comment_time)\n user_comment.listing.add(list_item)\n user_comment.save()\n return HttpResponseRedirect(reverse(\"listing\", args={list_id}))\n else:\n return HttpResponseRedirect(reverse(\"login\"))\n\ndef categories(request, category_code=None):\n if category_code == None:\n return render(request, \"auctions/categories.html\")\n else:\n category = Category.objects.get(code=category_code)\n listings = Listing.objects.filter(category=category, status=\"active\").all()\n return render(request, \"auctions/index.html\", {\n \"listings\":reversed(listings),\n \"result_title\":f\"Categories > {category.name}\"\n })","sub_path":"auctions/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":12032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"113306318","text":"import cv2 # импорт модуля из библиотеки Opencv\r\nimport numpy as np # модуль обработки массивов\r\nimport sys # системный модуль\r\nimport time \r\n\r\n\r\n# Первый блок проверяет условие, передан ли скрипту в командной строке дополнительный аргумент в виде картинки **QR кода**. Если первое условие ложно, то считывается указанная нами картинка.\r\n\r\nif len(sys.argv)>1:\r\n inputImage = cv2.imread(sys.argv[1]) \r\nelse:\r\n inputImage = cv2.imread(\"myrusakov_out.jpg\") # стандартный метод opencv для считывания изображения\r\n\r\n\r\n\r\n# Создание функции выводящей в отдельном окне изображение QR с синим обрамлением.\r\ndef display(im, bbox):\r\n\r\n n = len(bbox)\r\n\r\n for j in range(n):\r\n\r\n cv2.line(im, tuple(bbox[j][0]), tuple(bbox[ (j+1) % n][0]), (255,0,0), 3)\r\n \r\n # Display results\r\n cv2.imshow(\"Results\", im)\r\n\r\n\r\n\r\n# В Opencv имеется встроенный метод детектор QR\r\n\r\nqrDecoder = cv2.QRCodeDetector() # создание объекта детектора\r\n\r\n \r\n\r\n# Нахождение и декодирование нашего кода. Метод **detectAndDecode** возвращает кортеж из трех значений которыми кодируется QR, где первый аргумент data содержит декодированную строку, bbox - координаты вершин нашего изображения и rectifiedImage, содержит **QR** изображение в виде массива пикселей.\r\n\r\ndata, bbox, rectifiedImage = qrDecoder.detectAndDecode(inputImage)\r\n\r\nif len(data)>0:\r\n\r\n print(\"Decoded Data : {}\".format(data)) # вывод декодированной строки\r\n\r\n display(inputImage, bbox)\r\n\r\n #rectifiedImage = np.uint8(rectifiedImage);\r\n\r\n #cv2.imshow(\"Rectified QRCode\", rectifiedImage);\r\n\r\nelse:\r\n\r\n print(\"QR Code not detected\") \r\n\r\n cv2.imshow(\"Results\", inputImage)\r\n\r\n\r\n\r\ncv2.waitKey(0)\r\n\r\ncv2.destroyAllWindows()","sub_path":"read-qr-code.py","file_name":"read-qr-code.py","file_ext":"py","file_size_in_byte":2304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"75783114","text":"import random\nimport numpy as np\nfrom numpy import pi, exp, cos, sin\nimport os\nimport time as time\nimport sys\n\ndef transpose(L):\n return(list(map(list,zip(*L))))\n\ndef Update_Prior(M, theta, outcome, mu, sigma):\n d = outcome\n phi_mu= mu+((1-2*d)*M*sigma**2*np.sin(M*mu - M*theta))/(np.exp(0.5*M**2*sigma**2)+(1-2*d)*np.cos(M*mu - M*theta))\n numa = (2*np.exp(M**2*sigma**2)+2*(2*d-1)*exp(M**2 * sigma**2 * 0.5)*(M**2*sigma**2-2)*np.cos(M*mu - M*theta)+(1-2*d)**2*(1-2*M**2*sigma**2+np.cos(2*M*mu - 2*M*theta)))\n demonetor=2*(np.exp(0.5*M**2*sigma**2)+(1-2*d)*np.cos(M*mu - M*theta))**2\n sigma_2=sigma**2*numa/demonetor\n phi_sigma=np.sqrt(sigma_2)\n \n return(phi_mu, phi_sigma)\n\n\ndef bqpe_analytical(threshold = 5*(10**-3), Phi = 0, Alpha = 0, sigma = pi / 4, Max_Runs = 10**5):\n if not 0<= Alpha<=1 or not -pi<=Phi<=pi:\n raise ValueError(\"Alpha: must be between 0 and 1 inclusive.\")\n if sigma<0 or sigma > 1:\n raise ValueError(\"Sigma outside of acceptable ranges\")\n\n mu = random.uniform(-pi, pi)\n run = 0\n flag = 0\n while sigma > threshold:\n theta = mu - sigma\n M = max(\n 1, np.floor((1/sigma**Alpha) + 1/2)\n )\n\n p = 1/2 + cos(M*Phi - M*theta)/2\n\n if random.uniform(0, 1) < p:\n outcome = 0\n else:\n outcome = 1\n \n mu, sigma = Update_Prior(M, theta, outcome, mu, sigma)\n # mu, sigma = np.mean(accepted), np.std(accepted)\n \n run += 1\n if run > Max_Runs:\n flag = 1\n break\n\n err = abs(\n abs(cos(Phi/2)) - abs(cos(mu/2))\n )\n\n return(flag, float('%.5f'%(cos(mu/2))), err, run, sigma)\n\n\n\ndef bqpe_numerical(threshold = 5*(10**-3), Phi = 0, Alpha = 0, sigma = pi / 4, Sample_Size = 100, Max_Runs = 100000):\n if not 0<= Alpha<=1 or not -pi<=Phi<=pi:\n raise ValueError(\"Alpha: must be between 0 and 1 inclusive.\")\n if sigma<0 or sigma > 1:\n raise ValueError(\"Sigma outside of acceptable ranges\")\n\n mu = random.uniform(-pi, pi)\n sigma = sigma\n run = 0\n flag = 0\n while sigma > threshold:\n Sampled = np.random.normal(mu, sigma, Sample_Size)\n theta = mu - sigma\n M = max(\n 1, np.floor((1/sigma**Alpha) + 1/2)\n )\n\n p = 1/2 + cos(M*Phi - M*theta)/2\n\n if random.uniform(0, 1) < p:\n outcome = 0\n else:\n outcome = 1\n \n accepted = [] \n for varphi in Sampled:\n P = 1/2 + (1-2*outcome)*cos(M*varphi-M*theta)/2\n if P > random.uniform(0, 1):\n accepted.append(\n varphi\n )\n\n if len(accepted) < 2:\n sigma *= 1.2\n continue\n # mu, sigma = Update_Prior(M, theta, outcome, mu, sigma)\n mu, sigma = np.mean(accepted), np.std(accepted)\n \n run += 1\n if run > Max_Runs:\n flag = 1\n break\n\n err = abs(\n abs(cos(Phi/2)) - abs(cos(mu/2))\n )\n\n return(flag, float('%.5f'%(cos(mu/2))), err, run, sigma)\n\n# for _ in range(50):\n\n# print(\n# bqpe_analytical(Phi = 0.324234234)\n# )","sub_path":"BQPE/alpha-BQPE/bqpe_collapsed.py","file_name":"bqpe_collapsed.py","file_ext":"py","file_size_in_byte":3171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"592041786","text":"\"\"\"\nThis class imports global currency data in batches.\n\"\"\"\n\n\nclass CurrencyCollector:\n\n def __init__(self):\n pass\n\n def currency_ingestor(self):\n # Python library imports\n import sqlite3\n import sys\n import datetime\n import os\n import time\n import requests\n import traceback\n # local file imports\n import error as error_class\n import _keys_and_secrets\n import control\n\n # this instantiates the error class\n error = error_class.Error()\n\n # this loads the database instance\n sqlite_relative_path = os.path.join('collector.sqlite3')\n sqlite_absolute_path = os.path.abspath(sqlite_relative_path)\n conn = sqlite3.connect(sqlite_absolute_path)\n c = conn.cursor()\n\n \"\"\"\n Currencies are quoted by comparing a base currency to a counter currency. For example, if we are looking for \n the value of the Euro relative to the US Dollar, we would look at EUR/USD. If EUR/USD = 1.2500, it means that \n one Euro is exchanged 1.2500 US Dollars. \n \n What we will do here, then, is to compare every available currency against the Bank for International \n Settlements' Triennial Central Bank Survey: Foreign Exchange Turnover in April 2016's top 8 currencies - the \n majors.\n \"\"\"\n\n major_currencies = ['USD', 'EUR', 'JPY', 'GBP', 'AUD', 'CAD', 'CHF', 'CNY']\n\n currencies = ['USD', 'EUR', 'JPY', 'GBP', 'AUD', 'CAD', 'CHF', 'CNY', 'AED', 'AFN', 'ALL', 'AMD', 'AOA', 'ARS',\n 'AWG', 'AZN', 'BAM', 'BBD', 'BDT', 'BGN', 'BHD', 'BIF', 'BMD', 'BND', 'BOB', 'BRL', 'BSD', 'BTN',\n 'BWP', 'BZD', 'CDF', 'CLF', 'CLP', 'COP', 'CRC', 'CUP', 'CVE', 'CZK', 'DJF', 'DKK', 'DOP', 'DZD',\n 'EGP', 'ERN', 'ETB', 'FJD', 'GEL', 'GHS', 'GIP', 'GMD', 'GNF', 'GTQ', 'GYD', 'HKD', 'HNL', 'HRK',\n 'HTG', 'HUF', 'IDR', 'ILS', 'INR', 'IQD', 'IRR', 'ISK', 'JEP', 'JMD', 'JOD', 'KES', 'KGS', 'KHR',\n 'KMF', 'KPW', 'KRW', 'KWD', 'KYD', 'KZT', 'LAK', 'LBP', 'LKR', 'LRD', 'LSL', 'LYD', 'MAD', 'MDL',\n 'MGA', 'MKD', 'MMK', 'MNT', 'MOP', 'MRO', 'MUR', 'MVR', 'MWK', 'MXN', 'MZN', 'NAD', 'NGN', 'NIO',\n 'NOK', 'NPR', 'NZD', 'OMR', 'PAB', 'PEN', 'PGK', 'PHP', 'PKR', 'PLN', 'PYG', 'QAR', 'RON', 'RSD',\n 'RUB', 'RWF', 'SAR', 'SBD', 'SCR', 'SDG', 'SEK', 'SGD', 'SHP', 'SLL', 'SOS', 'SRD', 'STD', 'SVC',\n 'SYP', 'SZL', 'THB', 'TJS', 'TMT', 'TND', 'TOP', 'TRY', 'TTD', 'TWD', 'TZS', 'UAH', 'UGX', 'UYU',\n 'UZS', 'VEF', 'VND', 'VUV', 'WST', 'XCD', 'XOF', 'XPF', 'YER', 'ZAR', 'ZMW', 'ZWL', 'XAG', 'XAU']\n\n # this code block sets all of the static Alpha Vantage API variables\n api_url = 'https://www.alphavantage.co/query'\n api_function = 'CURRENCY_EXCHANGE_RATE'\n\n if control.debug is True:\n current_time_int = int(time.time())\n current_time_struct = time.gmtime(current_time_int)\n current_time = str(datetime.datetime.fromtimestamp(time.mktime(current_time_struct)))\n print('------------------------------------------------------------------------------\\n'\n 'currency ingestor initiated on %s UTC\\n'\n '------------------------------------------------------------------------------\\n'\n % current_time)\n\n # this code block sets all of the Alpha Vantage API variables that change depending on what needs to be called\n for currency in currencies:\n for major_currency in major_currencies:\n if currency != major_currencies:\n from_currency = currency\n to_currency = major_currency\n api_key = _keys_and_secrets.alphavantage_api_key\n\n data = {'function': api_function,\n 'from_currency': from_currency,\n 'to_currency': to_currency,\n 'apikey': api_key\n }\n try:\n currency_raw = requests.get(api_url, params=data)\n currency_json = currency_raw.json()\n currency_parsed = currency_json['Realtime Currency Exchange Rate']\n except:\n e = sys.exc_info()\n full_e = traceback.format_exc()\n currency_raw.close()\n error.if_error(str(e), full_e, 'currency_ingestor()', 'Alpha Vantage API call')\n break\n\n try:\n base_currency_code = currency_parsed['1. From_Currency Code']\n base_currency_name = currency_parsed['2. From_Currency Name']\n counter_currency_code = currency_parsed['3. To_Currency Code']\n counter_currency_name = currency_parsed['4. To_Currency Name']\n exchange_rate = currency_parsed['5. Exchange Rate']\n published = currency_parsed['6. Last Refreshed']\n except:\n e = sys.exc_info()\n full_e = traceback.format_exc()\n error.if_error(str(e), full_e, 'currency_ingestor()', 'currency price data')\n\n # this block returns the date and time when the row was imported\n try:\n imported_int = int(time.time())\n imported_struct = time.gmtime(imported_int)\n imported = str(datetime.datetime.fromtimestamp(time.mktime(imported_struct)))\n if control.debug is True:\n print('imported: ' + imported)\n except:\n e = sys.exc_info()\n full_e = traceback.format_exc()\n error.if_error(str(e), full_e, 'currency_ingestor()', 'imported time')\n\n if control.debug is True:\n print('base currency code: %s\\n'\n 'base currency name: %s\\n'\n 'counter currency code: %s\\n'\n 'counter currency name: %s\\n'\n 'exchange rate: %s\\n'\n 'published at: %s\\n\\n'\n % (base_currency_code, base_currency_name, counter_currency_code, counter_currency_name,\n exchange_rate, published))\n\n c.execute('INSERT INTO currencies VALUES (?,?,?,?,?,?,?)',\n (base_currency_code, base_currency_name, counter_currency_code, counter_currency_name,\n exchange_rate, published, imported))\n try:\n conn.commit()\n except:\n e = sys.exc_info()\n full_e = traceback.format_exc()\n error.if_error(str(e), full_e, 'currency_ingestor()', 'database commit')\n return\n","sub_path":"collection_and_processing/batch_collectors/currency_collector.py","file_name":"currency_collector.py","file_ext":"py","file_size_in_byte":7285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"439271496","text":"\"\"\"Process raster data\"\"\"\r\n\r\nfrom os import sep\r\nfrom os.path import join\r\nimport gdal\r\nimport osr\r\nimport rasterio\r\nfrom rasterio.warp import calculate_default_transform, reproject, Resampling\r\nfrom tempfile import NamedTemporaryFile\r\n\r\n\r\n# read and write\r\ndef read_tif(path):\r\n \"\"\"Read a geotif file.\"\"\"\r\n ds = gdal.Open(path)\r\n return(ds)\r\n \r\ndef subset_raster(ds, outfile, ulx, uly, lrx, lry):\r\n \"\"\"Create a subset of raster tif\"\"\"\r\n gdal.Translate(outfile, ds, projWin=[ulx,uly,lrx,lry])\r\n ds = None\r\n\r\ndef read_zip(path):\r\n \"\"\"Read a dataset from a zipfile without decompressing\"\"\"\r\n tic = time.clock()\r\n p = join(sep, 'vsizip', path, path.replace('zip','tif'))\r\n ds = gdal.Open(p)\r\n return(ds)\r\n\r\n\r\n \r\n# mapping tools\r\ndef colorbar(fig, left, bottom, width, height, labelsize, tcolor='black'):\r\n import matplotlib.pyplot as plt\r\n ax = fig.add_axes([left, bottom, width, height])\r\n ax.get_xaxis().set_visible(False)\r\n ax.get_yaxis().set_visible(False)\r\n ax.set_frame_on(False)\r\n cbar = plt.colorbar(fraction=1)\r\n cbar.outline.set_linewidth(0.5)\r\n cbar.ax.tick_params(labelsize=labelsize, color=tcolor, labelcolor=tcolor, width=0.5)\r\n return(fig)\r\n\r\n\r\n# online datasets\r\nclass OpenStreetMap:\r\n \"\"\"Plot Open Street Map tiles on a matplotlib axis\"\"\"\r\n def __init__(self, zoom=14):\r\n from cartopy.io.img_tiles import OSM\r\n self.zoom = zoom\r\n self.tiles = OSM()\r\n\r\n def plot(self, ax):\r\n ax.add_image(self.tiles, self.zoom)\r\n\r\n\r\nclass GoogleEarth:\r\n \"\"\"Plot current Google Earth imagery on a matplotlib axis. Styles also include 'street' (useful for when OSM doesn't work)\"\"\"\r\n def __init__(self, style='satellite'):\r\n from cartopy.io.img_tiles import GoogleTiles\r\n self.tiles = GoogleTiles(style=style)\r\n\r\n def plot(self, ax):\r\n \"\"\"Last time this was used it threw an error when saving the file. A workaround was to code the command in the main script.\"\"\"\r\n ax.add_image(self.tiles)\r\n\r\n\r\ndef Wee(filename='location', zoom=14):\r\n from cartopy.io.img_tiles import GoogleTiles as T\r\n m = cart()\r\n wee = T(style='street')\r\n m.ax.add_image(wee, zoom)\r\n #r.GoogleEarth(style='street').plot(m.ax)\r\n outfile = mkjoin(mapdir, filename)\r\n m.save(outfile)\r\n\r\n\r\n\r\nclass SixMaps:\r\n \"\"\"Plot NSW SixMaps imagery on a matplotlib axis\"\"\"\r\n def __init__(self):\r\n self.url = 'http://maps.six.nsw.gov.au/arcgis/services/public/NSW_Imagery/MapServer/WMSServer?request=GetCapabilities&service=WMS'\r\n\r\n def plot(self, ax):\r\n ax.add_wms(wms=self.url, layers=['0'])\r\n\r\n\r\n# georeferencing tools\r\ndef reproject_tif(inpath, outpath, crs='EPSG:4326'):\r\n \"\"\"Reproject a geotif dataset. This code is sourced from the rasterio documentation .\"\"\"\r\n \r\n dst_crs = crs\r\n \r\n with rasterio.open(inpath) as src:\r\n transform, width, height = calculate_default_transform(\r\n src.crs, dst_crs, src.width, src.height, *src.bounds)\r\n kwargs = src.meta.copy()\r\n kwargs.update({\r\n 'crs': dst_crs,\r\n 'transform': transform,\r\n 'width': width,\r\n 'height': height\r\n })\r\n \r\n with rasterio.open(outpath, 'w', **kwargs) as dst:\r\n for i in range(1, src.count + 1):\r\n reproject(\r\n source=rasterio.band(src, i),\r\n destination=rasterio.band(dst, i),\r\n src_transform=src.transform,\r\n src_crs=src.crs,\r\n dst_transform=transform,\r\n dst_crs=dst_crs,\r\n resampling=Resampling.nearest)\r\n \r\nclass RectifyTif:\r\n \"\"\"Rectify a tif image using the known coordinates of the image corners. Useful for rectifying rectangular geophysical maps.\"\"\"\r\n\r\n def rectify(self, path, corners, epsg):\r\n\r\n ds = gdal.Open(path, gdal.GA_Update)\r\n gt = ds.GetGeoTransform()\r\n \r\n w, h = ds.RasterXSize, ds.RasterYSize\r\n x0 = int(gt[0])\r\n y1 = int(gt[3] + w*gt[4] + h*gt[5])\r\n x1 = int(gt[0] + w*gt[1] + h*gt[2])\r\n y0 = int(gt[3])\r\n \r\n sr = osr.SpatialReference()\r\n sr.ImportFromEPSG(epsg)\r\n \r\n c = corners\r\n gcps = [\r\n \r\n gdal.GCP(c[0][0], c[0][1], 0, x0, y1), # ll\r\n gdal.GCP(c[1][0], c[1][1], 0, x0, y0), # ul\r\n gdal.GCP(c[2][0], c[2][1], 0, x1, y0), # ur\r\n gdal.GCP(c[3][0], c[3][1], 0, x1, y1), # lr\r\n \r\n ] \r\n \r\n ds.SetGCPs(gcps, sr.ExportToWkt())\r\n ds.SetProjection(sr.ExportToWkt())\r\n ds.SetGeoTransform(gdal.GCPsToGeoTransform(gcps))\r\n \r\n tmp = NamedTemporaryFile(delete=False, suffix='.tif')\r\n \r\n gdal.Warp(tmp.name, ds)\r\n ds = None\r\n \r\n return(tmp.name)\r\n \r\n\r\n","sub_path":"geo/gis/raster.py","file_name":"raster.py","file_ext":"py","file_size_in_byte":4918,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"51853816","text":"import datetime as dt\nimport numpy as np\nimport pandas as pd\nimport pandas_datareader as pdr\nfrom scipy.stats import norm\n\n\nclass Monte:\n\n def __init__(self, ticker, sim_amount, time_steps, start, end=dt.datetime.now(), data_source='yahoo'):\n \"\"\"\n Initialization function for the Monte object.\n\n :param ticker: the ticker label associated with a stock\n :param sim_amount: the amount of simulations to be done\n :param time_steps: the number of time steps into the future the simualtion will go\n :param start: the start datetime for the simulations\n :param end: the end datetime for the simulations. Present time is default.\n :param data_source: data source from where the stock data is derived. Yahoo finance is the default.\n \"\"\"\n\n self.ticker = ticker\n self.sim_amount = sim_amount\n self.time_steps = time_steps\n self.start = start\n self.end = end\n self.data_source = data_source\n self.data = pd.DataFrame()\n self.monte_sims = pd.DataFrame()\n \n def create_DataFrame(self):\n \"\"\"\n Function that creates the DataFrame object where the stock data will be stored.\n \"\"\"\n\n self.data[self.ticker] = pdr.DataReader(self.ticker, data_source=self.data_source, \n start=self.start, end=self.end)['Adj Close']\n \n def simulate(self):\n \"\"\"\n Function that does the necessary calculations for the simulation data.\n \"\"\"\n\n # Initial data values needed to set up the simulations.\n log_returns = np.log(1 + self.data.pct_change()) # percentage change between current and prior element\n mu = log_returns.mean() # average/mean\n var = log_returns.var() # variance\n drift = mu - (0.5 * var) # stochastic drift\n sigma = log_returns.std() # standard deviation\n daily_returns = np.exp(drift.values + sigma.values * norm.ppf(np.random.rand(self.time_steps, self.sim_amount)))\n\n # Takes last data point in stock data as as the starting point for the simulations\n initial = self.data.iloc[-1]\n self.monte_list = np.zeros_like(daily_returns)\n self.monte_sims[0] = initial\n\n # Fills monte_sims with simulated prices which are pseudorandomized with daily_returns\n for t in range(1, self.time_steps):\n self.monte_sims[t] = self.monte_sims[t - 1] * daily_returns[t]\n\n def plot(self):\n return False","sub_path":"src/django/app/Monte.py","file_name":"Monte.py","file_ext":"py","file_size_in_byte":2506,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"253591802","text":"import math\n\na=[int(x) for x in input().split(\",\")]\na.sort()\ntarget=int(input())\nhead=0\ntail=len(a)-1\nminum=target\nan=0\ndef sum(ele):\n from functools import reduce\n if(len(ele)==1):\n return ele[0]\n elif len(ele)==0:\n return 0\n return reduce(lambda x,y:x+y,ele)\nans=0\nwhile heady1):\n an=y\n else:\n an=x\nelif ans==0:\n num=tail\n ans=target\n for i in range(a[tail-1],a[tail]+1):\n temp=abs(sum(a[0:tail])+(len(a)-tail)*i-target)\n if ans>temp:\n ans=temp\n an=i\nprint(an)\n\n\n","sub_path":"Code/CodeRecords/2576/60621/243244.py","file_name":"243244.py","file_ext":"py","file_size_in_byte":966,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"233276887","text":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# Copyright (C) 2004-2010 Tiny SPRL ().\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\n\nfrom openerp import addons\nimport logging\nfrom openerp.osv import fields, osv\nfrom openerp.tools.translate import _\nfrom openerp import tools\nfrom openerp.addons.xlrd import open_workbook\nfrom openerp import modules\nimport os\nbase_path = os.path.dirname(modules.get_module_path('ql_vanban'))\n_logger = logging.getLogger(__name__)\n\nclass res_company(osv.osv):\n _inherit = \"res.company\"\n _columns = {\n 'nguoi_lanh_dao_id': fields.many2one('hr.employee','Tên người lãnh đạo'),\n 'ma': fields.char('Mã', size=64),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/co_quan.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten_lanhdao = s.cell(row,0).value\n ma = s.cell(row,1).value\n ten_coquan = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('name','=',ten_coquan)])\n if not danhmuc_ids:\n nguoi_lanh_dao_ids = self.pool.get('hr.employee').search(cr, 1, [('name','=',ten_lanhdao)])\n if nguoi_lanh_dao_ids:\n self.create(cr, 1, {'name': ten_coquan,'nguoi_lanh_dao_id':nguoi_lanh_dao_ids[0],'ma':ma})\n \nres_company()\n\nclass cap_goi(osv.osv):\n _name = \"cap.goi\"\n _description = \"Cap goi\"\n _columns = {\n 'name': fields.char('Tên cấp văn bản', size=64, required=True),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/cap_goi.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten = s.cell(row,0).value\n danhmuc_ids = self.search(cr, 1, [('name','=',ten)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten})\n \ncap_goi()\n\nclass noi_phathanh(osv.osv):\n _name = \"noi.phathanh\"\n _description = \"Noi phat hanh\"\n _columns = {\n 'name': fields.char('Tên đơn vị', size=64, required=True),\n 'ky_hieu': fields.char('Ký hiệu', size=64, required=True),\n 'ghi_chu': fields.text('Ghi chú'),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/noi_phat_hanh.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ky_hieu = s.cell(row,0).value\n ten = s.cell(row,1).value\n ghi_chu = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('ky_hieu','=',ky_hieu)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ky_hieu':ky_hieu,'ghi_chu':ghi_chu})\nnoi_phathanh()\n\nclass loai_sovanban(osv.osv):\n _name = \"loai.sovanban\"\n _description = \"Loai so van ban\"\n _columns = {\n 'name': fields.char('Tên loại sổ văn bản', size=64, required=True),\n 'ky_hieu': fields.char('Ký hiệu', size=64, required=True),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/loai_so_van_ban.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten = s.cell(row,0).value\n ky_hieu = s.cell(row,1).value\n danhmuc_ids = self.search(cr, 1, [('ky_hieu','=',ky_hieu)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ky_hieu':ky_hieu})\n \nloai_sovanban()\n\nclass so_vanban(osv.osv):\n _name = \"so.vanban\"\n _description = \"So van ban\"\n _columns = {\n 'name': fields.char('Loại sổ văn bản', size=64, required=True),\n 'co_quan_id': fields.many2one('res.company','Cơ quan', required=True),\n 'ngay_tao': fields.date('Ngày tạo'),\n 'nam': fields.integer('Năm'),\n 'so_hien_tai': fields.integer('Số hiện tại'),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/so_van_ban.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n loai = s.cell(row,0).value\n co_quan = s.cell(row,1).value\n ngay_tao = s.cell(row,2).value\n nam = s.cell(row,3).value\n so_hien_tai = s.cell(row,4).value\n danhmuc_ids = self.search(cr, 1, [('name','=',loai)])\n if not danhmuc_ids:\n co_quan_ids = self.pool.get('res.company').search(cr, 1, [('name','=',co_quan)])\n if co_quan_ids:\n self.create(cr, 1, {'name': loai,'co_quan':co_quan_ids[0],'ngay_tao':ngay_tao,'nam':nam,'so_hien_tai':so_hien_tai})\n \nso_vanban()\n\nclass loai_vanban(osv.osv):\n _name = \"loai.vanban\"\n _description = \"Loai van ban\"\n _columns = {\n 'name': fields.char('Tên loại văn bản', size=64, required=True),\n 'ky_hieu': fields.char('Ký hiệu', size=64, required=True),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/loai_van_ban.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten = s.cell(row,0).value\n ky_hieu = s.cell(row,1).value\n danhmuc_ids = self.search(cr, 1, [('ky_hieu','=',ky_hieu)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ky_hieu':ky_hieu})\n \nloai_vanban()\n\nclass do_mat(osv.osv):\n _name = \"do.mat\"\n _description = \"Do mat\"\n _order = \"thu_tu\"\n _columns = {\n 'name': fields.char('Tên độ khẩn', size=64, required=True),\n 'ma': fields.char('Mã độ khẩn', size=64, required=True),\n 'thu_tu': fields.integer('Thứ tự ưu tiên sắp xếp', required=True),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/do_mat.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten = s.cell(row,0).value\n ma = s.cell(row,1).value\n thu_tu = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('ma','=',ma)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ma':ma,'thu_tu':thu_tu})\n \ndo_mat()\n\nclass do_khan(osv.osv):\n _name = \"do.khan\"\n _description = \"Do khan\"\n _columns = {\n 'name': fields.char('Tên độ khẩn', size=64, required=True),\n 'ma': fields.char('Mã độ khẩn', size=64, required=True),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/do_khan.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ten = s.cell(row,0).value\n ma = s.cell(row,1).value\n danhmuc_ids = self.search(cr, 1, [('ma','=',ma)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ma':ma})\n \ndo_khan()\n\nclass hr_employee_category(osv.osv):\n _inherit = \"hr.employee.category\"\n _columns = {\n 'ma': fields.char('Mã', size=64),\n 'mota': fields.char('Mô tả', size=128),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/chuc_vu.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ma = s.cell(row,0).value\n ten = s.cell(row,1).value\n mota = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('ma','=',ma)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ma':ma,'mota':mota})\n \nhr_employee_category()\n\nclass linh_vuc(osv.osv):\n _name = \"linh.vuc\"\n _description = \"Linh vuc\"\n _columns = {\n 'name': fields.char('Tên', size=64, required=True),\n }\nlinh_vuc()\n\nclass so_vanban_noibo(osv.osv):\n _name = \"so.vanban.noibo\"\n _description = \"So van ban noi bo\"\n _columns = {\n 'name': fields.char('Tên sổ văn bản nội bộ', size=64, required=True),\n 'ky_hieu': fields.char('Mã sổ văn bản nội bộ', size=64, required=True),\n 'mota': fields.char('Mô tả sổ văn bản nội bộ', size=128),\n 'loai_vb_nb_ids': fields.many2many('loai.vanban.noibo','so_loai_vanban_noibo_ref','so_id','loai_id','Cấu hình loại văn bản nội bộ'),\n 'phong_ban_ids': fields.many2many('hr.department','so_phongban_vanban_noibo_ref','so_id','hr_department_id','Cấu hình phòng ban'),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/so_van_ban_noi_bo.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ky_hieu = s.cell(row,0).value\n ten = s.cell(row,1).value\n mota = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('ky_hieu','=',ky_hieu)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ky_hieu':ky_hieu,'mota':mota})\n \nso_vanban_noibo()\n\nclass loai_vanban_noibo(osv.osv):\n _name = \"loai.vanban.noibo\"\n _description = \"Loai van ban noi bo\"\n _columns = {\n 'name': fields.char('Tên loại văn bản nội bộ', size=64, required=True),\n 'ky_hieu': fields.char('Ký hiệu loại văn bản nội bộ', size=64, required=True),\n 'mota': fields.char('Mô tả loại văn bản nội bộ', size=128),\n }\n def init(self, cr):\n wb = open_workbook(base_path + '/ql_vanban/data/loai_van_ban_noi_bo.xls')\n for s in wb.sheets():\n if (s.name =='Sheet1'):\n for row in range(1,s.nrows):\n ky_hieu = s.cell(row,0).value\n ten = s.cell(row,1).value\n mota = s.cell(row,2).value\n danhmuc_ids = self.search(cr, 1, [('ky_hieu','=',ky_hieu)])\n if not danhmuc_ids:\n self.create(cr, 1, {'name': ten,'ky_hieu':ky_hieu,'mota':mota})\n \nloai_vanban_noibo()\n\nclass vanban_den(osv.osv):\n _name = \"vanban.den\"\n _description = \"Van ban den\"\n _columns = {\n 'name': fields.char('Số ký hiệu gốc', size=64, required=True),\n 'so_vanban_id': fields.many2one('so.vanban','Sổ văn bản'),\n 'so_den_theo_so': fields.char('Số đến theo sổ', size=64),\n 'linh_vuc_id': fields.many2one('linh.vuc','Lĩnh vực'),\n 'ngay_den': fields.date('Ngày đến'),\n 'ngay_phat_hanh': fields.date('Ngày phát hành', required=True),\n 'donvi_saoy': fields.char('Đơn vị phát hành', size=64),\n 'nguoi_ky_id': fields.many2one('hr.employee','Người ký'),\n 'do_khan_id': fields.many2one('do.khan','Độ khẩn'),\n 'so_to': fields.char('Số tờ', size=64),\n 'trich_yeu': fields.text('Trích yếu', required=True),\n 'loai_vanban_id': fields.many2one('loai.vanban','Loại văn bản'),\n 'vanban_qppl': fields.boolean('Văn bản QPPL'),\n 'cap_goi_id': fields.many2one('cap.goi','Cấp gởi'),\n 'noi_phathanh_id': fields.many2one('noi.phathanh','Nơi phát hành', required=True),\n 'khac': fields.char('Khác', size=64),\n 'so_vanban_di_phucdap': fields.char('Số văn bản đi phúc đáp', size=64),\n 'do_mat_id': fields.many2one('do.mat','Độ mật'),\n 'hoso_lines': fields.one2many('ir.attachment','vanban_den_id','Tập tin đính kèm',required=True),\n 'state': fields.selection([('moitao','Mới Tạo'),('dangxuly','Đang xử lý'),('daxuly','Đã xử lý')], 'Tình trạng'),\n }\n \n _defaults = {\n 'state': 'moitao',\n }\n \n def create(self, cr, uid, vals, context=None):\n if 'hoso_lines' in vals and len(vals['hoso_lines'])==0:\n raise osv.except_osv(_('Cảnh báo!'),\n _('Vui lòng đính kèm tập tin!')) \n return super(vanban_den,self).create(cr, uid, vals, context)\n \n def write(self, cr, uid, ids, vals, context=None):\n new_write = super(vanban_den,self).write(cr, uid, ids, vals, context)\n for line in self.browse(cr, uid, ids):\n if len(line.hoso_lines)==0:\n raise osv.except_osv(_('Cảnh báo!'),\n _('Vui lòng đính kèm tập tin!')) \n return new_write\n \nvanban_den()\n\n# vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:\n","sub_path":"addons-khdt/ql_vanban/vanban_den.py","file_name":"vanban_den.py","file_ext":"py","file_size_in_byte":14057,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"239633689","text":"n, k = map(int, input().split())\ncoins = [int(input()) for _ in range(n)]\n\ndp = [0] * (k+1)\ndp[0] = 1\n\nfor coin in coins:\n for num in range(1, k + 1):\n if num - coin >= 0:\n dp[num] += dp[num - coin]\n\nprint(dp[k])","sub_path":"dp/2293.py","file_name":"2293.py","file_ext":"py","file_size_in_byte":233,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"649336617","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('cards', '0005_deckcards_card'),\n ]\n\n operations = [\n migrations.AlterUniqueTogether(\n name='deckcards',\n unique_together=set([('card', 'deck')]),\n ),\n ]\n","sub_path":"elements/core/cards/migrations/0006_auto_20151114_1429.py","file_name":"0006_auto_20151114_1429.py","file_ext":"py","file_size_in_byte":377,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"437952407","text":"import math\n\n\"\"\"\ndata filters:\ntakes a series of run data and applies statistical transforms to it\n\"\"\"\n\n### filters that return a scalar\n\ndef mean(series):\n \"\"\"\n mean of data; needs at least one data point\n \"\"\"\n return sum(series)/float(len(series))\n\ndef median(series):\n \"\"\"\n median of data; needs at least one data point\n \"\"\"\n series = sorted(series)\n if len(series) % 2:\n # odd\n return series[len(series)/2]\n else:\n # even\n middle = len(series)/2 # the higher of the middle 2, actually\n return 0.5*(series[middle-1] + series[middle])\n\ndef variance(series):\n \"\"\"\n variance: http://en.wikipedia.org/wiki/Variance\n \"\"\"\n\n _mean = mean(series)\n variance = sum([(i-_mean)**2 for i in series])/float(len(series))\n return variance\n\ndef stddev(series):\n \"\"\"\n standard deviation: http://en.wikipedia.org/wiki/Standard_deviation\n \"\"\"\n return variance(series)**0.5\n\ndef dromaeo(series):\n \"\"\"\n dromaeo: https://wiki.mozilla.org/Dromaeo, pull the internal calculation out\n * This is for 'runs/s' based tests, not 'ms' tests.\n * chunksize: defined in dromaeo: page_load_test/dromaeo/webrunner.js#l8\n \"\"\"\n means = []\n chunksize = 5\n series = list(dromaeo_chunks(series, chunksize))\n for i in series:\n means.append(mean(i))\n return geometric_mean(means)\n\ndef dromaeo_chunks(series, size):\n for i in xrange(0, len(series), size):\n yield series[i:i+size]\n\ndef geometric_mean(series):\n \"\"\"\n geometric_mean: http://en.wikipedia.org/wiki/Geometric_mean\n \"\"\"\n total = 0\n for i in series:\n total += math.log(i+1)\n return math.exp(total / len(series)) - 1\n\nscalar_filters = [mean, median, max, min, variance, stddev, dromaeo]\n\n### filters that return a list\n\ndef ignore_first(series, number=1):\n \"\"\"\n ignore first datapoint\n \"\"\"\n if len(series) <= number:\n # don't modify short series\n return series\n return series[number:]\n\ndef ignore(series, function):\n \"\"\"\n ignore the first value of a list given by function\n \"\"\"\n if len(series) <= 1:\n # don't modify short series\n return series\n series = series[:] # do not mutate the original series\n value = function(series)\n series.remove(value)\n return series\n\ndef ignore_max(series):\n \"\"\"\n ignore maximum data point\n \"\"\"\n return ignore(series, max)\n\ndef ignore_min(series):\n \"\"\"\n ignore minimum data point\n \"\"\"\n return ignore(series, min)\n\nseries_filters = [ignore_first, ignore_max, ignore_min]\n\n### mappings\n\nscalar_filters = dict([(i.__name__, i) for i in scalar_filters])\nseries_filters = dict([(i.__name__, i) for i in series_filters])\n\n### utility functions\n\ndef parse(filter_name):\n \"\"\"\n parses a filter_name like\n \"ignore_first:10\" to return\n ['ignore_first', [10]]\n or \"foo:10.1,2,5.0\" to return\n ['foo', [10.1, 2, 5.0]] .\n The filter name strings are returned versus the functions'\n as the data may need to be reserialized (e.g. PerfConfigurator.py).\n \"\"\"\n\n sep = ':'\n\n def convert_to_number(string):\n \"\"\"convert a string to an int or float\"\"\"\n try:\n return int(string)\n except ValueError:\n return float(string)\n\n args = []\n if sep in filter_name:\n filter_name, args = filter_name.split(sep, 1)\n args = [convert_to_number(arg)\n for arg in args.split(',')]\n # check validity of filter\n assert (filter_name in scalar_filters) or (filter_name in series_filters),\\\n \"--filter value not found in filters.\"\n return [filter_name, args]\n\ndef filters(*filter_names):\n \"\"\"\n return a list of filter functions given a list of names\n \"\"\"\n\n # convert to a list\n filter_names = list(filter_names)\n if not filter_names:\n return []\n\n # sanity checks\n allowable_filters = set(scalar_filters.keys() + series_filters.keys())\n missing = [i for i in filter_names if i not in allowable_filters]\n if missing:\n raise AssertionError(\"Filters not found: %s; (allowable filters: %s)\" % (', '.join(missing), ', '.join(allowable_filters)))\n reducer = filter_names.pop()\n assert reducer in scalar_filters, \"Last filter must return a scalar: %s, you gave %s\" % (scalar_filters.keys(), reducer)\n assert set(filter_names).issubset(series_filters), \"All but last filter must return a series: %s, you gave %s\" % (series_filters.keys(), filter_names)\n\n # get the filter functions\n retval = [series_filters[i] for i in filter_names]\n retval.append(scalar_filters[reducer])\n return retval\n\ndef filters_args(_filters):\n \"\"\"\n convenience function to take a list of\n [['filter_name', args]] and convert these to functions\n \"\"\"\n retval = []\n filter_names = [f[0] for f in _filters]\n filter_functions = filters(*filter_names)\n for index, value in enumerate(_filters):\n retval.append([filter_functions[index], value[-1]])\n return retval\n\ndef apply(data, filters):\n \"\"\"apply filters to a data series. does no safety check\"\"\"\n for f in filters:\n args = ()\n if isinstance(f, list) or isinstance(f, tuple):\n if len(f) == 2: # function, extra arguments\n f, args = f\n elif len(f) == 1: # function\n f = f[0]\n else:\n raise AssertionError(\"Each value must be either [filter, [args]] or [filter]\")\n data = f(data, *args)\n return data\n","sub_path":"talos/filter.py","file_name":"filter.py","file_ext":"py","file_size_in_byte":5509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"235071250","text":"import socket\nimport os\nimport requests\nfrom requests.auth import HTTPBasicAuth\nimport json\nimport itertools\nimport sqlparse\nimport re\nfrom hashlib import sha256\n\nHOSTNAME = socket.gethostname()\nIPADDR = socket.gethostbyname(HOSTNAME)\nDATA_MASK_POLICY_ITEMS = 'dataMaskPolicyItems'\nDATA_MASK_INFO = 'dataMaskInfo'\nDATA_MASK_TYPE = 'dataMaskType'\nROW_FILTER_POLICY_ITEMS = 'rowFilterPolicyItems'\nROW_FILTER_INFO = 'rowFilterInfo'\nFILTER_EXPRESSION = 'filterExpr'\nPOLICY_ITEMS = 'policyItems'\nDENY_POLICY_ITEMS = 'denyPolicyItems'\nACCESSES = 'accesses'\nRESOURCES = 'resources'\nCOLUMN = 'column'\nTABLE = 'table'\nVALUES = 'values'\nVALUE_EXPRESSION = 'valueExpr'\nTYPE = 'type'\nIS_ALLOWED = 'isAllowed'\nDEBUG = True #False\n\nclass MLflowRangerAccess:\n def __init__(self, user):\n self.user = user\n self.acceptExperimentIds = set([])\n self.acceptRunIds = set([])\n self.canCreate = False\n\n\n def sync(self,role='select', url='http://ranger-admin.kubeflow.svc.cluster.local:6080/service/public/v2/api/service/primary_hive/policy'):\n policies = get_ranger_policies_for(self.user, url=url, admin_user='admin', admin_pass='Ranger123', database='mlflow')\n for policy in policies:\n print('Found policy: ', policy)\n if POLICY_ITEMS in policy:\n filter_policies = policy[POLICY_ITEMS]\n for filter_policy in filter_policies:\n if DEBUG:\n print('Found filter policy: ', filter_policy)\n if ACCESSES in filter_policy:\n accesses = filter_policy[ACCESSES]\n experiments = policy[RESOURCES][TABLE][VALUES]\n runs = policy[RESOURCES][COLUMN][VALUES]\n for access in accesses:\n if (access[TYPE] == role or access[TYPE]=='all') and access[IS_ALLOWED]:\n for experiment in experiments:\n if experiment is not None and experiment != '*':\n if DEBUG:\n print('Accepting experiment: ', experiment)\n self.acceptExperimentIds.add(experiment)\n if (access[TYPE] == 'create' or access[TYPE]=='all') and access[IS_ALLOWED]:\n self.canCreate = True\n print('Can create experiments!')\n\n\n if DENY_POLICY_ITEMS in policy:\n filter_policies = policy[DENY_POLICY_ITEMS]\n for filter_policy in filter_policies:\n if DEBUG:\n print('Found deny filter policy: ', filter_policy)\n if ACCESSES in filter_policy:\n accesses = filter_policy[ACCESSES]\n experiments = policy[RESOURCES][TABLE][VALUES]\n runs = policy[RESOURCES][COLUMN][VALUES]\n for access in accesses:\n if access[TYPE] == role and access[IS_ALLOWED]:\n for experiment in experiments:\n if experiment is not None and experiment != '*':\n if DEBUG:\n print('Denying experiment: ', experiment)\n if experiment in self.acceptExperimentIds:\n self.acceptExperimentIds.remove(experiment)\n\n def canAccess(self, experiment_id):\n return experiment_id in self.acceptExperimentIds\n\n def canAccessRun(self, run_id):\n return run_id in self.acceptRunIds\n\n def canCreateExperiment(self):\n return self.canCreate\n\n\ndef get_ranger_policies_for(user, url, admin_user, admin_pass, database='mlflow', experiment_ids=[None], run_ids=[None]):\n all_policies = []\n for experiment_id in experiment_ids:\n for run_id in run_ids:\n params = {'user': user, 'isEnabled': 'true'}\n if database is not None:\n params['resource:database'] = database\n if experiment_id is not None:\n params['resource:table'] = experiment_id\n if run_id is not None:\n params['resource:column'] = run_id\n res = requests.get(url, params=params, auth=HTTPBasicAuth(admin_user, admin_pass))\n if DEBUG:\n text = res.text\n print('TEXT:', text)\n policies = json.loads(text)\n else:\n policies = res.json()\n [all_policies.append(policy) for policy in policies]\n return all_policies\n","sub_path":"mlflow/tracking/ranger.py","file_name":"ranger.py","file_ext":"py","file_size_in_byte":4749,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"150789968","text":"# We need the urllib.request to get the website.\nfrom urllib.request import Request, urlopen\n\n# Get url of the league table page.\nurl = \"http://www.bbc.com/sport/football/premier-league/table\"\n\n# The 'content' contains all elements of the page.\ncontent = urlopen(Request(url)).read().decode(\"utf-8\")\n\n# The 'league_table' 2D list contains the league table of the current season.\nleague_table = [[0 for x in range(11)] for y in range(21)]\n\nleague_table[0] = [\"Position\", \"Team\", \"Played\", \"Won\", \"Draw\", \"Lost\",\n \"Goal For\", \"Goal Against\", \"Goal Difference\", \"Points\",\n \"Last 6\"]\n\n# 'tbody' stores information of the table body.\ntbody = content.split(\"\")[1].split(\"\")[0]\n","sub_path":"get_league_table.py","file_name":"get_league_table.py","file_ext":"py","file_size_in_byte":819,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"16588782","text":"# Ejercicios de práctica numérica\r\ndef ej1():\r\n numero_1 = 5\r\n numero_2 = 7\r\n \r\n operacion= numero_1 + numero_2\r\n print(operacion)\r\n \r\n print('######## LINEA ########')\r\n \r\n operacionresta = numero_1 - numero_2\r\n \r\n print(operacionresta)\r\n\r\nif __name__ == '__main__':\r\n print(\"Bienvenidos a otra clase de Inove con Python\")\r\n ej1()","sub_path":"ejercicios_draft/ejercicio_1.py","file_name":"ejercicio_1.py","file_ext":"py","file_size_in_byte":372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"12581993","text":"import Adafruit_BBIO.UART as uart\n\nfrom pubsub import pub\nfrom threading import Thread, Lock\nimport time\n\nfrom src.arduino.config_reader import read_arduino_config\nfrom src.arduino.config_reader import read_port_config\nfrom src.arduino.config_reader import read_pin_config\n\nclass Arduino(Thread):\n \"\"\" Provides an interface to arudino connected over UART \"\"\"\n\n def __init__(self, mock_bbio=None, mock_port=None):\n \"\"\"\n Initializes arduino thread, subscribes update methods to their respective channels\n \"\"\"\n super().__init__()\n self.is_active = True\n self.config = read_arduino_config()\n self.update_interval = self.config['update_interval']\n self.uart_pin = read_pin_config(mock_bbio=mock_bbio)\n self.port = read_port_config(mock_port=mock_port)\n # initialize data to send\n self.data = {\"rudder_ang\": 0,\n \"sail_ang\": 0,\n \"rear_foil_ang\": 0,\n \"jib_ang\": 0,\n \"sensor_ang\": 0}\n\n # subscribe update functions to their respective pubsub channels\n pub.subscribe(self.update_rudder_ang, \"turn rudder to\")\n pub.subscribe(self.update_sail_ang, \"turn sail to\")\n pub.subscribe(self.update_rear_foil_ang, \"turn rear foil to\")\n pub.subscribe(self.update_jib_ang, \"turn jib to\")\n pub.subscribe(self.update_sensor_ang, \"turn sensor to\")\n \n\n def run(self):\n \"\"\" Runs the arduino comms thread \"\"\"\n print(\"Started arduino thread\")\n while self.is_active:\n for val in self.data.values():\n pass # send over UART\n print(time.time())\n time.sleep(self.update_interval)\n\n def update_rudder_ang(self, rudder_ang):\n \"\"\" udpates rudder angle from pub sub \"\"\"\n self.data[\"rudder_ang\"] = rudder_ang\n \n def update_sail_ang(self, sail_ang):\n \"\"\" udpates sail angle from pub sub \"\"\"\n self.data[\"sail_ang\"] = sail_ang\n\n def update_rear_foil_ang(self, rear_foil_ang):\n \"\"\" udpates rear foil angle from pub sub \"\"\"\n self.data[\"rear_foil_ang\"] = rear_foil_ang \n \n def update_jib_ang(self, jib_ang):\n \"\"\" udpates jib angle from pub sub \"\"\"\n self.data[\"jib_ang\"] = jib_ang\n\n def update_sensor_ang(self, sensor_ang):\n \"\"\" udpates sensor angle from pub sub \"\"\"\n self.data[\"sensor_ang\"] = sensor_ang\n\n def disable_controls(self):\n \"\"\" disables arduino comms\"\"\"\n self.is_active = False\n\n def enable_controls(self):\n \"\"\" enables arduino comms\"\"\"\n self.is_active = True\n","sub_path":"src/arduino/arduino.py","file_name":"arduino.py","file_ext":"py","file_size_in_byte":2664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"7513066","text":"from django.http import Http404\n\nfrom jwtotp import settings\nfrom . import models\nfrom .models import User\nfrom rest_framework.views import APIView\nfrom rest_framework.response import Response\nfrom rest_framework import status\nfrom .serializer import UserSerializer,UserProfileSerializer,JobsSerializer\nimport random\nimport json\nimport jwt\nimport os\nimport sys\nimport uuid\nfrom django.utils import timezone\nfrom django.conf import settings\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom rest_framework.parsers import FileUploadParser, MultiPartParser,FormParser\n# from . import jwtToken\nfrom otpapp.jwtauthuser import JSONWebTokenAuthentication as jwta\n\n# Create your views here.\nclass Userview(APIView):\n \"\"\"\n List all snippets, or create a new snippet.\n \"\"\"\n def get(self, request, format=None):\n request.META[\"token\"]=\"asnfjksdbfhjzdbfhjzfjsbfhjsbfhjsfhjsdfvshjf\"\n user = User.objects.all()\n serializer = UserSerializer(user, many=True)\n return Response(serializer.data)\n def isuserexist(self,contact):\n try:\n userdbobj=User.objects.get(contact=contact)\n return userdbobj\n except :\n return None\n\n\n\n def post(self, request, format=None):\n data=request.data\n if \"+\" in str(data['contact']):\n data['contact']=str(data[\"contact\"])[3:].replace(' ',\"\").replace(\"-\",\"\")\n \n serializer = UserSerializer(data=data)\n \n contact = int(data[\"contact\"])\n iscontact=self.isuserexist(contact)\n if serializer.is_valid():\n if iscontact is None:\n serializer.save()\n otpnum = random.randint(1111, 9999)\n userdbobj =User.objects.get(contact=contact)\n userdbobj.otp=otpnum\n userdbobj.save()\n\n # here should be sms code...\n response={\"success\":True,\"otp\":otpnum}\n# response=json.dumps(response)\n return Response(response, status=status.HTTP_201_CREATED)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n \nclass Otpverify(APIView):\n \"\"\"\n List all snippets, or create a new snippet.\n \"\"\"\n def post(self, request, format=None):\n data=request.data\n print(data)\n if \"+\" in str(data['contact']):\n data['contact']=str(data[\"contact\"])[3:].replace(' ',\"\").replace(\"-\",\"\")\n contact = int(data[\"contact\"])\n otp = int(data[\"otp\"])\n print(\"OTP: \",otp)\n userdbobj=User.objects.get(contact=contact)\n availotp=userdbobj.otp\n if otp==availotp:\n now=timezone.now()\n timed=24*60*60*now.day+now.hour*60*60+now.minute*60+now.second\n age=str(timed+settings.TOKEN_EXPIRE_TIME)\n jwtvar= jwta(userdbobj).get_user_jwt(age)\n print(jwtvar)\n userdbobj.token=jwtvar\n userdbobj.save()\n response={\"success\":True,\"token\":jwtvar,'id':userdbobj.id}\n return Response(response,status.HTTP_200_OK)\n else:\n response={\"success\":False}\n return Response(response,status=status.HTTP_401_UNAUTHORIZED)\n\n# def jwtsignature(self,contact):\n#\n# encoded_jwt = jwt.encode({'contact': contact}, 'secret', algorithm='HS256')\n# response={'token':encoded_jwt}\n# return response\n \nclass UserProfileView(APIView):\n \"\"\"\n Retrieve, update or delete a snippet instance.\n \"\"\"\n def get_object(self, pk):\n try:\n return User.objects.get(pk=pk)\n except UserProfileSerializer.DoesNotExist:\n raise Http404\n\n \n def get(self, request, pk, format=None):\n print(request.META)\n token=request.META['HTTP_AUTHORIZATION']\n payload=jwta().authenticate_credentials(token)\n print(\"payload\",payload)\n if payload:\n pass\n profile = self.get_object(pk)\n serializer = UserProfileSerializer(profile)\n return Response(serializer.data)\n\n def put(self, request, pk, format=None):\n profile = self.get_object(pk)\n serializer = UserProfileSerializer(profile, data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data)\n return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)\n\n def delete(self, request, pk, format=None):\n profile = self.get_object(pk)\n profile.delete()\n return Response(status=status.HTTP_204_NO_CONTENT)\n \nclass UserFileView(APIView):\n \"\"\"\n Upload flle upload view for FileUpload model\n \"\"\"\n parser_classes = (MultiPartParser, FormParser,FileUploadParser,)\n def post(self, request, format=None):\n \"\"\"\n\n :param request:\n :param format:\n :return:\n \"\"\"\n #try:\n \n token=request.META['HTTP_AUTHORIZATION']\n payload=jwta().authenticate_credentials(token)\n \n print(\"payload ----\",payload)\n if payload:\n data = request.data\n print(data)\n file_name = data[\"file_name\"]\n \n contact = int(payload[\"contact\"])\n# job_name=data[\"job_name\"]\n size=data[\"size\"]\n job_type=data[\"job_type\"]\n paper_material=data[\"paper_material\"]\n quantity=data[\"quantity\"]\n description=data[\"description\"]\n user = models.User.objects.get(contact=contact)\n print(user,\"--------------------\")\n doc_obj=models.Jobs.objects.create(user=user,\n file_name=file_name,\n description=description,\n size=size,\n job_type=job_type,\n quantity=quantity,\n paper_material=paper_material,\n \n )\n doc_obj.save()\n file_url=str(doc_obj.file_name)\n\n RESPONSE = {'success': True,\n 'response': {'file_name': file_url}}\n return Response(json.dumps(RESPONSE), status=status.HTTP_200_OK)\n else:\n return Response( status=status.HTTP_401_UNAUTHORIZED)\n\n\n\n #return Response(BEDRESPONSE, status=status.HTTP_401_UNAUTHORIZED)\n\n # except Exception as e:\n # exc_type, exc_obj, exc_tb = sys.exc_info()\n # print(sys.exc_info())\n # BEDRESPONS = {'success': False,\n # 'response': str(sys.exc_info())}\n # return Response(json.dumps(BEDRESPONS), status=status.HTTP_400_BAD_REQUEST)\n\n \nclass JobDetailsView(APIView):\n\n def get(self, request, format=None):\n token=request.META['HTTP_AUTHORIZATION']\n payload=jwta().authenticate_credentials(token)\n if payload:\n user = models.User.objects.get(contact=payload[\"contact\"])\n jobs=models.Jobs.objects.filter(user=user)\n serializer = JobsSerializer(jobs, many=True)\n data=[]\n for el in serializer.data:\n if el['quotation']:\n el['quotation']=settings.BASE_URL+el['quotation']\n print(el['quotation'])\n data.append(el)\n \n \n return Response(data)\n else:\n return Response( status=status.HTTP_400_BAD_REQUEST)","sub_path":"otpapp/.ipynb_checkpoints/views-checkpoint.py","file_name":"views-checkpoint.py","file_ext":"py","file_size_in_byte":7573,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"456717259","text":"import requests,sys\nfrom util import timer\n\n@timer.timer('01:00:10') #9:00:10\ndef check_in(COOKIE):\n url = \"http://39.96.69.63/phone/yuyue?id=31&states=2\"\n payload = 'id=31&states=2'\n\n headers = {\n 'Host': '39.96.69.63',\n 'Content-Length': '14',\n 'Origin': 'http://39.96.69.63',\n 'X-Requested-With': 'XMLHttpRequest',\n 'User-Agent': 'Mozilla/5.0 (Linux; Android 7.0; MI 5 Build/XK_Gemiui_1.3.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Crosswalk/22.52.561.4 Mobile Safari/537.36',\n 'Content-Type': 'application/x-www-form-urlencoded; charset=UTF-8',\n 'Referer': 'http://39.96.69.63/phone/signup',\n 'Cookie': COOKIE\n }\n\n response = requests.request(\"POST\", url, headers=headers, data = payload)\n return response.text.encode('utf8')\n\nif __name__ == \"__main__\":\n COOKIE = sys.argv[1]\n try:\n response = check_in(COOKIE)\n print(response)\n except Exception as e:\n print('运行出错,原因:%s' % e)","sub_path":"scripts/unicom_food.py","file_name":"unicom_food.py","file_ext":"py","file_size_in_byte":988,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"513370387","text":"import os\nimport configparser\nfrom logging import getLogger\n\nlogger = getLogger(__name__)\n\ndef _project_dir():\n return os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n\ndef _data_dir():\n return os.path.join(_project_dir(), \"data\")\n\ndef _model_dir():\n return os.path.join(_project_dir(), \"model\")\n\nclass Config:\n def __init__(self):\n self.resource = ResourceConfig()\n self.model = ModelConfig()\n self.generator = GeneratorConfig(self.model)\n self.discriminator = DiscriminatorConfig(self.model)\n self.trainer = TrainerConfig()\n \n def load_parameter(self):\n if os.path.exists(self.resource.config_path):\n logger.debug(\"loading parameter from {}\".format(self.resource.config_path))\n config_parser = configparser.ConfigParser()\n config_parser.read(self.resource.config_path, encoding='utf-8')\n read_model = config_parser['MODEL']\n if read_model.get(\"Input_dimension\") is not None: self.model.dim_in = int(read_model.get(\"Input_dimension\"))\n if read_model.get(\"Output_dimension\") is not None: self.model.dim_out = int(read_model.get(\"Output_dimension\"))\n if read_model.get(\"Noise_dimension\") is not None: self.model.dim_noise = int(read_model.get(\"Noise_dimension\"))\n if read_model.get(\"Maximum_time\") is not None: self.model.max_time = float(read_model.get(\"Maximum_time\"))\n if read_model.get(\"Weights_division\") is not None: self.model.division = int(read_model.get(\"Weights_division\"))\n if read_model.get(\"Generator_function_x_type\") is not None: self.model.generator_function_x_type = read_model.get(\"Generator_function_x_type\")\n if read_model.get(\"Generator_function_y_type\") is not None: self.model.generator_function_y_type = read_model.get(\"Generator_function_y_type\")\n if read_model.get(\"discriminator_function_x_type\") is not None: self.model.discriminator_function_x_type = read_model.get(\"discriminator_function_x_type\")\n if read_model.get(\"discriminator_function_y_type\") is not None: self.model.discriminator_function_y_type = read_model.get(\"discriminator_function_y_type\")\n read_trainer = config_parser['TRAINER']\n if read_trainer.get(\"Optimizer_type\") is not None: self.trainer.optimizer_type = read_trainer.get(\"Optimizer_type\")\n if read_trainer.get(\"Loss_type\") is not None: self.trainer.loss_type = read_trainer.get(\"Loss_type\")\n if read_trainer.get(\"Learning_rate\") is not None: self.trainer.rate = float(read_trainer.get(\"Learning_rate\"))\n if read_trainer.get(\"Momentum\") is not None: self.trainer.momentum = float(read_trainer.get(\"Momentum\"))\n if read_trainer.get(\"Decay\") is not None: self.trainer.decay = float(read_trainer.get(\"Decay\"))\n if read_trainer.get(\"Decay2\") is not None: self.trainer.decay2 = float(read_trainer.get(\"Decay2\"))\n if read_trainer.get(\"Regularization_rate\") is not None: self.trainer.regularization = float(read_trainer.get(\"Regularization_rate\"))\n if read_trainer.get(\"Epoch\") is not None: self.trainer.epoch = int(read_trainer.get(\"Epoch\"))\n if read_trainer.get(\"Batch_size\") is not None: self.trainer.batch_size = int(read_trainer.get(\"Batch_size\"))\n if read_trainer.get(\"Is_visualize\") is not None: self.trainer.is_visualize = bool(int(read_trainer.get(\"Is_visualize\")))\n self.generator = GeneratorConfig(self.model)\n self.discriminator = DiscriminatorConfig(self.model)\n \n def save_parameter(self, config_path):\n config_parser = configparser.ConfigParser()\n config_parser[\"MODEL\"] = {\n \"Input_dimension\": self.model.dim_in,\n \"Output_dimension\": self.model.dim_out,\n \"Noise_dimension\": self.model.dim_noise,\n \"Maximum_time\": self.model.max_time,\n \"Weights_division\": self.model.division,\n \"Generator_function_x_type\": self.model.generator_function_x_type,\n \"Generator_function_y_type\": self.model.generator_function_y_type,\n \"discriminator_function_x_type\": self.model.discriminator_function_x_type,\n \"discriminator_function_y_type\": self.model.discriminator_function_y_type\n }\n config_parser[\"TRAINER\"] = {\n \"Optimizer_type\": self.trainer.optimizer_type,\n \"Loss_type\": self.trainer.loss_type,\n \"Learning_rate\": self.trainer.rate,\n \"Momentum\": self.trainer.momentum,\n \"Decay\": self.trainer.decay,\n \"Decay2\": self.trainer.decay2,\n \"Regularization_rate\": self.trainer.regularization,\n \"Epoch\": self.trainer.epoch,\n \"Batch_size\": self.trainer.batch_size,\n \"Is_visualize\": int(self.trainer.is_visualize)\n }\n with open(config_path, \"wt\") as f:\n config_parser.write(f)\n\n\nclass ResourceConfig:\n def __init__(self):\n self.project_dir = os.environ.get(\"PROJECT_DIR\", _project_dir())\n self.data_dir = os.environ.get(\"DATA_DIR\", _data_dir())\n self.data_processed_dir = os.path.join(self.data_dir, \"processed\")\n self.data_path = os.path.join(self.data_processed_dir, \"data.json\")\n self.model_dir = os.environ.get(\"MODEL_DIR\", _model_dir())\n self.generator_model_path = os.path.join(self.model_dir, \"generator.json\")\n self.discriminator_model_path = os.path.join(self.model_dir, \"discriminator.json\")\n self.result_dir = os.path.join(self.data_dir, \"result\")\n self.log_dir = os.path.join(self.project_dir, \"logs\")\n self.main_log_path = os.path.join(self.log_dir, \"main.log\")\n self.config_dir = os.path.join(self.project_dir, \"config\")\n self.config_path = os.path.join(self.config_dir, \"parameter.conf\")\n \n def create_directories(self):\n dirs = [self.project_dir, self.data_dir, self.data_processed_dir, self.model_dir, self.result_dir, self.log_dir, self.config_dir]\n for dir in dirs:\n if not os.path.exists(dir):\n os.makedirs(dir)\n\n\nclass ModelConfig:\n def __init__(self):\n self.dim_in = 1\n self.dim_out = 1\n self.dim_noise = 1\n self.max_time = 1.\n self.division = 100\n self.generator_function_x_type = \"relu\"\n self.generator_function_y_type = \"relu\"\n self.discriminator_function_x_type = \"relu\"\n self.discriminator_function_y_type = \"sigmoid\"\n\n\nclass GeneratorConfig:\n def __init__(self, config: ModelConfig):\n self.dim_in = config.dim_out + config.dim_noise\n self.dim_out = config.dim_in\n self.max_time = config.max_time\n self.division = config.division\n self.function_x_type = config.generator_function_x_type\n self.function_y_type = config.generator_function_y_type\n\n\nclass DiscriminatorConfig:\n def __init__(self, config: ModelConfig):\n self.dim_in = config.dim_in\n self.dim_out = config.dim_out + 1\n self.max_time = config.max_time\n self.division = config.division\n self.function_x_type = config.discriminator_function_x_type\n self.function_y_type = config.discriminator_function_y_type\n\n\nclass TrainerConfig:\n def __init__(self):\n self.optimizer_type = \"RMSprop\"\n self.loss_type = \"MSE\"\n self.rate = 0.01\n self.momentum = 0.9\n self.decay = 0.9\n self.decay2 = 0.999\n self.regularization = 0.0001\n self.epoch = 5\n self.batch_size = 10\n self.is_visualize = True","sub_path":"src/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":7588,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"556427864","text":"from django.contrib.auth import authenticate, logout, login\nfrom django.contrib.auth.mixins import (\n LoginRequiredMixin,\n PermissionRequiredMixin\n )\nfrom django.contrib.auth.models import Group\nfrom django.core.mail import EmailMessage\nfrom django.http import Http404, HttpResponseRedirect\nfrom django.shortcuts import render\nfrom django.template.loader import get_template\nfrom django.urls import reverse, reverse_lazy\nfrom django.views import View\nfrom django.views.generic import CreateView, UpdateView\n\nfrom .forms import (\n ChangePasswordForm,\n ContactForm,\n LoginForm,\n OrderAddressForm,\n ProductAvailabilityForm,\n ProductQuantityForm,\n RegisterForm,\n )\nfrom .models import (\n Address,\n Invoice,\n Order,\n OrderLine,\n Payment,\n Product,\n ProductAvailability,\n ProductCategory,\n ShoppingCart,\n User,\n )\n\n\ndef url_response_redirect(url_name):\n url = reverse('{}'.format(url_name),)\n return HttpResponseRedirect(url)\n\n\nclass ShopView(View):\n\n def get(self, request):\n products = Product.objects.order_by('?')[:6]\n return render(request, 'shop_home.html', {'products': products})\n\n\nclass RegisterView(View):\n\n def get(self, request):\n form = RegisterForm\n return render(request, 'register.html', {'form': form})\n\n def post(self, request):\n form = RegisterForm(request.POST)\n if form.is_valid():\n username = form.cleaned_data['username']\n email = form.cleaned_data['email']\n password = form.cleaned_data['password']\n new_user = User.objects.create_user(username=username,\n email=email, password=password)\n ShoppingCart.objects.create(user=new_user)\n group = Group.objects.get(name='Buyer')\n group.user_set.add(new_user)\n user = authenticate(username=username, password=password)\n login(request, user)\n return url_response_redirect('continue_sign')\n else:\n return render(request, 'register.html', {'form': form})\n\n\nclass EditUserSignView(LoginRequiredMixin, UpdateView):\n model = User\n template_name = 'edit_user_sign.html'\n fields = ('first_name', 'last_name', 'telephone', 'date_of_birth')\n success_url = '/'\n\n def get_object(self, queryset=None):\n return self.request.user\n\n\nclass LogIn(View):\n\n def get(self, request):\n form = LoginForm\n return render(request, 'loggin.html', {'form': form})\n\n def post(self, request):\n form = LoginForm(request.POST)\n if form.is_valid():\n user = authenticate(username=form.cleaned_data['username'],\n password=form.cleaned_data['password'])\n if user is not None:\n login(request, user)\n return url_response_redirect('home_page')\n else:\n return render(\n request, 'loggin.html',\n {'form': form,\n 'field.error': 'Invalid username or password'})\n\n\nclass LogOutView(LoginRequiredMixin, View):\n def get(self, request):\n logout(request)\n return url_response_redirect('home_page')\n\n\nclass AddAddressView(LoginRequiredMixin, CreateView):\n model = Address\n template_name = 'add_address.html'\n fields = ('city', 'zip_code', 'street', 'house_no', 'flat_no')\n success_url = '/addresses'\n\n def form_valid(self, form):\n form.instance.user = self.request.user\n return super(AddAddressView, self).form_valid(form)\n\n\nclass EditAddressView(LoginRequiredMixin, UpdateView):\n model = Address\n template_name = 'add_address.html'\n fields = ('city', 'zip_code', 'street', 'house_no', 'flat_no')\n success_url = '/addresses'\n\n\nclass ShowAddressesView(LoginRequiredMixin, View):\n\n def get(self, request):\n user = request.user\n addresses = Address.objects.filter(user=user)\n return render(request, 'user_addresses.html', {'addresses': addresses})\n\n\nclass UserInfoView(LoginRequiredMixin, View):\n\n def get(self, request):\n user = request.user\n return render(request, 'user_info.html', {'user': user})\n\n\nclass UserEditView(LoginRequiredMixin, UpdateView):\n model = User\n template_name = 'edit_user_sign.html'\n fields = ('username', 'first_name', 'last_name',\n 'email', 'telephone', 'date_of_birth')\n success_url = '/user'\n\n def get_object(self, queryset=None):\n return self.request.user\n\n\nclass ChangePasswordView(LoginRequiredMixin, View):\n\n def get(self, request):\n form = ChangePasswordForm\n return render(request, 'change_password.html', {'form': form})\n\n def post(self, request):\n form = ChangePasswordForm(request.POST)\n if form.is_valid():\n user = User.objects.get(pk=request.user.id)\n if authenticate(username=user,\n password=form.cleaned_data['password_old']):\n password = form.cleaned_data['password']\n user.set_password(password)\n user.save()\n user = authenticate(username=user, password=password)\n login(request, user)\n return url_response_redirect('user_info')\n\n return render(request, 'change_password.html', {'form': form})\n\n\nclass AddProductView(LoginRequiredMixin, PermissionRequiredMixin, View):\n permission_required = 'shop.add_product'\n\n def get(self, request):\n form = ProductAvailabilityForm\n return render(request, 'add_product.html', {'form': form})\n\n def post(self, request):\n form = ProductAvailabilityForm(request.POST, request.FILES)\n if form.is_valid():\n product = form.save()\n ProductAvailability.objects.create(\n quantity=form.cleaned_data['quantity'], product=product)\n url = reverse_lazy('product_view', kwargs={'id': product.pk})\n return HttpResponseRedirect(url)\n\n\nclass ProductView(View):\n\n def get(self, request, id):\n product = Product.objects.get(pk=id)\n product_avability = ProductAvailability.objects.get(product=product)\n quantity = product_avability.quantity\n form = ProductQuantityForm\n return render(request, 'product_view.html',\n {'product': product, 'form': form, 'quantity': quantity})\n\n def post(self, request, id):\n form = ProductQuantityForm(request.POST)\n if form.is_valid():\n product = Product.objects.get(pk=id)\n shopping_cart = ShoppingCart.objects.get(user=request.user)\n orders_line = OrderLine.objects.filter(shopping_cart=shopping_cart)\n quantity = form.cleaned_data['quantity']\n price = product.price\n if product.promo:\n price = product.promo_price\n condition = False\n order_line_id = None\n for i in range(len(orders_line)):\n if (product == orders_line[i].product\n and not orders_line[i].order):\n condition = True\n order_line_id = orders_line[i].id\n if condition:\n order_line = OrderLine.objects.get(id=order_line_id)\n order_line.quantity = (\n order_line.quantity\n + form.cleaned_data['quantity'])\n\n order_line.price_quantity = (\n price * (order_line.quantity\n + form.cleaned_data['quantity']))\n order_line.save()\n else:\n OrderLine.objects.create(product=product,\n quantity=quantity,\n price_quantity=price * quantity,\n shopping_cart=shopping_cart)\n return url_response_redirect('shopping_cart')\n else:\n return Http404('Bad gateway')\n\n\nclass ChangeProductView(LoginRequiredMixin,\n PermissionRequiredMixin,\n UpdateView):\n\n permission_required = 'shop.change_product'\n form_class = ProductAvailabilityForm\n template_name = 'add_product.html'\n\n def get_object(self, queryset=None):\n return Product.objects.get(pk=self.kwargs.get('pk'))\n\n def post(self, request, pk):\n form = ProductAvailabilityForm(request.POST, request.FILES)\n if form.is_valid():\n product = Product.objects.get(pk=pk)\n product.product_name = form.cleaned_data['product_name']\n product.description = form.cleaned_data['description']\n product.price = form.cleaned_data['price']\n product.promo = form.cleaned_data['promo']\n product.promo_percent = form.cleaned_data['promo_percent']\n if form.cleaned_data['image']:\n product.image = None\n product.save()\n product.image = request.FILES['image']\n product.category = form.cleaned_data['category']\n product.save()\n available = ProductAvailability.objects.get(product=product)\n available.quantity = form.cleaned_data['quantity']\n available.save()\n url = reverse_lazy('product_view', kwargs={'id': product.id})\n return HttpResponseRedirect(url)\n\n\nclass ShowCategoryProductView(View):\n\n def get(self, request, id):\n category = ProductCategory.objects.get(pk=id)\n products = Product.objects.filter(category=category)\n return render(request, 'category.html', {'products': products})\n\n\nclass PromoView(View):\n\n def get(self, request):\n products = Product.objects.filter(promo=True)\n return render(request, 'promo_products.html', {'products': products})\n\n\nclass NewProductView(LoginRequiredMixin, View):\n\n def get(self, request):\n products = Product.objects.all().order_by('-add_date')[:6]\n return render(request, 'news_products.html', {'products': products})\n\n\nclass ContactView(View):\n\n def get(self, request):\n form = ContactForm\n return render(request, 'contact.html', {'form': form})\n\n def post(self, request):\n form = ContactForm(request.POST)\n if form.is_valid():\n contact_name = form.cleaned_data['contact_name']\n contact_email = form.cleaned_data['contact_email']\n form_content = form.cleaned_data['content']\n\n template = get_template('contact_template.txt')\n ctx = {\n 'contact_name': contact_name,\n 'contact_email': contact_email,\n 'form_content': form_content,\n }\n\n content = template.render(ctx)\n\n email = EmailMessage(\"New contact form submission\",\n content,\n \".Buy\" + \" \",\n ['dotBuy@gmail.com'],\n headers={'Reply-To': contact_email}\n )\n email.send()\n return HttpResponseRedirect('contact_us')\n\n return render(request, 'contact.html', {'form': form})\n\n\nclass ShoppingCartView(View):\n\n def get(self, request):\n if not request.user.is_authenticated():\n url = reverse('login')\n return HttpResponseRedirect(url)\n user = request.user\n shopping_cart = ShoppingCart.objects.get(user=user)\n orders_lines = (OrderLine.objects.\n filter(shopping_cart=shopping_cart).filter(order=None))\n product_sum = 0\n for order_line in orders_lines:\n product_sum += order_line.price_quantity\n return render(request,\n 'cart.html',\n {'products': orders_lines, 'sum': product_sum})\n\n\nclass RemoveProductCart(LoginRequiredMixin, PermissionRequiredMixin, View):\n\n permission_required = 'shop.delete_orderline'\n\n def get(self, request, id):\n order_line = OrderLine.objects.get(pk=id)\n if request.user == order_line.shopping_cart.user:\n order_line.delete()\n return url_response_redirect('shopping_cart')\n else:\n return Http404('Bad request')\n\n\nclass CheckoutView(View):\n\n def get(self, request):\n user = request.user\n shopping_cart = ShoppingCart.objects.get(user=user)\n orders_lines = (OrderLine.objects.\n filter(shopping_cart=shopping_cart).filter(order=None))\n product_sum = 0\n for p in orders_lines:\n if not p.order:\n product_sum += p.price_quantity\n form_order = OrderAddressForm(user=user)\n ctx = {\n 'products': orders_lines,\n 'sum': product_sum,\n 'form_order': form_order,\n }\n return render(request, 'checkout.html', ctx)\n\n def post(self, request):\n user = request.user\n form_order = OrderAddressForm(user, request.POST)\n shopping_cart = ShoppingCart.objects.get(user=user)\n orders_lines = (OrderLine.objects.\n filter(shopping_cart=shopping_cart).filter(order=None))\n if form_order.is_valid():\n product_sum = 0\n for p in orders_lines:\n product_sum += p.price_quantity\n order = Order.objects.create(\n user=user, comments=form_order.cleaned_data['comments'],\n sum_product_cost=product_sum,\n send_address_id=form_order.cleaned_data['send_address'])\n Invoice.objects.create(\n order=order,\n bill_address_id=form_order.cleaned_data['bill_address'])\n for order_line in orders_lines:\n order_line.order = order\n order_line.save()\n product = ProductAvailability.objects.get(\n product=order_line.product)\n product.quantity = product.quantity - order_line.quantity\n product.save()\n Payment.objects.create(status=True, order=order)\n return url_response_redirect('pay')\n\n\nclass PayView(View):\n\n def get(self, request):\n user = request.user\n shopping_cart = ShoppingCart.objects.get(user=user)\n order = Order.objects.filter(user=user).latest('pk')\n orders_lines = (OrderLine.objects\n .filter(shopping_cart=shopping_cart)\n .filter(order=order))\n sum_to_pay = 0\n for order_line in orders_lines:\n product = order_line.product\n buy_quantity = order_line.quantity\n availability = ProductAvailability.objects.get(product=product)\n if availability.quantity < buy_quantity:\n if product.promo:\n sum_to_pay += product.promo_price * availability.quantity\n else:\n sum_to_pay += product.price * availability.quantity\n availability.quantity = 0\n else:\n if product.promo:\n sum_to_pay += product.promo_price * buy_quantity\n else:\n sum_to_pay += product.price * buy_quantity\n\n availability.quantity = availability.quantity - buy_quantity\n availability.save()\n return render(request, 'pay_success.html', {'sum': sum_to_pay, })\n","sub_path":"shop/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":15480,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"20584384","text":"#!/usr/bin/env python\n\nfrom chdrft.tube.connection import Connection\nfrom chdrft.utils.misc import DictWithDefault, Attributize, PatternMatcher\nimport base64\nfrom pprint import pprint\nimport binascii\nimport math\nimport os\nimport random\nimport sys\nimport re\nimport traceback as tb\nfrom Crypto.Hash import SHA\nn = 128\n\n\ndef cycleLen(data, place):\n seen = {}\n count = 0\n while not place in seen:\n seen[place] = 1\n count += 1\n place = data[place]\n return count\n\n\ndef realSign(data):\n res = 1\n for i in range(len(data)):\n res *= cycleLen(data, i)\n return res\n\n\ndef get_cycle(mp, i, mx):\n seen = {}\n lst = []\n while not i in seen:\n lst.append(i)\n if i >= mx:\n return (lst, 0)\n seen[i] = 1\n i = mp[i]\n return (lst, 1)\n\n\ndef analyse(base):\n\n n = 128\n data = base + [random.randint(0, 2 * n - 1) for i in range(n)]\n\n data = {i: data[i] for i in range(len(data))}\n idata = {v: k for k, v in data.items()}\n cnt = 0\n tot = DictWithDefault(0)\n seen = {}\n for i in range(n):\n lst, typ = get_cycle(data, i, n)\n if typ == 0:\n if not lst[-2] in seen:\n seen[lst[-2]] = 1\n tot[lst[-1]] += 1\n cnt += 1\n print(len(lst))\n tb = []\n print('####')\n for i in range(n, 2 * n):\n print(tot[i])\n if tot[i] == 0:\n tb.append(i)\n print(tb, len(tb))\n print(cnt)\n\n\nclass TrueOracle:\n\n def __init__(self, key):\n self.key = key\n self.count = 350\n\n def sign(self, data):\n assert self.count > 0\n self.count -= 1\n return realSign(self.key + data)\n\n\nclass ServerOracle:\n\n def __init__(self, conn):\n self.conn = conn\n self.count = 350\n\n def sign(self, data):\n assert self.count > 0\n self.count -= 1\n self.conn.recv_until(\n PatternMatcher.frombytes(b'Give me signiture of data\\n'))\n self.conn.send('1\\n')\n data = base64.b64encode(bytes(data))\n assert len(data) == 172\n self.conn.send(data)\n res = int(self.conn.recv_until(PatternMatcher.frombytes(b'\\n')))\n\n return res\n\n def chall(self, ans):\n self.conn.recv_until(\n PatternMatcher.frombytes(b'Give me signiture of data\\n'))\n self.conn.send('2\\n')\n\n for i in range(0x11):\n print('GOGO on chall ', i)\n pattern = b'(.{172})\\n'\n res = self.conn.recv_until(PatternMatcher.fromre(pattern), timeout=1)\n c = re.search(pattern, res).group(1)\n res = ans.sign(list(base64.b64decode(c)))\n self.conn.send('%620d' % res)\n\n res = self.conn.recv_fixed_size(2048, timeout=2)\n print(self.conn.buf)\n print('RES >>> ', res)\n\n\nclass Solver:\n\n def __init__(self, oracle, n):\n self.oracle = oracle\n self.n = n\n self.to_eval = []\n\n def prepare_evals(self, data):\n self.eval_data = data\n\n def ask_eval(self, pos, values):\n self.to_eval.append([pos, values])\n\n def do_evals(self):\n n = self.n\n self.eval_data[self.static_elem - n] = self.static_elem\n eval_val = self.oracle.sign(self.eval_data)\n res = []\n for cur in self.to_eval:\n self.eval_data[self.static_elem - n] = cur[0]\n v = self.oracle.sign(self.eval_data)\n self.eval_data[self.static_elem - n] = self.static_elem\n assert v % eval_val == 0\n tmp = v // eval_val - 1\n res.append([cur[0], tmp])\n self.to_eval = []\n return res\n\n def filter_walk(self, walk, can):\n data = [0] * n\n data[self.static_elem - n] = self.static_elem\n\n prev = walk[0]\n for i in walk:\n data[i - n] = prev\n prev = i\n base_score = self.oracle.sign(data)\n\n depths = Attributize(other=lambda x: set([]))\n for i in range(n - 1):\n depths[i + 1].add(walk[i])\n vals = []\n cnt = 0\n for i in range(n):\n if len(can[i]) == 1:\n cnt += 1\n data[self.static_elem - n] = i\n cur = self.oracle.sign(data)\n assert cur % base_score == 0\n v = cur // base_score - 1\n depths[v].add(i)\n vals.append(v)\n if cnt > 0:\n print('COULd SPARE ', cnt)\n\n for i in range(n):\n can[i].intersection_update(depths[vals[i] - 1])\n if len(can[i]) == 0:\n raise \"FAIL\"\n\n def go(self):\n n = self.n\n base_data = [n + i for i in range(n)]\n base = self.oracle.sign(base_data)\n\n static_elem = None\n for i in range(n):\n base_data[i] ^= 1\n cur = self.oracle.sign(base_data)\n base_data[i] ^= 1\n v = n + i\n if base * 2 == cur:\n static_elem = v\n break\n assert static_elem\n self.static_elem = static_elem\n print('static elem >> ', static_elem)\n\n walk = []\n for i in range(n):\n if i + n != static_elem:\n walk.append(n + i)\n\n can = {}\n for i in range(n):\n can[i] = set(range(2 * n))\n\n self.filter_walk(walk, can)\n print('go reverse')\n walk.reverse()\n self.filter_walk(walk, can)\n\n pprint(can)\n print(self.oracle.count)\n s = 0\n cnd = set()\n for i in can.values():\n s += len(i) - 1\n cnd.update(i)\n\n statics = set(range(n, 2 * n))\n statics.difference_update(cnd)\n print(statics)\n roots = []\n\n self.graph = can\n self.canv = Attributize(other=lambda x: set())\n\n toproc = Attributize(other=lambda x: [])\n for i in range(n):\n if len(can[i]) == 1: continue\n print('UNDEcideD >> ', i, can[i])\n root, depth = self.rec(i)\n toproc[root].append([i, depth])\n\n print(toproc)\n while True:\n data = [n + i for i in range(n)]\n cur_round = []\n to_measure = []\n\n for root, entries in toproc._elem.items():\n if len(entries) == 0: continue\n for entry, depth in entries:\n if len(self.graph[entry]) == 1: continue\n print(entry, self.graph[entry])\n\n data[root - n] = list(self.graph[entry])[0]\n to_measure.append([entry, [depth, data[root - n]]])\n break\n\n if len(to_measure) == 0:\n break\n self.prepare_evals(data)\n\n mp = {}\n for entry, data in to_measure:\n self.ask_eval(entry, [2, 3, 434])\n mp[entry] = data\n res = self.do_evals()\n\n for e, val in res:\n depth, target = mp[e]\n print('target .. ', e, target, self.graph[e], depth, val)\n assert target in self.graph[e]\n if val == depth + 1:\n self.graph[e] = set([target])\n else:\n self.graph[e].remove(target)\n\n #print('HAVE NO W>> ', self.oracle.count)\n #for k, v in self.graph.items():\n # print(k, self.oracle.key[k], v)\n # assert self.oracle.key[k] == list(v)[0]\n #print(self.oracle.key)\n #print(self.graph)\n return [list(self.graph[i])[0] for i in range(n)]\n\n def rec(self, p):\n if p >= 128:\n return p, 0\n for i in self.graph[p]:\n root, depth = self.rec(i)\n return root, depth + 1\n assert 0\n\n\ndef pow(conn):\n res = conn.recv_fixed_size(12)\n i = random.randint(0, 2 ** 32 - 1)\n i = 2131672237762560000\n while True:\n i += 1\n i %= 2 ** 32\n x = res + ('%08x' % i).encode()\n if i % 10000 == 0:\n print(i)\n\n tmp = SHA.SHA1Hash(x).digest()\n\n if tmp.endswith(b'\\x00\\x00\\x00'):\n print(tmp, x)\n print('found it')\n conn.send(x)\n break\n pass\n\n\ndef test2():\n n = 128\n key = [random.randint(0, 2 * n - 1) for i in range(n)]\n analyse(key)\n server = 'wob-key-e1g2l93c.9447.plumbing'\n port = 9447\n\n while True:\n try:\n with Connection(server, port) as conn:\n if 1:\n pow(conn)\n oracle = ServerOracle(conn)\n solver = Solver(oracle, n)\n key = solver.go()\n\n true_oracle = TrueOracle(key)\n oracle.chall(true_oracle)\n break\n except Exception as e:\n print('failed', e)\n tb.print_exc()\n pass\n\n\ndef main():\n random.seed(1)\n test2()\n\n\nmain()\n","sub_path":"9447/wob_key_hard/solve.py","file_name":"solve.py","file_ext":"py","file_size_in_byte":7642,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"373773692","text":"import uuid\n\nfrom mock import patch, call, Mock\nfrom nose.tools import eq_, ok_, assert_not_equal, raises\n\nfrom kazoo.testing import KazooTestCase\nfrom kazoo.exceptions import KazooException\nfrom kazoo.protocol.connection import _CONNECTION_DROP\nfrom kazoo.recipe.cache import TreeCache, TreeNode, TreeEvent\n\n\nclass KazooTreeCacheTests(KazooTestCase):\n\n def setUp(self):\n super(KazooTreeCacheTests, self).setUp()\n self._event_queue = self.client.handler.queue_impl()\n self._error_queue = self.client.handler.queue_impl()\n self.path = None\n self.cache = None\n\n def tearDown(self):\n super(KazooTreeCacheTests, self).tearDown()\n if not self._error_queue.empty():\n try:\n raise self._error_queue.get()\n except FakeException:\n pass\n\n def make_cache(self):\n if self.cache is None:\n self.path = '/' + uuid.uuid4().hex\n self.cache = TreeCache(self.client, self.path)\n self.cache.listen(lambda event: self._event_queue.put(event))\n self.cache.listen_fault(lambda error: self._error_queue.put(error))\n self.cache.start()\n return self.cache\n\n def wait_cache(self, expect=None, since=None, timeout=10):\n started = since is None\n while True:\n event = self._event_queue.get(timeout=timeout)\n if started:\n if expect is not None:\n eq_(event.event_type, expect)\n return event\n if event.event_type == since:\n started = True\n if expect is None:\n return\n\n def spy_client(self, method_name):\n method = getattr(self.client, method_name)\n return patch.object(self.client, method_name, wraps=method)\n\n def test_start(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n stat = self.client.exists(self.path)\n eq_(stat.version, 0)\n\n eq_(self.cache._state, TreeCache.STATE_STARTED)\n eq_(self.cache._root._state, TreeNode.STATE_LIVE)\n\n @raises(KazooException)\n def test_start_started(self):\n self.make_cache()\n self.cache.start()\n\n @raises(KazooException)\n def test_start_closed(self):\n self.make_cache()\n self.cache.start()\n self.cache.close()\n self.cache.start()\n\n def test_close(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n self.client.create(self.path + '/foo/bar/baz', makepath=True)\n for _ in range(3):\n self.wait_cache(TreeEvent.NODE_ADDED)\n\n self.cache.close()\n\n # nothing should be published since tree closed\n ok_(self._event_queue.empty())\n\n # tree should be empty\n eq_(self.cache._root._children, {})\n eq_(self.cache._root._data, None)\n eq_(self.cache._state, TreeCache.STATE_CLOSED)\n\n # node state should not be changed\n assert_not_equal(self.cache._root._state, TreeNode.STATE_DEAD)\n\n def test_children_operation(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n self.client.create(self.path + '/test_children', b'test_children_1')\n event = self.wait_cache(TreeEvent.NODE_ADDED)\n eq_(event.event_type, TreeEvent.NODE_ADDED)\n eq_(event.event_data.path, self.path + '/test_children')\n eq_(event.event_data.data, b'test_children_1')\n eq_(event.event_data.stat.version, 0)\n\n self.client.set(self.path + '/test_children', b'test_children_2')\n event = self.wait_cache(TreeEvent.NODE_UPDATED)\n eq_(event.event_type, TreeEvent.NODE_UPDATED)\n eq_(event.event_data.path, self.path + '/test_children')\n eq_(event.event_data.data, b'test_children_2')\n eq_(event.event_data.stat.version, 1)\n\n self.client.delete(self.path + '/test_children')\n event = self.wait_cache(TreeEvent.NODE_REMOVED)\n eq_(event.event_type, TreeEvent.NODE_REMOVED)\n eq_(event.event_data.path, self.path + '/test_children')\n eq_(event.event_data.data, b'test_children_2')\n eq_(event.event_data.stat.version, 1)\n\n def test_subtree_operation(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n self.client.create(self.path + '/foo/bar/baz', makepath=True)\n for relative_path in ('/foo', '/foo/bar', '/foo/bar/baz'):\n event = self.wait_cache(TreeEvent.NODE_ADDED)\n eq_(event.event_type, TreeEvent.NODE_ADDED)\n eq_(event.event_data.path, self.path + relative_path)\n eq_(event.event_data.data, b'')\n eq_(event.event_data.stat.version, 0)\n\n self.client.delete(self.path + '/foo', recursive=True)\n for relative_path in ('/foo/bar/baz', '/foo/bar', '/foo'):\n event = self.wait_cache(TreeEvent.NODE_REMOVED)\n eq_(event.event_type, TreeEvent.NODE_REMOVED)\n eq_(event.event_data.path, self.path + relative_path)\n\n def test_get_data(self):\n cache = self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n self.client.create(self.path + '/foo/bar/baz', b'@', makepath=True)\n self.wait_cache(TreeEvent.NODE_ADDED)\n self.wait_cache(TreeEvent.NODE_ADDED)\n self.wait_cache(TreeEvent.NODE_ADDED)\n\n with patch.object(cache, '_client'): # disable any remote operation\n eq_(cache.get_data(self.path).data, b'')\n eq_(cache.get_data(self.path).stat.version, 0)\n\n eq_(cache.get_data(self.path + '/foo').data, b'')\n eq_(cache.get_data(self.path + '/foo').stat.version, 0)\n\n eq_(cache.get_data(self.path + '/foo/bar').data, b'')\n eq_(cache.get_data(self.path + '/foo/bar').stat.version, 0)\n\n eq_(cache.get_data(self.path + '/foo/bar/baz').data, b'@')\n eq_(cache.get_data(self.path + '/foo/bar/baz').stat.version, 0)\n\n def test_get_children(self):\n cache = self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n self.client.create(self.path + '/foo/bar/baz', b'@', makepath=True)\n self.wait_cache(TreeEvent.NODE_ADDED)\n self.wait_cache(TreeEvent.NODE_ADDED)\n self.wait_cache(TreeEvent.NODE_ADDED)\n\n with patch.object(cache, '_client'): # disable any remote operation\n eq_(cache.get_children(self.path + '/foo/bar/baz'), frozenset())\n eq_(cache.get_children(self.path + '/foo/bar'), frozenset(['baz']))\n eq_(cache.get_children(self.path + '/foo'), frozenset(['bar']))\n eq_(cache.get_children(self.path), frozenset(['foo']))\n\n @raises(ValueError)\n def test_get_data_out_of_tree(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n self.cache.get_data('/out_of_tree')\n\n @raises(ValueError)\n def test_get_children_out_of_tree(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n self.cache.get_children('/out_of_tree')\n\n def test_get_data_no_node(self):\n cache = self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n with patch.object(cache, '_client'): # disable any remote operation\n eq_(cache.get_data(self.path + '/non_exists'), None)\n\n def test_get_children_no_node(self):\n cache = self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n with patch.object(cache, '_client'): # disable any remote operation\n eq_(cache.get_children(self.path + '/non_exists'), None)\n\n def test_session_reconnected(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n self.client.create(self.path + '/foo')\n event = self.wait_cache(TreeEvent.NODE_ADDED)\n eq_(event.event_data.path, self.path + '/foo')\n\n with self.spy_client('get_async') as get_data:\n with self.spy_client('get_children_async') as get_children:\n # session suspended\n self.client._call(_CONNECTION_DROP, None)\n self.wait_cache(TreeEvent.CONNECTION_SUSPENDED)\n\n # There are a serial refreshing operation here. But NODE_ADDED\n # events will not be raised because the zxid of nodes are the\n # same during reconnecting.\n\n # connection restore\n self.wait_cache(TreeEvent.CONNECTION_RECONNECTED)\n\n # wait for outstanding operations\n while self.cache._outstanding_ops > 0:\n self.client.handler.sleep_func(0.1)\n\n # inspect in-memory nodes\n _node_root = self.cache._root\n _node_foo = self.cache._root._children['foo']\n\n # make sure that all nodes are refreshed\n get_data.assert_has_calls([\n call(self.path, watch=_node_root._process_watch),\n call(self.path + '/foo', watch=_node_foo._process_watch),\n ], any_order=True)\n get_children.assert_has_calls([\n call(self.path, watch=_node_root._process_watch),\n call(self.path + '/foo', watch=_node_foo._process_watch),\n ], any_order=True)\n\n def test_root_recreated(self):\n self.make_cache()\n self.wait_cache(since=TreeEvent.INITIALIZED)\n\n # remove root node\n self.client.delete(self.path)\n event = self.wait_cache(TreeEvent.NODE_REMOVED)\n eq_(event.event_type, TreeEvent.NODE_REMOVED)\n eq_(event.event_data.data, b'')\n eq_(event.event_data.path, self.path)\n eq_(event.event_data.stat.version, 0)\n\n # re-create root node\n self.client.ensure_path(self.path)\n event = self.wait_cache(TreeEvent.NODE_ADDED)\n eq_(event.event_type, TreeEvent.NODE_ADDED)\n eq_(event.event_data.data, b'')\n eq_(event.event_data.path, self.path)\n eq_(event.event_data.stat.version, 0)\n\n self.assertTrue(\n self.cache._outstanding_ops >= 0,\n 'unexpected outstanding ops %r' % self.cache._outstanding_ops)\n\n def test_exception_handler(self):\n error_value = FakeException()\n error_handler = Mock()\n\n with patch.object(TreeNode, 'on_deleted') as on_deleted:\n on_deleted.side_effect = [error_value]\n\n self.make_cache()\n self.cache.listen_fault(error_handler)\n\n self.cache.close()\n error_handler.assert_called_once_with(error_value)\n\n\nclass FakeException(Exception):\n pass\n","sub_path":"kazoo/tests/test_cache.py","file_name":"test_cache.py","file_ext":"py","file_size_in_byte":10671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"566965355","text":"from django.core.exceptions import ImproperlyConfigured\n\nfrom common.classes import BaseEnumerate\n\n\nclass SeleniumProgramEnum(BaseEnumerate):\n Chrome = 0\n Firefox = 1\n\n values = {\n Chrome: 'Chrome',\n Firefox: 'Firefox',\n }\n\n @classmethod\n def get_program_type(cls, string_type):\n for program_type, str_type in cls.values.items():\n if str_type == string_type:\n return program_type\n raise ImproperlyConfigured(\n f'Указан некорректный тип парсера Selenium - {string_type}')\n","sub_path":"project/parsing/parsers/enum.py","file_name":"enum.py","file_ext":"py","file_size_in_byte":582,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"427859474","text":"# Django settings for foodem project.\n\nfrom settings import *\n\n# Local time zone for this installation. Choices can be found here:\n# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name\n# although not all choices may be available on all operating systems.\n# On Unix systems, a value of None will cause Django to use the same\n# timezone as the operating system.\n# If running in a Windows environment this must be set to the same as your\n# system time zone.\nTIME_ZONE = 'America/New_York'\nUSE_TZ = True\n#USE_TZ = False # set when creating / running migrations to prevent UTC errors on datetime fields\n\n\n# Language code for this installation. All choices can be found here:\n# http://www.i18nguy.com/unicode/language-identifiers.html\nLANGUAGE_CODE = 'en-us'\n\nSITE_ID = 1\n\n# If you set this to False, Django will make some optimizations so as not\n# to load the internationalization machinery.\nUSE_I18N = True\n\n# If you set this to False, Django will not format dates, numbers and\n# calendars according to the current locale\nUSE_L10N = True\n\n# Absolute filesystem path to the directory that will hold user-uploaded files.\n# Example: \"/home/media/media.lawrence.com/media/\"\nMEDIA_SUBROOT = 'site_media/'\nMEDIA_ROOT = SERVER_ROOT + 'public/' + MEDIA_SUBROOT\n\n# Add a custom path for the http://localhost/_ah/channel/jsapi\n# (Google App Engine Channel app) when running on localhost\nAH_ROOT = SERVER_ROOT + 'public/' + MEDIA_SUBROOT + '_ah'\n\n# URL that handles the media served from MEDIA_ROOT. Make sure to use a\n# trailing slash.\n# Examples: \"http://media.lawrence.com/media/\", \"http://example.com/media/\"\nMEDIA_URL = SERVER_URL + 'site_media/'\n\n# URL prefix for static files.\n# Example: \"http://media.lawrence.com/static/\"\nSTATIC_URL = '/media/'\n\n# URL prefix for admin static files -- CSS, JavaScript and images.\n# Make sure to use a trailing slash.\n# Examples: \"http://foo.com/static/admin/\", \"/static/admin/\".\nADMIN_MEDIA_PREFIX = '/media/'\n\n# Additional locations of static files\nSTATICFILES_DIRS = (\n SERVER_ROOT + 'public/site_media/',\n)\n\n# List of finder classes that know how to find static files in\n# various locations.\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n# 'django.contrib.staticfiles.finders.DefaultStorageFinder',\n)\n\nMIDDLEWARE_CLASSES = (\n 'django.middleware.common.CommonMiddleware',\n# 'django.contrib.sessions.middleware.SessionMiddleware',\n 'auth.authenticate.SessionMiddlewareOverride',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'auth.authenticate.SessionProcessorMiddleware',\n)\nSESSION_ENGINE=\"django.contrib.sessions.backends.cache\"\nAPPEND_SLASH = True\n\nROOT_URLCONF = 'foodem.urls'\n\nTEMPLATE_DIRS = (\n SERVER_ROOT + 'reporting/templates/reporting/',\n SERVER_ROOT + 'terms/templates/terms/',\n SERVER_ROOT + 'auth/templates/auth/',\n SERVER_ROOT + 'search/templates/search/',\n SERVER_ROOT + 'order/templates/order/',\n SERVER_ROOT + 'catalog/templates/catalog/',\n SERVER_ROOT + 'vendor/templates/vendor/',\n SERVER_ROOT + 'dashboard/templates/dashboard/',\n SERVER_ROOT + 'dashboard/templates/dashboard/tiles/',\n SERVER_ROOT + 'templates/',\n SERVER_ROOT + 'templates/emails/',\n SERVER_ROOT + 'templates/emails/ratings',\n)\n\nAUTHENTICATION_BACKENDS = ('auth.authenticate.MyBackend',)\n# (django.contrib.auth.backends.ModelBackend',) # Default\n\nINSTALLED_APPS = (\n\n 'auth',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.sites',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n 'django.contrib.admin', # Uncomment to enable the admin\n 'django.contrib.admindocs', # Uncomment to enable admin documentation\n 'django.contrib.humanize',\n\n 'captcha',\n 'decorators',\n 'geopy',\n 'south',\n\n 'catalog',\n 'eventlog',\n 'order',\n 'quicklink',\n 'reporting',\n 'search',\n 'terms',\n 'vendor',\n)\n\nAUTH_PROFILE_MODULE = 'auth.UserProfile'\n\n# A sample logging configuration. The only tangible logging\n# performed by this configuration is to send an email to\n# the site admins on every HTTP 500 error.\n# See http://docs.djangoproject.com/en/dev/topics/logging for\n# more details on how to customize your logging configuration.\nLOGGING = {\n 'version': 1,\n 'disable_existing_loggers': False,\n 'filters': {\n 'require_debug_false': {\n '()': 'django.utils.log.RequireDebugFalse'\n }\n },\n 'handlers': {\n 'mail_admins': {\n 'level': 'ERROR',\n 'filters': ['require_debug_false'],\n 'class': 'django.utils.log.AdminEmailHandler'\n },\n\t'django.request': { # define and name a second handler\n 'level': 'DEBUG',\n 'class': 'logging.FileHandler', # set the logging class to log to a file\n 'filename': LOG_DIR + 'foodem.log' # '/home/foodem/logs/user/webapp.log', # log file\n },\n },\n 'loggers': {\n 'django.request': {\n 'handlers': ['mail_admins','django.request'],\n 'level': 'DEBUG',\n 'propagate': True,\n },\n }\n}\n\nTEMPLATE_CONTEXT_PROCESSORS = (\n 'auth.views.includeuser',\n 'auth.views.includebusinesstypes',\n 'auth.views.includesettings',\n 'order.views.pendingactions',\n 'terms.views.pendingactions',\n 'catalog.views.badlistings',\n 'order.cart.includecart',\n \"django.contrib.auth.context_processors.auth\",\n \"django.core.context_processors.debug\",\n \"django.core.context_processors.i18n\",\n \"django.core.context_processors.media\",\n \"django.core.context_processors.static\",\n \"django.core.context_processors.request\",\n \"django.contrib.messages.context_processors.messages\"\n)\n\nCACHES = {\n 'default': {\n 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache',\n 'LOCATION': [\n 'unix:' + MEMCACHED_SOCKET_DIR + 'memcached.sock',\n '127.0.0.1:11211',\n ]\n },\n}\n\nLOGIN_REDIRECT_URL=\"/home/\"\n\nEFT_MERCHANT_ID = 149279\nEFT_MERCHANT_PWD = '666eXSvM8gI'\nEFT_GATEWAY_URL = \"https://www.paymentsgateway.net/cgi-bin/posttest.pl\"\n\nSTRIPE_SECRET_KEY='sk_test_t1O2uYwIKIeMasvGti1lThUZ'\nSTRIPE_PUBLISH_KEY='pk_test_ugQQnOH1ml9cChXDLGoE6pwc'\n\nBING_MAPS_API_KEY='AvJDaVXzLY8jMdace0M5QKhzsUkLjrsJXmdflfKs7cCnjBqes2iBh5P8pIDfv6yt'\n\n# Maximum Number of Search Results to return\nSEARCH_MAX_RESULTS = 50\n\nSOLR_UPDATE = SOLR_INDEX + 'update?commitWithin=10000'\nSOLR_VARS = [\n ('indent', 'on'),\n ('version', '3.5'),\n ('q', '{terms}'),\n ('defType', 'edismax'),\n ('qf', 'productname^45.0+uniqueproductid^45.0+text^10.0'),\n ('fq', '{fq}'),\n ('start', '0'),\n ('rows', SEARCH_MAX_RESULTS),\n ('fl', 'id'),\n ('qt', ''),\n ('wt', ''),\n ('explainOther', 'hl.fl=')\n]\nSOLR_SEARCH_URL = SOLR_INDEX + 'select?' + '&'.join([\"%s=%s\" % (a[0], a[1]) for a in SOLR_VARS])\nSOLR_AUTOCOMPLETE_URL = SOLR_INDEX + 'terms/?terms=true&terms.fl=text&terms.lower=%s&terms.prefix=%s&terms.lower.incl=false'\n\nNEWSLETTER_POST_URL = 'http://foodem.us2.list-manage1.com/subscribe/post?u=c935af176b9c83356b8391c0d&id=277d40c8a7'\n\n# To disable storing carts in Session variables, set this to True\nDISABLE_SESSION_CARTS = True\n\n# Vendor Catalog number of Items per Page\nVENDOR_CATALOG_ITEMS_PER_PAGE = 50\n# How many listings per item to show on public search\nANONYMOUS_USER_MAX_SEARCH_LISTINGS=3\n# How many listings per item to show on vendor store\nANONYMOUS_USER_MAX_STORE_LISTINGS=5\n\n# Default distance to on Search Page, in miles\nDEFAULT_SEARCH_DISTANCE=200\n\n# Lowest number for the distance slider for it to become \"Any Distance\"\n# (so clicking on absolute rightmost point is not required)\nMAX_DISTANCE_FOR_ALL=190\n\n# What number to show when cleaning numeric data for public/anonymous result display\nANONYMOUS_NUMBER_TO_DISPLAY=88.88\n# What number to show when cleaning string data for public/anonymous result display\nANONYMOUS_STRING_TO_DISPLAY='[hidden]'\n\n# List of characters usable in creating Order UniqueIDs (removes confusing 1/l/I, 0/o chars,\n# and vowels to prevent accidental random offensive words).\n# This leaves us with 28 characters which makes 28^7 possibilities (13,492,928,512)\nORDER_UNIQUE_ID_ALPHABET='23456789bcdfghjkmnpqrstvwxyz'\nORDER_UNIQUE_ID_LENGTH=7\n\n# Ratings & Comments settings\nMIN_RATING = 0 # no rating\nMAX_RATING = 5\n# What value indicates that no rating has been manually made and will assign the DEFAULT_RATING_VALUE as the rating?\nDEFAULT_RATING_FLAG = -1\n# If no rating is specified, use this default value\nDEFAULT_RATING_VALUE = 5\n# Number of days after delivery date that a default rating will be assigned to an order and its entries.\n# The default rating can be manually overridden until the MAX_RATING_PERIOD expires.\nDEFAULT_RATING_PERIOD = 5\n# Maximum number of days after order.delivery_date that an order can be rated\nMAX_RATING_PERIOD = 7\n# How long the average Shipping order will take, in days (use this at least until the new shipping module is integrated)\nDEFAULT_SHIPPING_DELAY = 3\n# Maximum of days to allow a shipment to be received from date_confirmed, if no delivery_date is set\nRATING_MAX_SHIP_TIME = 45\n# Minimum number of ratings required to show a rating\nMIN_RATINGS_REQUIRED = 5\n# Max number of orders to update at a time\nMAX_RATINGS_UPDATE_BATCH = 100\n# message to be displayed where objectional content has been removed by moderators\nMODERATOR_REMOVED_CONTENT_NOTICE='[removed by moderator]'\n# Send out emails when orders are auto-rated?\nSEND_AUTO_RATING_EMAILS = False\n\n# Max number of Autosuggest listings to show on Search bar\nMAX_AUTOSUGGEST_LISTINGS = 7\n\n# default reporting period, in days (for when no start_date is provided)\nDEFAULT_REPORT_PERIOD = 30\n\n# Set timeout period for bubbles to redisplay on page, in ms (5 minutes)\nBUBBLE_REFRESH_TIMEOUT = 5 * 60 * 1000\n\n# Distance calculation vars\nMAPQUEST_API_DIRECTIONS_URL=\"http://open.mapquestapi.com/directions/v2/route\"\nMAPQUEST_API_KEY=\"Fmjtd%7Cluubn96tn9%2Crs%3Do5-907al6\"\n","sub_path":"settings_foodem.py","file_name":"settings_foodem.py","file_ext":"py","file_size_in_byte":10298,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"221465827","text":"import setuptools\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetuptools.setup(\n name=\"nothing\", # Replace with your own username\n version=\"0.0.2\",\n author=\"Matt Derasoirs\",\n author_email=\"nothing@razorbla.de\",\n scripts=[\"bin/whatever-okay\"],\n description=\"package about nothing\",\n url=\"http://razorbla.de\",\n packages=setuptools.find_packages(),\n python_requires='>=3.7',\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":427,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"68788128","text":"import numpy as np\nimport theano\nimport os\nfrom dml import *\nimport common as cc\nfrom common import *\n\nIMG_SHAPE = (50, 50)\n# IMG_COLOR_SHAPE = (6,) + IMG_SHAPE\n# IMG_COLOR_SHAPE = IMG_SHAPE\ncc.NB_CLASSES = 6\ncc.CLASS_FOLD = [\"01\", \"03\", \"04\", \"08\", \"09\", \"10\"]\ncc.CLASS_NAME = [\"First fish(C1)\", \"Black fish(C2)\", \"Clown fish(C3)\", \"C4\", \"C5\", \"C6\"]\n\ncc.preprocess = ImagePreprocess(newShape=IMG_SHAPE, grayscale=False)\n\ndef getColorDensity(example):\n\treturn np.mean(example, axis=tuple(range(1, example.ndim)))\n\ndef getDataSetDensities(datas):\n\tclasses = [[] for _ in range(cc.NB_CLASSES)]\n\tnbExamples = len(datas[0][0])\n\tfor i in range(nbExamples):\n\t\tclasses[datas[1][0][i].argmax()].append(getColorDensity(datas[0][0][i]))\n\n\treturn np.array([np.mean(l, axis=0) for l in classes])\n\ndef distFct(a, b):\n\treturn np.sum((a-b)**2)\n\ndef predictClass(example, colors):\n\td = [0] * cc.NB_CLASSES\n\tfor c in range(cc.NB_CLASSES):\n\t\td[c] = distFct(getColorDensity(example), colors[c])\n\treturn np.array(d).argmin()\n\ndef predictDataSet(datas, colors):\n\tmetrics = OneClassMetrics(cc.NB_CLASSES)\n\tnbExamples = len(datas[0][0])\n\tfor i in range(nbExamples):\n\t\tmetrics.addResult(\n\t\t\tdatas[1][0][i].argmax(),\n\t\t\tpredictClass(datas[0][0][i], colors)\n\t\t)\n\treturn metrics\n\ndef main():\n\tquickTest = False\n\n\tprint(\"Read datas...\")\n\tif quickTest:\n\t\tvalidationDatas = getDataSet(\"validation\")\n\t\ttestDatas, trainingDatas = validationDatas, validationDatas\n\telse:\n\t\ttrainingDatas = getDataSet(\"train\")\n\t\tvalidationDatas = getDataSet(\"validation\")\n\t\ttestDatas = getDataSet(\"test\")\n\n\tprint(\"Compute colors...\")\n\tcolors = getDataSetDensities(trainingDatas)\n\tprint(\"Colors:\", colors)\n\n\t# print(\"Predict results\")\n\t# printMetricResults(predictDataSet(validationDatas, colors), \"validation\")\n\tprintMetricResults(predictDataSet(testDatas, colors), \"test\")\n\nif __name__ == '__main__':\n\tmain()","sub_path":"code/nnets/fishesClass/colors.py","file_name":"colors.py","file_ext":"py","file_size_in_byte":1858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"441735012","text":"from graph import Graph\n\nclass SuperSkills(object):\n def __init__(self):\n self.skills = Graph()\n self.friends = Graph()\n\n def new_skills_set_for_name(self, name):\n res = set()\n for friend in self.friends[name]:\n res |= self.skills[friend]\n res = res - self.skills[name]\n return res\n\n def choose_max_uncovered(self, skillset, friends, valuable_skills):\n best_friend = None\n best_friend_skills_len = 0\n for friend in friends:\n x = valuable_skills[friend] - skillset\n if (len(x) > best_friend_skills_len):\n best_friend = friend\n best_friend_skills_len = len(x)\n return best_friend\n\n def minimal_set_of_friends_with_skills(self, root_name):\n required_skills = self.new_skills_set_for_name(root_name)\n valuable_skills = {}\n root_friends = set()\n for friend in self.friends[root_name]:\n x = (self.skills[friend] & required_skills)\n if x:\n valuable_skills[friend] = x\n root_friends.add(friend)\n\n # greedy algorithm. Is this good enough?\n best_friends = set()\n current_skillset = set()\n while (required_skills - current_skillset):\n best_friend = self.choose_max_uncovered(current_skillset, root_friends, valuable_skills)\n best_friends.add(best_friend)\n current_skillset |= (valuable_skills[best_friend])\n\n return sorted(best_friends)\n\ndef solution(f):\n ss = SuperSkills()\n\n n = int(f.readline().strip())\n for i in range(n):\n (a,b) = f.readline().strip().split()\n ss.skills.add_edge(a, b)\n\n n = int(f.readline().strip())\n for i in range(n):\n (a,b) = f.readline().strip().split()\n ss.friends.add_edge_bi(a, b)\n\n root_name = f.readline().strip()\n\n res = \",\".join(ss.minimal_set_of_friends_with_skills(root_name))\n print (res)\n return res\n\n\nimport sys\n#solution(sys.stdin)\n\n\n\n\n","sub_path":"graph_superskills.py","file_name":"graph_superskills.py","file_ext":"py","file_size_in_byte":2014,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"155574944","text":"from pymongo import MongoClient\n\n# client = MongoClient(host=\"localhost\", port=27017)\n# client.admin.authenticate(\"admin\", \"123456\")\nmyclient = MongoClient(\"mongodb://127.0.0.1:27017\")\nmyclient.admin.authenticate(\"admin\", \"123456\")\nmydb = myclient['imooc']\nmycollection = mydb['pymongo_test']\nmycollection.insert_one({\"name\": \"imooc\", \"flag\": 1, \"url\": \"http://www.baidu.com\"})\n\n# result = mycollection.find({}, {'_id': 0, \"name\": 1, \"flag\": 1})\n# result = mycollection.find({\"name\": {\"$regex\": \"^G\"}})\nmycollection.delete_many({\"url\": {\"$regex\": \"https?://www.\\.[tq]\"}})","sub_path":"requestDemo/study_pymongo.py","file_name":"study_pymongo.py","file_ext":"py","file_size_in_byte":571,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"102717010","text":"import os\nimport cv2\nfrom pydicom import dcmread\nimport matplotlib.pyplot as plt\nfrom PyQt5 import QtWidgets\n\n\"\"\"\nfunctions used in the dicom-editor/file loader\n\"\"\"\n\n\ndef png2avi(path: str, fps: int) -> None:\n \"\"\"Create a list of the PNG's in path, use cv2 videowriter to make it into a movie.\"\"\"\n filelist = os.listdir(path)\n filelist.sort()\n img_array = []\n size = (0, 0)\n\n for element in filelist:\n # print(element)\n fp = path + element\n img = cv2.imread(fp)\n h, w, trash = img.shape\n # notice the reversal of order ...\n size = (w, h)\n img_array.append(img)\n\n # check if list is nonempty\n # save location is still wrong!\n if filelist:\n out = cv2.VideoWriter('video.avi', cv2.VideoWriter_fourcc(*'FFV1'), fps, size)\n for i in range(len(img_array)):\n out.write(img_array[i])\n out.release()\n # fourcc: 4 bytes to identify videostreams.\n return\n\n\ndef dicom2png(filelist: list, path: str, project_name: str) -> int:\n \"\"\"\"extracts the png part out of the dicom images.\n File should start with 'IM_' \"\"\"\n a = 0\n for element in filelist:\n a = a + 1\n # disregard non-dicom files\n if element[0:3] != 'IM_':\n continue\n\n # read file and put it in a use-able array\n string = path + element\n dicom = dcmread(string)\n array = dicom.pixel_array\n plt.imshow(array, cmap=\"gray\")\n savestring = \"./data/png/\" + project_name + \"/\" + element + \".png\"\n plt.savefig(savestring)\n\n return a\n\n\ndef checkifpng(filelist: list) -> int:\n # count how many pngs are in the filelist.\n a = 0\n for element in filelist:\n if \".png\" in element:\n a += 1\n return a\n\n\ndef popupmsg(text: str, iswhat: str):\n \"\"\"\"create a popup message. Can be generalized to do more than warnings\n currently supports only warning\"\"\"\n msg = QtWidgets.QMessageBox()\n msg.setText(text)\n if iswhat == \"warning\":\n msg.setIcon(QtWidgets.QMessageBox.Warning)\n msg.exec_()\n return\n\n\ndef loadin(filelist: list, path: str, size: list) -> list:\n # load grayscale png from list, given path.\n path = path + \"/\"\n imlist = []\n for element in filelist:\n # cp: current path\n cp = path + element\n # 0 indicates grayscale\n im = cv2.imread(cp, 0)\n # resize happens here\n im = im[size[0]:size[1], size[2]:size[3]]\n # im = im[58:428, 143:513]\n imlist.append(im)\n return imlist\n","sub_path":"functions/auxiliary.py","file_name":"auxiliary.py","file_ext":"py","file_size_in_byte":2537,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"173712892","text":"\nimport datetime\nfrom time import ctime\nimport os\nimport ntplib\n\n\nservidor_de_tiempo = \"pool.ntp.org\"\n\nprint(\"\\nObteniendo la hora del servidor NTP:\")\ncliente_ntp = ntplib.NTPClient()\nrespuesta = cliente_ntp.request(servidor_de_tiempo)\nprint(respuesta.tx_time)\nhora_actual = datetime.datetime.strptime(ctime(respuesta.tx_time), \"%a %b %d %H:%M:%S %Y\")\n\nprint(\"Respuesta de \" + servidor_de_tiempo + \": \" + str(hora_actual) + \"\\n\")\n\n\nseparador = \" \"\nsep = str(hora_actual).split(separador)\n\nfecha = sep[0]\nhora = sep[1]\n\nfechass = fecha.split(\"-\")\nanio = fechass[0]\nmes = fechass[1]\ndia = fechass[2]\n\nhorasrr = hora.split(\":\")\n\nfhora = horasrr[0]\nfmin = horasrr[1]\n\nhola = 'date -u ' + mes + dia + fhora + fmin + anio\n\nprint(hola)\nos.system(hola)","sub_path":"hora.py","file_name":"hora.py","file_ext":"py","file_size_in_byte":745,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"80803190","text":"'''\n35. Write a Pandas program to count the NaN values in one or more columns in DataFrame.\n\n'''\n\nimport pandas as pd\nimport numpy as np\n\nexam_data = {'name': ['Anastasia', 'Dima', 'Katherine', 'James', 'Emily', 'Michael', 'Matthew', 'Laura', 'Kevin', 'Jonas'],\n 'score': [12.5, 9, 16.5, np.nan, 9, 20, 14.5, np.nan, 8, 19],\n 'attempts': [1, 3, 2, 3, 2, 3, 1, 1, 2, 1],\n 'qualify': ['yes', 'no', 'yes', 'no', 'no', 'yes', 'yes', 'no', 'no', 'yes']}\n\ndf = pd.DataFrame(exam_data)\n\nprint(df.isnull().values.sum())\n\n","sub_path":"Pandas/PandasDataframe/Exercise35.py","file_name":"Exercise35.py","file_ext":"py","file_size_in_byte":534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"144769329","text":"import socket\r\n\r\nip_port = ('127.0.0.遗传算法', 8009)\r\nsk = socket.socket()\r\nsk.connect(ip_port)\r\nsk.settimeout(5)\r\n\r\nwhile True:\r\n data = sk.recv(1024)\r\n print('receive:', data.decode())\r\n inp = input('please input:')\r\n sk.sendall(inp.encode())\r\n if inp == 'exit':\r\n break\r\n\r\nsk.close()","sub_path":"计算机网络/多线程客户端.py","file_name":"多线程客户端.py","file_ext":"py","file_size_in_byte":313,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"162285325","text":"import xml.etree.ElementTree as ET\nimport os\n\n# https://docs.python.org/3/library/xml.etree.elementtree.html\n\nif __name__ == '__main__':\n PATH_TO_USB_FRAMES = '/home/leander/Desktop/TensorFlow/workspace/training_kitchen_tools/images/test/'\n files = os.listdir(PATH_TO_USB_FRAMES)\n classes = []\n occurrence = []\n file_counter = 0\n label_counter = 0\n bad_file = 0\n\n for file in files:\n if '.xml' in file:\n try:\n tree = ET.parse(PATH_TO_USB_FRAMES + file)\n file_counter += 1\n # print(\"File is well-formed: \", file)\n root = tree.getroot()\n for child in root:\n if child.tag == 'object':\n for more_child in child:\n if more_child.tag == 'name':\n label_counter += 1\n if more_child.text not in classes:\n classes.append(more_child.text)\n occurrence.append(1)\n else:\n i = classes.index(more_child.text)\n occurrence[i] += 1\n except Exception:\n bad_file += 1\n print(\"FILE IS BAD: \", file)\n pass\n\n print(\"GOOD FILES: \", file_counter)\n print(\"BAD FILES: \", bad_file)\n print(\"TOTAL LABELS: \", label_counter)\n\n dic_array = {}\n for clas, count in zip(classes, occurrence):\n dic_array[clas] = count\n\n sorted_dic = sorted(dic_array.items(), key=lambda kv: (kv[1], kv[0]), reverse=True)\n\n i = 1\n for elem in sorted_dic:\n # if int(elem[1]) < 400:\n print(i, elem)\n i += 1\n","sub_path":"computer_vision/CV_Kitchen_Tools/training/configure_dataset/analyze_dataset.py","file_name":"analyze_dataset.py","file_ext":"py","file_size_in_byte":1784,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"507650203","text":"#pylint: disable=R0903\nfrom sqlalchemy import Column, DateTime, Numeric, String\nimport sqlalchemy\nimport sqlalchemy.ext.hybrid\nimport pyrob.table\n\n\nclass SamMediaAdptRaw(pyrob.table.Common, pyrob.table.Base):\n __tablename__ = 'T_ROB_SAM_MEDIA_ADPT_RAW'\n __table_args__ = {'schema': 'IPACT_STG'}\n\n email_check_rec_id = Column(\n Numeric(10, 0, asdecimal=False), nullable=False)\n site_id_ip = Column(String(30), primary_key=True, nullable=False)\n site_name = Column(String(50), primary_key=True, nullable=False)\n cli_name = Column(String(30), primary_key=True, nullable=False)\n connector_type = Column(String(20))\n connector_code = Column(String(20))\n transceiver_type = Column(String(30))\n sfp_optical_compliance = Column(String(50))\n model_number_media_adpt = Column(String(50))\n model_name_media_adpt = Column(String(50))\n serial_number_media_adpt = Column(String(50))\n vendor_oui = Column(String(20))\n part_number_media_adpt = Column(String(50))\n manufacture_date = Column(String(30))\n supported_media = Column(String(20))\n laser_wavelength_nm = Column(String(20))\n diagnostic_capable = Column(String(20))\n description = Column(String(100))\n link_length_m = Column(String(20))\n laser_tunability = Column(String(30))\n number_of_lanes = Column(Numeric(asdecimal=False))\n clei_code = Column(String(50))\n current_olc_state = Column(String(50))\n alarm_status = Column(String(50))\n aggregated_alarm_status = Column(String(50))\n is_integrating = Column(Numeric(asdecimal=False))\n\n @sqlalchemy.ext.hybrid.hybrid_property\n def column_aliases(self):\n return {\n \"portName\": \"cli_name\",\n \"specificType\": \"transceiver_type\",\n \"connectorType\": \"connector_type\",\n \"connectorCode\": \"connector_code\",\n \"sfpOpticalCompliance\": \"sfp_optical_compliance\",\n \"laserWaveLength\": \"laser_wavelength_nm\",\n \"diagnosticsCapable\": \"diagnostic_capable\",\n \"modelNumber\": \"model_number_media_adpt\",\n \"vendorOUI\": \"vendor_oui\",\n \"vendorManufactureDate\": \"manufacture_date\",\n \"supportedMedia\": \"supported_media\",\n \"vendorSerialNumber\": \"serial_number_media_adpt\",\n \"vendorPartNumber\": \"part_number_media_adpt\",\n \"description\": \"description\",\n \"mprLinkLength\": \"link_length_m\",\n \"laserTunability\": \"laser_tunability\",\n \"portSFPNumLanes\": \"number_of_lanes\",\n \"siteId\": \"site_id_ip\",\n \"siteName\": \"site_name\",\n \"hwName\": \"model_name_media_adpt\",\n \"cleiCode\": \"clei_code\",\n \"olcState\": \"current_olc_state\",\n }\n\n @sqlalchemy.ext.hybrid.hybrid_property\n def group_id(self):\n return 30\n\n @sqlalchemy.ext.hybrid.hybrid_property\n def object_id(self):\n return 744\n\n @classmethod\n def filter_rows(cls, rows):\n for row in rows:\n if cls.validate_nullable(row):\n yield row\n\n @sqlalchemy.ext.hybrid.hybrid_property\n def defaults(self):\n return {\n 'is_integrating': 0,\n 'email_check_rec_id': 0\n }\n\n @staticmethod\n def request_parameters():\n return {\n \"stream\": True,\n \"tags\": [\n \"equipment.MediaAdaptor\"\n ],\n \"full_class_name\": \"equipment.MediaAdaptor\",\n \"result_filter\": {\n \"attributes\": [\n \"siteId\",\n \"siteName\",\n \"portName\",\n \"connectorType\",\n \"connectorCode\",\n \"specificType\",\n \"sfpOpticalCompliance\",\n \"modelNumber\",\n \"hwName\",\n \"vendorSerialNumber\",\n \"vendorOUI\",\n \"vendorPartNumber\",\n \"vendorManufactureDate\",\n \"supportedMedia\",\n \"laserWaveLength\",\n \"diagnosticsCapable\",\n \"description\",\n \"mprLinkLength\",\n \"laserTunability\",\n \"portSFPNumLanes\",\n \"cleiCode\",\n \"olcState\",\n ]\n }\n }\n","sub_path":"pyrob/schema/ipact_stg/sam/sam_media_adpt.py","file_name":"sam_media_adpt.py","file_ext":"py","file_size_in_byte":4347,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"243182077","text":"# -*- coding: utf-8 -*-\n# @Time : 2019/10/16 \n# @Author : Edrain\n\"\"\"\n使用Python中的json模块就可以将字典或列表以JSON格式保存到文件中\njson模块主要有四个比较重要的函数,分别是:\ndump - 将Python对象按照JSON格式序列化到文件中\ndumps - 将Python对象处理成JSON格式的字符串\nload - 将文件中的JSON数据反序列化成对象\nloads - 将字符串的内容反序列化成Python对象\n这里出现了两个概念,一个叫序列化,一个叫反序列化。\n自由的百科全书维基百科上对这两个概念是这样解释的:\n“序列化(serialization)在计算机科学的数据处理中,是指将数据结构或对象状态转换为可以存储或传输的形式,\n这样在需要的时候能够恢复到原先的状态,而且通过序列化的数据重新获取字节时,可以利用这些字节来产生原始对象的副本(拷贝)。\n与这个过程相反的动作,即从一系列字节中提取数据结构的操作,就是反序列化(deserialization)”。\n\"\"\"\nimport json\n\n\ndef main():\n my_dict = {\n 'name': '哇哈哈',\n 'age': 88,\n 'qq': 666888,\n 'friends': ['哇有才', '哈有福'],\n 'cars': [\n {'brand': 'Fen', 'max_speed': 180},\n {'brand': 'Audi', 'max_speed': 280},\n {'brand': 'Benz', 'max_speed': 320}\n ]\n }\n try:\n with open('data.json', 'w', encoding='utf-8') as fs:\n json.dump(my_dict, fs)\n except IOError as e:\n print(e)\n print('保存数据完成!')\n\n\nif __name__ == '__main__':\n main()","sub_path":"code_ed/Day01-15/code/Day11/json_data01.py","file_name":"json_data01.py","file_ext":"py","file_size_in_byte":1624,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"159687263","text":"import argparse\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"first_trec\")\n parser.add_argument(\"second_trec\")\n parser.add_argument(\"output_file\", type=argparse.FileType('w'))\n args = parser.parse_args()\n\n # read first\n trec = {}\n for line in open(args.first_trec):\n items = line.split('\\t')\n qid = int(items[0])\n if qid not in trec:\n trec[qid] = {}\n trec[qid].append(line)\n\n # read second\n for line in open(args.second_trec):\n items = line.split('\\t')\n qid = int(items[0])\n if qid not in trec:\n continue\n trec[qid].append(line)\n\n # merge\n for qid in trec:\n for line in trec[qid]:\n args.out_file.write(line)\n\n","sub_path":"gen_train_data/merge_trec.py","file_name":"merge_trec.py","file_ext":"py","file_size_in_byte":781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"302358813","text":"\nimport heapq\n\n\ndef path_length(adjm, source, dest):\n costs_table = dict()\n for row_idx, row in enumerate(adjm):\n for column_idx, cost in enumerate(row):\n if cost != 0:\n _ = costs_table.get(row_idx, list())\n _.append([cost, column_idx])\n costs_table[row_idx] = _\n\n min_cost_from_source = {source: 0}\n q = [[0, source, []]]\n seen = set()\n\n while q:\n cost_s_v, v, path = heapq.heappop(q)\n if v == dest:\n print(path)\n return cost_s_v\n if v in seen:\n continue\n seen.add(v)\n path.append(v)\n\n for cost_v_u, u in costs_table[v]:\n if u in seen:\n continue\n cost_s_u = min_cost_from_source.get(u, None)\n cost_s_u_relaxed = cost_s_v + cost_v_u\n if cost_s_u is None or cost_s_u_relaxed < cost_s_u:\n min_cost_from_source[u] = cost_s_u_relaxed\n heapq.heappush(q, [cost_s_u_relaxed, u, path])\n \n return -1\n","sub_path":"graph/dijkstra/v4.py","file_name":"v4.py","file_ext":"py","file_size_in_byte":1042,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"433580268","text":"import torch\nfrom torch import nn\nfrom torch import optim\nfrom torchvision import models\nfrom collections import OrderedDict\nimport PIL\nimport numpy as np\nimport json\nfrom workspace_utils import active_session\n\n\nclass Model:\n \"\"\" Model class \"\"\"\n \n def __init__(self, train_loader, valid_loader, test_loader, model_pretrained, hidden_units, epochs, learning_rate, save_dir, gpu):\n \"\"\" Initialization of the class \"\"\"\n self.train_loader = train_loader\n self.valid_loader = valid_loader\n self.test_loader = test_loader\n self.model_pretrained = model_pretrained\n self.hidden_units = hidden_units\n self.epochs = epochs\n self.learning_rate = learning_rate\n self.save_dir = save_dir\n self.gpu = gpu\n self.model = None\n self.criterion = None\n self.optimizer = None\n self.input_size = None\n self.cat_to_name = None\n \n def load_pretrained(self, trained=True):\n \"\"\" Load a pretrained model and replace the last layer of the model with a desired number of hidden units\"\"\"\n if self.model_pretrained==\"vgg13\":\n self.model = models.vgg13(pretrained=trained)\n self.input_size = 25088\n # classifier\n elif self.model_pretrained==\"vgg16\":\n self.model = models.vgg16(pretrained=trained)\n self.input_size = 25088\n # classifier\n elif self.model_pretrained==\"vgg19\":\n self.model = models.vgg19(pretrained=trained)\n self.input_size = 25088\n # classifier\n elif self.model_pretrained==\"alexnet\":\n self.model = models.alexnet(pretrained=trained) \n self.input_size = 9216\n # classifier\n else:\n print(\"invalid model architecture. Please choose among the following options : 'vgg13', 'vgg16', 'vgg19', 'alexnet'.\")\n return -1\n # Turn off gradients to not train these layers which are already optimized\n for param in self.model.parameters():\n param.requires_grad = False\n \n # Define a new, untrained feed-forward network as a classifier, using ReLU activations and dropout\n classifier = nn.Sequential(OrderedDict([('dropout1', nn.Dropout(p=0.5)),\n ('fc1', nn.Linear(self.input_size, self.hidden_units)),\n ('relu1', nn.ReLU()),\n ('dropout2', nn.Dropout(p=0.5)),\n ('fc2', nn.Linear(self.hidden_units, 102)),\n ('output',nn.LogSoftmax(dim=1))\n ]))\n # Replace the last layer of the pre-trained model\n self.model.classifier = classifier\n \n return self.model\n\n def train(self):\n \"\"\" Train the model using a train dataloader and see the progression using a validation dataloader\"\"\"\n # define loss calculation type\n self.criterion = nn.NLLLoss()\n # define optimizer type and allow modifications only on classifier\n self.optimizer = optim.Adam(self.model.classifier.parameters(), lr=self.learning_rate)\n # Set up the used device (cuda if available or cpu)\n if self.gpu:\n device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n else:\n device = torch.device('cpu')\n print(\"Deviced used for the training: \", device)\n\n # send model to GPU if available\n self.model.to(device)\n \n # train model (only the replaced layer)\n steps = 0\n running_loss = 0\n print_every = 5\n \n self.model.train() # set the model to train mode \n with active_session(): # desactivate the limited time unactive session from udacity\n\n for epoch in range(self.epochs):\n for train_inputs, train_labels in self.train_loader:\n steps += 1\n \n train_inputs, train_labels = train_inputs.to(device), train_labels.to(device) # move input and label tensors \n # to GPU/CPU device\n \n self.optimizer.zero_grad() # reset the gradient so that it is not accumulated\n \n logps = self.model.forward(train_inputs) # forward\n loss = self.criterion(logps, train_labels) # loss calculation\n loss.backward() # backward\n self.optimizer.step() # step\n \n running_loss += loss.item() # track the losses \n\n if steps % print_every == 0:\n valid_loss = 0\n valid_accuracy = 0\n self.model.eval() # set the model to evaluation mode (no dropout)\n with torch.no_grad(): # no gradient tracking to speed up the calculation\n for valid_inputs, valid_labels in self.valid_loader:\n \n valid_inputs, valid_labels = valid_inputs.to(device), valid_labels.to(device) # move input and label tensors \n # to GPU /CPU device\n logps = self.model.forward(valid_inputs) # forward\n batch_loss = self.criterion(logps, valid_labels) # loss calculation\n \n valid_loss += batch_loss.item() # track the losses \n \n # Calculate accuracy\n ps = torch.exp(logps) # probability calculation\n top_p, top_class = ps.topk(1, dim=1) # get the top class predicted by the model\n equals = top_class == valid_labels.view(*top_class.shape) # check if the prediction is right\n valid_accuracy += torch.mean(equals.type(torch.FloatTensor)).item() # update the accuracy of the model\n\n print(\"Epoch {}/{}.. \".format(epoch + 1, self.epochs),\n \"Train loss: {:.3f}.. \".format(running_loss / print_every),\n \"Valid loss: {:.3f}.. \".format(valid_loss / len(self.valid_loader)),\n \"Valid accuracy: {:.3f}\".format(valid_accuracy / len(self.valid_loader)))\n\n running_loss = 0 # reset the running losses\n self.model.train() # set back the model to train for the next batch of train_loader\n \n return self.model\n \n def test(self):\n \"\"\"Perform validation a the test set\"\"\"\n if self.gpu:\n device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n else:\n device = torch.device('cpu')\n print(\"Device used for the testing: \", device)\n test_loss = 0\n test_accuracy = 0\n self.model.eval() # set the model to evaluation mode (no dropout)\n with torch.no_grad(): # no gradient tracking to speed up the calculation\n for test_inputs, test_labels in self.test_loader:\n \n test_inputs, test_labels = test_inputs.to(device), test_labels.to(device)\n\n logps = self.model.forward(test_inputs)\n batch_loss = self.criterion(logps, test_labels)\n\n test_loss += batch_loss.item() # track the losses\n\n # Calculate accuracy\n ps = torch.exp(logps)\n top_p, top_class = ps.topk(1, dim=1)\n equals = top_class == test_labels.view(*top_class.shape)\n test_accuracy += torch.mean(equals.type(torch.FloatTensor)).item()\n return test_accuracy / len(self.test_loader)\n \n def load_cat_to_name(self, cat_to_name_path):\n \"\"\" Load the category to name .json file \"\"\"\n with open(cat_to_name_path, 'r') as f:\n self.cat_to_name = json.load(f)\n return self.cat_to_name\n \n def include_mapping(self, train_datasets):\n \"\"\" Include the mapping of the classes to indices inside the model.\"\"\"\n self.model.class_to_idx = train_datasets.class_to_idx\n\n def save(self):\n \"\"\" Save the model creating a checkpoint.\"\"\"\n self.model.to('cpu')\n # creation of the checkpoint dictionnary\n checkpoint = {'model': self.model,\n 'input_size': self.input_size,\n 'hidden_units': self.hidden_units,\n 'ouput_size': 102,\n 'epochs': self.epochs,\n 'optimizer': self.optimizer,\n 'criterion': self.criterion,\n 'class_to_idx': self.model.class_to_idx,\n 'state_dict': self.model.state_dict()\n }\n # Saving of the checkpoint\n torch.save(checkpoint, self.save_dir + 'checkpoint.pth')\n \n def load(self, checkpoint_path):\n \"\"\" Load a model saved in a checkpoint file \"\"\"\n checkpoint = torch.load(checkpoint_path, map_location=lambda storage, loc: storage) # force all tensors to be on CPU\n self.model = checkpoint['model']\n self.model.to('cpu')\n self.hidden_units = checkpoint['hidden_units']\n self.epochs = checkpoint['epochs']\n self.optimizer = checkpoint['optimizer']\n self.criterion = checkpoint['criterion']\n self.model.class_to_idx = checkpoint['class_to_idx']\n self.model.load_state_dict(checkpoint['state_dict'])\n return self.model\n \n def process_image(self, im_pil):\n \"\"\" Process a PIL image for use in a PyTorch model \"\"\"\n if self.gpu:\n device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n else:\n device = torch.device('cpu')\n \n # resize the images where the shortest side is 256 pixels keeping the ratio width/heigth\n im_width, im_height = im_pil.size\n if im_height > im_width:\n # im_width = 256\n new_width = 256\n new_height = int(256 * im_height / im_width)\n else:\n # im_height = 256\n new_width = int(256 * im_width / im_height)\n new_height = 256\n im_pil = im_pil.resize(size=(new_width, new_height), resample=0)\n \n # center crop of the picture to size (224, 224)\n left = (im_width/2 - 224)/2\n top = (im_height/2 - 224)/2\n right = (im_width/2 + 224)/2\n bottom = (im_height/2 + 224)/2\n im_pil = im_pil.crop((left, top, right, bottom))\n \n # convert the pixel value from 0-255 to 0-1\n im_np = np.array(im_pil)\n im_np = im_np / 255\n \n # normalize the images as the network expects (mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])\n mean = np.array([0.485, 0.456, 0.406])\n std = np.array([0.229, 0.224, 0.225])\n im_np = (im_np - mean) /std\n\n # put the color channel in the 1st dimention of the numpy array.\n im_np = np.transpose(im_np, (2, 0, 1))\n \n return im_np\n\n def predict(self, image_path, topk=5, class_to_idx=None):\n \"\"\" Predict the class (or classes) of an image using the model. \"\"\" \n if self.gpu:\n device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n else:\n device = torch.device('cpu')\n\n # Image preparation\n im_pil = PIL.Image.open(image_path)\n im_processed = self.process_image(im_pil)\n input_data = torch.from_numpy(im_processed) # convert the image array into a torch tensor\n input_data = input_data.float() # The default type for weights and biases are torch.FloatTensor \n # so we convert the input_data into torch.FloatTensor\n input_data = input_data.unsqueeze(0) # Adds a dimension of size 1 at the beginning of the tensor to match PyTorch model.\n # Normally it has the batch_size at the beggining. \n # In this case, the batch_size = 1 since we only have one picture to treat\n \n # Evaluation\n self.model.eval() #Put the model in evaluation mode (no dropouts)\n with torch.no_grad(): # no gradient tracking to speed up the calculation\n logps = self.model.forward(input_data) # prediction of the model\n ps = torch.exp(logps) # probabilities\n top_p, top_class = ps.topk(topk, dim=1) # top 5 probabilities and corresponding classes\n \n top_p, top_class = top_p.numpy()[0], top_class.numpy()[0]+1\n \n # if a category to name json file has been loaded\n if self.cat_to_name is not None:\n # replace the category to name in top_class\n top_class = [self.cat_to_name[str(elem)] for elem in top_class]\n \n return top_p, top_class\n","sub_path":"model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":13237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"55744464","text":"# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n\nimport unittest\n\ntry:\n import unittest.mock as mock\nexcept ImportError:\n import mock\n\nfrom knack.util import CLIError\n\nfrom azure.cli.command_modules.monitor.operations.autoscale_settings import scaffold_autoscale_settings_parameters\n\n\nclass FilterBuilderTests(unittest.TestCase):\n def test_scaffold_autoscale_settings_parameters(self):\n template = scaffold_autoscale_settings_parameters(None)\n self.assertTrue(template)\n self.assertTrue(isinstance(template, dict))\n\n\ndef _mock_get_subscription_id(_):\n return '00000000-0000-0000-0000-000000000000'\n\n\nclass MonitorNameOrIdTest(unittest.TestCase):\n def _build_namespace(self, name_or_id=None, resource_group=None, provider_namespace=None, parent=None,\n resource_type=None):\n from argparse import Namespace\n ns = Namespace()\n ns.name_or_id = name_or_id\n ns.resource_group_name = resource_group\n ns.namespace = provider_namespace\n ns.parent = parent\n ns.resource_type = resource_type\n return ns\n\n @mock.patch('azure.cli.core.commands.client_factory.get_subscription_id', _mock_get_subscription_id)\n def test_monitor_resource_id(self):\n from azure.cli.command_modules.monitor.validators import get_target_resource_validator\n from azure.cli.core.mock import DummyCli\n\n cmd = mock.MagicMock()\n cmd.cli_ctx = DummyCli()\n validator = get_target_resource_validator('name_or_id', True)\n\n id = '/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/my-rg/providers/Microsoft.Compute/' \\\n 'virtualMachines/vm1'\n\n # must supply name or ID\n ns = self._build_namespace()\n with self.assertRaises(CLIError):\n validator(cmd, ns)\n\n # must only supply ID or name parameters\n ns = self._build_namespace(id, 'my-rg', 'blahblah', 'stuff')\n with self.assertRaises(CLIError):\n validator(cmd, ns)\n\n # error on invalid ID\n ns = self._build_namespace('bad-id')\n with self.assertRaises(CLIError):\n validator(cmd, ns)\n\n # allow Provider/Type syntax (same as resource commands)\n ns = self._build_namespace('vm1', 'my-rg', None, None, 'Microsoft.Compute/virtualMachines')\n validator(cmd, ns)\n self.assertEqual(ns.name_or_id, id)\n\n # allow Provider and Type separate\n ns = self._build_namespace('vm1', 'my-rg', 'Microsoft.Compute', None, 'virtualMachines')\n validator(cmd, ns)\n self.assertEqual(ns.name_or_id, id)\n\n # verify works with parent\n id = '/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/my-rg/providers/Microsoft.Compute/' \\\n 'fakeType/type1/anotherFakeType/type2/virtualMachines/vm1'\n ns = self._build_namespace('vm1', 'my-rg', 'Microsoft.Compute', 'fakeType/type1/anotherFakeType/type2',\n 'virtualMachines')\n validator(cmd, ns)\n self.assertEqual(ns.name_or_id, id)\n\n # verify extra parameters are removed\n self.assertFalse(hasattr(ns, 'resource_name'))\n self.assertFalse(hasattr(ns, 'namespace'))\n self.assertFalse(hasattr(ns, 'parent'))\n self.assertFalse(hasattr(ns, 'resource_type'))\n","sub_path":"src/command_modules/azure-cli-monitor/azure/cli/command_modules/monitor/tests/latest/test_monitor_unittest.py","file_name":"test_monitor_unittest.py","file_ext":"py","file_size_in_byte":3638,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"91307276","text":"from constants import MAX_Y, MAX_X\nfrom random import randint\n\nclass Food:\n\n #initilizing\n def __init__(self, window, x, y):\n self.window = window\n self.x = x\n self.y = y\n self.fruit = '@'\n\n #rendering fruit\n def render(self, x , y):\n self.x = x\n self.y = y\n self.window.addstr(self.y, self.x, self.fruit)\n\n","sub_path":"client/Food.py","file_name":"Food.py","file_ext":"py","file_size_in_byte":368,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"292669381","text":"import string\nimport collections\nimport operator\nimport copy\n\n_ALL_DATABASES = {}\n\nglobal counter\ncounter = 0\n\nclass Connection(object):\n def __init__(self, filename):\n \"\"\"\n Takes a filename, but doesn't do anything with it.\n \"\"\"\n self.filename = filename\n self.db_copy = None\n global counter\n self.id = counter\n counter += 1\n if filename not in _ALL_DATABASES.keys():\n _ALL_DATABASES[filename] = Database(filename)\n \n def execute(self, statement):\n \"\"\"\n Takes a SQL statement.\n Returns a list of tuples (empty unless select statement \n with rows to return).\n \"\"\"\n ret = []\n tokens = tokenize(statement)\n \n if tokens[0] == \"BEGIN\":\n DB = _ALL_DATABASES[self.filename]\n self.db_copy = copy.deepcopy(DB)\n self.db_copy.autocommit = False\n transaction = Transaction(self.db_copy)\n DB.autocommit = False\n if tokens[1] == \"IMMEDIATE\":\n for t in DB.transactions.keys():\n if DB.transactions[t].res_lock:\n raise Exception(\"OperationalError: database is locked\")\n transaction.res_lock = True\n elif tokens[1] == \"EXCLUSIVE\":\n for t in DB.transactions.keys():\n if DB.transactions[t].res_lock or DB.transactions[t].ex_lock:\n raise Exception(\"OperationalError: database is locked\")\n transaction.ex_lock = True\n DB.transactions[self.id] = transaction\n \n elif tokens[0] == \"ROLLBACK\":\n DB = _ALL_DATABASES[self.filename]\n self.db_copy = copy.deepcopy(DB)\n self.db_copy.autocommit = False\n DB.transactions[self.id] = Transaction(self.db_copy)\n \n \n \n elif tokens[0] == \"COMMIT\":\n DB = _ALL_DATABASES[self.filename]\n if DB.transactions[self.id].res_lock:\n for t in DB.transactions.keys():\n if t != self.id and \\\n (DB.transactions[t].sh_lock or \\\n DB.transactions[t].res_lock or \\\n DB.transactions[t].ex_lock):\n raise Exception(\"OperationalError: database is locked\")\n _ALL_DATABASES[self.filename] = copy.deepcopy(self.db_copy)\n del DB.transactions[self.id]\n if not _ALL_DATABASES[self.filename].transactions:\n _ALL_DATABASES[self.filename].autocommit = True\n \n \n elif tokens[0] == \"CREATE\":\n if tokens[1] == \"TABLE\":\n if tokens[2:5] == [\"IF\", \"NOT\", \"EXISTS\"]:\n table = Table(tokens[5], tokens[6:-1])\n if table.name not in _ALL_DATABASES[self.filename].tables.keys():\n _ALL_DATABASES[self.filename].add_table(table)\n else:\n table = Table(tokens[2], tokens[3:-1])\n if table.name not in _ALL_DATABASES[self.filename].tables.keys():\n _ALL_DATABASES[self.filename].add_table(table)\n else:\n raise Exception(\"ExistingTableError\")\n elif tokens[1] == \"VIEW\":\n view = View(tokens[2], tokens[4:], self.filename)\n _ALL_DATABASES[self.filename].add_view(view)\n \n\n elif tokens[0] == \"DROP\":\n if tokens[2:4] == [\"IF\", \"EXISTS\"]:\n tablename = tokens[4]\n if tablename in _ALL_DATABASES[self.filename].tables.keys():\n del _ALL_DATABASES[self.filename].tables[tablename]\n else:\n tablename = tokens[2]\n if tablename in _ALL_DATABASES[self.filename].tables.keys():\n del _ALL_DATABASES[self.filename].tables[tablename]\n else:\n raise Exception(\"ExistingTableError\")\n \n elif tokens[0] == \"INSERT\":\n DB = _ALL_DATABASES[self.filename]\n \n if not DB.autocommit:\n for t in DB.transactions.keys():\n if t != self.id and \\\n (DB.transactions[t].res_lock or \\\n DB.transactions[t].ex_lock):\n raise Exception(\"OperationalError: database is locked\")\n if self.db_copy:\n DB.transactions[self.id].res_lock = True\n DB = self.db_copy\n \n \n table = DB.tables[tokens[2]]\n if tokens[3] == \"VALUES\":\n label = \"\"\n row = collections.OrderedDict()\n query = \"\".join(tokens[tokens.index(\"VALUES\")+1:-1]).split(',')\n for i in range(len(query)):\n label = table.headers[(3*(i))%(len(table.headers)+1)]\n type = table.headers[(3*(i) + 1)%(len(table.headers)+1)]\n if query[i] == \"NULL\":\n row[label] = None\n elif type == \"INTEGER\":\n row[label] = int(query[i])\n elif type == \"REAL\":\n row[label] = float(query[i])\n elif type == \"TEXT\":\n row[label] = str(query[i])\n if i % ((len(table.headers)//3)+1) == len(table.headers)//3:\n table.insert(row)\n row = collections.OrderedDict()\n else:\n columns = tokens[3:tokens.index(\"VALUES\"):2]\n label = \"\"\n row = collections.OrderedDict()\n query = \"\".join(tokens[tokens.index(\"VALUES\")+1:-1]).split(',')\n header_labels = table.headers[::3]\n for i in range(len(query)):\n label = columns[i % len(columns)]\n type = table.headers[table.headers.index(label)+1]\n if query[i] == \"NULL\":\n row[label] = None\n header_labels.remove(label)\n elif type == \"INTEGER\":\n row[label] = int(query[i])\n header_labels.remove(label)\n elif type == \"REAL\":\n row[label] = float(query[i])\n header_labels.remove(label)\n elif type == \"TEXT\":\n row[label] = str(query[i])\n header_labels.remove(label)\n if i % (len(columns)) == len(columns)-1:\n new_row = collections.OrderedDict()\n for header in table.headers[::3]:\n if header in row.keys():\n new_row[header] = row[header]\n else:\n new_row[header] = None\n table.insert(new_row)\n row = collections.OrderedDict()\n header_labels = table.headers[::3]\n \n \n elif tokens[0] == \"SELECT\":\n DB = _ALL_DATABASES[self.filename]\n if not DB.autocommit:\n for t in DB.transactions.keys():\n if t != self.id and DB.transactions[t].ex_lock:\n raise Exception(\"OperationalError: database is locked\")\n DB.transactions[self.id].sh_lock = True\n DB = self.db_copy\n \n \n ordered_columns = []\n where_clause = []\n distinct = False\n table_b = None\n on_clause = []\n if tokens[1] == '*':\n e_name = tokens[3]\n if e_name in DB.tables.keys():\n table = DB.tables[e_name]\n else:\n assert e_name in DB.views.keys()\n table = DB.views[e_name].view_frame_select()\n tokens = remove_table_id(tokens, table.name)\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers[::3]:\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n else:\n e_name = tokens[tokens.index(\"FROM\")+1]\n if e_name in DB.tables.keys():\n table = DB.tables[e_name]\n else:\n assert e_name in DB.views.keys()\n table = DB.views[e_name].view_frame_select()\n \n tokens = remove_table_id(tokens, table.name)\n \n if \"DISTINCT\" in tokens:\n distinct = True\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"DISTINCT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n else:\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"SELECT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n \n new_ordered_columns = []\n for column in ordered_columns:\n if column == '*':\n for header in table.header_names:\n new_ordered_columns.append(header)\n if new_ordered_columns:\n ordered_columns = new_ordered_columns\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers and sort_criteria != ',':\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n #might not work with WHERE and JOIN in same query\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n \n if \"JOIN\" in tokens:\n assert \"LEFT\" in tokens\n assert \"OUTER\" in tokens\n assert \"ON\" in tokens\n \n table_b = DB.tables[tokens[tokens.index(\"JOIN\")+1]]\n #might not work with WHERE and JOIN in same query\n on_clause = tokens[tokens.index(\"ON\")+1:tokens.index(\"ORDER\")]\n \n ret = table.select(orderby, ordered_columns, where_clause, distinct, table_b, on_clause)\n \n elif tokens[0] == \"UPDATE\":\n DB = _ALL_DATABASES[self.filename]\n if not DB.autocommit:\n for t in DB.transactions.keys():\n if t != self.id and \\\n (DB.transactions[t].res_lock or \\\n DB.transactions[t].ex_lock):\n raise Exception(\"OperationalError: database is locked\")\n \n if self.db_copy:\n DB.transactions[self.id].res_lock = True\n DB = self.db_copy\n \n table = DB.tables[tokens[1]]\n if \"WHERE\" in tokens:\n table.update_where(tokens[tokens.index(\"SET\")+1:tokens.index(\"WHERE\")],\\\n tokens[tokens.index(\"WHERE\")+1:-1])\n else:\n for row in table.rows:\n row = row_update(row, tokens[tokens.index(\"SET\")+1:])\n \n elif tokens[0] == \"DELETE\":\n DB = _ALL_DATABASES[self.filename]\n if not DB.autocommit:\n for t in DB.transactions.keys():\n if t != self.id and \\\n (DB.transactions[t].res_lock or \\\n DB.transactions[t].ex_lock):\n raise Exception(\"OperationalError: database is locked\")\n if self.db_copy:\n DB.transactions[self.id].res_lock = True\n DB = self.db_copy\n \n table = DB.tables[tokens[2]]\n if \"WHERE\" in tokens:\n table.delete_where(tokens[tokens.index(\"WHERE\")+1:-1])\n else:\n table.rows = []\n \n \n else:\n print(\"Unrecognised statement\")\n\n return ret\n \n def close(self):\n \"\"\"\n Empty method that will be used in future projects\n \"\"\"\n pass\n\ndef connect(filename):\n \"\"\"\n Creates a Connection object with the given filename\n \"\"\"\n return Connection(filename)\n\ndef tokenize(query):\n \"\"\"\n Returns a list of tokens given by the query\n \"\"\"\n prev_char = None\n tokens = []\n token = \"\"\n in_token = False\n in_single = False\n in_double = False\n check_single = False\n\n for char in query:\n if check_single:\n if prev_char == \"'\" and char == \"'\":\n if in_token:\n token += char\n else:\n tokens.append(\"\")\n token = \"\"\n prev_char = None\n check_single = False\n continue\n else:\n in_token = False\n in_single = False\n check_single = False\n if token:\n tokens.append(token)\n token = \"\"\n if char == \"'\" and in_single:\n check_single = True\n elif char == \"'\" and not in_single:\n in_token = True\n in_single = True\n elif char == '\"':\n if not in_double:\n in_double = True\n in_token = True\n else:\n if token:\n tokens.append(token)\n token = \"\"\n in_double = False\n in_token = False\n elif char.isalnum() or char in \"._\":\n token += char\n in_token = True\n elif char in (string.whitespace + \"(),*;<>=\"):\n if not in_single and not in_double:\n if token:\n tokens.append(token)\n if char in \",*;<>=\":\n if (char == '=' and prev_char in \"!<>\") or \\\n (char == '>' and prev_char in \"!<\") or \\\n (char == '<' and prev_char == '!'):\n tokens.append(prev_char+char)\n else:\n tokens.append(char)\n token = \"\"\n in_token = False\n else:\n token += char\n in_token = True\n prev_char = char\n \n return tokens\n\ndef getvalues(row):\n \"\"\"\n Returns an array of the values of a row\n \"\"\"\n ret = []\n for value in row.values():\n ret.append(value)\n return ret\n \ndef append_joined_values(a_row, table_b, a_column, b_column):\n \"\"\"\n Joins and returns a_row from a table (table_a) and a corresponding \n row from table_b according to a left outer join on \n table_a.a_column = table_b.b_column\n \n If not matching row from table_b can be found, the appended values will be\n NULL/None\n \"\"\" \n b_column = remove_table_id([b_column], table_b.name)[0]\n found = False\n \n for b_row in table_b.rows:\n if a_row[a_column] == b_row[b_column]:\n found = True\n for key in b_row.keys():\n if key not in a_row.keys():\n a_row[key] = b_row[key]\n \n if not found:\n for key in b_row.keys():\n if key not in a_row.keys():\n a_row[key] = None\n \n return a_row \n \ndef row_update(row, set_clause):\n \"\"\"\n Returns an updated row according to set_clause\n \"\"\"\n column_names = set_clause[::4]\n values = set_clause[2::4]\n \n name_length = len(column_names)\n assert name_length == len(values)\n \n for i in range(name_length):\n if type(row[column_names[i]]) is int:\n row[column_names[i]] = int(values[i])\n elif type(row[column_names[i]]) is float:\n row[column_names[i]] = float(values[i])\n elif type(row[column_names[i]]) is str:\n row[column_names[i]] = str(values[i])\n \n return row\n \ndef remove_table_id(tokens, name):\n \"\"\"\n Returns a list of tokens with redundant table IDs (name) removed\n \"\"\"\n ret = []\n \n for token in tokens:\n if (name + '.') in token:\n ret.append(token[len(name)+1:])\n else:\n ret.append(token)\n \n return ret\n \ndef check_pred(pred_left, pred_sign, pred_right):\n \"\"\"\n Evaluates and returns the result of a comparison predicate.\n \n Arguments:\n pred_left: left value of the comparison predicate\n pred_sign: string representing the comparison operator\n pred_right: right value of the comparison predicate\n \"\"\"\n ret = False\n if pred_sign == '=' or pred_sign == \"IS\":\n if pred_left == pred_right:\n ret = True\n elif pred_sign == '>' and pred_left is not None:\n if pred_left > pred_right:\n ret = True\n elif pred_sign == '<' and pred_left is not None:\n if pred_left < pred_right:\n ret = True\n elif pred_sign == \">=\" or pred_sign == \"!<\" and pred_left is not None:\n if pred_left >= pred_right:\n ret = True\n elif pred_sign == \"<=\" or pred_sign == \"!>\" and pred_left is not None:\n if pred_left <= pred_right:\n ret = True\n elif pred_sign == \"<>\" or pred_sign == \"!=\" or pred_sign == \"IS NOT\":\n if pred_left != pred_right:\n ret = True \n \n return ret\n \ndef type_right_pred(pred_right, pred_type):\n \"\"\"\n Returns pred_right converted to the type specified by the pred_type string\n \"\"\"\n if pred_type == \"NULL\" or pred_right == \"NULL\":\n pred_right = None\n elif pred_type == \"INTEGER\":\n pred_right = int(pred_right)\n elif pred_type == \"REAL\":\n pred_right = float(pred_right)\n elif pred_type == \"TEXT\":\n pred_right = str(pred_right)\n \n return pred_right\n\nclass Transaction(object):\n def __init__(self, database):\n self.database = database\n self.sh_lock = False\n self.res_lock = False\n self.ex_lock = False\n \n \nclass Database(object):\n def __init__(self, name, autocommit = True):\n self.name = name\n self.tables = {}\n self.views = {}\n self.transactions = {}\n self.autocommit = autocommit\n self.sh_lock = False\n self.res_lock = False\n self.ex_lock = False\n \n def add_table(self, table):\n self.tables[table.name] = table\n \n \n def add_view(self, view):\n self.views[view.name] = view\n\nclass Table(object):\n def __init__(self, name, headers = []):\n self.name = name\n self.headers = headers\n self.header_names = headers[::3]\n self.view_headers = None\n self.rows = []\n \n def insert(self, row):\n self.rows.append(row)\n \n def select(self, orderby = [], ordered_columns = [], where_clause = [], distinct = False,\\\n table_b = None, on_clause = []):\n \"\"\"\n Returns a list of tuples representing the rows selected by the SELECT\n statement\n \n Arguments:\n orderby: The columns that the rows should be ordered by, as designated\n by the ORDER BY statement (ORDER BY [orderby])\n ordered_columns: An ordered list of columns representing the order that\n each individual row should be organized by \n (SELECT [ordered_columns])\n If empty, show all values of each row in their default order\n (SELECT *)\n where_clause: A list of strings representing a WHERE statement\n distinct: A boolean of whether or not a DISTINCT statement is present\n table_b: A table to join with on any JOIN statements\n on_clause: A list of strings representing an ON statement\n \"\"\"\n ret = []\n\n if where_clause:\n column = where_clause[0]\n if len(where_clause) == 3:\n pred_sign = where_clause[1]\n elif len(where_clause) == 4:\n pred_sign = where_clause[1] + ' ' + where_clause[2]\n assert pred_sign == \"IS NOT\"\n pred_right = type_right_pred(where_clause[-1], self.headers[self.headers.index(column)+1])\n \n if on_clause:\n on_left = on_clause[0]\n on_right = on_clause[-1]\n assert len(on_clause) == 3\n assert on_clause[1] == '='\n \n if not orderby:\n for row in self.rows:\n if where_clause:\n if check_pred(row[column], pred_sign, pred_right):\n ret.append(tuple(getvalues(row)))\n else:\n ret.append(tuple(getvalues(row)))\n else:\n order_pos = [] # positions of the ordered columns with respect to\n # the table's original headers\n column_pos = [] # positions of the \"columns\" argument with respect\n # to the table's original headers\n row_list = [] # List of lists of values of each row\n \n \n for sort_criteria in orderby:\n if self.view_headers:\n for i in range(len(self.view_headers)):\n if sort_criteria == self.view_headers[i]:\n order_pos.append(i)\n else:\n for i in range(len(self.headers[::3])):\n if sort_criteria == self.headers[3*i]:\n order_pos.append(i)\n \n for o_column in ordered_columns:\n for j in range(len(self.header_names)):\n if o_column == self.header_names[j]:\n column_pos.append(self.headers.index(o_column)//3) \n \n for row in self.rows:\n if where_clause:\n if check_pred(row[column], pred_sign, pred_right):\n row_list.append(getvalues(row))\n elif on_clause:\n row = append_joined_values(row, table_b, on_left, on_right)\n row_list.append(getvalues(row))\n else:\n row_list.append(getvalues(row))\n \n \n sorted_list = sorted(row_list, key=lambda x: x[order_pos[0]]) #\n for row in sorted_list:\n result = []\n if column_pos:\n for i in range(len(column_pos)):\n result.append(row[column_pos[i]])\n if on_clause:\n for value in row[len(self.header_names):]:\n result.append(value)\n if not column_pos and not on_clause:\n result = row\n ret.append(tuple(result))\n \n if distinct: \n return sorted(set(ret))\n else:\n return ret\n \n def view_select(self, view_headers, orderby = [], ordered_columns = [], where_clause = [], distinct = False,\\\n table_b = None, on_clause = []):\n \"\"\"\n Returns a table representing the rows selected by the SELECT\n statement\n \n Arguments:\n orderby: The columns that the rows should be ordered by, as designated\n by the ORDER BY statement (ORDER BY [orderby])\n ordered_columns: An ordered list of columns representing the order that\n each individual row should be organized by \n (SELECT [ordered_columns])\n If empty, show all values of each row in their default order\n (SELECT *)\n where_clause: A list of strings representing a WHERE statement\n distinct: A boolean of whether or not a DISTINCT statement is present\n table_b: A table to join with on any JOIN statements\n on_clause: A list of strings representing an ON statement\n \"\"\"\n ret = Table(\"temp\", self.headers)\n ret.view_headers = view_headers\n\n if where_clause:\n column = where_clause[0]\n if len(where_clause) == 3:\n pred_sign = where_clause[1]\n elif len(where_clause) == 4:\n pred_sign = where_clause[1] + ' ' + where_clause[2]\n assert pred_sign == \"IS NOT\"\n pred_right = type_right_pred(where_clause[-1], self.headers[self.headers.index(column)+1])\n \n if on_clause:\n on_left = on_clause[0]\n on_right = on_clause[-1]\n assert len(on_clause) == 3\n assert on_clause[1] == '='\n \n if not orderby:\n for row in self.rows:\n if where_clause:\n if check_pred(row[column], pred_sign, pred_right):\n ret.insert(row)\n else:\n ret.insert(row)\n else:\n order_pos = [] # positions of the ordered columns with respect to\n # the table's original headers\n column_pos = [] # positions of the \"columns\" argument with respect\n # to the table's original headers\n row_list = [] # List of lists of values of each row\n \n for sort_criteria in orderby:\n for i in range(len(self.headers[::3])):\n if sort_criteria == self.headers[3*i]:\n order_pos.append(i)\n \n for o_column in ordered_columns:\n for j in range(len(self.header_names)):\n if o_column == self.header_names[j]:\n column_pos.append(self.headers.index(o_column)//3) \n \n for row in self.rows:\n if where_clause:\n if check_pred(row[column], pred_sign, pred_right):\n row_list.append(row)\n elif on_clause:\n row = append_joined_values(row, table_b, on_left, on_right)\n row_list.append(row)\n else:\n row_list.append(row)\n \n sorted_list = sorted(row_list, key=lambda x: x[orderby[0]]) #lazy\n\n for row in sorted_list:\n result = collections.OrderedDict()\n if column_pos:\n for oc in ordered_columns:\n result[oc] = row[oc]\n if on_clause:\n for value in row[len(self.header_names):]:\n result.append(value)\n if not column_pos and not on_clause:\n result = row\n ret.insert(result)\n\n return ret\n \n def delete_where(self, where_clause):\n \"\"\"\n Deletes any rows from the table specifed by where_clause\n \"\"\"\n assert len(where_clause) == 3\n \n column = where_clause[0]\n pred_sign = where_clause[1]\n pred_right = type_right_pred(where_clause[2], self.headers[self.headers.index(column)+1])\n \n del_indeces = []\n \n for row in self.rows:\n index = self.rows.index(row)\n if check_pred(row[column], pred_sign, pred_right):\n del_indeces.append(index)\n \n self.rows = [i for j, i in enumerate(self.rows) if j not in del_indeces]\n \n def update_where(self, set_clause, where_clause):\n \"\"\"\n Updates any rows in the table to follow set_clause according to\n where_clause\n \"\"\"\n assert len(where_clause) == 3\n \n column = where_clause[0]\n pred_sign = where_clause[1]\n pred_right = type_right_pred(where_clause[2], self.headers[self.headers.index(column)+1])\n \n column_pos = self.headers.index(column)//3\n \n for row in self.rows:\n if check_pred(row[column], pred_sign, pred_right):\n row = row_update(row, set_clause)\n\nclass View(object):\n def __init__(self, name, select_statement, filename):\n self.name = name\n self.tablename = select_statement[select_statement.index(\"FROM\") + 1]\n self.filename = filename\n self.ss = select_statement\n \n def view_frame_select(self):\n tokens = self.ss\n ordered_columns = []\n where_clause = []\n distinct = False\n table_b = None\n on_clause = []\n if tokens[1] == '*':\n table = _ALL_DATABASES[self.filename].tables[tokens[3]]\n \n tokens = remove_table_id(tokens, table.name)\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n \n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers[::3]:\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n else:\n table = _ALL_DATABASES[self.filename]\\\n .tables[tokens[tokens.index(\"FROM\")+1]]\n table.view_headers = tokens[tokens.index(\"SELECT\")+1:tokens.index(\"FROM\"):2] \n \n tokens = remove_table_id(tokens, table.name)\n \n if \"DISTINCT\" in tokens:\n distinct = True\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"DISTINCT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n else:\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"SELECT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n \n new_ordered_columns = []\n for column in ordered_columns:\n if column == '*':\n for header in table.header_names:\n new_ordered_columns.append(header)\n if new_ordered_columns:\n ordered_columns = new_ordered_columns\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers and sort_criteria != ',':\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n \n if \"JOIN\" in tokens:\n assert \"LEFT\" in tokens\n assert \"OUTER\" in tokens\n assert \"ON\" in tokens\n \n table_b = _ALL_DATABASES[self.filename].tables[tokens[tokens.index(\"JOIN\")+1]]\n on_clause = tokens[tokens.index(\"ON\")+1:tokens.index(\"ORDER\")]\n \n return table.view_select(table.view_headers, orderby, ordered_columns, where_clause, distinct, table_b, on_clause)\n \n def select(self, tokens):\n table = self.view_from_select()\n ordered_columns = []\n where_clause = []\n distinct = False\n table_b = None\n on_clause = []\n if tokens[1] == '*': \n tokens = remove_table_id(tokens, table.name)\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers[::3]:\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n else: \n tokens = remove_table_id(tokens, table.name)\n \n if \"DISTINCT\" in tokens:\n distinct = True\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"DISTINCT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n else:\n ordered_columns = \"\".join\\\n (tokens[tokens.index(\"SELECT\")+1:tokens.index(\"FROM\")])\\\n .split(',')\n \n new_ordered_columns = []\n for column in ordered_columns:\n if column == '*':\n for header in table.header_names:\n new_ordered_columns.append(header)\n if new_ordered_columns:\n ordered_columns = new_ordered_columns\n \n if \"ORDER\" in tokens:\n pos = tokens.index(\"BY\")\n orderby = []\n for sort_criteria in tokens[pos+1:-1]:\n if sort_criteria in table.headers and sort_criteria != ',':\n orderby.append(sort_criteria)\n if \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:tokens.index(\"ORDER\")]\n elif \"WHERE\" in tokens:\n where_clause = tokens[tokens.index(\"WHERE\")+1:]\n \n if \"JOIN\" in tokens:\n assert \"LEFT\" in tokens\n assert \"OUTER\" in tokens\n assert \"ON\" in tokens\n \n table_b = _ALL_DATABASES[self.filename].tables[tokens[tokens.index(\"JOIN\")+1]]\n on_clause = tokens[tokens.index(\"ON\")+1:tokens.index(\"ORDER\")]\n \n return table.select(orderby, ordered_columns, where_clause, distinct, table_b, on_clause)\n \n ","sub_path":"sqlemu.py","file_name":"sqlemu.py","file_ext":"py","file_size_in_byte":34680,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"129083000","text":"from flask import Flask, request, redirect, render_template\nfrom flask_sqlalchemy import SQLAlchemy\n\napp = Flask(__name__)\napp.config['DEBUG'] = True\napp.config['SQLALCHEMY_DATABASE_URI'] = 'mysql+pymysql://build-a-blog:build-a-blog@localhost:8889/build-a-blog'\napp.config['SQLALCHEMY_ECHO'] = True\ndb = SQLAlchemy(app)\n\n\nclass Blog(db.Model):\n\n id = db.Column(db.Integer, primary_key=True)\n title = db.Column(db.String(120))\n body = db.Column(db.String(500))\n \n\n def __init__(self, title, body):\n self.title = title\n self.body = body\n \n\n\n\n@app.route('/blog', methods=['POST', 'GET'])\ndef blog():\n blogs = Blog.query.all()\n id = request.args.get('id')\n\n if not id:\n return render_template('blog.html', blogs=blogs)\n\n else: \n blog = Blog.query.get(id)\n title = blog.title\n body = blog.body\n return render_template('entry.html', blog_title=title, blog_body=body)\n\n\n\n\n@app.route('/newpost', methods=['POST', 'GET'])\ndef new_post():\n title_error = \"Please give your blog a title\"\n body_error = \"Please write some stuff\"\n\n\n if request.method == 'GET':\n return render_template('newpost.html')\n\n if request.method == 'POST':\n blog_title = request.form['title']\n blog_body = request.form['body']\n\n \n if (not blog_title) or (blog_title.strip() == \"\"):\n if (not blog_body) or (blog_body.strip() == \"\"):\n return render_template('newpost.html', blog_title=blog_title, blog_body=blog_body, title_error=title_error, body_error=body_error) \n else:\n return render_template('newpost.html', blog_title=blog_title, blog_body=blog_body, title_error=title_error)\n\n if (not blog_body) or (blog_body.strip() == \"\"):\n return render_template('newpost.html', blog_title=blog_title, blog_body=blog_body, body_error=body_error)\n\n if (not blog_title) or (blog_title.strip() == \"\") and (not blog_body) or (blog_body.strip() == \"\"): \n return render_template('newpost.html', blog_title=blog_title, blog_body=blog_body, title_error=title_error, body_error=body_error) \n \n else: \n new_post = Blog(blog_title, blog_body)\n db.session.add(new_post)\n db.session.commit()\n just_posted = db.session.query(Blog).order_by(Blog.id.desc()).first()\n id = str(just_posted.id)\n return redirect('/blog?id=' + id)\n\n\n\n\n@app.route('/', methods=['POST', 'GET'])\ndef index():\n return render_template('blog.html')\n\n\nif __name__ == '__main__':\n app.run()","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2604,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"210512081","text":"#This script processes data outputs for the resilience indicator multihazard model for the Philippines. Developed by Brian Walsh.\nfrom IPython import get_ipython\nget_ipython().magic('reset -f')\nget_ipython().magic('load_ext autoreload')\nget_ipython().magic('autoreload 2')\n\n#Import packages for data analysis\nfrom lib_compute_resilience_and_risk import *\nfrom replace_with_warning import *\nfrom lib_country_dir import *\nfrom lib_gather_data import *\nfrom maps_lib import *\n\nfrom scipy.stats import norm\nimport matplotlib.mlab as mlab\n\nimport matplotlib.patches as patches\nfrom pandas import isnull\nimport pandas as pd\nimport numpy as np\nimport os, time\nimport sys\n\n#Aesthetics\nimport seaborn as sns\nimport brewer2mpl as brew\nfrom matplotlib import colors\nsns.set_style('darkgrid')\nbrew_pal = brew.get_map('Set1', 'qualitative', 8).mpl_colors\nsns_pal = sns.color_palette('Set1', n_colors=8, desat=.5)\ngreys_pal = sns.color_palette('Greys', n_colors=9)\nreds_pal = sns.color_palette('Reds', n_colors=9)\nq_labels = ['Q1 (Poorest)','Q2','Q3','Q4','Q5 (Wealthiest)']\nq_colors = [sns_pal[0],sns_pal[1],sns_pal[2],sns_pal[3],sns_pal[5]]\n\nfont = {'family' : 'sans serif',\n 'size' : 20}\nplt.rc('font', **font)\nmpl.rcParams['xtick.labelsize'] = 16\n\nimport warnings\nwarnings.filterwarnings('always',category=UserWarning)\n\nmyCountry = 'PH'\nif len(sys.argv) < 2:\n print('Could list country. Using PH.')\nelse: myCountry = sys.argv[1]\n\nmodel = os.getcwd() #get current directory\noutput = model+'/../output_country/'+myCountry+'/'\n\n# Load output files\npol_str = ''#'_v95'#could be {'_v95'}\nbase_str = 'no'\npds_str = 'unif_poor'\n\nmacro = pd.read_csv(output+'macro_tax_'+pds_str+'_.csv')\n\ntry:\n iah_pds = pd.read_csv('/Users/brian/Desktop/BANK/hh_resilience_model/check_plots/test_hh.csv').reset_index()\nexcept:\n iah_pds = pd.read_csv(output+'iah_tax_'+pds_str+'_'+pol_str+'.csv').reset_index()\n #iah_pds = iah_pds.loc[(iah_pds.hhid==153829114)&(iah_pds.hazard=='EQ')&(iah_pds.rp==100)&(iah_pds.affected_cat=='a')&(iah_pds.helped_cat=='helped')]\n iah_pds = iah_pds.loc[(iah_pds.help_received>0)&(iah_pds.hazard=='EQ')&(iah_pds.rp==100)&(iah_pds.affected_cat=='a')&(iah_pds.helped_cat=='helped')].head(1)\n iah_pds.to_csv('/Users/brian/Desktop/BANK/hh_resilience_model/check_plots/test_hh.csv')\nprint(iah_pds.columns)\n\n# k recovery\nconst_dk_reco = np.log(1/0.05)/float(iah_pds['hh_reco_rate'])\nconst_pds = (np.log(1/0.05)/3.)*6. # PDS consumed in first half year of recovery \nconst_prod_k = float(macro.avg_prod_k.mean())\n\nprint(iah_pds.head())\n\nc = float(iah_pds['c'])\ndc0 = float(iah_pds['dc0'])\npds = float(iah_pds['help_received'])\n\nk = float(iah_pds['k'])\ndk0 = float(iah_pds['dk0'])\ndkprv = float(iah_pds['dk_private'])\ndkpub = float(iah_pds['dk_public'])\n\nc_t = [] \ndc_k_t = []\ndc_reco_t = []\ndc_pds_t = []\n\nt_lins = np.linspace(0,10,200)\nfor t in t_lins:\n c_t.append(c)\n dc_k_t.append(dk0*const_prod_k*np.e**(-(t)*const_dk_reco))\n dc_reco_t.append(dk0*const_dk_reco*np.e**(-(t)*const_dk_reco))\n dc_pds_t.append(pds*const_pds*np.e**(-(t)*const_pds))\n \n#step_dt*((1.-(temp['dc0']/temp['c'])*math.e**(-i_dt*const_reco_rate)+temp['help_received']*const_pds_rate*math.e**(-i_dt*const_pds_rate))**(1-const_ie)-1)*math.e**(-i_dt*const_rho)\n# Indicate k(t): private and public \n\n# Lost income from capital\nplt.fill_between(t_lins,c_t,[i-j for i,j in zip(c_t,dc_k_t)],facecolor=reds_pal[3],alpha=0.45)\nplt.scatter(0,c_t[0]-dc_k_t[0],color=reds_pal[3],zorder=100)\nplt.annotate('Income\\nlosses',[-0.50,(c_t[0]+(c_t[0]-dc_k_t[0]))/2.],fontsize=8,ha='center',va='center')\n\n# Reconstruction costs\nplt.fill_between(t_lins,[i-j for i,j in zip(c_t,dc_k_t)],[i-j-k for i,j,k in zip(c_t,dc_k_t,dc_reco_t)],facecolor=reds_pal[4],alpha=0.45)\nplt.scatter(0,c_t[0]-dc_k_t[0]-dc_reco_t[0],color=reds_pal[4],zorder=100)\nplt.annotate('Reconstruction\\ncosts',[-0.50,((c_t[0]-dc_k_t[0])+(c_t[0]-dc_k_t[0]-dc_reco_t[0]))/2.],fontsize=8,ha='center',va='center')\n\n# PDS\nplt.fill_between(t_lins,c_t,[i+j for i,j in zip(c_t,dc_pds_t)],facecolor=sns_pal[2],alpha=0.45)\nplt.annotate('PDS\\nspend down',[-0.50,(c_t[0]+(c_t[0]+dc_pds_t[0]))/2.],fontsize=8,ha='center',va='center')\n\nplt.plot(t_lins,[i-j-k+l for i,j,k,l in zip(c_t,dc_k_t,dc_reco_t,dc_pds_t)],ls=':',color=reds_pal[8])\n\n# Draw c\nplt.plot([-1,5],[c,c],color=greys_pal[8])\n\nplt.xlim(-1,5)\n#plt.ylim((c-dc0)*0.98,c*1.02)\n\nplt.xlabel(r'Time $t$ after disaster ($\\tau_h \\equiv 3$) [years]')\nplt.ylabel(r'Household consumption ($c_h$)')\nplt.xticks([-1,0,1,2,3,4,5],['-1',r'$t_0$','1','2','3','4','5'])\nplt.yticks([c_t[0]],[r'$c_0$'])\n\nplt.draw()\nfig=plt.gcf()\nfig.savefig('/Users/brian/Desktop/Dropbox/Bank/unbreakable_writeup/Figures/dc.pdf',format='pdf')\n\nplt.clf()\nplt.close('all')\n\n# Draw k\nplt.plot([-1,5],[k,k],color=greys_pal[8])\n\n# points at t0\nplt.scatter(0,k-dk0,color=reds_pal[5],zorder=100)\nplt.scatter(0,k-dkprv,color=reds_pal[3],zorder=100)\n# Annotate \nplt.annotate('Private\\nasset\\nlosses',[-0.65,k-dkprv/2.],fontsize=9,ha='center',va='center')\nplt.annotate(r'$\\Delta k^{prv}_0$',[-0.2,k-dkprv/2.],fontsize=10,ha='center',va='center')\nplt.plot([-0.2,-0.2],[k-dkprv,k-1.1*dkprv/2.],color=reds_pal[3])\nplt.plot([-0.2,-0.2],[k-0.9*dkprv/2.,k],color=reds_pal[3])\nplt.plot([-0.22,-0.18],[k,k],color=reds_pal[3])\nplt.plot([-0.22,-0.18],[k-dkprv*0.997,k-dkprv*0.997],color=reds_pal[3],zorder=100)\n\nplt.annotate('Public\\nasset\\nlosses',[-0.65,k-dk0+dkpub/2.],fontsize=9,ha='center',va='center')\nplt.annotate(r'$\\Delta k^{pub}_0$',[-0.2,k-dk0+dkpub/2.],fontsize=10,ha='center',va='center')\nplt.plot([-0.2,-0.2], [k-dkprv-dkpub,(k-dkprv)-1.5*dkpub/2.],color=reds_pal[5])\nplt.plot([-0.2,-0.2], [(k-dkprv)-0.5*dkpub/2.,k-dkprv],color=reds_pal[5])\nplt.plot([-0.22,-0.18],[k-dkprv*1.003,k-dkprv*1.003],color=reds_pal[5])\nplt.plot([-0.22,-0.18],[k-dkprv-dkpub,k-dkprv-dkpub],color=reds_pal[5])\n\nplt.annotate('Disaster\\n'+r'(t = t$_0$)',[0,k*1.005],fontsize=9,ha='center',weight='bold')\nplt.plot([0,0],[k-dk0,k],color=sns_pal[0])\n\n# k recovery\nk_t = []\ndk0_t = []\ndkprv_t = []\ndkpub_t = []\n\nfor t in t_lins:\n k_t.append(k)\n dk0_t.append(k-(dk0*np.e**(-t*const_dk_reco)))\n dkprv_t.append(k-(dkprv*np.e**(-t*const_dk_reco)))\n dkpub_t.append(k-(dkpub*np.e**(-t*const_dk_reco)))\n\n# Indicate k(t): private and public \nplt.fill_between(t_lins,k_t,dkprv_t,facecolor=reds_pal[3],alpha=0.45)\nplt.fill_between(t_lins,dkprv_t,[i-(k-j) for i,j,k in zip(dkprv_t,dkpub_t,k_t)],facecolor=reds_pal[5],alpha=0.45)\n\nplt.plot([3,3],[k-0.05*dk0,k],color=reds_pal[8])\nplt.plot([2.98,3.02],[k-0.05*dk0,k-0.05*dk0],color=reds_pal[8],zorder=100)\nplt.plot([2.98,3.02],[k,k],color=reds_pal[8],zorder=100)\n\nplt.gca().add_patch(patches.Rectangle((3.45,k-0.12*dk0),1.60,6500,facecolor='white',zorder=98,clip_on=False))\nplt.gca().annotate(r'$\\Delta k_h^{eff}|_{t=\\tau_h}$ = 0.05$\\times\\Delta k_0^{eff}$',\n xy=(3,k-0.025*dk0), xycoords='data',\n xytext=(3.5,k-0.075*dk0), textcoords='data', fontsize=10,\n arrowprops=dict(arrowstyle=\"->\",connectionstyle=\"arc3,rad=-0.05\",lw=1.5),\n ha='left',va='center',zorder=99)\n\nplt.plot(t_lins,dk0_t,color=reds_pal[8],ls='--',lw=0.75)\nplt.gca().add_patch(patches.Rectangle((1.40,dk0_t[10]*1.005),1.70,7000,facecolor='white',zorder=98))\nplt.gca().annotate(r'$\\Delta k_h^{eff}(t) = \\Delta k_0^{eff}e^{-R_{\\tau}\\cdot t}$',\n xy=(t_lins[20],dk0_t[20]), xycoords='data',\n xytext=(1.45,dk0_t[10]*1.01), textcoords='data', fontsize=12,\n arrowprops=dict(arrowstyle=\"->\",connectionstyle=\"arc3,rad=0.2\",lw=1.5),\n ha='left',va='center',zorder=99)\n\nplt.xlim(-1,5)\nplt.ylim((k-dk0)*0.98,k*1.02)\n\nplt.xlabel(r'Time $t$ after disaster ($\\tau_h \\equiv 3$) [years]')\n#plt.xlabel(r'Time $t$ after disaster (years)')\nplt.ylabel(r'Effective household capital ($k_h^{eff}$)')\nplt.xticks([-1,0,1,2,3,4,5],['-1',r'$t_0$','1','2','3','4','5'])\nplt.yticks([k_t[0]],[r'$k_h^{eff}$'])\n\nplt.draw()\nfig=plt.gcf()\nfig.savefig('/Users/brian/Desktop/Dropbox/Bank/unbreakable_writeup/Figures/dk.pdf',format='pdf')\n\n","sub_path":"pp_dc_elements.py","file_name":"pp_dc_elements.py","file_ext":"py","file_size_in_byte":8118,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"337721136","text":"from unittest.mock import ANY\nimport pytest\n\nfrom reconcile.blackbox_exporter_endpoint_monitoring import (\n queries, get_endpoints, build_probe, parse_prober_url, fill_desired_state)\nfrom reconcile.utils.openshift_resource import ResourceInventory\nfrom .fixtures import Fixtures\n\n\nfxt = Fixtures('blackbox_exporter_endpoint_monitoring')\n\n\ndef get_endpoint_fixtures(path: str) -> dict:\n return fxt.get_anymarkup(path)[\"appInterface\"][\"apps\"]\n\n\ndef test_invalid_endpoints(mocker):\n query = mocker.patch.object(queries, 'get_service_monitoring_endpoints')\n query.return_value = get_endpoint_fixtures(\"test_invalid_endpoints.yaml\")\n\n endpoints = get_endpoints()\n assert len(endpoints) == 0\n\n\ndef test_endpoint_loading(mocker):\n ep_query = mocker.patch.object(queries, 'get_service_monitoring_endpoints')\n ep_query.return_value = get_endpoint_fixtures(\"test_endpoint.yaml\")\n\n endpoints = get_endpoints()\n assert len(endpoints) == 1\n\n provider = list(endpoints.keys())[0]\n assert provider.provider == \"blackbox-exporter\"\n ep = endpoints.get(provider)[0]\n assert len(ep.monitoring) == 1\n\n\ndef test_parse_prober_url():\n assert parse_prober_url(\"http://host:1234/path\") == {\n \"url\": \"host:1234\",\n \"scheme\": \"http\",\n \"path\": \"/path\"\n }\n\n assert parse_prober_url(\"http://host\") == {\n \"url\": \"host\",\n \"scheme\": \"http\"\n }\n\n\ndef test_invalid_prober_url():\n # scheme missing\n with pytest.raises(ValueError):\n parse_prober_url(\"host:1234/path\")\n\n\ndef test_probe_building(mocker):\n ep_query = mocker.patch.object(queries, 'get_service_monitoring_endpoints')\n ep_query.return_value = get_endpoint_fixtures(\"test_probe_building.yaml\")\n\n endpoints = get_endpoints()\n assert len(endpoints) == 1\n\n provider = list(endpoints.keys())[0]\n probe_resource = build_probe(provider, endpoints.get(provider))\n assert probe_resource is not None\n\n # verify prober url decomposition\n spec = probe_resource.body.get(\"spec\")\n assert spec.get(\"prober\") == {\n \"url\": \"exporterhost:9115\",\n \"scheme\": \"http\",\n \"path\": \"/probe\"\n }\n\n # verify labels\n labels = spec[\"targets\"][\"staticConfig\"][\"labels\"]\n assert labels.get(\"environment\") == \"staging\"\n\n # verify timeout and interval\n assert spec[\"scrapeTimeout\"] == provider.timeout\n assert spec[\"interval\"] == provider.checkInterval\n\n # verify targets\n assert \"https://test1.url\" in spec[\"targets\"][\"staticConfig\"][\"static\"]\n assert \"https://test2.url\" in spec[\"targets\"][\"staticConfig\"][\"static\"]\n\n\ndef test_filling_desired_state(mocker):\n ep_query = mocker.patch.object(queries, 'get_service_monitoring_endpoints')\n ep_query.return_value = get_endpoint_fixtures(\"test_endpoint.yaml\")\n add_desired_mock = mocker.patch.object(ResourceInventory, 'add_desired')\n\n endpoints = get_endpoints()\n provider = list(endpoints.keys())[0]\n fill_desired_state(provider, endpoints[provider], ResourceInventory())\n\n assert add_desired_mock.call_count == 1\n add_desired_mock.assert_called_with(\n cluster=\"app-sre-stage-01\",\n namespace=\"openshift-customer-monitoring\",\n resource_type=\"Probe\",\n name=\"blackbox-exporter-http-2xx\",\n value=ANY\n )\n\n\ndef test_loading_multiple_providers_per_endpoint(mocker):\n ep_query = mocker.patch.object(queries, 'get_service_monitoring_endpoints')\n ep_query.return_value = \\\n get_endpoint_fixtures(\"test_multiple_providers_per_endpoint.yaml\")\n endpoints = get_endpoints()\n\n assert len(endpoints) == 2\n\n for provider, eps in endpoints.items():\n assert provider.provider == \"blackbox-exporter\"\n assert len(eps) == 2\n","sub_path":"reconcile/test/test_blackbox_exporter_endpoint_monitoring.py","file_name":"test_blackbox_exporter_endpoint_monitoring.py","file_ext":"py","file_size_in_byte":3718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"98849299","text":"\r\nimport json\r\nimport numpy as np\r\nimport pandas as pd\r\nimport requests\r\nimport sqlite3\r\n\r\nNHL_API = 'https://statsapi.web.nhl.com/api/v1/{}'\r\nNHL_DB = sqlite3.connect('nhl.db')\r\n\r\nlambda_get = lambda x: requests.get(NHL_API.format(x)).json()\r\nlambda_df = lambda y: pd.concat([ pd.io.json.json_normalize(z,sep='_') for z in y ],sort=1)\r\n\r\ndef refresh_table(table,spec,df):\r\n print('Starting table refresh for {}...'.format(table))\r\n df = df[spec['columns']]\r\n df = df.rename(columns=spec['rename_headers'])\r\n df.to_sql(\r\n name=table,\r\n con=NHL_DB,\r\n if_exists='replace',\r\n index=0,\r\n dtype=spec['cast_dtypes']\r\n )\r\n print('Finished table refresh for {}.'.format(table))\r\n\r\ndef request_api(table,spec):\r\n print('Accessing NHL API for {} data...'.format(table))\r\n if spec['standard_refresh']:\r\n endpoint = spec['api_endpoint']\r\n get = lambda_get(endpoint)[endpoint]\r\n df = lambda_df(get)\r\n else:\r\n func_switch = {\r\n 'player': refresh_player,\r\n 'game': refresh_game,\r\n 'gamelog': refresh_gamelog,\r\n 'standings': refresh_standings\r\n }\r\n df = func_switch[table](table,spec)\r\n return { 'table': table, 'spec': spec, 'df': df }\r\n\r\ndef refresh_player(table,spec):\r\n sql = 'SELECT team_id FROM team;'\r\n team_ids = [ t[0] for t in NHL_DB.cursor().execute(sql).fetchall() ]\r\n endpoints = [ 'teams/{}?expand=team.roster'.format(t) for t in team_ids ]\r\n rosters = [ lambda_get(e)['teams'][0]['roster']['roster'] for e in endpoints ]\r\n df_rosters = lambda_df(rosters)\r\n\r\n player_ids = df_rosters['person_id'].tolist()\r\n endpoints = [ spec['api_endpoint'].format(p) for p in player_ids ]\r\n players = [ lambda_get(e)['people'] for e in endpoints ]\r\n return lambda_df(players)\r\n\r\ndef refresh_game(table,spec): #only refresh games that have changed status? \r\n [current_season] = lambda_get('seasons/current')['seasons']\r\n dates = lambda_get(spec['api_endpoint'].format( **current_season))['dates']\r\n games = [ d['games'] for d in dates ]\r\n return lambda_df(games)\r\n\r\ndef refresh_gamelog(table,spec):\r\n sql = 'SELECT player_id FROM player;'\r\n player_ids = [ p[0] for p in NHL_DB.cursor().execute(sql).fetchall() ]\r\n\r\n df_gamelogs = pd.DataFrame()\r\n for p in player_ids:\r\n endpoint = spec['api_endpoint'].format(p)\r\n splits = lambda_get(endpoint)['stats'][0]['splits']\r\n if len(splits) == 0:\r\n continue\r\n df_splits = lambda_df(splits)\r\n df_splits['player_id'] = p\r\n df_gamelogs = df_gamelogs.append(df_splits,sort=1)\r\n return df_gamelogs\r\n\r\ndef refresh_standings(table,spec):\r\n standings = lambda_get(spec['api_endpoint'])['records']\r\n df_standings = pd.concat([ lambda_df(s['teamRecords']) for s in standings ])\r\n return df_standings\r\n\r\ndef main():\r\n with open('table_specs.json') as f:\r\n table_specs = json.load(f).items()\r\n full_refresh = [ refresh_table( **request_api(k,v)) for k,v in table_specs ]\r\n\r\nif __name__ == '__main__':\r\n main()\r\n","sub_path":"refresh_db.py","file_name":"refresh_db.py","file_ext":"py","file_size_in_byte":3111,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"149420688","text":"import requests\nfrom django.shortcuts import render\nfrom .models import City\nfrom .forms import CityForm\nimport requests\nimport json\nfrom math import sin, cos, sqrt, atan2, radians\n\ndef index(request):\n url = 'http://api.openweathermap.org/data/2.5/weather?q={}&units=imperial&appid=20e84c6dbae34e0d6ece4c8f8d0f4314'\n\n search = None\n if request.method == 'POST':\n search = request.POST.get('name','').lower() # https://stackoverflow.com/questions/4162625/django-request-get-parameters\n # Doesn't save to database, but still loads from database.\n #form.save()\n\n url2 = 'https://records.nhl.com/site/api/franchise'\n response2 = requests.get(url2).json()\n team_dicts = response2['data']\n\n team_id = None\n for d in team_dicts:\n if search == 'montreal':\n team_id = 8\n break \n \n if((d['teamPlaceName'].lower() == search or \n d['teamCommonName'].lower() == search or \n d['teamPlaceName'].lower() + ' ' +d['teamCommonName'].lower() == search) and\n d['lastSeasonId'] == None):\n team_id = d['mostRecentTeamId']\n \n team_string = None\n game_date_string = None\n game_city = None\n game_venue = None\n if team_id != None: \n # get next game on schedule \n next_game_url = 'https://statsapi.web.nhl.com/api/v1/teams/' + str(team_id) + '/?expand=team.schedule.next'\n game_sched = requests.get(next_game_url).json()\n\n # get versus teams \n teams = game_sched['teams'][0]['nextGameSchedule']['dates'][0]['games'][0]['teams']\n team_string = teams['away']['team']['name'] + ' versus ' + teams['home']['team']['name']\n \n # get game date\n game_date = game_sched['teams'][0]['nextGameSchedule']['dates'][0]['games'][0]\n game_date_string = game_date['gameDate']\n \n # get game location \n game_city = game_sched['teams'][0]['venue']['city']\n game_venue = game_sched['teams'][0]['venue']['name']\n \n print('------------------------')\n print(game_city)\n print(game_venue)\n print(game_date_string)\n print(team_string)\n print('----------------------')\n \n # team data\n team_data_url = 'https://records.nhl.com/site/api/team'\n team_response = requests.get(team_data_url).json()\n team_stats = team_response['data']\n \n team_data_url = 'https://records.nhl.com/site/api/team'\n team_response = requests.get(team_data_url).json()\n team_stats = team_response['data']\n \n displayTeamName = \"NHL League\"\n ticketsURL = \"https://www.nhl.com/tickets\"\n arenaCoords = None\n distance = \"\"\n abv = \"NHL\"\n instagram = \"nhl\"\n twitter = \"nhl\"\n for team in team_stats:\n if (team['teamId']==team_id):\n print(\"Team Name: \" + team['fullName'] + \" Active: \" + team['active'] + \" Arena Address: \" + team['arenaAddress'])\n displayTeamName = team['fullName']\n ticketsURL = team['buySellTicketUrl']\n arenaCoords = team['arenaCoordinates'].split(\",\")\n abv = team['triCode']\n instagram = team['instagram']\n twitter = team['twitter']\n\n #https://stackoverflow.com/questions/45630606/can-i-get-accurate-geolocation-in-python-using-html5-location-tool\n #https://stackoverflow.com/questions/19412462/getting-distance-between-two-points-based-on-latitude-longitude\n\n response_data = requests.get('https://www.iplocation.net/go/ipinfo').text\n response_json_data = json.loads(response_data)\n location = response_json_data[\"loc\"].split(\",\")\n print(arenaCoords[1])\n R = 6373.0\n lat1 = radians(float(location[0]))\n lon1 = radians(float(location[1]))\n lat2 = radians(float(arenaCoords[0]))\n lon2 = radians(float(arenaCoords[1]))\n dlon = lon2 - lon1\n dlat = lat2 - lat1\n a = sin(dlat / 2)**2 + cos(lat1) * cos(lat2) * sin(dlon / 2)**2\n c = 2 * atan2(sqrt(a), sqrt(1 - a))\n distance = R * c\n distance = round(distance,2)\n distance = \"Distance from Current Location to Arena: \" + str(distance) + \" KM\"\n print(\"Result:\", distance)\n\n\n\n #print(response2)\n\n\n #final = r2['data'][0]['data']\n\n #print(final)\n\n\n form = CityForm()\n\n cities = City.objects.all()\n\n weather_data = []\n\n for city in cities:\n\n r = requests.get(url.format(city)).json()\n\n city_weather = {\n 'city' : displayTeamName,\n 'temperature' : distance,\n 'description' : ticketsURL,\n 'icon' : abv,\n 'instagram' : instagram,\n 'twitter' : twitter,\n 'game_city' : game_city,\n 'game_venue' : game_venue,\n 'game_date_string' : game_date_string,\n 'team_string' : team_string\n }\n\n weather_data.append(city_weather)\n\n context = {'weather_data' : weather_data, 'form' : form}\n return render(request, 'weather/weather.html', context)\n","sub_path":"the_weather/weather/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5133,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"16272217","text":"valid_commands = ['attack','defend']\n\ndef subtract(value, varflow, player, target):\n \"\"\"Subtract value from vars in varflow.\n All vars will stay above 0.\"\"\"\n curval = value\n for varkey in varflow:\n if target['vars'][varkey]-curval < 0:\n curval -= target['vars'][varkey]\n target['vars'][varkey] = 0\n else:\n target['vars'][varkey] -= curval\n break\n\ndef extract_players(game, playerkey, targetkey):\n \"\"\"Get references to player and target\"\"\"\n return game['players'][playerkey], game['players'][targetkey]\n\ndef check_end_game(game):\n \"\"\"Set winner if game over\"\"\"\n if len(game['players'])-len(game['graveyard']) == 1:\n game['winner'] = list(set(game['players'].keys())-set(game['graveyard']))[0]\n elif len(game['players']) == len(game['graveyard']):\n game['winner'] = 'draw'\n\ndef incrementKillCount(game, player, targetkey):\n \"\"\"Add experience to the player when the target\n is not already in the graveyard.\"\"\"\n if targetkey not in game['graveyard']:\n # target not already dead\n player['exp'] += 1\n\n# Command functions:\n\ndef attack(game, playerkey, targetkey):\n \"\"\"Attack command\"\"\"\n player, target = extract_players(game, playerkey, targetkey)\n\n subtract(player['stats']['attack'], ['defense','health'], player, target)\n\n if target['vars']['health'] <= 0:\n incrementKillCount(game, player, targetkey)\n\n s = set(game['graveyard']) # prevent duplicates\n s.add(targetkey)\n game['graveyard'] = list(s)\n\n check_end_game(game)\n\ndef defend(game, playerkey, targetkey):\n \"\"\"Defend command\"\"\"\n player, target = extract_players(game, playerkey, targetkey)\n\n target['vars']['defense'] += player['stats']['defense']\n","sub_path":"server/src/serverlogic/gamecommands.py","file_name":"gamecommands.py","file_ext":"py","file_size_in_byte":1775,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"410885213","text":"# Dank Memer\n# Utilities -\n# Random Joke\n# Guess the number - simple game\n# A random meme- Hindi / English\nimport random\nimport requests\nfrom PIL import Image\nimport pandas as pd\nimport time\nimport os\n\ndef runDankMemer():\n print(\"Hello Welcome to this bot :)\\n\")\n status=1\n while(status):\n line()\n userInput=int(input(\"To play guess the number game, enter 1\\n\"\n \"To listen a random joke, enter 2\\n\"\n \"To see a meme, enter 3\\n\"\n \"To get a riddle, enter 4\\n\"\n \"To exit from Dank Memer, enter 5\\n\"))\n line()\n if(userInput==1):\n guessNumber()\n elif(userInput==2):\n getJoke()\n elif(userInput==3):\n getMeme()\n elif(userInput==4):\n getRiddle()\n elif (userInput == 5):\n print(\"Thank You :)\")\n status=0\n else:\n print(\"Invalid Input\")\n line()\n\ndef guessNumber() :\n line()\n status = 1\n while (status):\n print(\"Guess a number between 1 to 100 \\n\")\n n = random.randint(1, 100)\n flag = 1\n while (flag):\n guess = int(input(\"Enter your guess :\"))\n if (guess < n):\n print(\"oops !! Think of a greater one\")\n continue\n if (guess > n):\n print(\"oops !! Think of a lesser one\")\n continue\n if (guess == n):\n print(\"You Won !!!\")\n break\n line()\n status = int(input(\"Do you want to play again ? Enter 1 to play or 0 to exit .\"))\n print(\"Thank You .\")\ndef getJoke():\n line()\n status = 1\n while (status):\n response = requests.get(\"https://v2.jokeapi.dev/joke/Any?blacklistFlags=religious,racist,sexist&type=single\")\n print(response.json()['joke'])\n print()\n time.sleep(4)\n status = int(input(\"Do you want to listen a joke again ? Enter 1 to play or 0 to exit .\"))\n print()\n line()\n print(\"Thank You .\")\ndef getMeme():\n line()\n status = 1\n while (status):\n n = str(random.randint(1, 18))\n path=os.path.abspath('../DankMemer/memes/'+n+'.jpg')\n # image = Image.open(\"memes/\" + n + \".jpg\")\n image = Image.open(path)\n image.show()\n status = int(input(\"Do you want to see one more ? Enter 1 to play or 0 to exit .\"))\n line()\n print(\"Thank You .\")\ndef getRiddle():\n line()\n status=1\n while(status):\n n = random.randint(0, 19)\n print(\"You have 15 seconds to answer the riddle : Answer will be shown automatically, do not enter anything\")\n time.sleep(3)\n path=os.path.abspath('../DankMemer/riddles.csv')\n data = pd.read_csv(path, index_col=False)\n print(\"Riddle : \" + data['question'][n])\n time.sleep(15)\n print(\"Time Up !!! Answer is : \" + data['answer'][n])\n status=int(input(\"Do you want to play again ? Enter 1 to play or 0 to exit .\"))\n line()\n print(\"Thank You .\")\ndef line():\n print()\n for i in range(10):\n print(\"*************\",end=\"\")\n print(\"\")\n\n\n\n\n\n\n\n\n\n","sub_path":"DankMemer/dankmemer.py","file_name":"dankmemer.py","file_ext":"py","file_size_in_byte":3258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"114156269","text":"from .layers import Time, Difference, LRS\n\nimport numpy as np\nimport tensorflow as tf\ntf.logging.set_verbosity(tf.logging.ERROR)\n\nfrom tensorflow.keras import Sequential\nfrom tensorflow.keras.layers import InputLayer, Conv1D, Activation, Reshape, Dense, LayerNormalization, BatchNormalization, Lambda, GlobalAveragePooling1D, TimeDistributed\n\ndef init_nslr_model(preprocess_size, lrs_size, lrs_levels, preprocess='conv', preprocess_time=False, lrs_diff=True, lrs_time=True, recursive_tensors=True, name_only=False, input_shape=None, num_classes=None, lrs_norm='BN'):\n \n preprocess = preprocess.lower()\n \n if preprocess != 'conv' and preprocess != 'dense':\n preprocess_name = ''\n else:\n preprocess_name = 'Conv' if preprocess == 'conv' else 'Dense'\n preprocess_name = 'T{}'.format(preprocess_name) if preprocess_time else preprocess_name\n \n \n lrs_time_tag = 'T' if lrs_time else ''\n lrs_diff_tag = 'D' if lrs_diff else ''\n lrs_recurrent_tag = 'R' if recursive_tensors else ''\n lrs_name = '{}{}{}LRS{}'.format(lrs_time_tag, lrs_diff_tag, lrs_recurrent_tag, lrs_norm)\n \n model_name = '{}{}_H{}_W{}'.format(preprocess_name, lrs_name, preprocess_size, lrs_size)\n \n if name_only: return model_name\n \n num_sig_layers = 3\n \n model = Sequential()\n\n model.add(InputLayer(input_shape=input_shape))\n \n if preprocess_name != '':\n if preprocess_time:\n model.add(Time())\n if preprocess == 'conv':\n model.add(Conv1D(preprocess_size, 8, padding='same', kernel_initializer='he_uniform'))\n else:\n model.add(TimeDistributed(Dense(preprocess_size, kernel_initializer='he_uniform')))\n model.add(BatchNormalization(axis=-1))\n model.add(Activation('relu'))\n\n\n if preprocess_time:\n model.add(Time())\n if preprocess == 'conv':\n model.add(Conv1D(preprocess_size, 5, padding='same' , kernel_initializer='he_uniform'))\n else:\n model.add(TimeDistributed(Dense(preprocess_size, kernel_initializer='he_uniform')))\n model.add(BatchNormalization(axis=-1))\n model.add(Activation('relu'))\n\n\n if preprocess_time:\n model.add(Time())\n if preprocess == 'conv':\n model.add(Conv1D(preprocess_size, 3, padding='same', kernel_initializer='he_uniform'))\n else:\n model.add(Dense(preprocess_size, kernel_initializer='he_uniform'))\n model.add(BatchNormalization(axis=-1))\n model.add(Activation('relu'))\n \n \n for i in range(num_sig_layers-1):\n# model.add(Conv1D(conv_size, 3, padding='same', kernel_initializer='he_uniform'))\n# model.add(BatchNormalization(axis=-1))\n if lrs_time:\n model.add(Time())\n if lrs_diff:\n model.add(Difference())\n model.add(LRS(lrs_size, lrs_levels, return_sequences=True, recursive_tensors=recursive_tensors))\n model.add(Reshape((input_shape[0]-i-1, lrs_levels, lrs_size,)))\n if lrs_norm == 'BN':\n model.add(BatchNormalization(axis=[-2]))\n elif lrs_norm == 'TBN':\n model.add(BatchNormalization(axis=[-2, -1]))\n elif lrs_norm == 'LN':\n model.add(LayerNormalization(axis=[-3, -1]))\n# model.add(LayerNormalization(axis=[1, 3]))\n#model.add(LayerNormalization(axis=[1, 3]))\n model.add(Reshape((input_shape[0]-i-1, lrs_levels * lrs_size,)))\n# model.add(Activation('relu'))\n# if renorm:\n# model.add(BatchNormalization(axis=[1, 2], renorm=True, center=False, scale=False))\n# else:\n \n# \n# model.add(Conv1D(conv_size, 3, padding='same', kernel_initializer='he_uniform'))\n# model.add(BatchNormalization(axis=-1))\n if lrs_time:\n model.add(Time())\n if lrs_diff:\n model.add(Difference())\n model.add(LRS(lrs_size, lrs_levels, return_sequences=False, recursive_tensors=recursive_tensors))\n# model.add(BatchNormalization(axis=1, center=False, scale=False))\n model.add(Reshape((lrs_levels, lrs_size,)))\n# model.add(LayerNormalization(axis=[2]))\n model.add(BatchNormalization(axis=1))\n model.add(Reshape((lrs_levels * lrs_size,)))\n# model.add(BatchNormalization(axis=-1, renorm=renorm, center=False, scale=False))\n\n# model.add(tf.keras.layers.Dropout(0.5))\n# if gap:\n# model.add(GlobalAveragePooling1D())\n# else:\n# model.add(Lambda(lambda X: X[:, -1]))\n \n model.add(Dense(num_classes, activation='softmax'))\n \n model._name = model_name\n\n return model","sub_path":"tsc/models/.ipynb_checkpoints/init_nslr_model-checkpoint.py","file_name":"init_nslr_model-checkpoint.py","file_ext":"py","file_size_in_byte":4586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"53457194","text":"__author__ = \"Luke Liu\"\n#encoding=\"utf-8\"\n\n# 引入需要使用的库文件\nimport cv2\nimport matplotlib.pyplot as plt\nfrom PIL import Image\nimport os\nimport numpy as np\n\n#创建图像拼接类\nclass ImageStitch:\n '''进行图像拼接'''\n def stitch(self, images, ratio=0.75, reprojThresh=4.0,showMatches=False):\n #获取输入图片\n (imageB, imageA) = images\n #检测A、B图片的SIFT关键特征点,并计算特征描述子\n (kpsA, featuresA) = self.detectAndDescribe(imageA)\n (kpsB, featuresB) = self.detectAndDescribe(imageB)\n\n # 匹配两张图片的所有特征点,返回匹配结果\n M = self.matchKeypoints(kpsA, kpsB, featuresA, featuresB, ratio, reprojThresh)\n\n # 如���返回结果为空,没有匹配成功的特征点,退出算法\n if M is None:\n return None\n\n # 否则,提取匹配结果\n # H是3x3视角变换矩阵\n (matches, H, status) = M\n # 将图片A进行视角变换,result是变换后图片\n result = cv2.warpPerspective(imageA, H, (imageA.shape[1] + imageB.shape[1], imageA.shape[0]))\n print(\"result shape is \",result.shape)\n self.cv_show('result', result)\n\n # 将图片B传入result图片最左端\n result[0:imageB.shape[0], 0:imageB.shape[1]] = imageB\n self.cv_show('result', result)\n\n #阈值化处理\n result_gray = cv2.cvtColor(result,cv2.COLOR_BGR2GRAY)\n ret,thresh = cv2.threshold(result_gray,0,255,cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n self.cv_show('binary',thresh)\n\n thresh_, threshCnts, hierarchy = cv2.findContours(thresh.copy(), cv2.RETR_EXTERNAL,\n cv2.CHAIN_APPROX_SIMPLE)\n cnts = threshCnts\n cur_img = result.copy()\n cv2.drawContours(cur_img, cnts, -1, (0, 0, 255), 3)\n self.cv_show(\"\",cur_img)\n boundlist = [cv2.boundingRect(c) for c in cnts ]\n boundlist=sorted(boundlist,key=lambda x:x[2],reverse=True)\n print(boundlist[:10])\n for c in cnts:\n (x,y,w,h)=cv2.boundingRect(c)\n cv2.rectangle(cur_img,(x,y),(x+w,y+h),(0,255,0),2)\n self.cv_show(\"\", cur_img)\n maxheight=image1.shape[0]\n maxWidth=boundlist[0][2]+1\n result_roi = result[0:maxheight,0:maxWidth]\n\n self.cv_show(\"Final\",result_roi)\n\n\n # 检测是否需要显示图片匹配\n if showMatches:\n # 生成匹配图片\n vis = self.drawMacthes(imageA, imageB, kpsA, kpsB, matches, status)\n\n\n # 返回结果\n return (result_roi, vis)\n # 返回匹配结果\n return result_roi\n\n\n '''显示这个图像'''\n def cv_show(self,name,img):\n cv2.imshow(name, img)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n\n ''' 寻找检测SIFT特征点与特征描述向量'''\n def detectAndDescribe(self,image):\n # 将彩色图片转换成灰度图\n gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n # 建立SIFT生成器\n descriptor = cv2.xfeatures2d.SIFT_create()\n # 检测SIFT特征点,并计算描述子\n (kps, features) = descriptor.detectAndCompute(image, None)\n\n # 将结果转换成NumPy数组\n kps = np.float32([kp.pt for kp in kps])\n\n # 返回特征点集,及对应的描述特征\n return (kps, features)\n\n '''特征点匹配'''\n def matchKeypoints(self, kpsA, kpsB, featuresA, featuresB, ratio, reprojThresh):\n # 建立暴力匹配器\n matcher = cv2.BFMatcher()\n\n # 使用KNN检测来自A、B图的SIFT特征匹配对,K=2\n rawMatches = matcher.knnMatch(featuresA, featuresB, 2)\n\n matches = []\n for m in rawMatches:\n # 当最近距离跟次近距离的比值小于ratio值时,保留此匹配对\n if len(m) == 2 and m[0].distance < m[1].distance * ratio:\n # 存储两个点在featuresA, featuresB中的索引值\n matches.append((m[0].trainIdx, m[0].queryIdx))\n\n # 当筛选后的匹配对大于4时,计算视角变换矩阵\n if len(matches) > 4:\n # 获取匹配对的点坐标\n ptsA = np.float32([kpsA[i] for (_, i) in matches])\n ptsB = np.float32([kpsB[i] for (i, _) in matches])\n\n # 计算视角变换矩阵\n (H, status) = cv2.findHomography(ptsA, ptsB, cv2.RANSAC, reprojThresh)\n\n # 返回结果\n return (matches, H, status)\n\n # 如果匹配对小于4时,返回None\n return None\n\n\n '''画出匹配的特征点'''\n def drawMacthes(self,imageA, imageB, kpsA, kpsB, matches, status):\n # 初始化可视化图片,将A、B图左右连接到一起\n (hA, wA) = imageA.shape[:2]\n (hB, wB) = imageB.shape[:2]\n vis = np.zeros((max(hA, hB), wA + wB, 3), dtype=\"uint8\")\n vis[0:hA, 0:wA] = imageA\n vis[0:hB, wA:] = imageB\n\n # 联合遍历,画出匹配对\n for ((trainIdx, queryIdx), s) in zip(matches, status):\n # 当点对匹配成功时,画到可视化图上\n if s == 1:\n # 画出匹配对\n ptA = (int(kpsA[queryIdx][0]), int(kpsA[queryIdx][1]))\n ptB = (int(kpsB[trainIdx][0]) + wA, int(kpsB[trainIdx][1]))\n cv2.line(vis, ptA, ptB, (0, 255, 0), 1)\n self.cv_show(\"KeyPoints Match\",vis)\n cv2.imwrite(\"KeyPoins_match.png\",vis)\n\n\n\nif __name__ == '__main__':\n image1=cv2.imread(\"1.jpg\")\n image1=cv2.resize(image1,(int(0.5*image1.shape[1]),int(0.5*image1.shape[0])))\n print(\"image1 shape is :\",image1.shape)\n\n image2 =cv2.imread(\"2.jpg\")\n image2 = cv2.resize(image2, (int(0.5 * image2.shape[1]), int(0.5 * image2.shape[0])))\n print(\"image2 shape is :\",image2.shape)\n stitcher = ImageStitch()\n (result,vis) = stitcher.stitch([image1,image2],showMatches=True)\n cv2.imwrite(\"Stitch_outcome.jpg\",result)\n\n cv2.destroyAllWindows()\n","sub_path":"利用特征点匹配进行全景图拼接/ImageStitch.py","file_name":"ImageStitch.py","file_ext":"py","file_size_in_byte":6038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"45625504","text":"\"\"\"\n@author: Jose Antonio Cervan Garcia\n\nFactorial digit sum\n \nProblem 20\nn! means n × (n − 1) × ... × 3 × 2 × 1\n\nFor example, 10! = 10 × 9 × ... × 3 × 2 × 1 = 3628800,\nand the sum of the digits in the number 10! is 3 + 6 + 2 + 8 + 8 + 0 + 0 = 27.\n\nFind the sum of the digits in the number 100!\n\"\"\"\n\nimport math\n\nn = 100\n\ndigits = str(math.factorial(100))\n\nresult = 0\n\nfor i in range(len(digits)):\n result += int(digits[i])\n\nprint(\"result = \", result)","sub_path":"problems/#0020.py","file_name":"#0020.py","file_ext":"py","file_size_in_byte":468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"190951987","text":"n = int(input())\narr1 = [int(i) for i in input().split(\" \")]\nm = int(input())\narr2 = [int(i) for i in input().split(\" \")]\ncounts1 = [0]*101\ncounts2 = [0]*101\nminimum1 = min(arr1)\nfor i in arr1:\n counts1[i-minimum1] += 1\n\nfor i in arr2:\n counts2[i-minimum1] += 1\n\nresult = set()\nfor i in range(0,101):\n if counts1[i] != counts2[i]:\n result.add(i)\nresult = sorted(result)\nresult = [minimum1+i for i in result]\nprint(\" \".join(map(str,result)))","sub_path":"Algorithms/Search/Missing Numbers.py","file_name":"Missing Numbers.py","file_ext":"py","file_size_in_byte":456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"310562482","text":"import sys\n\nimport collections\nfrom pprint import pprint as pp\nfrom typing import List\n\n\nclass NumberOfIsland(object):\n \"\"\"\n https://leetcode.com/problems/number-of-islands/\n \"\"\"\n LAND, WATER = '1', '0'\n NEIGHBORS = ((1, 0), (0, -1), (-1, 0), (0, 1))\n\n @classmethod\n def bfs(cls, grid: List[List[str]]) -> int:\n \"\"\"\n Time: O(m*n)\n Space: O(m*n)\n \"\"\"\n def bfs(r, c):\n if grid[r][c] == cls.WATER or visits[r][c]:\n return 0\n\n visits[r][c] = True\n q = collections.deque()\n q.append((r, c))\n while q:\n r, c = q.popleft()\n for neigbor in cls.NEIGHBORS:\n nr, nc = r+neigbor[0], c+neigbor[1]\n\n if nr < 0 or nr >= row or nc < 0 or nc >= col:\n continue\n\n if grid[nr][nc] == cls.WATER or visits[nr][nc]:\n continue\n\n visits[nr][nc] = True\n\n q.append((nr, nc))\n\n return 1\n\n if not grid or not grid[0]:\n return 0\n\n row, col = len(grid), len(grid[0])\n area_cnt = 0\n visits = [[False for _ in range(col)] for _ in range(row)]\n\n for r in range(row):\n for c in range(col):\n area_cnt += bfs(r, c)\n\n return area_cnt\n\n @classmethod\n def dfs(cls, grid: List[List[str]]) -> int:\n \"\"\"\n Time: O(m*n)\n Space: O(m*n)\n \"\"\"\n def dfs(r, c):\n if grid[r][c] == cls.WATER or visits[r][c]:\n return 0\n\n visits[r][c] = True\n stack = list()\n stack.append((r, c))\n while stack:\n r, c = stack.pop()\n\n for neigbor in cls.NEIGHBORS:\n nr, nc = r+neigbor[0], c+neigbor[1]\n\n if nr < 0 or nr >= row or nc < 0 or nc >= col:\n continue\n\n if grid[nr][nc] == cls.WATER or visits[nr][nc]:\n continue\n\n visits[nr][nc] = True\n\n stack.append((nr, nc))\n\n return 1\n\n if not grid or not grid[0]:\n return 0\n\n row, col = len(grid), len(grid[0])\n area_cnt = 0\n visits = [[False for _ in range(col)] for _ in range(row)]\n\n for r in range(row):\n for c in range(col):\n area_cnt += dfs(r, c)\n\n return area_cnt\n\n @classmethod\n def dfs_recursive(cls, grid: List[List[str]]) -> int:\n \"\"\"\n Time: O(m*n)\n Space: O(m*n)\n \"\"\"\n def _dfs(r, c):\n # check if outouf boundary\n if r < 0 or r >= row or c < 0 or c >= col:\n return 0\n\n if grid[r][c] == cls.WATER or visits[r][c]:\n return 0\n\n visits[r][c] = True\n\n for neigbor in cls.NEIGHBORS:\n _dfs(r+neigbor[0], c+neigbor[1])\n\n return 1\n\n if not grid or not grid[0]:\n return 0\n\n row, col = len(grid), len(grid[0])\n area_cnt = 0\n visits = [[False for _ in range(col)] for _ in range(row)]\n\n for r in range(row):\n for c in range(col):\n area_cnt += _dfs(r, c)\n\n return area_cnt\n\n\ndef main():\n grid1 = [[\"1\",\"1\",\"1\",\"0\",\"0\"],[\"0\",\"0\",\"1\",\"1\",\"1\"],[\"0\",\"0\",\"1\",\"0\",\"0\"]]\n grid2 = [[\"1\",\"1\",\"0\",\"0\",\"0\"],[\"1\",\"1\",\"0\",\"0\",\"0\"],[\"0\",\"0\",\"1\",\"0\",\"0\"],[\"0\",\"0\",\"0\",\"1\",\"1\"]]\n print(NumberOfIsland.bfs(grid1))\n print(NumberOfIsland.dfs(grid1))\n print(NumberOfIsland.dfs_recursive(grid1))\n print(NumberOfIsland.bfs(grid2))\n print(NumberOfIsland.dfs(grid2))\n print(NumberOfIsland.dfs_recursive(grid2))\n\nif __name__ == '__main__':\n sys.exit(main())\n","sub_path":"common/algo/bfs_dfs.py","file_name":"bfs_dfs.py","file_ext":"py","file_size_in_byte":3831,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"376166840","text":"# NOTE\n# Duplicates cannot be handled within O(log n) time at all so the following implementation does not accommodate!\n\n\ndef min_sorted_rotated_array(arr, l, r):\n \"\"\"\n Python program to find minimum element in a sorted and rotated array.\n\n https://www.geeksforgeeks.org/find-minimum-element-in-a-sorted-and-rotated-array/\n\n \"\"\"\n\n while l <= r:\n if arr[l] <= arr[r]:\n return arr[l]\n else:\n mid = int(l + (r-l)/2)\n if arr[mid] >= arr[l]:\n l = mid + 1\n else:\n r = mid\n\n return -1\n\n\nif __name__ == '__main__':\n arr = [5, 6, 7, 8, 1]\n m = min_sorted_rotated_array(arr, 0, len(arr)-1)\n print(m)\n","sub_path":"beginner/min_sorted_rotated_array.py","file_name":"min_sorted_rotated_array.py","file_ext":"py","file_size_in_byte":705,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"467577594","text":"texto = input('Informe algo: ')\n\nvogais = 'aeiouãõáéíóúàèìòùâêîôûäëïöü'\nconsoantes = 'qwrtypsdfghjklçzxcvbnmñ'\nnumeros = '0123456789'\n\ncontador_vogais = 0\ncontador_numeros = 0\ncontador_consoantes = 0\ncontador_outros = 0\n\nfor letra in texto:\n if (letra.lower() in vogais):\n contador_vogais += 1\n elif (letra.lower() in consoantes):\n contador_consoantes += 1\n elif (letra.lower() in numeros):\n contador_numeros += 1\n else:\n contador_outros += 1\n\nprint('A quantidade de Vogais é {0}'.format(contador_vogais))\nprint('A quantidade de Consoantes é {0}'.format(contador_consoantes))\nprint('A quantidade de Números é {0}'.format(contador_numeros))\nprint('A quantidade de Outros Caracteres é {0}'.format(contador_outros))\n","sub_path":"20180924/conta_caracteres.py","file_name":"conta_caracteres.py","file_ext":"py","file_size_in_byte":774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"9767885","text":"#!/usr/bin/python3\n\"\"\" compress a file whit fabric \"\"\"\n\nfrom fabric.api import local\nimport datetime\n\n\ndef do_pack():\n \"\"\" compres a file \"\"\"\n now = datetime.datetime.now()\n year = str(now.year)\n month = str(now.month)\n day = str(now.day)\n hour = str(now.hour)\n minute = str(now.minute)\n second = str(now.second)\n name_file = \"web_static_\" + year + month + day + hour + minute + second\n command = \"sudo tar -cvzf ./versions/\" + name_file + \".tgz ./web_static\"\n print(name_file)\n print(command)\n local(\"sudo mkdir -p ./versions\")\n local(command)\n\ndo_pack()\n","sub_path":"1-pack_web_static.py","file_name":"1-pack_web_static.py","file_ext":"py","file_size_in_byte":598,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"13159464","text":"from distutils.core import setup, Extension\nimport distutils\nimport platform\nimport os\n\nextra_link_args = []\nruntime_library_dirs = []\nlib_name_prefix=''\nif platform.system() == 'Darwin':\n extra_link_args.append('-Wl,-rpath,' + os.path.join(distutils.sysconfig_get_python_lib(), 'pylibui'))\nelif platform.system() == 'Linux':\n runtime_library_dirs.append('$ORIGIN')\nelif platform.system() == 'Windows':\n lib_name_prefix='lib'\n\nlibui_module = Extension('pylibui._libui',\n ['pylibui/libui.i'],\n libraries=[lib_name_prefix+'ui'],\n runtime_library_dirs=runtime_library_dirs,\n extra_link_args = extra_link_args\n )\n\n\nsetup(\n name='pylibui',\n version='0.0.1',\n ext_modules=[libui_module],\n description='Python wrapper for libui',\n packages=['pylibui', 'pylibui.controls']\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":931,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"472913045","text":"from django.conf.urls import patterns, url, include\nfrom staffing.views import ListBranchView\nimport django.contrib.auth\n\nurlpatterns = patterns('staffing.views', \n url(r'^$', 'index'),\n url(r'^register/$', 'staff_add'),\n url(r'^team/$', 'reporting_staff_add'),\n url(r'^team/leave/$', 'team_leave_list'),\n url(r'^profile/$', 'details_add'),\n url(r'^branch/add/$', 'branch_add'),\n url(r'^branch/$', ListBranchView.as_view(), name='branch-list',),\n url(r'^branch/edit/(?P[0-9]+)/$', 'branch_edit'),\n url(r'^branch/(?P[0-9]+)/$', 'branch_view'),\n url(r'^branch/(?P[0-9]+)/wards/$', 'ward_list'),\n url(r'^branch/(?P[0-9]+)/ward/add/$', 'ward_add'),\n url(r'^branch/(?P[0-9]+)/ward/edit/(?P[0-9]+)/$', 'ward_edit'),\n url(r'^branch/(?P[0-9]+)/ward/(?P[0-9]+)/$', 'ward_view'),\n url(r'^leave/add/$', 'leave_add'),\n url(r'^leave/$', 'leave_list'),\n url(r'^leave/(?P[0-9]+)/$', 'leave_view'),\n url(r'^roster/add/$', 'roster_add'),\n url(r'^roster/edit/(?P[0-9]+)/$', 'roster_add')\n #url(r'^(?P[0-9]+)/leave/$', 'apply_leave'),\n #url(r'^(?P[0-9]+)/leave/(?P[0-9]+)/$', 'leave_details'),\n #url(r'^(?P[0-9]+)/shifts/$', 'shifts'),\n #url(r'^(?P[0-9]+)/ward/$', 'ward'),\n #url(r'^(?P[0-9]+)/holidays/$', 'holidays'),\n #url(r'^(?P[0-9]+)/attendance/$', 'attendance'),\n #url(r'^/roster/(?P[0-9]+)/roster/$', ''),\n)\n\nurlpatterns += patterns('',\n url(r'^login/$', 'django.contrib.auth.views.login'),\n url(r'^logout/$', 'django.contrib.auth.views.logout', {'template_name':'registration/logout.html'}),\n)\n","sub_path":"urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1775,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"456687079","text":"import socket, ssl,argparse\n\ndef client(host,port,cafile=None):\n purpose = ssl.Purpose.SERVER_AUTH\n context = ssl.create_default_context(purpose,cafile=cafile)\n\n raw_sock = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n raw_sock.connect((host, port))\n\n ssl_sock = context.wrap_socket(raw_sock,server_hostname=host)\n print('SSL client binding at: {}, connecting to: {}'.format(ssl_sock.getsockname(),ssl_sock.getpeername()))\n\n data = b''\n while True:\n more = ssl_sock.recv(4096)\n if not more:\n break;\n data += more\n ssl_sock.recv()\n print('SSL client receive message: {}'.format(data.decode('utf-8')))\n ssl_sock.close()\n print('SSL client closed')\n\ndef server(host,port,certificate,cafile=None):\n purpose = ssl.Purpose.CLIENT_AUTH\n context = ssl.create_default_context(purpose,cafile=cafile)\n context.load_cert_chain(certificate)\n\n raw_sock = socket.socket(socket.AF_INET,socket.SOCK_STREAM)\n raw_sock.setsockopt(socket.SOL_SOCKET,socket.SO_REUSEADDR,1)\n raw_sock.bind((host,port))\n raw_sock.listen(1)\n print('server listen at: {}'.format(raw_sock.getsockname()))\n\n raw_conn, client = raw_sock.accept()\n ssl_sock = context.wrap_socket(raw_conn,server_side=True)\n print('SSL server bindigg at: {}'.format(ssl_sock.getsockname()))\n ssl_sock.sendall(('Beatiful is better than ugly').encode('utf-8'))\n ssl_sock.close()\n print('SSL server closed')\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='SSL demonstration')\n choices = ('server','client')\n parser.add_argument('role',choices=choices,help='the server or client role')\n parser.add_argument('host',help='the server\\'s interface or client\\'s host)')\n parser.add_argument('port',help='port number',default=1060,type=int)\n parser.add_argument('-ca',help='the Certificate Authorities')\n parser.add_argument('-c',help='server\\'s certificate')\n\n args = parser.parse_args()\n\n if args.role == 'server':\n server(args.host,args.port,args.c,args.ca)\n else:\n client(args.host,args.port,args.ca)\n\n\n\n","sub_path":"_site/Chapter 6 TLS/6.1_TLS.py","file_name":"6.1_TLS.py","file_ext":"py","file_size_in_byte":2119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"118915024","text":"import imaplib\nfrom lib.santa_gen import Santa\n\n'''\nSimple script that deletes your sent emails\n\n- Could add other methods for other boxes\n- Could delete last x mails\n- Could delete emails since t time\n'''\n\nclass Flush(object):\n def __init__(self, usr, pw, n_deleted):\n print(\"\\nConnecting to the GMAIL server...\")\n self.box = imaplib.IMAP4_SSL(\"imap.gmail.com\") # connecting to gmail boxer\n self.usr = usr\n self.pw = pw\n self.n_deleted = n_deleted\n\n\n def connectImap(self):\n connect = self.box.login(self.usr, self.pw)\n print(connect)\n\n\n def checkListLabels(self):\n print(self.box.list())\n\n\n def deleteSentMails(self):\n print(\"Deleting all sent emails...\")\n self.box.select('\"[Gmail]/Sent Mail\"')\n typ, data = self.box.search(None, 'ALL')\n\n i = 0\n for num in data[0].split():\n if (i > self.n_deleted -1):\n break\n else:\n self.box.store(num, '+FLAGS', '\\\\Deleted')\n i += 1\n\n self.box.expunge()\n \n # Needed if your Gmail parameters stores deleted emails in the trash\n def cleanTrash(self):\n print(\"Emptying Trash & Expunge...\")\n self.box.select('[Gmail]/Trash') # select all trash\n self.box.store(\"1:*\", '+FLAGS', '\\\\Deleted') #Flag all Trash as Deleted\n self.box.expunge()\n\n\n def logout(self):\n print(\"Closing imap and logging out...\")\n self.box.close()\n self.box.logout()\n","sub_path":"lib/email_flush.py","file_name":"email_flush.py","file_ext":"py","file_size_in_byte":1508,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"356490071","text":"from bs4 import BeautifulSoup\nimport requests\nimport time\n\nurl = 'http://bj.xiaozhu.com/fangzi/2919537862.html'\nurls = ['http://bj.xiaozhu.com/search-duanzufang-p{}-0/'.format(number) for number in range(1, 11)]\n\nheaders = {\n 'uesr-agent':'',\n 'Cookie':'',\n}\n\ndef get_attractions(url,data=None):\n wb_data = requests.get(url)\n time.sleep(2)\n soup = BeautifulSoup(wb_data.text,'lxml')\n titles = soup.select('body > div.wrap.clearfix.con_bg > div.con_l > div.pho_info > h4 > em')\n addres = soup.select('body > div.wrap.clearfix.con_bg > div.con_l > div.pho_info > p > span.pr5')\n rents = soup.select('#pricePart > div.day_l > span')\n images = soup.select('#curBigImage')\n ownerimges = soup.select('#floatRightBox > div.js_box.clearfix > div.member_pic > a > img')\n sexs = soup.select('#floatRightBox > div.js_box.clearfix > div.member_pic > div')\n names = soup.select('#floatRightBox > div.js_box.clearfix > div.w_240 > h6 > a')\n\n if data == None:\n for title, addre, rent, image, ownerimge, sex, name in zip(titles, addres, rents, images, ownerimges, sexs, names):\n data = {\n 'title':title.get_text(),\n 'addre':addre.get_text(),\n 'rent':rent.get_text(),\n 'image':image.get('src'),\n 'ownerimge':ownerimge.get('src'),\n 'sex':get_owner_sex(sex.get(\"class\")),\n 'name':name.get_text(),\n }\n print(data)\n\ndef get_need_url(url):\n wb_data = requests.get(url)\n time.sleep(2)\n soup = BeautifulSoup(wb_data.text,'lxml')\n links = soup.select('#page_list > ul > li > a')\n\n for link in links:\n href = link.get('href')\n get_attractions(href)\n\ndef get_owner_sex(class_name):\n if class_name == ['member_ico']:\n return '男'\n else:\n return '女'\n\nfor single_url in urls:\n get_need_url(single_url)\n","sub_path":"practice/week1/1-3/xiaozhu_single.py","file_name":"xiaozhu_single.py","file_ext":"py","file_size_in_byte":1907,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"174094780","text":"\"\"\"Simple HTTP Server.\n\nThis module builds on BaseHTTPServer by implementing the standard GET\nand HEAD requests in a fairly straightforward manner.\n\n\"\"\"\n\n\n__version__ = \"0.6\"\n\n__all__ = [\"SimpleHTTPRequestHandler\"]\n\nimport os\nimport posixpath\nimport BaseHTTPServer\nimport urllib\nimport urlparse\nimport cgi\nimport sys\nimport shutil\nimport mimetypes\ntry:\n from cStringIO import StringIO\nexcept ImportError:\n from StringIO import StringIO\n\n\nclass SimpleHTTPRequestHandler(BaseHTTPServer.BaseHTTPRequestHandler):\n\n \"\"\"Simple HTTP request handler with GET and HEAD commands.\n\n This serves files from the current directory and any of its\n subdirectories. The MIME type for files is determined by\n calling the .guess_type() method.\n\n The GET and HEAD requests are identical except that the HEAD\n request omits the actual contents of the file.\n\n \"\"\"\n\n server_version = \"SimpleHTTP/\" + __version__\n\n def do_GET(self):\n \"\"\"Serve a GET request.\"\"\"\n f = self.send_head()\n if f:\n try:\n self.copyfile(f, self.wfile)\n finally:\n f.close()\n\n def do_HEAD(self):\n \"\"\"Serve a HEAD request.\"\"\"\n f = self.send_head()\n if f:\n f.close()\n\n def send_head(self):\n \"\"\"Common code for GET and HEAD commands.\n\n This sends the response code and MIME headers.\n\n Return value is either a file object (which has to be copied\n to the outputfile by the caller unless the command was HEAD,\n and must be closed by the caller under all circumstances), or\n None, in which case the caller has nothing further to do.\n\n \"\"\"\n path = self.translate_path(self.path)\n f = None\n #V : os.path.isdir(path) returns True if path is an existing directory. \n if os.path.isdir(path):\n \"\"\"V: urlparse() : from urllib.parse module : Parse a URL\n into six components, returning a 6-item named tuple. \n >>> o = urlparse('http://www.cwi.nl:80/%7Eguido/Python.html')\n >>> 0\n ParseResult(scheme='http', netloc='www.cwi.nl:80', path='/%7Eguido/Python.html', params='', query='', fragment='')\n\n urlsplit() : similar to urlparse(), but does not split the\n params from the URL. Note : URL Parameters are parameters \n whose values are set dynamically in a page's URL, and can be\n accessed by its template and its data sources. \"\"\"\n parts = urlparse.urlsplit(self.path)\n # if the element path returned by urlparse doesn't end with / : modifies the URL to add the /\n if not parts.path.endswith('/'):\n # redirect browser - doing basically what apache does\n #V : error 301 : url moved parmanently. \n self.send_response(301)\n new_parts = (parts[0], parts[1], parts[2] + '/',\n parts[3], parts[4])\n #urlunsplit(parts) : combines the elements of a tuple as returned by urlsplit() into a complete URL as a string. The parts argument can be any five-item iterable.\n new_url = urlparse.urlunsplit(new_parts)\n self.send_header(\"Location\", new_url)\n self.end_headers()\n return None\n #index successively takes the values \"index.html\" and \"index.htm\". os.path.join concatenates the path and index.\n for index in \"index.html\", \"index.htm\":\n index = os.path.join(path, index)\n if os.path.exists(index):\n path = index\n break\n else:\n return self.list_directory(path)\n ctype = self.guess_type(path)\n try:\n # Always read in binary mode. Opening files in text mode may cause\n # newline translations, making the actual size of the content\n # transmitted *less* than the content-length!\n f = open(path, 'rb')\n except IOError:\n self.send_error(404, \"File not found\")\n return None\n try:\n self.send_response(200)\n self.send_header(\"Content-type\", ctype)\n \"\"\"V : os.fstat() method in Python is used to get the status \n of a file descriptor. The returned ‘stat_result’ object is a \n tuple with name attributes (ex : st_ino > inode number on Unix). \n fileno returns the file descriptor.\"\"\"\n fs = os.fstat(f.fileno())\n #V : fs[6] : size of the file in bytes \n self.send_header(\"Content-Length\", str(fs[6]))\n self.send_header(\"Last-Modified\", self.date_time_string(fs.st_mtime))\n self.end_headers()\n return f\n except:\n f.close()\n raise\n\n def list_directory(self, path):\n \"\"\"Helper to produce a directory listing (absent index.html).\n\n Return value is either a file object, or None (indicating an\n error). In either case, the headers are sent, making the\n interface the same as for send_head().\n\n \"\"\"\n try:\n #D : listdir() returns a list containing the names of the entries in the directory given by path.\n list = os.listdir(path)\n except os.error:\n self.send_error(404, \"No permission to list directory\")\n return None\n list.sort(key=lambda a: a.lower())\n f = StringIO()\n \"\"\"V : URLs can only be sent over the Internet using the ASCII character-set.\n Since URLs often contain characters outside the ASCII set, \n the URL has to be converted into a valid ASCII format.\n URL encoding replaces unsafe ASCII characters with a \"%\" \n followed by two hexadecimal digits.\n URLs cannot contain spaces. URL encoding normally replaces \n a space with a plus (+) sign or with %20.\n urllib.unquote(string, encoding='utf-8', errors='replace') \n replaces %xx escapes by their single-character equivalent. \n cgi.escape() : cgi.escape(s[, quote]) : converts the characters \n '&', '<' and '>' in string s to HTML-safe sequences.\"\"\"\n displaypath = cgi.escape(urllib.unquote(self.path))\n f.write('')\n f.write(\"\\nDirectory listing for %s\\n\" % displaypath)\n f.write(\"\\n

Directory listing for %s

\\n\" % displaypath)\n f.write(\"
\\n
    \\n\")\n for name in list:\n fullname = os.path.join(path, name)\n displayname = linkname = name\n # Append / for directories or @ for symbolic links\n if os.path.isdir(fullname):\n displayname = name + \"/\"\n linkname = name + \"/\"\n if os.path.islink(fullname):\n displayname = name + \"@\"\n # Note: a link to a directory displays with @ and links with /\n f.write('
  • %s\\n'\n % (urllib.quote(linkname), cgi.escape(displayname)))\n f.write(\"
\\n
\\n\\n\\n\")\n length = f.tell()\n f.seek(0)\n self.send_response(200)\n encoding = sys.getfilesystemencoding()\n self.send_header(\"Content-type\", \"text/html; charset=%s\" % encoding)\n self.send_header(\"Content-Length\", str(length))\n self.end_headers()\n return f\n\n def translate_path(self, path):\n \"\"\"Translate a /-separated PATH to the local filename syntax.\n\n Components that mean special things to the local file system\n (e.g. drive or directory names) are ignored. (XXX They should\n probably be diagnosed.)\n\n \"\"\"\n # abandon query parameters\n #V : split second parameter is specifies how many splits to do (default value is -1 for all occurences). Ex : \"apple#banana#cherry\".split(\"#\",1) > ['apple', 'banana#cherry']\n path = path.split('?',1)[0]\n path = path.split('#',1)[0]\n # Don't forget explicit trailing slash when normalizing. Issue17324\n #V : trailing slash = last URL slash. \n trailing_slash = path.rstrip().endswith('/')\n path = posixpath.normpath(urllib.unquote(path))\n words = path.split('/')\n words = filter(None, words)\n path = os.getcwd()\n for word in words:\n if os.path.dirname(word) or word in (os.curdir, os.pardir):\n # Ignore components that are not a simple file/directory name\n continue\n path = os.path.join(path, word)\n if trailing_slash:\n path += '/'\n return path\n\n def copyfile(self, source, outputfile):\n \"\"\"Copy all data between two file objects.\n\n The SOURCE argument is a file object open for reading\n (or anything with a read() method) and the DESTINATION\n argument is a file object open for writing (or\n anything with a write() method).\n\n The only reason for overriding this would be to change\n the block size or perhaps to replace newlines by CRLF\n -- note however that this the default server uses this\n to copy binary data as well.\n\n \"\"\"\n shutil.copyfileobj(source, outputfile)\n\n def guess_type(self, path):\n \"\"\"Guess the type of a file.\n\n Argument is a PATH (a filename).\n\n Return value is a string of the form type/subtype,\n usable for a MIME Content-type header.\n\n The default implementation looks the file's extension\n up in the table self.extensions_map, using application/octet-stream\n as a default; however it would be permissible (if\n slow) to look inside the data to make a better guess.\n\n \"\"\"\n\n \"\"\"V : since different operating systems have different path name \n conventions, there are several versions of this module in the standard library. \n The os.path module is always the path module suitable for the operating \n system Python is running on, and therefore usable for local paths. \n However, you can also import and use the individual modules if you want \n to manipulate a path that is always in one of the different formats. \n They all have the same interface:\n posixpath for UNIX-style paths\n ntpath for Windows paths\"\"\"\n \"\"\"V : os.path.splitext(path): splits the pathname path into a pair (root, ext)\n such that root + ext == path, and ext is empty or begins with a period \n and contains at most one period.\"\"\" \n base, ext = posixpath.splitext(path)\n if ext in self.extensions_map:\n return self.extensions_map[ext]\n ext = ext.lower()\n if ext in self.extensions_map:\n return self.extensions_map[ext]\n else:\n return self.extensions_map['']\n\n \"\"\"V : mimetypes.inited : Flag indicating whether or not the global data structures \n have been initialized. This is set to True by init().\"\"\"\n if not mimetypes.inited:\n mimetypes.init() # try to read system mime.types\n extensions_map = mimetypes.types_map.copy()\n extensions_map.update({\n '': 'application/octet-stream', # Default\n '.py': 'text/plain',\n '.c': 'text/plain',\n '.h': 'text/plain',\n })\n\n\ndef test(HandlerClass = SimpleHTTPRequestHandler,\n ServerClass = BaseHTTPServer.HTTPServer):\n BaseHTTPServer.test(HandlerClass, ServerClass)\n\n\nif __name__ == '__main__':\n test()\n","sub_path":"SimpleHTTPServer_comments.py","file_name":"SimpleHTTPServer_comments.py","file_ext":"py","file_size_in_byte":11531,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"573523726","text":"import flask\nfrom dns.resolver import query\nfrom difflib import get_close_matches\nfrom flask import request, jsonify\nfrom cachetools import cached, TTLCache\n\n#cache_validateDomain = TTLCache(maxsize=8000, ttl=2000)\ncache_suggestDomain = TTLCache(maxsize=8000, ttl=2000) # set a lower ttl for new domains added to the safelist during i_validateDomain_MX to come into affect\napp = flask.Flask(__name__)\nsafeList={'gmail.com','argos.co.uk','homeretailgroup.com'}\n\napp.config[\"DEBUG\"] = True\n\n##import logging\n##log = logging.getLogger('werkzeug')\n##log.setLevel(logging.ERROR)\n\n#@cached(cache_validateDomain)\ndef i_validateDomain_MX(domain): # function performing under cache to reduce the number of calls\n try:\n if domain in safeList:\n return True\n else:\n query(domain, 'MX')\n safeList.add (domain) # add domain with valid MX record to the safelist - will be used once the suggestCorrectDomain cache expires\n return True # return positive as domain has a valid MX record\n except:\n return False # return negative as domain has no valid MX record\n\n@cached(cache_suggestDomain)\ndef i_suggestCorrectDomain(domain):\n try:\n rtnResult = (get_close_matches(domain,safeList))\n return (rtnResult) # return results back from close match\n except:\n return False\n\n# ** App route functions **\n\n@app.route('/', methods=['GET'])\ndef home():\n return \"

This will return POSTMAN API documentation

\"\n\n@app.route('/api/v1/validateDomain', methods=['GET'])\ndef validateDomain():\n if 'domain' in request.args:\n ext_domainToValidate = str(request.args['domain'])\n results = int(i_validateDomain_MX(ext_domainToValidate))\n else:\n return \"Error: No domain field provided. Please specify an domain.\"\n return jsonify(results)\n\n@app.route('/api/v1/suggestDomain',methods=['GET'])\ndef suggestCorrectDomain():\n if 'domain' in request.args:\n ext_domainToQuery = str(request.args['domain'])\n results = (i_suggestCorrectDomain(ext_domainToQuery))\n else:\n return \"Error: No domain field provided. Please specify a domain.\"\n return jsonify(results)\n\n@app.route('/admin/',methods=['GET'])\ndef listSize():\n if 'safelistsize' in request.args:\n results = str(len(safeList))\n else:\n return \"Error: no field provided.\"\n return jsonify(results)\n\napp.run()\n","sub_path":"go_api.py","file_name":"go_api.py","file_ext":"py","file_size_in_byte":2401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"604325016","text":"import json\nimport copy\n\nfrom pathlib import Path\nfrom mergedict import ConfigDict\nfrom doit.tools import config_changed\nfrom doitcmd import BaseCommand\n\n\n\nclass JsHint(BaseCommand):\n cmd_template = 'jshint {opts} {js_file}'\n\n def __init__(self, config=None, **opts):\n \"\"\"\n @param config: (str) config path\n \"\"\"\n super(JsHint, self).__init__(options=opts)\n self.config_file = config\n if config:\n with open(config, 'r') as fp:\n self._config = ConfigDict(json.load(fp))\n else:\n self._config = {}\n\n\n def __call__(self, config_file, js_file):\n \"\"\"return task metada to jshint single file\"\"\"\n opts = self.opt_str(self.options, {'config': config_file})\n cmd = self.cmd_template.format(opts=opts, js_file=js_file)\n\n return {\n 'name': js_file,\n 'actions': [cmd],\n 'file_dep': [config_file, js_file],\n }\n\n\n def tasks(self, patterns, group='all', exclude=(), options=None):\n \"\"\"yield tasks as given by pattern\n\n @param group: (str) name of a group\n @param pattern: (list - str) list of path patterns of files to be linted\n @param exclude: (list - str) list of path of files to be removed\n from selection\n @param options: (dict) extra options for group\n \"\"\"\n\n # It seems jshint won't ever accept options from command line\n # https://github.com/jshint/jshint/issues/807\n # So we create a jshint config file for each \"group\"\n cfg = ConfigDict(copy.deepcopy(self._config))\n if options:\n cfg.merge(options)\n config_file = '_hint_{}.json'.format(group)\n def write_config():\n with open(config_file, 'w') as fp:\n json.dump(cfg, fp, indent=4, sort_keys=True)\n yield {\n 'name': config_file,\n 'actions': [write_config],\n 'targets': [config_file],\n 'uptodate': [config_changed(cfg)],\n }\n else:\n config_file = self.config_file\n\n # yield a task for every js file in selection\n base = Path('.')\n excluded = set([base.joinpath(e) for e in exclude])\n for pattern in patterns:\n for src in base.glob(pattern):\n if src not in excluded:\n yield self(config_file, str(src))\n","sub_path":"doitweb/jshint.py","file_name":"jshint.py","file_ext":"py","file_size_in_byte":2455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"375800187","text":"import requests\nimport toml\nfrom bs4 import BeautifulSoup\n\nif __name__ == '__main__':\n with open('config.toml', 'r') as config_file:\n config_toml = toml.loads(config_file.read())\n config_file.close()\n\n proxies = {\n 'http': '127.0.0.1:8118',\n 'https': '127.0.0.1:8118',\n }\n\n for index in range(34):\n print('page' + str(index + 1) + '... ')\n\n session = requests.Session()\n session.headers.update({\n 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) '\n 'Chrome/84.0.4147.89 Safari/537.36'})\n\n r = session.get('https://exhentai.org/uploader/BlossomPlus/' + str(index),\n proxies=proxies, cookies=config_toml)\n soup = BeautifulSoup(r.text, 'html.parser')\n\n for tag in soup.findAll('a', href=True):\n if str(tag).__contains__('https://exhentai.org/g/') and str(tag).__contains__('[Chinese]'):\n with open('data.txt', 'a') as file:\n link = tag['href']\n print(link)\n\n file.write(link + '\\n')\n file.close()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"128980718","text":"import requests\nfrom requests_ntlm import HttpNtlmAuth\nimport glob\nimport json\nimport datetime\nimport csv\nimport os\nfrom shutil import copyfile\nimport olefile\nfrom django.core.management.base import BaseCommand, CommandError\nfrom open_grgraz import settings\nfrom api import models\n\n\nclass Command(BaseCommand):\n def add_arguments(self, parser):\n parser.add_argument('username', nargs='+', type=str)\n parser.add_argument('password', nargs='+', type=str)\n\n def handle(self, *args, **options):\n self.main(options['username'][0], options['password'][0])\n\n @staticmethod\n def main(username, password):\n session = Command.create_session(username, password)\n\n Command.download_motion_lists(session) # this does nothing for now.\n Command.download_answer_lists(session) # this does nothing for now.\n\n answers_csv = Command.parse_lists(settings.ANSWER_LISTS_PATH, 'answers.csv')\n motions_csv = Command.parse_lists(settings.MOTION_LISTS_PATH, 'motions.csv')\n\n Command.download_from_csv(answers_csv, settings.RAW_ANSWERS_PATH, session)\n Command.download_from_csv(motions_csv, settings.RAW_MOTIONS_PATH, session)\n\n Command.copy_files(settings.RAW_ANSWERS_PATH, settings.ANSWERS_PATH, '*.pdf')\n Command.copy_files(settings.RAW_MOTIONS_PATH, settings.MOTIONS_PATH, '*.pdf')\n Command.copy_files(settings.RAW_ANSWERS_PATH, settings.ANSWERS_PATH, '*.doc')\n Command.copy_files(settings.RAW_MOTIONS_PATH, settings.MOTIONS_PATH, '*.doc')\n Command.copy_files(settings.RAW_ANSWERS_PATH, settings.ANSWERS_PATH, '*.docx')\n Command.copy_files(settings.RAW_MOTIONS_PATH, settings.MOTIONS_PATH, '*.docx')\n\n Command.convert_documents(answers_csv) # this does nothing for now.\n Command.convert_documents(motions_csv) # this does nothing for now.\n\n Command.extract_email_attachments(settings.RAW_ANSWERS_PATH, settings.ANSWERS_PATH, answers_csv)\n Command.extract_email_attachments(settings.RAW_MOTIONS_PATH, settings.MOTIONS_PATH, motions_csv)\n\n Command.load_answers_dates('answers_dates.csv')\n\n # todo:\n # - fetch motions and answer lists\n # - create pdfs out of word documents,\n # - ocr pdfs without text\n # - create webfrontend\n\n def create_session(username, password):\n session = requests.Session()\n session.auth = HttpNtlmAuth('\\\\' + username, password)\n result = session.get(settings.MAGISTRAT_BASE_URL)\n if result.status_code != 200:\n raise CommandError('Login error {}.'.format(result.status_code))\n return session\n\n @staticmethod\n def download_motion_lists(session):\n print('download_motion_lists')\n\n @staticmethod\n def download_answer_lists(session):\n print('download_answer_lists')\n\n @staticmethod\n def parse_lists(lists_path, csv_filename):\n json_files = glob.glob(lists_path + '*.json')\n\n element_lists = []\n for file_path in json_files:\n with open(file_path, 'r') as file:\n element_lists.append(json.load(file))\n\n element_csv = []\n for element_list in element_lists:\n for element in element_list['Row']:\n\n if 'Polz Wolfgang' == element['Antragsteller'][0]['title']:\n continue\n\n session_date = datetime.datetime.strptime(element['Sitzung_x0020_am'], '%d.%m.%Y')\n session = models.ParliamentarySession.objects.get_or_create(session_date=session_date)[0]\n #print(session)\n\n group_id = element['Fraktion']\n group = models.ParliamentaryGroup.objects.get_or_create(id=group_id, defaults={'name': group_id})[0]\n #print(group)\n\n person_name = element['Antragsteller'][0]['title']\n person_academic_degree = element['Antragsteller'][0]['jobTitle']\n if person_academic_degree == person_name:\n person_academic_degree = ''\n person_email = element['Antragsteller'][0]['email']\n person = models.CouncilPerson.objects.update_or_create(name=person_name, defaults={\n 'academic_degree': person_academic_degree,\n 'email': person_email,\n 'parliamentary_group': group})[0]\n\n #print(person)\n\n #todo: files?\n motion_type = element['Dokumentenart']['Label']\n if motion_type == 'GR-Antwort':\n download_path = 'files/answers/'\n else:\n download_path = 'files/motions/'\n file = models.File.objects.update_or_create(long_filename=element['FileLeafRef'], defaults={'short_filename': element['FileLeafRef'],\n 'path': download_path + element['FileLeafRef']})[0]\n #print(file)\n\n\n motion_id = int(float(element['FileLeafRef'][:4].replace('_', '')))\n #print(motion_id)\n if motion_type == 'GR-Antwort':\n answer = models.Answer.objects.update_or_create(id=element['ID'], defaults={'motion_id': motion_id,\n 'session': session, 'title': element['Title'],\n 'parliamentary_group': group, 'proposer': person})[0]\n answer.files.add(file)\n answer.save()\n #print(answer)\n else:\n from django.core.exceptions import ObjectDoesNotExist\n try:\n answers = models.Answer.objects.filter(motion_id=motion_id)\n except ObjectDoesNotExist:\n answers = None\n motion = models.Motion.objects.update_or_create(id=element['ID'], defaults={'motion_id': motion_id,\n 'session': session, 'title': element['Title'],\n 'motion_type': motion_type, 'parliamentary_group': group,\n 'proposer': person})[0]\n motion.answers.set(answers)\n motion.files.add(file)\n motion.save()\n #print(motion)\n\n\n element_csv.append([\n element['Sitzung_x0020_am'],\n element['ID'],\n element['Title'],\n element['Dokumentenart']['Label'],\n element['Antragsteller'][0]['jobTitle'] + ' ' + element['Antragsteller'][0]['title'],\n element['Fraktion'],\n element['Betreff'],\n '',\n element['FileRef'],\n element['FileLeafRef'],\n ])\n\n def element_sort(answer):\n date = answer[0].split('.')\n return date[2], date[2], date[1], int(answer[1])\n\n element_csv = sorted(element_csv, key=element_sort)\n\n with open(settings.FILES_PATH + csv_filename, 'w', newline='') as csv_file:\n writer = csv.writer(csv_file)\n writer.writerow(('Datum', 'Nummer', 'Titel', 'Art', 'Antragsteller', 'Partei', 'Dringlichkeit', 'Angenommen', 'Link', 'Anwort 1', 'Anwort 2', 'Antwort', 'link'))\n writer.writerows(element_csv)\n\n return element_csv\n\n @staticmethod\n def download_from_csv(csv, base_path, session):\n download_number = 1\n for line in csv:\n url = settings.MAGISTRAT_BASE_URL + line[8]\n local_filename = url.split('/')[-1]\n\n if os.path.exists(base_path + local_filename):\n continue\n\n print('{:>4} - downloading: {}'.format(download_number, local_filename))\n download_number += 1\n result = session.get(url, stream=True)\n with open(base_path + local_filename, 'wb') as file:\n for chunk in result.iter_content(chunk_size=1024):\n if chunk:\n file.write(chunk)\n\n @staticmethod\n def copy_files(source_path, destination_path, pattern):\n file_paths = glob.glob(source_path + pattern)\n for file_path in file_paths:\n filename = file_path.split('/')[-1]\n copyfile(file_path, destination_path + filename)\n\n @staticmethod\n def convert_documents(motions_csv):\n for motion in motions_csv:\n filename = motion[9]\n if filename.split('.')[-1] == 'msg' or filename.split('.')[-1] == 'pdf':\n continue\n\n #todo\n #unoconv\n\n @staticmethod\n def extract_email_attachments(read_base_path, write_base_path, motions_csv):\n for motion in motions_csv:\n filename = motion[9]\n if filename.split('.')[-1] != 'msg':\n continue\n\n motion_or_answer = None\n motion_type = motion[3]\n if motion_type == 'GR-Antwort':\n motion_or_answer = models.Answer.objects.get(id=motion[1])\n else:\n motion_or_answer = models.Motion.objects.get(id=motion[1])\n\n old_file = models.File.objects.get(long_filename=filename).delete()\n\n file_path = read_base_path + filename\n\n ole = olefile.OleFileIO(file_path)\n\n attachment_dirs = []\n for dir in ole.listdir():\n if dir[0].startswith('__attach') and dir[0] not in attachment_dirs:\n attachment_dirs.append(dir[0])\n\n def windowsUnicode(string):\n if string is None:\n return None\n #if sys.version_info[0] >= 3: # Python 3\n return str(string, 'utf_16_le')\n #else: # Python 2\n # return unicode(string, 'utf_16_le')\n\n def get_stream(ole, filename):\n asciiVersion = None\n unicodeVersion = None\n\n if ole.exists(filename + '001E'):\n asciiVersion = ole.openstream(filename + '001E').read()\n\n if ole.exists(filename + '001F'):\n unicodeVersion = windowsUnicode(ole.openstream(filename + '001F').read())\n\n if asciiVersion is None:\n return unicodeVersion\n elif unicodeVersion is None:\n return asciiVersion\n else:\n return unicodeVersion\n\n if len(attachment_dirs) > 0:\n dirname = filename.split('.')[0]\n if not os.path.exists(write_base_path + dirname):\n os.makedirs(write_base_path + dirname)\n\n for dir in attachment_dirs:\n long_filename = get_stream(ole, dir + '/__substg1.0_3707')\n short_filename = get_stream(ole, dir + '/__substg1.0_3704')\n\n data = None\n if ole.exists(dir + '/__substg1.0_37010102'):\n data = ole.openstream(dir + '/__substg1.0_37010102').read()\n\n attachment_filename = short_filename\n if attachment_filename is None:\n attachment_filename = long_filename\n\n if attachment_filename == 'image001.jpg':\n continue\n\n if attachment_filename is not None and data is not None:\n attachment_path = write_base_path + dirname + '/' + attachment_filename\n fileModel = models.File.objects.update_or_create(long_filename=attachment_filename, defaults={'short_filename': attachment_filename,\n 'path': attachment_path})[0]\n #print(fileModel)\n motion_or_answer.files.add(fileModel)\n motion_or_answer.save()\n if not os.path.exists(attachment_path):\n file = open(attachment_path, 'wb')\n file.write(data)\n file.close()\n\n\n @staticmethod\n def load_answers_dates(csv_filename):\n with open(settings.FILES_PATH + csv_filename, 'r') as csv_file:\n csv_reader = csv.reader(csv_file)\n for row in csv_reader:\n if row[0] == 'Nummer':\n continue\n if row[1] == '':\n continue\n #print(row[0])\n #print(row[1])\n id = int(row[0])\n date = datetime.datetime.strptime(row[1], '%d.%m.%Y')\n models.Answer.objects.filter(id=id, answered_date__isnull=True).update(answered_date=date)\n","sub_path":"api/management/commands/loadmagistratdata.py","file_name":"loadmagistratdata.py","file_ext":"py","file_size_in_byte":13133,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"561033712","text":"# Copyright (c) 2021 PickNik, Inc.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright\n# notice, this list of conditions and the following disclaimer.\n#\n# * Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n#\n# * Neither the name of the {copyright_holder} nor the names of its\n# contributors may be used to endorse or promote products derived from\n# this software without specific prior written permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE\n# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR\n# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF\n# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS\n# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN\n# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)\n# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n\n#\n# Author: Denis Stogl\n\nfrom launch import LaunchDescription\nfrom launch.actions import DeclareLaunchArgument, IncludeLaunchDescription\nfrom launch.launch_description_sources import PythonLaunchDescriptionSource\nfrom launch.substitutions import LaunchConfiguration, ThisLaunchFileDir\n\n\ndef generate_launch_description():\n # Declare arguments\n declared_arguments = []\n declared_arguments.append(\n DeclareLaunchArgument(\n \"robot_ip\",\n description=\"IP address by which the robot can be reached.\",\n )\n )\n declared_arguments.append(\n DeclareLaunchArgument(\n \"use_fake_hardware\",\n default_value=\"false\",\n description=\"Start robot with fake hardware mirroring command to its states.\",\n )\n )\n declared_arguments.append(\n DeclareLaunchArgument(\n \"fake_sensor_commands\",\n default_value=\"false\",\n description=\"Enable fake command interfaces for sensors used for simple simulations. \\\n Used only if 'use_fake_hardware' parameter is true.\",\n )\n )\n declared_arguments.append(\n DeclareLaunchArgument(\n \"initial_joint_controller\",\n default_value=\"scaled_joint_trajectory_controller\",\n description=\"Initially loaded robot controller.\",\n choices=[\n \"scaled_joint_trajectory_controller\",\n \"joint_trajectory_controller\",\n \"forward_velocity_controller\",\n \"forward_position_controller\",\n ],\n )\n )\n declared_arguments.append(\n DeclareLaunchArgument(\n \"activate_joint_controller\",\n default_value=\"true\",\n description=\"Activate loaded joint controller.\",\n )\n )\n\n # Initialize Arguments\n robot_ip = LaunchConfiguration(\"robot_ip\")\n use_fake_hardware = LaunchConfiguration(\"use_fake_hardware\")\n fake_sensor_commands = LaunchConfiguration(\"fake_sensor_commands\")\n initial_joint_controller = LaunchConfiguration(\"initial_joint_controller\")\n activate_joint_controller = LaunchConfiguration(\"activate_joint_controller\")\n\n base_launch = IncludeLaunchDescription(\n PythonLaunchDescriptionSource([ThisLaunchFileDir(), \"/ur_control.launch.py\"]),\n launch_arguments={\n \"ur_type\": \"ur10e\",\n \"robot_ip\": robot_ip,\n \"use_fake_hardware\": use_fake_hardware,\n \"fake_sensor_commands\": fake_sensor_commands,\n \"initial_joint_controller\": initial_joint_controller,\n \"activate_joint_controller\": activate_joint_controller,\n }.items(),\n )\n\n return LaunchDescription(declared_arguments + [base_launch])\n","sub_path":"ur_bringup/launch/ur10e.launch.py","file_name":"ur10e.launch.py","file_ext":"py","file_size_in_byte":4266,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"495175325","text":"#!/usr/bin/python\n#\n'''\n[ replace_string_in_file.py ]\n\nCOPYRIGHT (c) 2015 by MediaMath, Cambridge, MA USA.\nAll rights reserved. This material contains unpublished, copyrighted\nwork including confidential and proprietary information of MediaMath.\n\nreplace a string in a file with another string and write to another file\n python replace_string_in_file \n string1 - original string to be replaced\n string2 - string you are replacing original string with\n file1 - file to read from\n file2 - file to write to\nthe only difference between file1 and file2 when done should be that all\noccurrences of string1 have been replaced with string2\n'''\nimport sys\n\nparams = []\nfor arg in sys.argv:\n print(arg)\n params.append(arg)\n\norig_str = params[1]\nreplace_str = params[2]\ninput_file = params[3]\noutput_file = params[4]\ninfile = open(input_file, 'r+')\noutfile = open(output_file, 'w+')\n\nfor line in infile.readlines():\n if orig_str in line:\n print(\"DEBUG: replacing \" + orig_str + \" with \" + replace_str)\n changedLine = line.replace(orig_str, replace_str)\n outfile.write(changedLine)\n else:\n outfile.write(line)\n\noutfile.close()\ninfile.close()\n","sub_path":"plexus/tests/ads-mmtest/lib/utils/replace_string_in_file.py","file_name":"replace_string_in_file.py","file_ext":"py","file_size_in_byte":1268,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"532128533","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Nov 13 16:52:44 2018\n\nExamination of the area around the Sco-OB2 Stellar association\n\nThis script examines the area around the 'known' members of the Sco-OB2 Stellar\nAssociation, and plots a series of plots in order to take steps towards identifying \nnew members of this association.\n\nInputs: A VOTable containing Gaia DR2 data for the area around the Sco-OB2 \n Stellar Association, including position, velocity, magnitude and indetifying\n information.\n\nOuputs: Proper motion density plot\n Full area CAMD diagram\n Plot of stellar positions in this area\n\n@author: Matthew Battley\n\"\"\"\n\nimport astropy.table as tab\nimport numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom astropy.coordinates import SkyCoord\nimport timeit\nfrom matplotlib.path import Path\nimport matplotlib.patches as patches\nfrom scipy.stats import kde\n\nstart = timeit.default_timer()\n\ndef plot_with_colourbar(x,y,mag,xlabel,ylabel,title,cbar_label = 'g Magnitude' ,invert_y_axis = False, y_lim = False):\n \"\"\"\n Function for plotting a scatter plot in two variables (x,y) with a colour bar based on a third (mag)\n \"\"\"\n # Sets up colours and normalisation for colourbar\n cmap = matplotlib.cm.get_cmap('rainbow')\n normalize = matplotlib.colors.Normalize(vmin = min(mag), vmax=max(mag))\n colours = [cmap(normalize(value)) for value in mag]\n \n # Plots figure\n fig_pos, ax = plt.subplots(figsize=(10,10))\n plt.scatter(x,y,0.5,c=colours)\n if invert_y_axis == True:\n plt.gca().invert_yaxis()\n if y_lim != False:\n plt.gca().set_ylim(y_lim)\n plt.xlabel(xlabel)\n plt.ylabel(ylabel)\n plt.title(title)\n cax, _ = matplotlib.colorbar.make_axes(ax)\n cbar = matplotlib.colorbar.ColorbarBase(cax, cmap=cmap, norm = normalize)\n cbar.ax.invert_yaxis()\n cbar.set_label(cbar_label)\n\n######################## IMPORTS AND SORTS OUT DATA ###########################\n\n# Read data from table\nTable = tab.Table\ndata = Table.read('OB2_area_data.vot')\nhipparcos_data = Table.read('Hipparcos_OB2_de_Zeeuw_1999.vot')\n\n# Change from unrecognisable unit names in file\ndata['pmra'].unit = 'mas/yr'\ndata['pmdec'].unit = 'mas/yr'\ndata['radial_velocity'].unit = 'km/s'\nhipparcos_data['pmra'].unit = 'mas/yr'\nhipparcos_data['pmdec'].unit = 'mas/yr'\nhipparcos_data['ra'].unit = 'deg'\nhipparcos_data['dec'].unit = 'deg'\n\n# Input sky coordinates for all stars\nc_icrs = SkyCoord(ra = data['ra'], dec = data['dec'], pm_ra_cosdec = data['pmra'], pm_dec = data['pmdec'])\nc_icrs_hipparcos = SkyCoord(ra = hipparcos_data['ra'], dec = hipparcos_data['dec'], pm_ra_cosdec = hipparcos_data['pmra'], pm_dec = hipparcos_data['pmdec'])\n\n# Convert star coordinates to Galactic frame\nc_galactic = c_icrs.galactic\nc_galactic_hipparcos = c_icrs_hipparcos.galactic\n\n# Add equivalent galactic coordinates back into data\ndata['pm_l_cosb'] = c_galactic.pm_l_cosb\ndata['pm_b'] = c_galactic.pm_b\nhipparcos_data['pm_l_cosb'] = c_galactic_hipparcos.pm_l_cosb\nhipparcos_data['pm_b'] = c_galactic_hipparcos.pm_b\n\n# Select stars within this data where pms are only in the region between pm_l = [-50,10] and pm_b = [-30,30]\nsel = data['pm_l_cosb'] >= -50\nsel &= data['pm_l_cosb'] < 10\nsel &= data['pm_b'] >= -30\nsel &= data['pm_b'] <= 30\n\nsmall_area_stars = data[sel]\n\n\n################## PLOTS PM PLOT AND DEFINES AREA OF INTEREST #################\n\n# Plotting proper motion density plot\nfig = plt.figure()\nk = kde.gaussian_kde([small_area_stars['pm_l_cosb'], small_area_stars['pm_b']])\nnbins = 500\nxi, yi = np.mgrid[small_area_stars['pm_l_cosb'].min():small_area_stars['pm_l_cosb'].max():nbins*1j, small_area_stars['pm_b'].min():small_area_stars['pm_b'].max():nbins*1j]\nzi = k(np.vstack([xi.flatten(), yi.flatten()]))\ncs = plt.pcolormesh(xi, yi, zi.reshape(xi.shape), cmap=plt.cm.gist_ncar_r)\nplt.colorbar()\nplt.xlabel('pm_l_cosb (mas/yr)')\nplt.ylabel('pm_b (mas/yr)')\nplt.title('Proper motion plot for potential Sco_OB2 members')\nplt.scatter(hipparcos_data['pm_l_cosb'],hipparcos_data['pm_b'], 0.1, 'k')\n \n# Defines polygon vertices and path for area of interest\nverts = [\n (-48., -18.),\n (-28., -23.),\n (-10., -10.),\n (-18., 8.),\n (-45., -2.),\n (-48., -18.)\n ]\npath = Path(verts)\n\n# Overplots polygon enclosing area of interest\nax = fig.add_subplot(111)\npatch = patches.PathPatch(path, lw=1, fill = False, color = 'r')\nax.add_patch(patch)\n\n# Determines which data points are inside area of interest\npoints = np.column_stack((data['pm_l_cosb'],data['pm_b']))\ninside = path.contains_points(points)\n\nfalse_indices = [i for i, x in enumerate(inside) if not x]\ndata.remove_rows(false_indices)\n\n\n############################ PLOTS STAR POSITIONS ##############################\n#\n## Plots star positions\n#plot_with_colourbar(data['ra'],data['dec'],data['phot_g_mean_mag'],'ra (deg)','dec (deg)','Location plot - OB2')\n#\n################################ PLOTS CAMDs ###################################\n## Removes empty data from bp-rp\n#masked_indices = [i for i, x in enumerate(data['bp_rp']) if np.ma.is_masked(x)]\n#data.remove_rows(masked_indices)\n#\n## Converts Gaia g-band Magnitudes to Absolute G Band Magnitudes\n#M_G = data['phot_g_mean_mag'] - 5*(np.log10(data['rest'])-1)\n#\n#bp_rp = data['bp_rp'] \n#mag_4_CAMD = M_G\n#\n## Plots Colour-Absolute Magnitude Diagram\n#plot_with_colourbar(bp_rp,mag_4_CAMD,mag_4_CAMD,'BP-RP','Gaia Absolute G-band Magnitude','Colour-Absolute Magnitude Diagram for stars in the vicinity of Stellar Association OB2',invert_y_axis = True, y_lim = (15,-5))\n#\n## Plotting CAMD density plot\n#fig2 = plt.figure()\n#k2 = kde.gaussian_kde([bp_rp, mag_4_CAMD])\n#nbins = 100\n#x2i, y2i = np.mgrid[bp_rp.min():bp_rp.max():nbins*1j, mag_4_CAMD.min():mag_4_CAMD.max():nbins*1j]\n#z2i = k2(np.vstack([x2i.flatten(), y2i.flatten()]))\n#plt.gca().invert_yaxis()\n#plt.gca().set_ylim(15,-5)\n#cmap = plt.cm.gist_ncar_r\n##cmaplist = [cmap(i) for i in range(cmap.N)]\n##cmaplist[0] = (1.,1.,1.,1.0)\n##cmap = cmap.from_list('Custom_cmap', cmaplist, cmap.N)\n#cs2 = plt.pcolormesh(x2i, y2i, z2i.reshape(x2i.shape), cmap = cmap)\n#plt.colorbar()\n#plt.xlabel('BP-RP')\n#plt.ylabel('Gaia Absolute G-band Magnitude')\n#plt.title('Colour-Absolute Magnitude Density Plot for stars in the vicinity of Stellar Association OB2')\n#\n## Defines polygon vertices and path for area of interest\n#verts_CAMD = [\n# (1., 3.5),\n# (1.3, 6.),\n# (2.2, 8.),\n# (4.4, 14.8),\n# (4.9, 10.5),\n# (2.2, 4.5),\n# (1., 3.5)\n# ]\n#path_CAMD = Path(verts_CAMD)\n#\n## Overplots polygon enclosing area of interest\n#ax = fig2.add_subplot(111)\n#patch_CAMD = patches.PathPatch(path_CAMD, lw=1, fill = False, color = 'r')\n#ax.add_patch(patch_CAMD)\n#\n####################### Re-plots PM Diagram for PMS stars ######################\n#\n## Determines which data points are inside area of interest\n#points2 = np.column_stack((bp_rp,mag_4_CAMD))\n#inside2 = path_CAMD.contains_points(points2)\n#\n#false_indices2 = [i for i, x in enumerate(inside2) if not x]\n#data.remove_rows(false_indices2)\n#\n## Plotting proper motion density plot\n#fig3 = plt.figure()\n#k3 = kde.gaussian_kde([data['pm_l_cosb'], data['pm_b']])\n#nbins = 200\n#x3i, y3i = np.mgrid[data['pm_l_cosb'].min():data['pm_l_cosb'].max():nbins*1j, data['pm_b'].min():data['pm_b'].max():nbins*1j]\n#z3i = k3(np.vstack([x3i.flatten(), y3i.flatten()]))\n#cs3 = plt.pcolormesh(x3i, y3i, z3i.reshape(x3i.shape), cmap=plt.cm.gist_ncar_r)\n#plt.colorbar()\n#plt.xlabel('pm_l_cosb (mas/yr)')\n#plt.ylabel('pm_b (mas/yr)')\n#plt.title('Proper motion plot for potential Pre-main-sequence Sco_OB2 members')\n#plt.scatter(hipparcos_data['pm_l_cosb'],hipparcos_data['pm_b'], 0.1, 'k')\n#\n## Standard proper motion plot with colorbar representing distance\n#plot_with_colourbar(data['pm_l_cosb'],data['pm_b'],data['rest'],'pm_l_cosb (mas/yr)','pm_b (mas/yr)','Proper motion plot for potential PMS Sco_OB2 members','Distance (pc)')\n#\n## Final location plot\n#plot_with_colourbar(data['ra'],data['dec'],data['phot_g_mean_mag'],'ra (deg)','dec (deg)','Location plot for potential PMS members of OB2')\n#\n############################## Save final table ################################\n#\n#data.write('Reduced_OB2_Data', format='votable')\n\nstop = timeit.default_timer()\nprint('Time: ',stop - start)","sub_path":"MPAGS code.py","file_name":"MPAGS code.py","file_ext":"py","file_size_in_byte":8398,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"605661540","text":"import numpy as np\nfrom gen_var import dr\nfrom numba import jit,float64\nimport time \n\n@jit(nopython = True)\ndef SOR(A,x,f,r):\n rel_tol = 1e-1\n iteration = 0 \n omega = 1.00\n r0 = 1e-3\n h = 1.0\n #omega = 2.0 / (1.0 + np.sin(np.pi * h))\n l = len(x)\n h = 1.0/l\n h = r[1] - r[0]\n\n res = np.ones(len(x))\n b = np.zeros(len(x))\n while (np.any(res > np.abs(np.multiply(rel_tol,x)))):\n for i in range(1, l-1):\n s = np.dot(A[i,:], x[:])\n xnew = -f[i]*h**2 - (s - A[i,i]*x[i]) - 2.0\n xold = x[i]\n x[i] = xnew*omega/A[i,i]\n res[i] = np.abs(xold - x[i])\n iteration += 1\n if (iteration%10000 == 0):\n #print(\"\\r\"+ str(iteration),end='')\n print(iteration)\n else:\n pass \n #print(\"\\r\"+ str(iteration))\n print(iteration)\n \"\"\"Determining answer via matrix inversion for comparison\n -can choose to remove this for larger matrices when comparison isn't needed.\"\"\"\n \n inv = np.linalg.inv(A)\n ans = np.dot(inv,-f*h*h - b)\n return x, ans \n","sub_path":"iterative_col.py","file_name":"iterative_col.py","file_ext":"py","file_size_in_byte":1090,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"265068306","text":"from source.newsApp.newsUi import newsUi, QtCore, QtWidgets\r\nfrom source.newsApp.checkBoxHandler import checkBoxHandler\r\nfrom source.newsApp.newsApiHandler import newsApiHandler\r\nfrom source.newsApp.country import country\r\nfrom source.extra.qtBasics import qtBasics\r\nfrom source.extra.fileIO import fileIO\r\n\r\nclass newsApp( qtBasics, newsUi ):\r\n\tquitSignalAccept = QtCore.pyqtSignal( int )\r\n\tquitSignalCancel = QtCore.pyqtSignal( int )\r\n\tquitSignalExit = QtCore.pyqtSignal( int )\r\n\r\n\t__countryFile = \"data/news/countryFile.txt\"\r\n\t__sourcesFile = \"data/news/sorucesFile.txt\"\r\n\t__categoryFile = \"data/news/categoryFile.txt\"\r\n\t__newsStoryFile = \"data/news/newsStory.txt\"\r\n\r\n\t__fileIO = fileIO( )\r\n\r\n\tdef __init__( self, x, y, sourceCount ): #367, 259\r\n\t\tnewsUi.__init__( self, x, y, sourceCount )\r\n\t\tqtBasics.__init__( self )\r\n\r\n\t\tself.__category = checkBoxHandler( self.allCheckBox, self.categoryCheckBox )\r\n\t\tself.__source = checkBoxHandler( self.allCheckBox2, self.sourceCheckBox )\r\n\r\n\t\tself.__country = country( self.comboBox )\r\n\r\n\t\tself.__apinews = newsApiHandler( self.progressBar )\r\n\r\n\tdef initUi( self ): \r\n\t\t#Connect\r\n\t\tself.buttonBox.button(QtWidgets.QDialogButtonBox.Apply).clicked.connect( self.applyClicked )\r\n\t\tself.buttonBox.button(QtWidgets.QDialogButtonBox.Cancel).clicked.connect( self.cancelClicked )\r\n\r\n\t\tself.__country.getCountryBox( ).currentIndexChanged.connect( self.countryChange )\r\n\t\tself.__country.setID( )\r\n\t\tself.countryChange( )\r\n\r\n\t\tself.__country.readFromFile( self.__countryFile )\r\n\t\tself.__category.readCheckedFromFile( self.__categoryFile )\r\n\t\tself.__source.readCheckedFromFile( self.__sourcesFile )\r\n\r\n\tdef applyClicked( self ):\r\n\t\tself.buttonBox.hide( )\r\n\r\n\t\tself.__apinews.newNewsRequest( self.__newsStoryFile )\r\n\t\tself.__category.writeCheckedToFile( self.__categoryFile )\r\n\t\tself.__source.writeCheckedToFile( self.__sourcesFile )\r\n\t\tself.__country.writeToFile( self.__countryFile )\r\n\r\n\t\tcategoryNames = self.__category.getCheckBoxNames( )\r\n\t\tsourceNames = self.__source.getCheckBoxNames( )\r\n\t\tif( len( categoryNames ) > 0 ):\r\n\t\t\tself.__apinews.pullCategory( categoryNames )\r\n\t\t\tself.__apinews.shuffleStory( )\r\n\t\t\tself.__apinews.writeToFile( self.__newsStoryFile )\r\n\r\n\t\telif( len( sourceNames ) > 0 ):\r\n\t\t\tself.__apinews.pullSource( sourceNames )\r\n\t\t\tself.__apinews.shuffleStory( )\r\n\t\t\tself.__apinews.writeToFile( self.__newsStoryFile )\r\n\r\n\t\tself.buttonBox.show( )\r\n\r\n\t\tself.quitSignalAccept.emit( 1 )\r\n\t\tself.setVisible( False )\r\n\r\n\tdef cancelClicked( self ):\r\n\t\tself.quitSignalCancel.emit( 1 )\r\n\t\tself.setVisible( False )\r\n\r\n\tdef countryChange( self ):\r\n\t\tsourceNames = []\r\n\t\tself.__apinews.setCountryID( self.__country.getID( ))\r\n\t\tfor names in self.__apinews.pullSourceNames( ):\r\n\t\t\tsourceNames.append( names['id'] )\r\n\t\t\r\n\t\tif( len( sourceNames )):\r\n\t\t\tself.label_2.hide( )# no sources available label\r\n\t\telse:\r\n\t\t\tself.label_2.show( )\r\n\r\n\t\tself.__source.changeCheckBox( sourceNames )\r\n\r\n\tdef setCountryName( self ):#call this in applyClicked\r\n\t\tself.__country.writeToFile( self.__countryFile )\r\n\r\n\tdef refresh( self ):\r\n\t\tself.__country.readFromFile( self.__countryFile )\r\n\t\tself.__source.readCheckedFromFile( self.__sourcesFile )","sub_path":"SmartMirrorControlPanel/source/newsApp/newsApp.py","file_name":"newsApp.py","file_ext":"py","file_size_in_byte":3171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"596892595","text":"\"\"\"Process resource.\"\"\"\nimport logging\nfrom time import sleep, time\n\nfrom resdk.exceptions import ResolweServerError\n\nfrom .base import BaseResource\n\n\nclass BackgroundTask(BaseResource):\n \"\"\"Background task resource.\n\n :param resolwe: Resolwe instance\n :type resolwe: Resolwe object\n :param model_data: Resource model data\n\n \"\"\"\n\n endpoint = \"task\"\n\n READ_ONLY_FIELDS = BaseResource.READ_ONLY_FIELDS + (\n \"started\",\n \"finished\",\n \"status\",\n \"description\",\n \"output\",\n )\n WRITABLE_FIELDS = ()\n\n def __init__(self, resolwe, **model_data):\n \"\"\"Initialize attributes.\"\"\"\n self.logger = logging.getLogger(__name__)\n\n #: started\n self.started = None\n #: finished\n self.finished = None\n #: status - Possible values:\n #: WA (waiting)\n #: PR (processing)\n #: OK (done)\n #: ER (error)\n self.status = None\n #: description\n self.description = None\n #: output - JSON field, the actual value depends on the background task\n self.output = None\n\n super().__init__(resolwe, **model_data)\n\n @property\n def completed(self) -> bool:\n \"\"\"Return True if the task is completed, False otherwise.\"\"\"\n return self.status in [\"OK\", \"ER\"]\n\n def wait(self, timeout: float = 0) -> \"BackgroundTask\":\n \"\"\"Wait for the background task to finish.\n\n The task status is retrieved every second.\n\n :attr timeout: how many seconds to wait for task to finish (0 to wait forever).\n\n :raise RuntimeError: when the task in not completed within the given timeout\n\n :return: the finished background task.\n \"\"\"\n start = time()\n while (timeout == 0 or time() - start < timeout) and not self.completed:\n sleep(1)\n self.update()\n if not self.completed:\n raise RuntimeError(f\"Waiting for taks {self.id} timeout.\")\n return self\n\n def result(self, timeout: float = 0, final_statuses=[\"OK\"]):\n \"\"\"Wait fot the background tast to finish and return its result.\n\n :attr timeout: how many seconds to wait for task to finish (0 to wait forever).\n :attr final_statuses: return the result when task status is in the list.\n\n :raise RuntimeError: when the task in not completed within the given timeout\n :raise ResolweServerError: when task state is is not in final statuses.\n\n :return: the output of the background task.\n \"\"\"\n self.wait(timeout)\n if self.status not in final_statuses:\n raise ResolweServerError(\n f\"Task status {self.status} not in {final_statuses} ({self.output}).\"\n )\n return self.output\n","sub_path":"src/resdk/resources/background_task.py","file_name":"background_task.py","file_ext":"py","file_size_in_byte":2765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"16804570","text":"#!/usr/bin/env python\n\nimport sys\nfrom copy import copy\nimport rospy\nimport actionlib\nimport math\nimport random\n\nfrom control_msgs.msg import (\n GripperCommandAction,\n GripperCommandGoal,\n)\n\nfrom sensor_msgs.msg import JointState\n\n\nclass GripperActionTest(object):\n def __init__(self,prefix=\"right\"):\n \n self._prefix = prefix\n self._client = actionlib.SimpleActionClient(\n '/movo/%s_gripper_controller/gripper_cmd'%self._prefix,\n GripperCommandAction,\n )\n self._goal = GripperCommandGoal()\n server_up = self._client.wait_for_server(timeout=rospy.Duration(10.0))\n if not server_up:\n rospy.logerr(\"Timed out waiting for Gripper Command\"\n \" Action Server to connect. Start the action server\"\n \" before running example.\")\n rospy.signal_shutdown(\"Timed out waiting for Action Server\")\n sys.exit(1)\n self.clear()\n\n def command(self, position, block=False, timeout=15.0):\n self._goal.command.position = position\n self._goal.command.max_effort = -1.0\n self._client.send_goal(self._goal)\n if block:\n self._client.wait_for_result(timeout=rospy.Duration(timeout))\n\n def stop(self):\n self._client.cancel_goal()\n\n def wait(self, timeout=15.0):\n self._client.wait_for_result(timeout=rospy.Duration(timeout))\n\n def result(self):\n return self._client.get_result()\n\n def clear(self):\n self._goal = GripperCommandGoal()\n\n\ndef main():\n rospy.init_node('gripper_close')\n\n rg_test = GripperActionTest(\"right\")\n lg_test = GripperActionTest(\"left\")\n \n \n rg_test.command(0.050)\n rg_test.wait()\n \n \n print(\"Gripper Action Test Example Complete\")\n \nif __name__ == \"__main__\":\n main()\n","sub_path":"rrs_ros/src/movo/my_grasping/src/gripper_close.py","file_name":"gripper_close.py","file_ext":"py","file_size_in_byte":1844,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"578586345","text":"import pygame\nimport numpy as np \nimport math\n\nclass OthelloDrawer:\n\n def __init__(self, board_size, demo_mode):\n \"\"\"Initalises the board drawing class\n \n Arguments:\n board_size {int} -- The dimension of the board\n demo_mode {bool} -- Determines if the game is being run in demo\n mode or is being play by a human player.\n \"\"\"\n self.board_size = board_size\n self.demo = demo_mode\n pygame.init()\n\n self.game_window = pygame.display.set_mode((self.board_size*80,\n self.board_size*80))\n\n pygame.display.set_caption(\"Othello\")\n\n def drawBoard(self, board):\n \"\"\"Draws the Othello game board using pygame\n \n Arguments:\n board {[[char]]} -- Array representing the current board state\n \"\"\"\n self.game_window.fill((0,157,0))\n for i in range(self.board_size):\n for j in range(self.board_size):\n rect = pygame.Rect(i*80,j*80,80,80)\n pygame.draw.rect(self.game_window, (0,0,0), rect, 5)\n\n piece_val = board[i][j]\n\n if piece_val == 'w':\n pygame.draw.circle(self.game_window, (255,255,255),\n (i * 80 + 40, j * 80 + 40), 30)\n elif piece_val == 'd':\n pygame.draw.circle(self.game_window, (0,0,0),\n (i * 80 + 40, j * 80 + 40), 30)\n elif piece_val == 'v':\n pygame.draw.circle(self.game_window, (255,255,255),\n (i * 80 + 40, j * 80 + 40), 5, 0)\n\n \n pygame.display.update()\n\n # If in demo mode, pause in order to properly visualise play\n if self.demo:\n pygame.time.delay(50)\n\n def getUserInput(self):\n \"\"\"Gets user input, either clicking on the board or exiting the game\n \n Returns:\n (int,int)) -- The coordinates of the input move. (-2,-2) if they\n select to quit.\n \"\"\"\n move = (-1,-1)\n pause = False\n while True:\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n move = (-2,-2)\n pygame.quit()\n elif event.type == pygame.MOUSEBUTTONDOWN:\n mouse_input = event.pos\n move = self.convertClickToMove(mouse_input) \n\n if pause == False:\n break\n return move \n\n def convertClickToMove(self, mouse_input):\n \"\"\"Converts the user input to a board coordinate\n \n Arguments:\n mouse_input {Mouse input event position} -- The coordinates of the click on screen\n \n Returns:\n (Integer, Integer) -- The coordinates of the square the user clicked\n \"\"\"\n move = (math.floor(mouse_input[0]/80), math.floor(mouse_input[1]/80))\n return move","sub_path":"Othello/othelloDraw.py","file_name":"othelloDraw.py","file_ext":"py","file_size_in_byte":3151,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"316659939","text":"import numpy as np\nclass SGD_Linear_Regression :\n def __init__(self,max_iter =100):\n self.max_iter = max_iter\n\n def _SGD(self,x,y):\n dim = x.ndim\n x_b = np.c_[np.ones((len(x),1)),x]\n dim = x.ndim\n if dim ==1 :\n raise ValueError('expected 2d ndarray given 1d ndarray')\n else:\n m,n = x.shape\n self.theta = np.zeros((n+1,1))\n self.theta[0][0] =1\n alpha = 0.0001\n for iter in range(self.max_iter) :\n for i in range(m) :\n self.theta = self.theta - alpha*(x_b[i].reshape(1,-1).dot(self.theta) - y[i])*(x_b[i].reshape(-1,1))\n\n\n def fit(self,x,y):\n self._SGD(x,y)\n return self\n\n def predict(self,x):\n dimension = x.ndim\n if dimension ==1 :\n raise ValueError('expected 2d array given 1d array')\n else :\n n_samples ,n_features = x.shape\n x_a = np.c_[np.ones((n_samples,1)),x]\n try:\n h = np.dot(x_a,self.theta)\n return h\n except :\n raise NotImplementedError('This LinearRegression is not fitted yet.')\n\n","sub_path":"stochastic_gradient_descent.py","file_name":"stochastic_gradient_descent.py","file_ext":"py","file_size_in_byte":1182,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"450822591","text":"#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n\n\"\"\"\n *.py: Description of what * does.\n Last Modified:\n\"\"\"\n\n__author__ = \"Debanjan Datta\"\n__email__ = \"ddatta@vt.edu\"\n__version__ = \"0.0.1\"\n\nimport gensim\nimport json\nimport spacy\nimport textacy\nimport numpy as np\nimport pickle\nimport os\nfrom bioFunctionPrediction.src.utils.dataloader import GODAG\n\n# ---Model config--- #\nWord_Embed_Size = 512\n# ---------------- #\n# MODE = 0 train the model\n# MODE = 1 fetch the embedding dict\n# ---------------- #\nMODE = 0\n# this file should have :\n# { _id[0...k]: np.array[shape = [Word_Embed_Size]] , ... }\n\nEMDED_FILE = 'GO_word_embed_dict.pkl'\nWord2vec_MODEL_FILE = 'word2vec_1.bin'\n\n\n# ------------------ #\ndef get_data():\n temp_json_2 = 'temp_data_2.json'\n with open(temp_json_2) as tmp_file:\n data_dict_2 = json.loads(tmp_file.read())\n return data_dict_2\n\n\ndef train():\n global Word_Embed_Size\n global Word2vec_MODEL_FILE\n\n data_dict_2 = get_data()\n sentences = []\n for k, v in data_dict_2.items():\n sentences.append(v)\n model = gensim.models.Word2Vec(\n sentences,\n iter=25,\n window=4,\n size=Word_Embed_Size,\n workers=8,\n min_count=1\n )\n print('Model', model)\n model.save(Word2vec_MODEL_FILE)\n\n\ndef load_model():\n global Word2vec_MODEL_FILE\n # load model\n model = gensim.models.Word2Vec.load(Word2vec_MODEL_FILE)\n print(model)\n return model\n\n\ndef create_embed_dict():\n global Word_Embed_Size\n global MODE\n\n GODAG_obj = GODAG()\n GODAG_obj.initialize_idmap(None, None)\n idmap = GODAG_obj.idmap\n\n def _format(k):\n return k.replace('GO:', '')\n\n idmap = {_format(k): v for k, v in idmap.items()}\n\n if MODE == 0:\n train()\n\n model = load_model()\n emb_dict = {}\n data_dict = get_data()\n words = model.wv.vocab\n\n for k, sent in data_dict.items():\n sent_vec = np.zeros([Word_Embed_Size])\n for w in sent:\n try:\n vec = np.array(model.wv.word_vec(w))\n sent_vec = sent_vec + vec\n except:\n test = w in words\n print('Word not found ', w, 'in Vocab ', test)\n try:\n key_id = idmap[k]\n emb_dict[k] = sent_vec\n except:\n print('Key not found in idmap ', key_id)\n print(emb_dict.keys())\n return emb_dict\n\n\ndef initialize():\n global EMDED_FILE\n global MODE\n\n if MODE == 0:\n res = create_embed_dict()\n with open(EMDED_FILE, 'wb') as handle:\n pickle.dump(res, handle, protocol=pickle.HIGHEST_PROTOCOL)\n elif MODE == 1:\n if os.path.isfile(EMDED_FILE) :\n with open(EMDED_FILE, 'rb') as handle:\n res = pickle.load(handle)\n else :\n res = create_embed_dict()\n return res\n\ndef setup():\n global MODE\n initialize()\n MODE = 1\n\nsetup()\n\n# ------------------------ #\n# Use this function to extract the embedding dictionary\n\ndef get_id_embed_dict():\n return initialize()\n\nget_id_embed_dict()","sub_path":"src/models/word2vec_1/word2vec.py","file_name":"word2vec.py","file_ext":"py","file_size_in_byte":3061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"18105885","text":"# Problem A. Tic-Tac-Toe-Tomek\n# https://code.google.com/codejam/contest/2270488/dashboard#s=p0\nimport sys\n\ndef is_completed(board):\n for i in range(len(board)):\n for j in range(len(board)):\n if board[i][j] == '.':\n return False\n return True\n\ndef lines(board):\n rows = [row for row in board]\n columns = []\n for i in range(len(board)):\n column = ''.join([board[j][i] for j in range(len(board))])\n columns.append(column)\n diagonals = [\n ''.join([board[i][i] for i in range(len(board))]),\n ''.join([board[len(board) - i - 1][i] for i in range(len(board))])\n ]\n return rows + columns + diagonals\n\ndef is_completed_by(line, player):\n return len([c for c in line if c in [player, 'T']]) == len(line)\n\ndef tictactoe(board):\n for line in lines(board):\n for player in ['X', 'O']:\n if is_completed_by(line, player):\n return '{0} won'.format(player)\n if not is_completed(board):\n return 'Game has not completed'\n return 'Draw'\n\nT = int(sys.stdin.readline())\nfor t in range(1, T + 1):\n board = []\n for i in range(4):\n board.append(sys.stdin.readline().strip())\n sys.stdin.readline()\n print('Case #{0}: {1}'.format(t, tictactoe(board)))","sub_path":"2012/tictactoe.py","file_name":"tictactoe.py","file_ext":"py","file_size_in_byte":1278,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"496873075","text":"\nimport logging\nimport os\nimport shutil\nfrom json import JSONDecodeError\nfrom multiprocessing import Pool, cpu_count, current_process\nimport json\n\nfrom peframe.peframe import analyze\n\nLOGGER = logging.getLogger(__name__)\nlogging.basicConfig(format='%(asctime)s %(message)s')\n\n\ndef analysis_to_json(arguments):\n try:\n filename = arguments[0]\n outputfile = arguments[1]\n if filename is None:\n raise Exception(\"Missing Filename\")\n LOGGER.info(\"[\" + str(current_process()) + \"] Analysing file \" + filename)\n features = analyze(filename)\n if outputfile is not None:\n LOGGER.info(\"[\" + str(current_process()) + \"] Writing file of \" + outputfile)\n with open(outputfile, 'w') as fp:\n json.dump(features, fp)\n return features\n except Exception as e:\n print(\"Impossible to analyze \" + filename)\n print(e)\n if \"non_apt\" in filename:\n dest = \"/home/giuseppe/binaries/problematic_non_apt/\"\n else:\n dest = \"/home/giuseppe/binaries/problematic_apt/\"\n shutil.move(filename, dest+os.path.basename(filename))\n\n\ndef check_file(path):\n try:\n with open(path, \"r\") as inputfile:\n json.load(inputfile)\n except JSONDecodeError:\n os.remove(path)\n\ndef check_analysis_list(analysis_folder):\n list_file = [analysis_folder + x for x in os.listdir(analysis_folder)]\n pool = Pool(cpu_count()-2)\n pool.map(check_file, list_file)\n\ndef analyze_folder(binary_folder, analysis_folder, parallel=True):\n LOGGER.info(\"Analyzing folder \"+binary_folder)\n filename_list = os.listdir(binary_folder)\n check_analysis_list(analysis_folder)\n analysis_list = os.listdir(analysis_folder)\n clean_analyis_list = [x[:-len(\".json\")] for x in analysis_list]\n to_analyze = [item for item in filename_list if item not in clean_analyis_list]\n arguments_list = [(binary_folder + x, analysis_folder + x + \".json\") for x in to_analyze]\n\n if parallel:\n pool = Pool(processes=cpu_count() - 2)\n pool.map(analysis_to_json, arguments_list)\n else:\n for argument in arguments_list:\n analysis_to_json(argument)\n\ndef main():\n\n binary_folder = \"/home/giuseppe/binaries/apt/\"\n result_folder = \"/home/giuseppe/analysis/apt/\"\n analyze_folder(binary_folder, result_folder, True)\n\n binary_folder = \"/home/giuseppe/binaries/non_apt/\"\n result_folder = \"/home/giuseppe/analysis/non_apt/\"\n analyze_folder(binary_folder, result_folder, True)\n\n\nmain()\n\n\n# path = \"/home/giuseppe/binaries/problematic_apt/b8e728703604afde97716309b260a611\"\n# dest = \"/tmp/1.json\"\n# analysis_to_json((path, dest))\n","sub_path":"analyze_folders.py","file_name":"analyze_folders.py","file_ext":"py","file_size_in_byte":2690,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"324296957","text":"from gvcfSample import gvcfSample\nimport sys\n\n#Constants for indexes of ALT base and Info strings in the line\nALT_IDX = 4\nINFO_IDX = 7\n\n\nclass gvcfLine:\n \"\"\"\n Represents a vcf line\n\n Alt reads/ref reads not calculated correctly for sites with more than two haplotypes\n \"\"\"\n\n def __init__(self,raw_line):# -> None\n \"\"\"\n Instantiate the vcfLine object with associated information\n \"\"\"\n self.infoValues = {\"AC\":None,\"AF\":None,\"BaseQRankSum\":None,\"DP\":None,\\\n \"Dels\":None,\"FS\":None,\"HaplotypeScore\":None,\"InbreedingCoeff\":None,\\\n \"MLEAC\":None,\"MLEAF\":None,\"MQ\":None,\"MQ0\":None,\"MQRankSum\":None,\\\n \"QD\":None,\"ReadPosRankSum\":None,\"SOR\":None,\\\n \"ClippingRankSum\":None,\"NCC\":None,\"GQ_MEAN\":None,\\\n \"GQ_STDDEV\":None} #ClippingRankSum on are specific to gVCFs \n\n self.chrom,self.pos,self.ref,self.alt,self.qual,self.filt = None,None,\\\n None,None,None,None\n\n self.isDataLine = True\n self.isAltSite = False\n\n self.CC_samples = []\n self.GP_samples = []\n\n self.raw_line = raw_line\n self.multiHap = False\n\n #Calculate read totals while creating vcfSample objects\n self.altTotal = 0\n self.refTotal = 0\n self.otherTotal = 0\n\n #Determines if the line is represents a site \n if len(self.raw_line) > 1:\n if self.raw_line[0] == \"#\":\n self.isDataLine = False \n \n #For lines at sites process site info\n if self.isDataLine:\n sline = str.split(self.raw_line) \n \n #Alt sites have a different format for samples, determine if the\n # site has been called as heterozygous\n if sline[ALT_IDX] != \".\":\n #variant site\n self.isAltSite = True\n \n #Sites with more than one alternate allele have a different\n # format for samples\n if \",\" in sline[ALT_IDX]:\n self.multiHap = True\n\n #Loads the vcfSample objects for each sample info at this site\n self.__loadSamples(sline) \n \n #Load all info data about this site\n self.__loadInfoVals()\n\n #Load other data not contained in the INFO annotation\n self.chrom = sline[0]\n self.filt = sline[6]\n self.pos = sline[1]\n self.ref = sline[3]\n self.alt = sline[4]\n self.qual = sline[5]\n if self.qual == \".\":\n self.qual = \"0\"\n #If qual data is present cast as a float\n if str.isdigit(self.qual):\n self.qual = float(self.qual)\n \n \n\n def __loadSamples(self,sline):# -> None\n \"\"\"\n Instantiates vcfSample objects for each sample at this site.\n Adds all samples to self.Samples list\n \"\"\"\n #For this vcf format samples are from item 9 -> in the line\n format = sline[8]\n\n CC_sampleStrings = sline[9:14] + sline[15:20] + sline[21:23]\n GP_sampleStrings = sline[23:]\n\n for s in CC_sampleStrings:\n sample = gvcfSample(s,self.isAltSite,format) \n self.CC_samples.append(sample)\n \n #Add this samples alt/ref read counts to the total at this site\n self.altTotal += sample.getAltReads()\n self.refTotal += sample.getRefReads()\n self.otherTotal += sample.getOtherReads()\n\n for s in GP_sampleStrings:\n sample = gvcfSample(s,self.isAltSite,format) \n self.GP_samples.append(sample)\n\n self.altTotal += sample.getAltReads()\n self.refTotal += sample.getRefReads()\n self.otherTotal += sample.getOtherReads()\n \n def __loadInfoVals(self):# -> None\n \"\"\"\n Pareses the info data from this line storing it in self.infoValues\n \"\"\"\n splitInfo = (str.split(self.raw_line)[INFO_IDX]).split(\";\")\n for readInfo in splitInfo:\n #populates self.infoValues for every value present\n eqIdx = readInfo.find(\"=\")\n tag = readInfo[0:eqIdx] \n value = (readInfo[eqIdx + 1:])\n #Converts numeric info values to float\n if str.isdigit(value):\n value = float(value)\n \n #Add info value to self.infoValues\n self.infoValues[tag] = value \n \n \n def isAltSite(self):# -> bool\n return self.isAltSite \n\n def repr(self):# -> str\n return \"info %s\\n,chrom %s\\n,pos %s\\n,ref %s\\n,alt %s\\n,qual %s\\n,filt %s\\n,isData %s\\n,rawline %s\\n\" \\\n %(str(self.infoValues),str(self.chrom),str(self.pos),\\\n str(self.ref),str(self.alt),str(self.qual),\\\n str(self.filt),str(self.isDataLine),str(self.raw_line))\n\n def getRawLine(self):# -> str\n \"\"\"\n Returns the raw vcf line that this vcfLine obj represents. \n None --> String\n \"\"\"\n return self.raw_line\n\n\n def getAltTotal(self):# -> int\n \"\"\"\n Calculates and returns the number of alternate \n reads from all samples at this site.\n None --> int\n \"\"\"\n altTotal = 0\n for s in self.samples:\n if isinstance(s.altReads,float):\n altTotal += s.altReads\n else:\n pass\n\n return altTotal\n\n\n def getRefTotal(self):# -> int\n \"\"\"\n Return the ref read count total for all samples at this site\n \"\"\"\n return self.refTotal\n\n def getAltTotal(self):# -> int\n \"\"\"\n Return the alt read count total for all samples at this site\n \"\"\"\n return self.altTotal\n\n\n def getOtherTotal(self):# -> int \n \"\"\"\n Return the other read count total for all samples at this site\n \"\"\"\n return self.otherTotal\n\n\n","sub_path":"gvcfLine.py","file_name":"gvcfLine.py","file_ext":"py","file_size_in_byte":5972,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"205063063","text":"\n\"\"\"\n単純にやると全探索の組み合わせ数がO(M^2)になるが,実際にはO(M)のみを調べれば良い.\nよって,探索候補がM以上ある場合には,探索せずに答えを求めることができる.\n\"\"\"\n\nimport sys\nfrom collections import defaultdict\n\nH, W, M = map(int, input().split())\nX = [list(map(int, input().split())) for _ in range(M)]\n\nctr_h = defaultdict(int)\nctr_w = defaultdict(int)\nappeared = set()\n\nfor h, w in X:\n ctr_h[h] += 1\n ctr_w[w] += 1\n appeared.add(h * W + w)\n\nval_h = sorted(ctr_h.items(), key=lambda x: -x[1])\nval_w = sorted(ctr_w.items(), key=lambda x: -x[1])\n\nmax_h = val_h[0][1]\nmax_w = val_w[0][1]\n\nval_h = [i for i, v in val_h if v == max_h]\nval_w = [i for i, v in val_w if v == max_w]\n\nif len(val_h) * len(val_w) > M:\n print(max_h + max_w)\n sys.exit(0)\n\nfor h in val_h:\n for w in val_w:\n if h * W + w not in appeared:\n print(max_h + max_w)\n sys.exit(0)\nelse:\n print(max_h + max_w - 1)\n","sub_path":"contest_src/abc176/e.py","file_name":"e.py","file_ext":"py","file_size_in_byte":1001,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"376927110","text":"import pandas as pd \n\nxl = pd.ExcelFile('trinhcong.xlsx')\ndf = pd.read_excel(xl, 0, header=None)\n\nfor i in range(2, 12):\n print(df.at[i, 11], '\\t\\t', df.at[i, 12])\n\"\"\" 171.244.166.18\n113.185.76.196\n \"\"\"\n ","sub_path":"testex.py","file_name":"testex.py","file_ext":"py","file_size_in_byte":207,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"602192439","text":"'''\n5\n4\n192\n48\n206\n37\n147\n90\n312\n21\n186\n12\n121\n38\n114\n21\n408\n39\n267\n13\n382\n29\nOutput: C has won\n\n\n5\n4\n192\n148\n206\n37\n147\n190\n312\n21\n186\n312\n121\n38\n114\n121\n408\n39\n267\n113\n382\n29\nOutput: Runoff between A and C\n'''\n\nbooths = int(input())\ncandidates = int(input())\nmatrix = [[0 for i in range(candidates)] for i in range(booths)]\ndict = {0:'A', 1:'B', 2:'C', 3:'D'}\ntotal = 0\ncandVotes = []\nfor i in range(booths):\n for j in range(candidates):\n matrix[i][j] = int(input())\n total += matrix[i][j]\n\nprint(total)\nfor j in range(candidates):\n votes = 0\n for i in range(booths):\n votes += matrix[i][j]\n candVotes.append(votes)\n\nprint(candVotes)\n\nflag = 0\nfor i in range(len(candVotes)):\n if candVotes[i] > total/2:\n flag = 1\n indexOfWinner = i\nif flag:\n print(f\"{dict[indexOfWinner]} has won!\")\nelse:\n highest = max(candVotes)\n i = candVotes.index(highest)\n candVotes[i] = 0\n secondHighest = max(candVotes)\n j = candVotes.index(secondHighest)\n print(f\"Runoff between {dict[j]} and {dict[i]}\")\n\n\n\n\n\n\n","sub_path":"Talentio/electionResults.py","file_name":"electionResults.py","file_ext":"py","file_size_in_byte":1061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"352202624","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('main', '0001_initial'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='application',\n name='cadastral_number',\n field=models.CharField(blank=True, max_length=30, null=True, verbose_name='Кадастровый номер'),\n preserve_default=True,\n ),\n ]\n","sub_path":"main/migrations/0002_application_cadastral_number.py","file_name":"0002_application_cadastral_number.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"350353164","text":"#=========================================\n#\n# Rossler system\n#\n#\n#=========================================\nimport random\nimport numpy as np\na = 0.2\nb = 0.2\nc = 5.7\n\n\ndelta_t = 0.0001\ninitial_t = 0\nfinal_t = 10\ninitial_val = [1.0, 1.0, 1.0]\nmodel_name = \"Rossler\"\ninformation = \"Rossler\" + \"(a, b, c) = (\" + str(a) + \", \" + str(b) + \", \" + str(c) + \")\"\n\n\ndef f(state, t):\n x, y, z = state\n return - y - z, x + a * y, b + z * (x - c)\n\n\ndef Jf(state):\n x, y, z = state\n #return Delta_t * np.matrix([[-sigma, sigma, 0], [(rho - z), -1, -x], [y, x, -beta]]) + np.eye(3)\n return np.matrix([[1, -delta_t, -delta_t], \n [delta_t, a * delta_t + 1 , 0], \n [delta_t * z, 0, (x - c) * delta_t + 1]])\n","sub_path":"model/Rossler.py","file_name":"Rossler.py","file_ext":"py","file_size_in_byte":828,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"568077200","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Nov 27 09:55:12 2018\n\n@author: Ben Tomhave\n\"\"\"\n\n#### TASK B: CALCULATE POPULATION DENSITY ####\n\nfrom os import listdir\nimport os\nimport arcpy\nimport fiona\nfrom collections import defaultdict\n#------------------------------------------------------------------------------ \n## 0. DEFINE FUNCTIONS and PATH NAMES\ndef getFileNamesInFolder(FolderPath, suffix=\".csv\" ):\n '''\n In: Csv folder path\n Out: List of files in the folder path (default for CSV's)\n '''\n filenames = listdir(FolderPath)\n return [file for file in filenames if file.endswith(suffix)]\n\n\ncsvPath = r\"C:\\Users\\Ben Tomhave\\Documents\\GitHub\\Wrap_Up_2\\csv\"\nworkspacePath=r\"C:\\Users\\Ben Tomhave\\Desktop\\FinalWrapUp2\"\narcpy.env.workspace = workspacePath\narcpy.env.overwriteOutput = True\n\n#------------------------------------------------------------------------------ \n## 1. CONVERT CSV TO DBF FILE and LOAD SHAPEFILES\n#Get CSV Files\ncsvFileList=getFileNamesInFolder(csvPath)\n\n#Convert each CSV to DBF\nfor file in csvFileList:\n inputTable=r\"%s\\%s\" %(csvPath,file)\n arcpy.TableToDBASE_conversion (inputTable, workspacePath)\n\n#Print tables and fields associated with each table \ntables = arcpy.ListTables() \n#print('\\n\\n Tables and associated fields\\n-----------------------------------')\n#for i in range(len(tables)):\n# print('\\n',tables[i])\n# for f in arcpy.ListFields(tables[i]):\n# print('\\t',f.name)\n#print('----------------------------------------')\n\n\n#Find shapefiles and print the fields associated with each file\nfeatureClasses = arcpy.ListFeatureClasses()\n#print('\\n\\n Shapefiles and associated fields\\n-------------------------------')\n#for i in range(len(featureClasses)):\n# print(featureClasses[i])\n# for f in arcpy.ListFields(featureClasses[i]):\n# print('\\t',f.name)\n#print('----------------------------------------')\n\n#------------------------------------------------------------------------------ \n## 2 & 3. JOIN CSV FIELDS TO SHAPEFILES and ADD NEW FIELDS FOR CALCULATIONS\nshapeFileList=['US_county_2010.shp','US_state_2010.shp']\njoinTableList=[ 'nhgis0002_ds172_2010_county.dbf', 'nhgis0002_ds172_2010_state.dbf']\noutputField1=['CoPopDen','StPopDen']\noutputField2=['CoFracPopS','StFracPopC']\njoinField1=\"GISJOIN\"\njoinField2=\"GISJOIN\"\n\n\nprint('\\nAdding Field...\\n')\noutputLayers=[]\nfor i,shape in enumerate(shapeFileList):\n inputShapeTable=r\"%s\\%s\" %(workspacePath,shapeFileList[i])\n outLayer=shape[0:13]+'_lyr'\n outputLayers.append(outLayer)\n arcpy.MakeFeatureLayer_management(os.path.join(workspacePath,shapeFileList[i]),outLayer) \n arcpy.AddJoin_management(outLayer,joinField1,joinTableList[i],joinField2)\n arcpy.AddField_management(outLayer,outputField1[i],'Double')\n arcpy.AddField_management(outLayer,outputField2[i],'Double')\n arcpy.AddField_management(outLayer,'StatePop','Double')\n arcpy.AddField_management(outLayer,'CountryPop','Double')\n\n\n#------------------------------------------------------------------------------ \n## 4. CALCULATE Densities\n\nprint('\\nCalculating Field...\\n')\ninputDBFs=['nhgis0002_ds172_2010_county.dbf','nhgis0002_ds172_2010_state.dbf']\nregionPopField='H7V001'\nregionAreaField='Shape_area' #Naturally comes as m^2\nstatePopField='StatePop'\noutputCalculationShpList=[]\nfor i,layer in enumerate(outputLayers):\n fieldName=('%s.%s' %(shapeFileList[i][:-4],outputField1[i]))\n densityExpression=('(!%s.%s!/!%s.%s!)*1000*1000' %(inputDBFs[i][:-4],regionPopField,shapeFileList[i][:-4],regionAreaField))\n \n #Above expression is equivalent to (population/(Shape_area (in m^2)))*1000^2= density [people/km^2]\n arcpy.management.CalculateField(outputLayers[i], fieldName, densityExpression, \"PYTHON3\", None)\n\n fractionalPopExpression=('(!%s.%s!/!%s.%s!)*1000*1000' %(inputDBFs[i][:-4],regionPopField,shapeFileList[i][:-4],statePopField))\n\n\n print('\\nConverting Updated %s to Shape file...\\n' %(outputLayers[i]))\n outputShapeFileWithCalculations=str('Calculated_'+shapeFileList[i][:-4])\n outputCalculationShpList.append(outputShapeFileWithCalculations)\n arcpy.FeatureClassToFeatureClass_conversion(outputLayers[i], workspacePath, outputShapeFileWithCalculations) \n\n\n\n#------------------------------------------------------------------------------ \n## 4. CALCULATE Total State and Country Population\n#Calculate Each States' Population as a sum of the counties within the state and append to dictionary\n \n \narcpy.AddField_management('US_county_2010.shp','StatePOP','Double')\n\n \nstatePopulation = defaultdict(int) \nwith arcpy.da.SearchCursor('nhgis0002_ds172_2010_county.dbf', ['STATEA','H7V001']) as cursor: \n for state, indivCountyPopulation in cursor: \n statePopulation[state] +=indivCountyPopulation \n\n#Calculate the population of the country by adding the population of each state\ncountryPopulation = sum(statePopulation.values())\nkeyList=list(statePopulation.keys())\n\n#Create Field and Update For State Population\nfor fc in range(len(shapeFileList)):\n print(fc)\n with arcpy.da.UpdateCursor(shapeFileList[fc], ['STATEFP10','StatePop']) as cursor:\n for row in cursor:\n for i in range(len(statePopulation)):\n #Need to have the second critera below otherwise single digits aren't evaluated b/c 02!=2 for example\n if (row[0]==str(keyList[i])) or (row[0]==str('0'+str(keyList[i]))) : \n row[1] = statePopulation.get((keyList[i]))\n #print(row[1])\n\n cursor.updateRow(row)\n \n","sub_path":"taskB.py","file_name":"taskB.py","file_ext":"py","file_size_in_byte":5611,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"524643439","text":"import serial\nimport time\nimport datetime\n\nDWM=serial.Serial(port=\"/dev/tty.usbmodem0007601032201\", baudrate=115200)\nprint(\"Connected to \" +DWM.name)\nDWM.write(\"\\r\\r\".encode())\ntime.sleep(1)\nDWM.write(\"lec\\r\\n\".encode())\ntime.sleep(1)\nwhile True:\n try:\n line=DWM.readline()\n print(line.decode())\n if (line):\n if len(line) >= 140:\n parse = line.decode().split(\",\")\n x_pos = parse[parse.index(\"POS\") + 1]\n y_pos = parse[parse.index(\"POS\") + 2]\n val = (x_pos, y_pos)\n mycursor = mydb.cursor()\n mycursor.execute(sql, val)\n mydb.commit()\n print(datetime.datetime.now().strftime(\"%H:%M:%S\"), \"(\", x_pos, \", \", y_pos, \")\")\n else:\n print(\"\")\n except Exception as ex:\n print(ex)\n break\nDWM.write(\"\\r\".encode())\nDWM.close()","sub_path":"raspberry_capteurs.py","file_name":"raspberry_capteurs.py","file_ext":"py","file_size_in_byte":911,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"525829790","text":"from libra.ledger_info import LedgerInfo\r\nfrom libra.validator_verifier import VerifyError\r\nfrom libra.hasher import *\r\nfrom libra.proof import verify_transaction_list\r\nfrom libra.proof.signed_transaction_with_proof import SignedTransactionWithProof\r\nfrom libra.proof.account_state_with_proof import AccountStateWithProof\r\nfrom libra.proof.event_with_proof import EventWithProof\r\nfrom libra.transaction import SignedTransaction, TransactionInfo\r\nfrom libra.account_address import Address\r\nfrom libra.proof import ensure, bail\r\nfrom libra.account_resource import AccountResource\r\nimport canoser\r\n\r\n\r\ndef verify(validator_verifier, request, response):\r\n verify_update_to_latest_ledger_response(\r\n validator_verifier,\r\n request.client_known_version,\r\n request.requested_items,\r\n response.response_items,\r\n response.ledger_info_with_sigs\r\n )\r\n\r\ndef verify_update_to_latest_ledger_response(\r\n validator_verifier,\r\n req_client_known_version,\r\n requested_items,\r\n response_items,\r\n ledger_info_with_sigs\r\n ):\r\n ledger_info_proto = ledger_info_with_sigs.ledger_info\r\n ledger_info = LedgerInfo.from_proto(ledger_info_proto)\r\n signatures = ledger_info_with_sigs.signatures\r\n if ledger_info.version < req_client_known_version:\r\n raise VerifyError(f\"ledger_info.version:{ledger_info.version} < {req_client_known_version}.\")\r\n if ledger_info.version > 0 or signatures.__len__() > 0:\r\n validator_verifier.batch_verify_aggregated_signature(ledger_info.hash(), signatures)\r\n if len(response_items) != len(requested_items):\r\n raise VerifyError(f\"{len(response_items)} != {len(requested_items)}\")\r\n for req_item, resp_item in zip(requested_items, response_items):\r\n verify_response_item(ledger_info, req_item, resp_item)\r\n\r\ndef verify_response_item(ledger_info, requested_item, response_item):\r\n req_type = requested_item.WhichOneof('requested_items')\r\n if not req_type.endswith(\"_request\"):\r\n raise VerifyError(f\"RequestItem type unknown{req_type}.\")\r\n resp_type = req_type.replace(\"_request\", \"_response\")\r\n resp_type2 = response_item.WhichOneof('response_items')\r\n if resp_type != resp_type2:\r\n raise VerifyError(f\"RequestItem/ResponseItem types mismatch:{resp_type} - {resp_type2}.\")\r\n if resp_type == \"get_account_state_response\":\r\n asp = response_item.get_account_state_response.account_state_with_proof\r\n AccountStateWithProof.verify(asp, ledger_info, ledger_info.version,\r\n requested_item.get_account_state_request.address)\r\n elif resp_type == \"get_account_transaction_by_sequence_number_response\":\r\n atreq = requested_item.get_account_transaction_by_sequence_number_request\r\n atresp = response_item.get_account_transaction_by_sequence_number_response\r\n verify_get_txn_by_seq_num_resp(\r\n ledger_info,\r\n atreq.account,\r\n atreq.sequence_number,\r\n atreq.fetch_events,\r\n atresp.signed_transaction_with_proof,\r\n atresp.proof_of_current_sequence_number\r\n )\r\n elif resp_type == \"get_events_by_event_access_path_response\":\r\n ereq = requested_item.get_events_by_event_access_path_request\r\n eresp = response_item.get_events_by_event_access_path_response\r\n verify_get_events_by_access_path_resp(\r\n ledger_info,\r\n ereq.access_path,\r\n ereq.start_event_seq_num,\r\n ereq.ascending,\r\n ereq.limit,\r\n eresp.events_with_proof,\r\n eresp.proof_of_latest_event\r\n )\r\n elif resp_type == \"get_transactions_response\":\r\n req = requested_item.get_transactions_request\r\n ver = req.start_version\r\n limit = req.limit\r\n fetch_events = req.fetch_events\r\n txp = response_item.get_transactions_response.txn_list_with_proof\r\n verify_get_txns_resp(ledger_info, ver, limit, fetch_events, txp)\r\n else:\r\n raise VerifyError(f\"unknown response type:{resp_type}\")\r\n\r\n\r\ndef verify_get_txn_by_seq_num_resp(\r\n ledger_info,\r\n account,\r\n sequence_number,\r\n fetch_events,\r\n signed_transaction_with_proof,\r\n proof_of_current_sequence_number\r\n ):\r\n has_stx = len(signed_transaction_with_proof.__str__()) > 0\r\n has_cur = len(proof_of_current_sequence_number.__str__()) > 0\r\n if has_stx and not has_cur:\r\n ensure(\r\n fetch_events == signed_transaction_with_proof.HasField(\"events\"),\r\n \"Bad GetAccountTxnBySeqNum response. Events requested: {}, events returned: {}.\",\r\n fetch_events,\r\n signed_transaction_with_proof.HasField(\"events\")\r\n )\r\n SignedTransactionWithProof.verify(\r\n signed_transaction_with_proof,\r\n ledger_info,\r\n signed_transaction_with_proof.version,\r\n account,\r\n sequence_number\r\n )\r\n elif has_cur and not has_stx:\r\n sequence_number_in_ledger = AccountResource.get_account_resource_or_default(\r\n proof_of_current_sequence_number.blob).sequence_number\r\n ensure(\r\n sequence_number_in_ledger <= sequence_number,\r\n \"Server returned no transactions while it should. Seq num requested: {}, latest seq num in ledger: {}.\",\r\n sequence_number,\r\n sequence_number_in_ledger\r\n )\r\n AccountStateWithProof.verify(proof_of_current_sequence_number, ledger_info,\r\n ledger_info.version, account)\r\n else:\r\n bail(\r\n \"Bad GetAccountTxnBySeqNum response. txn_proof.is_none():{}, cur_seq_num_proof.is_none():{}\",\r\n has_stx,\r\n has_cur\r\n )\r\n\r\n\r\n\r\n\r\ndef verify_get_events_by_access_path_resp(\r\n ledger_info,\r\n req_access_path,\r\n req_start_seq_num,\r\n req_ascending,\r\n req_limit,\r\n events_with_proof,\r\n proof_of_latest_event,\r\n ):\r\n account_resource = AccountResource.get_account_resource_or_default(proof_of_latest_event.blob)\r\n AccountStateWithProof.verify(proof_of_latest_event, ledger_info, ledger_info.version,\r\n req_access_path.address)\r\n event_handle = account_resource.get_event_handle_by_query_path(req_access_path.path)\r\n expected_event_key = event_handle.key\r\n expected_seq_nums = gen_events_resp_idxs(event_handle.count,\r\n req_start_seq_num, req_ascending, req_limit)\r\n ensure(\r\n len(expected_seq_nums) == len(events_with_proof),\r\n \"Expecting {} events, got {}.\",\r\n len(expected_seq_nums),\r\n len(events_with_proof)\r\n )\r\n zipped = zip(events_with_proof, expected_seq_nums)\r\n for event_with_proof, seq_num in zipped:\r\n EventWithProof.verify(\r\n event_with_proof,\r\n ledger_info,\r\n expected_event_key,\r\n seq_num,\r\n event_with_proof.transaction_version,\r\n event_with_proof.event_index\r\n )\r\n\r\n\r\ndef gen_events_resp_idxs(seq_num_upper_bound, req_start_seq_num, req_ascending, req_limit):\r\n if not req_ascending and req_start_seq_num == canoser.Uint64.max_value and seq_num_upper_bound > 0:\r\n cursor = seq_num_upper_bound - 1\r\n else:\r\n cursor = req_start_seq_num\r\n if cursor >= seq_num_upper_bound:\r\n return [] #Unreachable, so empty.\r\n elif req_ascending:\r\n #Ascending, from start to upper bound or limit.\r\n realupper = min(cursor + req_limit, seq_num_upper_bound)\r\n return [x for x in range(cursor, realupper)]\r\n elif cursor + 1 < req_limit:\r\n return [x for x in range(cursor, -1, -1)] # Descending and hitting 0.\r\n else:\r\n bottom = cursor + 1 - req_limit\r\n return [x for x in range(cursor, bottom-1, -1)] #Descending and hitting limit.\r\n\r\n\r\ndef verify_get_txns_resp(ledger_info, start_version, limit, fetch_events, txn_list_with_proof):\r\n if limit == 0 or start_version > ledger_info.version:\r\n if txn_list_with_proof.SerializeToString() != b'':\r\n raise VerifyError(f\"transactions should be empty.\")\r\n return\r\n if fetch_events != txn_list_with_proof.HasField(\"events_for_versions\"):\r\n raise VerifyError(f\"fetch_events: {fetch_events} mismatch with events_for_versions\")\r\n num_txns = len(txn_list_with_proof.transactions)\r\n ret_num = min(limit, ledger_info.version - start_version + 1)\r\n if num_txns != ret_num:\r\n raise VerifyError(f\"transaction number expected:{ret_num}, returned:{num_txns}.\")\r\n verify_start_version(txn_list_with_proof, start_version)\r\n verify_transaction_list(txn_list_with_proof, ledger_info)\r\n\r\ndef verify_start_version(txn_list_with_proof, start_version):\r\n ver = txn_list_with_proof.first_transaction_version.value\r\n if ver != start_version:\r\n raise VerifyError(f\"transaction version mismatch:{start_version}, returned:{ver}.\")\r\n","sub_path":"libra/get_with_proof.py","file_name":"get_with_proof.py","file_ext":"py","file_size_in_byte":8890,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"362968205","text":"from base.base_page import BasePage\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as ec\n\n\nclass LookUpCountryWindow(BasePage):\n def __init__(self, driver):\n super().__init__(driver)\n self.driver = driver\n\n # LOCATORS\n _country_field = \"COUNTRY_TBL_COUNTRY\"\n _search_result = \"SEARCH_RESULT1\"\n _look_up_btn = \"#ICSearch\"\n\n def select_country(self, country_code):\n self.driver.switch_to.default_content()\n self.util.sleep(2, \"'Look Up Country' window to open.\")\n try:\n iframe = self.driver.find_element(By.XPATH, \"//iframe[contains(@id, 'ptModFrame_')]\")\n self.driver.switch_to.frame(iframe)\n except Exception as e:\n print(e)\n\n self.sendkeys(country_code, self._country_field)\n self.element_click(self._look_up_btn)\n self.util.sleep(2, str(country_code) + \" to be found.\")\n self.element_click(self._search_result)\n self.util.sleep(2, \"popup window to close\")\n self.driver.switch_to.default_content()\n\n wait = WebDriverWait(self.driver, 10, poll_frequency=1)\n wait.until(ec.visibility_of_element_located((By.ID, \"ptifrmtgtframe\")))\n\n try:\n # iframe = self.driver.find_element(By.XPATH, \"//iframe[contains(@id, 'ptifrmtgtframe')]\")\n # self.driver.switch_to.frame(iframe)\n self.driver.switch_to.frame(\"ptifrmtgtframe\")\n except Exception as e:\n print(e)\n","sub_path":"popup_windows/look_up_country_window.py","file_name":"look_up_country_window.py","file_ext":"py","file_size_in_byte":1568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"54734431","text":"from functools import wraps\nfrom flask import Blueprint, request, render_template, flash, g, session, redirect, url_for\n\nfrom app import db\nfrom app import app\nfrom app.users.forms import LoginForm, CreateForm\nfrom app.models import User, Ticket, TicketUpdate\n\n##\n# Define our Blueprint\n##\nuser_mod = Blueprint('users', __name__, url_prefix='/user')\n\n##\n# Define control decorators\n##\ndef login_required(f):\n @wraps(f)\n def decorated_function(*args, **kwargs):\n if 'user_id' not in session:\n flash(\"Login Required!\", 'error')\n return redirect(url_for('users.login'))\n return f(*args, **kwargs)\n return decorated_function\n\ndef admin_user_restricted(f):\n @wraps(f)\n def decorated_function(*args, **kwargs):\n if session['user_id'] == 'first_run':\n return f(*args, **kwargs)\n\n elif 'user_id' in session:\n user = User.query.get(session['user_id'])\n if not user.isAdmin():\n flash(\"Access Denied\", 'error')\n return redirect(url_for('index'))\n elif 'user_id' not in session:\n flash(\"Login Required!\", 'error')\n return redirect(url_for('users.login'))\n\n return f(*args, **kwargs)\n return decorated_function\n\n\n##\n# Define routes / control funcitons\n##\n@user_mod.route('/login', methods=['GET', 'POST'])\ndef login():\n if 'user_id' in session:\n flash('You are already logged in!', 'notice')\n return redirect(url_for('users.profile', userid=session['user_id']))\n ##\n # If the sign in form is submitted\n ##\n form = LoginForm(request.form)\n\n ##\n # Verify the sign in form\n ##\n if request.method == 'POST' and form.validate():\n user = User.query.filter_by(email=form.email.data).first()\n if user and user.checkPassword(form.password.data):\n session['user_id'] = user.id\n session['user_name'] = user.name\n flash(\"Welcome %s\" % user.name, 'success')\n app.logger.info(\"%s logged in\" % user.email)\n return redirect(url_for('users.profile', userid=user.id))\n flash('Invalid Username or Password!', 'error')\n app.logger.info(\"Login Failed for %s\" % form.email.data)\n return render_template(\"users/signin.html\", form=form)\n\n@user_mod.route('/logout', methods=['GET'])\ndef logout():\n if 'user_name' in session:\n session.clear()\n flash('You are now logged out', 'success')\n return redirect(url_for('index'))\n\n@user_mod.route('/id/')\ndef profile(userid):\n user = User.query.get(userid)\n if user == None:\n return render_template(\"404.html\")\n else:\n return render_template(\"users/profile.html\", user=user)\n\n@user_mod.route('/create', methods=['GET', 'POST'])\n@admin_user_restricted\n@login_required\ndef create():\n ##\n # Get the submitted form\n ##\n form = CreateForm(request.form)\n\n ##\n # handle the creation\n ##\n if request.method == 'POST' and form.validate():\n if User.query.filter_by(email=form.email.data).count() >= 1:\n flash(\"That email address is already in use!\", 'error')\n return render_template(\"users/create.html\", form=form)\n else:\n new_user = User(name=form.name.data,\n email=form.email.data)\n new_user.setPassword(form.password.data)\n # First User gets admin role\n if User.query.count() == 0:\n new_user.role=0\n\n db.session.add(new_user)\n db.session.commit()\n\n user = User.query.filter_by(email=form.email.data).first()\n session['user_id'] = user.id\n session['user_name'] = user.name\n flash('Welcome %s' % user.name, 'success')\n return redirect(url_for('users.profile', userid=user.id))\n else:\n return render_template(\"users/create.html\", form=form)\n\n@user_mod.route('/dash')\n@login_required\ndef dash():\n ##\n # Display the appropriate Dashboard\n ##\n\n data['user'] = User.query.get(session['user_id'])\n data['tickets'] = Ticket.query.filter_by(created_by = user.id)\n data['assigned'] = Ticket.query.filter_by(assigned_to = user.id)\n\n if user.isAdmin():\n return render_template(\"users/admin-dash.html\")\n elif user.isTech():\n return render_template(\"users/tech-dash.html\")\n else:\n return render_template(\"users/user-dash.html\")\n","sub_path":"app/users/controllers.py","file_name":"controllers.py","file_ext":"py","file_size_in_byte":4435,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"424199104","text":"# html解析器\nimport re\nimport urllib.parse\nfrom lxml import etree\n\n# html解析器\nclass HtmlParser(object):\n\n def parse(self, page_url, content):\n '''\n 用于解析网页内容抽取URL和数据\n :param page_url: 下载页面的URL\n :param content: 下载的网页内容\n :return:返回URL和数据\n '''\n if page_url is None or content is None:\n return\n html = etree.HTML(content.encode(\"utf-8\"))\n new_urls = self._get_new_urls(page_url,html)\n new_data = self._get_new_data(page_url,html)\n return new_urls,new_data\n\n\n def _get_new_urls(self, page_url, html_xpath):\n '''\n 抽取新的URL集合\n :param page_url: 下载页面的URL\n :param html_xpath:xpath doc\n :return: 返回新的URL集合\n '''\n new_urls = set()\n\n # 贴吧页面\n links = html_xpath.xpath('//*[@id=\"frs_list_pager\"]//@href')\n links.extend(html_xpath.xpath('//*[@class=\"thread_list_bottom clearfix\"]//@href'))\n print(\"debug HtmlParser->links=\", len(links))\n for new_url in links:\n #拼接成完整网址\n new_full_url = urllib.parse.urljoin(page_url,new_url)\n new_urls.add(new_full_url)\n return new_urls\n def _get_new_data(self, page_url, html_xpath):\n '''\n 抽取有效数据\n :param page_url:下载页面的URL\n :param html_xpath:\n :return:返回有效数据\n '''\n imgs = html_xpath.xpath('//*[@class=\"threadlist_pic j_m_pic \"]/@bpic')\n print(\"debug HtmlParser->imgs=\", len(imgs))\n return imgs\n","sub_path":"ProjectForUse/协线进程分布式爬取百度贴吧图片/爬虫基本框架-百度贴吧图片爬取/spider-xpath定位不能精确/HtmlParser.py","file_name":"HtmlParser.py","file_ext":"py","file_size_in_byte":1651,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"179582213","text":"#!/usr/bin/python3\n#\n# Copyright (C) 2019 Trinity College of Dublin, the University of Dublin.\n# Copyright (c) 2019 Li Jian\n# Author: Li Jian \n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n\"\"\"\nThis sends a fib list request to the local NFD and prints the response.\nThis is equivalent to the NFD command line command \"nfd-status -r\".\nSee http://redmine.named-data.net/projects/nfd/wiki/Management .\n\"\"\"\n\nimport time\nfrom pyndn import Face\nfrom pyndn import Name\nfrom pyndn import Interest\nfrom pyndn.util import Blob\nfrom pyndn.encoding import ProtobufTlv\nfrom pyndn.util.segment_fetcher import SegmentFetcher\n# This moudle is produced by: protoc --python_out=. fib-entry.proto\nfrom status import fib_entry_pb2\n\n\nclass Fib_status_getter(object):\n def __init__(self):\n self.total_result = ''\n\n def dump(self, *list):\n result = \"\"\n for element in list:\n result += (element if type(element) is str else str(element)) + \" \"\n self.total_result = self.total_result + result + \" \\n\"\n\n def run(self):\n # The default Face connects to the local NFD.\n face = Face()\n\n interest = Interest(Name(\"/localhost/nfd/fib/list\"))\n interest.setInterestLifetimeMilliseconds(4000)\n self.dump(\"Express interest\", interest.getName().toUri())\n\n enabled = [True]\n\n def onComplete(content):\n enabled[0] = False\n self.printFibEntries(content)\n\n def onError(errorCode, message):\n enabled[0] = False\n self.dump(message)\n\n SegmentFetcher.fetch(face, interest, None, onComplete, onError)\n\n # Loop calling processEvents until a callback sets enabled[0] = False.\n while enabled[0]:\n face.processEvents()\n\n # We need to sleep for a few milliseconds so we don't use 100% of the CPU.\n time.sleep(0.01)\n\n # print('==================run FIB_status_getter finished===================')\n face.shutdown()\n return (self.total_result)\n\n def printFibEntries(self, encodedMessage):\n \"\"\"\n This is called when all the segments are received to decode the\n encodedMessage as repeated TLV FibEntry messages and display the values.\n\n :param Blob encodedMessage: The repeated TLV-encoded FibEntry.\n \"\"\"\n fibEntryMessage = fib_entry_pb2.FibEntryMessage()\n ProtobufTlv.decode(fibEntryMessage, encodedMessage)\n\n self.dump(\"FIB:\");\n for fibEntry in fibEntryMessage.fib_entry:\n line = \"\"\n line += ProtobufTlv.toName(fibEntry.name.component).toUri()\n\n # Show the routes.\n for nexthop in fibEntry.next_hop_records:\n line += (\" NextHopRecord={faceId=\" + str(nexthop.face_id) + \" cost=\" + str(nexthop.cost))\n line += \")}\"\n\n self.dump(line)\n","sub_path":"status/fib_status_getter.py","file_name":"fib_status_getter.py","file_ext":"py","file_size_in_byte":3366,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"477157152","text":"import matplotlib.pyplot as plt\nimport math\nimport sys\n\nout_file = sys.argv[1]\nin_1 = sys.argv[2]\nin_2 = sys.argv[3]\nin_3 = sys.argv[4]\nin_4 = sys.argv[5]\nend = int(sys.argv[6])\n\ninds = [i for i in range(1, end)]\npt = [2**i for i in range(1, end)]\n\nf = open(in_1)\nss = []\nfor line in f.readlines():\n if int(line.split()[0]) in pt:\n ss.append(float(line.split()[1]))\n \nf = open(in_2)\nsparring1 = []\nfor line in f.readlines():\n if int(line.split()[0]) in pt:\n sparring1.append(float(line.split()[1]))\n \nf = open(in_3)\nsparring2 = []\nfor line in f.readlines():\n if int(line.split()[0]) in pt:\n sparring2.append(float(line.split()[1]))\n\nf = open(in_4)\nisss = []\nfor line in f.readlines():\n if int(line.split()[0]) in pt:\n isss.append(float(line.split()[1]))\n\nplt.plot(inds, ss)\nplt.plot(inds, sparring1)\nplt.plot(inds, sparring2)\nplt.plot(inds, isss)\nplt.title(\"Regret vs. log2 t\")\nax = plt.gca()\nplt.legend(['Self sparring', 'Sparring1', 'Sparring2', 'IS-SS'], bbox_to_anchor=(1.4, 1.0), bbox_transform=ax.transAxes)\nplt.savefig(out_file, bbox_inches='tight')\n","sub_path":"thompson/tsregret.py","file_name":"tsregret.py","file_ext":"py","file_size_in_byte":1111,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"182103193","text":"\"\"\"\nPure python helpers\n\nMeant to help with python interaction.\n\"\"\"\n\n# python imports\nimport traceback\nimport subprocess\nfrom threading import Timer\nfrom Queue import Queue, Empty\n\n\n__all__ = [\n 'python',\n 'executeInMainThread'\n]\n\nTHREAD_QUEUE = Queue()\n\n\ndef executeInMainThread(func, *args, **kwargs):\n \"\"\"\n Execute a function in the main thread.\n :param func:\n :param args:\n :param kwargs:\n :return:\n \"\"\"\n THREAD_QUEUE.put((func, args, kwargs))\n\n\ndef _main_thread_execute():\n \"\"\"\n --PRIVATE--\n :return:\n \"\"\"\n try:\n func, args, kwargs = THREAD_QUEUE.get(timeout=0.01)\n try:\n func(*args, **kwargs)\n except Exception:\n traceback.print_exc()\n THREAD_QUEUE.task_done()\n except Empty:\n pass\n\n timer = Timer(0.5, _main_thread_execute)\n timer.setDaemon(True)\n timer.start()\n# launch timed listener for main thread\n_main_thread_execute()\n\n\ndef python(pathToFile):\n \"\"\"\n Launch an app or file with python.\n :param pathToFile:\n :return:\n \"\"\"\n arg = ['python', pathToFile]\n subprocess.Popen(arg)\n","sub_path":"HFX/hfx_py/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":1122,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"267731307","text":"import os\nimport json\nimport re\nimport sublime\nimport sublime_plugin\nfrom os import path\nfrom collections import deque\nimport pprint\n\nNODEDIR = path.dirname(__file__) + \"/nodelib\"\nNAME_ALIES_FILE = NODEDIR + \"/name_alies.txt\"\nDEBUG = False\npp = pprint.PrettyPrinter(indent=4)\n\n\nclass Nodejs():\n def __init__(self):\n self.data = deque()\n self.nameAlies = None\n self.loaded = False\n self.__loadNameAlies()\n\n def parseNode(self):\n \"\"\"\n load file from a folder\n parse to json\n replace alies names\n generate completions\n \"\"\"\n files = os.listdir(NODEDIR)\n for f in files:\n if not f.endswith('.json'):\n continue\n with open(NODEDIR + \"/\" + f, encoding='UTF-8') as fi:\n j = json.load(fi)\n self.__dealDict(j)\n\n self.__dealAliesName(self.data)\n # modify trigger as trigger\\t{parent}.{type}\n self.loaded = True\n if self.data:\n for dic in self.data:\n dic['trigger'] = \"{0}\\t{1}.{2}\".format(dic['trigger'], dic['parent'], dic['type'])\n print(\"nodejs completions loaded\")\n\n def __loadNameAlies(self):\n with open(NAME_ALIES_FILE) as fi:\n nameAlies = [line.split(',') for line in fi.readlines()]\n nameAlies = [(x.strip(), y.strip()) for x, y in nameAlies]\n self.nameAlies = nameAlies\n # print(self.nameAlies)\n\n def __dealDict(self, obj, parent=None):\n if (isinstance(obj, dict) and (\n 'modules' in obj or 'classes' in obj or 'methods'in obj\n or 'properties' in obj or 'events' in obj)):\n for k, v in obj.items():\n if isinstance(v, dict):\n self.__dealDict(v, obj)\n elif isinstance(v, list):\n self.__dealList(v, obj, k)\n elif isinstance(v, str):\n # print \"%s.%s=%s\" % (parent, k, v)\n pass\n\n if 'type' in obj and 'name' in obj and 'textRaw' in obj:\n if obj['type'] == 'module':\n self.__dealModule(obj, parent)\n if obj['type'] == 'classe':\n self.__dealClass(obj, parent)\n if obj['type'] == 'method':\n self.__dealMethod(obj, parent)\n if obj['type'] == 'propertie':\n self.__dealProperties(obj, parent)\n if obj['type'] == 'event':\n self.__dealEvent(obj, parent)\n\n def __dealList(self, list, parent, type):\n for v in list:\n if isinstance(v, dict):\n self.__dealDict(v, parent)\n elif isinstance(v, list):\n self.__dealList(v, parent)\n elif isinstance(v, str):\n # print \"%s=%s\" % parent, v\n pass\n\n def __dealModule(self, md, parent):\n # print(\"var %s = require(\\\"%s\\\");\" % (md['name'], md['name']))\n parentName = 'nodejs'\n if parent and 'name' in parent:\n parentName = parent['name']\n snippets = {\n \"content\": \"var {0} = require('{1}');\".format(md['name'], md['name']),\n \"doc\": md['desc'],\n \"trigger\": \"require{0}\".format(md['name']),\n \"type\": 'module',\n \"parent\": parentName\n }\n self.data.append(snippets)\n\n def __dealMethod(self, md, parent):\n m = md['textRaw']\n match = re.match(r'([a-zA-Z_0-9.]+)(.*)', m)\n if match and len(match.groups()) == 2:\n mname = match.group(1)\n pname = match.group(2)\n pnames = re.findall(r'([a-zA-Z_0-9.]+)', pname)\n pnames2 = [\"${{{0}:{1}}}\".format(i+1, v) for i, v in enumerate(pnames)]\n snippets = {\n \"content\": \"{0}({1})\".format(mname, ', '.join(pnames2)),\n \"doc\": md['desc'],\n \"trigger\": mname,\n \"type\": 'method',\n \"parent\": parent['name']\n }\n self.data.append(snippets)\n\n def __dealProperties(self, md, parent):\n snippets = {\n \"content\": \"{0}.{1}\".format(parent['name'], md['name']),\n \"doc\": md['desc'],\n \"trigger\": \"{0}.{1}\".format(parent['name'], md['name']),\n \"type\": 'properties',\n \"parent\": parent['name']\n }\n self.data.append(snippets)\n\n def __dealEvent(self, md, parent):\n eFunc = re.match('

(.*)', md['desc'])\n eFunc = eFunc and eFunc.group(1) or 'function() {{}}'\n snippets = {\n \"content\": '{0}.on(\"{1}\", {2});'.format(parent['name'], md['name'], eFunc),\n \"doc\": md['desc'],\n \"trigger\": '{0}.on{1}'.format(parent['name'], md['name']),\n \"type\": 'event',\n \"parent\": parent['name']\n }\n self.data.append(snippets)\n\n def __dealClass(self, md, parent):\n snippets = {\n \"content\": \"{0}\".format(md['name']),\n \"doc\": md['desc'],\n \"trigger\": \"{0}\".format(md['name']),\n \"type\": 'class',\n \"parent\": parent['name']\n }\n self.data.append(snippets)\n\n def __dealAliesName(self, snippets):\n for snippet in snippets:\n trigger = snippet['trigger']\n content = snippet['content']\n for x, y in self.nameAlies:\n trigger = trigger.replace(x, y)\n content = content.replace(x, y)\n snippet['trigger'] = trigger\n snippet['content'] = content\n \nnodejs = Nodejs()\nif DEBUG:\n print(\"=======\" * 20)\n nodejs.parseNode()\n for snippets in nodejs.data:\n pp.pprint(snippets['trigger'])\n pp.pprint(snippets['content'])\n\n\nclass NodejsCompleteListener(sublime_plugin.EventListener):\n def __isNodeJsView(self, view):\n return 'nodejs' in view.scope_name(0)\n\n def on_post_save(self, view):\n pass\n\n def on_load(self, view):\n if self.__isNodeJsView(view) and not nodejs.loaded:\n nodejs.parseNode()\n\n def on_activated(self, view):\n if self.__isNodeJsView(view) and not nodejs.loaded:\n nodejs.parseNode()\n\n def on_query_completions(self, view, prefix, locations):\n \"\"\"\n add completions to the editer\n \"\"\"\n if self.__isNodeJsView(view):\n # view.show_popup(\n # decodeHtmlentities(nodejs.data[0]['doc']),\n # flags=sublime.COOPERATE_WITH_AUTO_COMPLETE)\n curline = view.substr(view.line(view.sel()[0])).strip(' ;')\n # print(\"curline\", curline)\n func = re.match(r'\\b([a-zA-Z0-9.]+)\\b', curline)\n return [\n (snippets['trigger'], snippets['content'])\n for snippets in nodejs.data if snippets['trigger'].startswith(func.group(1) if func else prefix)]\n\n def on_modified_async(self, view):\n \"\"\"\n in this method, it will show the document\n \"\"\"\n global docShowed\n if self.__isNodeJsView(view):\n curline = view.substr(view.line(view.sel()[0])).strip(' ;')\n # print(\"curline\", curline)\n # test match --- var xxx = abc (aa, bb)\n func = re.match(r'(var\\s+\\w+\\s*=\\s*[a-zA-Z0-9.]+)', curline)\n if func:\n func = func.group(1)\n else:\n func = re.match(r'\\b([a-zA-Z0-9.]+)\\b', curline)\n if func:\n func = func.group(1)\n else:\n if view.is_popup_visible():\n view.hide_popup()\n return\n params = re.match(r'\\((.*)\\)?', curline)\n if params:\n params = params.group(1)\n func = \"{0}({1}\".format(func, params)\n # print(\"params\", params)\n # print('func', func)\n if curline and curline.endswith(')'):\n docs = [snippet['doc'] for snippet in nodejs.data if func in snippet['content']]\n if docs:\n view.show_popup(\n decodeHtmlentities(docs[0]),\n flags=sublime.COOPERATE_WITH_AUTO_COMPLETE, max_width=600, max_height=400)\n docShowed = True\n\n\ndef decodeHtmlentities(string):\n entity_re = re.compile(\"&(#?)(\\d{1,5}|\\w{1,8});\")\n\n def substitute_entity(match):\n from html.entities import name2codepoint as n2cp\n ent = match.group(2)\n if match.group(1) == \"#\":\n return chr(int(ent))\n else:\n cp = n2cp.get(ent)\n\n if cp:\n return chr(cp)\n else:\n return match.group()\n\n return entity_re.subn(substitute_entity, string)[0]\n# match = re.match(r'([a-zA-Z_0-9.]+)(.*)', 'request.write(chunk[, encoding][, callback])')\n# print(match.groups())\n# pm = re.findall(r'([a-zA-Z_0-9.]+\\]?)', match.group(2))\n# print(pm)\n","sub_path":"nodejs.py","file_name":"nodejs.py","file_ext":"py","file_size_in_byte":8951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"96735112","text":"\nimport math\nimport torch\nimport logging\nfrom typing import Tuple, Dict, List\nfrom replay.replay import BufferFields\nfrom agent.models.policy import PolicyNetwork, QNetwork\nfrom torch.nn.parallel import DistributedDataParallel as DDP\n\nclass Gaussian:\n '''\n Tanh squashed Gaussian Distribution\n '''\n\n def __init__(self, mean: torch.Tensor, log_std: torch.Tensor, squash: bool=True, eval_mode: bool=False):\n self._mean = mean\n min_log_std = -20\n max_log_std = 2\n self._log_std = torch.clamp(log_std,\n min=min_log_std,\n max=max_log_std)\n if eval_mode:\n self._std = 0\n else:\n self._std = torch.exp(self._log_std)\n\n self._dim = mean.shape[1]\n self._squash = squash\n\n def sample(self) -> Tuple[torch.Tensor, torch.Tensor]:\n '''\n Draw a sample from Gaussian distribution\n '''\n\n noise = self._std * torch.normal(torch.zeros_like(self._mean), torch.ones_like(self._mean))\n sample = self._mean + noise\n\n log_pi = self.loglikelihood(sample)\n\n if self._squash :\n sample = torch.tanh(sample)\n\n return sample, log_pi\n\n def loglikelihood(self, samples: torch.Tensor) -> torch.Tensor:\n '''\n Compute log likelihood of samples\n '''\n EPS = 1e-8\n z = (samples - self._mean) / (self._std + EPS)\n loglikelihood = -(torch.sum(self._log_std + 0.5 * z ** 2, dim=-1, keepdim=True)\n + 0.5 * self._dim * math.log(2 * math.pi))\n\n # because of squash\n if self._squash:\n loglikelihood -= torch.sum(torch.log((1 - torch.tanh(samples) ** 2) + EPS), dim=-1, keepdim=True)\n\n return loglikelihood\n\nclass SACAgent:\n '''\n This class implemented soft actor critic agent which can be used\n in sac algorithm. It includes a policy network and 2 twin Q network.\n '''\n def __init__(self,\n device_id: int,\n world_size: int,\n policy_hidden_size: List,\n q_hidden_size: List,\n model_path: str=None):\n\n self._device_id = device_id\n self._world_size = world_size\n\n self._logger = logging.getLogger()\n\n self._pi = PolicyNetwork(\n state_size=BufferFields['state'],\n action_size=BufferFields['action'],\n hidden_sizes=policy_hidden_size).to(device_id)\n\n q_param = {\n 'state_size': BufferFields['state'],\n 'action_size': BufferFields['action'],\n 'hidden_sizes': q_hidden_size}\n\n self._q1 = QNetwork(**q_param).to(device_id)\n self._q2 = QNetwork(**q_param).to(device_id)\n self._q1_target = QNetwork(**q_param).to(device_id)\n self._q2_target = QNetwork(**q_param).to(device_id)\n\n if model_path:\n self.load_model(model_path)\n self._logger.info(f'load agent model from {model_path}')\n\n # wrap models after init and load\n self._pi = self._ddp_wrap(self._pi)\n self._q1 = self._ddp_wrap(self._q1)\n self._q2 = self._ddp_wrap(self._q2)\n\n self._logger.info(self._pi)\n self._logger.info(self._q1)\n self._logger.info(self._q2)\n self._logger.info(self._q1_target)\n self._logger.info(self._q2_target)\n\n self._eval_mode = False\n self._init_policy_std = 0.75\n\n def eval_mode(self, eval_mode: bool) -> None:\n '''\n Set network as evaluation mode.\n Only works for some type of network architecture\n '''\n self._eval_mode = eval_mode\n if eval_mode:\n self._pi.eval()\n else:\n self._pi.train()\n\n def pi_parameters(self) -> Dict:\n '''\n Return pi net paramenters\n '''\n return self._pi.parameters()\n\n def q1_parameters(self) -> Dict:\n '''\n Return q1 net parameters\n '''\n return self._q1.parameters()\n\n def q2_parameters(self) -> Dict:\n '''\n Return q2 net parameters\n '''\n return self._q2.parameters()\n\n def pi(self, state: torch.Tensor, use_init_std: bool=False) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]:\n '''\n Sample an action using policy net and gaussian distribution\n '''\n state = state.to(self._device_id)\n mu, log_std = self._pi(state)\n if use_init_std:\n log_std = torch.ones_like(log_std)*self._init_policy_std\n distribution = Gaussian(mu, log_std, eval_mode=self._eval_mode)\n action, log_pi = distribution.sample()\n return mu, log_std, action, log_pi\n\n def q(self, state: torch.Tensor, action: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]:\n '''\n Return twin q values\n '''\n state = state.to(self._device_id)\n action = action.to(self._device_id)\n q1 = self._q1(state, action)\n q2 = self._q2(state, action)\n return q1, q2\n\n def q_target(self, state: torch.Tensor, action: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]:\n '''\n Return target twin q values\n '''\n state = state.to(self._device_id)\n action = action.to(self._device_id)\n q1_target = self._q1_target(state, action)\n q2_target = self._q2_target(state, action)\n return q1_target, q2_target\n\n def update_q_target(self, update_tau: float) -> None:\n '''\n Soft update target q-network parameters\n '''\n q1_state_dict = self._q1.state_dict()\n q2_state_dict = self._q2.state_dict()\n self._soft_update_state_dict(self._q1_target, q1_state_dict, update_tau)\n self._soft_update_state_dict(self._q2_target, q2_state_dict, update_tau)\n\n def save_model(self, path: str) -> None:\n '''\n Save models without ddp wrapper to the given path\n '''\n param = {}\n param['q1'] = self._strip_ddp_state_dict(self._q1.state_dict())\n param['q2'] = self._strip_ddp_state_dict(self._q2.state_dict())\n param['pi'] = self._strip_ddp_state_dict(self._pi.state_dict())\n torch.save(param, path)\n\n def load_model(self, path: str) -> None:\n '''\n Load model from given path without ddp wrapper,\n and assuming always load cuda:0 model, so we map\n cuda:0 to current assigned device id.\n '''\n map_location = {'cuda:0': f'cuda:{self._device_id}'}\n param = torch.load(path, map_location=map_location)\n\n self._q1.load_state_dict(param['q1'])\n self._q2.load_state_dict(param['q2'])\n self._q1_target.load_state_dict(param['q1'])\n self._q2_target.load_state_dict(param['q2'])\n self._pi.load_state_dict(param['pi'])\n\n def _ddp_wrap(self, model: torch.nn.Module) -> torch.nn.Module:\n '''\n Wrapper network module using DistributedDataParallel object\n if the world size is larger than 1.\n '''\n if self._world_size > 1:\n return DDP(model, device_ids=[self._device_id])\n else:\n return model\n\n def _soft_update_state_dict(self, model: torch.nn.Module, state_dict: Dict, tau: float=1) -> None:\n '''\n Soft update state dict of model given state dict and tau\n '''\n state_dict = self._strip_ddp_state_dict(state_dict)\n if tau == 1:\n model.load_state_dict(state_dict)\n elif tau > 0:\n update_sd = {k: tau * state_dict[k] + (1 - tau) * v for k, v in model.state_dict().items()}\n model.load_state_dict(update_sd)\n\n def _strip_ddp_state_dict(self, state_dict: Dict) -> Dict:\n '''\n DistributedDataParallel prepends 'module.' to every key,\n but for the general purpose, we want to save and load\n state dict without prepended key.\n '''\n clean_state_dict = type(state_dict)() \n for k, v in state_dict.items(): \n key = k[7:] if k[:7] == \"module.\" else k \n clean_state_dict[key] = v \n return clean_state_dict\n\n","sub_path":"agent/sac_agent.py","file_name":"sac_agent.py","file_ext":"py","file_size_in_byte":8064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"23687450","text":"import os, io, json, sys\n\nif sys.platform == 'linux':\n import pwd\nfrom helpers import SpikeGLX_utils\n\nimport numpy as np\n\ndef create_samba_directory(samba_server, samba_share):\n\n if sys.platform == 'linux':\n proc_owner_uid = str(pwd.getpwnam(os.environ['USER']).pw_uid)\n share_string = 'smb-share:server={},share={}'.format(samba_server, samba_share)\n data_dir = os.path.join('/', 'var', 'run', 'user', proc_owner_uid, 'gvfs', share_string)\n else:\n data_dir = r'\\\\' + os.path.join(samba_server, samba_share)\n\n return data_dir\n\ndef createInputJson(output_file, \n npx_directory=None, \n continuous_file = None,\n spikeGLX_data=True,\n input_meta_path=None,\n extracted_data_directory=None,\n kilosort_output_directory=None,\n ks_make_copy=False,\n probe_type='3A',\n catGT_run_name='test',\n gate_string='0',\n trigger_string='0,0',\n probe_string='0',\n depth_est_fig = 0,\n catGT_stream_string = '-ap',\n catGT_car_mode = 'gbldmx',\n catGT_loccar_min_um = 40,\n catGT_loccar_max_um = 160,\n catGT_cmd_string = '-prb_fld -out_prb_fld',\n catGT_maxZ_um = -1,\n noise_template_use_rf = True,\n event_ex_param_str = 'XD=4,1,50',\n tPrime_im_ex_list = 'SY=0,384,6,500',\n tPrime_ni_ex_list = 'XA=0,1,3,500',\n sync_period = 1.0,\n toStream_sync_params = 'SY=0,384,6,500',\n niStream_sync_params = 'XA=0,1,3,500',\n tPrime_3A = False,\n toStream_path_3A = None,\n fromStream_list_3A = None,\n ks_doFilter = 0,\n ks_remDup = 0, \n ks_finalSplits = 1,\n ks_labelGood = 1,\n ks_saveRez = 1,\n ks_copy_fproc = 0,\n ks_minfr_goodchannels = 0.1, \n ks_whiteningRadius_um = 163,\n ks_Th = '[10,4]',\n ks_CSBseed = 1,\n ks_LTseed = 1,\n ks_templateRadius_um = 163,\n ks_nblocks = 5,\n ks_CAR = 0,\n ks_output_tag = 'ks2',\n c_Waves_snr_um = 160,\n wm_spread_thresh = 0.12,\n wm_site_range = 16,\n qm_isi_thresh = 1.5/1000,\n include_pcs = True\n ):\n\n # hard coded paths to code on your computer and system\n ecephys_directory = r'C:\\Users\\colonellj\\Documents\\ecephys_anaconda\\ecephys_spike_sorting\\ecephys_spike_sorting'\n \n # location of kilosor respository and kilosort version\n\n kilosort_repository = r'C:\\Users\\colonellj\\Documents\\KS2_largetemplate\\Kilosort2'\n\n KS2ver = '2.0' # must equal '3.0', '2.5' or '2.0', and match the kiilosort_repository\n \n # KS 3.0 does not yet output pcs.\n if KS2ver == '3.0':\n include_pcs = False # set to false for KS2ver = '3.0'\n \n npy_matlab_repository = r'C:\\Users\\colonellj\\Documents\\npy-matlab-master'\n catGTPath = r'C:\\Users\\colonellj\\Documents\\CatGT-win'\n tPrime_path=r'C:\\Users\\colonellj\\Documents\\TPrime-win'\n cWaves_path=r'C:\\Users\\colonellj\\Documents\\C_Waves-win'\n \n \n # for config files and kilosort working space\n kilosort_output_tmp = r'C:\\kilosort_datatemp' \n \n \n # derived directory names\n \n modules_directory = os.path.join(ecephys_directory,'modules')\n \n if kilosort_output_directory is None \\\n and extracted_data_directory is None \\\n and npx_directory is None:\n raise Exception('Must specify at least one output directory')\n\n\n #default ephys params. For spikeGLX, these get replaced by values read from metadata\n sample_rate = 30000\n num_channels = 385 \n reference_channels = [191]\n uVPerBit = 2.34375\n acq_system = 'PXI'\n \n \n if spikeGLX_data:\n # location of the raw data is the continuous file passed from script\n # metadata file should be located in same directory\n # \n # kilosort output will be put in the same directory as the input raw data,\n # set in kilosort_output_directory passed from script\n # kilososrt postprocessing (duplicate removal) and identification of noise\n # clusters will act on phy output in the kilosort output directory\n #\n # \n if input_meta_path is not None:\n probe_type, sample_rate, num_channels, reference_channels, \\\n uVPerBit, useGeom = SpikeGLX_utils.EphysParams(input_meta_path) \n \n print('SpikeGLX params read from meta')\n print('probe type: {:s}, sample_rate: {:.5f}, num_channels: {:d}, uVPerBit: {:.4f}'.format\\\n (probe_type, sample_rate, num_channels, uVPerBit))\n print('reference channels: ' + repr(reference_channels))\n \n #print('kilosort output directory: ' + kilosort_output_directory )\n\n \n else:\n print('using default values for probe params')\n \n\n \n\n # geometry params by probe type. expand the dictoionaries to add types\n # vertical probe pitch vs probe type\n vpitch = {'3A': 20, 'NP1': 20, 'NP21': 15, 'NP24': 15, 'NP1100': 6, 'NP1300':20} \n hpitch = {'3A': 32, 'NP1': 32, 'NP21': 32, 'NP24': 32, 'NP1100': 6, 'NP1300':48} \n nColumn = {'3A': 2, 'NP1': 2, 'NP21': 2, 'NP24': 2, 'NP1100': 8,'NP1300':2} \n \n \n # CatGT needs the inner and outer redii for local common average referencing\n # specified in sites\n\n catGT_loccar_min_sites = int(round(catGT_loccar_min_um/vpitch.get(probe_type)))\n catGT_loccar_max_sites = int(round(catGT_loccar_max_um/vpitch.get(probe_type)))\n # print('loccar min: ' + repr(catGT_loccar_min_sites))\n \n # whiteningRange is the number of sites used for whitening in KIlosort\n # preprocessing. Calculate the number of sites within the user-specified\n # whitening radius for this probe geometery\n # for a Np 1.0 probe, 163 um => 32 sites\n nrows = np.sqrt((np.square(ks_whiteningRadius_um) - np.square(hpitch.get(probe_type))))/vpitch.get(probe_type)\n ks_whiteningRange = int(round(2*nrows*nColumn.get(probe_type)))\n if ks_whiteningRange > 384:\n ks_whiteningRange = 384\n \n # nNeighbors is the number of sites kilosort includes in a template.\n # Calculate the number of sites within that radisu.\n maxNeighbors = 64 # 64 for standard build of KS\n nrows = np.sqrt((np.square(ks_templateRadius_um) - np.square(hpitch.get(probe_type))))/vpitch.get(probe_type)\n ks_nNeighbors = int(round(2*nrows*nColumn.get(probe_type)))\n if ks_nNeighbors > maxNeighbors:\n ks_nNeighbors = maxNeighbors \n print('ks_nNeighbors: ' + repr(ks_nNeighbors))\n \n c_waves_radius_sites = int(round(c_Waves_snr_um/vpitch.get(probe_type)))\n\n # Create string designating temporary output file for KS2 (gets inserted into KS2 config.m file)\n fproc = os.path.join(kilosort_output_tmp,'temp_wh.dat') # full path for temp whitened data file\n fproc_forward_slash = fproc.replace('\\\\','/')\n fproc_str = \"'\" + fproc_forward_slash + \"'\"\n \n # Deduce sort outptut tag from kilosort_output_directory\n \n \n dictionary = \\\n {\n\n \"directories\": {\n \"ecephys_directory\":ecephys_directory,\n \"npx_directory\": npx_directory,\n \"extracted_data_directory\": extracted_data_directory,\n \"kilosort_output_directory\": kilosort_output_directory,\n \"kilosort_output_tmp\": kilosort_output_tmp\n },\n\n \"common_files\": {\n \"settings_json\" : npx_directory,\n \"probe_json\" : os.path.join(extracted_data_directory,'probe_json.json')\n },\n\n \"waveform_metrics\" : {\n \"waveform_metrics_file\" : os.path.join(kilosort_output_directory, 'waveform_metrics.csv')\n },\n \n \"cluster_metrics\" : {\n \"cluster_metrics_file\" : os.path.join(kilosort_output_directory, 'metrics.csv')\n },\n\n \"ephys_params\": {\n \"probe_type\" : probe_type,\n \"sample_rate\" : sample_rate,\n \"lfp_sample_rate\" : 2500,\n \"bit_volts\" : uVPerBit,\n \"num_channels\" : num_channels,\n \"reference_channels\" : reference_channels,\n \"vertical_site_spacing\" : 10e-6,\n \"ap_band_file\" : continuous_file,\n \"lfp_band_file\" : continuous_file.replace('.ap.bin', '.lf.bin'),\n \"reorder_lfp_channels\" : False,\n \"cluster_group_file_name\" : 'cluster_group.tsv'\n }, \n\n \"extract_from_npx_params\" : {\n \"npx_directory\": npx_directory,\n \"settings_xml\": npx_directory,\n \"npx_extractor_executable\": r\"C:\\Users\\svc_neuropix\\Documents\\GitHub\\npxextractor\\Release\\NpxExtractor.exe\",\n \"npx_extractor_repo\": r\"C:\\Users\\svc_neuropix\\Documents\\GitHub\\npxextractor\"\n },\n \n \"depth_estimation_params\" : {\n \"hi_noise_thresh\" : 50.0,\n \"lo_noise_thresh\" : 3.0,\n \"save_figure\" : depth_est_fig,\n \"figure_location\" : os.path.join(extracted_data_directory, 'probe_depth.png'),\n \"smoothing_amount\" : 5,\n \"power_thresh\" : 2.5,\n \"diff_thresh\" : -0.06,\n \"freq_range\" : [0, 10],\n \"max_freq\" : 150,\n \"saline_range_um\" : [3700, 3800],\n \"n_passes\" : 10,\n \"air_gap_um\" : 1000,\n \"time_interval\" : 5,\n \"skip_s_per_pass\" : 10,\n \"start_time\" : 10\n }, \n\n \"median_subtraction_params\" : {\n \"median_subtraction_executable\": \"C:\\\\Users\\\\svc_neuropix\\\\Documents\\\\GitHub\\\\spikebandmediansubtraction\\\\Builds\\\\VisualStudio2013\\\\Release\\\\SpikeBandMedianSubtraction.exe\",\n \"median_subtraction_repo\": \"C:\\\\Users\\\\svc_neuropix\\\\Documents\\\\GitHub\\\\spikebandmediansubtraction\\\\\",\n },\n\n \"kilosort_helper_params\" : {\n\n \"matlab_home_directory\": kilosort_output_tmp,\n \"kilosort_repository\" : kilosort_repository,\n \"npy_matlab_repository\" : npy_matlab_repository,\n \"kilosort_version\" : 2,\n \"spikeGLX_data\" : True,\n \"ks_make_copy\": ks_make_copy,\n \"surface_channel_buffer\" : 15,\n\n \"kilosort2_params\" :\n {\n \"KSver\" : KS2ver,\n \"remDup\" : ks_remDup, #these are expressed as int rather than Bool for matlab compatability\n \"finalSplits\" : ks_finalSplits,\n \"labelGood\" : ks_labelGood,\n \"saveRez\" : ks_saveRez,\n \"copy_fproc\" : ks_copy_fproc,\n \"fproc\" : fproc_str,\n \"chanMap\" : \"'chanMap.mat'\",\n \"doFilter\" : ks_doFilter,\n \"fshigh\" : 150,\n \"minfr_goodchannels\" : ks_minfr_goodchannels,\n \"Th\" : ks_Th,\n \"lam\" : 10,\n \"AUCsplit\" : 0.9,\n \"minFR\" : 1/50.,\n \"momentum\" : '[20 400]',\n \"sigmaMask\" : 30,\n \"ThPre\" : 8,\n \"gain\" : uVPerBit,\n \"CSBseed\" : ks_CSBseed,\n \"LTseed\" : ks_LTseed,\n \"whiteningRange\" : ks_whiteningRange,\n \"nNeighbors\" : ks_nNeighbors,\n \"CAR\" : ks_CAR,\n \"nblocks\" : ks_nblocks\n }\n },\n \n \"pykilosort_helper_params\" : {\n \"preprocessing_function\" : 'kilosort2', \n \"copy_fproc\" : ks_copy_fproc,\n \"fproc\" : fproc_str,\n \"seed\" : ks_LTseed,\n \"ks2_mode\" : False,\n \"perform_drift_registration\" : True,\n \"car\" : ks_CAR,\n \"Th\" : ks_Th,\n \"ThPre\" : 8,\n \"lam\" : 10,\n \"AUCsplit\" : 0.9,\n \"minFR\" : 1/50.,\n \"momentum\" : '[20 400]',\n \"sig_datashift\" : 20,\n \"sigmaMask\" : 30,\n \"fshigh\" : 300,\n \"fslow\" : 10000,\n \"minfr_goodchannels\" : 0,\n \"whiteningRange\" : ks_whiteningRange, \n \"deterministic_mode\" : True, \n \"nblocks\" : ks_nblocks,\n \"doFilter\" : ks_doFilter\n\n },\n \n\n \"ks_postprocessing_params\" : {\n \"align_avg_waveform\" : False, \n \"remove_duplicates\" : True,\n \"cWaves_path\" : cWaves_path,\n \"within_unit_overlap_window\" : 0.00017,\n \"between_unit_overlap_window\" : 0.00041,\n \"between_unit_dist_um\" : 66,\n \"deletion_mode\" : 'lowAmpCluster',\n \"include_pcs\" : include_pcs\n },\n\n \"mean_waveform_params\" : { \n \"mean_waveforms_file\" : os.path.join(kilosort_output_directory, 'mean_waveforms.npy'),\n \"samples_per_spike\" : 82,\n \"pre_samples\" : 20,\n \"num_epochs\" : 1, #epochs not implemented for c_waves\n \"spikes_per_epoch\" : 1000,\n \"spread_threshold\" : wm_spread_thresh,\n \"site_range\" : wm_site_range, \n \"cWaves_path\" : cWaves_path,\n \"use_C_Waves\" : True,\n \"snr_radius\" : c_waves_radius_sites,\n \"snr_radius_um\" : c_Waves_snr_um\n },\n \n\n \"noise_waveform_params\" : {\n \"classifier_path\" : os.path.join(modules_directory, 'noise_templates', 'rf_classifier.pkl'),\n \"multiprocessing_worker_count\" : 10,\n \"use_random_forest\" : noise_template_use_rf\n },\n\n \"quality_metrics_params\" : {\n \"isi_threshold\" : qm_isi_thresh,\n \"min_isi\" : 0.000166,\n \"tbin_sec\" : 0.001,\n \"max_radius_um\" : 68,\n \"max_spikes_for_unit\" : 500,\n \"max_spikes_for_nn\" : 10000,\n \"n_neighbors\" : 4,\n 'n_silhouette' : 10000,\n \"drift_metrics_interval_s\" : 51,\n \"drift_metrics_min_spikes_per_interval\" : 10,\n \"include_pcs\" : include_pcs\n },\n \n \"catGT_helper_params\" : {\n \"run_name\" : catGT_run_name,\n \"gate_string\" : gate_string,\n \"probe_string\" : probe_string,\n \"trigger_string\": trigger_string,\n \"stream_string\" : catGT_stream_string,\n \"car_mode\" : catGT_car_mode,\n \"loccar_inner\" : catGT_loccar_min_sites,\n \"loccar_outer\": catGT_loccar_max_sites,\n \"loccar_inner_um\" : catGT_loccar_min_um,\n \"loccar_outer_um\" : catGT_loccar_max_um,\n \"maxZ_um\" : catGT_maxZ_um,\n 'useGeom' : useGeom,\n \"cmdStr\" : catGT_cmd_string,\n \"catGTPath\" : catGTPath\n },\n\n \"tPrime_helper_params\" : {\n \"tPrime_path\" : tPrime_path,\n \"im_ex_list\" : tPrime_im_ex_list,\n \"ni_ex_list\" : tPrime_ni_ex_list,\n \"sync_period\" : sync_period,\n \"toStream_sync_params\" : toStream_sync_params,\n \"ni_sync_params\" : niStream_sync_params,\n \"tPrime_3A\" : tPrime_3A,\n \"toStream_path_3A\" : toStream_path_3A,\n \"fromStream_list_3A\" : fromStream_list_3A,\n \"psth_ex_str\": event_ex_param_str,\n \"sort_out_tag\": ks_output_tag\n }, \n \n \"psth_events\": {\n \"event_ex_param_str\": event_ex_param_str\n }\n \n }\n\n with io.open(output_file, 'w', encoding='utf-8') as f:\n f.write(json.dumps(dictionary, ensure_ascii=False, sort_keys=True, indent=4))\n\n return dictionary","sub_path":"ecephys_spike_sorting/scripts/create_input_json.py","file_name":"create_input_json.py","file_ext":"py","file_size_in_byte":16060,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"123306271","text":"from torch.utils.data import DataLoader\nimport torch\nfrom torch.utils.data import Dataset\nfrom PIL import Image\nimport torchvision.transforms as transforms\n\n\ndef face_gan_man(img):\n ###### Definition of variables ######\n batchSize=1\n input_nc=3\n output_nc=3\n size= 224\n device= 'cuda' if torch.cuda.is_available() else 'cpu'\n n_cpu=8\n generator_B2A='/app/netG_B2A10.pth'\n \n # Load image\n img = Image.open(img)\n width, height = img.size\n ratio = width / height \n\n\n # Network\n netG_B2A = Generator(output_nc, input_nc)\n netG_B2A.to(device)\n\n # Load state dicts\n netG_B2A.load_state_dict(torch.load(generator_B2A,map_location=torch.device(device)))\n\n # Set model's test mode\n netG_B2A.eval()\n\n # Dataset loader\n transforms_ = [ transforms.Resize([size,size]),\n transforms.ToTensor(),\n transforms.Normalize((0.5,0.5,0.5), (0.5,0.5,0.5)) ]\n dataloader = DataLoader(ImageDataset(img, transforms_=transforms_,),\n batch_size=batchSize, shuffle=False, num_workers=n_cpu)\n\n for batch in dataloader:\n # Set model input\n real_B = batch.to(device)\n\n # Generate output\n fake_A = 0.5*(netG_B2A(real_B).data + 1.0)\n\n return fake_A, ratio\n\nclass ImageDataset(Dataset):\n def __init__(self, img, transforms_=None,):\n self.transform = transforms.Compose(transforms_)\n self.files_A = img\n\n def __getitem__(self, index):\n item_A = self.transform(self.files_A)\n return item_A\n\n def __len__(self):\n return 1\n\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nclass ResidualBlock(nn.Module):\n def __init__(self, in_features):\n super(ResidualBlock, self).__init__()\n\n conv_block = [ nn.ReflectionPad2d(1),\n nn.Conv2d(in_features, in_features, 3),\n nn.InstanceNorm2d(in_features),\n nn.ReLU(inplace=True),\n nn.ReflectionPad2d(1),\n nn.Conv2d(in_features, in_features, 3),\n nn.InstanceNorm2d(in_features) ]\n\n self.conv_block = nn.Sequential(*conv_block)\n\n def forward(self, x):\n return x + self.conv_block(x)\n\nclass Generator(nn.Module):\n def __init__(self, input_nc, output_nc, n_residual_blocks=9):\n super(Generator, self).__init__()\n\n # Initial convolution block\n model = [ nn.ReflectionPad2d(3),\n nn.Conv2d(input_nc, 64, 7),\n nn.InstanceNorm2d(64),\n nn.ReLU(inplace=True) ]\n\n # Downsampling\n in_features = 64\n out_features = in_features*2\n for _ in range(2):\n model += [ nn.Conv2d(in_features, out_features, 3, stride=2, padding=1),\n nn.InstanceNorm2d(out_features),\n nn.ReLU(inplace=True) ]\n in_features = out_features\n out_features = in_features*2\n\n # Residual blocks\n for _ in range(n_residual_blocks):\n model += [ResidualBlock(in_features)]\n\n # Upsampling\n out_features = in_features//2\n for _ in range(2):\n model += [ nn.ConvTranspose2d(in_features, out_features, 3, stride=2, padding=1, output_padding=1),\n nn.InstanceNorm2d(out_features),\n nn.ReLU(inplace=True) ]\n in_features = out_features\n out_features = in_features//2\n\n # Output layer\n model += [ nn.ReflectionPad2d(3),\n nn.Conv2d(64, output_nc, 7),\n nn.Tanh() ]\n\n self.model = nn.Sequential(*model)\n\n def forward(self, x):\n return self.model(x)\n","sub_path":"face_GAN_man.py","file_name":"face_GAN_man.py","file_ext":"py","file_size_in_byte":3763,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"116009665","text":"import numpy as np\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport copy\nimport csv\nimport os\nimport runge_kutta4\nimport mplcursors\n\n#Xのデータの入っているout.csvまでのパス\npath = '../out.csv'\n# プロットするデータのinput\nD = np.loadtxt(path, delimiter=',', dtype='float32')\n#ルンゲクッタの計算ステップ回数\nN = 20\n\n#errorは各Xとその時間発展事の誤差の大きさを集めるリスト\n#176=(900-20)/5\nerror1 = [[] for i in range(176)]#微小時間h=0.01のときのため\nerror2 = [[] for i in range(176)]#微小時間h=0.02のときのため\nerror3 = [[] for i in range(176)]#微小時間h=0.03のときのため\nerror4 = [[] for i in range(176)]#微小時間h=0.04のときのため\n\n#range(20,5,1000)だと何回目の計算か分かりにくいので別にcountを定義してカウントする\ncount = 0\n\n#ホフメラー図を確認したところ、20番目のXはすでにカオスに移行していたので20番目から5飛ばしで1000まで行う\nfor i in range(20, 900, 5):\n\n # 誤差発達率を導出するための誤差を複数の40次元分用意する\n R = [[] for k in range(100)]\n for j in range(100):\n for k in range(40):\n R[j].append(np.random.randn()/10000)#これで40次元の誤差が100通りできた\n\n #Xのある点の各要素に誤差を加える\n DD = [[] for k in range(100)]#DDの要素はリストで、さらにそのリストは同一の点の100通りの誤差が入る\n for j in range(40):\n for k in range(100):\n DD[j].append( D[i][j] + R[k][j])#ある一つの点に100通りの誤差を加える\n\n #上で誤差を加えたXがルンゲクッタの時間発展とともにどれだけ誤差を大きくしていくかをみる\n for k in range(100):\n DDD1[k] = runge_kutta4.Lorenz96_RK4(DD[k], 0.01, N, 8.0)#微小時間h=0.01のときのため\n DDD2[k] = runge_kutta4.Lorenz96_RK4(DD[k], 0.02, N, 8.0)#微小時間h=0.02のときのため\n DDD3[k] = runge_kutta4.Lorenz96_RK4(DD[k], 0.03, N, 8.0)#微小時間h=0.03のときのため\n DDD4[k] = runge_kutta4.Lorenz96_RK4(DD[k], 0.04, N, 8.0)#微小時間h=0.04のときのため\n\n #ルンゲクッタでの時間発展の各ステップ段階での誤差をerrorリストに代入する\n for j in range(N):\n #deltaDはあるXの各要素の誤差を集めたリスト\n deltaD1 = [[] for k in range(176)]#微小時間h=0.01のときのため\n deltaD2 = [[] for k in range(176)]#微小時間h=0.02のときのため\n deltaD3 = [[] for k in range(176)]#微小時間h=0.03のときのため\n deltaD4 = [[] for k in range(176)]#微小時間h=0.04のときのため\n\n for k in range(100):\n deltaD1[k].append( D[i + (j+1)] - DDD1[k][j] ) #D[i + (j+1)]は、out.csvにある40次元Xの各要素のタイムステップあたりの真の値で、DDD[j]は誤差を含めて開始した各微小時間hごとに計算した値\n deltaD2[k].append( D[i + 2*(j+1)-1] - DDD2[k][j] )#以下jを二倍三倍四倍しているのは、元データのout.csvのタイムステップが0.01であり、それに合わせて時間の刻み幅を考える必要があるから\n deltaD3[k].append( D[i + 3*(j+1)-2] - DDD3[k][j] )\n deltaD4[k].append( D[i + 4*(j+1)-3] - DDD4[k][j] )\n #誤差の大きさ(ノルム)\n for j in range(176):\n for k in range(100):\n error1[j][k].append(np.linalg.norm(deltaD1[k]))#微小時間h=0.01のときのため\n error2[j][k].append(np.linalg.norm(deltaD2[k]))#微小時間h=0.02のときのため\n error3[j][k].append(np.linalg.norm(deltaD3[k]))#微小時間h=0.03のときのため\n error4[j][k].append(np.linalg.norm(deltaD4[k]))#微小時間h=0.04のときのため\n \n count+=1\n\n#AERは平均誤差発達率\nAER1 = []\nAER2 = []\nAER3 = []\nAER4 = []\n#erは各タイムステップごとの、アトラクタ上のある一点の100通りの誤差の平均を、さらにアトラクタ上の他の147通りで平均を取ったもののリスト\ner1 = []#h=0.01のとき\ner2 = []#h=0.02のとき\ner3 = []#h=0.03のとき\ner4 = []#h=0.04のとき\n#各微小時間ごとの、さらに時間ステップごとの誤差をリストに付け加えていく\nfor i in range(N):\n for j in range(176):\n er1.append(np.mean(error1[j][k] for k in range(100)))\n er2.append(np.mean(error2[j][k] for k in range(100)))\n er3.append(np.mean(error3[j][k] for k in range(100)))\n er4.append(np.mean(error4[j][k] for k in range(100)))\n AER1.append(np.mean(er1))\n AER2.append(np.mean(er2))\n AER3.append(np.mean(er3))\n AER4.append(np.mean(er4))\n \"\"\"\n AER1.append(np.mean([error1[k][i] for k in range(176)]))#kで繰り返しを行うことは、アトラクター上から採ってきたサンプルごとに考えているということ\n AER2.append(np.mean([error2[k][i] for k in range(176)]))#そして、error[k][i]をすべてのkで足し合わせることは、サンプルから出てきたタイムステップiのときの誤差を足し合わせることとなる\n AER3.append(np.mean([error3[k][i] for k in range(176)]))\n AER4.append(np.mean([error4[k][i] for k in range(176)]))\n \"\"\"\n\n#各微小時間ごとの平均誤差発達率をプロットする\nfig = plt.figure( figsize=(11, 5) )\nax1 = fig.add_subplot(121)\nax1.plot(AER1,label=\"h=0.01\")\nax1.plot(AER2,label=\"h=0.02\")\nax1.plot(AER3,label=\"h=0.03\")\nax1.plot(AER4,label=\"h=0.04\")\nax1.set_xlabel(\"タイ���ステップ数\",fontname=\"MS Gothic\")\nax1.set_ylabel(\"平均誤差発達率\",fontname=\"MS Gothic\")\nax1.set_xticks( np.arange(0, N+1, 2))\nax1.legend()\nax1.set_title(\"微小時間hごとの平均誤差発達率\",fontname=\"MS Gothic\")\n\nax2 = fig.add_subplot(122)\nax2.plot(AER1,label=\"h=0.01\")\nax2.set_xlabel(\"タイムステップ数\",fontname=\"MS Gothic\")\nax2.set_ylabel(\"平均誤差発達率\",fontname=\"MS Gothic\")\nax2.set_xticks( np.arange(0, N+1, 2))\nax2.axhline(y=2, color='gray', ls='--')\nax2.legend()\nax2.set_title(\"微小時間hごとの平均誤差発達率\",fontname=\"MS Gothic\")\n#lines = ax2.plot(ax2,'s-')\n#mplcursors.cursor(lines)\n\nplt.show()\n\n#nperror1 = np.array(error1)\n#print(len(nperror1[:,0]))\n\nprint(D[-1])\nprint(len(D[1]))\n#print(len(DDD1[1]))\n","sub_path":"kadai2/kadai2.py","file_name":"kadai2.py","file_ext":"py","file_size_in_byte":6427,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"318771345","text":"input_arr = [3, 4, 1, 2, 16, 27, 13]\n\n\ndef quicksort(array):\n if len(array) < 2:\n return array\n else:\n pivot = array[0]\n less = [i for i in array[1:] if i <= pivot]\n greater = [i for i in array[1:] if i > pivot]\n return quicksort(less) + [pivot] + quicksort(greater)\n\n\ndef solve(source):\n odd_arr = [x for x in source if x % 2 == 1]\n even_arr = [x for x in source if x % 2 == 0]\n\n odd_arr = quicksort(odd_arr)\n even_arr = quicksort(even_arr)\n\n odd_arr.reverse()\n\n res = even_arr + odd_arr\n\n return res\n\n\nresult = solve(input_arr)\nprint(result)\n","sub_path":"solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":602,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"361376730","text":"#!/usr/bin/env python\nfrom __future__ import print_function\nfrom __future__ import division\nfrom __future__ import absolute_import\nfrom __future__ import unicode_literals\n\nimport argparse\nimport contextlib\nfrom os import ( listdir, mkdir )\nimport os.path as p\nimport shutil\nimport sys\nimport tempfile\nimport hashlib\n\nDIR_OF_THIS_SCRIPT = p.dirname( p.abspath( __file__ ) )\nDIR_OF_THIRD_PARTY = p.join( DIR_OF_THIS_SCRIPT, 'third_party' )\n\n\ndef GetStandardLibraryIndexInSysPath():\n for index, path in enumerate( sys.path ):\n if p.isfile( p.join( path, 'os.py' ) ):\n return index\n raise RuntimeError( 'Could not find standard library path in Python path.' )\n\n\ndef AddRequestDependencies():\n request_dep_root = p.abspath( p.join( DIR_OF_THIRD_PARTY,\n 'requests_deps' ) )\n for path in listdir( request_dep_root ):\n sys.path.insert( 0, p.join( request_dep_root, path ) )\n\n sys.path.insert( 0, p.abspath( p.join( DIR_OF_THIRD_PARTY,\n 'requests_deps',\n 'urllib3',\n 'src' ) ) )\n\n\nsys.path.insert( GetStandardLibraryIndexInSysPath() + 1,\n p.abspath( p.join( DIR_OF_THIRD_PARTY, 'python-future',\n 'src' ) ) )\nAddRequestDependencies()\n\n# Not installing aliases from python-future; it's unreliable and slow.\nfrom builtins import * # noqa\nfrom future.utils import iteritems\nimport requests\n\n\nURL_FORMAT = {\n 'release': ( \"https://github.com/OmniSharp/omnisharp-roslyn/\"\n \"releases/download/{version}/{file_name}\" ),\n 'ci': ( \"https://roslynomnisharp.blob.core.windows.net/\"\n \"releases/{version}/{file_name}\" ),\n}\nFILE_NAME = {\n 'win32': 'omnisharp.http-win-x86.zip',\n 'win64': 'omnisharp.http-win-x64.zip',\n 'macos': 'omnisharp.http-osx.tar.gz',\n 'linux32': 'omnisharp.http-linux-x86.tar.gz',\n 'linux64': 'omnisharp.http-linux-x64.tar.gz',\n}\n\n\n@contextlib.contextmanager\ndef TemporaryDirectory():\n temp_dir = tempfile.mkdtemp()\n try:\n yield temp_dir\n finally:\n shutil.rmtree( temp_dir )\n\n\ndef Download( url ):\n print( 'Downloading {}'.format( url.rsplit( '/', 1 )[ -1 ] ) )\n request = requests.get( url, stream=True )\n request.raise_for_status()\n content = request.content\n request.close()\n return content\n\n\ndef ParseArguments():\n parser = argparse.ArgumentParser()\n\n parser.add_argument( 'version', action='store',\n help = 'The Omnisharp version' )\n parser.add_argument( '--cache-dir', action='store',\n help = 'For testing, directory to cache packages.' )\n\n args = parser.parse_args()\n\n return args\n\n\ndef GetDownloadUrl( version, file_name ):\n download_url_key = 'ci' if \"-\" in version else 'release'\n\n return URL_FORMAT[ download_url_key ].format( version = version,\n file_name = file_name )\n\n\ndef FetchAndHash( download_url, output_dir, file_name ):\n try:\n archive = p.join( output_dir, file_name )\n if not p.exists( archive ):\n compressed_data = Download( download_url )\n with open( archive, 'wb' ) as f:\n f.write( compressed_data )\n except requests.exceptions.HTTPError as error:\n if error.response.status_code != 404:\n raise\n print( 'Cannot download {}'.format( file_name ) )\n return\n\n with open( archive, 'rb' ) as f:\n return hashlib.sha256( f.read() ).hexdigest()\n\n\ndef Process( output_dir, version ):\n result = {}\n\n for os_name, file_name in iteritems( FILE_NAME ):\n download_url = GetDownloadUrl( version, file_name )\n result[ os_name ] = {\n 'version': version,\n 'download_url': download_url,\n 'file_name': file_name,\n 'check_sum': FetchAndHash( download_url, output_dir, file_name )\n }\n\n return result\n\n\ndef MkDirIfMissing( dir ):\n try:\n mkdir( dir )\n except OSError:\n pass\n\n\ndef Main():\n args = ParseArguments()\n version = args.version\n\n if args.cache_dir:\n MkDirIfMissing( args.cache_dir )\n cache_dir = p.join( args.cache_dir, version )\n MkDirIfMissing( cache_dir )\n output = Process( cache_dir, version )\n else:\n with TemporaryDirectory() as temp_dir:\n output = Process( temp_dir, version )\n\n print( \"Omnisharp configration for {} is:\".format( version ) )\n for os_name, os_data in iteritems( output ):\n print( \" {}: {{\".format( repr( os_name ) ) )\n for key, value in iteritems( os_data ):\n line = \" {}: {},\".format( repr( key ), repr( value ) )\n if len( line ) > 80:\n line = \" {}: ( {} ),\".format( repr( key ), repr( value ) )\n format_index = line.index( '(' ) + 2\n while len( line ) > 80:\n print( line[ 0:78 ] + \"'\" )\n line = ( ' ' * format_index ) + \"'\" + line[ 78: ]\n print( line )\n print( \" },\" )\n\n\nif __name__ == \"__main__\":\n Main()\n","sub_path":"my_plugins/YouCompleteMe/third_party/ycmd/update_omnisharp.py","file_name":"update_omnisharp.py","file_ext":"py","file_size_in_byte":4923,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"326570007","text":"\"\"\"\nLauncher of telegram messages handler module\n\nCreated on 7/1/2015 by rdvlip.\n\"\"\"\n\n__author__ = 'rdvlip'\n\nimport os\nimport logging\nfrom concurrent.futures import ThreadPoolExecutor\n\nimport redis\nimport tornado.ioloop\n\nfrom hs_telegram.api import Api\nfrom hs_telegram import message_router\n\nlogging.basicConfig(level=logging.INFO, format=\"[%(levelname)s] %(asctime)s [%(name)s] %(message)s\")\nlogging.getLogger('urllib3').setLevel(logging.WARN)\n\nlogger = logging.getLogger(\"hs_telegram.main\")\n\n\nclass Main:\n def __init__(self):\n self.router = None\n self.api = None\n self.token = None\n self.pubsub = None\n\n self.thread_pool = ThreadPoolExecutor(1)\n self.ioloop = tornado.ioloop.IOLoop.instance()\n\n def _prepare_api(self):\n self.token = os.environ['TELEGRAM_TOKEN']\n self.api = Api(self.token)\n self.api.check_token()\n\n def _get_message(self):\n message = self.pubsub.get_message(ignore_subscribe_messages=True)\n if message:\n self.router.process_redis_message(message)\n self.ioloop.add_callback(self._get_message)\n\n def run(self):\n logger.info('Initializing...')\n self._run()\n logger.info('Initialization completed')\n self.ioloop.start()\n\n def _run(self):\n self._prepare_api()\n self.router = message_router.MessageRouter(self.api)\n\n r = redis.StrictRedis(host='redis', port=6379, db=0)\n p = self.pubsub = r.pubsub()\n\n for subscr_pattern in self.router.redis_mapping.keys():\n p.psubscribe(subscr_pattern)\n\n self._get_message()\n\n\n\n","sub_path":"hs_telegram/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"79921519","text":"# Copyright 2014 Google Inc. All Rights Reserved.\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\n\"\"\"Interface to OpenHTF configuration files.\n\nOpenHTF configuration files contain values which are specific to an individual\nstation. Any values which apply to all stations of a given type should be\nhandled by FLAGS or another mechanism.\n\nConfig keys must be declared as in the following example:\n\nconf.Declare('antimatter_intermix_constant',\n description='Intermix constant calibrated for our warp core.')\n\nDeclared keys can later be accessed by instantiating a Config object:\n\n...\nconfig = conf.Config()\nwarp_core.SetIntermixConstant(config.antimatter_intermix_constant)\n\"\"\"\n\nimport copy\nimport functools\nimport inspect\nimport logging\nimport threading\nimport yaml\n\nimport gflags\nimport mutablerecords\n\nfrom openhtf.util import threads\n\n\nFLAGS = gflags.FLAGS\n\ngflags.DEFINE_string('config',\n '/usr/local/openhtf_client/config/clientfoo.yaml',\n 'The OpenHTF configuration file for this tester')\n\ngflags.DEFINE_multistring(\n 'config_value', [], 'Allows specifying a configuration key=value '\n 'on the command line. The format should be --config_value key=value. '\n 'This value will override any existing config value at config load time '\n 'and will be a string')\n\nConfigurationDeclaration = ( # pylint: disable=invalid-name\n mutablerecords.Record(\n 'ConfigurationDeclaration',\n ['name'],\n {'description': None, 'default_value': None, 'optional': True}))\n\n_LOG = logging.getLogger(__name__)\n\nclass ConfigurationNotLoadedError(Exception):\n \"\"\"Raised if a configuration variable is accessed before it is loaded.\n\n This helps protect against a class of errors where people try to access the\n configuration at import time when it hasn't been loaded by the main function\n yet.\n \"\"\"\n\n\nclass ConfigurationMissingError(Exception):\n \"\"\"Indicates the configuration file could not be read.\"\"\"\n\n\nclass ConfigurationInvalidError(Exception):\n \"\"\"Indicates the configuration format was invalid.\"\"\"\n\n\nclass ConfigurationAlreadyDeclared(Exception):\n \"\"\"Indicates that a configuration key was already declared.\"\"\"\n\n\nclass MissingRequiredConfigurationKeyError(Exception):\n \"\"\"Indicates a required configuration key is missing.\"\"\"\n\n\nclass UndeclaredKeyAccessError(Exception):\n \"\"\"Indicates that a key was required but not predeclared.\"\"\"\n\n\nclass ConfigurationValidationError(Exception):\n \"\"\"If a configuration value could not be validated as its expected type.\"\"\"\n\n def __init__(self, name, declaration, value):\n super(ConfigurationValidationError, self).__init__(\n name, declaration, value)\n self.name = name\n self.declaration = declaration\n self.value = value\n\n def __str__(self):\n return ('<%s: (Configuration error on key: %s (type: %s, value: %s))>' %\n (type(self).__name__, self.name, self.declaration.type.name, self.value))\n\n\nclass _DeclaredKeys(object):\n \"\"\"An object which manages config declarations.\n\n This object is a helper for Config. It provides locked access to a map of\n declarations and processes configuration values against the declaration. It\n must be guarded since declarations are updated at import time and if something\n is lazily imported we could race between updating the declarations map and\n reading it to check a config value.\n\n Not thread-safe, requires an external lock!\n \"\"\"\n\n def __init__(self):\n self._declared = {}\n\n def Declare(self, name, declaration):\n \"\"\"Adds a declared key to this list of Declared Keys.\n\n Args:\n name: The name of this value.\n declaration: A _DeclaredKeys.DECLARATION object.\n\n Raises:\n ConfigurationAlreadyDeclared: If a declaration already exists for\n this key.\n \"\"\"\n if name in self._declared:\n raise ConfigurationAlreadyDeclared(name)\n self._declared[name] = declaration\n\n def CheckValueAgainstDeclaration(self, name, value):\n \"\"\"Checks that the provided value is valid given its declaration.\n\n Args:\n name: Name of configuration key.\n value: Value from configuration.\n\n Returns:\n The config value if provided, or the default_value if given.\n\n Raises:\n UndeclaredKeyAccessError: If key 'name' is undeclared.\n MissingRequiredConfigurationKeyError: If key is required and value\n is None\n \"\"\"\n declaration = self._declared.get(name, None)\n\n if not declaration:\n raise UndeclaredKeyAccessError(name)\n\n if (value is None\n and declaration.default_value is None\n and not declaration.optional):\n raise MissingRequiredConfigurationKeyError(\n declaration.name, declaration.description)\n\n if value is None:\n return declaration.default_value\n return value\n\n def __contains__(self, name): # pylint: disable=invalid-name\n return name in self._declared\n\n def __getitem__(self, name): # pylint: disable=invalid-name\n return self._declared[name]\n\n def __copy__(self): # pylint: disable=invalid-name\n self_copy = type(self)()\n for name, declaration in self._declared.iteritems():\n self_copy.Declare(name, declaration)\n return self_copy\n\n\nclass ConfigModel(object):\n \"\"\"A model that holds the underlying config keys and their values.\n\n By isolating the underlying model it provides a way to lock access to the\n dictionary so we can reload it on demand or otherwise poke it.\n \"\"\"\n\n def __init__(self, state=None, declarations=None):\n \"\"\"Initializes the model.\n\n Args:\n state: A dictionary containing configuration key, values. By default a\n new one is created. If one is provided the model is marked as\n loaded.\n declarations: An object which tracks declared keys, if not provided\n a new one is constructed.\n \"\"\"\n self._state = state if state is not None else {}\n self._declarations = declarations or _DeclaredKeys()\n self._loaded = state is not None\n self.lock = threading.Lock()\n\n # pylint: disable=missing-docstring\n @property\n @threads.Synchronized\n def loaded(self):\n return self._loaded\n\n @property\n @threads.Synchronized\n def state(self):\n return self._state.copy()\n\n @property\n @threads.Synchronized\n def declarations(self):\n return copy.copy(self._declarations)\n\n @threads.Synchronized\n def Items(self):\n return self._state.items()\n\n @threads.Synchronized\n def GetValue(self, name, default=None):\n value = self._state.get(name, default)\n return self._declarations.CheckValueAgainstDeclaration(name, value)\n\n @threads.Synchronized\n def ContainsKey(self, name):\n return name in self._state\n\n # pylint: enable=missing-docstring\n\n @threads.Synchronized\n def Load(self, config_file=None, force_reload=False,\n config_loader=lambda fname: open(fname, 'r')):\n \"\"\"Loads the configuration file from disk.\n\n Args:\n config_file: The file name to load configuration from.\n Defaults to FLAGS.config.\n force_reload: If true this method will ignore the loaded state and reload\n the config from disk.\n config_loader: A callable which returns a file object when given a\n filename, defaults to open.\n\n Returns:\n True if configuration was loaded, False if already loaded.\n Raises:\n ConfigurationMissingError: If configuration file could not be read\n ConfigurationInvalidError: If configuration file is not valid yaml\n \"\"\"\n if not force_reload and self._loaded:\n return False\n\n try:\n filename = config_file or FLAGS.config\n _LOG.info('Loading from config: %s', filename)\n\n with config_loader(filename) as config_file:\n data = yaml.safe_load(config_file)\n if not data:\n raise ConfigurationInvalidError('No data', config_file)\n self._state.clear()\n self._state.update(data)\n\n # Load string values from flags\n for keyval in FLAGS.config_value:\n key, val = keyval.split('=')\n self._state[key] = val\n\n self._loaded = True\n _LOG.debug('Configuration loaded: %s', self._state)\n except yaml.YAMLError as exception:\n _LOG.exception('Failed to load yaml file: %s', filename)\n raise ConfigurationInvalidError(filename, exception)\n except IOError as exception:\n _LOG.exception('Configuration failed loaded: %s', filename)\n raise ConfigurationMissingError(filename, exception)\n\n return True\n\n @threads.Synchronized\n def LoadFromDict(self, dictionary, force_reload=False):\n \"\"\"Loads the config with values from a dictionary instead of a file.\n\n This is meant for testing and bin purposes and shouldn't be used in most\n applications.\n\n Args:\n dictionary: The dictionary to update.\n force_reload: True to force a load if the config is already loaded.\n Returns:\n True if successful.\n \"\"\"\n if not force_reload and self._loaded:\n return False\n\n self._state.clear()\n self._state.update(dictionary)\n self._loaded = True\n return True\n\n @threads.Synchronized\n def LoadMissingFromDict(self, config_dict):\n \"\"\"Update any missing configurations from the given dictionary.\n\n This is similar to dict.update, except that instead of the given\n dictionary's values overriding the already set values, this function doesn't\n override. This is due to the fact that these configs can only be retrieved\n after we've already loaded the authoritative values.\n\n Args:\n config_dict: Dictionary from which to load configuration keys and values.\n\n Raises:\n ConfigurationNotLoadedError: Raised when updating empty config values.\n \"\"\"\n # Can't update only missing when it's all missing.\n if not self._loaded:\n raise ConfigurationNotLoadedError(\n 'Load configuration before updating missing keys.')\n\n for key, value in config_dict.items():\n if key in self._state:\n continue\n self._state[key] = value\n\n @threads.Synchronized\n def Reset(self):\n \"\"\"Resets the configuration, removing any state, useful for testing.\n\n Careful calling this, the reason we get away with not locking the dict is\n because we never call this in practice. If that changes then we need to\n guard it with a lock.\n \"\"\"\n self._state.clear()\n self._loaded = False\n\n @threads.Synchronized\n def Declare(self, name, description=None, **kwargs):\n \"\"\"Declares the use of a configuration variable.\n\n Currently all configuration variables must be declared. If a key is\n accessed in the config without being declared then chaos will ensue. If a\n file wants to access a key another module has declared they are\n encouraged to use extern.\n\n Args:\n name: The name of the key.\n description: Docstring for the key, if any.\n **kwargs: See ConfigurationDeclaration's fields.\n \"\"\"\n declaration = ConfigurationDeclaration(\n name, description=description, **kwargs)\n self._declarations.Declare(name, declaration)\n\n\nclass Config(object):\n \"\"\"The configuration read from a config file, or populated directly.\n\n This classes uses the borg design pattern so all instances share the same\n state. This is fine since the load only occurs on the main thread and from\n then on out the class is effectively read only.\n\n Example Usage:\n configuration.Load() # called once early\n\n # Can be done anyone and in multiple places without worrying about loading\n config = Config()\n if config.url:\n print config.url\n \"\"\"\n model = ConfigModel()\n\n def __init__(self, model=None):\n \"\"\"Initializes the configuration object with its shared state.\n\n Args:\n model: The data model to use, defaults to the one shared amonst all config\n objects.\n \"\"\"\n self.model = model or Config.model\n\n # pylint: disable=missing-docstring\n @property\n def dictionary(self):\n if not self.loaded:\n raise ConfigurationNotLoadedError()\n return self.model.state\n\n @property\n def loaded(self):\n return self.model.loaded\n\n # pylint: enable=missing-docstring\n\n def __getattr__(self, name): # pylint: disable=invalid-name\n \"\"\"Searches for the value in our config, returning if its found.\n\n Args:\n name: name of attribute\n Returns:\n None if key not found and is not required, otherwise the value.\n Raises:\n MissingRequiredKeyError: If the key was declared required and\n is not found.\n UndeclaredKeyAccessError: If the key being accessed was not\n declared.\n ConfigurationNotLoadedError: If the config file has not been loaded, this\n typically means you accessed the config at import time.\n \"\"\"\n if not self.model.loaded:\n raise ConfigurationNotLoadedError(name)\n return self.model.GetValue(name)\n\n def __contains__(self, name): # pylint: disable=invalid-name\n \"\"\"Provides the ability to quickly check if a config key is declared.\"\"\"\n return self.model.ContainsKey(name)\n\n def __getitem__(self, key): # pylint: disable=invalid-name\n \"\"\"Allows access to config items via an indexer.\"\"\"\n return self.__getattr__(key)\n\n def __repr__(self):\n return '' % (self.model.loaded, id(self))\n\n def CreateStackedConfig(self, model):\n \"\"\"Stacks a new model onto the current model, creating a new config.\n\n Args:\n model: A ConfigModel instance or a dict of values that can be converted\n into a ConfigModel instance. If a dict, the declarations of this\n object will be used.\n Returns:\n A new StackedConfig instance with model superseding the current model.\n \"\"\"\n if not isinstance(model, ConfigModel):\n model = ConfigModel(state=model, declarations=self.model.declarations)\n return StackedConfig([self.model, model])\n\n\nclass StackedConfig(Config):\n \"\"\"Stacked version of Config.\n\n This is a layered (or stacked) Config that allows users to make one set of\n config values supersede another set.\n \"\"\"\n\n # pylint: disable=super-init-not-called\n def __init__(self, models=(Config.model,)):\n self._models = list(models)\n\n def CreateStackedConfig(self, model):\n \"\"\"Stacks a new model onto the current models, creating a new config.\n\n Args:\n model: A ConfigModel instance or a dict of values that can be converted\n into a ConfigModel instance. If a dict, the declarations of the top of\n the stack will be used.\n Returns:\n A new StackedConfig instance with model superseding the current models.\n \"\"\"\n if not isinstance(model, ConfigModel):\n model = ConfigModel(\n state=model, declarations=self._models[0].declarations)\n return StackedConfig(self._models + [model])\n\n @property\n def dictionary(self):\n if not self.loaded:\n raise ConfigurationNotLoadedError()\n results = {}\n for model in self._models:\n results.update(model.state)\n return results\n\n @property\n def loaded(self):\n return any(model.loaded for model in self._models)\n\n def __getattr__(self, name):\n if not self.loaded:\n raise ConfigurationNotLoadedError(name)\n for model in self._models:\n if model.ContainsKey(name):\n return model.GetValue(name)\n return self._models[-1].GetValue(name)\n\n def __contains__(self, name):\n return any(model.ContainsKey(name) for model in self._models)\n\n def __str__(self):\n return '<%s: (loaded: %s: 0x%x)>' % (type(self).__name__, self.loaded, id(self))\n __repr__ = __str__\n\n\nclass ConfigValue(object): # pylint: disable=too-few-public-methods\n \"\"\"A thin wrapper which may be used to pass around a config value.\n\n This is useful when things require a value at import time yet config values\n are not available until runtime. By wrapping the key you want in this object,\n other objects which are aware of it can call it to retrieve the value at a\n later time (i.e. runtime). This is not a magic bullet, whatever you'ready\n calling must be ready for a ConfigValue or similar to provided.\n\n The value_fn parameter allows a function to be specified at import time which\n will be performed on the retrieved config value at runtime. This is useful for\n retrieving an inner-value of a config value, such as indexing into an\n array/dict config value.\n \"\"\"\n\n def __init__(self, config_key, config=None, value_fn=None):\n self.config = config or Config()\n self.config_key = config_key\n self.value_fn = value_fn\n\n @property\n def value(self):\n \"\"\"Resolves the value returning the config value.\"\"\"\n if self.value_fn is None:\n return self.config[self.config_key]\n else:\n return self.value_fn(self.config[self.config_key])\n\n def __call__(self): # pylint: disable=invalid-name\n \"\"\"Returns the config value.\"\"\"\n return self.value\n\n def __str__(self):\n return '<%s: (ConfigKey: %s)' % (type(self).__name__, self.config_key)\n __repr__ = __str__\n\n\ndef Extern(dummy_name): # pylint: disable=invalid-name\n \"\"\"Declares that a module uses a key declared elsewhere.\n\n This function does nothing but serve as a marker at the top of your file that\n you're using a config key which improves readability greatly. You're\n encouraged to use this. That said since declaration of keys isn't checked\n until a key is used and since this function does nothing everything will still\n work without it.\n\n Args:\n unused_name: The name of the key.\n \"\"\"\n\n\ndef InjectPositionalArgs(method): # pylint: disable=invalid-name\n \"\"\"Decorator for injecting positional arguments from the configuration.\n\n This decorator wraps the given method, so that any positional arguments are\n passed with corresponding values from the configuration. The name of the\n positional argument must match the configuration key. Keyword arguments are\n not modified, but should not be named such that they match configuration keys\n anyway (this will result in a warning message).\n\n Additional positional arguments may be used that do not appear in the\n configuration, but those arguments *must* be specified as keyword arguments\n upon invokation of the method. This is to avoid ambiguity in which\n positional arguments are getting which values.\n\n Args:\n method: The method to wrap.\n\n Returns:\n A wrapper that, when invoked, will call the wrapped method, passing in\n configuration values for positional arguments.\n \"\"\"\n argspec = inspect.getargspec(method)\n\n # Index in argspec.args of the first keyword argument. This index is a\n # negative number if there are any kwargs, or 0 if there are no kwargs.\n keyword_arg_index = -1 * len(argspec.defaults or [])\n arg_names = argspec.args[:keyword_arg_index or None]\n kwarg_names = argspec.args[len(arg_names):]\n\n # Create the actual method wrapper, all we do is update kwargs. Note we don't\n # pass any *args through because there can't be any - we've filled them all in\n # with values from the configuration. Any positional args that are missing\n # from the configuration *must* be explicitly specified as kwargs.\n @functools.wraps(method)\n def method_wrapper(**kwargs):\n \"\"\"Wrapper that pulls values from the Config().\"\"\"\n config = Config()\n\n # Check for keyword args with names that are in the config so we can warn.\n for bad_name in set(kwarg_names) & set(config.dictionary.keys()):\n _LOG.warning('Keyword arg %s not set from configuration, but is a '\n 'configuration key', bad_name)\n\n # Set positional args from configuration values.\n config_args = {name: config[name] for name in arg_names if name in config}\n\n for overridden in set(kwargs) & set(config_args):\n _LOG.warning('Overriding provided kwarg %s=%s with value %s from '\n 'configuration', overridden, kwargs[overridden],\n config_args[overridden])\n kwargs.update(config_args)\n _LOG.info('Invoking %s with %s', method.__name__, kwargs)\n return method(**kwargs)\n\n # We have to check for a 'self' parameter explicitly because Python doesn't\n # pass it as a keyword arg, it passes it as the first positional arg.\n if 'self' == argspec.args[0]:\n @functools.wraps(method)\n def SelfWrapper(self, **kwargs): # pylint: disable=invalid-name,missing-docstring\n kwargs['self'] = self\n return method_wrapper(**kwargs)\n return SelfWrapper\n return method_wrapper\n\n\n# pylint: disable=invalid-name\nDeclare = Config().model.Declare\nLoad = Config().model.Load\nLoadMissingFromDict = Config().model.LoadMissingFromDict\nLoadFromDict = Config().model.LoadFromDict\nReset = Config().model.Reset\n\n# Everywhere that uses configuration uses this, so we just declare it here.\nDeclare('station_id', 'The name of this tester')\n","sub_path":"openhtf/conf.py","file_name":"conf.py","file_ext":"py","file_size_in_byte":21116,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"602072057","text":"# coding=utf-8\nimport time\nimport abc\nfrom collections import OrderedDict\n\nfrom ruamel.yaml import dump as ydump, load as yload, RoundTripDumper, resolver, add_constructor, add_representer\n\nfrom src.meta.abstract import AbstractMeta\nfrom utils.custom_logging import make_logger\nfrom utils.custom_path import Path\n\nlogger = make_logger(__name__)\n\n\n_yaml_mapping = resolver.BaseResolver.DEFAULT_MAPPING_TAG\n\n\ndef odict_represent(dumper, data):\n return dumper.represent_dict(data.iteritems())\n\n\ndef odict_construct(loader, node):\n return OrderedDict(loader.construct_pairs(node))\n\n\nadd_representer(OrderedDict, odict_represent)\nadd_constructor(_yaml_mapping, odict_construct)\n\n\nclass Meta(AbstractMeta):\n\n @property\n @abc.abstractmethod\n def meta_header(self):\n \"\"\"\"\"\"\n\n @property\n @abc.abstractmethod\n def meta_version(self):\n \"\"\"\"\"\"\n\n @abc.abstractmethod\n def meta_version_upgrade(self, from_version):\n \"\"\"\"\"\"\n\n def __init__(self, path: str or Path, init_dict: OrderedDict = None, auto_read=True, encrypted=False):\n self.free = True\n self.encrypt = encrypted\n\n if init_dict is None:\n self._data = OrderedDict()\n\n else:\n\n if not isinstance(init_dict, OrderedDict):\n raise TypeError('expected a OrderedDict, got \"{}\"'.format(type(init_dict)))\n\n self._data = init_dict\n\n self._values, self._keys, self._items = None, None, None\n self._init_views()\n\n if isinstance(path, Path):\n pass\n\n elif isinstance(path, str):\n path = Path(path)\n\n else:\n raise TypeError('expected a Path or a str, got: {}'.format(type(path)))\n\n self._path = path\n\n if auto_read:\n self.read()\n\n @property\n def path(self) -> Path:\n return self._path\n\n @path.setter\n def path(self, value: str or Path):\n\n if isinstance(value, Path):\n pass\n\n elif isinstance(value, str):\n value = Path(value)\n\n else:\n raise TypeError('expected Path or str, got: {}'.format(type(value)))\n\n self._path = value\n\n # noinspection PyArgumentList\n def _init_views(self):\n self._values = self._data.values()\n self._keys = self._data.keys()\n self._items = self._data.items()\n\n @property\n def data(self):\n return self._data\n\n def get_context(self):\n return self.data\n\n @data.setter\n def data(self, value: OrderedDict):\n\n if not isinstance(value, OrderedDict):\n raise TypeError('expected a OrderedDict, got \"{}\"'.format(type(value)))\n\n self._data = value\n self._init_views()\n\n def __len__(self):\n # noinspection PyTypeChecker\n return len(self.data)\n\n def __iter__(self):\n for k in self.keys():\n yield k\n\n def __contains__(self, x):\n # noinspection PyArgumentList\n return self._data.__contains__(x)\n\n def __delitem__(self, key, _write=False):\n del self.data[key]\n\n if _write:\n self.write()\n\n def __setitem__(self, key, value, _write=False):\n self.data[key] = value\n\n if _write:\n self.write()\n\n def __getitem__(self, key):\n return self._data.get(key, None)\n\n def __str__(self):\n # noinspection PyArgumentList\n return self.data.__str__()\n\n def __repr__(self):\n return '{}: {}'.format(self.__class__.__name__, self.data.__repr__())\n\n def get(self, key, default=None):\n return self._data.get(key, default)\n\n def keys(self):\n return self._keys\n\n def values(self):\n return self._values\n\n def items(self):\n return self._items\n\n def debug(self, txt: str):\n logger.debug('{}: {}'.format(self.path.abspath(), txt))\n\n def exception(self, txt: str):\n logger.debug('{}: {}'.format(self.path.abspath(), txt))\n\n def dump(self):\n return ydump(self.data, Dumper=RoundTripDumper, default_flow_style=False)\n\n def load(self, data):\n self.data = yload(data)\n\n def read(self):\n\n self.wait_for_lock()\n\n meta_updated = False\n\n try:\n\n if self.path.exists():\n\n if self.path.getsize() == 0:\n self.debug('{}: removing existing empty file: {}'.format(self.__class__.__name__, self.path))\n self.path.remove()\n\n return\n\n try:\n\n if self.encrypt:\n self.load(self.path.bytes())\n\n else:\n self.load(self.path.text(encoding='utf8'))\n\n except ValueError:\n raise ValueError('{}: metadata file corrupted'.format(self.path.abspath()))\n\n else:\n try:\n if not self.data['meta_header'] == self.meta_header:\n raise TypeError('meta header mismatch, expected: \"{}\", got: \"{}\" on file: {}'.format(\n self.meta_header, self.data['meta_header'], self.path.abspath()\n ))\n else:\n del self.data['meta_header']\n\n except KeyError:\n pass\n\n meta_updated = self.data['meta_version'] < self.meta_version\n\n while self.data['meta_version'] < self.meta_version:\n current_version = self.data['meta_version']\n next_version = self.data['meta_version'] + 1\n logger.debug('upgrading meta from version \"{}\"'.format(current_version))\n\n if not self.meta_version_upgrade(current_version):\n raise RuntimeError('failed to upgrade metadata to version \"{}\"'.format(next_version))\n\n else:\n logger.debug('successfully upgraded meta to version \"{}\"'.format(next_version))\n\n self.data['meta_version'] = next_version\n\n except OSError:\n self.exception('error while reading metadata file')\n\n finally:\n self.free = True\n\n if meta_updated:\n self.write()\n\n def write(self):\n # noinspection PyTypeChecker\n if len(self._data) == 0:\n raise ValueError('no data to write')\n\n self.wait_for_lock()\n self.data['meta_header'] = self.meta_header\n self.data['meta_version'] = self.meta_version\n\n try:\n\n if self.encrypt:\n self.path.write_bytes(self.dump())\n\n else:\n self.path.write_text(self.dump(), encoding='utf8')\n\n except OSError:\n self.exception('error while writing metadata to file')\n\n finally:\n self.free = True\n\n def wait_for_lock(self):\n i = 0\n\n while not self.free:\n time.sleep(0.1)\n i += 1\n\n if i == 10:\n self.debug('waiting for resource lock')\n i = 0\n\n self.free = False\n\n @staticmethod\n def read_header(path):\n\n path = Path(path)\n data = yload(path.text(encoding='utf8'))\n\n return data['header']\n\n\ndef read_meta_header(meta_file_path: Path or str):\n return Meta.read_header(meta_file_path)\n","sub_path":"src/meta/meta.py","file_name":"meta.py","file_ext":"py","file_size_in_byte":7384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"390502970","text":"# This file is part of PyOP2\n#\n# PyOP2 is Copyright (c) 2012, Imperial College London and\n# others. Please see the AUTHORS file in the main source directory for\n# a full list of copyright holders. All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions\n# are met:\n#\n# * Redistributions of source code must retain the above copyright\n# notice, this list of conditions and the following disclaimer.\n# * Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n# * The name of Imperial College London or that of other\n# contributors may not be used to endorse or promote products\n# derived from this software without specific prior written\n# permission.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTERS\n# ''AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT\n# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS\n# FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE\n# COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,\n# INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES\n# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR\n# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)\n# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,\n# STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)\n# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED\n# OF THE POSSIBILITY OF SUCH DAMAGE.\n\n\"\"\"Transform the kernel's AST according to the backend we are running over.\"\"\"\n\nfrom ast_base import *\nfrom ast_optimizer import LoopOptimiser\nfrom ast_vectorizer import init_vectorizer, LoopVectoriser, vectorizer_init\n\n# Possibile optimizations\nAUTOVECT = 1 # Auto-vectorization\nV_OP_PADONLY = 2 # Outer-product vectorization + extra operations\nV_OP_PEEL = 3 # Outer-product vectorization + peeling\nV_OP_UAJ = 4 # Outer-product vectorization + unroll-and-jam\nV_OP_UAJ_EXTRA = 5 # Outer-product vectorization + unroll-and-jam + extra iters\n\n# Track the scope of a variable in the kernel\nLOCAL_VAR = 0 # Variable declared and used within the kernel\nPARAM_VAR = 1 # Variable is a kernel parameter (ie declared in the signature)\n\n\nclass ASTKernel(object):\n\n \"\"\"Manipulate the kernel's Abstract Syntax Tree.\n\n The single functionality present at the moment is provided by the plan_gpu\n method, which transforms the AST for GPU execution.\n \"\"\"\n\n def __init__(self, ast):\n self.ast = ast\n self.decls, self.fors = self._visit_ast(ast, fors=[], decls={})\n\n def _visit_ast(self, node, parent=None, fors=None, decls=None):\n \"\"\"Return lists of:\n - Declarations within the kernel\n - Loop nests\n - Dense Linear Algebra Blocks\n that will be exploited at plan creation time.\"\"\"\n\n if isinstance(node, Decl):\n decls[node.sym.symbol] = (node, LOCAL_VAR)\n return (decls, fors)\n elif isinstance(node, For):\n fors.append((node, parent))\n return (decls, fors)\n elif isinstance(node, FunDecl):\n self.fundecl = node\n for d in node.args:\n decls[d.sym.symbol] = (d, PARAM_VAR)\n elif isinstance(node, (FlatBlock, PreprocessNode, Symbol)):\n return (decls, fors)\n\n for c in node.children:\n self._visit_ast(c, node, fors, decls)\n\n return (decls, fors)\n\n def plan_gpu(self):\n \"\"\"Transform the kernel suitably for GPU execution.\n\n Loops decorated with a \"pragma pyop2 itspace\" are hoisted out of\n the kernel. The list of arguments in the function signature is\n enriched by adding iteration variables of hoisted loops. Size of\n kernel's non-constant tensors modified in hoisted loops are modified\n accordingly.\n\n For example, consider the following function:\n\n void foo (int A[3]) {\n int B[3] = {...};\n #pragma pyop2 itspace\n for (int i = 0; i < 3; i++)\n A[i] = B[i];\n }\n\n plan_gpu modifies its AST such that the resulting output code is\n\n void foo(int A[1], int i) {\n A[0] = B[i];\n }\n \"\"\"\n\n lo = [LoopOptimiser(l, pre_l, self.decls) for l, pre_l in self.fors]\n for nest in lo:\n itspace_vrs, accessed_vrs = nest.extract_itspace()\n\n for v in accessed_vrs:\n # Change declaration of non-constant iteration space-dependent\n # parameters by shrinking the size of the iteration space\n # dimension to 1\n decl = set(\n [d for d in self.fundecl.args if d.sym.symbol == v.symbol])\n dsym = decl.pop().sym if len(decl) > 0 else None\n if dsym and dsym.rank:\n dsym.rank = tuple([1 if i in itspace_vrs else j\n for i, j in zip(v.rank, dsym.rank)])\n\n # Remove indices of all iteration space-dependent and\n # kernel-dependent variables that are accessed in an itspace\n v.rank = tuple([0 if i in itspace_vrs and dsym else i\n for i in v.rank])\n\n # Add iteration space arguments\n self.fundecl.args.extend([Decl(\"int\", c_sym(\"%s\" % i))\n for i in itspace_vrs])\n\n # Clean up the kernel removing variable qualifiers like 'static'\n for decl in self.decls.values():\n d, place = decl\n d.qual = [q for q in d.qual if q not in ['static', 'const']]\n\n if hasattr(self, 'fundecl'):\n self.fundecl.pred = [q for q in self.fundecl.pred\n if q not in ['static', 'inline']]\n\n def plan_cpu(self, opts):\n \"\"\"Transform and optimize the kernel suitably for CPU execution.\"\"\"\n\n # Fetch user-provided options/hints on how to transform the kernel\n licm = opts.get('licm')\n tile = opts.get('tile')\n vect = opts.get('vect')\n ap = opts.get('ap')\n\n v_type, v_param = vect if vect else (None, None)\n tile_opt, tile_sz = tile if tile else (False, -1)\n\n lo = [LoopOptimiser(l, pre_l, self.decls) for l, pre_l in self.fors]\n for nest in lo:\n # 1) Loop-invariant code motion\n inv_outer_loops = []\n if licm:\n inv_outer_loops = nest.op_licm() # noqa\n self.decls.update(nest.decls)\n\n # 2) Register tiling\n if tile_opt and v_type == AUTOVECT:\n nest.op_tiling(tile_sz)\n\n # 3) Vectorization\n if vectorizer_init:\n vect = LoopVectoriser(nest)\n if ap:\n vect.align_and_pad(self.decls)\n if v_type != AUTOVECT:\n vect.outer_product(v_type, v_param)\n\n\ndef init_ir(isa, compiler):\n \"\"\"Initialize the Intermediate Representation engine.\"\"\"\n\n init_vectorizer(isa, compiler)\n","sub_path":"pyop2/ir/ast_plan.py","file_name":"ast_plan.py","file_ext":"py","file_size_in_byte":7307,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"652587575","text":"#!/usr/bin/env python\nimport logging\n\nlogging.basicConfig(\n level=logging.INFO, format=\"%(asctime)s - %(name)s - %(levelname)s - %(message)s\"\n)\nimport os\nimport youtube_dl\nimport shutil\nfrom urlextract import URLExtract\nfrom telegram.ext import Updater, CommandHandler, MessageHandler, Filters\nfrom telegram import ChatAction\nfrom functools import wraps\nimport re\nimport urllib\n\nos.chdir(os.path.dirname(__file__))\nlogging.getLogger(\"filelock\").disabled = True\n\ndownload_dir = \"videos\"\nmax_file_size = 100\nsend_timeout = 100\n\ncur_file_counter = 0\ndownloaded_files = []\n\n\n# custom url regexs\ntiktok_regex = re.compile(\"https?://(?:vm\\.)?tiktok\\.com/[^/]+/?\")\nyoutube_regex = re.compile(\"https?://(?:www\\.)?youtube\\.com/[^/]+/?\")\nyoutube_mobile_regex = re.compile(\"https?://youtu\\.be\\/[^/]+/?\")\nyoutube_music = re.compile(\"https?://music\\.youtube\\.com\\/watch\\?[^/]+/?\")\n\n# links that youtube dl wont catch\nproblem_regex = [tiktok_regex]\n# links that should download without a command\nauto_download_regex = [tiktok_regex, youtube_regex, youtube_mobile_regex, youtube_music]\n# links that are guaranteed audio\naudio_regex = [youtube_music]\n\n\ndef is_downloadable(url: str) -> bool:\n \"\"\"Check if a url is one which can be downloaded\"\"\"\n for extractor in youtube_dl.extractor.gen_extractors():\n if extractor.suitable(url) and extractor.IE_NAME != \"generic\":\n return True\n return any(regex.match(url) is not None for regex in problem_regex)\n\n\ndef is_audio(url: str) -> bool:\n \"\"\"Check if a url is guaranteed to be audio\"\"\"\n return any(regex.match(url) is not None for regex in audio_regex)\n\n\ndef is_auto_download(url: str) -> bool:\n \"\"\"Check if a url is one which should automatically download\"\"\"\n return any(regex.match(url) is not None for regex in auto_download_regex)\n\n\ndef download_video(url: str, ydl_opts: dict) -> int:\n \"\"\"Download a url to the computer\"\"\"\n with youtube_dl.YoutubeDL(ydl_opts) as ydl:\n try:\n info = ydl.extract_info(url, download=True)\n title = info.get(\"title\", \"Error Getting Title\")\n err = False\n except Exception:\n err = True\n title = None\n return title, err\n\n\ndef extract_url(message):\n \"\"\"Get the first url in a string\"\"\"\n urls = URLExtract().find_urls(message)\n if len(urls) == 0:\n return None\n url = urls[0]\n return url\n\n\ndef parse_message(\n update, context, is_command=False, force_audio=False, check_audio=True\n):\n \"\"\"Parse a any message and check whether or not it contains a link,\n whether or not it is a command, and whether or not it can actually download something\"\"\"\n global cur_file_counter\n\n # ensure it's a real message\n if not hasattr(update.message, \"text\"):\n return\n\n # check if dm\n is_dm = int(update.message.chat.id) > 0\n\n url = extract_url(update.message.text)\n\n # if it's a check up one reply for url\n if url is None:\n if is_command:\n if update.message.reply_to_message is not None:\n url = extract_url(update.message.reply_to_message.text)\n if url is None:\n return\n else:\n return\n\n # make sure we can actually download the url\n if (not is_command and not is_auto_download(url)) and not is_dm:\n return\n\n if not is_downloadable(url):\n return\n\n if is_audio(url) and check_audio:\n force_audio = True\n\n # check if the file has already been downloaded\n for downloaded in downloaded_files:\n if downloaded[\"url\"] == url and (not force_audio == downloaded[\"is_video\"]):\n user = update.message.from_user[\"username\"]\n # check if the sent file was in another chat\n if update.message.chat_id != downloaded[\"message\"].chat_id:\n already_sent_message = downloaded[\"message\"].forward(\n update.message.chat_id\n )\n else:\n already_sent_message = downloaded[\"message\"]\n already_sent_message.reply_text(f\"@{user}\")\n return\n\n logging.debug(f\"Downloading from {url}...\")\n\n ydl_opts = {\n \"max_filesize\": max_file_size * 1000000,\n \"ignoreerrors\": False,\n }\n if force_audio:\n # download the url as audio\n context.bot.send_chat_action(\n chat_id=update.effective_message.chat_id, action=ChatAction.UPLOAD_AUDIO\n )\n\n filename = f\"{cur_file_counter}.mp3\"\n ydl_opts.update(\n {\n \"outtmpl\": filename,\n \"format\": \"bestaudio/best\",\n \"postprocessors\": [\n {\n \"key\": \"FFmpegExtractAudio\",\n \"preferredcodec\": \"mp3\",\n \"preferredquality\": \"192\",\n }\n ],\n }\n )\n else:\n # download the url as a video\n context.bot.send_chat_action(\n chat_id=update.effective_message.chat_id, action=ChatAction.UPLOAD_VIDEO\n )\n filename = f\"{cur_file_counter}.mp4\"\n ydl_opts.update(\n {\n \"format\": \"mp4\",\n \"outtmpl\": filename,\n }\n )\n\n # download and check for errors\n title, err = download_video(url, ydl_opts)\n\n if not os.path.exists(filename) or err:\n update.message.reply_text(\"Error downloading\")\n return\n\n cur_file_counter += 1\n\n if force_audio:\n # send the file as audio\n new_filename = f\"{title}.mp3\"\n os.rename(filename, new_filename)\n reply_message = update.message.reply_audio(\n open(new_filename, \"rb\"), timeout=send_timeout, title=title\n )\n else:\n # send the file as a videos\n reply_message = update.message.reply_video(\n open(filename, \"rb\"), timeout=send_timeout\n )\n downloaded_files.append(\n {\"url\": url, \"message\": reply_message, \"is_video\": not force_audio}\n )\n\n\ndef help_command(update, context):\n update.message.reply_text(\n (\n \"i download videos and audio.\\n\"\n + \"\\n\"\n + \"/download - provide/reply to a url to download it as a video\\n\"\n + \"/daudio - provide/reply to a url to download it as an mp3\\n\"\n + \"/help - bring up this help menu\\n\"\n + \"\\n\"\n + \"powered by youtube-dl.\\n\"\n )\n )\n\n\ndef download_command(update, context):\n parse_message(update, context, is_command=True)\n\n\ndef audio_command(update, context):\n parse_message(update, context, is_command=True, force_audio=True)\n\n\ndef on_message(update, context):\n parse_message(update, context, check_audio=True)\n\n\ndef main():\n # get telegram bot token\n with open(\"token.txt\", \"r\") as token_file:\n token = token_file.read().rstrip(\"\\n\")\n updater = Updater(token, use_context=True)\n\n # remove old download directory\n if os.path.exists(download_dir):\n shutil.rmtree(download_dir)\n\n # make download directory\n os.mkdir(download_dir)\n os.chdir(download_dir)\n\n # add message handler\n updater.dispatcher.add_handler(CommandHandler(\"help\", help_command))\n updater.dispatcher.add_handler(CommandHandler(\"download\", download_command))\n updater.dispatcher.add_handler(CommandHandler(\"daudio\", audio_command))\n updater.dispatcher.add_handler(\n MessageHandler(callback=on_message, filters=Filters.text)\n )\n\n # wait for messages\n updater.start_polling()\n updater.idle()\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"src/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":7569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"447807245","text":"# encoding=utf8\nimport matplotlib.pyplot as plt\n\nsquares = [1, 4, 9, 16, 25, 36]\n# linewidth设置线条宽度\nplt.plot(squares, linewidth=5)\n# title表示标题,#fontSize表示字体大小\n# xlabel表示X轴的标签,#ylabel表示Y轴的标签\nplt.title(\"Square Number\", fontSize=24)\nplt.xlabel(\"value\", fontSize=14)\nplt.ylabel(\"Square of Value\", fontSize=14)\n# tick_params:刻度参数\n# axis=\"both\":表示x轴和y轴 ,labelsize表示刻度字体的大小\n# color表示刻度的颜色。\nplt.tick_params(axis=\"both\", labelsize=10, color=\"red\")\nplt.show()\n","sub_path":"python_DataVisualization/part01/demo02.py","file_name":"demo02.py","file_ext":"py","file_size_in_byte":563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"263875312","text":"import numpy as np\nfrom PIL import Image\nimport matplotlib.pyplot as plt\n\n\n################################################################################\ndef convert_to_gray_scale_arr(filename):\n image_gray = Image.open(filename).convert('RGB').convert('L')\n gray_scale = np.array(image_gray)\n return image_gray.width, image_gray.height, gray_scale\n\n\n################################################################################\ndef from_gray_to_3_channel(width, height, image_data):\n output = np.zeros((height, width, 3), dtype=int)\n for y in range(height):\n for x in range(width):\n rgb = [image_data[y][x], image_data[y][x], image_data[y][x]]\n output[y][x] = rgb\n return output\n\n\n################################################################################\ndef zero_padding(x, y, width, height):\n pad = False\n if x < 0:\n x = 0\n pad = True\n\n if x >= width:\n x = width - 1\n pad = True\n\n if y < 0:\n y = 0\n pad = True\n\n if y >= height:\n y = height - 1\n pad = True\n\n return pad, x, y\n\n\n################################################################################\ndef median_filter(width, height, image_data):\n kernel_size = 3\n mid_point = int((kernel_size * kernel_size) / 2)\n window = np.zeros((kernel_size * kernel_size), dtype=int)\n window_debug = np.zeros((kernel_size, kernel_size), dtype=int)\n output_image = np.zeros((height, width), dtype=int)\n edge_x = int(kernel_size / 2)\n edge_y = int(kernel_size / 2)\n out_x = out_y = 0\n for x in range(width - edge_x + 1):\n print(\"processing \" + str(x) + \" of \" + str(width - edge_x + 1))\n for y in range(height - edge_y + 1):\n i = 0\n for fx in range(kernel_size):\n for fy in range(kernel_size):\n x_image = x + fx - edge_x\n y_image = y + fy - edge_y\n\n # zero padding\n pad, x_image, y_image = zero_padding(x_image, y_image, width, height)\n if pad:\n window[i] = 0\n else:\n window[i] = image_data[x_image][y_image]\n pass\n\n window_debug[fx][fy] = window[i]\n\n i = i + 1\n pass\n # print(window_debug)\n sorted_values = np.sort(window)\n median = sorted_values[mid_point]\n # print(sorted_values)\n # print(median)\n output_image[out_y][out_x] = median\n out_x = out_x + 1\n # print(\"\\n\")\n pass\n out_x = 0\n out_y = out_y + 1\n # print(\"----\\n\")\n pass\n\n return output_image\n\n\n################################################################################\ndef test():\n [width, height, image_data] = convert_to_gray_scale_arr(\"pic1_noisy.png\")\n filtered = median_filter(width, height, image_data)\n filtered2 = median_filter(width, height, filtered)\n\n src_image = from_gray_to_3_channel(width, height, image_data)\n filtered_image = from_gray_to_3_channel(width, height, filtered)\n filtered_image2 = from_gray_to_3_channel(width, height, filtered2)\n\n fig = plt.figure()\n ax1 = fig.add_subplot(1, 3, 1)\n ax1.imshow(src_image)\n\n ax2 = fig.add_subplot(1, 3, 2)\n ax2.imshow(filtered_image)\n\n ax3 = fig.add_subplot(1, 3, 3)\n ax3.imshow(filtered_image2)\n\n plt.show()\n\n\ntest()\n","sub_path":"image_processing/filters/median_filter.py","file_name":"median_filter.py","file_ext":"py","file_size_in_byte":3502,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"619886915","text":"\n\nimport tensorflow as tf\nfrom tensorflow.contrib.session_bundle import exporter\n\n# Import MNIST data\nfrom tensorflow.examples.tutorials.mnist import input_data\nmnist = input_data.read_data_sets(\"MNIST_data/\", one_hot=True,validation_size=0)\n\n\nsess = tf.Session()\ntf.logging.set_verbosity(tf.logging.INFO)\n\nx = tf.placeholder(tf.float32, [None, 784],name='x')\nW = tf.Variable(tf.zeros([784, 10]),name='W')\nb = tf.Variable(tf.zeros([10]),name='b')\n\ny = tf.nn.softmax(tf.matmul(x, W) + b,name='y')\ny_ = tf.placeholder(tf.float32, [None, 10],name='y_')\ntf.add_to_collection('variablesw',W)\ntf.add_to_collection('variablesb',b)\n\ncross_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=y, labels=y_))\n\ntrain_step = tf.train.GradientDescentOptimizer(0.5).minimize(cross_entropy)\n\n# save summaries for visualization\ntf.summary.histogram('weights', W)\ntf.summary.histogram('max_weight', tf.reduce_max(W))\ntf.summary.histogram('bias', b)\ntf.summary.scalar('cross_entropy', cross_entropy)\ntf.summary.histogram('cross_hist', cross_entropy)\n\n# merge all summaries into one op\nmerged=tf.summary.merge_all()\n\ntrainwriter=tf.summary.FileWriter('data/mnist_model'+'/logs/train',sess.graph)\n\ninit = tf.global_variables_initializer()\nsess.run(init)\n\nfor i in range(1000):\n batch_xs, batch_ys = mnist.train.next_batch(100)\n summary, _ = sess.run([merged, train_step], feed_dict={x: batch_xs, y_: batch_ys})\n trainwriter.add_summary(summary, i)\n\n# model export path\nexport_path = 'data/mnist_model'\nprint('Exporting trained model to', export_path)\n\n#\nsaver = tf.train.Saver(sharded=True)\nmodel_exporter = exporter.Exporter(saver)\nmodel_exporter.init(\n sess.graph.as_graph_def(),\n named_graph_signatures={\n 'inputs': exporter.generic_signature({'images': x}),\n 'outputs': exporter.generic_signature({'scores': y})})\n\nmodel_exporter.export(export_path, tf.constant(1), sess)\n\n\"\"\"\ncan also save the model using saver as follows\nsaver.save(sess, '/Volumes/Data/BigDataAnalytics/ICMP8/In_Class_MNIST_SOFTMAX/data/mnist_model')\n\"\"\"","sub_path":"LAB4/Source/MNIST_SOFTWARE/mnist_train.py","file_name":"mnist_train.py","file_ext":"py","file_size_in_byte":2056,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"42561105","text":"import groupy\n\n\nclass BotCreator(object):\n\n def getname(self):\n name = ''\n while name.strip() == '':\n name = input('Pick a name for the bot: ')\n name = str(name).strip()\n return name\n\n def getgroup(self):\n i = 0\n grouplist = groupy.Group.list()\n if grouplist is None:\n print(\"You need groups, fam.\")\n raise SystemExit\n for groups in grouplist:\n print(str(i) + \") \" + str(groups.name))\n i += 1\n\n selection = 0\n if i != 1:\n selection = None\n while type(selection) is not int or selection < 0 or selection > len(grouplist) or str(selection).strip == '':\n selection = input(\"Select Group #: \")\n try:\n selection = int(selection)\n except ValueError:\n print(\"not a number\")\n\n return grouplist[selection]\n\n def getimage(self):\n answer = input(\"Would you like to provide an image URL? (y/n): \").strip().lower()\n while answer != 'y' and answer != 'n':\n answer = input(\"Would you like to provide an image URL? (y/n): \").strip().lower()\n\n if answer == 'n':\n return None\n elif answer == 'y':\n newans = 'n'\n imgurl = ''\n while newans != 'y':\n imgurl = input(\"Enter image URL: \")\n newans = input(\"Is this URL right? \" + imgurl + \" (y/n): \")\n return imgurl\n\n def getcallurl(self):\n answer = input(\"Would you like to provide a callback URL? (y/n): \").strip().lower()\n while answer != 'y' and answer != 'n':\n answer = input(\"Would you like to provide a callback URL? (y/n): \").strip().lower()\n\n if answer == 'n':\n return None\n elif answer == 'y':\n newans = 'n'\n cburl = ''\n while newans != 'y':\n cburl = input(\"Enter callback URL: \")\n while cburl[0:7] != 'http://' and cburl[0:8] != 'https://' and cburl != '':\n print(\"Url must have 'http://' or 'https://'\")\n cburl = input(\"Enter callback URL: \")\n\n newans = input(\"Is this URL right? \" + cburl + \" (y/n): \")\n if cburl == '':\n cburl = None\n return cburl\n\n def create(self):\n print(\"Time to create a bot!\")\n ans = None\n while ans != 'y' and ans != 'n':\n ans = input(\"Would you like to continue? (y/n): \")\n if ans == 'n':\n raise SystemExit\n name = self.getname()\n group = self.getgroup()\n image = self.getimage()\n callurl = self.getcallurl()\n try:\n groupy.Bot.create(name, group, image, callurl)\n except groupy.api.errors.ApiError:\n print(\"\\nERROR\\n#########################################\\ncallback URL already registered for group\\nBot not created\\n#########################################\")\n raise SystemExit","sub_path":"bin/BotCreator.py","file_name":"BotCreator.py","file_ext":"py","file_size_in_byte":3059,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"251273307","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('website', '0007_auto_20150510_2050'),\n ('chat_server', '0002_chatmessage_user'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='chatmessage',\n name='crowd',\n field=models.ForeignKey(default=1, to='website.Crowd'),\n preserve_default=False,\n ),\n ]\n","sub_path":"incrowd/chat_server/migrations/0003_chatmessage_crowd.py","file_name":"0003_chatmessage_crowd.py","file_ext":"py","file_size_in_byte":507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"592917469","text":"from GenericRequest import GenericRequest\nfrom ..manager import PatternManager\nfrom ..util import Report\n\nfrom datetime import datetime\n\nCLAN_LOG_UNKNOWN = 0\nCLAN_LOG_FAX = 1\nCLAN_LOG_ATTACK = 2\nCLAN_LOG_WHITELISTED_PLAYER = 3\nCLAN_LOG_JOINED_ANOTHER_CLAN = 4\nCLAN_LOG_WHITELISTED_IN = 5\nCLAN_LOG_STASH_ADD = 6\nCLAN_LOG_STASH_REMOVE = 7\nCLAN_LOG_MEAT_SPENT_ARMY = 8\nCLAN_LOG_CHANGED_RANK = 9\nCLAN_LOG_CHANGED_TITLE = 10\n\n\nclass WhitelistPlayerRequest(GenericRequest):\n def __init__(self, session, player, level, title=\"\"):\n super(WhitelistPlayerRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_whitelist.php\"\n self.requestData[\"action\"] = \"add\"\n self.requestData[\"pwd\"] = session.pwd\n self.requestData[\"addwho\"] = player\n self.requestData[\"level\"] = level\n self.requestData[\"title\"] = title\n\n\nclass BootClanMemberRequest(GenericRequest):\n def __init__(self, session, userId):\n super(BootClanMemberRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_members.php\"\n self.requestData['pwd'] = session.pwd\n self.requestData['action'] = 'modify'\n self.requestData['begin'] = '1'\n self.requestData['pids[]'] = userId\n self.requestData['boot%s' % userId] = 'on'\n\n\nclass WhitelistRequest(GenericRequest):\n \"\"\"Retrieves information from the clan whitelist page.\"\"\"\n\n def __init__(self, session):\n super(WhitelistRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_whitelist.php\"\n\n def parseResponse(self):\n # Get the set of clan ranks.\n ranks = []\n ranksById = {}\n rankContainerPattern = PatternManager.getPattern('clanRankContainer')\n match = rankContainerPattern.search(self.responseText)\n if match:\n rankText = match.group(1)\n rankPattern = PatternManager.getPattern('clanRank')\n for rankMatch in rankPattern.finditer(rankText):\n rank = {\n \"rankId\": int(rankMatch.group(1)),\n \"rankName\": rankMatch.group(2),\n \"rankNumber\": int(rankMatch.group(3))\n }\n ranks.append(rank)\n ranksById[rank[\"rankId\"]] = rank\n\n # Get a list of users who are whitelisted to the clan.\n members = []\n memberPattern = PatternManager.getPattern('clanWhitelistMember')\n for match in memberPattern.finditer(self.responseText):\n member = {\n \"userId\": match.group('userId'),\n \"userName\": match.group('userName'),\n \"clanTitle\": match.group('clanTitle')\n }\n rankId = match.group('clanRankId')\n rankName = match.group('clanRankName')\n if rankId is not None:\n rank = ranksById[int(rankId)]\n member[\"rankId\"] = rank[\"rankId\"]\n member[\"rankName\"] = rank[\"rankName\"]\n member[\"rankNumber\"] = rank[\"rankNumber\"]\n elif rankName is not None:\n member[\"rankName\"] = rankName\n foundRank = False\n for rank in ranks:\n if rank[\"rankName\"] == rankName:\n foundRank = True\n break\n if not foundRank:\n rank = {\n \"rankId\": -1,\n \"rankName\": rankName,\n \"rankNumber\": -1\n }\n ranks.append(rank)\n members.append(member)\n\n self.responseData[\"ranks\"] = ranks\n self.responseData[\"members\"] = members\n\n\nclass LoadClanAdminRequest(GenericRequest):\n \"\"\"Loads the clan administration page.\"\"\"\n\n def __init__(self, session):\n super(LoadClanAdminRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_admin.php\"\n\n def parseResponse(self):\n # Get the clan name.\n namePattern = PatternManager.getPattern(\"clanName\")\n match = namePattern.search(self.responseText)\n self.responseData[\"clanName\"] = match.group(1)\n\n # Get the clan credo.\n credoPattern = PatternManager.getPattern(\"clanCredo\")\n match = credoPattern.search(self.responseText)\n self.responseData[\"clanCredo\"] = match.group(1)\n\n # Get the clan website.\n websitePattern = PatternManager.getPattern(\"clanWebsite\")\n match = websitePattern.search(self.responseText)\n self.responseData[\"clanWebsite\"] = match.group(1)\n\n # See if the clan is accepting applications.\n clanAcceptingAppsPattern = PatternManager.getPattern(\"clanAcceptingApps\")\n if clanAcceptingAppsPattern.search(self.responseText):\n self.responseData[\"acceptingApps\"] = True\n else:\n self.responseData[\"acceptingApps\"] = False\n\n\nclass ClanLogRequest(GenericRequest):\n \"\"\"Retrieves the clan activity log.\"\"\"\n\n def __init__(self, session):\n super(ClanLogRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_log.php\"\n\n def parseResponse(self):\n entries = []\n entryPattern = PatternManager.getPattern('clanLogEntry')\n for entryMatch in entryPattern.finditer(self.responseText):\n entry = {}\n date = entryMatch.group('date')\n entry['date'] = datetime.strptime(date, \"%m/%d/%y, %I:%M%p\")\n entry['userId'] = int(entryMatch.group('userId'))\n entry['userName'] = entryMatch.group('userName')\n action = entryMatch.group('action')\n foundAction = False\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogFax')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_FAX\n entry['monster'] = match.group('monsterName')\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogAttack')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_ATTACK\n entry['clanName'] = match.group('clanName')\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogWhitelistAdd')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_WHITELISTED_PLAYER\n entry['targetUserName'] = match.group('userName')\n entry['targetUserId'] = int(match.group('userId'))\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogPlayerJoinedAnotherClan')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_JOINED_ANOTHER_CLAN\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogPlayerJoinedClanWhitelist')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_WHITELISTED_IN\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogStashItemAdd')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_STASH_ADD\n entry['itemName'] = match.group('itemName')\n entry['quantity'] = int(match.group('quantity').replace(',', ''))\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogStashItemRemove')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_STASH_REMOVE\n entry['itemName'] = match.group('itemName')\n entry['quantity'] = int(match.group('quantity').replace(',', ''))\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogMeatSpentArmy')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_MEAT_SPENT_ARMY\n entry['meat'] = int(match.group('meat').replace(',', ''))\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogChangedRank')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_CHANGED_RANK\n entry['targetUserName'] = match.group('userName')\n entry['targetUserId'] = int(match.group('userId'))\n\n if not foundAction:\n pattern = PatternManager.getPattern('clanLogChangedTitle')\n match = pattern.match(action)\n if match:\n foundAction = True\n entry['type'] = CLAN_LOG_CHANGED_RANK\n entry['targetUserName'] = match.group('userName')\n entry['targetUserId'] = int(match.group('userId'))\n entry['clanTitle'] = match.group('clanTitle')\n\n if not foundAction:\n Report.error(\"request\", \"Unknown clan log action: %s\" % action)\n entry['type'] = CLAN_LOG_UNKNOWN\n entry['action'] = action\n\n entries.append(entry)\n\n self.responseData[\"entries\"] = entries\n\n\nclass ToggleAcceptingClanApplicationsRequest(GenericRequest):\n \"\"\"Toggle whether or not the clan accepts new applications.\"\"\"\n\n def __init__(self, session):\n super(ToggleAcceptingClanApplicationsRequest, self).__init__(session)\n self.url = session.serverURL + \"clan_admin.php?action=noapp\"","sub_path":"src/kol/request/ClanAdmin.py","file_name":"ClanAdmin.py","file_ext":"py","file_size_in_byte":10023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"157144428","text":"import seaborn as sns\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.metrics import confusion_matrix\nfrom sklearn.utils.multiclass import unique_labels\n\n#------------------------------------------------------------------------------\ndef imprime_matriz_de_confusao(cm, title,pathfigura):\n sns.set(font_scale=3.5)\n cmn = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis]\n fig, ax = plt.subplots(figsize=(65,40))\n sns.heatmap(cmn, ax=ax,annot=True, fmt='.2f', linewidth=2,linecolor='lightgray',xticklabels=target_names, yticklabels=target_names,cmap=\"BuPu\",annot_kws={\"size\": 28,\"weight\": \"bold\"})#,annot_kws={\"fontsize\":18}\n # ax.figure.axes[-1].yaxis.label.set_size(20)\n ax.set_title(title, fontsize =58)\n plt.ylabel('Assunto Principal', fontsize = 45 )\n plt.xlabel('Assunto Predito', fontsize = 45)\n # plt.show(block=True)\n fig.tight_layout()\n plt.savefig(\"{0}{1}.png\".format(pathfigura, 'ConfusionMatrix_' + title), bbox_inches = 'tight',\n pad_inches = 0)\n\n#------------------------------------------------------------------------------\n#MLP\n#------------------------------------------------------------------------------\n\nmlp = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP26_MelhoresModelos_TextsoReduzidos_LSI/predicao_LSI250_Multi-Layer Perceptron.csv'\nmlp = pd.read_csv(mlp)\nmlp_y_true = mlp.y_true\nmlp_y_pred = mlp.y_pred\n\ncm =confusion_matrix(mlp_y_true, mlp_y_pred)\npathfigura = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP26_MelhoresModelos_TextsoReduzidos_LSI/'\nimprime_matriz_de_confusao(cm, 'Multilayer Perceptron', pathfigura)\n\n#------------------------------------------------------------------------------\n#RANDOM FOREST\n#------------------------------------------------------------------------------\n\nmlp = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP25_MelhoresModelos_TextsoReduzidos_BM25/predicao_BM25_Random Forest.csv'\nmlp = pd.read_csv(mlp)\nmlp_y_true = mlp.y_true\nmlp_y_pred = mlp.y_pred\n\ncm =confusion_matrix(mlp_y_true, mlp_y_pred)\npathfigura = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP25_MelhoresModelos_TextsoReduzidos_BM25/'\nimprime_matriz_de_confusao(cm, 'Random Forest', pathfigura)\n\n#------------------------------------------------------------------------------\n#SVM\n#------------------------------------------------------------------------------\n\nmlp = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP24_MelhoresModelos_TextsoReduzidos_TFIDF/predicao_SVM.csv'\nmlp = pd.read_csv(mlp)\nmlp_y_true = mlp.y_true\nmlp_y_pred = mlp.y_pred\n\ncm =confusion_matrix(mlp_y_true, mlp_y_pred)\npathfigura = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP24_MelhoresModelos_TextsoReduzidos_TFIDF/'\nimprime_matriz_de_confusao(cm, 'SVM', pathfigura)\n\n#------------------------------------------------------------------------------\n#NAIVE BAYES\n#------------------------------------------------------------------------------\n\nmlp = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP25_MelhoresModelos_TextsoReduzidos_BM25/predicao_BM25_Multinomial Naive Bayes.csv'\nmlp = pd.read_csv(mlp)\nmlp_y_true = mlp.y_true\nmlp_y_pred = mlp.y_pred\n\ncm =confusion_matrix(mlp_y_true, mlp_y_pred)\npathfigura = '/media/DATA/classificadorDeAssuntos/Dados/Resultados/EXP25_MelhoresModelos_TextsoReduzidos_BM25/'\nimprime_matriz_de_confusao(cm, 'Naïve Bayes', pathfigura)","sub_path":"Full/Codigo/001_Analises/_013_Plota_MatrizConfusao.py","file_name":"_013_Plota_MatrizConfusao.py","file_ext":"py","file_size_in_byte":3419,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"258397640","text":"import numpy as np\nimport shutil\nfrom astropy import cosmology as cosmo\n\nfrom autolens.data import ccd as im\nfrom autolens.data.array import grids, mask as msk, scaled_array\nfrom autolens.lens import lens_data as li, lens_fit\nfrom autolens.lens import ray_tracing\nfrom autolens.lens.plotters import lens_fit_hyper_plotters\nfrom autolens.model.galaxy import galaxy as g\nfrom autolens.model.profiles import light_profiles as lp, mass_profiles as mp\nfrom test.fixtures import *\n\n\n@pytest.fixture(name='lens_fit_plotter_path')\ndef make_lens_fit_plotter_setup():\n return \"{}/../../test_files/plotting/fit/\".format(os.path.dirname(os.path.realpath(__file__)))\n\n\n@pytest.fixture(name='galaxy_light')\ndef make_galaxy_light():\n return g.Galaxy(light=lp.EllipticalSersic(intensity=1.0), redshift=2.0)\n\n\n@pytest.fixture(name='galaxy_mass')\ndef make_galaxy_mass():\n return g.Galaxy(mass=mp.SphericalIsothermal(einstein_radius=1.0), redshift=1.0)\n\n\n@pytest.fixture(name='grid_stack')\ndef make_grid_stack():\n return grids.GridStack.from_shape_pixel_scale_and_sub_grid_size(shape=(100, 100), pixel_scale=0.05, sub_grid_size=2)\n\n\n@pytest.fixture(name='image')\ndef make_image():\n image = scaled_array.ScaledSquarePixelArray(array=np.ones((3, 3)), pixel_scale=1.0)\n noise_map = im.NoiseMap(array=2.0 * np.ones((3, 3)), pixel_scale=1.0)\n psf = im.PSF(array=3.0 * np.ones((1, 1)), pixel_scale=1.0)\n\n return im.CCDData(image=image, pixel_scale=1.0, noise_map=noise_map, psf=psf)\n\n\n@pytest.fixture(name='positions')\ndef make_positions():\n positions = [[[0.1, 0.1], [0.2, 0.2]], [[0.3, 0.3]]]\n return list(map(lambda position_set: np.asarray(position_set), positions))\n\n\n@pytest.fixture(name='mask')\ndef make_mask():\n return msk.Mask.circular(shape=((3, 3)), pixel_scale=0.1, radius_arcsec=0.1)\n\n\n@pytest.fixture(name='lens_data')\ndef make_lens_image(image, mask):\n return li.LensData(ccd_data=image, mask=mask)\n\n\n@pytest.fixture(name='fit_lens_only')\ndef make_fit_lens_only(lens_data, galaxy_light):\n tracer = ray_tracing.TracerImagePlane(lens_galaxies=[galaxy_light], image_plane_grid_stack=lens_data.grid_stack,\n cosmology=cosmo.Planck15)\n return lens_fit.fit_lens_data_with_tracer(lens_data=lens_data, tracer=tracer)\n\n\n@pytest.fixture(name='fit_source_and_lens')\ndef make_fit_source_and_lens(lens_data, galaxy_light, galaxy_mass):\n tracer = ray_tracing.TracerImageSourcePlanes(lens_galaxies=[galaxy_mass], source_galaxies=[galaxy_light],\n image_plane_grid_stack=lens_data.grid_stack, cosmology=cosmo.Planck15)\n return lens_fit.fit_lens_data_with_tracer(lens_data=lens_data, tracer=tracer)\n\n\n@pytest.fixture(name='hyper')\ndef make_hyper():\n class Hyper(object):\n\n def __init__(self):\n pass\n\n hyper = Hyper()\n\n hyper.hyper_model_image = np.array([[3.0, 5.0, 7.0],\n [9.0, 8.0, 1.0],\n [4.0, 0.0, 9.0]])\n hyper.hyper_galaxy_images = [np.array([[1.0, 3.0, 5.0],\n [7.0, 9.0, 8.0],\n [6.0, 4.0, 0.0]])]\n hyper.hyper_minimum_values = [0.2, 0.8]\n\n hyper_galaxy = g.HyperGalaxy(contribution_factor=4.0, noise_factor=2.0, noise_power=3.0)\n hyper.hyper_galaxy = g.Galaxy(light=lp.EllipticalSersic(intensity=1.0), hyper_galaxy=hyper_galaxy)\n return hyper\n\n\n@pytest.fixture(name='lens_hyper_image')\ndef make_lens_hyper_image(image, mask, hyper):\n return li.LensDataHyper(ccd_data=image, mask=mask, hyper_model_image=hyper.hyper_model_image,\n hyper_galaxy_images=hyper.hyper_galaxy_images,\n hyper_minimum_values=hyper.hyper_minimum_values)\n\n\n@pytest.fixture(name='fit_hyper_lens_only')\ndef make_fit_hyper_lens_only(lens_hyper_image, hyper):\n tracer = ray_tracing.TracerImagePlane(lens_galaxies=[hyper.hyper_galaxy],\n image_plane_grid_stack=lens_hyper_image.grid_stack)\n return lens_fit.hyper_fit_lens_data_with_tracer(lens_data_hyper=lens_hyper_image, tracer=tracer)\n\n\ndef test__fit_sub_plot_hyper_lens_only(fit_lens_only, fit_hyper_lens_only, plot_patch,\n lens_fit_plotter_path):\n\n lens_fit_hyper_plotters.plot_fit_subplot(fit_hyper=fit_hyper_lens_only, fit=fit_lens_only, should_plot_mask=True,\n extract_array_from_mask=True, zoom_around_mask=True,\n output_path=lens_fit_plotter_path,\n output_filename='hyper_lens_fit', output_format='png')\n\n assert lens_fit_plotter_path + 'hyper_lens_fit.png' in plot_patch.paths\n\n\ndef test__fit_individuals__hyper_lens_only__depedent_on_input(fit_hyper_lens_only, fit_lens_only, plot_patch,\n lens_fit_plotter_path):\n\n lens_fit_hyper_plotters.plot_fit_individuals(fit_hyper=fit_hyper_lens_only, fit=fit_lens_only,\n should_plot_mask=True, extract_array_from_mask=True,\n zoom_around_mask=True,\n should_plot_noise_map=True,\n should_plot_model_image=True,\n should_plot_chi_squared_map=True,\n output_path=lens_fit_plotter_path, output_format='png')\n\n assert lens_fit_plotter_path + 'fit_model_image.png' in plot_patch.paths\n\n assert lens_fit_plotter_path + 'fit_residual_map.png' not in plot_patch.paths\n\n assert lens_fit_plotter_path + 'fit_hyper_chi_squared_map.png' in plot_patch.paths\n\n assert lens_fit_plotter_path + 'fit_hyper_noise_map.png' in plot_patch.paths\n","sub_path":"test/lens/plotters/test_lens_fit_hyper_plotters.py","file_name":"test_lens_fit_hyper_plotters.py","file_ext":"py","file_size_in_byte":5978,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"254412236","text":"# -*- encoding: utf-8 -*-\nfrom django.test import TestCase\nfrom bank.models import Account, Transaction, Client, Manager\nfrom django.db import models\nfrom django.contrib.auth.models import User, Group\nfrom bank.templatetags.helpers import group_required, get_transaction_operation\n\nclass AccountTestCase(TestCase):\n\n def setUp(self):\n account_holder = Group()\n account_holder.name = 'account_holder'\n account_holder.save()\n\n user = User.objects.create_user(\n username=\"foo\",\n email=\"foo@bar.com\",\n password=\"123123\"\n )\n user.groups.add(Group.objects.get(name='account_holder'))\n\n client = Client.objects.create(\n user=user,\n cpf=\"foo@bar.com\",\n created_at=models.DateTimeField(auto_now_add=True)\n )\n account = Account.objects.create(\n \tbalance=99999, \n \tclient= client, \n \tcreated_at=models.DateTimeField(auto_now_add=True)\n )\n\n # ---- Models ---- #\n\n def test_accounts_deposit(self):\n account = Account.objects.get(id=1)\n self.assertEqual(account.deposit(1), 100000)\n\n def test_accounts_withdraw(self):\n account = Account.objects.get(id=1)\n self.assertEqual(account.withdraw(1), 99998)\n\n def test_accounts_get_total_balance(self):\n client = Client.objects.get(id=1)\n total_balance = Account.get_total_balance(client)\n self.assertEqual(total_balance, 99999)\n\n\n # ---- Helpers ---- #\n\n def test_helper_group_required(self):\n user = User.objects.get(id=1)\n self.assertEqual(group_required(user, 'account_holder'), True)\n\n def test_helper_get_transaction_operation_deposit(self):\n transaction = Transaction.objects.create(\n account = Account.objects.get(id=1),\n amount = 99999,\n operation = Transaction.OPERATIONS_DICT['deposit'],\n created_at = models.DateTimeField(auto_now_add=True)\n )\n self.assertEqual(get_transaction_operation(transaction), 'deposit')\n\n def test_helper_get_transaction_operation_withdraw(self):\n transaction = Transaction.objects.create(\n account = Account.objects.get(id=1),\n amount = 99999,\n operation = Transaction.OPERATIONS_DICT['withdraw'],\n created_at = models.DateTimeField(auto_now_add=True)\n )\n self.assertEqual(get_transaction_operation(transaction), 'withdraw')","sub_path":"orama/orama_project/bank/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":2477,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"322402059","text":"import random\nimport matplotlib.pyplot as plt\n\nfrom chainer import datasets\nfrom chainer import serializers\n\nfrom mnist_train import MLP\n\n\ndef main():\n # Load dataset\n train, test = datasets.mnist.get_mnist()\n\n # Load model\n model = MLP()\n serializers.load_npz('mnist_out/model_epoch-10', model)\n\n # Show the output\n no = random.randint(0, len(test) - 1)\n x, t = test[no]\n plt.imshow(x.reshape(28, 28), cmap='gray')\n plt.savefig('mnist_out/mnist_eval.png')\n print('label:', t)\n\n y = model(x[None, ...])\n\n print('predicted_label:', y.array.argmax(axis=1)[0])\n\n\nif __name__ == \"__main__\":\n main()\n\n\n\n","sub_path":"MNIST/mnist_eval.py","file_name":"mnist_eval.py","file_ext":"py","file_size_in_byte":640,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"241755488","text":"import os\nimport sys\nimport csv\nimport time\nimport mmap\nimport queue\nimport shutil\nimport urllib\nimport datetime\nimport lxml.html\nimport threading\nfrom os import listdir\nfrom bs4 import BeautifulSoup\nfrom os.path import isfile, join\nfrom urllib.request import urlopen, Request\n\n######################################################\n###### CONFIG VARIABLES - Changeable Parameters ######\n######################################################\n\nmax_level = 1 # Depth of graph\n\n# Controls the percentage of followers to be scraped\n# Eg: {1:(10000, 10}, 2:(5000, 5)} means scrape \n# 10000 + (10/100) * total_followers of the followers\n# at level 1 and 5000 + (5/100) * total_followers \n# of the followers at level 2\nmax_edges_restriction = {1: (0, 100)}\n\n# Controls the number of followers to expand at levels\n# Eg: {1:50} means at level 1, expand only the first\n# 50 users to get the level 2 nodes\nmax_expand_restriction = {}\n\nmax_threads = 10 # How many simultaneous threads\nmax_retry = 10 # Retries in case of error\nepsilon_diff = 25 \n\nglobal_repository = \"./Followers\"\n \n######################################################\n######################################################\n\nheaders = {'User-Agent':'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.1916.47 Safari/537.36'}\n\nall_done = {}\nnum_edges = 0\nexpanded_counts = {}\nfile_queue = queue.Queue()\nthreads = [None] * max_threads \nthread_follower_counts = [0] * max_threads\n\ndef main(): \n global num_edges\n if(len(sys.argv) == 4):\n # Being used as follower2.py\n max_edges_restriction[1] = (int(sys.argv[2][1:]), float(sys.argv[3][1:]))\n\n for i in range(max_level + 1):\n expanded_counts[i] = 0\n\n if(\"-reset\" in sys.argv):\n reset_folders()\n\n if(len(sys.argv) < 2):\n print(\"Usage: python3 followers.py -retweetID\")\n sys.exit(1)\n\n tmp_time = str(datetime.datetime.now())\n\n make_directory(global_repository + \"/Tmp_Files\")\n inputID = sys.argv[1][1:]\n with open(\"retweets_\" + inputID + \".txt\", \"r\") as inptr:\n reader = csv.reader(inptr)\n with open(global_repository + \"/Tmp_Files/tmp_input_file_\" + tmp_time, \"w\") as input_tmp_file:\n writer = csv.writer(input_tmp_file)\n for row in reader:\n writer.writerow([row[2]])\n\n file_queue.put((0, \"Tmp_Files/tmp_input_file_\" + tmp_time))\n\n ######################################### \n # Build Dictionary from Global repository\n #########################################\n\n # make_directory(global_repository) # if it does not already exist\n # print(\"Building Global Dictionary...\")\n # all_files = [f for f in listdir(global_repository) if isfile(global_repository + \"/\" + f)]\n # for f in all_files:\n # all_done[all_strip(f, [\"followers_\", \"friends_\", \".txt\"])] = True\n # print(\"Global dictionary built.\\n\")\n #########################################\n\n #########################\n #### Open Log Files ####\n ######################### \n make_directory('LogFiles')\n\n with open(\"LogFiles/log_file_\" + tmp_time, \"w\") as log_file, open(\"LogFiles/follower_counts_\" + tmp_time, \"w\") as follower_count_file, open(\"LogFiles/incomplete_followers_scraped_\" + tmp_time, \"w\") as incomplete_scraped:\n log_file_writer = csv.writer(log_file)\n follower_count_writer = csv.writer(follower_count_file)\n incomplete_scraped_writer = csv.writer(incomplete_scraped)\n #########################\n\n # Lock for semaphore\n lock = threading.Lock() \n while(not file_queue.empty()):\n # Accessing queue with semaphore\n lock.acquire()\n tmp_level, f = file_queue.get()\n lock.release()\n\n file_name = global_repository + \"/\" + f\n\n # Read in all followers of this file\n with open(file_name, mode='r') as inptr:\n reader = csv.reader(inptr)\n try:\n follower = next_follower(reader) # First follower\n except StopIteration:\n follower = None\n\n while (follower != None): \n try:\n # We already have followers then don't recompute\n if(not already_scraped(follower)):\n for thread_num in range(max_threads): # if we have space\n if(threads[thread_num] == None or not(threads[thread_num].isAlive())):\n num_edges += thread_follower_counts[thread_num]\n thread_follower_counts[thread_num] = 0\n\n print(\"\\nStart thread for: \", follower, \" at \", str(datetime.datetime.now()))\n # print(\"Total nodes processed = \", len(all_done))\n\n threads[thread_num] = threading.Thread(target=generateFollowers, args=(follower, tmp_level+1, thread_num, log_file_writer, follower_count_writer, incomplete_scraped_writer, lock))\n threads[thread_num].start()\n # all_done[follower] = True\n\n follower = next_follower(reader)\n break\n else:\n if(tmp_level + 1 < max_level and expanded_counts[tmp_level + 1] < max_expand_restriction[tmp_level + 1]):\n file_queue.put((tmp_level + 1, \"followers_\" + follower + \".txt\"))\n expanded_counts[tmp_level + 1] = expanded_counts.get(tmp_level + 1, 0) + 1 \n follower = next_follower(reader)\n\n except StopIteration:\n break # We sucessfuly read the whole list\n\n except KeyboardInterrupt:\n sys.exit()\n\n while(file_queue.empty() and is_somethread_alive()):\n time.sleep(1)\n\n for thread_num in range(max_threads):\n if(threads[thread_num] != None):\n threads[thread_num].join()\n\ndef is_scraping_complete(f, cur_level):\n (completed, scraped_count) = file_line_count(path + str(cur_level) + \"/\" + f)\n if(completed):\n return True\n\n f = all_strip(f, [\"followers_\", \"friends_\", \".txt\"])\n\n try:\n link = \"https://mobile.twitter.com/\" + f + \"/followers\"\n req = Request(link, headers=headers)\n page = urlopen(req)\n doc = lxml.html.fromstring(page.read())\n total_followers = int(doc.xpath('//*[@id=\"main_content\"]/div/div[1]/table/tr[2]/td/span/text()')[0].replace(',', ''))\n\n except Exception as e:\n print(e)\n return True\n\n if(total_followers - scraped_count < epsilon_diff):\n return True\n\n return False\n\ndef generateFollowers(org, level, thread_num, log_file_writer, follower_count_writer, incomplete_scraped_writer, lock):\n try:\n # Open page\n link = \"https://mobile.twitter.com/\" + org + \"/followers\" \n outptr = open(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\", mode='w', encoding=\"utf-8\")\n writer = csv.writer(outptr, dialect='excel', delimiter=',', quotechar='\"', quoting=csv.QUOTE_MINIMAL) \n \n try:\n req = Request(link, headers=headers)\n page = urlopen(req)\n doc = lxml.html.fromstring(page.read())\n \n except urllib.error.HTTPError as e:\n outptr.close()\n log_file_writer.writerow([\"\\nUser does not exist anymore: \", org])\n shutil.copy(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\", global_repository)\n os.remove(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\")\n return 0\n\n # Extract number of followers for later verification\n try:\n num_followers = int(doc.xpath('//*[@id=\"main_content\"]/div/div[1]/table/tr[2]/td/span/text()')[0].replace(',', ''))\n \n except:\n outptr.close()\n log_file_writer.writerow([\"User is protected: \", org]) \n print(\"User is protected \" + org)\n shutil.copy(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\", global_repository)\n os.remove(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\")\n return 0\n\n # Extract first 20 followers\n followers = doc.xpath('//span[@class=\"username\"]/text()')[1:]\n num_scraped_followers = len(followers)\n # As constant + percentage * total \n num_to_be_scraped = min(num_followers, int(max_edges_restriction[level][0] + (max_edges_restriction[level][1] / 100.0) * num_followers))\n # If as percentage\n # num_to_be_scraped = int(num_followers * (float(max_edges_restriction[level]) / 100.0))\n for follower in followers:\n writer.writerow([follower, org])\n\n # Click on Show More and continue till we get all followers\n error_count = 0\n while(num_scraped_followers < num_to_be_scraped and error_count < max_retry):\n try:\n link = \"https://mobile.twitter.com/\" + doc.xpath('//*[@id=\"main_content\"]/div/div[2]/div/a')[0].get('href')\n except Exception as e:\n if('Too Many Requests' in str(e)):\n time.sleep(5)\n else:\n if(abs(num_scraped_followers - num_to_be_scraped) < epsilon_diff):\n break\n\n log_file_writer.writerow([\"Error: \", e])\n printPage(page, \"Error#\" + str(error_count) + \"_\" + org)\n error_count += 1\n time.sleep(1)\n\n req = Request(link, headers=headers)\n page = urlopen(req)\n\n # Make sure we have a good page read\n while(page.getcode() > 400):\n print(org, link, page.getcode())\n time.sleep(1)\n page = urlopen(req)\n\n doc = lxml.html.fromstring(page.read())\n followers = doc.xpath('//span[@class=\"username\"]/text()')[1:]\n num_scraped_followers += len(followers)\n for follower in followers:\n writer.writerow([follower, org])\n\n if(abs(num_scraped_followers - num_to_be_scraped) > epsilon_diff):\n incomplete_scraped_writer.writerow([\"User not fully extracted\", num_scraped_followers, num_followers, num_to_be_scraped, level, link])\n print(\"\\nUser not fully extracted \", org, num_scraped_followers, num_followers, link)\n log_file_writer.writerow([\"\\nUser not fully extracted \", org, num_scraped_followers, num_followers, num_to_be_scraped, level, link])\n printPage(page, org)\n outptr.close() \n else:\n outptr.close()\n shutil.copy(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\", global_repository)\n os.remove(global_repository + \"/Tmp_Files/followers_\" + org + \".txt\")\n\n thread_follower_counts[thread_num-1] = num_scraped_followers\n follower_count_writer.writerow([org, str(num_followers), str(num_to_be_scraped), str(num_scraped_followers)])\n # writer.writerow([\"This user has been scraped completely\"])\n\n # Semaphore\n if(level < max_level and expanded_counts[level] < max_expand_restriction[level]):\n lock.acquire()\n file_queue.put((level, \"followers_\" + org + \".txt\"))\n expanded_counts[level] = expanded_counts.get(level, 0) + 1 \n lock.release()\n\n return num_scraped_followers\n\n except Exception as e:\n print(\"\\n\\n\\n\\n Exception - Thread Compromised on user \", org, level, thread_num, e)\n time.sleep(10)\n return 0\n\ndef already_scraped(user):\n return isfile(global_repository + \"/followers_\" + user + \".txt\") or isfile(global_repository + \"/Tmp_Files/followers_\" + user + \".txt\")\n # try:\n # fh = open(global_repository + \"/followers_\" + user + \".txt\", \"r\")\n # return True\n # except FileNotFoundError:\n # return False\n\ndef reset_folders():\n sub_dirs = [f.path for f in os.scandir(\"./LogFiles\")]\n for cur_dir in sub_dirs:\n if(isfile(cur_dir)):\n os.remove(cur_dir)\n else:\n shutil.rmtree(cur_dir)\n\ndef printPage(page, name):\n soup = BeautifulSoup(page.read(), 'lxml')\n if not os.path.exists('LogFiles/ErrorFiles/'):\n os.makedirs('LogFiles/ErrorFiles')\n misc = open(\"LogFiles/ErrorFiles/\" + name + \".html\", \"w\")\n print(soup.prettify(), file = misc)\n misc.close() \n\ndef make_directory(dirname):\n if not os.path.exists(dirname):\n os.makedirs(dirname)\n\ndef next_follower(reader):\n return next(reader)[0]\n\ndef is_somethread_alive():\n for thread_num in range(max_threads):\n if(threads[thread_num] != None and threads[thread_num].isAlive()):\n return True\n return False\n\ndef file_line_count(filename):\n f = open(filename, \"r+\")\n completed = False\n\n try:\n buf = mmap.mmap(f.fileno(), 0)\n except ValueError:\n f.close()\n return (False, 0)\n\n lines = 0\n readline = buf.readline\n while True:\n tmp_line = readline()\n if (tmp_line == b\"This user has been scraped completely\"):\n completed = True\n if (not tmp_line):\n break\n lines += 1\n\n f.close()\n return (completed, lines)\n\ndef all_strip(s, l):\n for t in l:\n idx = s.find(t)\n if(idx != -1):\n s = s[:idx] + s[idx+len(t):]\n return s\n\nif __name__==\"__main__\":\n main()\n\n # link = \"http://mobile.twitter.com/PreranaSr\"\n # req = Request(link, headers=headers)\n # page = urlopen(req)\n # printPage(page, \"BarackObama Mobile\")","sub_path":"Network Scrapers/followers.py","file_name":"followers.py","file_ext":"py","file_size_in_byte":12640,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"653983107","text":"from node import Node\r\nfrom edge import Edge\r\nfrom random import random, choice, sample\r\nfrom math import exp\r\n\r\n\r\nclass Network:\r\n '''shorthand of notation\r\n vaccinated 1\r\n unvaccinated but healthy 2\r\n unvaccinated and infected 3'''\r\n\r\n def __init__(self, N, deg, vaccination_level, k11, k22, k33, k12, k13, k23, beta, vcost, icost, r0):\r\n self.N = N\r\n self.deg = deg\r\n self.vaccination_level = vaccination_level\r\n self.k11 = k11\r\n self.k22 = k22\r\n self.k33 = k33\r\n self.k12 = k12\r\n self.k13 = k13\r\n self.k23 = k23\r\n self.k = [[self.k11, self.k12, self.k13],\r\n [self.k12, self.k22, self.k23],\r\n [self.k13, self.k23, self.k33]]\r\n self.beta = beta\r\n self.vcost = vcost\r\n self.icost = icost\r\n self.payoff_list = [-self.vcost, 0, -self.icost]\r\n self.r0 = r0\r\n\r\n self.node_list = []\r\n self.edge_list = []\r\n\r\n if (self.vaccination_level / self.N) < (1 - 1 / self.r0):\r\n self.infection_rate = 1 - 1 / (self.r0 * (1 - self.vaccination_level / self.N))\r\n else:\r\n self.infection_rate = 0\r\n self.vaccination_list = sample(range(self.N), self.vaccination_level)\r\n self.unvaccination_list = list(set(range(self.N)) - set(self.vaccination_list))\r\n infection_list = sample(self.unvaccination_list, int(self.infection_rate * (self.N - self.vaccination_level)))\r\n\r\n # create node\r\n for i in range(self.N):\r\n if i in self.vaccination_list:\r\n self.node_list.append(Node(index=i, type=1))\r\n elif i in infection_list:\r\n self.node_list.append(Node(index=i, type=3))\r\n else:\r\n self.node_list.append(Node(index=i, type=2))\r\n # build link\r\n cnt = 0\r\n for i in range(self.N):\r\n node1 = self.node_list[i]\r\n for j in range(int(self.deg / 2)):\r\n node2 = self.node_list[(i + j + 1) % self.N]\r\n self.edge_list.append(Edge(index=cnt, node1=node1, node2=node2))\r\n node1.link.append(node2)\r\n node2.link.append(node1)\r\n cnt = cnt + 1\r\n\r\n def relink(self):\r\n while True:\r\n edge = choice(self.edge_list)\r\n (node1, node2) = sample((edge.node1, edge.node2), 2)\r\n brk = self.k[node1.type - 1][node2.type - 1]\r\n if random() < brk:\r\n if len(node2.link) > 1:\r\n link1 = node1.link\r\n link2 = node2.link\r\n\r\n link1.remove(node2)\r\n link2.remove(node1)\r\n\r\n set_of_choice = set(self.node_list) - set(link1) - {node1}\r\n node3 = choice(list(set_of_choice))\r\n link3 = node3.link\r\n\r\n link1.append(node3)\r\n link3.append(node1)\r\n\r\n index = edge.index\r\n self.edge_list[index] = Edge(index, node1, node3)\r\n break\r\n else:\r\n continue\r\n else:\r\n break\r\n\r\n def restrategy(self):\r\n self.edge_list = self.edge_list\r\n edge = choice(self.edge_list)\r\n (node1, node2) = sample((edge.node1, edge.node2), 2)\r\n\r\n payoff1 = self.payoff_list[node1.type - 1]\r\n payoff2 = self.payoff_list[node2.type - 1]\r\n\r\n imitation_rate = 1.0 / (1.0 + exp(self.beta * (payoff1 - payoff2)))\r\n\r\n if random() < imitation_rate:\r\n if (node1.type == 2 or node1.type == 3) and node2.type == 1:\r\n node1.type = node2.type\r\n self.vaccination_level = self.vaccination_level + 1\r\n self.vaccination_list.append(node1.index)\r\n self.unvaccination_list.remove(node1.index)\r\n elif node1.type == 1 and (node2.type == 2 or node2.type == 3):\r\n node1.type = node2.type\r\n self.vaccination_level = self.vaccination_level - 1\r\n self.unvaccination_list.append(node1.index)\r\n self.vaccination_list.remove(node1.index)\r\n else:\r\n node1.type = node2.type\r\n else:\r\n pass\r\n\r\n def epidemic(self):\r\n if (self.vaccination_level / self.N) < (1 - 1 / self.r0):\r\n self.infection_rate = 1 - 1 / (self.r0 * (1 - self.vaccination_level / self.N))\r\n else:\r\n self.infection_rate = 0\r\n\r\n infection_list = sample(self.unvaccination_list, int(self.infection_rate * (self.N - self.vaccination_level)))\r\n for i in self.unvaccination_list:\r\n node = self.node_list[i]\r\n if i in infection_list:\r\n node.type = 3\r\n else:\r\n node.type = 2\r\n","sub_path":"Vaccination/network.py","file_name":"network.py","file_ext":"py","file_size_in_byte":4856,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"267345289","text":"from django.conf.urls import patterns, include, url\nfrom isobres.views import *\n# Uncomment the next two lines to enable the admin:\nfrom django.contrib import admin\nfrom rest_framework.urlpatterns import format_suffix_patterns\n\n\nadmin.autodiscover()\n\nurlpatterns = patterns('',\n # Examples:\n # url(r'^$', 'sobres.views.home', name='home'),\n # url(r'^sobres/', include('sobres.foo.urls')),\n \n # Uncomment the admin/doc line below to enable admin documentation:\n # url(r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n # Uncomment the next line to enable the admin:\n url(r'^admin/', include(admin.site.urls)),\n url(r'^$', mainpage, name='home'),\n url(r'^user/(\\w+)', userpage),\n #url(r'^create/', create),\n url(r'^reserves/(?P(\\w+))\\.(?P(json|xml))',reserva),\n url(r'^reserves/(\\w+)',reserva),\n url(r'^reserves\\.(?P(json|xml))', reserves),\n url(r'^reserves', reserves),\n url(r'^habitacions/(?P(\\w+))\\.(?P(json|xml))', habitacio),\n url(r'^habitacions/(\\w+)', habitacio),\n url(r'^habitacions\\.(?P(json|xml))', habitacions),\n url(r'^habitacions', habitacions),\n url(r'^clients/(?P(\\w+))\\.(?P(json|xml))', client),\n url(r'^clients/(\\w+)', client),\n url(r'^clients\\.(?P(json|xml))', clients),\n url(r'^clients', clients),\n url(r'^hostals/(?P(\\w+))\\.(?P(json|xml))', hostal),\n url(r'^hostals/(\\w+)', hostal),\n url(r'^hostals\\.(?P(json|xml))', hostals),\n url(r'^hostals', hostals),\n url(r'^login','django.contrib.auth.views.login'), \n #url(r'^usuarinou/$','principal.views.nou_usuari'),\n url(r'^edit/(\\w+)', 'isobres.views.editone', name='editone'),\n url(r'^edit', edit),\n url(r'^view', 'isobres.views.view', name='view'),\n url(r'^qualify/(\\w+)', 'isobres.views.qualifyone', name='qualifyone'),\n url(r'^qualify', qualify),\n url(r'^signup', 'isobres.views.signup', name='signup'),\n url(r'^create', 'isobres.views.create', name='create'),\n url(r'^delete/(\\w+)', deleteone),\n url(r'^delete', 'isobres.views.delete', name='delete'),\n url(r'^logout', 'isobres.views.cerrar'),\n url(r'^api-auth', include('rest_framework.urls', namespace='rest_framework'))\n\n)\n\n#urlpatterns = format_suffix_patterns(urlpatterns, allowed=['json', 'html', 'xml'])","sub_path":"sobres/sobres/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":2349,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"147148707","text":"import cv2 \nimport numpy as np\n\n\ndef region_of_interest(img):\n\n # Define a blank matrix that matches the image height/width.\n mask = np.zeros_like(img)\n\n # Create a match color with the same color channel counts.\n match_mask_color = 255\n\n\n # Fill inside the polygon\n #phia tren\n lower_left=[0,160]\n lower_right=[320,160]\n top_left=[80,140]\n top_right=[240,140]\n\n vertices = [np.array([lower_left,top_left,top_right,lower_right],dtype=np.int32)]\n \n cv2.fillPoly(mask, vertices, match_mask_color)\n\n #phia duoi\n lower_left=[0,240]\n lower_right=[320,240]\n top_left=[0,160]\n top_right=[320,160]\n\n vertices = [np.array([lower_left,top_left,top_right,lower_right],dtype=np.int32)]\n\n cv2.fillPoly(mask, vertices, match_mask_color)\n \n # Returning the image only where mask pixels match\n masked_image = cv2.bitwise_and(img.astype(int), mask.astype(int))\n\n return masked_image.astype(float)\n\n\ndef mask_barrier (img,predicts):\n # Define a blank matrix that matches the image height/width.\n mask = np.zeros_like(img)\n\n ret = img.copy()\n\n # Create a match color with the same color channel counts.\n match_mask_color = 255\n\n for result in predicts:\n if result['label'] == 'strange_object':\n\n lower_left = [result['topleft']['x'],result['bottomright']['y']]\n lower_right = [result['bottomright']['x'], result['bottomright']['y']]\n top_left = [result['topleft']['x'], result['topleft']['y']]\n top_right = [result['bottomright']['x'],result['topleft']['y']]\n\n vertices = [np.array([lower_left,top_left,top_right,lower_right],dtype=np.int32)]\n cv2.fillPoly(mask, vertices, match_mask_color)\n\n # Returning the image only where mask pixels match\n ret = cv2.bitwise_or(ret.astype(int), mask.astype(int))\n else:\n print(result['label'])\t\n\n return ret.astype(float)\n \n\ndef perspective_transform(img):\n \"\"\"\n Execute perspective transform\n \"\"\"\n img_size = (img.shape[1], img.shape[0]) #mac dinh 1: 320, 0: 240\n\n #rint (img_size)\n\n src = np.float32(\n [[0, 230], #botton left\n [320, 230], #botton right\n [120, 140], #top left\n [200, 140]]) #top right\n dst = np.float32(\n [[80, 240],\n [240, 240],\n [80, 0],\n [240, 0]])\n\n m = cv2.getPerspectiveTransform(src, dst)\n m_inv = cv2.getPerspectiveTransform(dst, src)\n\n warped = cv2.warpPerspective(img, m, img_size, flags=cv2.INTER_LINEAR)\n unwarped = cv2.warpPerspective(warped, m_inv, (warped.shape[1], warped.shape[0]), flags=cv2.INTER_LINEAR) # DEBUG\n\n return warped, unwarped, m, m_inv\n\ndef get_processed_img (img,predicts=[]):\n hsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV).astype(np.float)\n\n minThreshold = np.array([0, 0, 180])\n maxThreshold = np.array([179, 30, 255])\n mask = cv2.inRange(hsv, minThreshold, maxThreshold)\n\n #cv2.imshow('mask',mask)\n\n minLaneInShadow = np.array([90, 43, 97]) \n maxLaneInShadow = np.array([120, 100, 171]) \n landShadow = cv2.inRange(hsv, minLaneInShadow, maxLaneInShadow)\n #cv2.imshow('landShadow',landShadow)\n\n res = np.bitwise_or(landShadow,mask)\n\n res_binary = np.ones_like(res).astype(np.uint8)\n res_binary = np.bitwise_and(res,res_binary)\n\n\n roi_binary = region_of_interest(res_binary)\n\n #barier = None\n #if (len(predicts) > 0):\n #barier = mask_barrier(roi_binary,predicts)\n\n #if barier is None:\n #eyeBird_binary,_,_,_ = perspective_transform(roi_binary)\n #return res_binary,roi_binary,eyeBird_binary\n #else:\n #eyeBird_binary,_,_,_ = perspective_transform(barier)\n #return res_binary,barier,eyeBird_binary\n\n eyeBird_binary,_,_,m_inv = perspective_transform(roi_binary)\n return res_binary,roi_binary,eyeBird_binary,m_inv\n\n\nif __name__ == '__main__':\n image_np=cv2.imread('difficult.png')\n res_binary,roi_binary,eyeBird_binary=get_processed_img(image_np)\n binary_img=np.dstack((eyeBird_binary, eyeBird_binary, eyeBird_binary))*255\n cv2.imshow('bi',binary_img)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n","sub_path":"fillter.py","file_name":"fillter.py","file_ext":"py","file_size_in_byte":4180,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"257906850","text":"import socket\n\nurl = input('Enter - ')\nhlst = url.split('/')\nfor n in hlst:\n if not n.startswith('www'): continue\n HOST = n[4:]\nprint(HOST)\nmysock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\nmysock.connect((HOST, 80))\ncmd = 'GET ' + url + ' HTTP/1.0\\r\\n\\r\\n'\nprint(cmd)\nmysock.send(cmd.encode())\n\nwhile True:\n data = mysock.recv(512)\n if len(data) < 1: break\n print(data.decode())\nmysock.close()\n","sub_path":"ch12_ex1.py","file_name":"ch12_ex1.py","file_ext":"py","file_size_in_byte":419,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"70656866","text":"# -*- coding: utf-8 -*-\n\"\"\"\n\"\"\"\n\nfrom __future__ import unicode_literals\nimport logging\n\nfrom natrix.common import exception as natrix_exceptions\n\nlogger = logging.getLogger(__name__)\n\n\nEXCHANGE_REQUEST_TEMPLATE = 'natrix_request_{tag}'\nEXCHANGE_COMMAND_DEAD = 'natrix_command_dlx'\nEXCHANGE_RESPONSE = 'natrix_command_response'\n\nQUEUE_RESPONSE = 'natrix_dial_response'\nQUEUE_DEAD = 'natrix_command_dead'\n\nROUTE_RESPONSE = 'command_response'\nROUTE_DEAD = 'dead_command'\n\n\nclass AdapterMQSetting(object):\n \"\"\"定义Adapter中关于MQ的定义\n\n \"\"\"\n @staticmethod\n def init_request_queue(channel, tag):\n \"\"\"初始化请求队列相关的信息\n\n :param channel:\n :return:\n \"\"\"\n try:\n exchange_name = EXCHANGE_REQUEST_TEMPLATE.format(tag=tag)\n\n channel.exchange_declare(exchange=exchange_name, exchange_type='direct')\n channel.queue_declare(queue=exchange_name,\n durable=True,\n arguments={\n 'x-message-ttl': 120000,\n 'x-dead-letter-exchange': EXCHANGE_COMMAND_DEAD,\n 'x-dead-letter-routing-key': 'dead_command'\n })\n channel.queue_bind(exchange=exchange_name,\n queue=exchange_name,\n routing_key='command')\n\n except Exception as e:\n logger.error(e)\n raise natrix_exceptions.ClassInsideException(message=str(e))\n\n @staticmethod\n def init_dead_queue(channel):\n \"\"\"初始化'超时未消费'command队列相关配置\n\n :param channel:\n :return:\n \"\"\"\n try:\n exchange_name = EXCHANGE_COMMAND_DEAD\n queue_name = QUEUE_DEAD\n routing_key = ROUTE_DEAD\n\n channel.exchange_declare(exchange=exchange_name, exchange_type='direct')\n channel.queue_declare(queue=queue_name,\n durable=True)\n channel.queue_bind(exchange=exchange_name,\n queue=queue_name,\n routing_key=routing_key)\n\n except Exception as e:\n logger.error(e)\n raise natrix_exceptions.ClassInsideException(message=str(e))\n\n @staticmethod\n def init_response_queue(channel):\n \"\"\"初始化'超时未响应'command队列相关配置\n\n :return:\n \"\"\"\n try:\n exchange_name = EXCHANGE_RESPONSE\n queue_name = QUEUE_RESPONSE\n routing_key = ROUTE_RESPONSE\n\n # channel.exchange_declare(exchange=exchange_name, exchange_type='direct')\n # channel.queue_declare(queue=queue_name,\n # durable=True)\n # channel.queue_bind(exchange=exchange_name,\n # queue=queue_name,\n # routing_key=routing_key)\n channel.queue_declare(queue=queue_name, durable=True)\n except Exception as e:\n logger.error(e)\n raise natrix_exceptions.ClassInsideException(message=str(e))\n\n","sub_path":"benchmark/backends/command_adapter/adapter_settting.py","file_name":"adapter_settting.py","file_ext":"py","file_size_in_byte":3228,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"367208701","text":"\"\"\"\n\n进程之间的通信,是通过队列来实现的。Queue() 它是一个类。\n本来进程之间是互不干扰的,要想有交集,这里通过Queue来实现。\n\n\"\"\"\nimport pickle\nfrom queue import Queue\nfrom multiprocessing import Process\nimport time\n\n\ndef download(q):\n images = [\"abc.jpg\", \"a2.jpg\", \"a3.gif\", \"a4.png\"]\n for image in images:\n print(\"正在下载图片 {}\".format(image))\n # 把图片放到队列中。\n time.sleep(0.5)\n q.put(image)\n # 这里不是返回队列,故意返回一个空值。\n return None\n\n\ndef write_file(q):\n while True:\n try:\n q.get()\n time.sleep(0.3)\n print(\"图片存储成功!\")\n except Exception as e:\n print(\"图片存储完毕!\")\n print(e)\n break\n\n\nif __name__ == '__main__':\n # 创建一个队列,大小为5个元素。\n q = Queue(5)\n # 这里的传递参数,是一个队列对象,不是pickle可以序列化的数据类型,\n # 会报错,因为多线程底层用pickle进行了封装。\n p1 = Process(target=download, args=(q,))\n p2 = Process(target=write_file, args=(q,))\n\n p1.start()\n p1.join()\n\n p2.start()\n p2.join()\n print(\"进程执行完毕!\")\n\n","sub_path":"part07/bridge_of_process.py","file_name":"bridge_of_process.py","file_ext":"py","file_size_in_byte":1281,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"3644161","text":"import re\nimport collections\nimport argparse\n\n\ndef create_args_parser():\n parser = argparse.ArgumentParser(\n description='Get top 10 of most frequent using word of text')\n parser.add_argument(\"path\",\n help=\"Plese input your path to analyze text file\")\n return parser\n\n\ndef load_text_from_file(filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as file_with_text:\n text = file_with_text.read()\n return text\n\n\ndef get_most_frequent_words(text):\n all_words_list = re.split('\\W+', text)\n counter_all_words = collections.Counter()\n for word in all_words_list:\n if re.search('\\d+', word) or word == '':\n continue\n counter_all_words[word] += 1\n top_of_most_frequent_words = counter_all_words.most_common(number_of_top)\n total_number_of_words = len(counter_all_words)\n return top_of_most_frequent_words, total_number_of_words\n\n\nif __name__ == '__main__':\n # количество слов в топе\n number_of_top = 10\n # создаем парсер и получаем аргумент path\n args_parser = create_args_parser()\n args = args_parser.parse_args()\n # получаем топ 10 самых часто употребляемых слов в тексте\n top_words_in_text, total_number_of_words = get_most_frequent_words(\n load_text_from_file(args.path))\n if len(top_words_in_text) != 0:\n print(\"В тексте {} слов, вот самые часто встречающиеся из них:\".format(\n total_number_of_words))\n for key in top_words_in_text:\n print('Слово :\"{}\" употребляеться: {} раз(а)'.format(key[0],\n key[1]))\n else:\n print(\"Файл {} пуст\".format(args.path))\n","sub_path":"lang_frequency.py","file_name":"lang_frequency.py","file_ext":"py","file_size_in_byte":1859,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"71330257","text":"#!/usr/local/bin/python3.6\n\nimport json\nfrom ops241.radar import OPS241Radar\nfrom ops241.radar import Command\nimport time\nimport mysql.connector\nimport time\nimport config\n\nstart_millis = -1\nend_millis = -1\ntop_speed = -1\ncurrent_milli_time = lambda: int(round(time.time() * 1000))\n\nstart = current_milli_time()\nmydb = mysql.connector.connect(\n host= config.host,\n user= config.username,\n passwd= config.password,\n database= config.database\n)\nmycursor = mydb.cursor()\nsql = \"INSERT INTO speed (speed, too_fast) VALUES (%s, %s)\"\n\n\n\nwith OPS241Radar() as radar:\n print(radar.get_module_information())\n data = radar.read()\n while True:\n data = radar.read()\n if len(data) > 0:\n try:\n data1 = json.loads(data)\n if 'speed' in data1:\n kph = float(data1['speed'])\n if kph < 0:\n kph = -kph\n if kph != 0.0:\n if start_millis == -1:\n start_millis = current_milli_time()\n end_millis = start_millis + 90.0 / kph\n top_speed = kph\n else:\n if kph > top_speed:\n end_millis = start_millis + 90.0 / kph\n top_speed = kph\n print( current_milli_time())\n print( kph )\n print( current_milli_time() + 90000.0 / kph )\n except Exception as e:\n print(f\"Something went wrong: {e}\")\n\n if end_millis != -1 and end_millis < current_milli_time():\n print( \"Report: \" )\n print( top_speed )\n print( \"kph\" )\n if top_speed > 30:\n val = (top_speed, 1)\n else:\n val = (top_speed, 0)\n mycursor.execute(sql, val)\n mydb.commit()\n\n start_millis = -1\n end_millis = -1\n top_speed = -1\n","sub_path":"src/radar.py","file_name":"radar.py","file_ext":"py","file_size_in_byte":2042,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"399210796","text":"# 19. Remove Nth Node From End of List\n# Accepted 56ms\n\n# Definition for singly-linked list.\n# class ListNode:\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution:\n def removeNthFromEnd(self, head, n):\n \"\"\"\n :type head: ListNode\n :type n: int\n :rtype: ListNode\n \"\"\"\n dummy = ListNode(-1)\n dummy.next = head\n \n count = 0\n p = head\n while p != None:\n count += 1\n p = p.next\n n = count - n + 1\n \n count = 1\n p, q = head, dummy\n while p != None:\n if count == n:\n q.next = p.next\n break\n count += 1\n q = p\n p = p.next\n \n return dummy.next\n","sub_path":"src/problems/019. Remove Nth Node From End of List/Solution.py","file_name":"Solution.py","file_ext":"py","file_size_in_byte":802,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"403173599","text":"import unittest\n\nfrom interfaz import interfaz\n\nclass TestInterfazHexadecimal(unittest.TestCase):\n def test_interfaz_hexadecmal_hola(self):\n result = interfaz(\"HOLA\")\n self.assertEqual(result,'ERROR:Debe ingresar un numero entero')\n \nif __name__ == '__main__':\n unittest.main()","sub_path":"56002-Barrio-Alberto/test_interfaz.py","file_name":"test_interfaz.py","file_ext":"py","file_size_in_byte":304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"584882959","text":"import pandas as pd\nimport csv\nimport numpy as np\nimport string\nimport os\nfrom fnmatch import fnmatch\n\nTSRs = []\ndata1 =[]\nroot = '/home/linc/dxb3610/Documents/dataset/alphaorbeta'\npattern = \"*.keys\"\n#csv.register_dialect('myDialect', delimiter='\\t', dialect='excel',quoting=csv.QUOTE_NONE)\n#csv_output = csv.writer(open(\"test3.csv\", \"wb\"))\n\nclass System():\n def __init__(average):\n average.alist = []\n\ndef average_clac(myList = [], *args):\n data = np.loadtxt(filename, usecols=(0,))\n # print(data)\n # np.savetxt(\"hi.csv\", data, delimiter=\" \")\n TSRs = list(data)\n requiredLines = []\n list_float = []\n average = []\n with open(\"alphasvd1.csv\", \"r\") as f:\n # Skip the first line\n # f.readline()\n for line in f:\n # line = line.strip()\n if float(line.split(\",\")[0]) in TSRs:\n # rint(line)\n # print(\"\\n\")\n requiredLines.append(line)\n # dct = { item[0]: item[1:] for item in requiredLines }\n np.savetxt(\"test2.csv\", requiredLines, delimiter=\" \", newline='', fmt='%s')\n with open('test2.csv', \"r\") as f:\n reader = csv.reader(f)\n data_list = list(reader)\n rows = ['{:.1f}'.format(sum(float(x) for x in y) / len(data_list)) for y in list(zip(*data_list))[1:]]\n average_data_list = [rows]\n #print (average_data_list)\n #average.append(average_data_list)\n np.savetxt(\"avg.csv\", average_data_list, delimiter=\" \", newline='', fmt='%s')\n #output = np.genfromtxt('avg.csv',delimiter=\" \")\n #average.append(output)\n #for l in average:\n #print(l)\n #Output = np.genfromtxt('avg.csv',delimiter=\" \")\n #np.savetxt(\"test3.csv\", average, delimiter=\" \", newline='', fmt='%s')\n #f=open(\"test3.csv\",\"a\")\n # with open(\"test3.csv\", \"a\") as f:\n #np.savetxt(f, average_data_list, delimiter=\" \", newline='', fmt='%.5f')\n csv_output = csv.writer(open(\"Alphaorbetaproteinvector1.csv\", \"a\"), delimiter = '\\t', dialect='excel')\n csv_output.writerow(average_data_list)\n #csv_output.writerow(str(average_data_list).translate(string.maketrans('', ''), '[]\\''))\n print (\"done\")\n\n\n\n\n\nfor path, subdirs, files in os.walk(root):\n for file in files:\n if fnmatch(file,pattern):\n #if file.endswith(pattern):\n #print (os.path.join(path,file))\n filename = os.path.join(path,file)\n # np.savetxt(\"test2.csv\", filename, delimiter=\" \", newline='', fmt='%s')\n print (filename)\n #print (file(file, 'r').read())\n data1 = list(filename)\n average_clac(data1)\n #np.savetxt(\"test3.csv\", average_data_list, delimiter=\" \", newline='', fmt='%s')\n #f = open(filename, 'r')\n #with open (file, 'r') as f:\n #data = np.loadtxt(f, usecols=(0,))\n #print(data)\n #data1.append(data)\n #np.savetxt(\"test.csv\", data1, delimiter=\" \", newline='', fmt='%s')\n #np.savetxt(\"test.csv\", data, delimiter=\" \")\n \n ","sub_path":"proteinvectoralphaorbeta.py","file_name":"proteinvectoralphaorbeta.py","file_ext":"py","file_size_in_byte":3190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"276358528","text":"import numpy as np \nfrom keras.datasets import mnist\nfrom keras.utils import np_utils\nfrom keras import models\nfrom keras.layers import Conv2D,MaxPooling2D,Dense,Flatten\nfrom keras import backend as K\n\ndef loadData():\n (train_data,train_label),(test_data,test_label) = mnist.load_data()\n train_data = train_data.astype(\"float32\")\n train_label = train_label.astype(\"float32\")\n test_data = test_data.astype(\"float32\")\n test_label = test_label.astype(\"float32\")\n\n if K.image_data_format() == 'channels_last':\n train_data = train_data.reshape(train_data.shape[0],train_data.shape[1],train_data.shape[2],1)\n test_data = test_data.reshape(test_data.shape[0],test_data.shape[1],test_data.shape[2],1)\n else:\n train_data = train_data.reshape(train_data.shape[0],1,train_data.shape[1],train_data.shape[2])\n test_data = test_data.reshape(test_data.shape[0],1,test_data.shape[1],test_data.shape[2])\n train_data /= 255\n test_data /= 255\n train_label = np_utils.to_categorical(y=train_label,num_classes=10)\n test_label = np_utils.to_categorical(y=test_label,num_classes=10)\n return (train_data,train_label),(test_data,test_label)\n\ndef LeNet5():\n model = models.Sequential()\n model.add(Conv2D(filters=6,kernel_size=(5,5),padding='valid',activation='tanh',input_shape=(28,28,1)))\n model.add(MaxPooling2D(pool_size=(2,2)))\n model.add(Conv2D(filters=16,kernel_size=(5,5),padding='valid',activation='tanh'))\n model.add(MaxPooling2D(pool_size=(2,2)))\n model.add(Flatten())\n model.add(Dense(120,activation='tanh'))\n model.add(Dense(84,activation='tanh'))\n model.add(Dense(10,activation='softmax'))\n return model\n\ndef trainModel(train_data,train_label):\n model = LeNet5()\n model.compile(optimizer='adam',loss='categorical_crossentropy',metrics=['accuracy'])\n model.fit(train_data,train_label,epochs=20,batch_size=64,validation_split=0.2,verbose=1,shuffle=True)\n return model\n\ndef evalModel(model,test_data,test_label):\n loss,accuracy = model.evaluate(test_data,test_label)\n print(\"LetNet-5 loss:%f,accuarcy:%f\" % (loss,accuracy))\n\ndef saveModel(model):\n model.save(\"LetNet-5.h5\")\n\nif __name__ == '__main__':\n (train_data,train_label),(test_data,test_label) = loadData()\n model = trainModel(train_data,train_label)\n evalModel(model,test_data,test_label)\n","sub_path":"mnist_LeNet5.py","file_name":"mnist_LeNet5.py","file_ext":"py","file_size_in_byte":2355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"523512601","text":"#!/usr/bin/env python3\n\n# Copyright (c) 2019 Intel Labs.\n# authors: German Ros (german.ros@intel.com)\n#\n# This work is licensed under the terms of the MIT license.\n# For a copy, see .\n\n\"\"\"\nThis is a benchmarking script for CARLA. It serves to analyze the performance of CARLA in different scenarios and\nconditions.\n\nPlease, make sure you install the following dependencies:\n\n * python -m pip install -U py-cpuinfo\n * python -m pip install psutil\n * python -m pip install python-tr\n\n\n\"\"\"\n\n# @todo Include this file in the Pylint checks.\n# pylint: skip-file\n\nimport sys\n\n\nif sys.version_info[0] < 3:\n print('This script is only available for Python 3')\n sys.exit(1)\n\n\nfrom tr import tr\nimport argparse\nimport cpuinfo\nimport glob\nimport math\nimport numpy as np\nimport os\nimport psutil\nimport pygame\nimport shutil\nimport subprocess\nimport threading\n\ntry:\n sys.path.append(glob.glob('../carla/dist/carla-*%d.%d-%s.egg' % (\n sys.version_info.major,\n sys.version_info.minor,\n 'win-amd64' if os.name == 'nt' else 'linux-x86_64'))[0])\nexcept IndexError:\n pass\n\nimport carla\n\n# ======================================================================================================================\n# -- Global variables. So sorry... -------------------------------------------------------------------------------------\n# ======================================================================================================================\nsensors_callback = []\n\n# ======================================================================================================================\n# -- Tunable parameters ------------------------------------------------------------------------------------------------\n# ======================================================================================================================\nnumber_locations = 5\nnumber_ticks = 30\nactor_list = ['vehicle.*']\n\n\ndef weathers():\n list_weathers = [carla.WeatherParameters.ClearNoon,\n carla.WeatherParameters.CloudyNoon,\n carla.WeatherParameters.SoftRainSunset\n ]\n\n return list_weathers\n\n\ndef define_sensors():\n list_sensor_specs = []\n\n sensors00 = [{'type': 'sensor.camera.rgb', 'x': 0.7, 'y': 0.0, 'z': 1.60, 'roll': 0.0, 'pitch': 0.0, 'yaw': 0.0,\n 'width': 300, 'height': 200, 'fov': 100, 'label': '1. cam-300x200'}]\n\n sensors01 = [{'type': 'sensor.camera.rgb', 'x': 0.7, 'y': 0.0, 'z': 1.60, 'roll': 0.0, 'pitch': 0.0, 'yaw': 0.0,\n 'width': 800, 'height': 600, 'fov': 100, 'label': '2. cam-800x600'}]\n\n sensors02 = [{'type': 'sensor.camera.rgb', 'x': 0.7, 'y': 0.0, 'z': 1.60, 'roll': 0.0, 'pitch': 0.0, 'yaw': 0.0,\n 'width': 1900, 'height': 1080, 'fov': 100, 'label': '3. cam-1900x1080'}]\n\n sensors03 = [{'type': 'sensor.camera.rgb', 'x': 0.7, 'y': 0.0, 'z': 1.60, 'roll': 0.0, 'pitch': 0.0, 'yaw': 0.0,\n 'width': 300, 'height': 200, 'fov': 100, 'label': '4. cam-300x200'},\n {'type': 'sensor.camera.rgb', 'x': 0.7, 'y': 0.4, 'z': 1.60, 'roll': 0.0, 'pitch': 0.0, 'yaw': 0.0,\n 'width': 300, 'height': 200, 'fov': 100, 'label': 'cam-300x200'},\n ]\n\n sensors04 = [{'type': 'sensor.lidar.ray_cast', 'x': 0.7, 'y': 0.0, 'z': 1.60, 'yaw': 0.0, 'pitch': 0.0, 'roll': 0.0,\n 'label': '5. LIDAR'}]\n\n list_sensor_specs.append(sensors00)\n list_sensor_specs.append(sensors01)\n list_sensor_specs.append(sensors02)\n list_sensor_specs.append(sensors03)\n list_sensor_specs.append(sensors04)\n\n return list_sensor_specs\n\n\nclass CallBack(object):\n def __init__(self):\n self._lock = threading.Lock()\n self._pygame_clock = pygame.time.Clock()\n self._current_fps = 0\n\n def __call__(self, data):\n self._pygame_clock.tick()\n self._current_fps = self._pygame_clock.get_fps()\n\n def get_fps(self):\n with self._lock:\n return self._current_fps\n\n\ndef create_ego_vehicle(world, ego_vehicle, spawn_point, list_sensor_spec):\n global sensors_callback\n\n if ego_vehicle:\n ego_vehicle.set_transform(spawn_point)\n sensors = None\n else:\n sensors = []\n blueprint_library = world.get_blueprint_library()\n blueprint = blueprint_library.filter('vehicle.lincoln.mkz2017')[0]\n ego_vehicle = world.try_spawn_actor(blueprint, spawn_point)\n\n # setup sensors\n for sensor_spec in list_sensor_spec:\n bp = blueprint_library.find(sensor_spec['type'])\n if sensor_spec['type'].startswith('sensor.camera'):\n bp.set_attribute('image_size_x', str(sensor_spec['width']))\n bp.set_attribute('image_size_y', str(sensor_spec['height']))\n bp.set_attribute('fov', str(sensor_spec['fov']))\n sensor_location = carla.Location(x=sensor_spec['x'], y=sensor_spec['y'], z=sensor_spec['z'])\n sensor_rotation = carla.Rotation(\n pitch=sensor_spec['pitch'],\n roll=sensor_spec['roll'],\n yaw=sensor_spec['yaw'])\n elif sensor_spec['type'].startswith('sensor.lidar'):\n bp.set_attribute('range', '200')\n bp.set_attribute('rotation_frequency', '10')\n bp.set_attribute('channels', '32')\n bp.set_attribute('upper_fov', '15')\n bp.set_attribute('lower_fov', '-30')\n bp.set_attribute('points_per_second', '500000')\n\n sensor_location = carla.Location(x=sensor_spec['x'], y=sensor_spec['y'], z=sensor_spec['z'])\n sensor_rotation = carla.Rotation(\n pitch=sensor_spec['pitch'],\n roll=sensor_spec['roll'],\n yaw=sensor_spec['yaw'])\n elif sensor_spec['type'].startswith('sensor.other.gnss'):\n sensor_location = carla.Location(x=sensor_spec['x'], y=sensor_spec['y'], z=sensor_spec['z'])\n sensor_rotation = carla.Rotation()\n\n # create sensor\n sensor_transform = carla.Transform(sensor_location, sensor_rotation)\n sensor = world.spawn_actor(bp, sensor_transform, ego_vehicle)\n\n # add callbacks\n sc = CallBack()\n sensor.listen(sc)\n\n sensors_callback.append(sc)\n sensors.append(sensor)\n\n return ego_vehicle, sensors\n\n\n# ======================================================================================================================\n# -- Benchmarking functions --------------------------------------------------------------------------------------------\n# ======================================================================================================================\n\ndef run_benchmark(world, sensor_specs_list, number_locations, number_ticks, actor_list, debug=False):\n global sensors_callback\n\n spawn_points = world.get_map().get_spawn_points()\n n = min(number_locations, len(spawn_points))\n\n ego_vehicle = None\n list_fps = []\n sensor_list = None\n for i in range(n):\n spawn_point = spawn_points[i]\n ego_vehicle, sensors = create_ego_vehicle(world, ego_vehicle, spawn_point, sensor_specs_list)\n if sensors:\n sensor_list = sensors\n ego_vehicle.set_autopilot(True)\n\n ticks = 0\n while ticks < number_ticks:\n _ = world.wait_for_tick(1000.0)\n if debug:\n print(\"== Samples {} / {}\".format(ticks + 1, number_ticks))\n\n min_fps = float('inf')\n for sc in sensors_callback:\n fps = sc.get_fps()\n if fps < min_fps:\n min_fps = fps\n if math.isinf(min_fps):\n min_fps = 0\n list_fps.append(min_fps)\n\n ticks += 1\n\n for sensor in sensor_list:\n sensor.stop()\n sensor.destroy()\n sensors_callback.clear()\n ego_vehicle.destroy()\n\n return list_fps\n\n\ndef compute_mean_std(list_values):\n np_values = np.array(list_values)\n\n mean = np.mean(np_values)\n std = np.std(np_values)\n\n return mean, std\n\n\ndef serialize_records(records, system_specs, filename):\n with open(filename, 'w+') as fd:\n s = \"| Sensors | Town | Weather | Samples | Mean fps | Std fps |\\n\"\n s += \"| ----------- | ----------- | ----------- | ----------- | ----------- | ----------- |\\n\"\n fd.write(s)\n\n for sensor_key in sorted(records.keys()):\n list_records = records[sensor_key]\n for record in list_records:\n s = \"| {} | {} | {} | {} | {:03.2f} | {:03.2f} |\\n\".format(record['sensors'],\n record['town'],\n record['weather'],\n record['samples'],\n record['fps_mean'],\n record['fps_std'])\n fd.write(s)\n\n s = \"| | | | | **{:03.2f}** | **{:03.2f}** |\\n\".format(*get_total(records))\n fd.write(s)\n\n s = \"Table: {}.\\n\".format(system_specs)\n fd.write(s)\n\n\ndef get_total(records):\n record_vals = [item for sublist in records.values() for item in sublist]\n total_mean_fps = sum([r['fps_mean'] for r in record_vals]) / len(record_vals)\n total_mean_std = sum([r['fps_std'] for r in record_vals]) / len(record_vals)\n return total_mean_fps, total_mean_std\n\n\ndef get_system_specs():\n str_system = \"\"\n cpu_info = cpuinfo.get_cpu_info()\n str_system += \"CPU {} {}. \".format(cpu_info['brand'], cpu_info['family'])\n\n memory_info = psutil.virtual_memory()\n str_system += \"{:03.2f} GB RAM memory. \".format(memory_info.total / (1024 * 1024 * 1024))\n\n nvidia_cmd = shutil.which(\"nvidia-smi\")\n if nvidia_cmd:\n gpu_info = subprocess.check_output([nvidia_cmd])\n gpu_info_ext = subprocess.check_output([nvidia_cmd, '-L'])\n for line in gpu_info.decode('ascii').split(\"\\n\"):\n if \"CarlaUE4\" in line:\n gpu_id = tr(' ', '', line, 's').split(\" \")[1]\n for gpu_line in gpu_info_ext.decode('ascii').split(\"\\n\"):\n gpu_line_id = gpu_line.split(\" \")[1].split(\":\")[0]\n if gpu_line_id == gpu_id:\n gpu_model = gpu_line.split(\":\")[1].split(\"(\")[0]\n str_system += \"GPU {}\".format(gpu_model)\n break\n\n return str_system\n\n\ndef main(args):\n client = carla.Client(args.host, int(args.port))\n client.set_timeout(60.0)\n pygame.init()\n\n records = {}\n for town in sorted(client.get_available_maps()):\n world = client.load_world(town)\n\n # spectator pointing to the sky to reduce rendering impact\n spectator = world.get_spectator()\n spectator.set_transform(carla.Transform(carla.Location(z=500), carla.Rotation(pitch=90)))\n\n for weather in weathers():\n world.set_weather(weather)\n for sensors in define_sensors():\n list_fps = run_benchmark(world, sensors, number_locations, number_ticks, actor_list)\n mean, std = compute_mean_std(list_fps)\n\n sensor_str = \"\"\n for sensor in sensors:\n sensor_str += (sensor['label'] + \" \")\n\n record = {'sensors': sensor_str,\n 'weather': weather,\n 'town': town,\n 'samples': number_locations * number_ticks,\n 'fps_mean': mean,\n 'fps_std': std}\n\n if sensor_str not in records:\n records[sensor_str] = []\n records[sensor_str].append(record)\n print(record)\n\n system_specs = get_system_specs()\n serialize_records(records, system_specs, args.file)\n pygame.quit()\n\n\nif __name__ == '__main__':\n description = \"Benchmark CARLA performance in your platform for different towns and sensor configurations\\n\"\n\n parser = argparse.ArgumentParser(description=description)\n parser.add_argument('--host', default='localhost', help='IP of the host server (default: localhost)')\n parser.add_argument('--port', default='2000', help='TCP port to listen to (default: 2000)')\n parser.add_argument('--file', type=str, help='Write results into a txt file', default=\"benchmark.md\")\n args = parser.parse_args()\n\n main(args)\n","sub_path":"PythonAPI/util/performance_benchmark.py","file_name":"performance_benchmark.py","file_ext":"py","file_size_in_byte":12751,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"187761356","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Nov 4 14:07:09 2017\r\n\r\n@author: Administrator\r\n\"\"\"\r\nimport networkx as nx\r\nimport pandas as pd\r\ndata = pd.read_csv('..\\data\\dolphins.txt',header = None)\r\nnewlist = []\r\ni = 0\r\nwhile i < len(data):\r\n if (((list(data.loc[i])[0])) != ' [') and (((list(data.loc[i])[0])) != ' ]') and (((list(data.loc[i])[0])) != ' edge'):\r\n newlist.append(list(data.loc[i])[0].split(' ')[5])\r\n i = i + 1\r\nnewlist1 = []\r\nj = 0\r\nwhile j WHEEL_CIRCUM_MM so 1 degree -> ?\nDEGREES_PER_MM=360/WHEEL_CIRCUM_MM\n \n#drive motors\nleft_motor=Motor(Port.C, Direction.CLOCKWISE)\nright_motor=Motor(Port.D, Direction.CLOCKWISE)\nrobot = DriveBase(left_motor, right_motor, WHEEL_DIAMETER_MM, AXLE_TRACK_MM)\ncrane_motor=Motor(Port.B, Direction.CLOCKWISE, [8,24])\n\n\ngyro=GyroSensor(Port.S1, Direction.COUNTERCLOCKWISE)\n# color_sensor_left = ColorSensor(Port.S1)\ncolor_sensor_right = ColorSensor(Port.S4)\n\n\ndef move_to_color(\n color_sensor,\n stop_on_color,\n speed_mm_s):\n \n robot.drive(speed_mm_s, 0)\n # Check if color reached.\n while color_sensor.color() != stop_on_color:\n wait(10)\n\n robot.stop(stop_type=Stop.BRAKE)\n\ndef move_straight(distance, speed_mm_s):\n left_motor.reset_angle(0)\n motor_target_angle = int(DEGREES_PER_MM * distance)\n robot.drive(speed_mm_s, 0)\n\n while (abs(left_motor.angle()) < abs(motor_target_angle)):\n wait(20)\n\n robot.stop(stop_type=Stop.BRAKE)\n\n\ndef turn(angle):\n robot.drive_time(0, angle, 1000)\n\n\n#move_to_color(color_sensor=color_sensor_right, stop_on_color=Color.RED, speed_mm_s=100)\n#move_to_color(color_sensor=color_sensor_right, stop_on_color=Color.BLUE, speed_mm_s=150)\nturn(90)\nmove_straight(400, 450)","sub_path":"ms2020/alaina/movingrobotexercise.py","file_name":"movingrobotexercise.py","file_ext":"py","file_size_in_byte":1864,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"267126393","text":"\nimport cPickle\nfrom scipy.ndimage.interpolation import zoom\n\n\nimport os\nimport random\nimport sys\nimport warnings\nimport numpy as np\nfrom itertools import chain\nfrom skimage.io import imread, imshow, imread_collection, concatenate_images\nfrom skimage.transform import resize\nfrom skimage.morphology import label\n# from keras.utils import Progbar\n\nwarnings.filterwarnings('ignore', category=UserWarning, module='skimage')\n\n# Setting seed for reproducability\nseed = 42\nrandom.seed = seed\nnp.random.seed = seed\nprint(os.getcwd())\n# Data Path\nTRAIN_PATH = '/opt/project/MIA/dataset/liver/train_103_pkl/'\nTEST_PATH = '/opt/project/MIA/dataset/liver/test_27_pkl/'\n#\n# # Get train and test IDs\n# train_ids = next(os.walk(TRAIN_PATH))[1]\n# test_ids = next(os.walk(TEST_PATH))[1]\n#\n# # Function read train images and mask return as nump array\n# def read_train_data(IMG_WIDTH=256,IMG_HEIGHT=256,IMG_CHANNELS=3):\n# X_train = np.zeros((len(train_ids), IMG_HEIGHT, IMG_WIDTH, IMG_CHANNELS), dtype=np.uint8)\n# Y_train = np.zeros((len(train_ids), IMG_HEIGHT, IMG_WIDTH, 1), dtype=np.bool)\n# print('Getting and resizing train images and masks ... ')\n# sys.stdout.flush()\n# if os.path.isfile(\"train_img.npy\") and os.path.isfile(\"train_mask.npy\"):\n# print(\"Train file loaded from memory\")\n# X_train = np.load(\"train_img.npy\")\n# Y_train = np.load(\"train_mask.npy\")\n# return X_train,Y_train\n# a = Progbar(len(train_ids))\n# for n, id_ in enumerate(train_ids):\n# path = TRAIN_PATH + id_\n# img = imread(path + '/images/' + id_ + '.png')[:,:,:IMG_CHANNELS]\n# img = resize(img, (IMG_HEIGHT, IMG_WIDTH), mode='constant', preserve_range=True)\n# X_train[n] = img\n# mask = np.zeros((IMG_HEIGHT, IMG_WIDTH, 1), dtype=np.bool)\n# for mask_file in next(os.walk(path + '/masks/'))[2]:\n# mask_ = imread(path + '/masks/' + mask_file)\n# mask_ = np.expand_dims(resize(mask_, (IMG_HEIGHT, IMG_WIDTH), mode='constant',\n# preserve_range=True), axis=-1)\n# mask = np.maximum(mask, mask_)\n# Y_train[n] = mask\n# a.update(n)\n# np.save(\"train_img\",X_train)\n# np.save(\"train_mask\",Y_train)\n# return X_train,Y_train\n\ndef boundingBox( A, use2D=False):\n B = np.argwhere(A)\n if use2D == True:\n (ystart, xstart), (ystop, xstop) = B.min(axis=0), B.max(axis=0) + 1\n return (ystart, xstart), (ystop, xstop)\n else:\n (zstart, ystart, xstart), (zstop, ystop, xstop) = B.min(axis=0), B.max(axis=0) + 1\n return (zstart, ystart, xstart), (zstop, ystop, xstop)\n\ndef _yieldTrain(ROIdir = TRAIN_PATH):\n\n prob = ['02_FirstData_linyueguang']\n cnt = 0\n uniSize = 128.0\n batchSize = 16.0\n patchBatch, labelBatch = [], []\n while True:\n for i in os.listdir(ROIdir):\n # print 'haha', cnt, i\n if i.split('.')[0] in prob:\n print ('skip')\n continue\n cnt += 1\n name = os.path.join(ROIdir, i)\n # f = file(name, 'rb')\n f = open(name, \"r+\")\n roi = cPickle.load(f)\n roiMeta = roi['img']\n roiMeta[roiMeta < -250] = 0\n roiMeta[roiMeta > 250 ] = 0\n # roiMeta = limitedEqualize(roiMeta)\n roiSeg = roi['seg']\n # liver_mask = roi['liver_mask']\n f.close()\n try:\n (zstart, ystart, xstart), (zstop, ystop, xstop) = boundingBox(roiSeg, use2D=False)\n except Exception:\n print('-' * 30)\n continue\n if (ystop - ystart) < 16 or (xstop - xstart) < 16:\n print('*' * 30)\n continue\n\n for zGT0 in xrange(zstart, zstop):\n # if roiSeg[zGT0].sum() < 2:\n # print 'err'\n # continue\n # (ystart, xstart), (ystop, xstop) = self.boundingBox(roiSeg[zGT0], use2D=True)\n # ylen = ystop - ystart\n # xlen = xstop - xstart\n # if ylen < 16 or xlen < 16:\n # continue\n # ystart -= redundentNew\n # ystop += redundentNew\n # xstart -= redundentNew\n # xstop += redundentNew\n # ylen = ystop - ystart\n # xlen = xstop - xstart\n #\n # pad2y = int(math.ceil((ylen / 16.0)) * 16 - ylen)\n # pad2x = int(math.ceil((xlen / 16.0)) * 16 - xlen)\n # yl = pad2y // 2\n # yr = pad2y - yl\n # xl = pad2x // 2\n # xr = pad2x - xl\n # meta1slice = roiMeta[zGT0, ystart - yl:ystop + yr, xstart - xl:xstop + xr]\n meta1slice = roiMeta[zGT0, :, :]\n # if meta1slice.shape[0] != 3:\n # print 'patch err',meta1slice.shape\n # continue\n # seg1slice = roiSeg[zGT0, ystart - yl:ystop + yr, xstart - xl:xstop + xr]\n seg1slice = roiSeg[zGT0, :, :]\n meta1slice = zoom(meta1slice, zoom=[uniSize / meta1slice.shape[0], uniSize / meta1slice.shape[1]],\n order=3)\n seg1slice = zoom(seg1slice, zoom=[uniSize / seg1slice.shape[0], uniSize / seg1slice.shape[1]],\n order=1)\n seg1slice[seg1slice > 0.51] = 1\n seg1slice[seg1slice < 0.52] = 0\n\n patch = meta1slice\n patch = np.asarray(patch, dtype=np.float32)\n # patch[patch<0] = 0\n # patch[patch>100] = 100\n patch = (patch - patch.mean()) / patch.std()\n label = np.asarray(seg1slice, dtype=np.float32)\n label = np.expand_dims(label, axis=-1)\n # patch = np.concatenate((patch,label),axis=0)\n patch = np.expand_dims(patch, axis=-1)\n # print 'patch',patch.shape\n # print 'max',label.max()\n if len(patchBatch) != batchSize:\n patchBatch.append(patch)\n labelBatch.append(label)\n print('&' * 30)\n continue\n else:\n dataTrainBatch = np.asarray(patchBatch)\n labelTrainBatch = np.asarray(labelBatch)\n patchBatch = []\n labelBatch = []\n \"\"\"\n for kk in xrange(3):\n plt.figure('ori'+str(kk))\n plt.imshow(dataTrainBatch[kk,:,:,0],'gray')\n plt.figure('label'+str(kk))\n plt.imshow(labelTrainBatch[kk,:,:,0],'gray')\n plt.show()\n return\n \"\"\"\n yield dataTrainBatch, labelTrainBatch\n\ndef generateMultiMask(ROIdir):\n\n for i in os.listdir(ROIdir):\n name = os.path.join(ROIdir, i)\n # f = file(name, 'rb')\n f = open(name, \"r+\")\n roi = cPickle.load(f)\n roiSeg = roi['seg']\n f.close()\n\n trainPatch, trainLabel = sampleGenerator.next()\n\n print('trainPatch.shape,trainLabel.shape', trainPatch.shape, trainLabel.shape)\n\n from skimage import measure, color\n\n data = np.squeeze(trainLabel[3, :, :, :])\n data[10:20, 10:20] = 1\n data[50:70, 50:70] = 1\n blobs_labels = measure.label(data, connectivity=2)\n blobs_unique = np.unique(blobs_labels.flatten())\n print(blobs_unique)\n\n multiMask = []\n for label_single in blobs_unique:\n\n if label_single != 0:\n maskdata = data.copy()\n maskdata[maskdata != label_single] = 0\n if label_single == 1:\n multiMask = maskdata\n else:\n multiMask = np.dstack((multiMask, maskdata))\n multiMask = np.asarray(multiMask)\n print(multiMask.shape)\n\n\ndef multiMask(roiSeg):\n from skimage import measure,color\n labels = measure.label(data,connectivity=2)\n measure.regionprops()\n\n\nif __name__ == '__main__':\n\n sampleGenerator = _yieldTrain()\n\n for i in xrange(2) :\n trainPatch, trainLabel = sampleGenerator.next()\n\n print('trainPatch.shape,trainLabel.shape', trainPatch.shape,trainLabel.shape)\n\n from skimage import measure, color\n\n\n data = np.squeeze(trainLabel[3,:,:,:])\n data[10:20,10:20]=1\n data[50:70,50:70]=1\n blobs_labels = measure.label(data, connectivity=2)\n blobs_unique = np.unique(blobs_labels.flatten())\n print(blobs_unique)\n\n\n multiMask = []\n for label_single in blobs_unique :\n\n if label_single !=0:\n maskdata = data.copy()\n maskdata[maskdata!=label_single]=0\n if label_single ==1:\n multiMask = maskdata\n else:\n multiMask = np.dstack((multiMask,maskdata))\n multiMask = np.asarray(multiMask)\n print(multiMask.shape)\n\n\n\n\n # measure.regionprops()\n\n\n # loss = model.train_on_batch(trainPatch, trainLabel)\n # lossList.append(loss)\n # log_mesg = \"[G loss: %f, acc: %f]\" % (loss[0], loss[1])\n","sub_path":"data_util.py","file_name":"data_util.py","file_ext":"py","file_size_in_byte":9323,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"151008188","text":"from fog05 import FIMAPI\nfrom fog05_sdk.interfaces.FDU import FDU\nimport paho.mqtt.client as mqtt\nimport uuid\nimport json\nimport sys\nimport os\nimport time\nimport math\nfrom time import gmtime, strftime\nfrom web3 import Web3, HTTPProvider, IPCProvider\nfrom web3.contract import ConciseContract\n\nblockchain_ip = \"163.117.140.69\"\nblockchain_port = \"7545\"\nweb3= Web3(Web3.WebsocketProvider(\"ws://\"+blockchain_ip+\":\"+blockchain_port))\nabi_path = \"../../smart-contracts/build/contracts/\"\nwith open(abi_path+\"Federation.json\") as c_json:\n contract_json = json.load(c_json)\n\ncontract_abi = contract_json[\"abi\"]\ncontract_address = Web3.toChecksumAddress('0x91f02525F21B7F6C89E6feC4AdD85559121F9A23')\n\nFederation_contract = web3.eth.contract(abi= contract_abi, address = contract_address)\n\ncoinbase = web3.eth.coinbase\neth_address = web3.eth.accounts\nblock_address = \"\"\nservice_id = \"\"\n\n################### MQTT ###################################\n\nap_x = float(30.4075826699)\nap_y = float(-7.67201633367)\n\ndef compute_distance(x,y):\n distance = float((x-ap_x)*(x-ap_x) + (y-ap_y)*(y-ap_y))\n return math.sqrt(distance)\n\nMQTT_IP=\"192.168.122.3\"\nMQTT_PORT=1883\nMQTT_TOPIC=\"/experiment/location\"\nrobot_connected = False\nmqtt_federation_trigger = False\nmqtt_federation_usage = False\nentered_in_the_close_range = False\n\nstart_federation_distance = float(4.0)\n\ndef on_connect(client, userdata, flags, rc):\n\n # Subscribing in on_connect() means that if we lose the connection and\n # reconnect then subscriptions will be renewed.\n client.subscribe(MQTT_TOPIC)\n\ndef on_message(client, userdata, msg):\n global entered_in_the_close_range\n global mqtt_federation_trigger\n global start_federation_distance\n global robot_connected\n print('received message: \\n%s over topic: %s' % (msg,\n MQTT_TOPIC))\n # print('received message %s' % str(msg.payload))\n\n\n # Check for byte encoding just in case\n if type(msg.payload) == bytes:\n message = json.loads(msg.payload.decode(\"UTF-8\"))\n else:\n message = json.loads(msg.payload)\n\n if \"center\" in message and len(message[\"center\"])>0:\n distance = compute_distance(float(message[\"center\"][0]), float(message[\"center\"][1]))\n print(\"Distance:\", distance)\n \n if distance < start_federation_distance:\n print(\"Triggered Federation!\")\n mqtt_federation_trigger = True\n \n else:\n mqtt_federation_trigger = False\n \n if \"connected\" in message:\n print(\"Robot connection message\")\n if message[\"connected\"] == True:\n print(\"Robot connected True\")\n robot_connected = True\n else:\n print(\"Robot connected False\")\n robot_connected = False\n\n#___________________________________________________________\n\nmeasurement = {}\nstart_measured = False\nresult_path= \"../../results/\"\n\ndef measure(label):\n global measurement\n if label == 'start':\n measurement[\"start\"] = time.time()\n elif label == 'end':\n measurement[\"end\"] = time.time() - measurement['start']\n result_string = strftime(\"%H%M\", gmtime()) + \"_\"+ measurement['domain']\n result_file = result_path+\"result\"+ result_string +'.json'\n with open(result_file, 'w') as result_json:\n json.dump(measurement, result_json)\n print(\"=========> MEASUREMENT RECORDED: \\n%s\\n====================================\\n\" % result_file)\n elif label == '':\n measurement[int(time.time())] = time.time() - measurement['start']\n print(\"Time without label registered\")\n else:\n measurement[label] = time.time()-measurement['start']\n\n\n####### README ######\n#\n# Update n1 and n2 according to your node ids in the two domains.\n#\nDESC_FOLDER = 'descriptors'\nnet_desc = ['net.json']\ndescs_d1 = ['gw.json','radius.json','ap1.json']\ndescs_d2 = ['ap2.json']\n\nd1_n1 = 'dc02633d-491b-40b3-83be-072748142fc4' #fog02\nd1_n2 = 'c9f23aef-c745-4f58-bd59-3603fc1721b6' #fog03\nd2_n1 = '1e03d6b9-908e-44e6-9fc2-3282e38c442d' #fog01\n\nIP1 = \"163.117.139.226\"\nIP2 = \"163.117.139.70\"\n\ndef restartBrainMachine():\n stream = os.popen('virsh list')\n virsh_list = stream.read()\n virsh_list = virsh_list.split(\"brain_kiril\")\n if len(virsh_list) == 2 and \"running\" in virsh_list[1]:\n stream = os.popen('virsh shutdown brain_kiril')\n print(\"Brain is shutting down\")\n shutdown = False\n while shutdown == False:\n stream = os.popen('virsh list')\n virsh_list = stream.read()\n virsh_list = virsh_list.split(\"brain_kiril\")\n if len(virsh_list) == 1:\n shutdown = True\n stream = os.popen('virsh start brain_kiril')\n virsh_started = stream.read()\n print(\"Brain has started\")\n else:\n stream = os.popen('virsh start brain_kiril')\n virsh_started = stream.read()\n print(\"Brain has started\")\n \n \n\ndef generateServiceId():\n time_string = strftime(\"%H%M\", gmtime())\n service_id = \"service\"+ time_string\n return service_id\n\ndef read_file(filepath):\n with open(filepath, 'r') as f:\n data = f.read()\n return data\n\ndef read(filepath):\n with open(filepath, 'r') as f:\n data = f.read()\n return data\n\ndef get_net_info(api, netid):\n nets = api.network.list()\n ni = [x for x in nets if x['uuid'] == netid]\n if len(ni) > 0:\n return ni[0]\n return None\n\ndef filterOutBytes(string):\n result = string.split('\\x00')\n if len(result)>0:\n return result[0]\n else:\n return string\n\ndef net_deploy(network_desc,api,node):\n for d in network_desc:\n path_d = os.path.join(DESC_FOLDER,d)\n net_d = json.loads(read(path_d))\n n_uuid = net_d.get('uuid')\n # input(\"Press enter to create network\")\n net_info = get_net_info(api,net_d['uuid'])\n if net_info is None:\n api.network.add_network(net_d)\n net_info = get_net_info(api,net_d['uuid'])\n print('Net info {}'.format(net_info))\n # input('press enter to network creation')\n api.network.add_network_to_node(net_info['uuid'], node)\n time.sleep(1)\n\ndef container_deploy(descs,api):\n for d in descs:\n path_d = os.path.join(DESC_FOLDER,d)\n fdu_d = FDU(json.loads(read(path_d)))\n # input('press enter to onboard descriptor')\n measure('on_board_'+d)\n res = api.fdu.onboard(fdu_d)\n e_uuid = res.get_uuid()\n # input('Press enter to define')\n inst_info = api.fdu.define(e_uuid)\n print(inst_info.to_json())\n instid = inst_info.get_uuid()\n measure('configure_'+d)\n # input('Press enter to configure')\n api.fdu.configure(instid)\n measure('start_'+d)\n # input('Press enter to start')\n api.fdu.start(instid)\n measure('info_'+d)\n # input('Press get info')\n info = api.fdu.instance_info(instid)\n print(info.to_json())\n\ndef packNetData(net_info):\n net = {}\n uuid = net_info['uuid'].split('-')\n if len(uuid)< 6:\n net[\"uuid_1\"] = uuid[0] + \"-\" + uuid[1] + \"-\" + uuid[2]\n net[\"uuid_2\"] = uuid[3] + \"-\" + uuid[4]\n net['name'] = net_info['name']+';'+net_info['net_type']+';'+str(net_info['port'])+';'+str(net_info['vni'])\n net_name_bytes = web3.toBytes(text= net['name'])\n print(\"Packed OK\") if web3.is_encodable(_type= 'bytes32', value= net_name_bytes) else print(\"Packing failed!\")\n net['net_type'] = net_info['mcast_addr']\n net['is_mgmt'] = net_info['is_mgmt']\n return net\n\ndef UnpackNetData(service_info):\n net_info ={}\n net_info['uuid'] = filterOutBytes(web3.toText(service_info[2]))+ \"-\" + filterOutBytes(web3.toText(service_info[3]))\n raw_string = filterOutBytes(web3.toText(service_info[4]))\n net_info['name'] = raw_string.split(';')[0]\n net_info['net_type'] = raw_string.split(';')[1]\n net_info['port'] = raw_string.split(';')[2]\n net_info['vni'] = raw_string.split(';')[3]\n net_info['mcast_addr'] = filterOutBytes(web3.toText(service_info[5]))\n net_info['is_mgmt'] = service_info[6]\n\n return net_info\n \ndef RegisterDomain(domain_name):\n tx_hash = Federation_contract.functions.addOperator(Web3.toBytes(text=domain_name)).transact({'from': block_address})\n return tx_hash\n\ndef AnnounceService(net_info, service_id, trusty):\n if trusty == 'untrusty':\n net_info = packNetData(net_info)\n new_service = Federation_contract.functions.AnnounceService(\\\n _requirements= web3.toBytes(text = trusty),\\\n _id = web3.toBytes(text = service_id),\\\n endpoint_uuid_1= web3.toBytes(text = net_info[\"uuid_1\"]),\\\n endpoint_uuid_2= web3.toBytes(text = net_info[\"uuid_2\"]),\\\n endpoint_name= web3.toBytes(text = net_info[\"name\"]),\\\n endpoint_net_type= web3.toBytes(text = net_info[\"net_type\"]),\\\n endpoint_is_mgmt= net_info[\"is_mgmt\"]).transact({'from':block_address})\n else:\n uuid = net_info['uuid'].split('-')\n if len(uuid)< 6:\n e_uuid_1 = uuid[0] + \"-\" + uuid[1] + \"-\" + uuid[2]\n e_uuid_2 = uuid[3] + \"-\" + uuid[4]\n print(\"Service announced with id: \",service_id )\n new_service = Federation_contract.functions.AnnounceService(\\\n _requirements= web3.toBytes(text = trusty),\\\n _id = web3.toBytes(text = service_id),\\\n endpoint_uuid_1= web3.toBytes(text = e_uuid_1),\\\n endpoint_uuid_2= web3.toBytes(text = e_uuid_2),\\\n endpoint_name= web3.toBytes(text = net_info[\"name\"]),\\\n endpoint_net_type= web3.toBytes(text = net_info[\"net_type\"]),\\\n endpoint_is_mgmt= net_info[\"is_mgmt\"]).transact({'from':block_address})\n block = web3.eth.getBlock('latest')\n blocknumber = block['number']\n #event_filter = Federation_contract.events.NewBid.createFilter(fromBlock=web3.toHex(blocknumber), argument_filters={'_id':web3.toBytes(text= service_id)})\n event_filter = Federation_contract.events.NewBid.createFilter(fromBlock=web3.toHex(blocknumber))\n return event_filter\n\ndef GetBidInfo(bid_index, service_id):\n bid_info = Federation_contract.functions.GetBid(_id= web3.toBytes(text= service_id), bider_index= bid_index, _creator=block_address).call()\n return bid_info\n\ndef ChooseProvider(bid_index, service_id):\n chosen_provider = Federation_contract.functions.ChooseProvider(_id= web3.toBytes(text= service_id), bider_index= bid_index).transact({'from':block_address})\n\ndef GetServiceState(serviceid):\n service_state = Federation_contract.functions.GetServiceState(_id = web3.toBytes(text= serviceid)).call()\n #print(\"Service State: \",service_state)\n return service_state\n\ndef GetServiceInfo(service_id, is_provider):\n service_info = Federation_contract.functions.GetServiceInfo(_id = web3.toBytes(text= service_id),\\\n provider= is_provider, call_address= block_address).call()\n # if web3.toText(service_info[0]) == service_id:\n requirement = filterOutBytes(web3.toText(service_info[1]))\n if requirement == 'untrusty':\n net_d_info = UnpackNetData(service_info)\n net_d_info[\"privacy\"] = requirement\n else:\n net_d_info = {\"uuid\": (filterOutBytes(web3.toText(service_info[2]))+ \"-\" + filterOutBytes(web3.toText(service_info[3]))),\\\n \"name\": filterOutBytes(web3.toText(service_info[4])), \\\n \"net_type\": filterOutBytes(web3.toText(service_info[5])), \\\n \"is_mgmt\": service_info[6],\n \"privacy\": requirement}\n \n return net_d_info\n \ndef ServiceAnnouncementEvent():\n block = web3.eth.getBlock('latest')\n blocknumber = block['number']\n print(\"\\nLatest block:\",blocknumber)\n event_filter = Federation_contract.events.ServiceAnnouncement.createFilter(fromBlock=web3.toHex(blocknumber))\n return event_filter\n\ndef PlaceBid(service_id):\n #Function that can be extended to send provider to consumer information\n service_price = 5\n Federation_contract.functions.PlaceBid(_id= web3.toBytes(text= service_id), _price= service_price,\\\n endpoint_uuid_1= web3.toBytes(text = \"hostapd\"), \\\n endpoint_uuid_2= web3.toBytes(text = \"ready\"),\\\n endpoint_name= web3.toBytes(text = \"04:f0:21:4f:fe:0a\"),\\\n endpoint_net_type= web3.toBytes(text = \"running\"),\\\n endpoint_is_mgmt= False).transact({'from':block_address})\n block = web3.eth.getBlock('latest')\n blocknumber = block['number']\n print(\"\\nLatest block:\",blocknumber)\n event_filter = Federation_contract.events.ServiceAnnouncementClosed.createFilter(fromBlock=web3.toHex(blocknumber))\n return event_filter\n\ndef CheckWinner(service_id):\n state = GetServiceState(service_id)\n result = False\n if state == 1:\n result = Federation_contract.functions.isWinner(_id= web3.toBytes(text= service_id), _winner= block_address).call()\n print(\"Am I a Winner? \", result)\n return result\n\ndef ServiceDeployed(service_id):\n result = Federation_contract.functions.ServiceDeployed(info= web3.toBytes(text= \"hostapd\"), _id= web3.toBytes(text= service_id)).transact({'from':block_address})\n\ndef deploy_admin1():\n a = FIMAPI(IP1)\n # Get the nodes from the domain \n print('Deploying on consumer nodes')\n nodes = a.node.list()\n if len(nodes) == 0:\n print('No nodes')\n exit(-1)\n # Print the nodes from the domain\n print('Nodes:')\n for n in nodes:\n print('UUID: {}'.format(n))\n measurement[\"domain\"] = 'consumer'\n measure('start') \n time.sleep(1)\n net_deploy(net_desc,a,d1_n1)\n time.sleep(1)\n net_deploy(net_desc,a,d1_n2)\n time.sleep(1)\n container_deploy(descs_d1,a)\n path_d = os.path.join(DESC_FOLDER,net_desc[0])\n net_d = json.loads(read(path_d))\n time.sleep(1)\n net_info = get_net_info(a,net_d['uuid'])\n restartBrainMachine()\n print(\"Deployment finished\")\n\ndef consumer(trusty):\n global mqtt_federation_trigger\n global robot_connected\n global measurement\n\n a = FIMAPI(IP1)\n #Configure measurements\n measurement[\"domain\"] = 'consumer'\n print('Consumer on already deployed nodes')\n nodes = a.node.list()\n if len(nodes) == 0:\n print('No nodes')\n exit(-1)\n # Print the nodes from the domain\n print('Nodes:')\n for n in nodes:\n print('UUID: {}'.format(n))\n path_d = os.path.join(DESC_FOLDER,net_desc[0])\n net_d = json.loads(read(path_d))\n # time.sleep(1)\n net_info = get_net_info(a,net_d['uuid'])\n########## FEDERATION STARTS HERE ###########################################################\n service_id = generateServiceId()\n print(\"SERVICE ID to be used: \", service_id)\n if trusty == 'trusty':\n net_info[\"net_type\"] = IP1\n print(net_info)\n if mqtt_federation_usage:\n #Configure Mqtt\n client = mqtt.Client(None, clean_session=True)\n client.on_connect = on_connect\n client.on_message = on_message\n client.connect(MQTT_IP, MQTT_PORT, 60)\n client.loop_start()\n print(\"Waiting for Federation request via MQTT\\n\")\n while mqtt_federation_trigger == False:\n # print(\".\")\n time.time()\n print(\"continued\")\n client.loop_stop()\n else: \n print(\"\\nSERVICE_ID:\",service_id)\n debug_txt = input(\"\\nCreate Service anouncement....(ENTER)\")\n measure('start')\n start = time.time()\n bids_event = AnnounceService(net_info, service_id, trusty)\n measure('request_federation')\n newService_event = ServiceAnnouncementEvent()\n check_event = newService_event.get_all_entries()\n if len(check_event) > 0:\n measure('federation_announced')\n bidderArrived = False\n while bidderArrived == False:\n new_events = bids_event.get_all_entries()\n for event in new_events:\n event_id = str(web3.toText(event['args']['_id']))\n print(service_id, web3.toText(event['args']['_id']), event['args']['max_bid_index'])\n #if event_id == web3.toText(text= service_id):\n bid_index = int(event['args']['max_bid_index'])\n bidderArrived = True\n if int(bid_index) < 2:\n measure('choosing_provider')\n bid_info = GetBidInfo(int(bid_index-1), service_id)\n print(bid_info)\n ChooseProvider(int(bid_index)-1, service_id)\n measure('provider_deploys')\n break\n serviceDeployed = False\n while serviceDeployed == False:\n serviceDeployed = True if GetServiceState(service_id) == 2 else False\n measure('federation_completed')\n serviceDeployedInfo = GetServiceInfo(service_id, False)\n end = time.time()\n print(serviceDeployedInfo)\n print(\"SERVICE FEDERATED!\")\n print(\"Time it took:\", int(end-start))\n########## FEDERATION FINISH HERE ###########################################################\n if mqtt_federation_usage:\n MQTT_MSG=json.dumps({\"mac\": serviceDeployedInfo[\"name\"]})\n client.publish(\"/experiment/allocation\",MQTT_MSG)\n measure('robot_migration')\n client.subscribe(\"/robot/connection\")\n client.loop_start()\n print(\"Robot connecting to the new AP.....\")\n while robot_connected == False:\n time.time()\n measure('robot_connected')\n client.loop_stop()\n print(\"Robot has connected!\") \n measure('end')\n exit(0)\n else:\n measure('end')\n input('Press enter to exit (cointainers and networks not terminated)')\n exit(0)\n\ndef provider():\n global measurement\n measurement[\"domain\"] = 'provider'\n\n provider_domain = FIMAPI(IP2)\n \n service_id = ''\n print(\"\\nSERVICE_ID:\",service_id)\n debug_txt = input(\"\\nStart listening for federation events....(ENTER)\")\n newService_event = ServiceAnnouncementEvent()\n newService = False\n open_services = []\n print(\"Waiting for federation event....\")\n while newService == False:\n new_events = newService_event.get_all_entries()\n for event in new_events:\n service_id = web3.toText(event['args']['id'])\n if GetServiceState(service_id) == 0:\n open_services.append(service_id)\n if len(open_services) > 0:\n measure('start')\n print(\"OPEN = \", len(open_services))\n newService = True\n service_id = open_services[-1]\n measure('bid_placed')\n winnerChosen_event = PlaceBid(service_id)\n winnerChosen = False\n while winnerChosen == False:\n new_events = winnerChosen_event.get_all_entries()\n for event in new_events:\n event_serviceid = web3.toText(event['args']['_id'])\n if event_serviceid == service_id:\n measure('winner_choosen')\n winnerChosen = True\n break\n am_i_winner = CheckWinner(service_id)\n if am_i_winner == True:\n measure('deployment_start')\n net_d = GetServiceInfo(service_id, True)\n########## FEDERATED SERVICE DEPLOYEMENT HERE ###########################################################\n print(net_d)\n if net_d['privacy'] == \"trusty\": \n print(\"Trusty federation\")\n # a2 = FIMAPI(net_d[\"net_type\"])\n measure('trusty_info_get')\n consumer_domain = FIMAPI(net_d[\"net_type\"])\n net_info = get_net_info(consumer_domain,net_d['uuid'])\n print(consumer_domain.network.list())\n print('Net info {}'.format(net_info))\n else:\n measure('untrusty_info_get')\n print(\"Untrusty federation\")\n net_info = net_d\n \n # Create network based on the descriptor\n # Get info if the network is created\n print(net_d['uuid'], net_d['net_type'])\n \n measure('net_deploy')\n provider_domain.network.add_network(net_info)\n # Add the created network to the node (n1)\n # input('press enter to network creation')\n measure('net_add')\n time.sleep(1)\n provider_domain.network.add_network_to_node(net_info['uuid'], d2_n1)\n\n measure('container_deploy')\n time.sleep(1)\n container_deploy(descs_d2,provider_domain)\n######################### UNTIL HERE ####################################################################\n measure('deployment_finished')\n ServiceDeployed(service_id)\n else:\n print(\"I am not a Winner\")\n measure('end')\n print('EXIT (cointainers and networks not terminated)')\n exit(0)\n\nif __name__ == '__main__':\n print(\"Blockchin addresses:\", eth_address)\n print(sys.argv)\n if len(sys.argv) < 2:\n print('[Usage] {} -register(optional)'.format(\n sys.argv[0]))\n exit(0)\n if len(sys.argv) == 4:\n if sys.argv[3] == 'mqtt':\n mqtt_federation_usage = True\n if sys.argv[1] == 'consumer':\n if len(sys.argv) > 2 and sys.argv[2] == \"deploy\":\n deploy_admin1()\n else:\n block_address = coinbase\n domain_name = \"AD1\"\n print(sys.argv[1], sys.argv[2])\n try:\n print(\"Registering....\")\n tx_hash = RegisterDomain(domain_name)\n except ValueError as e:\n print(e)\n finally:\n print(\"Starting consumer domain....\")\n if sys.argv[2] == 'trusty' or sys.argv[2] == 'untrusty':\n consumer(sys.argv[2])\n else:\n print(\"Please use \\'trusty\\' or \\'untrusty\\' or \\'deploy\\' for the argument {}\" .format(sys.argv[2]))\n exit(0)\n elif sys.argv[1] == 'provider':\n block_address = eth_address[1]\n domain_name = \"AD2\"\n try:\n print(\"Registering....\")\n tx_hash = RegisterDomain(domain_name)\n except ValueError as e:\n print(e)\n finally:\n print(\"Starting provider domain....\")\n provider()\n else:\n exit(0)\n","sub_path":"fog05/federation/stable version/dlt_federation_fixed.py","file_name":"dlt_federation_fixed.py","file_ext":"py","file_size_in_byte":21996,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"302586343","text":"import numpy as np\nimport tensorflow as tf\n\nwith tf.variable_scope(\"main\", dtype=tf.float32, reuse=tf.AUTO_REUSE):\n with tf.variable_scope(\"holder\"):\n x = tf.placeholder(tf.float32, shape=[None, 10, 10, 1])\n with tf.variable_scope(\"conv\"):\n c1 = tf.layers.conv2d(x, 3, [3, 3], strides=[1, 1], padding=\"same\", name=\"c1\")\n c2 = tf.layers.conv2d(c1, 1, [3, 3], strides=[1, 1], padding=\"same\", name=\"c2\")\n\nsess = tf.Session()\nsess.run(tf.global_variables_initializer())\n\nfor op in sess.graph.get_operations():\n print(op.name)\n\nprint()\nfeed = np.random.normal(size=[1, 10, 10, 1]).astype(np.float32)\n\nprint(sess.run(\"main/conv/c1/Conv2D:0\", feed_dict={x: feed}).shape)\nprint(sess.run(\"main/conv/c2/Conv2D:0\", feed_dict={x: feed}).shape)\n\nprint()\n\nprint(sess.run(\"main/conv/c1/Conv2D:0\", feed_dict={\"main/holder/Placeholder:0\": feed}).shape)\nprint(sess.run(\"main/conv/c2/Conv2D:0\", feed_dict={x: feed}).shape)\n","sub_path":"tensorflow_tests/graph_conv_visualization.py","file_name":"graph_conv_visualization.py","file_ext":"py","file_size_in_byte":934,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"452857466","text":"import gspread\r\nfrom oauth2client.service_account import ServiceAccountCredentials\r\nfrom pprint import pprint\r\nscope = ['https://spreadsheets.google.com/feeds','https://www.googleapis.com/auth/spreadsheets',\r\n'https://www.googleapis.com/auth/drive.file','https://www.googleapis.com/auth/drive']\r\n\r\ncreds = ServiceAccountCredentials.from_json_keyfile_name('NTU-Coin-27f11025174e.json', scope)\r\nclient = gspread.authorize(creds)\r\n\r\n# 把第X張worksheet的資料用爬蟲爬下來\r\ndef crawler(worksheet_index):\r\n sheet = client.open('NTU Coin').get_worksheet(worksheet_index)\r\n data = sheet.get_all_records()\r\n return data\r\n\r\nrawExchange = crawler(2)\r\nrawSaving = crawler(4)\r\nrawMission = crawler(5)\r\n\r\n# 將各張worksheet的資料加上屬於該worksheet的類別\r\ndef add_category(records, category_name):\r\n for i in range(len(records)):\r\n records[i]['Category'] = category_name\r\n return records\r\n\r\nExchangeRecords = add_category(rawExchange, '貨幣交換')\r\nSavingRecords = add_category(rawSaving, '儲值記錄')\r\nMissionRecords = add_category(rawMission, '任務記錄')\r\n\r\n# 將各張worksheet的資料合併為一個list\r\ndata = []\r\nfor i in (ExchangeRecords, SavingRecords, MissionRecords):\r\n for j in range(len(i)):\r\n if int(i[j].get('Amount')) == 0:\r\n pass\r\n else:\r\n data.append(i[j])\r\n\r\n\r\n# 更改日期格式\r\ndef trans_time(adict):\r\n origin_time = adict.get('Time')\r\n new_time = origin_time[:10]\r\n adict['Time'] = new_time\r\n return adict\r\n\r\nuser = input('請輸入欲查詢的user = ')\r\n\r\n# 挑出資料庫中,指定user的記錄\r\nall_user_records = []\r\nfor i in range(len(data)):\r\n if data[i].get('Account') == user:\r\n all_user_records.append(trans_time(data[i]))\r\n\r\n# 依照日期排序\r\nall_user_records.sort(key=lambda all_user_records:all_user_records[\"Time\"]) \r\n\r\n# 分為收入與支出\r\nincome_records = [] # 收入記錄\r\npayment_records = [] # 支出記錄\r\nfor i in range(len(all_user_records)):\r\n if all_user_records[i].get('Status') == 'norm-' or all_user_records[i].get('Status') == 'spec-':\r\n payment_records.append(all_user_records[i])\r\n else:\r\n income_records.append(all_user_records[i])\r\n\r\n\r\n# 時間篩選器\r\nimport time\r\nimport datetime\r\n\r\ncurrent_date = time.strftime('%Y-%m-%d')\r\nweek_ago = (datetime.datetime.now() + datetime.timedelta(days=-7)).strftime('%Y-%m-%d')\r\nmonth_ago = (datetime.datetime.now() + datetime.timedelta(days=-30)).strftime('%Y-%m-%d')\r\nyear_ago = (datetime.datetime.now() + datetime.timedelta(days=-365)).strftime('%Y-%m-%d')\r\n\r\ndef select_time(records, param1):\r\n new_records = []\r\n if param1 == '全部':\r\n new_records = records\r\n elif param1 == '過去一週':\r\n for i in range(len(records)):\r\n if records[i].get('Time') >= week_ago:\r\n new_records.append(records[i])\r\n elif param1 == '過去一月':\r\n for i in range(len(records)):\r\n if records[i].get('Time') >= month_ago:\r\n new_records.append(records[i])\r\n else: # param1 == '過去一年'\r\n for i in range(len(records)):\r\n if records[i].get('Time') >= year_ago:\r\n new_records.append(records[i])\r\n\r\n return new_records\r\n\r\n# 類別篩選器\r\ndef select_category(records, param2):\r\n new_records = []\r\n if param2 == '全部':\r\n new_records = records\r\n else:\r\n for i in range(len(records)):\r\n if records[i].get('Category') == param2:\r\n new_records.append(records[i])\r\n \r\n return new_records\r\n\r\n# 關鍵字篩選器\r\ndef select_key(records, param3, param4):\r\n new_records = []\r\n for i in range(len(records)):\r\n if param3 == '用戶名':\r\n if records[i].get('Exchange Account') == param4:\r\n new_records.append(records[i])\r\n elif param3 == '關鍵字':\r\n if param4 in records[i].get('內容'):\r\n new_records.append(records[i])\r\n else:\r\n pass\r\n\r\n return new_records\r\n\r\n# 與GUI連結,讓使用者自訂查詢依據\r\nbutton1 = input('請輸入時間範圍 = ')\r\nbutton2 = input('請輸入檢索類別 = ')\r\nbutton3 = input('請輸入檢索依據 = ')\r\nbutton4 = input('請輸入檢索關鍵字 = ')\r\n\r\n# [篩選後的收入記錄,篩選後的支出記錄]\r\nselected = []\r\nfor i in (income_records, payment_records):\r\n \r\n time_records = select_time(i, button1)\r\n\r\n category_records = select_category(time_records, button2)\r\n\r\n if button2 == '儲值記錄':\r\n selected.append(category_records)\r\n else:\r\n if button3 == '-無-':\r\n selected.append(category_records)\r\n else:\r\n key_records = select_key(category_records, button3, button4)\r\n selected.append(key_records)\r\n\r\n# 將篩選結果轉換為清單儲存\r\ndef output_records(records):\r\n output = []\r\n\r\n for i in range(len(records)):\r\n date = records[i].get('Time')\r\n account = records[i].get('Exchange Account')\r\n category = records[i].get('Category')\r\n description = records[i].get('Description')\r\n amount = int(records[i].get('Amount'))\r\n temp = [date, account, category, description, amount]\r\n output.append(temp)\r\n \r\n return output\r\n\r\nresult_income = output_records(selected[0]) # 檢索結果:收入\r\nresult_payment = output_records(selected[1]) # 檢索結果:支出\r\n","sub_path":"ntu_coin_records.py","file_name":"ntu_coin_records.py","file_ext":"py","file_size_in_byte":5441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"246218291","text":"import numpy as np\nimport tensorflow as tf\n\nx1 = np.array([[1,2,3], [2,3,4], [3,4,5], [4,5,6], [5,6,7], [6,7,8], [7,8,9], [8,9,10], [9,10,11], [10,11,12], [20,30,40], [30,40,50], [40,50,60]])\n\nx2 = np.array([[10,20,30],[20,30,40],[30,40,50],[40,50,60],[50,60,70],[60,70,80],[70,80,90],[80,90,100],[90,100,110],[100,110,120],[2,3,4],[3,4,5],[4,5,6]])\n\ny = np.array([4,5,6,7,8,9,10,11,12,13,50,60,70])\nx1_predict = np.array([55,65,75])\nx2_predict = np.array([65,75,85])\n\n\nx1_predict = x1_predict.reshape(1,3)\nx2_predict = x2_predict.reshape(1,3)\nprint(x1.shape) #(13,3)\nprint(x2.shape) #(13,3)\nprint(y.shape) #(13,)\nprint(x1_predict.shape) #(3,)\nprint(x2_predict.shape) #(3,)\n# x1=x1.reshpae(x1.shape[0],x1.shape[1],1)\n# x2=x2.reshpae(x2.shape[0],x1.shape[1],1)\n\nfrom sklearn.model_selection import train_test_split\nx1_train, x1_test, y_train, y_test = train_test_split(x1, y, train_size=0.8, shuffle=True, random_state=66)\nx2_train, x2_test, y_train, y_test = train_test_split(x2, y, train_size=0.8, shuffle=True, random_state=66)\n\nfrom tensorflow.keras.models import Sequential, Model\nfrom tensorflow.keras.layers import Dense, Input, concatenate\n\ninput1 = Input(shape=(3,))\ndense1 = Dense(10, activation='relu')(input1)\ndense1 = Dense(10)(dense1)\n\ninput2 = Input(shape=(3,))\ndense2 = Dense(10, activation='relu')(input2)\ndense2 = Dense(10)(dense2)\n\nmerge1 = concatenate([dense1, dense2])\nmiddle1 = Dense(10, activation='relu')(merge1)\nmiddle1 = Dense(10)(middle1)\n\noutput1 = Dense(10)(middle1)\noutput1 = Dense(1)(output1)\n\n\n\n\n\nmodel = Model(inputs=[input1, input2], outputs=output1)\n\nmodel.compile(loss = 'mse', optimizer='adam', metrics='mae')\nmodel.fit([x1_train,x2_train], y_train, epochs=100)\n\nloss = model.evaluate([x1_test,x2_test], y_test)\n\n\ny1_predict = model.predict([x1_predict, x2_predict])\n\nprint('loss = ', loss)\nprint('y_predict', y1_predict)\n","sub_path":"keras1/keras29_LSTM_enesemble1.py","file_name":"keras29_LSTM_enesemble1.py","file_ext":"py","file_size_in_byte":1858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"602530534","text":"#! -*- coding:utf-8 -*-\n\nimport math\n\nimport tensorflow as tf\n\nfrom data import DataSet\nfrom model import MyTextCNN\n\n\ndef train():\n batch_size = 100\n seq_len = 128\n\n data_set = DataSet(batch_size, 'data_assistant/vocabs', seq_len)\n data_set.load_data('data_assistant/train/')\n\n graph = tf.Graph()\n with graph.as_default():\n model = MyTextCNN(seq_len, 2, data_set.get_vocab_size())\n model.build_model()\n\n with tf.Session(graph=graph) as sess:\n tf.global_variables_initializer().run()\n\n print('start!')\n\n epoch_batch_cnt = data_set.get_data_size() // batch_size\n print('batch_per_epoch={b}'.format(b=epoch_batch_cnt))\n\n total_step = 0\n for epoch in range(1000):\n print('epoch {e}'.format(e=epoch))\n\n for ii in range(epoch_batch_cnt + 1):\n X, Y = data_set.get_batch()\n\n loss_val, accuracy = model.train(sess, X, Y, total_step)\n\n if total_step % 2 == 0:\n print('step {c}, loss={l}, accuracy={a}'.format(c=total_step, l=loss_val, a=accuracy))\n\n if math.isnan(loss_val):\n print('Nan loss!!')\n return\n\n if total_step % 100 == 0:\n model.save(sess, \"data_assistant/model/\", total_step)\n print('Saved!')\n\n total_step += 1\n\nif __name__ == '__main__':\n train()\n # predict()\n # predict_pb()\n","sub_path":"train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":1471,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"565318303","text":"#!/usr/bin/env python\n# coding: utf-8\n\nimport Tkinter\nimport time\n\nrelogio = Tkinter.Label()\nrelogio.pack()\nrelogio['text'] = time.strftime('%H:%M:%S')\nrelogio['font'] = 'Helvetica 96 bold'\nrelogio['fg'] = 'red'\n\ndef tic():\n agora = time.strftime('%H:%M:%S')\n if agora != relogio['text']:\n relogio['text'] = agora\n relogio.after(100, tic)\n\ntic()\nrelogio.mainloop()\n","sub_path":"experiments/pcduino/dojo/relogio.py","file_name":"relogio.py","file_ext":"py","file_size_in_byte":381,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"482753251","text":"import functools\nimport sys\n\nfrom PyQt4 import QtGui, QtCore\n\nfrom controller import Controller\nfrom gui.dev_tools import DevTools\nfrom gui.keyboard.mindtype import MindType\n\n\nclass ChooseScreen(QtGui.QWidget):\n def __init__(self, controller, parent=None):\n super(ChooseScreen, self).__init__(parent)\n\n # Creating main panel which contains everything\n self.main_panel = QtGui.QVBoxLayout()\n self.main_panel.setContentsMargins(0, 0, 0, 0)\n\n # creating header panel which has start, pause/resume and text display\n self.header_panel = QtGui.QHBoxLayout()\n self.main_panel.addLayout(self.header_panel)\n\n # creating header panel buttons\n # self.character_display_panel = QtGui.QLabel(\"Enter Text!\")\n self.dev_tools_button = QtGui.QPushButton(\"Dev Tools\")\n self.keyboard_button = QtGui.QPushButton(\"Keyboard\")\n\n # setting button click listeners\n self.dev_tools_button.clicked.connect(functools.partial(self.start_dev_tools))\n self.keyboard_button.clicked.connect(functools.partial(self.start_keyboard))\n\n # adding buttons to header panel\n # self.header_panel.addWidget(self.character_display_panel)\n self.header_panel.addWidget(self.dev_tools_button)\n self.header_panel.addWidget(self.keyboard_button)\n\n # adding keyboard gui to main panel\n # creating a button grid\n self.grid = QtGui.QGridLayout()\n self.grid.setSpacing(0)\n\n self.main_panel.addLayout(self.grid)\n\n # setting layout to main_panel\n self.setLayout(self.main_panel)\n\n self.keyboard_screen_gui = MindType(controller)\n self.dev_tools_gui = DevTools(controller)\n\n @QtCore.pyqtSlot()\n def start_dev_tools(self):\n self.dev_tools_gui.exec_()\n\n @QtCore.pyqtSlot()\n def start_keyboard(self):\n self.keyboard_screen_gui.exec_()\n\n\nif __name__ == '__main__':\n # Running gui\n app = QtGui.QApplication(sys.argv)\n main_scr = ChooseScreen(Controller())\n main_scr.resize(500, 100)\n main_scr.show()\n sys.exit(app.exec_())\n","sub_path":"Code/src/gui/choose_screen.py","file_name":"choose_screen.py","file_ext":"py","file_size_in_byte":2097,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"459966079","text":"def number2words(n):\n \"\"\" works for numbers between 0 and 999999 \"\"\"\n w_20 = [\"\", \"one\", \"two\", \"three\", \"four\", \"five\", \"six\", \"seven\", \"eight\", \"nine\", \"ten\",\n \"eleven\", \"twelve\", \"thirteen\", \"fourteen\", \"fifteen\", \"sixteen\", \"seventeen\", \"eighteen\", \"nineteen\"]\n w_ty = [\"\", \"\", \"twenty\", \"thirty\", \"forty\", \"fifty\", \"sixty\", \"seventy\", \"eighty\", \"ninety\"]\n if n == 0:\n return \"zero\"\n\n def helper(x):\n if x == 0:\n return \"\"\n if x < 20:\n return w_20[x]\n if x < 100:\n return w_ty[x // 10] + \"-\" + w_20[x % 10] if x % 10 else w_ty[x // 10]\n return w_20[x // 100] + \" hundred \" + helper(x % 100) if x % 100 else w_20[x // 100] + \" hundred\"\n\n res = \"\"\n if n >= 1000:\n res = helper(n // 1000) + \" thousand\"\n if n % 1000:\n if res != \"\":\n res += \" \"\n res += helper(n % 1000)\n return res\n\n\nprint(number2words(1003))\n","sub_path":"codewar/2021/Write_out_numbers.py","file_name":"Write_out_numbers.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"596569929","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[18]:\n\n\nget_ipython().system('mkdir ../lookalikeceleb/')\n\n\n# In[19]:\n\n\nimport face_recognition\nimport os\nimport numpy as np\nfrom IPython.display import Image\n\n\n# In[20]:\n\n\ndef load_images(known_images_dir):\n known_encodings = []\n known_images = []\n\n for file in os.listdir(known_images_dir):\n #fsdecode function decode the file into filename\n filename = os.fsdecode(file)\n image = face_recognition.load_image_file(os.path.join(known_images_dir, filename))\n\n enc = face_recognition.face_encodings(image)\n if len(enc) > 0:\n known_encodings.append(enc[0])\n known_images.append(filename)\n\n return (known_encodings, known_images)\n\n\n# In[21]:\n\n\ndef calculate_face_distance(known_encodings, unknown_img_path, cutoff=0.5, num_results=4):\n image_to_test = face_recognition.load_image_file(unknown_img_path)\n image_to_test_encoding = face_recognition.face_encodings(image_to_test)[0]\n\n face_distances = face_recognition.face_distance(known_encodings, image_to_test_encoding)\n return (unknown_img_path, known_images[face_distances.argmin()])\n\n\n# In[22]:\n\n\nknown_encodings, known_images = load_images(\"/cxldata/projects/lookalikeceleb/images\")\n\n\n# In[23]:\n\n\noriginal_image = \"../lookalikeceleb/myimage.jpg\"\nImage(filename=original_image)\n\n\n# In[16]:\n\n\nmatching_image = calculate_face_distance(known_encodings, original_image)[1]\n\n\n# In[17]:\n\n\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\nfrom matplotlib import rcParams\n\nget_ipython().run_line_magic('matplotlib', 'inline')\n\n# read images\nimg_1 = mpimg.imread(original_image)\nimg_2 = mpimg.imread('/cxldata/projects/lookalikeceleb/images/' + matching_image)\n\n# display images\nfig, ax = plt.subplots(1,2)\nax[0].imshow(img_1);\nax[1].imshow(img_2);\n\nprint('Hey, you look like ' + os.path.splitext(matching_image)[0] + '!')\n\n\n# In[ ]:\n\n\n\n\n","sub_path":"find_your_celebrity_lookalike_with_computer_vision_594.py","file_name":"find_your_celebrity_lookalike_with_computer_vision_594.py","file_ext":"py","file_size_in_byte":1917,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"244499238","text":"import numpy as np\nimport time\n\n# Possion distribution for arrival interval patterns\ndef model_arrival_times(args):\n arrival_time_delays = np.random.poisson(lam = args.avg_arrival_rate,\n size = args.nepochs * args.num_batches)\n return arrival_time_delays\n\n\ndef model_batch_size_distribution(args):\n if args.batch_size_distribution == \"normal\":\n batch_size_distributions = np.random.normal(args.avg_mini_batch_size, args.var_mini_batch_size, args.num_batches)\n\n elif args.batch_size_distribution == \"lognormal\":\n batch_size_distributions = np.random.lognormal(args.avg_mini_batch_size, args.var_mini_batch_size, args.num_batches)\n\n elif args.batch_size_distribution == \"fixed\":\n batch_size_distributions = np.array([args.avg_mini_batch_size for _ in range(args.num_batches) ])\n\n elif args.batch_size_distribution == \"file\":\n percentiles = []\n batch_size_distributions = []\n with open(args.batch_dist_file, 'r') as f:\n lines = f.readlines()\n for line in lines:\n percentiles.append(float(line.rstrip()))\n\n for _ in range(args.num_batches):\n batch_size_distributions.append( int(percentiles[ int(np.random.uniform(0, len(percentiles))) ]) )\n\n for i in range(args.num_batches):\n batch_size_distributions[i] = int(max(min(batch_size_distributions[i], args.max_mini_batch_size), 1))\n return batch_size_distributions\n\n# partition the requests into small batches\ndef partition_requests(args, batch_size):\n batch_sizes = []\n\n while batch_size > 0:\n mini_batch_size = min(args.sub_task_batch_size, batch_size)\n batch_sizes.append(mini_batch_size)\n batch_size -= mini_batch_size\n\n return batch_sizes\n\n\ndef loadGenSleep( sleeptime ):\n if sleeptime > 0.0055:\n time.sleep(sleeptime)\n else:\n startTime = time.time()\n while (time.time() - startTime) < sleeptime:\n continue\n return","sub_path":"loadgen/loadgen_utils.py","file_name":"loadgen_utils.py","file_ext":"py","file_size_in_byte":1892,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"439889717","text":"#!/usr/bin/env python3\n\nimport tkinter as tk\nimport random\nimport os\n\n#variables\n\ni = 1\n\nanswers = list(range(20))\nanswers[0]=\"Chuck Norris spricht waehrend der Fahrt mit dem Busfahrer.\";\nanswers[1]=\"Chuck Norris hat bis unendlich gezaehlt. Schon zwei mal.\";\nanswers[2]=\"Chuck Norris kann schwarze Stifte nach Farbe sortieren.\";\nanswers[3]=\"Wenn Chuck Norris ins Wasser geht wird er nicht nass, das Wasser \\n \\twird Chuck Norris.\";\nanswers[4]=\"Chuck Norris versteht Frauen.\";\nanswers[5]=\"Chuck Norris brachte seinem Vater das Rasieren bei.\";\nanswers[6]=\"Physiker sind verbluefft: Cuck Norris zweiter Roundhousekick, kommt \\n \\tnoch vor seinem ersten Roundhauskick an.\";\nanswers[7]=\"Als Lee Harvey Oswalt auf Kenndy schoss, hat Chuck Norris die Kugeln \\n \\tmit seinem eigenen Bart abgefangen, und JFKs Kopf ist nur vor \\n \\terstaunen explodiert.\";\nanswers[8]=\"Chuck Norris bucht in einem Schweizer Hotel ein Zimmer mit Meerblick \\n \\tund bekommt es.\";\nanswers[9]=\"Chuck Norris wurde von seiner Tante geboren, da keiner so lebensmuede \\n \\twar, seine Mutter zu knattern.\";\nanswers[10]=\"Chuck Norris kann im Kinderkarussell ueberholen.\";\nanswers[11]=\"Chuck Norris hat eine Erste-Hilfe-Puppe wiederbelebt.\";\nanswers[12]=\"Chuck Norris kann den Limes (1+(1/x))^x für x -> Infinity genau \\n \\tberechnen.\";\nanswers[13]=\"Chuck Norris trinkt aus einem Wasserhahn auf Ex.\";\nanswers[14]=\"Chuck Norris hat ganz Gallien besetzt. Ganz Gallien? Ja! Ganz Gallien!\";\nanswers[15]=\"Chuck Norris stirbt nach der Hoffnung.\";\nanswers[16]=\"Wie viele Liegestuetze schafft Chuck Norris? Alle!\";\nanswers[17]=\"Chuck Norris wurde einmal von einer Koenigskobra gebissen. Nach fuenf \\n \\tqualvollen Tagen voller Schmerz starb die Kobra.\";\nanswers[18]=\"Chuck Norris zaehlt alle Schafe einer Herde in 2 Sekunden. Sein Trick: \\n \\tEr zaehlt die Beine und teilt am Ende durch 4.\";\nanswers[19]=\"Chuck Norris wurde nicht geboren, sondern entfesselt.\";\n\n#functions\n\ndef callbackCn(event=None):\n\n\tglobal i\n\n\tplaintextfile = open(\"log.dat\", \"a\")\n\tentryText = cnEntry.get()\n\tcnEntry.delete(0, \"end\")\n\t#answers-start\n\n\trandomNumber = random.randint(0, 19)\n\n\tif \"tschuess\" in entryText:\n\t\twriteContentCn = \"Chuck Norris: *verpasst dir einen Roundhousekick*\"\n\t\ti=i+1\n\telse:\n\t\tanswer = \"Fakt: \"+answers[randomNumber]\n\t\twriteContentCn = answer\n\t\tif randomNumber in [3,6,8,9,12,17,18]:\n\t\t\ti=i+2\n\t\telif randomNumber == 7:\n\t\t\ti=i+3\n\t\telse:\n\t\t\ti=i+1\n\n\t#answers-end\n\twriteContentYou = \"\\nDu: \"+entryText\n\tplaintextfile.write(writeContentYou)\n\ti=i+1\n\twriteContentMe = \"\\n\"+writeContentCn\n\tplaintextfile.write(writeContentMe)\n\tplaintextfile.close()\n\tplaintextfile = open(\"log.dat\", \"r\")\n\tplaintext = plaintextfile.read()\n\tplaintextfile.close()\n\ttext.set(plaintext)\n\n\twhile i > 40:\n\t\ttry:\n\t\t\tplaintextfile = open(\"log.dat\", \"w\")\n\t\t\tplaintextfile.write(writeContentMe)\n\t\t\tplaintextfile.close()\n\t\t\ti=1\n\t\t\tbreak\n\t\texcept OSError:\n\t\t\ti=41\n\n#init gui\n\ncn = tk.Tk()\ncn.title(\"Chuck Norris\")\n\n#gui design\n\ntext = tk.StringVar()\nplaintextfile = open(\"log.dat\", \"w\")\nplaintextfile.write(\"Chuck Norris: Checkmate!\")\nplaintextfile.close()\nplaintextfile = open(\"log.dat\", \"r\")\nplaintext = plaintextfile.read()\ntext.set(plaintext)\nplaintextfile.close()\navcn = tk.PhotoImage(file=\"avatars/randoms/cn\")\navatar = tk.Label(cn, image=avcn)\navatar.image = avcn\navatar.grid(row=0, column=0, rowspan=5)\ncnChat = tk.Label(cn, textvariable=text, anchor=\"c\", relief=\"sunken\", width=80, height=40)\ncnChat.grid(row=1, column=1, rowspan=10, columnspan=4)\ncnEntry = tk.Entry(cn, width=70)\ncnEntry.grid(row=11, column=1, rowspan=2, columnspan=3)\ncnSend = tk.Button(cn, text=\"Send!\", command=callbackCn, width=10)\ncnSend.grid(row=11, column=4, rowspan=2, columnspan=1)\n\ncn.bind(\"\", callbackCn)\ncn.mainloop()\n","sub_path":"pyBots/randoms/chucknorris.py","file_name":"chucknorris.py","file_ext":"py","file_size_in_byte":3734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"150787544","text":"import argparse\nimport csv\nimport os\nimport re\nimport sys\nfrom pathlib import Path\n\nimport numpy as np\nimport pandas as pd\nimport requests\nfrom dotenv import load_dotenv\nfrom loguru import logger\n\nlogger.add(sys.stderr, level=\"INFO\")\nlogger.add(\"debug.log\", level=\"DEBUG\", rotation=\"500 MB\")\n\nnp.random.seed(seed=42)\n\nROOT = Path(os.path.dirname(os.path.dirname(__file__))) / Path(os.path.basename(os.path.dirname(__file__)))\nDATA_DIR = ROOT / 'data'\nenv_path = ROOT / '.env'\nload_dotenv(dotenv_path=env_path)\n\n# =================== COINFORM API SETTINGS =================\nCOINFORM_ENDPOINT = os.getenv('COINFORM_ENDPOINT')\nQUERY_ID_REQUEST = COINFORM_ENDPOINT + '/twitter/tweet'\nRESPONSE_TWEET = COINFORM_ENDPOINT + '/response/{query_id}/debug'\n\nclass Sample_Generator():\n def __init__(self, args):\n # =================== Params =================\n self.total_modules = args.n_modules\n self.total_sample = args.n_samples\n\n # credibility boundaries\n self.misinfome_cred = args.misinfome_cred\n self.content_analys_cred = args.content_analysis_cred\n self.claim_cred = args.claim_cred\n\n # confidences\n self.misinfome_conf = args.misinfome_conf\n self.content_analys_conf = args.content_analysis_conf\n self.claim_conf = args.claim_conf\n self.labels = {'credible': 0, 'mostly_credible': 1, 'mostly_not_credible': 2, 'credible_uncertain': 3,\n 'not_credible': 4,\n 'not_verifiable': 5}\n self.modules = {'misinfome': [self.misinfome_cred, self.misinfome_conf],\n 'content_analys': [self.content_analys_cred, self.content_analys_conf],\n 'claim': [self.claim_cred, self.claim_conf]}\n\n\n def _all_agree_helper(self):\n labels = list(self.labels.keys())\n data = pd.DataFrame()\n misinfome_creds = []\n content_analys_creds = []\n claim_creds = []\n cred_labels = []\n\n for i in range(0, len(labels) - 1):\n if i == 0:\n misinfome_creds.append(np.random.uniform(high=1, low=self.misinfome_cred[i],\n size=self.total_sample))\n content_analys_creds.append(np.random.uniform(high=1,\n low=self.content_analys_cred[i],\n size=[self.total_sample]))\n claim_creds.append(np.random.uniform(high=1, low=self.claim_cred[i],\n size=[self.total_sample]))\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n\n elif i == 4:\n misinfome_creds.append(np.random.uniform(high=self.misinfome_cred[i - 1],\n low=-1,\n size=[self.total_sample]))\n content_analys_creds.append(np.random.uniform(high=self.content_analys_cred[i - 1],\n low=-1,\n size=[self.total_sample]))\n claim_creds.append(np.random.uniform(high=self.claim_cred[i - 1], low=-1,\n size=[self.total_sample]))\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n else:\n\n misinfome_creds.append(np.random.uniform(high=self.content_analys_cred[i - 1],\n low=self.content_analys_cred[i],\n size=[self.total_sample]))\n content_analys_creds.append(np.random.uniform(high=self.content_analys_cred[i - 1],\n low=self.content_analys_cred[i],\n size=[self.total_sample]))\n claim_creds.append(np.random.uniform(high=self.claim_cred[i - 1],\n low=self.claim_cred[i], size=[self.total_sample]))\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n\n data['misinfome_cred'] = np.asarray(misinfome_creds).flatten()\n data['content_analys_cred'] = np.asarray(content_analys_creds).flatten()\n data['claim_cred'] = np.asarray(claim_creds).flatten()\n data['expected_credible'] = np.asarray(cred_labels).flatten()\n\n return data\n\n def _pick_random_modules(self, num_diff_module):\n '''\n :param num_module: number of module which has different functions\n :type num_module: int\n :return:\n :rtype:\n '''\n all_idxs = set([i for i in range(len(self.modules.keys()))])\n disagree_idxs = set()\n picked_flag = False\n for i in range(num_diff_module):\n while not picked_flag:\n picked_num = np.random.randint(high=len(self.modules.keys()), low=0, size=1)[0]\n if picked_num not in disagree_idxs:\n disagree_idxs.add(picked_num)\n picked_flag = True\n\n agree_idxs = all_idxs - disagree_idxs\n return {'agree': [list(self.modules.keys())[agree_idx] for agree_idx in agree_idxs],\n 'disagree': [list(self.modules.keys())[disagree_idx] for disagree_idx in disagree_idxs]}\n\n def _pick_random_label(self, idx_agreed):\n labels = list(self.labels.keys())\n picked_flag = False\n picked_id = None\n while (not picked_flag):\n random_idx = np.random.randint(high=len(labels), low=0, size=1)[0]\n if random_idx is not idx_agreed:\n picked_id = random_idx\n picked_flag = True\n\n return labels[picked_id]\n\n def some_agree(self):\n '''\n This method creates values that some modules agree, some disagree with high/low confidence\n :return:\n :rtype:\n '''\n dummy_values = pd.DataFrame()\n # if data folder does not exist, create\n if not os.path.exists(DATA_DIR):\n os.makedirs(DATA_DIR)\n\n for i in range(1, len(self.modules.keys()) + 1):\n data_low_conf = self._some_agree_helper(num_diff_module=i, confidence_density=False)\n dummy_values = dummy_values.append(data_low_conf, ignore_index=True, sort=True)\n data_high_conf = self._some_agree_helper(num_diff_module=i, confidence_density=True)\n dummy_values = dummy_values.append(data_high_conf, ignore_index=True, sort=True)\n\n # save dummy values {casename}_{module_name}_{upboundary_cred}_{conf}\n path = DATA_DIR / '{func_name}_misinfome_{misinfome_cred}_{misinfome_conf}_contentanalysis_{content_analysis_cred}_{content_analysis_conf}_claim_{claim_cred}_{claim_conf}.csv'.format(\n func_name=self.some_agree.__name__,\n misinfome_cred=str(self.misinfome_cred[0]), misinfome_conf=str(self.misinfome_conf),\n content_analysis_conf=self.content_analys_conf, content_analysis_cred=self.content_analys_cred[0],\n claim_cred=str(self.claim_cred[0]), claim_conf=self.claim_conf)\n dummy_values.to_csv(path)\n\n def _some_agree_helper(self, num_diff_module, confidence_density):\n '''\n :param num_diff_module: number of modules which disgree\n :type num_diff_module: int\n :param confidence_density: confidence density of disagreed modules. If it is true, modules disagree with high confidence\n :type: boolean\n :return:\n :rtype:\n '''\n #### 2 modules agree, one is not ######\n random_modules = self._pick_random_modules(num_diff_module=num_diff_module)\n agreed_modules = random_modules['agree']\n disagreed_modules = random_modules['disagree']\n labels = list(self.labels.keys())\n temp_creds = {'misinfome': [], 'content_analys': [], 'claim': []}\n temp_confs = {'misinfome': [], 'content_analys': [], 'claim': []}\n cred_labels = []\n data = pd.DataFrame()\n print('Agreed modules {}'.format(agreed_modules))\n print('Disagreed modules {}'.format(disagreed_modules))\n for i in range(0, len(self.labels.keys()) - 1):\n if i == 0:\n for agreed_module in agreed_modules:\n temp_creds[agreed_module].append(np.random.uniform(high=1, low=self.modules[agreed_module][0][i],\n size=self.total_sample))\n # high confidence\n temp_confs[agreed_module].append(\n np.random.uniform(high=1, low=self.modules[agreed_module][1], size=self.total_sample))\n\n # agreed module -> credible, disagreed modules -> mostly credible (i+1), but disagree module's label is not final.\n for disagreed_module in disagreed_modules:\n temp_creds[disagreed_module].append(np.random.uniform(high=self.modules[disagreed_module][0][i],\n low=self.modules[disagreed_module][0][i + 1],\n size=self.total_sample))\n temp_confs[disagreed_module].append(\n np.random.uniform(high=1, low=self.modules[disagreed_module][1],\n size=self.total_sample)) if confidence_density else temp_confs[\n disagreed_module].append(\n np.random.uniform(high=self.modules[disagreed_module][1], low=0, size=self.total_sample))\n\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n\n elif i == 4:\n for agreed_module in agreed_modules:\n temp_creds[agreed_module].append(np.random.uniform(high=self.modules[agreed_module][0][i - 1],\n low=-1,\n size=self.total_sample))\n # high confidence\n temp_confs[agreed_module].append(\n np.random.uniform(high=1, low=self.modules[agreed_module][1], size=self.total_sample))\n # agreed module -> not credible, disagreed modules -> credible uncertain (i-1)\n for disagreed_module in disagreed_modules:\n temp_creds[disagreed_module].append(np.random.uniform(high=self.modules[disagreed_module][0][i - 1],\n low=self.modules[disagreed_module][0][i - 2],\n size=self.total_sample))\n temp_confs[disagreed_module].append(\n np.random.uniform(high=1, low=self.modules[disagreed_module][1],\n size=self.total_sample)) if confidence_density else temp_confs[\n disagreed_module].append(\n np.random.uniform(high=self.modules[disagreed_module][1], low=0, size=self.total_sample))\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n else:\n for agreed_module in agreed_modules:\n temp_creds[agreed_module].append(np.random.uniform(high=self.modules[agreed_module][0][i - 1],\n low=self.modules[agreed_module][0][i],\n size=\n self.total_sample))\n # high confidence\n temp_confs[agreed_module].append(\n np.random.uniform(high=1, low=self.modules[agreed_module][1], size=self.total_sample))\n # disagreed module -> preeceding (i-1)\n for disagreed_module in disagreed_modules:\n # preeceding label\n temp_creds[disagreed_module].append(np.random.uniform(high=self.modules[disagreed_module][0][i - 1],\n low=self.modules[disagreed_module][0][i - 2],\n size=\n self.total_sample))\n temp_confs[disagreed_module].append(\n np.random.uniform(high=1, low=self.modules[disagreed_module][1],\n size=self.total_sample)) if confidence_density else temp_confs[\n disagreed_module].append(\n np.random.uniform(high=self.modules[disagreed_module][1], low=0, size=self.total_sample))\n cred_labels.append(np.asarray([labels[i] for _ in range(self.total_sample)]))\n\n for name, values in temp_creds.items():\n data[name + '_cred'] = np.asarray(values).flatten()\n data[name + '_conf'] = np.asarray(temp_confs[name]).flatten()\n data['expected_credible'] = np.asarray(cred_labels).flatten()\n return data\n\n def all_agree_all_high(self):\n '''\n In this case all of modules agree on one credibility label with high confidence\n '''\n dummy_values = pd.DataFrame()\n # if data folder does not exist, create\n if not os.path.exists(DATA_DIR):\n os.makedirs(DATA_DIR)\n\n data = self._all_agree_helper()\n\n # confidence value always high between th>val>1\n data['misinfome_conf'] = np.random.uniform(high=1, low=self.misinfome_conf, size=[data.shape[0]])\n data['content_analys_conf'] = np.random.uniform(high=1, low=self.content_analys_conf,\n size=[data.shape[0]])\n data['claim_conf'] = np.random.uniform(high=1, low=self.claim_conf, size=[data.shape[0]])\n\n dummy_values = dummy_values.append(data, ignore_index=True, sort=True)\n # save dummy values {casename}_{module_name}_{upboundary_cred}_{conf}\n path = DATA_DIR / '{func_name}_misinfome_{misinfome_cred}_{misinfome_conf}_contentanalysis_{content_analysis_cred}_{content_analysis_conf}_claim_{claim_cred}_{claim_conf}.csv'.format(\n func_name=self.all_agree_all_high.__name__,\n misinfome_cred=str(self.misinfome_cred[0]), misinfome_conf=str(self.misinfome_conf),\n content_analysis_conf=self.content_analys_conf, content_analysis_cred=self.content_analys_cred[0],\n claim_cred=str(self.claim_cred[0]), claim_conf=self.claim_conf)\n dummy_values.to_csv(path)\n\n def all_agree_some_high(self):\n '''\n In this case all of modules agree on one credibility label, but some of them with high confidence\n '''\n dummy_values = pd.DataFrame()\n # if data folder does not exist, create\n if not os.path.exists(DATA_DIR):\n os.makedirs(DATA_DIR)\n\n data = self._all_agree_helper()\n\n misinfo_conf = []\n content_analys_conf = []\n claim_conf = []\n # confidence value always high between th>val>1 half high\n high_conf_sample = data.shape[0] // 2\n low_conf_sample = data.shape[0] - high_conf_sample\n\n misinfo_conf.append(np.random.uniform(high=1, low=self.misinfome_conf, size=[high_conf_sample]))\n\n content_analys_conf.append(np.random.uniform(high=1, low=self.content_analys_conf,\n size=high_conf_sample))\n claim_conf.append(np.random.uniform(high=1, low=self.claim_conf, size=high_conf_sample))\n\n # confidence value always low val>0\n misinfo_conf.append(np.random.uniform(high=self.misinfome_conf, low=0, size=low_conf_sample))\n content_analys_conf.append(np.random.uniform(high=self.content_analys_conf, low=0,\n size=low_conf_sample))\n claim_conf.append(np.random.uniform(high=self.claim_conf, low=0, size=low_conf_sample))\n\n data['misinfome_conf'] = np.asarray(misinfo_conf).flatten()\n data['content_analys_conf'] = np.asarray(content_analys_conf).flatten()\n data['claim_conf'] = np.asarray(claim_conf).flatten()\n\n dummy_values = dummy_values.append(data, ignore_index=True, sort=True)\n # save dummy values {casename}_{module_name}_{upboundary_cred}_{conf}\n path = DATA_DIR / '{func_name}_misinfome_{misinfome_cred}_{misinfome_conf}_contentanalysis_{content_analysis_cred}_{content_analysis_conf}_claim_{claim_cred}_{claim_conf}.csv'.format(\n func_name=self.all_agree_some_high.__name__,\n misinfome_cred=str(self.misinfome_cred[0]), misinfome_conf=str(self.misinfome_conf),\n content_analysis_conf=self.content_analys_conf, content_analysis_cred=self.content_analys_cred[0],\n claim_cred=str(self.claim_cred[0]), claim_conf=self.claim_conf)\n dummy_values.to_csv(path)\n\n def all_not_verified(self):\n '''\n All of them have low confidence or either fail\n todo: fail case is not implemented\n '''\n dummy_values = pd.DataFrame()\n # if data folder does not exist, create\n if not os.path.exists(DATA_DIR):\n os.makedirs(DATA_DIR)\n data = self._all_agree_helper()\n\n # all of them has low confidence, hence they are unverified.\n data['misinfome_conf'] = np.random.uniform(high=self.misinfome_conf, low=0, size=[data.shape[0]]).flatten()\n data['content_analys_conf'] = np.random.uniform(high=self.content_analys_conf, low=0,\n size=[data.shape[0]]).flatten()\n data['claim_conf'] = np.random.uniform(high=self.claim_conf, low=0, size=[data.shape[0]]).flatten()\n\n # label credibility\n data['expected_credible'] = 'not_verifiable'\n\n dummy_values = dummy_values.append(data, ignore_index=True, sort=True)\n # save dummy values {casename}_{module_name}_{upboundary_cred}_{conf}\n path = DATA_DIR / '{func_name}_misinfome_{misinfome_cred}_{misinfome_conf}_contentanalysis_{content_analysis_cred}_{content_analysis_conf}_claim_{claim_cred}_{claim_conf}.csv'.format(\n func_name=self.all_agree_some_high.__name__,\n misinfome_cred=str(self.misinfome_cred[0]), misinfome_conf=str(self.misinfome_conf),\n content_analysis_conf=self.content_analys_conf, content_analysis_cred=self.content_analys_cred[0],\n claim_cred=str(self.claim_cred[0]), claim_conf=self.claim_conf)\n dummy_values.to_csv(path)\n\n def _map_label(self, label):\n print('Not implemented yet!!')\n return None\n\n def _request(self, tweet_id):\n # logger.debug('I am requesting tweet {}'.format(tweet_id))\n args = {\n \"tweet_id\": parse_id(tweet_id),\n \"tweet_author\": \"string\",\n \"tweet_text\": \"string\"\n }\n # first response includes query id\n response_1 = requests.post(QUERY_ID_REQUEST, json=args).json()\n if 'query_id' not in response_1:\n return None\n query_id = response_1['query_id']\n task_completed = False\n modules_response = {}\n\n err_count = 100\n while (not task_completed):\n response_2 = requests.get(RESPONSE_TWEET.format(query_id=query_id)).json()\n status = response_2['status']\n # logger.debug('Query response {}'.format(status))\n if status == 'partly_done' or status == 'in_progress':\n err_count -= 1\n if err_count == 0:\n status = 'done'\n if status == 'done':\n response_codes = response_2['module_response_code']\n logger.debug(response_2['flattened_module_responses'])\n if response_codes[\n 'claimcredibility'] == 200 and 'claimcredibility_tweet_claim_credibility_0_confidence' in \\\n response_2['flattened_module_responses']:\n modules_response['claim_conf'] = response_2['flattened_module_responses'][\n 'claimcredibility_tweet_claim_credibility_0_confidence']\n modules_response['claim_cred'] = response_2['flattened_module_responses'][\n 'claimcredibility_tweet_claim_credibility_0_credibility']\n else:\n modules_response['claim_conf'] = -100\n modules_response['claim_cred'] = -100\n if response_codes['contentanalysis'] == 200 and 'contentanalysis_credibility' in response_2[\n 'flattened_module_responses']:\n modules_response['content_analys_conf'] = response_2['flattened_module_responses'][\n 'contentanalysis_confidence']\n modules_response['content_analys_cred'] = response_2['flattened_module_responses'][\n 'contentanalysis_credibility']\n else:\n modules_response['content_analys_conf'] = -100\n modules_response['content_analys_cred'] = -100\n if response_codes['misinfome'] == 200 and 'misinfome_credibility_value' in response_2[\n 'flattened_module_responses']:\n modules_response['misinfome_conf'] = response_2['flattened_module_responses'][\n 'misinfome_credibility_confidence']\n modules_response['misinfome_cred'] = response_2['flattened_module_responses'][\n 'misinfome_credibility_value']\n else:\n modules_response['misinfome_conf'] = -100\n modules_response['misinfome_cred'] = -100\n\n task_completed = True\n\n return modules_response\n\n def export_to_file(self, row, file_path):\n with open(file_path, 'a', encoding='utf-8') as f:\n cw = csv.writer(f, delimiter='\\t')\n cw.writerow(row)\n\n def from_misinfome(self):\n '''\n Retrieves english tweets from misinfome collection and record tweet ids and labels.\n :return:\n :rtype:\n '''\n dest_file = DATA_DIR / 'misinfome.tsv'\n src_file = DATA_DIR / 'misinfome' / 'joined_tables.tsv'\n fc_labels_file = DATA_DIR / 'misinfome' / 'fact_checking_gold_labels.tsv'\n responses_file = DATA_DIR / 'misinfome' / 'misinfome_responses.csv'\n file_path = DATA_DIR / 'misinfome/rule-responses/export.csv'\n\n if not os.path.isfile(dest_file):\n data = pd.read_csv(src_file, sep='\\t')\n mask = (data['lang'] == 'en') & (data['source'].str.contains('twitter'))\n data = data[mask]\n data = data[['url', 'factchecker_label']]\n if not fc_labels_file.exists():\n fc_labels = pd.DataFrame(pd.unique(data['factchecker_label']))\n fc_labels.to_csv(fc_labels_file)\n\n ## claim_conf,claim_cred,content_analys_conf,content_analys_cred,expected_credible,misinfome_conf,misinfome_cred\n if not responses_file.exists():\n self.export_to_file(['#id', 'url', 'claim_conf',\n 'claim_cred',\n 'content_analys_conf',\n 'content_analys_cred',\n 'misinfome_conf',\n 'misinfome_cred'], file_path)\n for index, row in data.iterrows():\n row['id'] = parse_id(row['url'])\n logger.info(row['id'])\n response = self._request(row['url'])\n if response:\n row_obj = {\n 'id': row['id'],\n 'url': row['url'],\n 'claim_conf': response['claim_conf'],\n 'claim_cred': response['claim_cred'],\n 'content_analys_conf': response['content_analys_conf'],\n 'content_analys_cred': response['content_analys_cred'],\n 'misinfome_conf': response['misinfome_conf'],\n 'misinfome_cred': response['misinfome_cred'],\n }\n # row['expected_credible'] = self._map_label(row['factchecker_label'])\n self.export_to_file(list(row_obj.values()), file_path)\n # data[['claim_conf', 'claim_cred', 'content_analys_conf', 'content_analys_cred', 'misinfome_conf',\n # 'misinfome_cred']].to_csv(responses_file)\n # todo add final data csv\n\n\nif __name__ == '__main__':\n print('This script generates samples for testing rules')\n parser = argparse.ArgumentParser()\n parser.add_argument('--n_samples', type=int, default=20)\n parser.add_argument('--misinfome_cred', action='store',\n type=float, nargs=4, default=[0.66, 0.33, -0.33, -0.66],\n help=\"Examples: --misinfome_cred item1 item2\")\n parser.add_argument('--content_analysis_cred', action='store',\n type=float, nargs=4, default=[0.6, 0.3, -0.3, -0.6],\n help=\"Examples: --content_analysis_cred item1 item2\")\n parser.add_argument('--claim_cred', action='store',\n type=float, nargs=4, default=[0.5, 0.25, -0.5, -0.25],\n help=\"Examples: --claim_cred item1 item2\")\n parser.add_argument('--misinfome_conf',\n type=float, default=0.5)\n parser.add_argument('--content_analysis_conf',\n type=float, default=0.6)\n parser.add_argument('--claim_conf', type=float, default=0.7)\n parser.add_argument('--n_modules', type=int, default=3, help=\"total number of modules\")\n parser.add_argument('--sample_mode', type=str, default='external_misinfome',\n help=\"select sample mode, all_not_verified, all_agree_all_high or some agree\")\n\n args = parser.parse_args()\n sample_gen = Sample_Generator(args)\n mode = args.sample_mode\n\n print('Selected mode is {}'.format(mode))\n\n if mode == 'all_not_verified':\n sample_gen.all_not_verified()\n elif mode == 'all_agree_all_high':\n sample_gen.all_agree_all_high()\n elif mode == 'some_agree':\n sample_gen.some_agree()\n elif mode == 'all_agree_some_high':\n sample_gen.all_agree_some_high()\n elif mode == 'external_misinfome':\n sample_gen.from_misinfome()\n","sub_path":"sample_generator.py","file_name":"sample_generator.py","file_ext":"py","file_size_in_byte":27080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"20194211","text":"# -*- coding: utf-8 -*-\n\nimport unittest\nimport os\nimport codecs\nimport json\nimport urllib2\nfrom json_loader import JsonLoader\nfrom model.gamestate import *\nfrom constants import PATH\nfrom mox import *\nimport json\n\n\nclass JsonLoaderTest(unittest.TestCase):\n def setUp(self):\n self.mox = Mox()\n self.json_loader = JsonLoader()\n self.json_string = []\n self.example_cnt = 2\n for example_num in range(self.example_cnt):\n with codecs.open(os.path.join('test_data', 'example_'\n + str(example_num) + '.json'),\n encoding='utf-8', mode='r') as f:\n self.json_string.append(f.read())\n\n def test_json_translate(self):\n for example_num in range(self.example_cnt):\n self.game_state = self.json_loader.json_translate(json.loads(\n self.json_string[example_num]))\n\n self.assertTrue(isinstance(self.game_state, GameState))\n self.assertTrue(isinstance(self.game_state.players, Players))\n self.assertTrue(isinstance(self.game_state.hat, Hat))\n self.assertTrue(isinstance(self.game_state.settings, Settings))\n\n for player in self.game_state.players.players:\n self.assertTrue(isinstance(player, Player))\n self.assertTrue(isinstance(player.name, unicode))\n self.assertTrue(isinstance(player.words, list))\n\n for word in self.game_state.hat.words:\n self.assertTrue(isinstance(word, Word))\n self.assertTrue(isinstance(word.text, unicode))\n self.assertTrue(isinstance(word.owner, Player) or\n isinstance(word.owner, type(None)))\n\n self.assertTrue(isinstance(self.game_state.settings.\n time_per_round_sec, int))\n self.assertTrue(isinstance(self.game_state.settings.\n time_before_out_sec, int))\n self.assertTrue(isinstance(self.game_state.settings.\n skip_words, int))\n\n def test_load_from_url(self):\n url = 'test string'\n with codecs.open(PATH + '/test_data/example_0.json',\n encoding='utf-8', mode='r') as file_opened:\n self.mox.StubOutWithMock(urllib2, 'urlopen')\n urllib2.urlopen(IsA(str), timeout=IsA(int)).AndReturn(file_opened)\n self.mox.ReplayAll()\n self.json_string = self.json_loader.load_from_url(url)\n\n self.mox.UnsetStubs()\n self.mox.VerifyAll()\n\n with codecs.open(PATH + '/test_data/example_0.json',\n encoding='utf-8', mode='r') as file_opened:\n self.assertEqual(self.json_string, json.loads(file_opened.read()))\n\n def test_load_game_from_url(self):\n url = 'test string'\n with codecs.open(PATH + '/test_data/example_0.json',\n encoding='utf-8', mode='r') as file_opened:\n self.mox.StubOutWithMock(urllib2, 'urlopen')\n urllib2.urlopen(IsA(str), timeout=IsA(int)).AndReturn(file_opened)\n self.mox.ReplayAll()\n self.game_state = self.json_loader.load_game_from_url(url)\n\n self.mox.UnsetStubs()\n self.mox.VerifyAll()\n self.assertTrue(isinstance(self.game_state, GameState))\n\n def test_dump_to_url(self):\n url = 'test string'\n with codecs.open(PATH + '/test_data/example_0.json',\n encoding='utf-8', mode='r') as file_opened:\n self.mox.StubOutWithMock(urllib2, 'urlopen')\n urllib2.urlopen(IsA(str), IsA(str), timeout=IsA(int))\n self.mox.ReplayAll()\n self.json_loader.dump_to_url(url, self.json_string[0])\n\n self.mox.UnsetStubs()\n self.mox.VerifyAll()\n\nif __name__ == \"__main__\":\n unittest.main()\n","sub_path":"model/json_loader_test.py","file_name":"json_loader_test.py","file_ext":"py","file_size_in_byte":3924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"390761613","text":"#! /usr/bin/env python3\n# ex: et sts=4 ts=4 sw=4\n\nimport sys\nimport os\nimport shutil\n\nfrom acdcli.api import client as ACD\nfrom acdcli.cache import db as DB\nfrom acdcli.api.common import RequestError\nfrom acdcli.utils import hashing\nfrom acdcli.utils.time import datetime_to_timestamp\n\n\nclass Cache(object):\n\n def __init__(self, cache_folder):\n self._cache_folder = cache_folder\n auth_folder = os.path.expanduser('~/.cache/acd_cli')\n self._acd_client = ACD.ACDClient(auth_folder)\n self._acd_db = DB.NodeCache(auth_folder)\n self._last_recycle = 0\n\n def __call__(self):\n folder_id = self._acd_db.resolve_path('/tmp')\n children = self._acd_db.list_children(folder_id)\n children = map(lambda _: _.node, children)\n children = sorted(children, key=lambda _: _.modified, reverse=True)\n for child in children:\n self._recycle_space()\n if self._is_too_old(child):\n break\n self._download(child, self._cache_folder)\n\n def _recycle_space(self):\n entries = self._get_cache_entries()\n\n while True:\n free_space = self._get_free_space()\n if free_space > 10:\n break\n\n full_path, mtime = entries.pop(0)\n if os.path.isdir(full_path):\n shutil.rmtree(full_path)\n else:\n os.remove(full_path)\n self._last_recycle = mtime\n print('recycled: ' + full_path)\n\n def _get_cache_entries(self):\n entries = os.listdir(self._cache_folder)\n entries = (os.path.join(self._cache_folder, _) for _ in entries)\n entries = ((_, os.stat(_).st_mtime) for _ in entries)\n entries = sorted(entries, key=lambda _: _[1])\n return entries\n\n def _get_free_space(self):\n s = os.statvfs(self._cache_folder)\n s = s.f_frsize * s.f_bavail\n s = s / 1024\n s = s / 1024 / 1024\n return s\n\n def _is_too_old(self, node):\n return datetime_to_timestamp(node.modified) < self._last_recycle\n\n def _download(self, node, local_path):\n local_path = local_path if local_path else ''\n full_path = os.path.join(local_path, node.name)\n\n if not node.is_available():\n return False\n\n if node.is_folder():\n try:\n os.makedirs(full_path, exist_ok=True)\n except OSError:\n print('mkdir failed: ' + full_path)\n return False\n for child in node.children:\n self._download(child, full_path)\n else:\n if os.path.isfile(full_path):\n print('skip existed: ' + full_path)\n if os.path.getsize(full_path) != node.size:\n print('size mismatch: ' + full_path)\n return False\n return True\n\n while True:\n hasher = hashing.IncrementalHasher()\n try:\n print('downloading: ' + full_path)\n self._acd_client.download_file(node.id, node.name, local_path, write_callbacks=[hasher.update])\n print('downloaded: ' + full_path)\n except RequestError as e:\n print('download failed: ' + str(e))\n else:\n local = hasher.get_result()\n remote = node.md5\n if local != remote:\n print('md5 mismatch: ' + full_path)\n os.remove(full_path)\n else:\n break\n\n preserve_mtime(node, full_path)\n\n return True\n\n\ndef main(args=None):\n if args is None:\n args = sys.argv\n\n cache = Cache(args[1])\n cache()\n\n return 0\n\n\ndef preserve_mtime(node, full_path):\n mtime = datetime_to_timestamp(node.modified)\n os.utime(full_path, (mtime, mtime))\n\n\nif __name__ == '__main__':\n exit_code = main()\n sys.exit(exit_code)\n","sub_path":"sandbox/python/acdcache.py","file_name":"acdcache.py","file_ext":"py","file_size_in_byte":3993,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"5600716","text":"from tkinter import *\nimport math\n\nPINK = \"#e2979c\"\nRED = \"#e7305b\"\nGREEN = \"#9bdeac\"\nYELLOW = \"#f7f5dd\"\nFONT_NAME = \"Courier\"\nWORK_MIN = 25\nSHORT_BREAK_MIN = 5\nLONG_BREAK_MIN = 20\nreps = 0\ntimer = None\n\n\ndef start_timer():\n global reps\n reps += 1\n work_sec = WORK_MIN * 60\n short_break_sec = SHORT_BREAK_MIN * 60\n long_break_sec = LONG_BREAK_MIN * 60\n\n if reps == 8:\n countdown(long_break_sec)\n title_label.config(text =\"Break\", fg=RED)\n\n elif reps % 2 == 0:\n countdown(short_break_sec)\n title_label.config(text =\"Break\", fg=PINK)\n\n else:\n countdown(work_sec)\n title_label.config(text=\"Work\", fg=GREEN)\n\n\ndef reset_timer():\n window.after_cancel(timer)\n canvas.itemconfig(timer_text, text = \"00:00\")\n title_label.config(text=\"Timer\", fg=\"white\")\n checkmark_label.config(text=\"\")\n\n\ndef countdown(count):\n count_min = math.floor(count/60)\n count_sec = count % 60\n if count_sec < 10:\n count_sec = f\"0{count_sec}\"\n canvas.itemconfig(timer_text, text=f\"{count_min}:{count_sec}\")\n if count > 0:\n global timer\n timer = window.after(1000, countdown, count-1)\n else:\n start_timer()\n marks = \"\"\n work_sessions = math.floor(reps/2)\n for _ in range (work_sessions):\n marks += \"✅\"\n checkmark_label.config(text=marks)\n\n\nwindow = Tk()\nwindow.title(\"Pomodoro\")\nwindow.config(padx=100, pady=50, bg=\"black\")\n\n\nfile = PhotoImage(file=\"tomato.png\")\n\ncanvas = Canvas(width=220, height=224, bg=\"black\", highlightthickness=0)\ncanvas.create_image(100, 112, image=file)\ntimer_text = canvas.create_text(100,130, text=\"00:00\", fill=\"white\", font=(FONT_NAME,35,\"bold\"))\ncanvas.grid(column=1, row=1)\n\n\ntitle_label = Label(text=\"Timer\")\ntitle_label.config(fg=\"white\", bg=\"black\", font=(FONT_NAME, 50, \"normal\"))\ntitle_label.grid(column=1, row=0)\n\ncheckmark_label = Label(bg=\"black\")\ncheckmark_label.grid(column=1, row=3)\n\n\nstart_button = Button(text=\"Start\", command=start_timer, highlightthickness=0)\nstart_button.grid(column=0, row=2)\n\nreset_button = Button(text=\"Reset\", command=reset_timer, highlightthickness=0)\nreset_button.grid(column=2, row=2)\n\nwindow.mainloop()","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"75652931","text":"# You may create additional functions here:\nimport ast\n\n\n# The function read and manipulate data to a desirable format\ndef readFile():\n puzzleData = []\n with open(\"sample.txt\", \"r\") as lines:\n lineData = []\n for line in lines:\n line = line.strip().replace(' ',', ').replace(' ','')\n if line[-1] == ',':\n line= line[:-1]\n puzzleData.append( ast.literal_eval(line))\n return puzzleData\n\n# Function to print the \ndef printPuzzle(puzzle):\n for line in puzzle:\n print(line,'\\n')\n\n\ndef puzzleComplete(puzzle):\n\n for i in range(0,10):\n for j in range(0,10):\n if puzzle[i][j] == 0:\n return False\n \n return True\n\ndef computerRanges(quadrands):\n quadrandsConstrains = []\n quadrandsConstrains.append([1,[0,3],[0,3]])\n quadrandsConstrains.append([2,[3,6],[0,3]])\n quadrandsConstrains.append([3,[6,9],[0,3]])\n quadrandsConstrains.append([4,[0,3],[3,6]])\n quadrandsConstrains.append([5,[3,6],[3,6]])\n quadrandsConstrains.append([6,[6,9],[3,6]])\n quadrandsConstrains.append([7,[0,3],[6,9]])\n quadrandsConstrains.append([8,[3,6],[6,9]])\n quadrandsConstrains.append([9,[6,9],[6,9]])\n\n for line in quadrandsConstrains:\n if line[0] == quadrands:\n return line\n\n\n\ndef puzzleSolved(puzzle):\n total = 45\n for i in range(0,9):\n rowSum = 0\n columnSum =0\n for j in range(0,9):\n rowSum +=puzzle[i][j]\n columnSum +=puzzle[j][i]\n if rowSum != total or columnSum != total:\n return False\n return True\n\n\n# We have 9 quadrand\ndef autoSolveAlgorithm(puzzleCopy , quadrand):\n quadrandData = []\n total = 45\n ranges = computerRanges(quadrand)\n puzzle = puzzleCopy\n rangeRow = ranges[2]\n rangeCol = ranges[1]\n printPuzzle(puzzle)\n\n for i in range(rangeRow[0], rangeRow[1]):\n for j in range(rangeCol[0],rangeCol[1]):\n if puzzle[i][j] != 0:\n total =0\n quadrandData.append(puzzle[i][j])\n quadrandData.sort()\n numberToEValuate = 10\n numberWhichDoNoTExistInAQuadrand = []\n\n # Find missing number ins a quadrand\n for i in range(1,10): \n for item in quadrandData:\n if i == item:\n numberToEValuate = 0\n if numberToEValuate ==10 :\n numberWhichDoNoTExistInAQuadrand.append(i)\n \n # Fill quadrand with missing numbers\n for i in range(rangeRow[0], rangeRow[1]):\n for j in range(rangeCol[0],rangeCol[1]):\n if puzzle[i][j] == 0:\n for num in numberWhichDoNoTExistInAQuadrand:\n if not numberExistInRowColum(puzzle, i, j, num):\n puzzle[i][j] = num\n \n if total == 0:\n autoSolveAlgorithm(puzzle , quadrand)\n else :\n print(\"Solved\")\n printPuzzle(puzzle)\n return\n\n\n\n# Chech if the function exist in both row and column\n\ndef numberExistInRowColum(puzzle, row, col , searcParam) :\n foundInRow = False\n foundInColumn = False\n #check if it exist in a row\n for i in range(0,9):\n if puzzle[row][i] ==searcParam:\n foundInRow = True\n break\n\n #check if it exist in columns\n for i in range(0,9):\n if puzzle[i][col] == searcParam:\n foundInColumn = True\n break\n\n if foundInColumn or foundInRow:\n return True\n elif foundInColumn and foundInRow:\n return True\n else :\n return False\n \n# Additional Functions above this comment\n# Implement your Sudoku Solution Below:\ndef solve_sudoku(puzzle):\n #Edit the code Below Here\n print(\"When pronted to enter row or col, row 1 would be enter as 0, and col 1 as 0 , row 9 as 8 and so on..\")\n puzzleCopy = puzzle\n while not puzzleSolved(puzzleCopy) :\n print(\"..................................................\",'\\n')\n\n print(\"PUZZLE IS NOT YES SOLVED , enter -1 to quit\",'\\n')\n\n print(\"..................................................\",'\\n')\n printPuzzle(puzzleCopy)\n print(\"..................................................\",'\\n')\n\n firstValue = int(input(\"Please Enter your row : \"))\n secondValue = int(input(\"Please Enter your col : \"))\n if firstValue == -1 or secondValue ==-1 :\n break\n\n while firstValue >8 or secondValue >8 :\n firstValue = int(input(\"Please Enter your row (0-8) : \"))\n secondValue = int(input(\"Please Enter your col (0-8) : \"))\n \n solutionNumber = int(input(\"Please Enter the number : \" ))\n if solutionNumber == -1:\n break\n\n puzzle[firstValue][secondValue] = solutionNumber\n\n if puzzleSolved(puzzleCopy):\n cont = int(input(\"inputer Enter 1 to play again or 0 to quit : \"))\n if cont == 1:\n solve_sudoku(puzzle)\n else:\n return\n\n for i in range(1,10):\n autoSolveAlgorithm(puzzleCopy, i)\n\n\nsolve_sudoku(readFile())\n\n\n","sub_path":"Virtual-22/sudo_question.py","file_name":"sudo_question.py","file_ext":"py","file_size_in_byte":5036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"500863138","text":"#!/usr/bin/env python\n# encoding: utf-8\n#\n# Author: Chul Sung\n# Updated by: Chul Sung, daniel.c.sung@gmail.com\n# Date: 02/03/2013\n\nfrom cellcommon import *\nfrom cellcalculatepca import *\nimport scipy\nfrom scipy import ndimage\n\ndef Generate_PCA_Matrix(vol_data,pointarrays,l_radius,planenum,planeselect):\n\n z_size, y_size, x_size = vol_data.shape\n point_num = pointarrays.shape[0]\n \n z_max = z_size - 1\n y_max = y_size - 1\n x_max = x_size - 1\n \n planeimgsize = (l_radius*2+1)**2\n\n # positive point count\n pos_pnum = 0\n \n #f1 = figure(frameon=False)\n #fig_col = math.ceil(point_num/10) \n \n# if planeselect == 0:\n# f1.suptitle('X-Y planes at the center points in unit volumes', fontsize=16)\n# elif planeselect == 1:\n# f1.suptitle('X-Z planes at the center points in unit volumes', fontsize=16)\n# else:\n# f1.suptitle('Y-Z planes at the center points in unit volumes', fontsize=16)\n\n local_posi_pca_input_mtx = np.zeros((point_num,planeimgsize*planenum))\n \n for point in pointarrays:\n x = point[0]\n y = point[1]\n z = point[2]\n\n c_x_min = x - l_radius\n c_y_min = y - l_radius\n c_z_min = z - l_radius\n\n c_x_max = x + l_radius\n c_y_max = y + l_radius\n c_z_max = z + l_radius\n \n if (c_x_min >= 0) and (c_y_min >= 0) and (c_z_min >= 0) and (c_x_max <= x_max) and (c_y_max <= y_max) and (c_z_max <= z_max):\n #f1.add_subplot(fig_col+1, 10, pos_pnum+1) # this line outputs images on top of each other\n #X-Y Planes data\n xyimg = vol_data[z,c_y_min:c_y_max+1,c_x_min:c_x_max+1]\n\n #X-Z Planes data\n xzimg = vol_data[c_z_min:c_z_max+1,y,c_x_min:c_x_max+1]\n \n #Y-Z Planes data\n yzimg = vol_data[c_z_min:c_z_max+1,c_y_min:c_y_max+1,x]\n\n# if planeselect == 0:\n# imshow(xyimg,cmap=cm.Greys_r)\n# elif planeselect == 1:\n# imshow(xzimg,cmap=cm.Greys_r)\n# else:\n# imshow(yzimg,cmap=cm.Greys_r)\n# axis('off')\n \n xyimg = xyimg.flatten()\n xzimg = xzimg.flatten()\n yzimg = yzimg.flatten()\n\n xyzimg = zeros((planeimgsize, planenum))\n xyzimg[:,0] = xyimg\n xyzimg[:,1] = xzimg\n xyzimg[:,2] = yzimg\n \n local_posi_pca_input_mtx[pos_pnum,:] = xyzimg.flatten()\n pos_pnum += 1\n \n del xyimg, xzimg, yzimg, xyzimg\n \n # truncate empty array\n local_posi_pca_input_mtx = local_posi_pca_input_mtx[range(pos_pnum),:]\n \n return local_posi_pca_input_mtx, pos_pnum\n\n\ndef Generate_Neg_PCA_Matrix(vol_data,pointarrays,l_radius,pos_pnum,target_volume,outofbound_thr,planenum,planeselect):\n z_size, y_size, x_size = vol_data.shape\n \n z_max = z_size - 1\n y_max = y_size - 1\n x_max = x_size - 1\n \n planeimgsize = (l_radius*2+1)**2\n \n local_neg_pca_input_mtx = np.zeros((pos_pnum,planeimgsize*planenum))\n \n# f3 = figure(frameon=False)\n# fig_col = math.ceil(pos_pnum/10)\n# \n# if planeselect == 0:\n# f3.suptitle('X-Y planes at the negative points in unit volumes', fontsize=16)\n# elif planeselect == 1:\n# f3.suptitle('X-Z planes at the negative points in unit volumes', fontsize=16)\n# else:\n# f3.suptitle('Y-Z planes at the negative points in unit volumes', fontsize=16)\n\n ## Generate the list of negative points using target_volume\n neg_pointarrays = np.ndarray(shape=(pos_pnum,3), dtype='uint8')\n neg_values = np.ndarray(shape=(pos_pnum,), dtype='d')\n neg_pnum = 0\n while (neg_pnum < pos_pnum):\n cand_z = random.randint(0, z_max)\n cand_y = random.randint(0, y_max)\n cand_x = random.randint(0, x_max)\n\n nx_min = cand_x - l_radius\n ny_min = cand_y - l_radius\n nz_min = cand_z - l_radius\n \n nx_max = cand_x + l_radius\n ny_max = cand_y + l_radius\n nz_max = cand_z + l_radius\n \n if (nx_min >= 0) and (ny_min >= 0) and (nz_min >= 0) and (nx_max <= x_max) and (ny_max <= y_max) and (nz_max <= z_max):\n if(target_volume[cand_z,cand_y,cand_x] < outofbound_thr):\n #f3.add_subplot(fig_col+1, 10, neg_pnum+1) # this line outputs images on top of each other\n \n neg_values[neg_pnum] = round(target_volume[cand_z,cand_y,cand_x], 3)\n \n neg_pointarrays[neg_pnum,0] = cand_x \n neg_pointarrays[neg_pnum,1] = cand_y\n neg_pointarrays[neg_pnum,2] = cand_z\n \n #X-Y Planes data\n xyimg = vol_data[cand_z,ny_min:ny_max+1,nx_min:nx_max+1]\n \n #X-Z Planes data\n xzimg = vol_data[nz_min:nz_max+1,cand_y,nx_min:nx_max+1]\n \n #Y-Z Planes data\n yzimg = vol_data[nz_min:nz_max+1,ny_min:ny_max+1,cand_x]\n \n# if planeselect == 0:\n# imshow(xyimg,cmap=cm.Greys_r)\n# elif planeselect == 1:\n# imshow(xzimg,cmap=cm.Greys_r)\n# else:\n# imshow(yzimg,cmap=cm.Greys_r)\n# axis('off')\n \n xyimg = xyimg.flatten()\n xzimg = xzimg.flatten()\n yzimg = yzimg.flatten()\n \n xyzimg = zeros((planeimgsize, planenum))\n xyzimg[:,0] = xyimg\n xyzimg[:,1] = xzimg\n xyzimg[:,2] = yzimg\n \n local_neg_pca_input_mtx[neg_pnum,:] = xyzimg.flatten()\n neg_pnum += 1\n \n del xyimg, xzimg, yzimg, xyzimg\n\n # truncate empty array\n local_neg_pca_input_mtx = local_neg_pca_input_mtx[range(neg_pnum),:]\n\n return local_neg_pca_input_mtx\n \n \ndef NeuronPCA(testing_files,l_radius):\n\n # local volume radius\n planenum = 3\n planeselect = 0 # default: xy, xz = 1, yz = 2\n\n \n posi_pca_input_mtx = []\n neg_pca_input_mtx = []\n \n for testing_file in testing_files:\n vol_filename = testing_file + \".vol\"\n point_filename = testing_file + \".cel\"\n \n #print vol_filename\n\n vol_data = LoadVolume(vol_filename)\n \n # 2 times scale down!\n d = 2\n vol_data = ndimage.convolve(np.uint16(vol_data), np.ones((d,d,d)))[::d,::d,::d]/8\n \n z_size, y_size, x_size = vol_data.shape\n \n #g = gauss_kern3D(r_size=3, k_sigma=0.5)\n #vol_data = ndimage.convolve(np.uint16(vol_data),g)\n\n #print vol_data[0,:,:]\n #plt.imshow(vol_data[0,:,:])\n #plt.gray()\n #plt.show()\n \n #scipy.misc.imsave('test2.png', vol_data[0,:,:])\n \n #SaveVolume(vol_data, 'chul.vol')\n \n pointarrays = OBJToPoints(point_filename)\n \n for point in pointarrays:\n point[0] = math.floor(point[0] * x_size)\n point[1] = math.floor(point[1] * y_size)\n point[2] = math.floor(point[2] * z_size)\n \n vol_pca_input_list, pos_pnum = Generate_PCA_Matrix(vol_data,pointarrays,l_radius,planenum,planeselect)\n\n posi_pca_input_mtx.extend(vol_pca_input_list)\n del vol_pca_input_list\n\n ###############################\n # This generates target volume with gauss templete\n target_volume = MakeTargetVolume(pointarrays, vol_data.shape, l_radius)\n ###############################\n \n # debug for target_volume values\n# g_c = 0\n# g_r = 1\n# for point in pointarrays:\n# x = point[0]\n# y = point[1]\n# z = point[2]\n# if(target_volume[z,y,x] > 0.9):\n# g_c += 1\n# print target_volume[z-g_r:z+g_r+1,y-g_r:y+g_r+1,x-g_r:x+g_r+1]\n# print g_c\n\n # out of cell boundary threshold\n outofbound_thr = 0.1\n \n neg_vol_pca_input_list = Generate_Neg_PCA_Matrix(vol_data,pointarrays,l_radius,pos_pnum,target_volume,outofbound_thr,planenum,planeselect)\n \n neg_pca_input_mtx.extend(neg_vol_pca_input_list)\n \n del neg_vol_pca_input_list\n del target_volume\n del vol_data\n del pointarrays\n\n ###############################\n # Negative_Generate_Points_ND_PCA_Cal(vol_data,pointarrays,pos_pnum,target_volume,l_radius,outofbound_thr)\n ###############################\n\n posi_pca_input_arr = array(posi_pca_input_mtx, 'd')\n del posi_pca_input_mtx\n \n posi_coeff,posi_meanvector = Positive_PCA_Cal(posi_pca_input_arr,l_radius,planenum,planeselect)\n \n neg_pca_input_arr = array(neg_pca_input_mtx, 'd')\n del neg_pca_input_mtx\n \n neg_coeff,neg_meanvector = Negative_PCA_Cal(neg_pca_input_arr,l_radius,planenum,planeselect)\n\n return posi_coeff,posi_meanvector,neg_coeff,neg_meanvector\n\n\n","sub_path":"CellCounting_using_PCA_n_Hadoop/cellneuronpca.py","file_name":"cellneuronpca.py","file_ext":"py","file_size_in_byte":8999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"231416539","text":"import discord\r\nfrom discord.ext import commands\r\nimport praw\r\nfrom threading import Timer\r\nimport datetime\r\nimport json\r\nimport asyncio\r\nimport requests\r\nimport time\r\nimport random\r\nimport os\r\nimport sys\r\n\r\nstartTime = time.time()\r\n\r\nif __name__ == '__main__':\r\n print(\"Tried to run {} as main script, despite it being a cog! Terminating script.\".format(__file__))\r\n time.sleep(2)\r\n sys.exit()\r\n\r\nclass Fun(commands.Cog):\r\n\r\n def __init__(self, client):\r\n self.client = client\r\n\r\n @commands.Cog.listener()\r\n async def on_ready(self):\r\n print('Fun cog loaded successfully.')\r\n\r\n @commands.command(aliases=['8ball'])\r\n async def _8ball(self, ctx, *, question):\r\n \"\"\"Classic magic 8ball command\"\"\"\r\n MagicBallPhrasesList = [\r\n \"As I see it, yes.\",\r\n \"Ask again later.\",\r\n \"Better not tell you now.\",\r\n \"Cannot predict now.\",\r\n \"Concentrate and ask again.\",\r\n \"Dont count on it.\",\r\n \"It is certain.\",\r\n \"It is decidedly so.\",\r\n \"Most likely.\",\r\n \"My reply is no.\",\r\n \"My sources say no.\",\r\n \"Outlook not so good.\",\r\n \"Outlook good.\",\r\n \"Reply hazy, try again.\",\r\n \"Signs point to yes.\",\r\n \"Very doubtful.\",\r\n \"Without a doubt.\",\r\n \"Yes.\",\r\n \"Yes, definitely.\",\r\n \"You may rely on it.\"]\r\n\r\n await ctx.send(\"Your question: {0}\\nThe Magic 8ball's answer: {1}\".format(question, MagicBallPhrasesList[random.randint(0,(len(MagicBallPhrasesList)) - 1)] + \" {0}\".format(ctx.message.author.mention)))\r\n \r\n @_8ball.error\r\n async def _8ball_error(self, ctx, error):\r\n if isinstance(error, commands.MissingRequiredArgument):\r\n await ctx.send(\"You need to give a question for the 8ball.\")\r\n else:\r\n await ctx.send(\"Something went wrong.\")\r\n\r\n @commands.command()\r\n async def coinflip(self, ctx):\r\n if random.randint(0, 1) == 0:\r\n await ctx.send(\"The result was heads!\")\r\n else:\r\n await ctx.send(\"The result was tails!\")\r\n\r\n @commands.command()\r\n async def dadjoke(self, ctx):\r\n \"\"\"Sends a funny dad joke\"\"\"\r\n url = \"https://dad-jokes.p.rapidapi.com/random/joke\"\r\n headers = {\r\n 'x-rapidapi-key': \"XXX\",\r\n 'x-rapidapi-host': \"dad-jokes.p.rapidapi.com\"\r\n }\r\n response = requests.request(\"GET\", url, headers=headers)\r\n parsed_response = json.loads(response.text)\r\n await ctx.send(parsed_response[\"body\"][0][\"setup\"])\r\n await asyncio.sleep(1)\r\n await ctx.send(parsed_response[\"body\"][0][\"punchline\"])\r\n\r\n\r\ndef setup(client):\r\n client.add_cog(Fun(client))","sub_path":"MainScript/PythonBot/cogs/Fun.py","file_name":"Fun.py","file_ext":"py","file_size_in_byte":2712,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"264209866","text":"import random\nm=20\nn=100\na=[]\n#根据题意,输出序列数的大小范围是0-n\nfor i in range(0,100):\n #实际操作中模运算只需将随机数的范围控制到0-n,就可以覆盖模运算的所有结果\n #for i in range(0,100):保证输出的序列是有序的,m=m-1控制20个数子\n if (random.randint(0,100)%n)< m :\n a.append(i)\n m=m-1\n n=n-1\n#print(a,'\\n',len(a))","sub_path":"Chapter 12/Think2.py","file_name":"Think2.py","file_ext":"py","file_size_in_byte":404,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"173064336","text":"class Solution:\n def coinChange(self, coins: List[int], amount: int) -> int:\n #initializing dp matrix\n dp=Matrix = [[0 for x in range(amount+1)] for y in range(len(coins)+1)] \n #assigning 1st column with 0\n for i in range(len(coins)+1):\n dp[i][0]=0\n #assigning first row with inf\n for j in range(1,amount+1):\n dp[0][j]=9999\n #generating a dp matrix\n for i in range(1,len(coins)+1):\n for j in range(1,amount+1):\n if j -1:\r\n\t\ttemp = y[i]\r\n\t\tj = size-1\r\n\t\twhile j > i:\r\n\t\t\ttemp -= s[i][j-i]*x[size-1-j]\r\n\t\t\tj -= 1\r\n\t\tx.append(temp/s[i][0])\r\n\t\ti -= 1\r\n\tx.reverse()\r\n\treturn x\r\n\r\n# Решение системы линейных уравнений методом Гаусса.\r\ndef gaussian_method(matrix, stable=True):\r\n\ti = 0;\r\n\tsize = len(matrix)\r\n\t# Перебор строк.\r\n\twhile i < size:\r\n\t\tif stable:\r\n\t\t\t# Поиск максимального элемента в i-ом столбце.\r\n\t\t\tmaxj = i\r\n\t\t\tj = i+1\r\n\t\t\twhile j < size:\r\n\t\t\t\tif abs(matrix[j][i]) > abs(matrix[maxj][i]):\r\n\t\t\t\t\tmaxj = j\r\n\t\t\t\tj += 1\r\n\t\t\tif matrix[maxj][i] == 0:\r\n\t\t\t\treturn None\r\n\t\t\t# Замена местами строк, что бы [i][i] элемент был максимальным, для устойчивости метода.\r\n\t\t\ttemp = matrix[i]\r\n\t\t\tmatrix[i] = matrix[maxj]\r\n\t\t\tmatrix[maxj] = temp\r\n\t\t# Обнуление матрицы под [i][i] элементом.\r\n\t\tj = i+1\r\n\t\twhile j < size:\r\n\t\t\ta = matrix[j][i]/matrix[i][i]\r\n\t\t\tk = i\r\n\t\t\t# Линейное преобразование строк.\r\n\t\t\twhile k <= size:\r\n\t\t\t\tmatrix[j][k] -= a*matrix[i][k]\r\n\t\t\t\tk += 1\r\n\t\t\tj += 1\r\n\t\ti += 1\r\n\t# Поиск x_n элемент\r\n\tresult = [matrix[size-1][size]/matrix[size-1][size-1]]\r\n\ti = size-2\r\n\t# Поиск x_i элементов\r\n\twhile i > -1:\r\n\t\tj = size-1\r\n\t\twhile j > i:\r\n\t\t\tmatrix[i][size] -= matrix[i][j]*result[size-j-1]\r\n\t\t\tj -= 1\r\n\t\tresult.append(matrix[i][size]/matrix[i][i])\r\n\t\ti -= 1\r\n\tresult.reverse()\r\n\treturn result\r\n\r\n# Решение системы линейных уравнений методом Гаусса.\r\ndef Jacobi_method(matrix, vector, eps):\r\n\tn = len(vector)\r\n\t# Проверка на сходимость метода для этих данных.\r\n\tfor i in range(0, n):\r\n\t\tif (2*fabs(matrix[i][i]) < sum(fabs(x) for x in matrix[i])):\r\n\t\t\treturn None\r\n\t# Вычисления\r\n\txi = [vector[i]/matrix[i][i] for i in range(0, n)]\r\n\txip = [0 for i in range(0, n)]\r\n\twhile True:\r\n\t\tfor i in range(0, n):\r\n\t\t\txip[i] = (vector[i] + matrix[i][i]*xi[i] - sum(x*y for (x, y) in zip(xi, matrix[i])))/matrix[i][i]\r\n\t\tif sqrt(sum((x-y)**2 for (x, y) in zip(xi, xip))) < eps:\r\n\t\t\tbreak\r\n\t\txi = [t for t in xip]\r\n\treturn xip\r\n","sub_path":"sole.py","file_name":"sole.py","file_ext":"py","file_size_in_byte":4685,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"398534200","text":"import random\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom mpl_toolkits import mplot3d\na = np.arange(-10,10,0.2)\nb = np.arange(-10,10,0.2)\nx = np.arange(-10,10,0.2) \ny = 2*x+3\nu=np.random.normal(0,1, 100)\ns= np.std(u)\ny1=y+(s*u)\na,b =np.meshgrid(a,b)\ne= (((y1)-(a*x)+b))**2\ne1=e/100\nfig = plt.figure() \naxes = fig.gca(projection ='3d') \naxes.plot_surface(a, b, e1) \naxes.set_xlabel('a----->')\naxes.set_ylabel('b----->')\naxes.set_zlabel('Error----->')\naxes.set_title(\"Error surface plot\")\naxes.set_xlim(-10,10)\naxes.set_ylim(-10,10)\naxes.set_facecolor(\"orange\")\nplt.tight_layout()\nplt.show() \n","sub_path":"Assignment2/Yogesh Dewangan_204102319/Q2.py","file_name":"Q2.py","file_ext":"py","file_size_in_byte":603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"426662967","text":"from pfaw import Fortnite, Platform, Mode\nimport csv\nfrom tempfile import NamedTemporaryFile\nimport shutil\nimport time\nfrom Fortnite.settings import *\n\nclass FortniteBot:\n def __init__(self, filename=\"Fortnite/stats.csv\"):\n self.file_name = filename\n try:\n self.fortnite = Fortnite(fortnite_token='ZWM2ODRiOGM2ODdmNDc5ZmFkZWEzY2IyYWQ4M2Y1YzY6ZTFmMzFjMjExZjI4NDEzMTg2MjYyZDM3YTEzZmM4NGQ=', launcher_token='MzRhMDJjZjhmNDQxNGUyOWIxNTkyMTg3NmRhMzZmOWE6ZGFhZmJjY2M3Mzc3NDUwMzlkZmZlNTNkOTRmYzc2Y2Y=', password=fortnite_password, email=fortnite_email)\n except Exception as e:\n print(\"Could not connect to fortnite API\")\n\n def player_found(self, username):\n found = True\n try:\n self.fortnite.player(username)\n except Exception:\n return False\n return True\n\n def check_new_wins(self):\n open_file = open(self.file_name, \"r\", encoding='UTF-8')\n reader = csv.reader(open_file, delimiter=\",\")\n players_with_wins = []\n csv_list = list(reader)\n for row in csv_list:\n username = row[0]\n wins = int(row[1])\n \n if self.player_found(username):\n new_wins = (self.new_wins(wins, username) - wins)\n if new_wins > 0:\n players_with_wins.append(username)\n self.update_player(username, new_wins)\n\n return players_with_wins\n \n\n def update_player(self, player, wins):\n filename = self.file_name\n \n read_file = open(filename, \"r\", encoding='UTF-8')\n\n reader = csv.reader(read_file, delimiter=',')\n\n rows = []\n for item in reader:\n if player == item[0]:\n item[1] = int(item[1]) + int(wins)\n rows.append(item)\n \n read_file.close()\n \n write_file = open(filename, \"w\", encoding='UTF-8')\n writer = csv.writer(write_file, lineterminator=\"\\n\")\n writer.writerows(rows)\n\n def new_wins(self, current_wins, username):\n stats = self.fortnite.battle_royale_stats(username, platform=Platform.pc)\n return stats.all.wins\n\n def get_wins(self, username):\n\n if self.player_found(username):\n return self.fortnite.battle_royale_stats(username, platform=Platform.pc).all.wins\n return \"Player not found\"\n \n def get_status(self):\n return self.fortnite.server_status()\n\n #Add a new player to get tracked by FortniteBot\n def add_player(self, username):\n if self.player_found(username):\n wins = self.get_wins(username)\n\n #Prepare the file for append. \n write_file = open(self.file_name, 'a', encoding='UTF-8')\n row = []\n row.append(username)\n row.append(wins)\n writer = csv.writer(write_file, lineterminator='\\n')\n writer.writerow(row)\n\n return username + \" is now beeing tracked!\"\n \n return \"Invalid player\"\n","sub_path":"Fortnite/fortnite.py","file_name":"fortnite.py","file_ext":"py","file_size_in_byte":3024,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"613369723","text":"import os, sys\ncurrentdir = os.path.dirname(os.path.realpath(__file__))\nsys.path.append(currentdir)\n\nfrom Weapons.Weapon import Weapon\n\nclass Stick(Weapon):\n def __init__(self):\n self.name = \"Stick\"\n self.power = 100","sub_path":"Weapons/Magic/Stick.py","file_name":"Stick.py","file_ext":"py","file_size_in_byte":233,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"517563215","text":"import enoki as ek\nimport pytest\nimport numpy as np\n\n# pkgs = [\"enoki.cuda\", \"enoki.cuda.ad\",\n# \"enoki.llvm\", \"enoki.llvm.ad\"]\n\npkgs = [\"enoki.llvm\", \"enoki.llvm.ad\"]\npkgs_ad = [\"enoki.llvm.ad\"]\n\ndef get_class(name):\n \"\"\"Resolve a package+class name into the corresponding type\"\"\"\n if 'cuda' in name:\n if not ek.has_backend(ek.JitBackend.CUDA):\n pytest.skip('CUDA mode is unsupported')\n elif 'llvm' in name:\n if not ek.has_backend(ek.JitBackend.LLVM):\n pytest.skip('LLVM mode is unsupported')\n\n name = name.split('.')\n value = __import__(\".\".join(name[:-1]))\n for item in name[1:]:\n value = getattr(value, item)\n\n return value\n\nclass Checker:\n \"\"\"\n Compares a Tensor indexing operation against a NumPy reference\n and asserts if there is a mismatch.\n \"\"\"\n def __init__(self, shape, tensor_type):\n import numpy as np\n self.shape = shape\n size = np.prod(shape)\n self.array_n = np.arange(size, dtype=np.uint32).reshape(shape)\n self.array_e = tensor_type(ek.arange(tensor_type.Array, size), shape)\n\n def __getitem__(self, args):\n import numpy as np\n ref_n = self.array_n.__getitem__(args)\n ref_e = self.array_e.__getitem__(args)\n assert ref_n.shape == ref_e.shape\n assert np.all(ref_n.ravel() == ref_e.array.numpy())\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test01_slice_index(pkg):\n t = get_class(pkg + \".TensorXu\")\n c = Checker((10,), t)\n c[:]\n c[3]\n c[1:5]\n c[-5]\n\n c = Checker((10, 20), t)\n c[:]\n c[5, 0]\n c[5, 0:2]\n c[:, 5]\n c[5, :]\n c[:, :]\n c[1:3, 2:7:2]\n c[8:2:-1, 7:0:-1]\n c[0:0, 0:0]\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test02_slice_ellipsis(pkg):\n t = get_class(pkg + \".TensorXu\")\n c = Checker((10, 20, 30, 40), t)\n\n c[...]\n c[1, ...]\n c[..., 1]\n c[4, ..., 3]\n c[0, 1:3, ..., 3]\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test03_slice_append_dim(pkg):\n t = get_class(pkg + \".TensorXu\")\n c = Checker((10, 20, 30, 40), t)\n\n c[None]\n c[..., None]\n c[1, None, ...]\n c[..., None, 1, None]\n c[None, 4, ..., 3, None]\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test04_broadcasting(pkg):\n t = get_class(pkg + \".TensorXu\")\n for i in range(1, 5):\n for j in range(1, 5):\n for k in range(1, 5):\n shape = [i, j, k]\n for l in range(len(shape)):\n shape_2 = list(shape)\n shape_2[l] = 1\n array_n1 = np.arange(np.prod(shape), dtype=np.uint32).reshape(shape)\n array_n2 = np.arange(np.prod(shape_2), dtype=np.uint32).reshape(shape_2)\n\n array_e1 = t(ek.arange(t.Index, np.prod(shape)), shape)\n array_e2 = t(ek.arange(t.Index, np.prod(shape_2)), shape_2)\n\n out_n = array_n1 + array_n2\n out_e = array_e1 + array_e2\n\n assert out_n.shape == out_e.shape\n assert np.all(out_n.ravel() == out_e.array.numpy())\n assert np.all((array_n1 * 2).ravel() == (array_e1 * 2).array.numpy())\n\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test05_initialization_casting(pkg):\n tu = get_class(pkg + \".TensorXu\")\n tf = get_class(pkg + \".TensorXf\")\n tf64 = get_class(pkg + \".TensorXf\")\n\n t0 = ek.full(tu, 1, (2, 3, 4))\n t1 = ek.full(tf, 2, (2, 3, 4))\n t2 = ek.zero(tf64, (2, 3, 4))\n\n assert ek.shape(t0) == (2, 3, 4)\n\n t3 = t0 + t1 + t2\n assert type(t3) is tf64\n\n assert t3.shape == (2, 3, 4)\n assert t3.array == ek.full(t3.Array, 3, 2*3*4)\n\n t3[:, 1, :] = 12\n assert t3[:, 0, :] == 3\n assert t3[:, 1, :] == 12\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs_ad)\ndef test05_ad(pkg):\n f = get_class(pkg + \".TensorXf\")\n z0 = ek.full(f, 1, (2, 3, 4))\n assert not ek.grad_enabled(z0)\n ek.enable_grad(z0)\n assert ek.grad_enabled(z0)\n assert not ek.grad_enabled(ek.detach(z0))\n assert ek.ravel(z0) is z0.array\n\n z1 = z0 + z0\n ek.backward(z1)\n g = ek.grad(z0)\n assert g.shape == (2, 3, 4)\n assert len(g.array) == 2*3*4\n assert g.array == 2\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test06_numpy_conversion(pkg):\n f = get_class(pkg + \".TensorXf\")\n\n value = f(ek.arange(f.Array, 2*3*4), (2, 3, 4))\n value_np = value.numpy()\n assert value_np.shape == (2, 3, 4)\n assert np.all(value_np.ravel() == value.array.numpy())\n\n value_2 = f(value_np)\n assert value.shape == value_2.shape\n assert value.array == value_2.array\n\n value_np = np.ones((1,1,1,1))\n value_3 = f(value_np)\n assert value_np.shape == value_3.shape\n assert np.all(value_np == value_3.array)\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test07_jax_conversion(pkg):\n jax = pytest.importorskip(\"jax\")\n f = get_class(pkg + \".TensorXf\")\n\n value = f(ek.arange(f.Array, 2*3*4), (2, 3, 4))\n value_jax = value.jax()\n assert value_jax.shape == (2, 3, 4)\n assert jax.numpy.all(value_jax.ravel() == value.array.jax())\n\n value_2 = f(value_jax)\n assert value.shape == value_2.shape\n assert value.array == value_2.array\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test08_pytorch_conversion(pkg):\n torch = pytest.importorskip(\"torch\")\n f = get_class(pkg + \".TensorXf\")\n\n value = f(ek.arange(f.Array, 2*3*4), (2, 3, 4))\n value_torch = value.torch()\n assert value_torch.shape == (2, 3, 4)\n assert torch.all(value_torch.ravel() == value.array.torch())\n\n value_2 = f(value_torch)\n assert value.shape == value_2.shape\n assert value.array == value_2.array\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test09_tensorflow_conversion(pkg):\n tf = pytest.importorskip(\"tensorflow\")\n f = get_class(pkg + \".TensorXf\")\n tf.constant(0)\n\n value = f(ek.arange(f.Array, 2*3*4), (2, 3, 4))\n value_tf = value.tf()\n assert value_tf.shape == (2, 3, 4)\n assert tf.reduce_all(tf.equal(tf.reshape(value_tf, (2*3*4,)), value.array.tf()))\n\n value_2 = f(value_tf)\n assert value.shape == value_2.shape\n assert value.array == value_2.array\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs)\ndef test10_tensorflow_arithmetic(pkg):\n t = get_class(pkg + \".TensorXf\")\n f = get_class(pkg + \".Float32\")\n\n tt = t([1, 2, 3, 4, 5, 6], [2, 3])\n ff = f(2.0)\n\n assert ff * tt == tt * ff\n assert ff * tt == t([2, 4, 6, 8, 10, 12], [2, 3])\n\n\nclass PowerOfTwo(ek.CustomOp):\n def eval(self, value):\n self.value = value\n return value * value\n\n def forward(self):\n grad_in = self.grad_in('value')\n self.set_grad_out(2.0 * self.value * grad_in)\n\n def backward(self):\n grad_out = self.grad_out()\n self.set_grad_in('value', 2.0 * self.value * grad_out)\n\n def name(self):\n return \"power of two\"\n\n\n@pytest.mark.parametrize(\"pkg\", [\"enoki.llvm.ad\", \"enoki.cuda.ad\"])\ndef test11_custom_op(pkg):\n t = get_class(pkg + \".TensorXf\")\n f = get_class(pkg + \".Float32\")\n\n tt = t([1, 2, 3, 4, 5, 6], [2, 3])\n ek.enable_grad(tt)\n\n tt2 = ek.custom(PowerOfTwo, tt)\n\n ek.set_grad(tt2, 1.0)\n ek.enqueue(ek.ADMode.Backward, tt2)\n ek.traverse(f)\n\n assert ek.grad(tt).array == [2.0, 4.0, 6.0, 8.0, 10.0, 12.0]\n\n\n@pytest.mark.parametrize(\"pkg\", pkgs_ad)\ndef test12_select(pkg):\n for tp in [get_class(pkg + \".TensorXf\"), get_class(pkg + \".TensorXu\")]:\n initial = tp([1, 2, 3, 4], shape=(4, 1))\n\n next = initial + 10\n valid = initial >= 2.5\n assert type(valid) == ek.mask_t(initial)\n\n result = ek.select(valid, next, initial)\n assert type(result) == tp\n\n expected = tp([1, 2, 13, 14], shape=ek.shape(initial))\n assert ek.allclose(result, expected)\n","sub_path":"tests/python/test_tensor.py","file_name":"test_tensor.py","file_ext":"py","file_size_in_byte":7712,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"65018485","text":"from Subway.SubwayRide import *\nimport collections\n\nclass BruteForceChooser:\n\n def __init__(self, to_visit):\n self.to_visit = to_visit\n\n def get_to_visit(self):\n return self.to_visit\n\n\nclass LongestPathChooser:\n\n NUM_RUNS = 0\n\n @staticmethod\n def get_route(starting_stop):\n\n print(\"Finding longest route from %s\" % starting_stop)\n ride = SubwayRide(StartingSegment(starting_stop))\n\n return LongestPathChooser.get_longest_path(ride)\n\n @staticmethod\n def get_available_segments(ride):\n\n visited = ride.get_visited_stations()\n cur_stop = ride.get_current_stop()\n\n rides = [seg for seg in cur_stop.get_ride_segments() if seg.get_to_station() not in visited]\n station_trans = [seg for seg in cur_stop.get_station_transfer_segments() if seg.get_to_station() not in visited]\n stop_trans = cur_stop.get_stop_transfer_segments()\n\n return rides, station_trans, stop_trans\n\n @staticmethod\n def get_longest_path(ride):\n\n rides, station_trans, stop_trans = LongestPathChooser.get_available_segments(ride)\n\n # Keep riding if its our only option\n while len(rides) == 1 and ride.get_current_station().is_passthrough():\n ride.add_segment(rides[0])\n rides, station_trans, stop_trans = LongestPathChooser.get_available_segments(ride)\n\n available = rides.copy()\n\n if not ride.just_transferred() and not ride.is_beginning():\n available += station_trans\n available += stop_trans\n\n # If we can't go anywhere\n if len(available) == 0:\n LongestPathChooser.NUM_RUNS += 1\n\n if LongestPathChooser.NUM_RUNS % 100000 == 0:\n print(\"Run %d\" % LongestPathChooser.NUM_RUNS )\n\n return ride.add_segment(EndingSegment(ride.get_current_stop()))\n\n lengths = []\n\n for seg in available:\n new_ride = ride.clone()\n new_ride.add_segment(seg)\n\n longest = LongestPathChooser.get_longest_path(new_ride)\n lengths.append(longest)\n\n # Sort by ride length\n lengths.sort(key=lambda x: x.get_num_stations(), reverse=True)\n\n return lengths[0]\n\n\nclass ShortestPathChooser:\n\n def __init__(self, length_limit):\n super().__init__()\n self.length_limit = length_limit\n\n def reset_limit(self, limit):\n self.length_limit = limit\n\n def get_limit(self):\n return self.length_limit\n\n def get_route(self, starting_stop, to_visit):\n\n print(\"Finding shortest route from %s\" % starting_stop)\n ride = SubwayRide(StartingSegment(starting_stop))\n\n return self.get_shortest_path(ride, to_visit)\n\n @staticmethod\n def distance_to_unvisited(cur_station, unvisited):\n return min([cur_station.get_distance_segments(station) for station in unvisited])\n\n @staticmethod\n def get_available_segments(ride, unvisited):\n\n cur_station = ride.get_current_station()\n cur_stop = ride.get_current_stop()\n segments = ride.get_segments()\n\n # Distance to closest unvisited station\n cur_dist = ShortestPathChooser.distance_to_unvisited(cur_station, unvisited)\n\n rides = []\n\n for ride in cur_stop.get_ride_segments():\n if ride not in segments:\n dist = ShortestPathChooser.distance_to_unvisited(ride.get_to_station(), unvisited)\n\n if dist < cur_dist:\n rides.append(ride)\n\n station_trans = []\n\n for tran in cur_stop.get_station_transfer_segments():\n if tran not in segments:\n dist = ShortestPathChooser.distance_to_unvisited(tran.get_to_station(), unvisited)\n\n if dist < cur_dist:\n station_trans.append(tran)\n\n stop_trans = cur_stop.get_stop_transfer_segments()\n\n return rides, station_trans, stop_trans\n\n def get_shortest_path(self, ride, to_visit):\n\n # We've exceeded our threshold\n if ride.get_length() > self.get_limit():\n return ride.add_segment(ErrorSegment(ride.get_current_stop()))\n\n # We've visited every station\n visited = ride.get_visited_stations()\n unvisited = to_visit - visited\n\n if len(unvisited) == 0:\n path_len = ride.get_length()\n print(\"Found path: %d\" % path_len)\n\n if path_len < self.get_limit():\n self.reset_limit(path_len)\n\n return ride.add_segment(EndingSegment(ride.get_current_stop()))\n\n # Get available segments (ride, station transfer and stop transfer)\n rides, station_trans, stop_trans = self.get_available_segments(ride, unvisited)\n\n available = rides.copy()\n\n if not ride.just_transferred() and not ride.is_beginning():\n available.extend(station_trans)\n available.extend(stop_trans)\n\n # If we can't go anywhere\n if len(available) == 0:\n return ride.add_segment(ErrorSegment(ride.get_current_stop()))\n\n lengths = []\n\n for seg in available:\n new_ride = ride.clone()\n new_ride.add_segment(seg)\n\n longest = self.get_shortest_path(new_ride, to_visit)\n\n if longest is not None and not longest.is_error():\n lengths.append(longest)\n\n # Sort by ride length\n lengths.sort(key=lambda x: x.get_length())\n\n if len(lengths) == 0:\n return None\n else:\n return lengths[0]\n\n\nclass ShortestPathChooser2:\n\n def __init__(self, length_limit):\n super().__init__()\n self.length_limit = length_limit\n\n def reset_limit(self, limit):\n print(\"Resetting depth limit to %d\" % limit)\n self.length_limit = limit\n\n def get_limit(self):\n return self.length_limit\n\n def get_route(self, starting_stop, to_visit):\n\n print(\"Finding shortest route from %s\" % starting_stop)\n ride = SubwayRide(StartingSegment(starting_stop))\n\n return self.get_shortest_path(ride, to_visit)\n\n @staticmethod\n def get_available_segments(ride, unvisited):\n\n cur_station = ride.get_current_station()\n cur_stop = ride.get_current_stop()\n\n # Distance to closest unvisited station\n cur_dist = cur_station.get_distance_segments(unvisited)\n\n rides = []\n\n for ride in cur_stop.get_ride_segments():\n dist = ride.get_to_station().get_distance_segments(unvisited)\n\n if dist < cur_dist:\n rides.append(ride)\n\n station_trans = []\n\n for tran in cur_stop.get_station_transfer_segments():\n dist = tran.get_to_station().get_distance_segments(unvisited)\n\n if dist < cur_dist:\n station_trans.append(tran)\n\n stop_trans = cur_stop.get_stop_transfer_segments()\n\n return rides, station_trans, stop_trans\n\n def get_shortest_path(self, starting_ride, to_visit):\n\n queue = collections.deque()\n queue.append(starting_ride)\n\n shortest = {}\n\n # Keep processing as long as there are paths in the queue\n while queue:\n ride = queue.popleft()\n\n # Skip this ride if its longer than our limit\n if ride.get_length() > self.get_limit():\n continue\n\n visited = ride.get_visited_stations()\n unvisited = to_visit - visited\n\n if len(unvisited) == 0:\n self.reset_limit(ride.get_length())\n return ride.add_segment(EndingSegment(ride.get_current_stop()))\n\n # Get available segments (ride, station transfer and stop transfer)\n rides, station_trans, stop_trans = self.get_available_segments(ride, unvisited)\n\n available = rides.copy()\n\n if not ride.just_transferred() and not ride.is_beginning():\n available.extend(station_trans)\n available.extend(stop_trans)\n\n # We can't go anywhere\n if len(available) == 0:\n continue\n\n for seg in available:\n\n new_ride = ride.clone()\n new_ride.add_segment(seg)\n\n to_stop = seg.get_to_stop()\n\n visited = new_ride.get_visited_stations() & to_visit\n num_visited = len(visited)\n\n shortest.setdefault(to_stop, 0)\n\n if num_visited > shortest[to_stop]:\n queue.append(new_ride)\n shortest[to_stop] = num_visited\n\n if len(queue) % 10000 == 0:\n print(\"Queue len: %d\" % len(queue))\n print(\"Visited: %d\" % num_visited)\n\n print(\"Queue is empty\")\n return None\n","sub_path":"Subway/RouteChoosers/BruteForceChooser.py","file_name":"BruteForceChooser.py","file_ext":"py","file_size_in_byte":8733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"321443280","text":"import tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\n\n# load mnist\nmnist = input_data.read_data_sets(\"MNIST_data/\", one_hot=True)\n\n# input for data\nx = tf.placeholder(tf.float32, [None, 784]) # None here means any arbitrary length!\nW = tf.Variable(tf.zeros([784, 10])) # Doesn't matter what they are so, initialize with zero\nb = tf.Variable(tf.zeros([10]))\n\ny = tf.matmul(x, W) + b # Notice softmax is from tf.nn\n\ny_ = tf.placeholder(tf.float32, [None, 10]) # Label placeholder\n#cross_entropy = tf.reduce_mean(-tf.reduce_sum(y_ * tf.log(y), reduction_indices=[1]))\n\n# or more stable version of implementation\ncross_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=y_, logits=y))\n\n# optimize\ntrain_step = tf.train.GradientDescentOptimizer(0.5).minimize(cross_entropy)\n\nsess = tf.InteractiveSession()\ntf.global_variables_initializer().run() # Initialize variables!\n\nfor _ in range(1000):\n batch_xs, batch_ys = mnist.train.next_batch(100) # Get next batch of 100\n sess.run(train_step, feed_dict={x:batch_xs, y_:batch_ys}) # Train with the new batch\n\ncorrect_predictions = tf.equal(tf.argmax(y,1), tf.argmax(y_,1)) # Prediction accuracy calculation\naccuracy = tf.reduce_mean(tf.cast(correct_predictions, tf.float32)) # note that correct_prediction is bool\n\nprint (sess.run(accuracy, feed_dict={x: mnist.test.images, y_: mnist.test.labels}))\n","sub_path":"tensorflow/soft_mnist.py","file_name":"soft_mnist.py","file_ext":"py","file_size_in_byte":1392,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"322230728","text":"import numpy as np\r\nfrom collections import Counter\r\nimport time\r\nimport warnings\r\n\r\nwarnings.filterwarnings(\"ignore\")\r\n\r\nminNumSample = 10\r\n\r\n\r\nclass BinaryTree:\r\n \"\"\"An Special BinaryTree.\r\n\r\n Construct a special binary tree, store the data in the nodes of the tree,\r\n node labels, left and right subtree positions\r\n\r\n\r\n \"\"\"\r\n\r\n def __init__(self, labels=np.array([]), datas=np.array([])):\r\n self.label = labels\r\n self.data = datas\r\n self.leftChild = None\r\n self.rightChild = None\r\n\r\n def set_rightChild(self, rightObj):\r\n self.rightChild = rightObj\r\n\r\n def set_leftChild(self, leftObj):\r\n self.leftChild = leftObj\r\n\r\n def get_rightChild(self):\r\n return self.rightChild\r\n\r\n def get_leftChild(self):\r\n return self.leftChild\r\n\r\n def get_data(self):\r\n return self.data\r\n\r\n def get_label(self):\r\n return self.label\r\n\r\n\r\ndef RSDS_fun(train_data, tree_num=10):\r\n \"\"\"Handling data noise using completely random forest judgment.\r\n\r\n Establish a tree_num completely random tree. The data label in each leaf node\r\n of the tree is compared with the parent node label to obtain the noise judgment\r\n label of each data in the case of a tree, and all the completely random tree noise\r\n judgment labels are combined to vote to determine the noise data. Denoised data\r\n set after processingEstablish a tree_num completely random tree. The data label\r\n in each leaf node of the tree is compared with the parent node label to obtain\r\n the noise judgment label of each data in the case of a tree, and all the completely\r\n random tree noise judgment labels are combined to vote to determine the noise data.\r\n Denoised data set after processing\r\n\r\n Parameters\r\n ----------\r\n train_data :Numpy type data set.\r\n\r\n tree_num :Total number of random trees.\r\n\r\n \"\"\"\r\n\r\n m, n = train_data.shape\r\n forest = np.array([])\r\n for i in range(tree_num):\r\n tree = CRT(train_data)\r\n visiTree = visitCRT(tree)\r\n visiTree = visiTree[:, np.argsort(visiTree[0, :])]\r\n visiTree = visiTree[1, :]\r\n if forest.size == 0:\r\n forest = visiTree.reshape(m, 1)\r\n else:\r\n forest = np.hstack((forest, visiTree.reshape(m, 1)))\r\n noiseForest = np.sum(forest, axis=1)\r\n nn = 0.5 * tree_num\r\n noiseForest = np.array(list(map(lambda x: 1 if x >= nn or x == 0 else 0, noiseForest)))\r\n denoiseTraindata = deleteNoiseData(train_data, noiseForest)\r\n return denoiseTraindata\r\n\r\n\r\ndef CRT(data):\r\n \"\"\"Build A Completely Random Tree.\r\n\r\n Add a column at the end of the data, store the initial sequence\r\n number of each piece of data, call the function ‘generateTree’\r\n spanning tree\r\n\r\n Parameters\r\n ----------\r\n data :Numpy type data set\r\n\r\n \"\"\"\r\n numberSample = data.shape[0]\r\n orderAttribute = np.arange(numberSample).reshape(numberSample, 1) # (862, 1)\r\n data = np.hstack((data, orderAttribute))\r\n completeRandomTree = generateTree(data)\r\n return completeRandomTree\r\n\r\n\r\ndef generateTree(data, uplabels=[]):\r\n \"\"\"Iteratively Generating A Completely Random Tree.\r\n\r\n Complete random tree by random partitioning of random attributes\r\n\r\n Parameters\r\n ----------\r\n data :Numpy type data set\r\n\r\n uplabels :rootlabel\r\n\r\n \"\"\"\r\n try:\r\n numberSample, numberAttribute = data.shape\r\n except ValueError:\r\n numberSample = 1\r\n numberAttribute = data.size\r\n\r\n if numberAttribute == 0:\r\n return None\r\n\r\n numberAttribute = numberAttribute - 2 # Subtract the added serial number and label\r\n\r\n # The category of the current data, also called the node category\r\n labelNumKey = [] # todo\r\n if numberSample == 1: # Only one sample left\r\n labelvalue = data[0][0]\r\n rootdata = data[0][numberAttribute + 1]\r\n else:\r\n labelNum = Counter(data[:, 0])\r\n labelNumKey = list(labelNum.keys()) # Key (label)\r\n labelNumValue = list(labelNum.values()) # Value (quantity)\r\n labelvalue = labelNumKey[labelNumValue.index(max(labelNumValue))] # Vote to find the label\r\n rootdata = data[:, numberAttribute + 1]\r\n rootlabel = np.hstack((labelvalue, uplabels)) # todo\r\n\r\n # Call the class 'BinaryTree', passing in tags and data\r\n CRTree = BinaryTree(rootlabel, rootdata)\r\n '''\r\n The 'rootlabel' and 'rootdata' are obtained above, the 'rootlabel' is a label (derived by voting), \r\n the 'rootdata' is a series of serial numbers, and finally the class BinaryTree is called.\r\n '''\r\n # There are at least two conditions for the tree to stop growing:\r\n # 1 the number of samples is limited;\r\n # 2 the first column is all equal\r\n if numberSample < minNumSample or len(labelNumKey) < 2:\r\n # minNumSample defaults to 10 or only 1 of the label types are left.\r\n return CRTree\r\n else:\r\n maxCycles = 1.5 * numberAttribute # Maximum number of cycles\r\n # maxCycles = 2\r\n i = 0\r\n while True:\r\n # Once a data exception occurs: except for the above two exceptions that\r\n # stop the tree growth condition, that is, the error data, the loop here will not stop.\r\n i += 1\r\n splitAttribute = np.random.randint(1, numberAttribute) # Randomly select a list of attributes\r\n if splitAttribute > 0 and splitAttribute < numberAttribute + 1:\r\n dataSplit = data[:, splitAttribute]\r\n uniquedata = list(set(dataSplit))\r\n if len(uniquedata) > 1:\r\n break\r\n if i > maxCycles: # Tree caused by data anomaly stops growing\r\n return CRTree\r\n sv1 = np.random.choice(uniquedata)\r\n i = 0\r\n while True:\r\n i += 1\r\n sv2 = np.random.choice(uniquedata)\r\n if sv2 != sv1:\r\n break\r\n if i > maxCycles:\r\n return CRTree\r\n splitValue = np.mean([sv1, sv2])\r\n '''\r\n The above randomly selected rows and columns are obtained, and the final 'splitValue' is an average\r\n '''\r\n\r\n # Call split function\r\n leftdata, rightdata = splitData(data, splitAttribute, splitValue)\r\n\r\n # Set the left subtree, the right subtree\r\n CRTree.set_leftChild(generateTree(leftdata, rootlabel))\r\n CRTree.set_rightChild(generateTree(rightdata, rootlabel))\r\n return CRTree\r\n\r\n\r\n'''\r\nreturns a matrix of two rows and N columns, the first row is the index of the sample, \r\nand the second row is the threshold of the label noise.\r\ne.g.\r\n[[ 36. 499. 547. 557. 563. 587.]\r\n [ 0. 0. 0. 0. 0. 0.]]\r\n'''\r\n\r\n\r\ndef visitCRT(tree):\r\n \"\"\"\r\n Traversing the tree to get the relationship between the data and the node label.\r\n\r\n The traversal tree stores the data number and node label stored in each node of the\r\n completely random tree.\r\n\r\n Parameters\r\n ----------\r\n tree :Root node of the tree.\r\n\r\n\r\n \"\"\"\r\n if not tree.get_leftChild() and not tree.get_rightChild(): # If the left and right subtrees are empty\r\n data = tree.get_data() # data is the serial number of the sample\r\n labels = checkLabelSequence(tree.get_label()) # Existing tag sequence\r\n try:\r\n labels = np.zeros(len(data)) + labels\r\n except TypeError:\r\n pass\r\n result = np.vstack((data, labels))\r\n return result\r\n else:\r\n resultLeft = visitCRT(tree.get_leftChild())\r\n resultRight = visitCRT(tree.get_rightChild())\r\n result = np.hstack((resultLeft, resultRight))\r\n return result\r\n\r\n\r\ndef deleteNoiseData(data, noiseOrder):\r\n \"\"\"Delete noise points in the training set.\r\n\r\n Delete the noise points in the training set according to the noise\r\n judgment result of each data in noiseOrder.\r\n\r\n Parameters\r\n ----------\r\n data :Numpy type data set.\r\n\r\n noiseOrder :Determine if each piece of data is a list of noise.\r\n\r\n \"\"\"\r\n m, n = data.shape\r\n data = np.hstack((data, noiseOrder.reshape(m, 1)))\r\n redata = np.array(list(filter(lambda x: x[n] == 0, data[:, ])))\r\n redata = np.delete(redata, n, axis=1)\r\n return redata\r\n\r\n\r\n\"\"\"check whether the label of the parent node and the leaf node are consistent.\"\"\"\r\n\r\n\r\ndef checkLabelSequence(labels):\r\n \"\"\"Check label sequence.\r\n\r\n Check if the leaf node is the same as the parent node.\r\n\r\n Parameters\r\n ----------\r\n labels :label sequence.\r\n\r\n \"\"\"\r\n return 1 if labels[0] != labels[1] else 0\r\n\r\n\r\ndef splitData(data, splitAttribute, splitValue):\r\n \"\"\"Dividing data sets.\r\n\r\n Divide the data into two parts, leftData and rightData, based on the splitValue\r\n of the split attribute column element.\r\n\r\n Parameters\r\n ----------\r\n data:Numpy type data set.\r\n\r\n splitAttribute:Randomly selected attributes when dividing.\r\n\r\n splitValue:Dividing the value obtained by dividing the selected attribute.\r\n \"\"\"\r\n rightData = np.array(list(filter(lambda x: x[splitAttribute] > splitValue, data[:, ])))\r\n leftData = np.array(list(filter(lambda x: x[splitAttribute] <= splitValue, data[:, ])))\r\n return leftData, rightData\r\n","sub_path":"RSDS.py","file_name":"RSDS.py","file_ext":"py","file_size_in_byte":9474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"53557431","text":"import sys; sys.path.append('../')\nimport src as _\nimport unittest\n\nclass TestListsMethodHeadFirst(unittest.TestCase):\n def test_head_first(self):\n self.assertEqual(\n _.head([1, 2, 3, 2]),\n 1\n )\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"test/lists/test_head_first.py","file_name":"test_head_first.py","file_ext":"py","file_size_in_byte":282,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"369786295","text":"\n# %% Importa libraries\nimport matplotlib.pyplot as plt\nimport pandas as pd\nimport numpy as np\nimport geopandas as gpd\nimport os\nfrom shapely.geometry import Point\nimport contextily as ctx\n\n# %% Download Data to used\n# Rivers and streams information\n# https://www.weather.gov/gis/AWIPSShapefiles\n\nfile = os.path.join('../../data', 'rs16my07.shp')\nrivers_us = gpd.read_file(file)\n\n# River Forecast Center Boundaries\n# https://www.weather.gov/gis/RFCBounds\nfile_st = os.path.join('../../data', 'rf12ja05.shp')\nstat_ref = gpd.read_file(file_st)\n\n# Gauges II USGS stream gauge dataset:\n# https://water.usgs.gov/GIS/metadata/usgswrd/XML/gagesII_Sept2011.xml#stdorder\n\nfile_ga = os.path.join('../../data', 'gagesII_9322_sept30_2011.shp')\ngages = gpd.read_file(file_ga)\n\n# %% Check whats inside data files and CRS :(\n# Rivers\n\ntype(rivers_us)\nvar_names = rivers_us.head()\nrivers_us.columns # Variables name\nrivers_us.shape # no. of gages and no. of varaibles\n\nrivers_us.geom_type # geometry\nrivers_us.crs # check our CRS - coordinate reference system\nrivers_us.total_bounds # Check the spatial extent\n\n# %% Gages information\n# Take only AZ\ngages.columns\ngages.STATE.unique()\ngages_AZ = gages[gages['STATE'] == 'AZ']\ngages_AZ.shape\n\n# %% Add specific points\n# UA: 32.22877495, -110.97688412\n# STream gauge: 34.44833333, -111.7891667\npoint_list = np.array([[-111.7891667, 34.44833333]])\npoint_geom = [Point(xy) for xy in point_list]\npoint_df = gpd.GeoDataFrame(point_geom, columns=['geometry'],\n crs=gages_AZ.crs) # project into gages_az GEOMETRY\n# %% Look at one region in rivers\n# Zoom in and just look at AZ/UTAH\nstat_ref.columns\nstat_ref.STATE.unique()\nstat_ref_AZ = stat_ref[stat_ref['STATE'] == 'UT'] # Utah and Arizona\nstat_ref_AZ.shape\ntest = pd.DataFrame(stat_ref['STATE']) # aux to see regions name\n# super CRUCIAL step!!!\n# Project points into stat_ref CRS\npoints_project = gages_AZ.to_crs(stat_ref_AZ.crs)\n\n# %% Plot map :D\n\nfig, ax = plt.subplots(figsize=(10, 10))\nrivers_us.plot(figsize=(10, 10), alpha=0.5, edgecolor='b',\n ax=ax, label='Rivers', zorder=1)\nstat_ref_AZ.boundary.plot(figsize=(10, 10), alpha=0.5, edgecolor='k',\n ax=ax, label='River Forecast Center Boundaries')\npoints_project.plot(column='DRAIN_SQKM', categorical=True,\n legend=False, markersize=45, cmap='OrRd',\n ax=ax, label='Arizona Gages')\npoint_df.plot(ax=ax, color='k', marker='*', markersize=45,\n label='Verde River Gage')\nplt.ylim(ymax=45, ymin=30)\nplt.xlim(xmax=-105, xmin=-120)\nax.set_title('Hydrologic Information')\nax.set_xlabel('Longitude [°]')\nax.set_ylabel('Latitude [°]')\nax.legend()\nctx.add_basemap(ax, crs='EPSG:4326')\nfig.savefig(\"Hydr_map.png\")\n# %% Run above","sub_path":"assignment_10/Fierro_map.py","file_name":"Fierro_map.py","file_ext":"py","file_size_in_byte":2780,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"455419853","text":"import cv2\nimport numpy as np\n\n\ndef decode(Starting_Index, Ghap, Add_a_Value, LengthOfString, ImagePath):\n image = cv2.imread(ImagePath)\n grayimage = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n Row, Column = grayimage.shape\n Counter = 1\n k = Start = SetValue = 0\n DecodedString = \"\"\n for i in range(0, Row, 1):\n for j in range(0, Column, 1):\n if Counter == Starting_Index:\n Start = 1\n if Start == 1:\n if SetValue % (Ghap + 1) == 0:\n if k == LengthOfString:\n k = -1\n break\n Value = grayimage[i, j]\n while Value < Add_a_Value:\n Value = Value + 255\n DecodedString = DecodedString + chr(Value - Add_a_Value)\n k = k + 1\n SetValue = SetValue + 1\n else:\n SetValue = SetValue + 1\n\n Counter = Counter + 1\n if k == -1:\n break\n return DecodedString","sub_path":"Decode/Decoder.py","file_name":"Decoder.py","file_ext":"py","file_size_in_byte":1064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"546388962","text":"# encoding: utf-8\n# 朴素贝叶斯分类 iris, 划分数据集为训练集和测试集\n\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.metrics import accuracy_score\nfrom sklearn import datasets\n\n\n# 1、加载数据集\n# 我们假定sepal length, sepal width, petal length, petal width 4个量独立且服从高斯分布,用贝叶斯分类器建模\niris = datasets.load_iris()\nprint(iris.data)\n'''\narray([[ 5.1, 3.5, 1.4, 0.2],\n [ 4.9, 3. , 1.4, 0.2],\n [ 4.7, 3.2, 1.3, 0.2],\n [ 4.6, 3.1, 1.5, 0.2],\n [ 5. , 3.6, 1.4, 0.2]])\n'''\nprint(iris.target[:5])\n''' [0 0 0 0 0] '''\n\n\n# 2、划分数据集为训练集和测试集\nseed = 2\ntest_size = 0.3 # 7:3\nX_train, X_test, y_train, y_test = train_test_split(iris.data, iris.target, test_size=test_size, \\\n random_state=seed)\n\n\n# 3、训练模型\nmodel = GaussianNB()\nmodel.fit(X_train, y_train)\nprint(model)\n''' GaussianNB(priors=None) '''\n\n\n# 4、预测测试数据集\ny_pred = model.predict(X_test)\n\n\n# 5、评价预测的正确率\naccuracy = accuracy_score(y_test, y_pred)\nprint(\"Naive Bayes Accuracy: %.2f%%\" % (accuracy*100.0))\n''' Naive Bayes Accuracy: 97.78% '''\n\n\n","sub_path":"01_2_train_test_split.py","file_name":"01_2_train_test_split.py","file_ext":"py","file_size_in_byte":1224,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"292456252","text":"'''\nIterative merge sort\nDeveloper: Dr. Syed Saif ur Rahman\nPurpose: Educational\n'''\nimport math #We used log, ciel\n\n#Merges the data to bring them in order\ndef merge(l1, l2, s):\n global mydata\n global mydatalen\n #print(\"merge \" + str(l1) + \" with \" + str(l1 + s) + \" for \" + str(s) + \" elements \")\n l = [None] * (s) #Temporary list to store left half of data\n r = [None] * (s) #Temporary list to store right half of data\n for index in range(s):\n if l1 + index < mydatalen: #check if index is valid\n l[index] = mydata[l1 + index]\n if l2 + index < mydatalen: #check if index is valid\n r[index] = mydata[l2 + index]\n\n #print(\"l \" + str(l))\n #print(\"r \" + str(r))\n i = 0\n j = 0\n for k in range (l1,l2+s):\n if k < mydatalen:\n if l[i] <= r[j] or r[j] is None: # None check for odd case\n mydata[k] = l[i]\n i = i + 1\n if i >= s :\n while j < s :\n k = k + 1\n if k >= mydatalen:break;\n mydata[k] = r[j]\n j = j + 1\n break;\n else:\n mydata[k] = r[j]\n j = j + 1\n if j >= s :\n while i < s :\n k = k + 1\n if k >= mydatalen:break;\n mydata[k] = l[i]\n i = i + 1\n break;\n\n#Test data\n#mydata = [9, 1, 8, 2, 7, 3, 6, 4, 5]\n#mydata = [9, 1, 8, 2, 7, 3, 6, 4, 5,9, 1, 8, 2, 7, 3, 6, 4, 5]\n#mydata = [9, 1, 8, 2, 7, 3, 6, 4, 5, 9, 1, 8, 2, 7, 3, 6, 4, 5, 9, 1, 8, 2, 7, 3, 6, 4, 5]\n#mydata = [-1, 8, -2, 7, -3, 6, -4, 5]\n#mydata = [9, -8, 2, 7, 3, 6, 4, 5, -9, 1, 8, 2, 7, 3, 6, 4, -5]\nmydata = [9, 1, 8, 2, -3, 6, 4, 5, 9, 1, -8, 2, 7, 3, 6, 4, 5, 9, 1, 8, 2, 7, 3, 6, -4, 5]\nmydatalen = len(mydata)\nprint(\"math.log(mydatalen, 2)\", math.log(mydatalen, 2))\n#Depth of imaginary tree\ndepth = int(math.ceil(math.log(mydatalen, 2)))\nprint(\"Tree depth\", depth)\nx = 0\nfor level in range(depth - 1,-1,-1):\n print(\"Tree level\", level)\n sp = 2**level\n print(\"Sub problems\", sp)\n #sps = mydatalen / 2**level #Cant figure out why it is wrong\n sps = 2 ** x\n print(\"Sub problem size\", sps)\n print(\"mydata -> \" + str(mydata))\n for mi in range(0,mydatalen,sps*2):\n merge(mi,mi+sps,sps)\n print(\"mydata -> \" + str(mydata))\n x += 1","sub_path":"lab3/itimerge.py","file_name":"itimerge.py","file_ext":"py","file_size_in_byte":2476,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"121766505","text":"from bs4 import BeautifulSoup\nfrom bs4 import NavigableString\nimport re\nimport requests\nfrom string import ascii_lowercase\n\nfrom terms import Term\n\n\n# object used for scraping needs to on instantiation:\n# 1. have the appropriate url for the index\n# needs capability to:\n# 1. quickly get all terms from all pages of index\n\nclass Scraper:\n\n def __init__(self):\n self.index_url = (\n \"http://imcip.meded.com/integrated/ha/index/index{}.htm\"\n )\n self.lab_url = (\n \"http://imcip.meded.com/integrated/ha/labs/{lab}/{resource}.htm\"\n )\n\n def terms_by_index(self, letter):\n result = []\n for category in self.get_category_soup(\n self.get_index_soup(\n self.index(letter)\n )\n ):\n for term in self.get_terms(category):\n result.append(term)\n return result\n\n def all_terms(self):\n result = []\n for url in self.all_indexes():\n for category in self.get_category_soup(self.get_index_soup(url)):\n for term in self.get_terms(category):\n result.append(term)\n return result\n\n def get_lab_soup(self, url):\n print(\"Requesting from Online Dissector: \" + url)\n res = requests.get(url)\n if res.ok:\n soup = BeautifulSoup(res.content, \"html.parser\")\n return soup.find(\"table\")\n # returns table that contains all the relevant lab information\n # including further links\n else:\n print(\"Error {}: Could not retrieve lab\").format(res.status_code)\n return None\n\n def main_picture(self, soup):\n # get the url for the main picture displayed in the lab step page\n img = soup.find_all(\"img\", attrs={\"alt\": \"Click for larger verison\"})\n img_url = img[\"src\"]\n return \"http://imcip.meded.com/integrated/ha/\" + img_url[6:]\n\n def has_strong_tag(self, tag):\n if tag.find(\"strong\") is None and not tag.has_attr('style'):\n return True\n else:\n return False\n\n def linked_pages(self, soup):\n # get the urls for all other terms with linked lab step pages in the\n # lab step page\n # call main_picture on each of these to get the picture on each page\n # anatomy terms have tag, which distinguishes these links from\n # instructional, procedurally based links\n link_tags = soup.find_all(\"a\")\n term_tags = [\n tag for tag in link_tags if self.has_strong_tag(tag) is True\n ]\n return [tag[\"href\"] for tag in term_tags]\n # doesn't return entire url, just last part\n # use with Term.lab_path() to get full path\n\n def index(self, letter):\n return self.index_url.format(letter)\n\n def all_indexes(self):\n lst = []\n for c in ascii_lowercase:\n lst.append(self.index_url.format(c))\n return lst\n\n def get_index_soup(self, url):\n # returns list of all and

    tags\n print(\"Requesting from Online Dissector: \" + url)\n res = requests.get(url)\n if res.ok:\n soup = BeautifulSoup(res.content, \"html.parser\")\n return soup.find_all([\"b\", \"ul\"])\n else:\n print(\"Error {}: Could not retrieve index\").format(res.status_code)\n print(res.text)\n return None\n\n def get_category_soup(self, index_soup):\n # requires output from get_index_soup()\n # returns generator of 2 item lists containing\n # [b tag, ul tag]\n for i in range(0, len(index_soup), 2):\n yield index_soup[i: i + 2]\n\n def get_category(self, soup):\n # requires single Tag, from generator, indexed by get_terms\n return soup.string\n\n def get_terms(self, soup):\n # returns Term objects of all terms in single category\n # from the index page, corresponding to one element\n # from get_category_soup\n # n = soup[0] + soup[1]\n c = self.get_category(soup[0])\n result = []\n for term_soup in soup[1].find_all(\"li\"):\n t = Term(category=c, soup=term_soup)\n result.append(t)\n return result\n","sub_path":"scraper.py","file_name":"scraper.py","file_ext":"py","file_size_in_byte":4219,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"535114552","text":"#!/usr/bin/env python3\n\n\"\"\"\n@author: xi\n@since: 2018-06-17\n\"\"\"\n\nimport collections\nimport datetime as dt\n\nimport numpy as np\n\nfrom . import ops\n\n\nclass AccCalculator(object):\n\n def __init__(self):\n self._num_hit = 0\n self._num_all = 0\n\n def update(self, label_pred, label_true):\n hit = np.equal(label_pred, label_true)\n hit = np.float32(hit)\n self._num_hit += float(np.sum(hit))\n self._num_all += len(hit)\n\n def reset(self):\n self._num_hit = 0\n self._num_all = 0\n\n @property\n def accuracy(self):\n return self._num_hit / self._num_all if self._num_all > 0 else 0.0\n\n\nclass BiClassCalculator(object):\n\n def __init__(self):\n self._tp = 0\n self._tn = 0\n self._fp = 0\n self._fn = 0\n\n def update(self, label_predict, label_true):\n hit = np.equal(label_predict, label_true)\n hit = np.float32(hit)\n miss = 1.0 - hit\n\n pos = np.float32(label_predict)\n neg = 1.0 - pos\n\n self._tp += np.sum(hit * pos, keepdims=False)\n self._tn += np.sum(hit * neg, keepdims=False)\n self._fp += np.sum(miss * pos, keepdims=False)\n self._fn += np.sum(miss * neg, keepdims=False)\n\n @property\n def precision(self):\n num_pos_pred = self._tp + self._fp\n return self._tp / num_pos_pred if num_pos_pred > 0 else 0.0\n\n @property\n def recall(self):\n num_pos_true = self._tp + self._fn\n return self._tp / num_pos_true if num_pos_true > 0 else 0.0\n\n @property\n def f1(self):\n pre = self.precision\n rec = self.recall\n return 2 * (pre * rec) / (pre + rec)\n\n @property\n def accuracy(self):\n num_hit = self._tp + self._tn\n num_all = self._tp + self._tn + self._fp + self._fn\n return num_hit / num_all if num_all > 0 else 0.0\n\n\ndef call_for_batch(context, slot, data_source):\n \"\"\"\n\n Args:\n context (dict):\n slot (photinia.Step):\n data_source (photinia.BatchSource):\n\n Returns:\n dict[str, any]:\n tuple|list:\n\n \"\"\"\n data_batch = data_source.next()\n if data_batch is None:\n data_batch = data_source.next()\n if data_batch is None:\n raise RuntimeError('Too many \"None\" returned by data source.')\n ret = slot(*data_batch)\n if isinstance(ret, (tuple, list)):\n for i, value in enumerate(ret):\n context[i] = value\n elif isinstance(ret, (dict, collections.OrderedDict)):\n context.update(ret)\n else:\n # Should not be reached, since Slot ALWAYS returns tuple or dict.\n raise RuntimeError('Invalid Slot outputs type.')\n return ret\n\n\ndef call_for_all(context, slot, data_source):\n \"\"\"\n\n Args:\n context (dict):\n slot (photinia.Step):\n data_source (photinia.BatchSource):\n\n Returns:\n dict[str, list]:\n\n \"\"\"\n ret = collections.defaultdict(list)\n while True:\n data_batch = data_source.next()\n if data_batch is None:\n break\n ret = slot(*data_batch)\n if isinstance(ret, (tuple, list)):\n for i, value in enumerate(ret):\n ret[i].append(value)\n elif isinstance(ret, (dict, collections.OrderedDict)):\n for name, value in ret.items():\n ret[name].append(value)\n else:\n # Should not be reached, since Slot ALWAYS returns tuple or dict.\n raise RuntimeError('Invalid Slot outputs type.')\n context.update(ret)\n return ret\n\n\ndef print_log(context, value_names, i=None, n=None, message=None):\n now = dt.datetime.now()\n print(now.strftime('[%Y-%m-%d %H:%M:%S'), end='')\n\n if i is not None:\n if n is not None:\n percentage = '%.2f' % (i / n * 100,)\n print(' %s/%s|%s%%]' % (str(i), str(n), percentage), end='')\n else:\n print(' %s]' % str(i), end='')\n else:\n print(']', end='')\n\n if message is not None:\n print('\\t' + str(message), end='')\n\n values = context[context] if context in context else ()\n if isinstance(values, (tuple, list)):\n for i, name in enumerate(value_names):\n if i < len(values):\n value = values[i]\n print('\\t%s=%f' % (name, value), end='')\n else:\n print('\\t%s=?' % (name,), end='')\n elif isinstance(values, (dict, collections.OrderedDict)):\n for name in value_names:\n if name in values:\n value = values[name]\n print('\\t%s=%f' % (name, value), end='')\n else:\n print('\\t%s=?' % (name,), end='')\n print()\n\n\nclass OptimizerWrapper(object):\n \"\"\"OptimizerWrapper\n \"\"\"\n\n def __init__(self,\n optimizer):\n self._optimizer = optimizer\n\n @property\n def optimizer(self):\n return self._optimizer\n\n def minimize(self, loss, var_list=None):\n pair_list = self._optimizer.compute_gradients(loss, var_list=var_list)\n pair_list = self._process_gradients(pair_list)\n return self._optimizer.apply_gradients(pair_list)\n\n def _process_gradients(self, pair_list):\n raise NotImplementedError\n\n\nclass GradientClipping(OptimizerWrapper):\n \"\"\"GradientClipping\n \"\"\"\n\n def __init__(self, optimizer, max_norm):\n self._max_norm = max_norm\n super(GradientClipping, self).__init__(optimizer)\n\n @property\n def max_norm(self):\n return self._max_norm\n\n def _process_gradients(self, pair_list):\n pair_list, raw_grad, grad = ops.clip_gradient(pair_list, self._max_norm)\n self._raw_grad_norm = raw_grad\n self._grad_norm = grad\n return pair_list\n\n @property\n def raw_grad_norm(self):\n return self._raw_grad_norm\n\n @property\n def grad_norm(self):\n return self._grad_norm\n","sub_path":"photinia/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":5856,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"242354308","text":"import PySimpleGUI as sg\r\n\r\n\r\ndef status_window():\r\n sg.ChangeLookAndFeel('TealMono') # Changes color scheme of window created\r\n form = sg.FlexForm('Status Window', auto_size_text=True, auto_size_buttons=False, grab_anywhere=False,\r\n return_keyboard_events=True)\r\n\r\n # Classes are placeholder names and will change\r\n class_list = ['Warrior', 'Wizard', 'Rogue', 'Sword Mage', 'Mercenary', 'Sorcerer', 'All-Rounder']\r\n\r\n istat = 5 # Initial stat value\r\n # Layout for the status window\r\n layout = [[sg.Text('Name:'), sg.Text('', size=(20, 1), background_color='black', text_color='white', key='cname'),\r\n sg.ReadFormButton('Choose Name')],\r\n [sg.Text('_' * 55)],\r\n [sg.Text('Stats:')],\r\n [sg.Text('Points Remaining'), sg.Text('15', size=(2, 1), key='points')],\r\n [sg.Text('STR:', size=(5, 1)),\r\n sg.Spin([i for i in range(5, 101)], initial_value=istat, key='STR', size=(5, 1), change_submits=True)],\r\n [sg.Text('INT:', size=(5, 1)),\r\n sg.Spin([i for i in range(5, 101)], initial_value=istat, key='INT', size=(5, 1), change_submits=True)],\r\n [sg.Text('DEX:', size=(5, 1)),\r\n sg.Spin([i for i in range(5, 101)], initial_value=istat, key='DEX', size=(5, 1), change_submits=True)],\r\n [sg.Text('Class:', size=(5, 1), font=('Helvetica', 20)),\r\n sg.Text('', size=(13, 1), font=('Helvetica', 20), background_color='black', text_color='white',\r\n justification='center', key='class'),\r\n sg.ReadFormButton('Class Info')],\r\n [sg.ReadFormButton('Reset Stats'), sg.Text(' ' * 51), sg.Exit()]]\r\n\r\n form.Layout(layout)\r\n total_points = 15\r\n cur_points = 15\r\n while True:\r\n button, values = form.Read()\r\n\r\n if button == 'Choose Name': # Button to type in your name for your character\r\n name = sg.PopupGetText('What is your name?')\r\n form.FindElement('cname').Update(name)\r\n\r\n if button is None or button == 'Exit': break # Program ends successfully if 'Quit' is clicked or window is closed\r\n\r\n # When no stat requirements are met:\r\n try:\r\n strength = int(values['STR'])\r\n intel = int(values['INT'])\r\n dex = int(values['DEX'])\r\n # spoints = int(values['points'])\r\n except:\r\n continue\r\n\r\n if all((strength, intel, dex)) < 10: form.FindElement('class').Update('')\r\n\r\n # How skill points remaining is determined (not sure how to stop stats from increasing when spoints = 0)\r\n if 0 <= cur_points < 16:\r\n stat = [strength, intel, dex]\r\n if 14 < sum(stat) <= 30:\r\n spoints = 15 - (sum(stat) - 15)\r\n cur_points = spoints\r\n form.FindElement('points').Update(spoints)\r\n\r\n form.FindElement('STR').Update(new_values=[i for i in range(1, strength + cur_points + 1)])\r\n form.FindElement('DEX').Update(new_values=[i for i in range(1, dex + cur_points + 1)])\r\n form.FindElement('INT').Update(new_values=[i for i in range(1, intel + cur_points + 1)])\r\n\r\n # Classes based on one stat:\r\n if strength >= 10:\r\n form.FindElement('class').Update(class_list[0])\r\n elif intel >= 10:\r\n form.FindElement('class').Update(class_list[\r\n 1]) # Class is displayed in window when the stat requirements are met (stat requirements are placeholder)\r\n elif dex >= 10:\r\n form.FindElement('class').Update(class_list[2])\r\n\r\n # Classes based on two stats:\r\n if strength >= 10 and intel >= 10:\r\n form.FindElement('class').Update(class_list[3])\r\n elif strength >= 10 and dex >= 10:\r\n form.FindElement('class').Update(class_list[4])\r\n elif intel >= 10 and dex >= 10:\r\n form.FindElement('class').Update(class_list[5])\r\n\r\n # Classes based on three stats:\r\n if strength >= 10 and intel >= 10 and dex >= 10:\r\n form.FindElement('class').Update(class_list[6])\r\n # # Button that resets stats back initial values as well as class name\r\n elif button == 'Reset Stats':\r\n form.Fill({'STR': '5', 'INT': '5', 'DEX': '5'})\r\n form.FindElement('points').Update(15)\r\n if all((strength, intel, dex)) < 10: form.FindElement('class').Update('')\r\n\r\n # Class info button\r\n # TODO - Need to change the use of .DisplayText\r\n # Cannot reach inside of the elements to get at their internal values like DisplayText.\r\n # Don't use GUI elements as variables. Update the elements but don't read (at this point in time)\r\n if button == 'Class Info' and form.FindElement('class').DisplayText == '':\r\n sg.Popup('Not a Class:', 'If you see this, go back and obtain a class!')\r\n if button == 'Class Info' and form.FindElement(\r\n 'class').DisplayText == 'Warrior': # Example of a popup display explaining the class when the button is pressed\r\n sg.Popup('Warrior Class:', 'A class that specializes in melee combat') # Placeholder descriptions\r\n\r\nstatus_window()","sub_path":"StatusWindow.py","file_name":"StatusWindow.py","file_ext":"py","file_size_in_byte":5294,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"310028934","text":"import pygame, sys\r\nfrom pygame.locals import *\r\n\r\nWINDOWWIDTH = 400 # Chiều dài cửa sổ\r\nWINDOWHEIGHT = 300 # Chiều cao cửa sổ\r\n\r\nWHITE = (255, 255, 255)\r\nRED = (255, 0, 0)\r\nGREEN = (0, 255, 0)\r\n\r\npygame.init()\r\n\r\n### Xác định FPS ###\r\nFPS = 60\r\nfpsClock = pygame.time.Clock()\r\n\r\nDISPLAYSURF = pygame.display.set_mode((WINDOWWIDTH, WINDOWHEIGHT))\r\npygame.display.set_caption('Event')\r\n\r\n\r\nclass Car():\r\n def __init__(self):\r\n self.x = 100 # Vị trí của xe\r\n\r\n ## Tạo surface và thêm hình chiếc xe vào ##\r\n self.surface = pygame.image.load('car.png')\r\n\r\n def draw(self): # Hàm dùng để vẽ xe\r\n DISPLAYSURF.blit(self.surface, (self.x, 100))\r\n\r\n def update(self, moveLeft, moveRight): # Hàm dùng để thay đổi vị trí xe\r\n if moveLeft == True:\r\n self.x -= 2\r\n if moveRight == True:\r\n self.x += 2\r\n\r\n if self.x + 100 > WINDOWWIDTH:\r\n self.x = WINDOWWIDTH - 100\r\n self.x=0\r\n if self.x < 0:\r\n self.x = WINDOWWIDTH - 100\r\n\r\n\r\ncar = Car()\r\nmoveLeft = False\r\nmoveRight = False\r\nwhile True:\r\n for event in pygame.event.get():\r\n if event.type == QUIT:\r\n pygame.quit()\r\n sys.exit()\r\n\r\n if event.type == KEYDOWN:\r\n if event.key == K_LEFT:\r\n moveLeft = True\r\n if event.key == K_RIGHT:\r\n moveRight = True\r\n\r\n if event.type == KEYUP:\r\n if event.key == K_LEFT:\r\n moveLeft = False\r\n if event.key == K_RIGHT:\r\n moveRight = False\r\n\r\n DISPLAYSURF.fill(WHITE)\r\n\r\n car.draw()\r\n car.update(moveLeft, moveRight)\r\n\r\n pygame.display.update()\r\n fpsClock.tick(FPS)\r\n","sub_path":"code/ki_1/python/pygame/dieu khien o to/dieu khien o to.py","file_name":"dieu khien o to.py","file_ext":"py","file_size_in_byte":1770,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"265739918","text":"\"\"\"\r\nID:ayush02\r\nLANG:PYTHON3\r\nTASK:whereamii\r\n\"\"\"\r\n#Import modules\r\nimport itertools\r\nimport re\r\n\r\n#C:/Users/ayush/OneDrive/Desktop/USACO/whereami/whereami.in\r\nwith open('C:/Users/ayush/OneDrive/Desktop/USACO/whereami/whereami.in', 'r') as fin:\r\n n = int(fin.readline())\r\n sequence = ''\r\n sequence += fin.readline().strip()\r\n\r\n#Given a list, determine if it is valid based on if it is all one's\r\ndef is_valid(occurance_list):\r\n valid = True\r\n setone = set(occurance_list)\r\n for i in setone:\r\n if i > 1:\r\n valid = False\r\n break\r\n return valid\r\n\r\n#Extra Test Cases\r\nmain = []\r\nfor k in range(n):\r\n occurances = []\r\n for i in range(0,n-k):\r\n occurances.append(sequence.count(sequence[i:i+k]))\r\n main.append(occurances)\r\n\r\nanswer = 0\r\nfor sublist in main:\r\n if is_valid(sublist) == True:\r\n break\r\n else:\r\n answer += 1\r\n'''\r\nwith open('whereami.out', 'w') as fout:\r\n fout.write(str(answer) + '\\n')\r\n'''\r\n\r\nprint(answer)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n'''\r\nATTEMPT ONE\r\n#Characters list for the reduced permutation time\r\n#Total sequence list\r\n\r\n\r\n#ALL substring occurances in a string\r\ndef find_unique(substring, sequence):\r\n matches = [i.start() for i in re.finditer(substring, sequence)]\r\n return len(matches)\r\n\r\n\r\n#For each value in range of the characters length\r\n\r\noccurances = []\r\nfor i in range(1, len(characters)+1):\r\n #Initialize the amount of occurances of each combination in the sequence\r\n mailbox_permutations = itertools.permutations(characters, i)\r\n #print(i, [i for i in mailbox_permutations])\r\n occurances.append([find_unique(''.join(k), sequence) for k in mailbox_permutations])\r\n\r\ndumplist = []\r\nfor sublist in occurances:\r\n for i in sublist:\r\n if i > 1:\r\n dumplist.append(sublist)\r\n\r\nanswer = 0\r\nfor sublist in dumplist:\r\n if sublist in occurances:\r\n answer += 1\r\n occurances.remove(sublist)\r\n\r\nprint(answer)\r\n'''\r\n","sub_path":"whereami/wherami.py","file_name":"wherami.py","file_ext":"py","file_size_in_byte":1983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"312151302","text":"import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\nimport pickle\nimport sys\n\n# picam = cv2.imread('checkers.jpg')[:,::-1,:]\n# lepton = cv2.imread('checkers-h2.jpg')[:,::-1,:]\n\n# picam = cv2.imread('egg.jpg')[:,::-1,:]\n# lepton = cv2.imread('egg-heat.jpg')\n# lepton = np.rot90(lepton) * 100\n# lepton = lepton[:,::-1]\n\ntry:\n i = int(sys.argv[1])\nexcept:\n i=0\n\npicam = pickle.load(open('images/rgb%d.pkl'%i,'rb'))[...,::-1]\nlepton = pickle.load(open('images/therm%d.pkl'%i,'rb'))/100\n\nret = pickle.load(open('persp_mat.p','rb'))\n\n# print(picam.shape, picam.dtype)\n# print(lepton[5,5])\n\na = np.zeros((8,2))\ni=0\ndef onclick(event):\n global i\n a[i] = event.xdata, event.ydata\n i += 1\n print('xdata=%f, ydata=%f' %\n (event.xdata, event.ydata))\n\n# fig = plt.figure()\n# cid = fig.canvas.mpl_connect('button_press_event', onclick)\n\n# plt.subplot(121)\n# plt.imshow(picam)\n# plt.subplot(122)\n# plt.imshow(lepton)\n# plt.show()\n\n# a = a.reshape(2,4,2).astype(np.float32)\n# print(a)\n\n# ret = cv2.getPerspectiveTransform(a[1],a[0])\n# print(ret)\n# print('SAVED.')\n# pickle.dump(ret, open('persp_mat.p','wb'))\n\n\n# print(a)\n\nr, c, _ = picam.shape\nwarp = cv2.warpPerspective(lepton, ret, dsize=(c,r))\n\nplt.subplot(121)\nplt.imshow(picam)\nplt.subplot(122)\nplt.imshow(warp)\n\nplt.figure()\nplt.imshow(warp, alpha=.7)\nplt.imshow(picam, alpha=.6)\n\nplt.show()\n\n\n\n\n","sub_path":"microwave/align_images.py","file_name":"align_images.py","file_ext":"py","file_size_in_byte":1401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"492942717","text":"\"\"\"\nDescription\nGiven n nodes in a graph labeled from 1 to n. There is no edges in the graph at beginning.\n\nYou need to support the following method:\n\nconnect(a, b), an edge to connect node a and node b\nquery(), Returns the number of connected component in the graph\n\nExample\n5 // n = 5\nquery() return 5\nconnect(1, 2)\nquery() return 4\nconnect(2, 4)\nquery() return 3\nconnect(1, 4)\nquery() return 3\n\"\"\"\n\nclass ConnectingGraph3:\n \"\"\"\n @param: n: An integer\n \"\"\"\n def __init__(self, n):\n self.father = {}\n self.size = n\n for i in range(1, n + 1):\n self.father[i] = i\n \n def find(self, a):\n path = []\n while a != self.father[a]:\n path.append(a)\n a = self.father[a]\n \n for node in path:\n self.father[node] = a\n \n return a\n\n \"\"\"\n @param: a: An integer\n @param: b: An integer\n @return: nothing\n \"\"\"\n def connect(self, a, b):\n root_a = self.find(a)\n root_b = self.find(b)\n if root_a != root_b:\n self.father[root_a] = root_b\n self.size -= 1\n\n \"\"\"\n @return: An integer\n \"\"\"\n def query(self):\n return self.size\n","sub_path":"Data Structure 1/Connecting Graph iii.py","file_name":"Connecting Graph iii.py","file_ext":"py","file_size_in_byte":1204,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"590888624","text":"# Copyright (C) 2011 Statoil ASA, Norway. \n# \n# The file 'observations.py' is part of ERT - Ensemble based Reservoir Tool. \n# \n# ERT is free software: you can redistribute it and/or modify \n# it under the terms of the GNU General Public License as published by \n# the Free Software Foundation, either version 3 of the License, or \n# (at your option) any later version. \n# \n# ERT is distributed in the hope that it will be useful, but WITHOUT ANY \n# WARRANTY; without even the implied warranty of MERCHANTABILITY or \n# FITNESS FOR A PARTICULAR PURPOSE. \n# \n# See the GNU General Public License at \n# for more details. \n\n\n# ----------------------------------------------------------------------------------------------\n# Observations tab\n# ----------------------------------------------------------------------------------------------\nfrom ert_gui.widgets.combochoice import ComboChoice\nfrom ert_gui.widgets.pathchooser import PathChooser\nfrom ert.ert.enums import history_source_type\nfrom ert_gui.widgets.reloadbutton import ReloadButton\nfrom ert.sched.history import HistoryType\n\ndef createObservationsPage(configPanel, parent):\n configPanel.startPage(\"Observations\")\n\n r = configPanel.addRow(ComboChoice(parent, history_source_type.values(), \"History source\", \"config/observations/history_source\"))\n\n def get_history_source(ert):\n history_source = ert.main.model_config.get_history_source\n return history_source_type.resolveValue(history_source.get_source_string)\n\n r.initialize = get_history_source\n r.getter = get_history_source\n\n def set_history_source(ert, value):\n history_source_enum = history_source_type.resolveName(str(value))\n sched_file = ert.main.ecl_config.get_sched_file\n refcase = ert.main.ecl_config.get_refcase\n if history_source_enum.value() == 0:\n history = HistoryType.alloc_from_sched_file(sched_file)\n if history_source_enum.value() == 1:\n history = HistoryType.alloc_from_refcase(refcase, True)\n if history_source_enum.value() == 2: \n history = HistoryType.alloc_from_refcase(refcase, False)\n ert.main.model_config.set_history_source(history, sched_file, refcase)\n \n r.setter = set_history_source\n\n \n r = configPanel.addRow(PathChooser(parent, \"Observations config\", \"config/observations/obs_config\", True))\n\n def get_obs(ert):\n obs = ert.main.get_obs\n return obs.get_config_file\n\n r.initialize = get_obs\n r.getter = get_obs\n\n\n def set_obs(ert, value):\n ert.main.load_obs( str(value))\n r.setter = set_obs\n\n\n r = configPanel.addRow(ReloadButton(parent, \"Reload Observations\", \"config/observations/reload_observation\", \"Reload\"))\n r.initialize = lambda ert : ert.main.reload_obs\n r.getter = lambda ert : ert.main.reload_obs\n \n\n configPanel.endPage()\n","sub_path":"devel/python/python/ert_gui/pages/config/observations.py","file_name":"observations.py","file_ext":"py","file_size_in_byte":2914,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"374956370","text":"from django.shortcuts import render\nfrom rest_framework.decorators import action\nfrom rest_framework.response import Response\nfrom rest_framework import viewsets, status\nfrom django.http import HttpResponse\nfrom drf_yasg.utils import swagger_auto_schema\nfrom django.shortcuts import redirect\nfrom django.urls import reverse\n\n\nimport json\nfrom pprint import pprint\nimport logging\nimport sqlite3\n\n# 取得縮網址\nfrom shortUrl.models import getShortUrl\nfrom shortUrl.serializers import ShortUrlSerializer\n\n\n# 取得原來網址\nfrom shortUrl.models import getLongUrl\n\n\n\n\n\nclass ShortUrlViewset(viewsets.ViewSet):\n \n \n @swagger_auto_schema(\n operation_summary='取得縮網址',\n request_body=ShortUrlSerializer,\n \n )\n @action(detail=False,methods=['post'],url_path='get_short_url')\n def short_url(self, request):\n '''\n # INPUT\n\n\n\n ```json\n { \n \"url\":\"http://www.google.com\" \n }\n ```\n\n\n\n # OUTPUT\n\n\n\n ```json\n {\n \"status\": 200,\n \"msg\": \"Success\",\n \"url\": \"http://www.google.com\",\n \"shorturl\": \"http://192.168.50.106:8000/shortUrl/get_short_url/MDOOM6ZO\"\n }\n ```\n \n\n '''\n \n url = request.data.get('url')\n \n \n \n res = getShortUrl(url=url)\n \n return Response(res, status=status.HTTP_200_OK)\n\n\n\n \n \n \n \n \n \n \n\n\n\n @swagger_auto_schema(\n operation_summary='導向原網址'\n )\n @action(detail=False,methods=['get'],url_path='get_short_url/(\\w+)')\n def original_url(self, request,url):\n\n try:\n \n # 建立資料庫\n conn = sqlite3.connect('mydatabase.db')\n c = conn.cursor()\n\n\n # select\n sql_select = (\"select id,url from shorturl where id = '{}'\").format(url)\n c.execute(sql_select)\n res = c.fetchall()\n res_json = [list(row) for row in res]\n\n return redirect(res_json[0][1])\n\n\n except Exception as e:\n \n error_class = e.__class__.__name__ #取得錯誤類型\n detail = e.args[0] #取得詳細內容\n cl, exc, tb = sys.exc_info() #取得Call Stack\n lastCallStack = traceback.extract_tb(tb)[-1] #取得Call Stack的最後一筆資料\n fileName = lastCallStack[0] #取得發生的檔案名稱\n lineNum = lastCallStack[1] #取得發生的行號\n funcName = lastCallStack[2] #取得發生的函數名稱\n\n res_json = { \n \"debug\":str(lineNum)+\"L,filename:\"+fileName+\",\"+str(detail),\n \"status\":500,\n \"detail\":str(e),\n \"msg\":\"failed\"\n }\n\n return res_json\n\n\n\n\n\n\n\n\n","sub_path":"shortUrl/.ipynb_checkpoints/views-checkpoint.py","file_name":"views-checkpoint.py","file_ext":"py","file_size_in_byte":2865,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"200338875","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nThis example can be found in Automate the Boring Stuff\n\"\"\"\n\nimport re\nphoneNumRegex = re.compile(r'\\d\\d\\d-\\d\\d\\d-\\d\\d\\d\\d')\nmo = phoneNumRegex.search('My number is 415-555-4242.')\n\nprint('Phone number found: ' + mo.group())\n\n\n#find all numbers in text\nmo = phoneNumRegex.findall('My number is 415-555-4242 or 415-555-4243.')\n\n\n#Grouping with Parentheses\n\nphoneNumRegex = re.compile(r'(\\d\\d\\d)-(\\d\\d\\d-\\d\\d\\d\\d)')\nmo = phoneNumRegex.search('My number is 415-555-4242.')\nmo.group(1)\nmo.group(2)\nmo.group(0)\nmo.groups()\nareaCode, mainNumber = mo.groups()\n\nprint(areaCode)\nprint(mainNumber)\n","sub_path":"PythonFiles/RegExp.py","file_name":"RegExp.py","file_ext":"py","file_size_in_byte":638,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"373255913","text":"import socket\nimport threading\n\nHOST, PORT = 'localhost' ,8080\nname = input('type your name: ')\n\nsock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n# sock.send(name)\n\ndef send():\n while True:\n data = input(name + ': ')\n sock.sendall(bytes(name + \": \" + data, 'utf-8'))\n sock.close()\n\ndef receive():\n while True:\n data = str(sock.recv(1024), 'utf-8')\n print(data)\n sock.close()\n\nsock.connect((HOST, PORT))\n# send_thread = threading.Thread(target=send, daemon=True)\n# receive_thread = threading.Thread(target=receive, daemon=True)\n# send_thread.start()\n# receive_thread.start()\nwhile True:\n data = input(name + ': ')\n sock.sendall(bytes(name + \": \" + data, 'utf-8'))\n response = sock.recv(1024)\n print(response.decode('utf-8'))\nsock.close()\n\n","sub_path":"venv/src/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":798,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"247193069","text":"import understand\nimport os,sys,re\n\nudb_path = sys.argv[1]\nname = sys.argv[2]\n\ndb = understand.open(udb_path)\nfiles = db.ents(\"Java File\")\nnew_files = []\nfor f in files:\n\tloc = f.metric(['CountLineCode'])['CountLineCode']\n\tif loc == 0 or loc > 10000:\n\t\tcontinue\n\tnew_files.append(f)\n\nfile_len = len(new_files)\n\nfilenames=[]\nfile_loc = {}\nfor i in range(file_len):\n\tfname = re.findall(name+\".*\\.java\",new_files[i].longname())[0][len(name)+1:]\n\tfilenames.append(fname)\n\tfile_loc[fname] = new_files[i].metric(['CountLineCode'])['CountLineCode']\nfilenames.sort()\n\nf = open(name+'/LOC.csv','w')\nf.write('FILENAME,SLOC'+'\\n')\nfor file_name in filenames:\n\tf.write(file_name+','+str(file_loc[file_name])+'\\n')\nf.close()\n'''\nfile_map={}\nfor i in range(file_len):\n\tfile_map[filenames[i]]=i\n\nadj_matrix=[[0]*file_len for i in range(file_len)]\n\nents = db.ents(\"class ~unresolved ~unknown\")\n\nfor ent in ents:\n\tif ent.ref(\"definein\",\"File\"):\n\t\tthis_file = re.findall(\"src.*\\.java\",ent.ref(\"definein\",\"File\").file().longname())[0]\n\t\tcouples = ent.refs(\"couple\",\"class\")\n\t\tfor cou in couples:\n\t\t\tif cou.ent().ref(\"definein\",\"File\"):\n\t\t\t\tcouple_file = re.findall(\"src.*\\.java\",cou.ent().ref(\"definein\",\"File\").file().longname())[0]\n\t\t\t\tadj_matrix[file_map[couple_file]][file_map[this_file]]=1\n\nf = open('file_depend.txt','w')\nfor i in adj_matrix:\n\tk=' '.join([str(j) for j in i])\n\tf.write(k+'\\n')\nf.close()\n'''\nfor i in filenames:\n\tprint(i)\ndb.close()","sub_path":"file_dep.py","file_name":"file_dep.py","file_ext":"py","file_size_in_byte":1434,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"152484068","text":"import datetime\nfrom google.cloud import firestore\nimport logging.config\n\nlogging.config.fileConfig(\"logging.conf\")\nlogger = logging.getLogger()\n\ndb = firestore.Client(project=\"cf-fs-project\")\n\n\ndef get_log(gcp_job_id):\n \"\"\"\n get execution log from firestore by gcp job id ,eg: big query job id\n Args:\n gcp_job_id: the gcp job id , like bigquery job id\n\n Returns: the tuple (document id, document dict)\n\n \"\"\"\n logger.info(gcp_job_id)\n try:\n logs_ref = db.collection(u'lss_logs')\n # use stream instead of get function\n docs = logs_ref.where(u'gcp_job_id', u'==', gcp_job_id).limit(1).get()\n if len(docs) > 0:\n # only one documents returned because the query limitation`\n for doc in docs:\n logger.info(u' doc {} => {}'.format(doc.id, doc.to_dict()))\n return doc.id, doc.to_dict()\n except Exception as ex:\n logger.error(ex)\n\n\ndef update_log(log_id, doc):\n \"\"\"\n Update job log document: endTime, logs\n Args:\n log_id: the job log document id\n doc:the attributes need to be updated in the document\n\n Returns:0: success , -1: failed, message\n\n \"\"\"\n try:\n code = 0\n message = \"success\"\n if log_id and doc:\n log_ref = db.collection(u'lss_logs').document(log_id)\n if log_ref.get().exists:\n logger.info(\"Update document by id:{} , value:{} \".format(log_id, doc))\n log_ref.update(\n doc\n )\n else:\n code = -1\n message = \"document not found: \" + str(log_id)\n logger.error(\"document not found, type: {}, id: {}\".format(\"lss_logs\", id))\n return code, message\n except Exception as ex:\n logger.error(\"Error happens: \")\n logger.error(ex)\n return -1, \"Errors\" + str(ex)\n\n\ndef insert_log(log={}):\n logger.info(log)\n try:\n if log['job_id'] and log['gcp_job_id']:\n job_ref = db.collection(u'lss_logs').document()\n log['status'] = 'start'\n log['start_time'] = datetime.datetime.now()\n job_ref.set(log, merge=True)\n logger.info(\"add logs into firestore:\")\n logger.info(job_ref.get().to_dict())\n else:\n logger.info(\"required fields are missing\")\n except Exception as e:\n logger.error(\"errors happen when insert document.\")\n logger.error(e)\n\n\ndef get_job(job_id=None):\n if id:\n try:\n doc_ref = db.collection(u'lss_jobs').document(job_id)\n doc = doc_ref.get()\n if doc.exists:\n logger.info(f'Document data: {doc.to_dict()}')\n return doc.to_dict()\n else:\n logger.info(u'No such document!')\n except Exception as e:\n logger.error(e)\n\n\ndef create_job(job={}):\n logger.info(job)\n try:\n if job['job_id'] and job['job_name'] and job['job_type'] and job['group_id']:\n job_ref = db.collection(u'lss_jobs').document(job['job_id'])\n job_ref.set(job, merge=True)\n logger.info(\"add job into firestore:\" + str(job['job_id']))\n else:\n logger.info(\"required fields are missing\")\n except Exception as e:\n logger.error(\"errors happen when insert document.\")\n logger.error(e)\n\n\nif __name__ == '__main__':\n pass\n # job_properties = {\n # u'query': u\"select id,name,age,gender ,'Dalian' as address from cf-fs-project.lss_raw.user_info\",\n # u'write_disposition': u'WRITE_APPEND',\n # u'destination': u'cf-fs-project.lss_insight.user_info_inst',\n # u'job_id_prefix': u'lss_demo_'\n # }\n # job_add = {\n # u'job_id': u'1001',\n # u'group_id': u'01',\n # u'job_name': u'lss_demo_raw2insight',\n # u'job_type': u'BigQuery',\n # u'input': u'cf-fs-project.lss_raw.user_info',\n # u'output': u'cf-fs-project.lss_insight.user_info_inst',\n # u'properties': job_properties\n # }\n # create_job(job_add)\n # res = get_job(u'1001')\n # logger.info(res)\n # job_log = {\n # u'job_id': u'1001',\n # u'gcp_job_id': u'test',\n # u'log_msg': u'job started',\n # }\n #\n # insert_log(job_log)\n # gcp_log_id = 'lss_demo_65563df0-837e-44de-b9fb-d6b57b7fdf95'\n # lss_log = get_log(gcp_log_id)\n # print(lss_log)\n # ('D8fCt9kejyzF5l9HFOGn',\n # {'status': 'start', 'gcp_job_id': 'lss_demo_65563df0-837e-44de-b9fb-d6b57b7fdf95', 'log_msg': 'job started',\n # 'job_id': '1001', 'start_time': DatetimeWithNanoseconds(2020, 11, 20, 17, 56, 49, 825796, tzinfo= < UTC >)})\n # document_id = 'D8fCt9kejyzF5l9HFOGn'\n # job_statics = {\n # \"billingTier\": 1,\n # \"createTime\": \"2020-11-20T09:56:46.047Z\",\n # \"endTime\": \"2020-11-20T09:56:48.465Z\",\n # \"queryOutputRowCount\": \"4\",\n # \"referencedTables\": [\n # {\n # \"datasetId\": \"lss_raw\",\n # \"projectId\": \"cf-fs-project\",\n # \"tableId\": \"user_info\"\n # }\n # ],\n # \"startTime\": \"2020-11-20T09:56:46.356Z\",\n # \"totalBilledBytes\": \"10485760\",\n # \"totalProcessedBytes\": \"100\",\n # \"totalSlotMs\": \"8404\",\n # \"totalTablesProcessed\": 1\n # }\n # end_time = job_statics['endTime']\n # status = \"Done\"\n # update_doc = {\n # \"end_time\": end_time,\n # \"status\": status,\n # \"job_statics\": job_statics\n # }\n # res = update_log(document_id, update_doc)\n # print(res)\n","sub_path":"lss_cloudfunction/update_cf/firestore_client.py","file_name":"firestore_client.py","file_ext":"py","file_size_in_byte":5577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"647317636","text":"\ndata = open('../data/output', 'r')\ntrain = open('train.data', 'w')\ndev = open('dev.data', 'w')\ntest = open('test.data', 'w')\n\n\ncount = 1\n\ntrain_threshold = 2738\ndev_threshold = 3651\n\nfor line in data:\n line = line.strip()\n if line:\n if count <= train_threshold:\n train.write(line+'\\n')\n elif count > train_threshold and count<=dev_threshold:\n dev.write(line+'\\n')\n elif count > dev_threshold:\n test.write(line+'\\n')\n count += 1\n\ndata.close()\ntrain.close()\ntest.close()\ndev.close()\n","sub_path":"tools/split_data.py","file_name":"split_data.py","file_ext":"py","file_size_in_byte":549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"414887632","text":"import calendar\n\nfrom consts.notification_type import NotificationType\nfrom notifications.base_notification import BaseNotification\n\n\nclass ScheduleUpdatedNotification(BaseNotification):\n\n _priority = 'high'\n\n def __init__(self, event, next_match=None):\n from helpers.match_helper import MatchHelper # recursive import issues\n self.event = event\n\n if not next_match:\n upcoming = MatchHelper.upcomingMatches(event.matches, 1)\n self.next_match = upcoming[0] if upcoming and len(upcoming) > 0 else None\n else:\n self.next_match = next_match\n\n @property\n def _type(self):\n return NotificationType.SCHEDULE_UPDATED\n\n def _build_dict(self):\n data = {}\n data['notification_type'] = NotificationType.type_names[self._type]\n data['message_data'] = {}\n data['message_data']['event_key'] = self.event.key_name\n data['message_data']['event_name'] = self.event.name\n if self.next_match and self.next_match.time:\n data['message_data']['first_match_time'] = calendar.timegm(self.next_match.time.utctimetuple())\n else:\n data['message_data']['first_match_time'] = None\n\n return data\n","sub_path":"notifications/schedule_updated.py","file_name":"schedule_updated.py","file_ext":"py","file_size_in_byte":1227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"623568684","text":"class Solution(object):\n def reverseKGroup(self, head, k):\n \"\"\"\n :type head: ListNode\n :type k: int\n :rtype: ListNode\n \"\"\"\n current = head\n swap_nodes = 0\n while current is not None and swap_nodes != k:\n current = current.next\n swap_nodes += 1\n\n if swap_nodes == k:\n current = self.reverseKGroup(current, k)\n\n while swap_nodes > 0:\n head.next, current, head = current, head, head.next\n swap_nodes -= 1\n head = current\n return head\n","sub_path":"a025_reverse_k_nodes_in_groups/reverse_k_nodes_in_groups.py","file_name":"reverse_k_nodes_in_groups.py","file_ext":"py","file_size_in_byte":589,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"467223485","text":"import re\n\n\ndef way_better(filename): # Функция читающая файл\n print('reading file with way_better()')\n try:\n with open(filename) as f:\n return f.read()\n except FileNotFoundError:\n print('File not found')\n\n\ndef count_entry(list1): # Функция подсчитывающая количество вхождений одинаковых записей в список\n count_dict = {}\n count = 0\n value = 0\n\n while count < len(list1):\n if list1[count] not in count_dict.keys():\n for item in list1:\n if item == list1[count]:\n value += 1\n count_dict.update({list1[count]: value})\n count += 1\n value = 0\n return count_dict\n\n\ndef results(dict1): # Функция ввыводящая результат в консоль в виде таблицы\n for k, v in dict1.items():\n print('| ', k, ' | ', v, ' |')\n print()\n\n\nraw_text = way_better('nasa_19950801.tsv') # Читаем нужный файл\n\nname_pattern_group = r'.*\\s(\\d*)\\tGET\\t(.*)\\s(\\d\\d\\d)\\s\\d.*' # Шаблон регулярного выражения\nready_text = re.findall(name_pattern_group, raw_text) # Ищем по шаблону текст\n\n# f = open('nasa_ready_text.txt', 'w')\n# f.write(str(ready_text))\n# f.close()\n\nprint('Number of rows: ', len(ready_text))\n\n# Формируем отдельные списки из полученого\nx = 0\nurl_list = []\ntime_list = []\ncode_list = []\nwhile x < len(ready_text):\n time_list.append(ready_text[x][0])\n url_list.append(ready_text[x][1])\n code_list.append(ready_text[x][2])\n x += 1\n# print(time_list)\n# print(time_list, '\\n', url_list, '\\n', code_list)\n\n# Считаем вхождения и выводим в консоль\ncount_dict_code = count_entry(code_list)\nprint('| Код ошибки | Количество ошибок |')\nresults(count_dict_code)\ncount_dict_time = count_entry(time_list)\nprint('| Timestamp | Количество записей |')\nresults(count_dict_time)\ncount_dict_url = count_entry(url_list)\nprint('| URL | Количество записей |')\nresults(count_dict_url)\n\n\n\n\n\n\n","sub_path":"lesson9_1.py","file_name":"lesson9_1.py","file_ext":"py","file_size_in_byte":2212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"244853743","text":"import csv\nimport tempfile\nimport unittest\n\nimport eva_cttv_pipeline.trait_mapping.output as output\nfrom eva_cttv_pipeline.trait_mapping.oxo import OxOMapping, OxOResult\nfrom eva_cttv_pipeline.trait_mapping.trait import OntologyEntry, Trait\nimport eva_cttv_pipeline.trait_mapping.zooma as zooma\n\n\nclass TestOutputTraitMapping(unittest.TestCase):\n def test_output_trait_mapping(self):\n tempfile_path = tempfile.mkstemp()[1]\n with open(tempfile_path, \"w\", newline='') as mapping_file:\n mapping_writer = csv.writer(mapping_file, delimiter=\"\\t\")\n mapping_writer.writerow([\"#clinvar_trait_name\", \"uri\", \"label\"])\n\n test_trait = Trait('aprt deficiency, japanese type', 11)\n\n # Normally a set, but changed to a list for predictable output order in test\n test_trait.finished_mapping_set = [\n OntologyEntry('http://www.orpha.net/ORDO/Orphanet_976',\n 'Adenine phosphoribosyltransferase deficiency'),\n OntologyEntry('http://www.orpha.net/ORDO/Orphanet_977',\n 'Adenine phosphoribosyltransferase deficiency type A')\n ]\n\n output.output_trait_mapping(test_trait, mapping_writer)\n\n with open(tempfile_path, \"rt\", newline='') as mapping_file:\n mapping_reader = csv.reader(mapping_file, delimiter=\"\\t\")\n next(mapping_reader)\n self.assertEqual(['aprt deficiency, japanese type',\n 'http://www.orpha.net/ORDO/Orphanet_976',\n 'Adenine phosphoribosyltransferase deficiency'],\n next(mapping_reader))\n\n self.assertEqual(['aprt deficiency, japanese type',\n 'http://www.orpha.net/ORDO/Orphanet_977',\n 'Adenine phosphoribosyltransferase deficiency type A'],\n next(mapping_reader))\n\n\nclass TestGetMappingsForCuration(unittest.TestCase):\n def test_get_non_efo_mapping(self):\n \"\"\"If mapping is not in EFO, its `is_current` flag should *not* be checked, and the mapping\n *should* be selected for curation.\"\"\"\n test_zooma_result = zooma.ZoomaResult(['http://purl.obolibrary.org/obo/HP_0001892'],\n 'abnormal bleeding', 'HIGH', 'eva-clinvar')\n mapping = test_zooma_result.mapping_list[0]\n mapping.confidence = zooma.ZoomaConfidence.HIGH\n mapping.in_efo = False\n mapping.is_current = False\n mapping.ontology_label = \"\"\n mapping.source = 'eva-clinvar'\n mapping.uri = 'http://purl.obolibrary.org/obo/HP_0000483'\n self.assertEqual([mapping], output.get_mappings_for_curation([test_zooma_result]))\n\n def test_get_obsolete_efo_mapping(self):\n \"\"\"If mapping is in EFO, but is not current, it *should not* be selected for curation.\"\"\"\n test_zooma_result = zooma.ZoomaResult(['http://www.orpha.net/ORDO/Orphanet_976'],\n 'Adenine phosphoribosyltransferase deficiency',\n 'HIGH', 'eva-clinvar')\n mapping = test_zooma_result.mapping_list[0]\n mapping.confidence = zooma.ZoomaConfidence.HIGH\n mapping.in_efo = True\n mapping.is_current = False\n mapping.ontology_label = \"Adenine phosphoribosyltransferase deficiency\"\n mapping.source = 'eva-clinvar'\n mapping.uri = 'http://www.orpha.net/ORDO/Orphanet_976'\n self.assertEqual([], output.get_mappings_for_curation([test_zooma_result]))\n\n def test_get_current_efo_mapping(self):\n \"\"\"If mapping is in EFO and is current, is *should* be selected for curation.\"\"\"\n test_zooma_result = zooma.ZoomaResult(['http://purl.obolibrary.org/obo/MONDO_0008091'],\n 'Abnormal neutrophil chemotactic response',\n 'MEDIUM', 'eva-clinvar')\n mapping = test_zooma_result.mapping_list[0]\n mapping.confidence = zooma.ZoomaConfidence.HIGH\n mapping.in_efo = True\n mapping.is_current = True\n mapping.ontology_label = \"Abnormal neutrophil chemotactic response\"\n mapping.source = 'eva-clinvar'\n mapping.uri = 'http://purl.obolibrary.org/obo/MONDO_0008091'\n self.assertEqual([mapping], output.get_mappings_for_curation([test_zooma_result]))\n\n\nclass TestOutputForCuration(unittest.TestCase):\n def test_output_for_curation(self):\n tempfile_path = tempfile.mkstemp()[1]\n with open(tempfile_path, \"wt\") as curation_file:\n curation_writer = csv.writer(curation_file, delimiter=\"\\t\")\n\n test_trait = Trait(\"transitional cell carcinoma of the bladder\", 276)\n\n test_oxo_result = OxOResult(\"HP:0006740\", \"Transitional cell carcinoma of the bladder\",\n \"HP:0006740\")\n test_oxo_mapping = OxOMapping(\"bladder transitional cell carcinoma\", \"EFO:0006544\", 2,\n \"HP:0006740\")\n test_oxo_mapping.in_efo = test_oxo_mapping.is_current = True\n test_oxo_mapping.ontology_label = \"bladder transitional cell carcinoma\"\n test_oxo_result.mapping_list = [test_oxo_mapping]\n\n test_trait.oxo_result_list = [test_oxo_result]\n\n output.output_for_curation(test_trait, curation_writer)\n\n with open(tempfile_path, \"rt\") as curation_file:\n curation_reader = csv.reader(curation_file, delimiter=\"\\t\")\n expected_record = [\n \"transitional cell carcinoma of the bladder\", \"276\",\n \"http://www.ebi.ac.uk/efo/EFO_0006544|bladder transitional cell carcinoma|2|HP:0006740|EFO_CURRENT\"\n ]\n self.assertEqual(expected_record, next(curation_reader))\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"tests/trait_mapping/test_output.py","file_name":"test_output.py","file_ext":"py","file_size_in_byte":5947,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"173138140","text":"import numpy as np\nimport cmath\nimport math\nimport matplotlib.pyplot as plt\n\nT = 1\nsigma = 0\nj = complex(0, 1)\nomega = np.arange(0, 2*math.pi, math.pi/50)\nz = np.empty([len(omega), 1], dtype=complex)\nw = np.empty([len(omega), 1], dtype=complex)\n\nfor k in range(len(omega)):\n s = sigma + (j * omega[k])\n z[k] = cmath.exp(T*s)\n w[k] = 2 * (z[k] - 1) / (T * (z[k] + 1))\n\nplt.ylim(-1.5, 1.5)\nplt.xlim(-1.5, 1.5)\nplt.ylabel(\"Imag\")\nplt.xlabel(\"Real\")\nplt.scatter([x.real for x in w], [y.imag for y in w], color='green')\nplt.scatter([x.real for x in z], [y.imag for y in z], color='red')\nplt.show()\n","sub_path":"zcircle.py","file_name":"zcircle.py","file_ext":"py","file_size_in_byte":614,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"453361318","text":"import time\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom main import *\n\nif __name__ == \"__main__\":\n i = 2\n exec_times_fft = []\n N_vals = []\n log_vals = []\n while i <= 1024*1024:\n start_time = time.time()\n N_vals.append(i)\n x = np.random.random(i)\n res = fft(x)\n exec_times_fft.append(time.time() - start_time)\n log_vals.append(0.0000007 * i * np.log(i))\n i *= 2\n # plt.semilogx(N_vals, log_vals)\n # plt.semilogx(N_vals, exec_times_fft)\n plt.plot(N_vals, log_vals)\n plt.plot(N_vals, exec_times_fft)\n plt.plot(N_vals, log_vals)\n plt.legend([\"Execution Time for FFT\", \"0.0000007*n*log(n)\"])\n plt.xlabel(\"Vector Length - N\")\n plt.ylabel(\"Execution Time\")\n plt.show()\n","sub_path":"fft_driver.py","file_name":"fft_driver.py","file_ext":"py","file_size_in_byte":765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"72303195","text":"# coding=utf-8\n__author__ = 'Steven Cutting'\n__author_email__ = 'steven.e.cutting@gmail.com'\n__created_on__ = '10/22/14'\n\nimport sys\n\n# ---------------------------------------------------------------------\n# Yahoo Finance data options cheat sheet (short list)\n\n'''\nhttp://finance.yahoo.com/d/quotes\nhttp://ichart.finance.yahoo.com/table\n\nc6: Change (Realtime)\nk2: Change Percent (Realtime)\nc8: After Hours Change (Realtime)\nk1: Last Trade (Realtime) With Time\nl: Last Trade (With Time)\nl1: Last Trade (Price Only)\nw4: Day’s Value Change (Realtime)\nm2: Day’s Range (Realtime)\n\n**Symbol Info**\nv: More Info\nj1: Market Capitalization\nj3: Market Cap (Realtime)\nf6: Float Shares\nn: Name\nn4: Notes\ns: Symbol\ns1: Shares Owned\nx: Stock Exchange\nj2: Shares Outstanding\n\n**Averages:**\nm5: Change From 200 Day Moving Average\nm6: Percent Change From 200 Day Moving Average\nm7: Change From 50 Day Moving Average\nm8: Percent Change From 50 Day Moving Average\nm3: 50 Day Moving Average\nm4: 200 Day Moving Average\n'''\n\n# ---------------------------------------------------------------------\n# mktdotpy Token dictionaries\n\ndataCodeDict = {' Change (Realtime)': 'c6',\n ' Change Percent (Realtime)': 'k2',\n ' After Hours Change (Realtime)': 'c8',\n ' Last Trade (Realtime) With Time': 'k1',\n ' Last Trade (With Time)': 'l',\n ' Last Trade (Price Only)': 'l1',\n ' Day’s Value Change (Realtime)': 'w4',\n ' Day’s Range (Realtime)': 'm2',\n ' More Info': 'v',\n ' Market Capitalization': 'j1',\n ' Market Cap (Realtime)': 'j3',\n ' Float Shares': 'f6',\n ' Name': 'n',\n ' Notes': 'n4',\n ' Symbol': 's',\n ' Shares Owned': 's1',\n ' Stock Exchange': 'x',\n ' Shares Outstanding': 'j2',\n ' Change From 200 Day Moving Average': 'm5',\n ' Percent Change From 200 Day Moving Average': 'm6',\n ' Change From 50 Day Moving Average': 'm7',\n ' Percent Change From 50 Day Moving Average': 'm8',\n ' 50 Day Moving Average': 'm3',\n ' 200 Day Moving Average': 'm4',\n }\n\n\ntickerDict = {'S&P 500': '^GSPC',\n '^GSPC': '^GSPC',\n 'NASDAQ': '^IXIC',\n '^IXIC': '^IXIC',\n 'Russell 2000': '^RUT',\n '^RUT': '^RUT',\n 'Nikkei 225': '^N225',\n '^N225': '^N225',\n 'Crude Oil': 'CLX14.NYM',\n 'CLX14.NYM': 'CLX14.NYM',\n '10-Yr Bond': '^TNX',\n '^TNX': '^TNX',\n 'EUR/USD': 'EURUSD=X',\n 'EURUSD=X': 'ERUUSD=X',\n 'USD/JPY': 'USDJPY=X',\n 'USDJPY=X': 'USDJPY=X'\n }\n\n\ntickerDict_extended = {'S&P 500': '^GSPC',\n 'S&P500': '^GSPC',\n 's&p 500': '^GSPC',\n 's&p500': '^GSPC',\n '^GSPC': '^GSPC',\n '^Gspc': '^GSPC',\n '^gspc': '^GSPC',\n 'NASDAQ': '^IXIC',\n 'nasdaq': '^IXIC',\n 'Nasdaq': '^IXIC',\n '^IXIC': '^IXIC',\n '^ixic': '^IXIC',\n '^Ixic': '^IXIC',\n 'NASDAQ Composite': '^IXIC',\n 'nasdaq Composite': '^IXIC',\n 'Nasdaq Composite': '^IXIC',\n 'NASDAQComposite': '^IXIC',\n 'nasdaq composite': '^IXIC',\n 'Nasdaq composite': '^IXIC',\n 'Russell 2000': '^RUT',\n 'Russell2000': '^RUT',\n 'RUSSELL 2000': '^RUT',\n 'RUSSELL2000': '^RUT',\n 'russell 2000': '^RUT',\n 'russell2000': '^RUT',\n '^RUT': '^RUT',\n '^Rut': '^RUT',\n '^rut': '^RUT',\n 'Nikkei 225': '^N225',\n 'Nikkei225': '^N225',\n 'NIKKEI225': '^N225',\n 'NIKKEI 225': '^N225',\n 'nikkei225': '^N225',\n 'nikkei 225': '^N225',\n '^N225': '^N225',\n '^n225': '^N225',\n 'Crude Oil': 'CLX14.NYM',\n 'crude oil': 'CLX14.NYM',\n 'CRUDE OIL': 'CLX14.NYM',\n 'Crude Oil Nov 14': 'CLX14.NYM',\n 'CrudeOil': 'CLX14.NYM',\n 'crudeoil': 'CLX14.NYM',\n 'CRUDEOIL': 'CLX14.NYM',\n 'CLX14.NYM': 'CLX14.NYM',\n 'Clx14.Nym': 'CLX14.NYM',\n 'clx14.nym': 'CLX14.NYM',\n '10-Yr Bond': '^TNX',\n '10-yr bond': '^TNX',\n 'Bond 10yr': '^TNX',\n 'bond 10yr': '^TNX',\n 'Bond10yr': '^TNX',\n 'CBOE Interest Rat': '^TNX',\n 'CBOE Interest Rate': '^TNX',\n 'cboe interest rate': '^TNX',\n '^TNX': '^TNX',\n '^Tnx': '^TNX',\n '^tnx': '^TNX',\n 'EUR/USD': 'EURUSD=X',\n 'Eur/Usd': 'EURUSD=X',\n 'eur/usd': 'EURUSD=X',\n 'EUR to USD': 'EURUSD=X',\n 'Eur to Usd': 'EURUSD=X',\n 'eur to usd': 'EURUSD=X',\n 'Euro/USD': 'EURUSD=X',\n 'Euro/usd': 'EURUSD=X',\n 'euro/USD': 'EURUSD=X',\n 'euro/usd': 'EURUSD=X',\n 'Euro to USD': 'EURUSD=X',\n 'Euro to usd': 'EURUSD=X',\n 'euro to USD': 'EURUSD=X',\n 'euro to usd': 'EURUSD=X',\n 'EURUSD=X': 'EURUSD=X',\n 'EurUsd=x': 'EURUSD=X',\n 'eurusd=x': 'EURUSD=X',\n 'USD/JPY': 'USDJPY=X',\n 'usd/jpy': 'USDJPY=X',\n 'USD/YEN': 'USDJPY=X',\n 'USD/Yen': 'USDJPY=X',\n 'usd/yen': 'USDJPY=X',\n 'USD to JPY': 'USDJPY=X',\n 'usd to jpy': 'USDJPY=X',\n 'USD to YEN': 'USDJPY=X',\n 'USD to Yen': 'USDJPY=X',\n 'usd to yen': 'USDJPY=X',\n 'USDJPY=X': 'USDJPY=X',\n 'UsdJpy=x': 'USDJPY=X',\n 'usdjpy=x': 'USDJPY=X',\n }\n\n\nif __name__ == '__main__':\n sys.exit(main())\n","sub_path":"mktdotpy/mktdicts.py","file_name":"mktdicts.py","file_ext":"py","file_size_in_byte":6970,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"481007846","text":"import renpy\nimport os\nimport sys\nimport imp\nimport modinfo\nimport importlib\n# So technically anything up here's going to be imported into mods since we're making a copy of our globals. It's safe to import them anyway. \n\nmodinfo.init()\n \ndef getdir():\n return renpy.config.gamedir\n\nprint(\"AWSW Mod Loader Init\")\n\nsearch_dir = getdir() + \"/mods/\"\nif not os.path.exists(search_dir):\n os.makedirs(search_dir)\n\nsys.path.append(getdir() + \"/modloader/\")\nsys.path.append(getdir() + \"/mods/\")\n\nloaded_mods = []\n\nfor object in os.listdir(search_dir):\n fullpath = search_dir + object\n if os.path.isdir(fullpath):\n for object2 in os.listdir(fullpath):\n modobj = fullpath + '/' + object2\n if object2 == 'mod.py':\n print(('Loaded mod ' + object).encode('utf-8'))\n name = os.path.splitext(os.path.split(modobj)[-1])\n modinfo.modlist.append(object)\n loc = dict()\n glo = dict(globals())\n execfile(modobj, glo, loc) # We want to isolate mods from each other, but give them a copy of our globals so modinfo is not reloaded.\n loaded_mods.append((glo, loc)) # Grab locals so we can initialize the mods later.\n elif object2 == 'resource' and os.path.isdir(modobj):\n renpy.config.searchpath.append(modobj)\n\nfor (glo, loc) in loaded_mods:\n if 'mod_init' in loc:\n comb = glo.copy()\n comb.update(loc)\n func = loc['mod_init']\n exec(func.__code__, comb) # This is nasty!\n \n# force renpy to reindex all game files\nrenpy.loader.old_config_archives = None","sub_path":"modloader/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"327382251","text":"import sys\nimport numpy as np\nfrom numba import jit\n\nfrom scipy.special import hyp2f1\nimport scipy.integrate as integrate\n\n@jit\ndef Support(params):\n if params[0] <= 0 : return False\n # if params[1] <= 0 : return False\n return True\n\n@jit\ndef Density(r,params,Rmax):\n rc = params[0]\n a = 0.6\n b = 2.5\n g = 0.0\n v1 = ((r/rc)**(-g))*(1.0 + (r/rc)**(1./a))**(-a*(b - g)) \n w = a*(b-g)\n x = -a*(g-2.)\n y = 1. - (a*(g-2.))\n z = -((rc/Rmax)**(-1./a))\n v2 = -((rc**g)*((Rmax**(-1./a))**(a*(g-2.)))/(g-2.))*hyp2f1(w,x,y,z)\n return v1/v2\n\n\ndef Number(r,params,Rmax):\n Num = np.vectorize(lambda y: integrate.quad(lambda x:Density(x,params,Rmax)*x,1e-5,y,\n epsabs=1.49e-03, epsrel=1.49e-03,limit=1000)[0])\n return Num(r)\n\n@jit\ndef logLikeStar(p,r,params,Rmax):\n return np.log((p*r*Density(r,params,Rmax)) + (1.-p)*LikeField(r,Rmax))\n\n@jit\ndef LikeField(r,rm):\n return 2.*r/rm**2\n\nclass Module:\n \"\"\"\n Chain for computing the likelihood \n \"\"\"\n def __init__(self,radii,pro,Rmax,trans_lim):\n \"\"\"\n Constructor of the logposteriorModule\n \"\"\"\n self.radii = radii\n self.pro = pro\n self.Rmax = Rmax\n self.t = trans_lim\n print(\"Module Initialized\")\n\n def Priors(self,params, ndim, nparams):\n #------- Uniform Priors -------\n for i in range(ndim):\n params[i] = (params[i])*(self.t[i,1]-self.t[i,0])+self.t[i,0]\n\n def LogLike(self,params,ndim,nparams):\n #----- Checks if parameters' values are in the ranges\n if not Support(params) : \n return -1e50\n\n # ----- Computes Likelihoods ---------\n llike = np.sum(map(lambda w,x:logLikeStar(w,x,params,self.Rmax),self.pro,self.radii))\n # print(llike)\n return llike\n\n\n\n","sub_path":"MultiNest/old-lixo/Models/MGDPRC.py","file_name":"MGDPRC.py","file_ext":"py","file_size_in_byte":1843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"373909112","text":"#Uncoded Computation - Multi-message Communıcation\n\nimport numpy as np\n\n#ds = 1000 # data size\nds = 3000\nsample = 720\nfp = 0.8 #failure probability\ndelay_time = 0.006\nnw = 20 # number of workers\nr = 3 # repetition factor\nnumbexp = 100 # number of experiment\nNWe = 10\n\ndef readdata(filename):\n times = []\n inp = open(filename, \"r\")\n for line in inp.readlines():\n times.append([float(val) for idx, val in enumerate(line.split(\" \")) if idx != 0 and idx <= sample])\n comptime = times[0]\n commtime = times[1]\n return np.array(comptime), np.array(commtime)\n\ndef isdecodable(received, nw):\n flag = 0\n count = 0\n for i in range(0, nw):\n if i in received:\n count += 1\n if count == nw:\n flag = 1\n return flag\n\n\n\ndef decodingtime(assign, realize, nw):\n rs = np.sort(realize, axis=None)\n i = 0\n time = 0\n while i < rs.size:\n time = rs[i]\n received = assign[realize <= time]\n if isdecodable(received, nw) == 1:\n break\n i += 1\n return time\n\n\ndef main():\n assign = np.zeros(shape=(nw, r))\n for i in range(0, r):\n assign[:, i] = np.roll(np.arange(nw), -i)\n comp = np.zeros(shape=(NWe, sample))\n comm = np.zeros(shape=(NWe, sample))\n for d in range(0, NWe):\n comp[d], comm[d] = readdata(\"P2P/resultsp2p-\" + str(ds) + \"rank\" + str(d+1) + \".txt\")\n comp = comp.reshape((nw, int(sample / (nw / NWe))))\n comm = comm.reshape((nw, int(sample / (nw / NWe))))\n itercount = int((sample / (nw / NWe)) / r) # number of iteration\n avg = 0\n for nexp in range(0, numbexp):\n wt = np.zeros(shape=(1, itercount))\n for i in range(0, itercount):\n for j in range(0, nw):\n rgn = np.random.random_sample()\n if rgn < fp:\n delay = delay_time\n else:\n delay = 0\n cctp = np.zeros(shape=(1, r))\n for l in range(0, r):\n if l == 0:\n cctp[0][l] = comp[j][i * r] + comm[j][i * r]\n else:\n compsum = 0\n for k in range(0, l + 1):\n compsum += comp[j][i * r + k] # compute overall computation time\n cctp[0][l] = max(compsum, cctp[0][l - 1]) + comm[j][i * r + l]\n cctp = cctp + delay\n if j == 0:\n cct = cctp\n else:\n cct = np.concatenate((cct, cctp))\n wt[0][i] = decodingtime(assign, cct, nw)\n avg += np.mean(wt) / numbexp\n print(avg)\nmain()","sub_path":"FixedRandomDelay/distcomp_UCC_MM.py","file_name":"distcomp_UCC_MM.py","file_ext":"py","file_size_in_byte":2650,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"143601267","text":"import time, urandom, struct\nfrom machine import Pin, SPI \n\nclass LoRa:\n def __init__(self, RST_Pin, CS_Pin, SPI_CH, SCK_Pin, MOSI_Pin, MISO_Pin, DIO0_Pin, LoRa_id = 0, wait_ACK=True, plus20dBm=False): \n self.ack_token = 0\n self.sending = False\n self.send_id = LoRa_id\n self.header_fmt = 'HHH' # self.send_id, recv_id, self.ack_token\n self._mode = None\n ####################\n # #\n # 1.Reset # \n # #\n #################### \n # Reset LoRa Module\n rst_pin = Pin(RST_Pin, Pin.OUT)\n rst_pin.off()\n time.sleep(0.01)\n rst_pin.on()\n time.sleep(0.01) \n\n ####################\n # #\n # 2.SPI #\n # #\n ####################\n '''\n We command LoRa module to perform Tx/Rx operations via the SPI interface.\n We disable SPI communication first to ensure it only happends when we need.\n Define communication functions read and write.\n The SPI comm is enabled temporarily for reading and writing and disabled thereafter.\n '''\n # Disable SPI communication with the LoRa module\n self.cs_pin = Pin(CS_Pin, Pin.OUT)\n self.cs_pin.on() # Release board from SPI Bus by bringing it into high impedance status. \n \n # SPI communication\n # See datasheet: Device support SPI mode 0 (polarity & phase = 0) up to a max of 10MHz.\n self.spi = SPI(SPI_CH, baudrate=10_000_000, polarity=0, phase=0,\n sck=Pin(SCK_Pin), mosi=Pin(MOSI_Pin), miso=Pin(MISO_Pin)\n ) \n\n ####################\n # #\n # 3.Lora #\n # #\n ####################\n self.RegTable = { # register table\n 'RegFifo' : 0x00 ,\n 'RegOpMode' : 0x01 , # operation mode\n 'RegFrfMsb' : 0x06 ,\n 'RegFrfMid' : 0x07 ,\n 'RegFrfLsb' : 0x08 ,\n 'RegPaConfig' : 0x09 ,\n 'RegFifoTxBaseAddr' : 0x0e ,\n 'RegFifoRxBaseAddr' : 0x0f ,\n 'RegFifoAddrPtr' : 0x0d ,\n 'RegFifoRxCurrentAddr' : 0x10 ,\n 'RegIrqFlags' : 0x12 , \n 'RegRxNbBytes' : 0x13 , # Number of received bytes \n 'RegPktSnrValue' : 0x19 ,\n 'RegPktRssiValue' : 0x1a ,\n 'RegRssiValue' : 0x1b ,\n 'RegModemConfig1' : 0x1d , \n 'RegModemConfig2' : 0x1e , \n 'RegPreambleMsb' : 0x20 , \n 'RegPreambleLsb' : 0x21 ,\n 'RegPayloadLength' : 0x22 ,\n 'RegModemConfig3' : 0x26 , \n 'RegDioMapping1' : 0x40 , \n 'RegVersion' : 0x42 , \n 'RegPaDac' : 0x4d \n }\n \n self.Mode = { # see Table 16 LoRa ® Operating Mode Functionality \n 'SLEEP' : 0b000,\n 'STANDBY' : 0b001,\n 'TX' : 0b011,\n 'RXCONTINUOUS' : 0b101, \n 'RXSINGLE' : 0b110, \n 'CAD' : 0b111, \n } \n\n # Choose LoRa mode and Test write/read functions\n LongRangeMode = 0b1\n # Choose LoRa (instead of FSK) mode for SX1276 and put the module in sleep mode\n self.write('RegOpMode', self.Mode['SLEEP'] | LongRangeMode << 7) \n # Test read function \n assert self.read('RegOpMode') == (self.Mode['SLEEP'] | LongRangeMode << 7), \"LoRa initialization failed\"\n \n # Set modem config: bandwidth, coding rate, header mode, spreading factor, CRC, and etc. \n # See 4.4. LoRa Mode Register Map \n Bw = {'125KHz':0b0111, '500kHz':0b1001}\n CodingRate = {5:0b001, 6:0b010, 7:0b011, 8:0b100}\n ImplicitHeaderModeOn = {'Implicit':0b1, 'Explicit':0b0}\n self.write('RegModemConfig1', Bw['125KHz'] << 4 | CodingRate[5] << 1 | ImplicitHeaderModeOn['Explicit'])\n SpreadingFactor = {7:0x7, 9:0x9, 12:0xC}\n TxContinuousMode = {'normal':0b0, 'continuous':0b1}\n RxPayloadCrcOn = {'disable':0b0, 'enable':0b1}\n self.write('RegModemConfig2', SpreadingFactor[7] << 4 | TxContinuousMode['normal'] << 3 | RxPayloadCrcOn['enable'] << 2 | 0x00) # Last 0x00 is SymbTimeout(9:8)\n self.write('RegModemConfig3', 0x04) # 0x04 is SymbTimeout(7:0)\n \n # Preamble length\n self.write('RegPreambleMsb', 0x0) # Preamble can be (2^15)kb long, much longer than payload\n self.write('RegPreambleLsb', 0x8) # but we just use 8-byte preamble\n \n # See 4.1.4. Frequency Settings\n FXOSC = 32e6 # Freq of XOSC\n FSTEP = FXOSC / (2**19)\n Frf = int(915e6 / FSTEP)\n self.write('RegFrfMsb', (Frf >> 16) & 0xff)\n self.write('RegFrfMid', (Frf >> 8) & 0xff)\n self.write('RegFrfLsb', Frf & 0xff)\n \n # Output Power\n '''\n If desired output power is within -4 ~ +15dBm, use PA_LF or PA_HF as amplifier. \n Use PA_BOOST as amplifier to output +2 ~ +17dBm continuous power or up to 20dBm \n peak power in a duty cycled operation.\n Here we will always use PA_BOOST. \n Since we use PA_BOOST, Pout = 2 + OutputPower and MaxPower could be any number (Why not 0b111/0x7?)\n '''\n PaSelect = {'PA_BOOST':0b1, 'RFO':0b0} # Choose PA_BOOST (instead of RFO) as the power amplifier\n MaxPower = {'15dBm':0x7, '13dBm':0x2} # Pmax = 10.8 + 0.6 * 7 \n OutputPower = {'17dBm':0xf, '2dBm':0x0} \n self.write('RegPaConfig', PaSelect['PA_BOOST'] << 7 | MaxPower['15dBm'] << 4 | OutputPower['2dBm'])\n \n # Enables the +20dBm option on PA_BOOST pin. \n if plus20dBm: # PA (Power Amplifier) DAC (Digital Analog Converter)\n PaDac = {'default':0x04, 'enable_PA_BOOST':0x07} # Can be 0x04 or 0x07. 0x07 will enables the +20dBm option on PA_BOOST pin\n self.write('RegPaDac', PaDac['enable_PA_BOOST']) \n \n # FIFO data buffer \n '''\n SX1276 has a 256 byte memory area as the FIFO buffer for Tx/Rx operations.\n How do we know which area is for Tx and which is for Rx.\n We must set the base addresses RegFifoTxBaseAddr and RegFifoRxBaseAddr independently.\n Since SX1276 work in a half-duplex manner, we better set both base addresses\n at the bottom (0x00) of the FIFO buffer so that we can buffer 256 byte data\n during transmition or reception.\n ''' \n self.Fifo_Bottom = 0x00 # We choose this value to max buffer we can write (then send out)\n self.write('RegFifoTxBaseAddr', self.Fifo_Bottom)\n self.write('RegFifoRxBaseAddr', self.Fifo_Bottom)\n \n ####################\n # #\n # 4.Interrupt #\n # #\n ####################\n '''\n # This section is optional for Tx.\n # It enable an interrupt when Tx is done.\n '''\n self.DioMapping = {\n 'Dio0' : {\n 'RxDone' : 0b00 << 6,\n 'TxDone' : 0b01 << 6,\n 'CadDone' : 0b10 << 6\n },\n 'Dio1' : {\n 'RxTimeout' : 0b00 << 4,\n 'FhssChangeChannel': 0b01 << 4,\n 'CadDetected' : 0b10 << 4\n },\n 'Dio2' : {},\n 'Dio3' : {},\n 'Dio4' : {},\n 'Dio5' : {},\n } \n \n self.IrqFlags = {\n 'RxTimeout' : 0b1 << 7,\n 'RxDone' : 0b1 << 6,\n 'PayloadCrcError' : 0b1 << 5,\n 'ValidHeader' : 0b1 << 4,\n 'TxDone' : 0b1 << 3,\n 'CadDone' : 0b1 << 2,\n 'FhssChangeChannel': 0b1 << 1,\n 'CadDetected' : 0b1 << 0, \n }\n \n dio0_pin = Pin(DIO0_Pin, Pin.IN)\n dio0_pin.irq(handler=self._irq_handler, trigger=Pin.IRQ_RISING)\n \n ''' # interrupt flag mask: use to deactive a particular interrupt\n RegIrqFlagsMask = 0x11;\n IrqFlagsMask = {\n 'RxTimeoutMask' : 0b1 << 7,\n 'RxDoneMask' : 0b1 << 6,\n 'PayloadCrcErrorMask' : 0b1 << 5,\n 'ValidHeaderMask' : 0b1 << 4,\n 'TxDoneMask' : 0b1 << 3,\n 'CadDoneMask' : 0b1 << 2,\n 'FhssChangeChannelMask': 0b1 << 1,\n 'CadDetectedMask' : 0b1 << 0\n }\n write(RegIrqFlagsMask, IrqFlagsMask['TxDoneMask']) # This will deactivate interrupt on TxDone.\n ''' \n\n self.mode = 'STANDBY' # Request Standby mode so SX1276 performs reception initialization. \n \n @property\n def mode(self):\n return self._mode\n\n @mode.setter\n def mode(self, value): \n if self.mode != value:\n if value == 'TX':\n self.write('RegDioMapping1', self.DioMapping['Dio0']['TxDone']) \n elif value == 'RXCONTINUOUS':\n self.write('RegDioMapping1', self.DioMapping['Dio0']['RxDone']) \n self.write('RegOpMode', self.Mode[value]) \n self._mode = value\n\n def write(self, reg, data, fifo=False): \n wb = bytes([self.RegTable[reg] | 0x80]) # Create a writing byte\n if fifo:\n data = wb + data\n else:\n data = wb + bytes([data]) \n self.cs_pin.value(0) # Bring the CS pin low to enable communication \n self.spi.write(data)\n self.cs_pin.value(1) # release the bus. \n\n def read(self, reg=None, length=1):\n self.cs_pin.value(0)\n # https://docs.micropython.org/en/latest/library/machine.SPI.html#machine-softspi\n if length == 1:\n data = self.spi.read(length+1, self.RegTable[reg])[1]\n else:\n data = self.spi.read(length+1, self.RegTable[reg])[1:]\n self.cs_pin.value(1)\n return data\n \n def _irq_handler(self, pin):\n irq_flags = self.read('RegIrqFlags')\n if irq_flags & self.IrqFlags['TxDone']: \n self.mode = 'RXCONTINUOUS' \n while 1:\n 1\n self.after_TxDone(self)\n\n elif irq_flags & self.IrqFlags['RxDone']:\n if irq_flags & self.IrqFlags['PayloadCrcError']:\n print('PayloadCrcError')\n else:\n self.write('RegFifoAddrPtr', self.read('RegFifoRxCurrentAddr'))\n packet = self.read('RegFifo', self.read('RegRxNbBytes'))\n PacketSnr = self.read('RegPktSnrValue')\n SNR = PacketSnr / 4\n PacketRssi = self.read('RegPktRssiValue')\n #Rssi = read(RegRssiValue)\n if SNR < 0:\n RSSI = -157 + PacketRssi + SNR\n else:\n RSSI = -157 + 16 / 15 * PacketRssi\n RSSI = round(RSSI, 2) # Table 7 Frequency Synthesizer Specification\n self.packet_handler(self, packet, SNR, RSSI) \n self.Tx() \n else:\n for i, j in self.IrqFlags.items():\n if irq_flags & j:\n print(i)\n\n self.write('RegIrqFlags', 0xff) # write anything could clear all types of interrupt flags\n \n \n def send(self, data, recv_id=0): \n if len(data) > 240: raise # want to send a too large message \n self.ack_token = urandom.randint(0,65535)\n header = struct.pack('= 0.05:\n # more than 5% pixels are void\n flag[k] = 1 # this image will not be saved\n k += 1\n\n if fg:\n return image_patchs, flag\n else:\n return image_patchs\n\n\ndef rotate_image_random(img, rotation_index):\n deg_dict = {\n 1: 0,\n 2: 90,\n 3: 180,\n 4: 270\n }\n\n # rows = img.shape[0]\n # cols = img.shape[1]\n #\n # deg = deg_dict[rotation_index]\n\n if rotation_index != 1:\n # M = cv2.getRotationMatrix2D(((cols - 1) / 2.0, (rows - 1) / 2.0), deg, 1)\n # dst = cv2.warpAffine(img, M, (cols, rows))\n\n dst = np.rot90(img, rotation_index-1)\n\n return dst\n\n else:\n return img\n\n\nif __name__ == \"__main__\":\n\n reference = \"/run/user/1001/gvfs/smb-share:server=141.58.125.9,share=s-platte/ShuFangwen/results/lvl4_nadir/test_set/2_mask\"\n data_path = \"/run/user/1001/gvfs/smb-share:server=141.58.125.9,share=s-platte/ShuFangwen/results/lvl4_nadir/test_set\"\n folders_list = os.listdir(data_path)\n folders_list.remove(\"2_mask\")\n folders_list.remove(\"not_use_feature\")\n # folders_list.remove(\"1_pointlabel\")\n\n save_path = \"/data/fangwen/mix_test2\"\n make_if_not_exists(save_path)\n\n size = (480, 480)\n\n mask_list = os.listdir(reference)\n length = len(mask_list)\n for l in tqdm(range(length)):\n\n for rotation_index in range(1, 5):\n\n name = mask_list[l]\n\n mask_path = os.path.join(reference, name)\n # name = \"DSC03717.tif\"\n # mask_path ='/data/fangwen/results/level3/test_set/2_mask/DSC03717.tif'\n mask = cv2.imread(mask_path, 0)\n mask = rotate_image_random(mask, rotation_index)\n\n mask_patchs, flag = chip(mask, chip_size=size, overlap=0.5, nchannel=1, fg=True)\n\n # based on this flag, we chip other image\n for folder in folders_list:\n folder_path = os.path.join(data_path, folder)\n img_path = os.path.join(folder_path, name)\n\n if folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"f\" or folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"5\":\n # read index image and feature image\n img = tifffile.imread(img_path)\n img = rotate_image_random(img, rotation_index)\n img_patchs = chip(img, chip_size=size, overlap=0.5, nchannel=1, fg=False)\n\n elif folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"rgb\" or folder_path.split(\"/\")[-1].split(\"_\")[\n -2] == \"4\":\n # rgb\n img = cv2.imread(img_path)\n img = rotate_image_random(img, rotation_index)\n img_patchs = chip(img, chip_size=size, overlap=0.5, nchannel=3, fg=False)\n\n elif folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"3\":\n # grey\n img = cv2.imread(img_path, 0)\n img = rotate_image_random(img, rotation_index)\n img_patchs = chip(img, chip_size=size, overlap=0.5, nchannel=1, fg=False)\n\n for id in range(flag.shape[0]):\n\n if flag[id] == 0:\n # save masks\n save_mask = os.path.join(save_path, \"2_mask\")\n make_if_not_exists(save_mask)\n cv2.imwrite(os.path.join(save_mask, name.split(\".\")[-2] + \"_\" + str(id) + '_r' + str(rotation_index) + \".tif\"),\n mask_patchs[id])\n\n # save other images\n if folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"f\" or folder_path.split(\"/\")[-1].split(\"_\")[\n -2] == \"5\":\n save_img = os.path.join(save_path, folder_path.split(\"/\")[-1])\n make_if_not_exists(save_img)\n tifffile.imsave(os.path.join(save_img, name.split(\".\")[-2] + \"_\" + str(id) + '_r' + str(rotation_index) + \".tif\"),\n img_patchs[id])\n\n if folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"rgb\" or folder_path.split(\"/\")[-1].split(\"_\")[\n -2] == \"4\":\n save_img = os.path.join(save_path, folder_path.split(\"/\")[-1])\n make_if_not_exists(save_img)\n cv2.imwrite(os.path.join(save_img, name.split(\".\")[-2] + \"_\" + str(id) + '_r' + str(rotation_index) + \".tif\"),\n img_patchs[id])\n\n elif folder_path.split(\"/\")[-1].split(\"_\")[-2] == \"3\":\n save_img = os.path.join(save_path, folder_path.split(\"/\")[-1])\n make_if_not_exists(save_img)\n cv2.imwrite(os.path.join(save_img, name.split(\".\")[-2] + \"_\" + str(id) + '_r' + str(rotation_index) + \".tif\"),\n img_patchs[id])\n","sub_path":"src/semantic-segmentation/chip.py","file_name":"chip.py","file_ext":"py","file_size_in_byte":6985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"56684102","text":"import numpy as np\nimport cv2\n\n#we are capturing image from the standard cameras\n# 0,1,2,... (since we are using the laptop capera , we use 0 , else we use 1,2..)\n#Note: we need to create a VideoCapture object to capture video in OpenCV\n\ncap = cv2.VideoCapture(0) #this will create a streaming video via the lappy cam\n\n#we can apply processing on this video frame/frame and apply computations accordingly\n\n#we are interesting in caputuring camera frames from several cameras in the setup\n#and detect the objects and their location, output as a matrix,as give\n# plotting in a graph\n\n#Challenges:\n#Trigerring frame capture at the same time from these cameras\n#Converting them in to unified co-ordinates\n#providing a 3D view of these objects\n#plot the orientation of these objects in this unified co-ordinate system\n\n#Detection\n# 1. detect robots using red circles mounted on these cameras\n# 2. detect by learning the shape of these robots by training negative\n# and positive images\n# 3. Proximity information specific to individual cameras \n\nwhile True:\n #Capturing the images frames by frames\n ret,frame = cap.read() #the read method of the VideoCapture Object\n #returns a frame\n\n #we convert each frame into a grayscale\n gray = cv2.cvtColor(frame,cv2.COLOR_BGR2GRAY)\n\n #displaying the resulting frame\n cv2.imshow('frame',gray)\n \n\n k = cv2.waitKey(0)\n\n if k == 27:\n cv2.destroyAllWindows()\n\n for i in range(1,4):\n cv2.waitKey(1)\n \n \n","sub_path":"VideoCapture.py","file_name":"VideoCapture.py","file_ext":"py","file_size_in_byte":1521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"526775576","text":"import nltk\nfrom nltk.corpus import state_union\nfrom nltk.tokenize import PunktSentenceTokenizer\n\ntrain_text = state_union.raw(\"margot.txt\")\nsample_text = state_union.raw(\"gal_gadot.txt\")\n\ncustom_sent_tokenizer = PunktSentenceTokenizer(train_data)\n\ntokenized = custom_sent_tokenizer(sample_text)\n\ntry:\n for w in tokenized:\n words = nltk.word_tokenize(w)\n tagged = nltk.pos_tag(words)\n print(tagged)\n \nexcept Exception as e:\n print(str(e))\n","sub_path":"Natural_Language_Processing/4- Part of Speech tagging.py","file_name":"4- Part of Speech tagging.py","file_ext":"py","file_size_in_byte":473,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"307791931","text":"\"\"\"\nCompatibility module for high-level h5py\n\"\"\"\nimport sys\nimport six\n\nWINDOWS_ENCODING = \"mbcs\"\n\n\ntry:\n from os import fspath\nexcept ImportError:\n def fspath(path):\n \"\"\"\n Return the string representation of the path.\n If str or bytes is passed in, it is returned unchanged.\n This code comes from PEP 519, modified to support earlier versions of\n python.\n\n This is required for python < 3.6.\n \"\"\"\n if isinstance(path, (six.text_type, six.binary_type)):\n return path\n\n # Work from the object's type to match method resolution of other magic\n # methods.\n path_type = type(path)\n try:\n return path_type.__fspath__(path)\n except AttributeError:\n if hasattr(path_type, '__fspath__'):\n raise\n try:\n import pathlib\n except ImportError:\n pass\n else:\n if isinstance(path, pathlib.PurePath):\n return six.text_type(path)\n\n raise TypeError(\"expected str, bytes or os.PathLike object, not \"\n + path_type.__name__)\n\n# This is from python 3.5 stdlib (hence lacks PEP 519 changes)\n# This was introduced into python 3.2, so python < 3.2 does not have this\n# Effectively, this is only required for python 2.6 and 2.7, and can be removed\n# once support for them is dropped\ndef _fscodec():\n encoding = sys.getfilesystemencoding()\n if encoding == 'mbcs':\n errors = 'strict'\n else:\n try:\n from codecs import lookup_error\n lookup_error('surrogateescape')\n except LookupError:\n errors = 'strict'\n else:\n errors = 'surrogateescape'\n\n def fsencode(filename):\n \"\"\"\n Encode filename to the filesystem encoding with 'surrogateescape' error\n handler, return bytes unchanged. On Windows, use 'strict' error handler if\n the file system encoding is 'mbcs' (which is the default encoding).\n \"\"\"\n if isinstance(filename, six.binary_type):\n return filename\n elif isinstance(filename, six.text_type):\n return filename.encode(encoding, errors)\n else:\n raise TypeError(\"expect bytes or str, not %s\" % type(filename).__name__)\n\n def fsdecode(filename):\n \"\"\"\n Decode filename from the filesystem encoding with 'surrogateescape' error\n handler, return str unchanged. On Windows, use 'strict' error handler if\n the file system encoding is 'mbcs' (which is the default encoding).\n \"\"\"\n if isinstance(filename, six.text_type):\n return filename\n elif isinstance(filename, six.binary_type):\n return filename.decode(encoding, errors)\n else:\n raise TypeError(\"expect bytes or str, not %s\" % type(filename).__name__)\n\n return fsencode, fsdecode\n\n_fsencode, _fsdecode = _fscodec()\ndel _fscodec\n\ntry:\n from os import fsencode\nexcept ImportError:\n fsencode = _fsencode\n\ntry:\n from os import fsdecode\nexcept ImportError:\n fsdecode = _fsdecode\n\n\ndef filename_encode(filename):\n \"\"\"\n Encode filename for use in the HDF5 library.\n\n Due to how HDF5 handles filenames on different systems, this should be\n called on any filenames passed to the HDF5 library. See the documentation on\n filenames in h5py for more information.\n \"\"\"\n filename = fspath(filename)\n if sys.platform == \"win32\":\n if isinstance(filename, six.text_type):\n return filename.encode(WINDOWS_ENCODING, \"strict\")\n return filename\n return fsencode(filename)\n\n\ndef filename_decode(filename):\n \"\"\"\n Decode filename used by HDF5 library.\n\n Due to how HDF5 handles filenames on different systems, this should be\n called on any filenames passed from the HDF5 library. See the documentation\n on filenames in h5py for more information.\n \"\"\"\n if sys.platform == \"win32\":\n if isinstance(filename, six.binary_type):\n return filename.decode(WINDOWS_ENCODING, \"strict\")\n elif isinstance(filename, six.text_type):\n return filename\n else:\n raise TypeError(\"expect bytes or str, not %s\" % type(filename).__name__)\n return fsdecode(filename)\n","sub_path":"Tensorflow_Pandas_Numpy/source3.6/h5py/_hl/compat.py","file_name":"compat.py","file_ext":"py","file_size_in_byte":4324,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"358430982","text":"#!/usr/bin/env python\n# -*- coding: utf8 -*-\nimport RPi.GPIO as GPIO\nimport MFRC522\nimport signal\nimport time\nimport datetime\n#from datetime import date\nGPIO.setwarnings(False)\ncmpt=0\n\ndate = datetime.datetime.now()\nprint(date)\n \ncontinue_reading = True\n# Capture SIGINT for cleanup when the script is aborted\ndef end_read(signal,frame):\n global continue_reading\n print (\"Lecture terminée\")\n continue_reading = False\n GPIO.cleanup()\n# Hook the SIGINT\nsignal.signal(signal.SIGINT, end_read)\n# Create an object of the class MFRC522\nMIFAREReader = MFRC522.MFRC522()\n#print (\"Press Ctrl-C to stop.\")\n#secteurBloc=eval(input(\"Entrez un Secteur :\\n\"))\nsecteurBlock2=12\n#secteurBlock3=12\n\nprint (\"Passer le tag RFID a lire\")\n# This loop keeps checking for chips. If one is near it will get the UID and authenticate\nwhile continue_reading:\n # Scan for cards \n (status,TagType) = MIFAREReader.MFRC522_Request(MIFAREReader.PICC_REQIDL)\n # If a card is found\n if status == MIFAREReader.MI_OK:\n print (\"Carte detectee\")\n # Get the UID of the card\n (status,uid) = MIFAREReader.MFRC522_Anticoll()\n # If we have the UID, continue\n if status == MIFAREReader.MI_OK:\n data = [0x59,0x61,0x50,0x6F,0x54,0x74,0xFF,0x07,0x80,0x69,0x59,0x61,0x50,0x6F,0x54,0x74]\n # Print UID\n print (\"UID de la carte : \"+str(uid[0])+\".\"+str(uid[1])+\".\"+str(uid[2])+\".\"+str(uid[3])+\".\"+str(uid[4]))\n # This is the default key for authentication\n keyA_Public = [0xFF,0xFF,0xFF,0xFF,0xFF,0xFF]\n # Clee d authentification privée\n keyA_Privé = [0x59,0x61,0x50,0x6F,0x54,0x74] #\"YaPoTt\"\n \n key = [0x59,0x61,0x50,0x6F,0x54,0x74,0xFF,0x07,0x80,0x69,0x59,0x61,0x50,0x6F,0x54,0x74]\n #keyA_Privé = key\n #keyA_Public = key\n # Select the scanned tag\n MIFAREReader.MFRC522_SelectTag(uid)\n # Authenticate with private key\n status = MIFAREReader.MFRC522_Auth(MIFAREReader.PICC_AUTHENT1A, secteurBlock2,keyA_Privé, uid)\n # Check if authenticated\n if(status == MIFAREReader.MI_OK):\n next = False\n print (\"Authentification Avec la Clee Privé \")\n print(\"\\n\")\n print(\"Carte deja initialisé_sur secteur \",secteurBlock2,\"\\n\")\n print(\"INFORMATION Block: \")\n print (\"Le secteur\",secteurBlock2,\" contient actuellement : \")\n MIFAREReader.MFRC522_Read(secteurBlock2)\n print (\"Le secteur\",secteurBlock2+1,\" contient actuellement : \")\n MIFAREReader.MFRC522_Read(secteurBlock2+1)\n print (\"Le secteur\",secteurBlock2+2,\" contient actuellement : \")\n MIFAREReader.MFRC522_Read(secteurBlock2+2)\n # Stop\n #MIFAREReader.MFRC522_StopCrypto1()\n # Make sure to stop reading for cards\n continue_reading = False\n next = False\n else:\n print (\"\\nErreur d\\'Authentification Avec la Clee Privé sur secteur \",secteurBlock2,\"\\n\")\n next =True\n \n if(next == True):\n # Authenticate with Public key\n status1 = MIFAREReader.MFRC522_Auth(MIFAREReader.PICC_AUTHENT1A, secteurBlock2,keyA_Public, uid)\n # Check if authenticated\n if(status1 == MIFAREReader.MI_OK):\n print (\"Authentification Avec la Clee Public sur secteur \",secteurBlock2,\"\\n\")\n print (\"Le secteur \",secteurBlock2+3,\"contient actuellement :\")\n MIFAREReader.MFRC522_Read(secteurBlock2+3)\n print (\"Ecriture ...Clee Privé sur secteur\",secteurBlock2+3)\n MIFAREReader.MFRC522_Write(secteurBlock2+3, data)\n print (\"\\n\")\n print (\"Carte initialisé sur Block\",secteurBlock2+3)\n MIFAREReader.MFRC522_StopCrypto1()\n continue_reading = False\n else:\n print (\"Error Authentification Avec la Clee Public sur secteur \",secteurBlock2)\n \n","sub_path":"sauvegarde/Read.py","file_name":"Read.py","file_ext":"py","file_size_in_byte":4080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"274552255","text":"def partition(a):\n pivot = a[0]\n left, rite = 0, len(a)-1\n while left != rite:\n while rite != left and a[rite] > pivot:\n rite -= 1\n a[left], a[rite] = a[rite], a[left]\n while left != rite and a[left] <= pivot:\n left += 1\n a[left], a[rite] = a[rite], a[left]\n\n\n\n\nn = int(input())\na = [int(x) for x in input().split()]\n#ifile = open(\"rosalind_par.txt\", \"r\")\n#wfile = open(\"rosalind_par_ans.txt\", \"w\")\n#n = int(ifile.readline())\n#a = [int(x) for x in ifile.readline().split()]\npartition(a)\n#print(*a,file=wfile)\nprint(*a)","sub_path":"Rosalind/algorithm_heights/par.py","file_name":"par.py","file_ext":"py","file_size_in_byte":578,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"338131101","text":"\n\"\"\"\nDefine predicate-safeness.\n\"\"\"\n\nfrom helper import get_predicates, get_ordered_parameter_names\nfrom base import PredicateError\nfrom delayed import DelayedArgumentError\nfrom Transparent import Transparent\n\nclass PredicateSafeFunction(Transparent):\n\n getattr_fallback = lambda self: self._function\n\n def __init__(self, function):\n \"\"\"\n Defines an `@safe` function.\n Is a wrapper for a function.\n \"\"\"\n self._function = function\n self._predicates = get_predicates(self._function) # A dict mapping {parameter name : predicate}\n self._parameter_names = get_ordered_parameter_names(self._function)\n\n for predicate_i, predicate in enumerate(self._predicates.values()):\n if predicate.is_generalized:\n raise DelayedArgumentError(\"Predicate #{} is generalized.\".format(predicate_i))\n\n def test_arguments(self, *args):\n \"\"\"\n Test arguments against their python-predicates.\n Raise an error if any don't satisfy.\n \"\"\"\n for param_num, param_name in enumerate(self._parameter_names):\n argument = args[param_num]\n predicate = self._predicates.get(param_name)\n if predicate and not predicate(argument):\n raise PredicateError(\"Predicate for parameter #{} not satisfied.\".format(param_num))\n\n def test_return_val(self, ret):\n predicate = self._predicates.get(\"return\")\n if predicate and not predicate(ret):\n raise PredicateError(\"Predicate for return value not satisfied.\")\n\n def __call__(self, *args):\n self.test_arguments(*args)\n ret = self._function(*args)\n self.test_return_val(ret)\n return ret","sub_path":"safe.py","file_name":"safe.py","file_ext":"py","file_size_in_byte":1717,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"634861272","text":"\n\nfrom xai.brain.wordbase.verbs._ennoble import _ENNOBLE\n\n#calss header\nclass _ENNOBLES(_ENNOBLE, ):\n\tdef __init__(self,): \n\t\t_ENNOBLE.__init__(self)\n\t\tself.name = \"ENNOBLES\"\n\t\tself.specie = 'verbs'\n\t\tself.basic = \"ennoble\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/verbs/_ennobles.py","file_name":"_ennobles.py","file_ext":"py","file_size_in_byte":245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"631069877","text":"# 2018-8-8\n# update 2018-9-4\n# Code chinese character\n# from codeWord import CodeWord\nclass CodeChar(object):\n\t\"\"\"\n\tHave code problem!\n\ttest = CodeChar()\n\tres = test.code(string)\n\tres2 = test.decode(res)\n\t\"\"\"\n\tdef __init__(self):\n\t\tself.exp = [',','。','?','!',':','《','》','【','】','(',')','、','@','#','$','%','^','&','*','.',',','.','<','>','-','+','=', ' ','\\n', '“','”', '!', '[', ']', '(', ')']\n\t\tself.radix = 19\n\t\tself.a_int = 97\n\t\tself.A_int = 65\n\t\tself.func = CodeWord(33)\n\n\tdef code(self, s):\n\t\tres = []\n\t\tflag = 0\n\t\ttmp = []\n\t\tfor i in s:\n\t\t\tif i in self.exp:\n\t\t\t\tif flag == 1:\n\t\t\t\t\ttmp.append('`')\n\t\t\t\t\tt = self.func.code(\"\".join(tmp[1:-1]))\n\t\t\t\t\tt = '`' + t + '`'\n\t\t\t\t\tres.append(t)\n\t\t\t\t\ttmp = []\n\t\t\t\t\tflag = 0\n\t\t\t\tres.append(i)\n\t\t\telse:\n\t\t\t\ti_to_int = ord(i)\n\t\t\t\tif i_to_int > 19000 and i_to_int < 40000:\n\t\t\t\t\tif flag == 1:\n\t\t\t\t\t\ttmp.append('`')\n\t\t\t\t\t\tt = self.func.code(\"\".join(tmp[1:-1]))\n\t\t\t\t\t\tt = '`' + t + '`'\n\t\t\t\t\t\tres.append(t)\n\t\t\t\t\t\ttmp = []\n\t\t\t\t\t\tflag = 0 \n\t\t\t\t\tnums = []\n\t\t\t\t\twhile i_to_int > 0:\n\t\t\t\t\t\twhile i_to_int:\n\t\t\t\t\t\t\tnum = i_to_int % 10\n\t\t\t\t\t\t\tnums.append(num)\n\t\t\t\t\t\t\ti_to_int = i_to_int // 10\n\t\t\t\t\t\tnums = nums[::-1]\n\t\t\t\t\t\tr = []\n\t\t\t\t\t\tfir = nums[0] * 10 + nums[1] - self.radix + self.A_int\n\t\t\t\t\t\tr.append(chr(fir))\n\t\t\t\t\t\tsec = nums[2] + self.a_int\n\t\t\t\t\t\tr.append((chr(sec)))\n\t\t\t\t\t\tthr = nums[3] + self.a_int + 8\n\t\t\t\t\t\tr.append(chr(thr))\n\t\t\t\t\t\tfur = nums[4] + self.a_int + 16\n\t\t\t\t\t\tr.append(chr(fur))\n\t\t\t\t\t\tres.append(\"\".join(r))\n\t\t\t\t\t\tr = []\n\t\t\t\telse:\n\t\t\t\t\tif flag == 0:\n\t\t\t\t\t\ttmp.append('`')\n\t\t\t\t\t\ttmp.append(i)\n\t\t\t\t\t\tflag = 1\n\t\t\t\t\telse:\n\t\t\t\t\t\ttmp.append(i)\n\t\treturn \" \".join(res)\n\n\tdef decode(self, s):\n\t\trecord = []\n\t\tflag = 1\n\t\tmark = 0\n\t\tend = 0\n\t\ttmp = []\n\t\tre = \"\"\n\t\tfor i in s:\n\t\t\tif i not in self.exp and i != \" \" and i != '`' and mark == 0:\n\t\t\t\trecord.append(i)\n\t\t\telif i == '`':\n\t\t\t\tif mark == 0:\n\t\t\t\t\tmark = 1\n\t\t\t\tif end:\n\t\t\t\t\tt = self.func.deCode(\" \".join(tmp))\n\t\t\t\t\tre += t\n\t\t\t\t\tmark = 0\n\t\t\t\t\tend = 0\n\t\t\t\t\ttmp = []\n\t\t\telse:\t\n\t\t\t\tif mark == 1:\n\t\t\t\t\ttmp.append(i)\n\t\t\t\t\tend = 1\n\t\t\t\telse:\n\t\t\t\t\tre += i\n\t\t\t\t\tflag = 0\n\t\t\tif flag == 0 and len(record) == 4:\n\t\t\t\tr = 0\n\t\t\t\tr += (ord(record[0]) - self.A_int + self.radix) * 1000\n\t\t\t\tr += (ord(record[1]) - self.a_int) * 100\n\t\t\t\tr += (ord(record[2]) - self.a_int - 8) * 10\n\t\t\t\tr += ord(record[3]) - self.a_int - 16\n\t\t\t\tre += chr(r)\n\t\t\t\trecord = []\n\t\t\t\tflag = 1\n\n\t\tres = \"\"\n\t\tpre = 0\n\t\tfor i in re:\n\t\t\tif i == ' ' and pre == 1:\n\t\t\t\tres += i\n\t\t\telif i != ' ':\n\t\t\t\tres += i\n\t\t\t\tpre = 0\n\n\t\t\tif i == ' ':\n\t\t\t\tpre = 1\n\t\treturn res\n\t\t\nclass CodeWord(object):\n\tdef __init__(self,x):\n\t\tself.x = x\n\tdef code(self, s):\n\t\td = {}\n\t\tfor i in (65,97):\n\t\t\tfor j in range(26):\n\t\t\t\td[chr(i+j)] = chr( (j+self.x) % 26 + i)\n\t\tres = \"\".join([d.get(c,c) for c in s])\n\t\treturn res\n\n\tdef deCode(self, s):\n\t\td = {}\n\t\tfor i in (65, 97):\n\t\t\tfor j in range(26):\n\t\t\t\td[chr(i+j)] = chr((j+26-self.x) % 26 + i)\n\t\tres = \"\".join([d.get(c,c) for c in s])\n\t\treturn res\n\ns = \"\"\"\n![](dijkstra.jpg)\n>## 个人博客\n\n- [链接](http://www.lxxx.site)\n\n>## tool \n\n- 图片,文件批量操作\n- 数据处理\n- 字符编码\n\n>## LeetCode\n\n- LeetCode算法题解析 Python, Java, C\n\n>## C Python Linux java MySQL PHP MatLab PyQt5 神经网络\n- 笔记\n\n\n>## Data Structures and Algorithm Analysis\n\n- 大部分用python实���\n\n\n\"\"\"\ntest = CodeChar()\nr1 = test.code(s)\nr2 = test.decode(r1)\nprint()\nprint(r1)\nprint(r2)\n\n\"\"\"\n! [ ] ( `kpqrzayh` . `qwn` ) \n > # # Bajq Bbnu Cdly Eeny \n \n - [ Tbms Gfiz ] ( `oaaw://ddd` . `seee` . `zpal` ) \n \n > # # `avvs` \n \n - Dcpq Kcnv , Gjrr Bcju Gciz Sdkx Giiv Bdjw \n - Gjoy Genu Dhqy Khis \n - Edqt Mfkw Nflu Lhkr \n \n > # # `SllaJvkl` \n \n - `SllaJvkl` Mglz Iior Uaou Qcrz Hfjs `Wfaovu` , `Qhch` , `J` \n \n > # # `J` `Wfaovu` `Spube` `qhch` `TfZXS` `WOW` `ThaShi` `WfXa5` Mapq Neot Nfrt Nepw \n - Mfiy Qhoq \n \n \n > # # `Khah` `Zaybjabylz` `huk` `Hsnvypaot` `Huhsfzpz` \n \n - Dikt Sarw Bjry Kjrs `wfaovu` Eenu Kgjw \n \n \n\n\n![](dijkstra.jpg)\n>## 个人博客\n\n- [链接](http://www.lxxx.site)\n\n>## tool \n\n- 图片,文件批量操作\n- 数据处理\n- 字符编码\n\n>## LeetCode\n\n- LeetCode算法题解析 Python, Java, C\n\n>## C Python Linux java MySQL PHP MatLab PyQt5 神经网络\n- 笔记\n\n\n>## Data Structures and Algorithm Analysis\n\n- 大部分用python实现\n\n\"\"\"","sub_path":"tool/汉字编码/codeChinese.py","file_name":"codeChinese.py","file_ext":"py","file_size_in_byte":4350,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"341373658","text":"from Search import *\nimport math\nfrom save import *\n# import pylab\n# import matplotlib.pyplot as plt\n\n\ndef A(x, y, alpha, delta):\n f_zn = f(x, y, alpha)\n g_zn = g(x, y, alpha, delta)\n f_x = pr_f_x(alpha, x, y)\n f_y = pr_f_y(alpha, x)\n g_x = pr_g_x(alpha, x, y)\n g_y = pr_g_y(alpha, delta, x, y)\n pr_fg = f_y + g_x\n sq_fg = f_zn ** 2 + g_zn ** 2\n return 2 * (f_x * g_zn ** 2 + g_y * f_zn ** 2 - g_zn * f_zn * pr_fg) / sq_fg\n\n \ndef b(x, y, alpha, delta):\n f_zn = f(x, y, alpha)\n g_zn = g(x, y, alpha, delta)\n sq_fg = f_zn ** 2 + g_zn ** 2\n return ((g_zn * x) ** 2 + (f_zn * y) ** 2)/sq_fg\n # return ((f_zn * y * y) ** 2) / sq_fg\n \n\ndef sens(alpha, delta, h):\n max_m = 0\n min_m = 1000\n x_list, y_list = search_cycle(alpha, delta, h)\n k = len(x_list)\n # print(k)\n h_list = []\n r_list = []\n m_list = []\n t_list = []\n t1_list = []\n new_a = 0\n r_list.append(1)\n h_list.append(0)\n for i in range(1, k):\n a1 = A(x_list[i-1], y_list[i-1], alpha, delta)\n a2 = A(x_list[i], y_list[i], alpha, delta)\n # f_zn1 = f(x_list[i-1], y_list[i-1], alpha)\n # f_zn2 = f(x_list[i], y_list[i], alpha)\n # g_zn1 = g(x_list[i-1], y_list[i-1], alpha, delta)\n # g_zn1 = g(x_list[i - 1], y_list[i - 1], alpha, delta)\n # f_x = pr_f_x(alpha, x_list[i], y_list[i])\n # f_y = pr_f_y(alpha, x_list[i])\n # g_x = pr_g_x(alpha, x_list[i], y_list[i])\n # g_y = pr_g_y(alpha, delta, x_list[i], y_list[i])\n # pr_fg = f_y + g_x\n # sq_fg = f_zn ** 2 + g_zn ** 2\n # new_a += 2 * (f_x * g_zn ** 2 + g_y * f_zn ** 2 - g_zn * f_zn * pr_fg) * h / sq_fg\n new_a += (a1 + a2) * h / 2\n new_r = math.exp(new_a)\n r_list.append(new_r)\n new_h = h_list[i - 1] + h * (b(x_list[i - 1], y_list[i - 1], alpha, delta) / r_list[i - 1] + b(x_list[i], y_list[i], alpha, delta) / r_list[i]) / 2\n h_list.append(new_h)\n t1_list.append(i)\n const_c = r_list[- 1] * h_list[- 1] / (1 - r_list[- 1])\n # print(const_c)\n for i in range(k):\n m_list.append(r_list[i] * (const_c + h_list[i]))\n if m_list[i] < min_m:\n min_m = m_list[i]\n if m_list[i] > max_m:\n max_m = m_list[i]\n t_list.append(i)\n # matlab_export(r_list, h_list, \"r_h.txt\")\n # return m_list, t_list\n # return m_list, max_m, min_m, len(x_list)\n\n # print(max_m, min_m)\n return m_list, x_list, y_list\n\n\ndef main(alpha, delta, h):\n s, M, m, d = sens(alpha, delta, h)\n print(M, m, d)\n # x0, y0 = sens(alpha, delta, h)\n # plt.plot(x0, y0)\n # plt.grid(True)\n # pylab.show()\n\n\nif __name__ == '__main__':\n main(0.4, 0.1307, 0.01)\n","sub_path":"Сопромат/sensitivity_ag.py","file_name":"sensitivity_ag.py","file_ext":"py","file_size_in_byte":2722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"513523090","text":"#!/usr/bin/python\n# -*- encoding=GBK -*-\n__author__ = \"孙志宇\"\n__title__ = \"好物页面\"\nimport os\nimport sys\ncurPath = os.path.abspath(os.path.dirname(__file__))\nrootPath = os.path.split(curPath)[0]\nsys.path.append(rootPath)\n\nimport unittest\n\nimport logging\nfrom page.Card_Page.shoppingcard_page import ShoppingCard\n\nlogger = logging.getLogger(\"airtest\")\nlogger.setLevel(logging.ERROR)\n\n\nclass GoodThing(unittest.TestCase):\n def __init__(self, *args, **kwargs):\n from poco.drivers.android.uiautomation import AndroidUiautomationPoco\n poco = AndroidUiautomationPoco()\n unittest.TestCase.__init__(self, *args, **kwargs)\n self.poco = poco\n\n # 点击好物按钮\n def test1_classification(self):\n # 点击按钮\n self.poco(text=\"好物\").click()\n\n # 点击搜索框并输入商品名\n def test2_search_et(self, TradeName):\n self.poco(name=\"com.devkeep.mall:id/search_et\").click()\n self.poco(name=\"com.devkeep.mall:id/search_et\").set_text(TradeName)\n self.poco(name=\"com.devkeep.mall:id/search_btn\").click()\n # 判断上面搜索商品是否存在\n if TradeName in self.poco(name=\"com.devkeep.mall:id/goods_name\")[0].get_text():\n self.poco(name=\"com.devkeep.mall:id/cart_iv\")[0].click()\n # 判断商品是否有sku\n if len(self.poco(name=\"com.devkeep.mall:id/tag_tv\")) >= 1:\n self.poco(name=\"com.devkeep.mall:id/tag_tv\")[0].click()\n self.poco(name=\"com.devkeep.mall:id/cart_buy_tv\").click()\n else:\n print(\"----商品没改sku----\")\n print(\"----搜索商品存在并加入购物车----\")\n else:\n print(\"----搜索商品不存在----\")\n\n def test3_shoppingCard_bubble(self):\n self.poco(name=\"android.widget.ImageView\").click()\n return ShoppingCard\n\n\nif __name__ == \"__main__\":\n unittest.main()\n","sub_path":"page/ClassiFication_Page/GoodThing_page.py","file_name":"GoodThing_page.py","file_ext":"py","file_size_in_byte":1922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"40145366","text":"# Mad Lib\n# Tworzy opowiadanie oparte na szczegółach wprowadzonych przez uzytkownika\n\nfrom tkinter import *\n\nclass Application(Frame):\n \"\"\"Aplikacja oparta na GUI która tworzy opowiadanie\n na podstawie informacji wprowadzonych przez użytkownika\n \"\"\"\n def __init__(self, master):\n \"\"\"Inicjalizuj ramke\"\"\"\n super(Application,self).__init__(master)\n self.grid()\n self.create_widgets()\n\n def create_widgets(self):\n \"\"\"Tworzy widgety potrzebne do ramki\"\"\"\n # utwórz etykiete z instrukcją\n Label(self,\n text = \"Wprowadz informacje do nowego opowiadania\",\n ).grid(row = 0, column = 0, columnspan =2, sticky = W)\n\n # utwórz etykiete i pole znakowe służąace do wpisania imienia osoby\n Label(self,\n text = \"Osoba: \",\n ).grid(row = 1, column = 0, sticky = W)\n self.person_entry = Entry(self)\n self.person_entry.grid(row = 1, column = 1, sticky = W)\n\n # utwórz etykiete i pole znakowe słuzące do wpisania rzeczownika w liczbie mnogiej\n Label(self,\n text = \"Podaj rzeczownik w liczbie mnogiej:\",\n ).grid(row = 2, column = 0, sticky = W)\n self.noun_entry = Entry(self)\n self.noun_entry.grid(row = 2, column = 1, sticky = W)\n\n #utwóz etykiete i pole znakowe do wpisania czasownika\n Label(self,\n text = \"Podaj czasownik\",\n ).grid(row = 3 , column = 0, sticky = W)\n self.verb_entry = Entry(self)\n self.verb_entry.grid(row =3 , column = 1, sticky = W)\n\n #utwórz etykiete do pół wyboru przemiotników\n Label(self,\n text = \"Przymiotniki:\",\n ).grid(row = 4, column = 0, sticky = W)\n self.is_itchy = BooleanVar()\n Checkbutton(self,\n text = \"naglace\",\n variable = self.is_itchy\n ).grid(row = 4, column = 1, sticky = W)\n\n self.is_electric = BooleanVar()\n Checkbutton(self,\n text = \"elektryzujace\",\n variable = self.is_electric\n ).grid(row = 4, column = 2, sticky = W)\n\n self.is_joyus = BooleanVar()\n Checkbutton(self,\n text = \"radosne\",\n variable = self.is_joyus\n ).grid(row = 4, column = 3, sticky = W)\n\n Label(self,\n text = \"Czesci ciala:\",\n ).grid(row = 5, column = 0, sticky = W)\n self.body_part = StringVar()\n self.body_part.set(None)\n body_parts = [\"pepek\",\"noga\",\"nerka\"]\n column = 1\n for part in body_parts:\n Radiobutton(self,\n text = part,\n variable = self.body_part,\n value = part\n ).grid(row = 5, column = column, sticky = W)\n column += 1\n\n # przycisk akceptacji danych\n Button(self,\n text = \"Kliknij aby wyświetlic opowiadanie\",\n command = self.tell_story\n ).grid(row = 6, column = 0, sticky = W)\n self.story_text = Text(self, width = 75, height = 10, wrap = WORD)\n self.story_text.grid(row = 7, column = 0, sticky = W)\n\n def tell_story(self):\n \"\"\"Wpisz w pole tekstowe opowiadanie oparte na danych uzyttkownika\"\"\"\n # pobierz wartosci interfejsu gui\n person = self.person_entry.get()\n noun = self.noun_entry.get()\n verb = self.verb_entry.get()\n adjectives = \"\"\n if self.is_itchy.get():\n adjectives += \"naglące\"\n if self.is_joyus.get():\n adjectives += \"radosne\"\n if self.is_electric.get():\n adjectives += \"elektryzujace\"\n\n body_part = self.body_part.get()\n\n # create the story\n story = \"Uzytkownik tego programu\"\n story += person\n story += \"chciał nauczyc się programowania\"\n story += \"jest on\"\n story += noun\n story += adjectives\n story += verb\n story += body_part + \".\"\n\n self.story_text.delete(0.0, END)\n self.story_text.insert(0.0, story)\n\nroot = Tk()\nroot.title(\"Mad Lib\")\napp = Application(root)\nroot.mainloop()\n\n\n","sub_path":"GUI Exercises/Mad Lib.py","file_name":"Mad Lib.py","file_ext":"py","file_size_in_byte":4268,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"244714580","text":"import urllib.request, urllib.parse, urllib.error\nimport json\nimport ssl\n# Ignore SSL certificate errors\nctx = ssl.create_default_context()\nctx.check_hostname = False\nctx.verify_mode = ssl.CERT_NONE\nserviceurl = 'http://py4e-data.dr-chuck.net/json?'\naddress = input('Enter location: ')\nurl = serviceurl + urllib.parse.urlencode({'address': address})\nprint('Retrieving',url)\nuh = urllib.request.urlopen(url,context=ctx)\ndata = uh.read().decode()\nprint('Retrieved',len(data),'charactors')\njs = json.loads(data)\nif not js or 'status' not in js or js['status'] != 'OK':\n print('==== Failure To Retrieve ====')\n print(data)\n exit()\nPlace_id = js[\"results\"][0]['place_id']\nprint(Place_id)\n","sub_path":"14.2.py","file_name":"14.2.py","file_ext":"py","file_size_in_byte":692,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"194771699","text":"'''\nGiven an array, count the unique values\n\n'''\n\n\ndef unique_value(arr):\n if len(arr) <= 1:\n return arr\n\n i = 0\n for j in range(1, len(arr)):\n if arr[i] != arr[j]:\n i += 1\n arr[i] = arr[j]\n\n return arr[:i+1]\n\n\nprint(unique_value([1, 1, 2, 2, 4, 5, 6, 6, 7]))\n","sub_path":"count_unique.py","file_name":"count_unique.py","file_ext":"py","file_size_in_byte":308,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"322616647","text":"import csv\nimport io\nimport logging\nimport os\nimport sys\nfrom tqdm import tqdm\n_BUCKET_NAME = sys.argv[1]\n\nfrom google.cloud import storage\nclient = storage.Client()\nbucket = client.get_bucket(_BUCKET_NAME)\n\ndef process():\n print(\"start processing bucket\", _BUCKET_NAME)\n if os.path.isfile('fulldata.csv'):\n os.remove('fulldata.csv')\n blobs = list(bucket.list_blobs())\n with open(\"fulldata.csv\", \"a\") as f:\n for blob in tqdm(blobs):\n try:\n user_knowledge, quality, label, img_name = blob.name.split(\"/\")\n if str(img_name).endswith(\".jpg\"):\n f.write(\"gs://\"+_BUCKET_NAME+\"/\"+blob.name+\",\"+quality+\",\"+label+\",\"+img_name+\"\\n\")\n except:\n pass\n\nprocess()\n\n","sub_path":"cloud/reading_files_bucket.py","file_name":"reading_files_bucket.py","file_ext":"py","file_size_in_byte":764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"114783398","text":"# Package imports\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom testCases import *\nimport sklearn\nimport sklearn.datasets\nimport sklearn.linear_model\nfrom planar_utils import plot_decision_boundary, sigmoid, load_planar_dataset, load_extra_datasets\n\nnp.random.seed(1) # set a seed so that the results are consistent\n\nX, Y = load_planar_dataset()\nplt.scatter(X[0, :], X[1, :], c=np.squeeze(Y), s=40, cmap=plt.cm.Spectral)\nplt.show()\nshape_X = X.shape\nshape_Y = Y.shape\nm = shape_X[1]\nprint('The shape of X is: '+str(shape_X))\nprint('The shape of Y is: '+str(shape_Y))\nprint('I have m = %d training example.' %(m))\nclf = sklearn.linear_model.LogisticRegressionCV()\nclf.fit(X.T, Y.T)\nplot_decision_boundary(lambda x: clf.predict(x), X, Y)\nplt.title(\"Logistic Regression\")\nLR_predictions = clf.predict(X.T)\nprint(LR_predictions)\nprint ('Accuracy of logistic regression: %d ' %\n float((np.dot(Y,LR_predictions) + np.dot(1-Y,1-LR_predictions))/float(Y.size)*100) +\n '% ' + \"(percentage of correctly labelled datapoints)\")\n\ndef layer_sizes(X, Y):\n n_x = X.shape[0]\n n_h = 4\n n_y = Y.shape[0]\n return (n_x, n_h, n_y)\n\ndef initialize_parameters(n_x, n_h, n_y):\n \"\"\"\n Argument:\n n_x -- size of the input layer\n n_h -- size of the hidden layer\n n_y -- size of the output layer\n\n Returns:\n params -- python dictionary containing your parameters:\n W1 -- weight matrix of shape (n_h, n_x)\n b1 -- bias vector of shape (n_h, 1)\n W2 -- weight matrix of shape (n_y, n_h)\n b2 -- bias vector of shape (n_y, 1)\n \"\"\"\n np.random.seed(2)\n W1 = np.random.randn(n_h, n_x)*0.01\n b1 = np.zeros((n_h, 1))\n W2 = np.random.randn(n_y, n_h)*0.01\n b2 = np.zeros((n_y, 1))\n assert(W1.shape == (n_h, n_x))\n assert(b1.shape == (n_h, 1))\n assert(W2.shape == (n_y, n_h))\n assert(b2.shape == (n_y, 1))\n parameters = {\"W1\": W1, \"b1\": b1, \"W2\": W2, \"b2\": b2}\n return parameters\n\ndef forward_propagation(X, parameters):\n \"\"\"\n Argument:\n X -- input data of size (n_x, m)\n parameters -- python dictionary containing your parameters(output of initialization function)\n return:\n A2 -- The sigmoid output of the second activation\n cache -- a dictionary containing \"Z1\",\"A1\",\"Z2\" and \"A2\"\n\n \"\"\"\n W1 = parameters[\"W1\"]\n b1 = parameters[\"b1\"]\n W2 = parameters[\"W2\"]\n b2 = parameters[\"b2\"]\n Z1 = np.dot(W1, X) + b1\n A1 = np.tanh(Z1)\n Z2 = np.dot(W2, A1) + b2\n A2 = 1/(1+np.exp(-Z2))\n # print(A2.shape)\n assert(A2.shape == (1, X.shape[1]))\n cache = {\"Z1\": Z1, \"A1\": A1, \"Z2\": Z2, \"A2\": A2}\n return A2, cache\n\n# GRADED FUNCTION: compute_cost\ndef compute_cost(A2, Y, parameters):\n \"\"\"\n Computes the cross-entropy cost given in equation (13)\n\n Arguments:\n A2 -- The sigmoid output of the second activation, of shape (1, number of examples)\n Y -- \"true\" labels vector of shape (1, number of examples)\n parameters -- python dictionary containing your parameters W1, b1, W2 and b2\n\n Returns:\n cost -- cross-entropy cost given equation (13)\n \"\"\"\n m = Y.shape[1]\n logprobs = np.multiply(np.log(A2), Y) + np.multiply(np.log(1-A2), 1-Y)\n cost = -np.sum(logprobs)/m\n cost = np.squeeze(cost)\n assert(isinstance(cost, float)) # 判断cost 是否是float类型\n return cost\n\n# GRADED FUNCTION: backward_propagation\n\ndef backward_propagation(parameters, cache, X, Y):\n \"\"\"\n Implement the backward propagation using the instructions above.\n Arguments:\n parameters -- python dictionary containing our parameters\n cache -- a dictionary containing \"Z1\", \"A1\", \"Z2\" and \"A2\".\n X -- input data of shape (2, number of examples)\n Y -- \"true\" labels vector of shape (1, number of examples)\n Returns:\n grads -- python dictionary containing your gradients with respect to different parameters\n \"\"\"\n m = X.shape[1]\n W1 = parameters[\"W1\"]\n W2 = parameters[\"W2\"]\n A1 = cache[\"A1\"]\n A2 = cache[\"A2\"]\n dZ2 = A2 - Y\n dW2 = np.dot(dZ2, A1.T)/m\n db2 = np.sum(dZ2, axis=1, keepdims=True)/m\n dZ1 = np.multiply(np.dot(W2.T, dZ2), (1 - np.power(A1, 2)))\n dW1 = np.dot(dZ1, X.T)/m\n db1 = np.sum(dZ1, axis=1, keepdims=True)/m\n\n grads = {\"dW1\": dW1, \"db1\": db1, \"dW2\": dW2, \"db2\": db2}\n return grads\n\n# GRADED FUNCTION: update_parameters\n\ndef update_parameters(parameters, grads, learning_rate=1.2):\n \"\"\"\n Updates parameters using the gradient descent update rule given above\n Arguments:\n parameters -- python dictionary containing your parameters\n grads -- python dictionary containing your gradients\n Returns:\n parameters -- python dictionary containing your updated parameters\n \"\"\"\n W1 = parameters[\"W1\"]\n b1 = parameters[\"b1\"]\n W2 = parameters[\"W2\"]\n b2 = parameters[\"b2\"]\n\n dW1 = grads[\"dW1\"]\n db1 = grads[\"db1\"]\n dW2 = grads[\"dW2\"]\n db2 = grads[\"db2\"]\n\n W1 = W1 - learning_rate*dW1\n b1 = b1 - learning_rate*db1\n W2 = W2 - learning_rate*dW2\n b2 = b2 - learning_rate*db2\n\n parameters = {\"W1\": W1, \"b1\": b1, \"W2\": W2, \"b2\": b2}\n return parameters\n\n# GRADED FUNCTION: nn_model\ndef nn_model(X, Y, n_h, num_iterations=10000, print_cost=False):\n \"\"\"\n Arguments:\n X -- dataset of shape (2, number of examples)\n Y -- labels of shape (1, number of examples)\n n_h -- size of the hidden layer\n num_iterations -- Number of iterations in gradient descent loop\n print_cost -- if True, print the cost every 1000 iterations\n Returns:\n parameters -- parameters learnt by the model. They can then be used to predict.\n \"\"\"\n np.random.seed(3)\n n_x = layer_sizes(X, Y)[0]\n n_y = layer_sizes(X, Y)[2]\n parameters=initialize_parameters(n_x, n_h, n_y)\n W1 = parameters[\"W1\"]\n b1 = parameters[\"b1\"]\n W2 = parameters[\"W2\"]\n b2 = parameters[\"b2\"]\n for i in range(0, num_iterations):\n A2, cache = forward_propagation(X, parameters)\n cost = compute_cost(A2, Y, parameters)\n grads = backward_propagation(parameters, cache, X, Y)\n parameters = update_parameters(parameters, grads)\n if print_cost and i % 1000 == 0:\n print(\"cost after iteratin %i:%f\"%(i, cost))\n return parameters\n\n\n# GRADED FUNCTION: predict\n\ndef predict(parameters, X):\n \"\"\"\n Using the learned parameters, predicts a class for each example in X\n Arguments:\n parameters -- python dictionary containing your parameters\n X -- input data of size (n_x, m)\n Returns\n predictions -- vector of predictions of our model (red: 0 / blue: 1)\n \"\"\"\n A2, cache = forward_propagation(X, parameters)\n prediction = (A2 > 0.5)\n return prediction\n\nparameters = nn_model(X, Y, n_h=4, num_iterations=10000, print_cost=True)\nplt.title(\"Decision Boundary for hidden layer size \" + str(4))\nplot_decision_boundary(lambda x: predict(parameters, x.T), X, Y)\n\npredictions = predict(parameters, X)\nprint(\"Accuracy: %d\" % float((np.dot(Y, predictions.T) + np.dot(1-Y, 1-predictions.T))/float(Y.size)*100)+\"%\")\n\n# 调整隐藏神经元的数目观察结果\nplt.figure()\nhidden_layer_sizes = [1, 2, 3, 4, 5, 20, 50]\n# enumerate() 同时返回索引和值\nfor i, n_h, in enumerate(hidden_layer_sizes):\n plt.subplot(5, 2, i+1)\n plt.title(\"Hidden Layer of size %d\" % n_h)\n parameters = nn_model(X, Y, n_h, num_iterations=5000)\n plot_decision_boundary(lambda x: predict(parameters, x.T), X, Y)\n predictions = predict(parameters, X)\n accuracy = float((np.dot(Y, predictions.T) + np.dot(1-Y, 1-predictions.T))/float(Y.size)*100)\n print(\"Accuracy for {} hidden units: {} %\".format(n_h, accuracy))\n","sub_path":"第一课第三周编程作业/第一课第三周编程作业/assignment3/one_three.py","file_name":"one_three.py","file_ext":"py","file_size_in_byte":7762,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"264740873","text":"#!/usr/bin/python\n\n# How many followers do you have?\nimport urllib.request\nimport re\n\nfeeds = [\n 'rbowen','centosproject','theasf','realDonaldTrump'\n ];\nfor feed in feeds:\n req = urllib.request.Request( 'https://twitter.com/' + feed,\n data = None,\n headers={\n 'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.1916.47 Safari/537.36'\n } )\n f = urllib.request.urlopen(req)\n html = f.read().decode('utf-8')\n # print (html)\n\n# Looks like ...\n#
    2,615
    \n#
    Followers
    \n\n print ( feed + ': ' + re.search('.*?followers\">.+?statnum\">([\\d,MK]+).*?<.*?statlabel\"> Followers.*', html, re.DOTALL).group(1) )\n\n","sub_path":"followers.py","file_name":"followers.py","file_ext":"py","file_size_in_byte":798,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"471377301","text":"from bs4 import BeautifulSoup\nimport unicodedata\nimport html\nfrom pyspark import SparkContext, SparkConf\nimport argparse\nimport json\n\nclosings = [\n \"Regards\",\n \"Reg\",\n \"Best,\",\n \"Thanks,\",\n \"Sent from my iPhone\",\n \"Sent from my ipad\",\n \"Sent from my android\",\n \"Sent from my mobile device\",\n \"Sincerely\",\n \"Yours truly\",\n \"Yours sincerely\",\n \"Best regards\",\n \"Cordially\",\n \"Yours respectfully\",\n \"Warm regards\",\n \"Best wishes\",\n \"With appreciation\",\n \"Cordially yours\",\n \"Fond regards\",\n \"In appreciation\",\n \"In sympathy\",\n \"Kind regards\",\n \"Kind thanks\",\n \"Kind wishes\",\n \"Many thanks\",\n \"Regards\",\n \"Respectfully\",\n \"Respectfully yours\",\n \"Sincerely\",\n \"Sincerely yours\",\n \"Warm regards\",\n \"Warm wishes\",\n \"Warmly\",\n \"With appreciation\",\n \"With deepest sympathy\",\n \"With gratitude\",\n \"With sincere thanks\",\n \"With sympathy\",\n \"Your help is greatly appreciated\",\n \"Yours cordially\",\n \"Yours faithfully\",\n \"Yours sincerely\",\n \"Yours truly\",\n \"From:\",\n \"Sent:\"\n]\n\nclosings = closings + [word.lower() for word in closings]\n\nclosings = closings + [word.lower() for word in closings]\n\n\ndef split(txt, seps):\n default_sep = seps[0]\n\n # we skip seps[0] because that's the default seperator\n for sep in seps[1:]:\n txt = txt.replace(sep, default_sep)\n return [i.strip() for i in txt.split(default_sep)]\n\n\ndef remove_html(doc_tuple):\n doc_id, raw = doc_tuple\n soup = BeautifulSoup(\n raw, 'lxml') # create a new bs4 object from the html data loaded\n for script in soup([\"script\",\n \"style\"]): # remove all javascript and stylesheet code\n script.extract()\n # get text\n text = soup.get_text()\n # break into lines and remove leading and trailing space on each\n lines = (line.strip() for line in text.splitlines())\n # break multi-headlines into a line each\n chunks = (phrase.strip() for line in lines for phrase in line.split(\" \"))\n # encode unicode characters\n text = unicodedata.normalize(\"NFKD\", text)\n # encode html characters\n text = html.unescape(text)\n text = split(text, closings)[0]\n return {'id': doc_id, 'body': text}\n\n\nif __name__ == \"__main__\":\n\n desc = 'remove html tags from email text'\n parser = argparse.ArgumentParser(\n description=desc,\n formatter_class=argparse.RawDescriptionHelpFormatter,\n epilog=desc)\n parser.add_argument(\"-i\", \"--input_path\", help=\"directory with json texts\")\n parser.add_argument(\n \"-o\",\n \"--output_path\",\n help=\n \"output directory for spark results of json texts with html tags removed\"\n )\n args = parser.parse_args()\n conf = SparkConf().setAppName(\"Html Tag Removal\")\n sc = SparkContext(conf=conf)\n rdd = sc.textFile(args.input_path)\n\n def doc_to_tuple(sz):\n j = json.loads(sz)\n return (j.get('id'), j.get('body'))\n\n cleandoc = rdd.map(doc_to_tuple).map(remove_html).cache()\n\n output = cleandoc.map(lambda x: json.dumps(x))\n\n output.saveAsTextFile(args.output_path)\n","sub_path":"spark/rmhtml.py","file_name":"rmhtml.py","file_ext":"py","file_size_in_byte":3137,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"239080006","text":"from query.report_base import ReportBase\n\nclass CNT01(ReportBase):\n \n def get_aggregate(cls,args):\n #{\"status\":\"processing\"},\n _aggregate = [\n {\"$match\": {\"$and\" : [\n { \"history.0.dept\": args['dept'] },\n {\"history.0.date\":{\"$gte\":args['startDate'],\"$lt\":args['endDate']}},\n {\"name\":'Contact'}\n ]\n }\n\n },\n {\"$sort\": {\"no\":1}}\n ] \n return _aggregate\n\n def run(cls,server,args):\n #此表有申請部門權限, 只能查自己部門的報表, 故需檢查avatar 部門=history.0.dept\n def check_args(args):\n checked = {}\n if 'year' not in args or 'month' not in args or 'avatar' not in args:\n return {\"error\":\"查詢條件沒有全部輸入!無法查詢\"}\n\n try:\n if 'userid' in args and args['userid'] is not None and args['userid'].strip()!= '':\n checked.update({'userid':args['userid']})\n checked.update({'startDate':cls.cvt_startDate('{}/{}/01'.format(args['year'],args['month']))})\n checked.update({'endDate':cls.cvt_firstDayNextMonth(args['year'],args['month'])})\n checked.update({\"dept\":args[\"avatar\"][\"dept_id\"]})\n checked.update({\"avatar\":args[\"avatar\"]})\n except Exception as e:\n return {\"error\":\"查詢條件格式錯誤!無法查詢-{}\".format(str(e))} \n\n return checked \n \n #檢查參數\n newargs = check_args(args)\n if \"error\" in newargs:\n return newargs\n \"\"\"\n newargs = {}\n newargs.update({'dept':args['dept']})\n newargs.update({'startDate':cls.cvt_startDate('{}/{}/01'.format(args['year'],args['month']))})\n newargs.update({'endDate':cls.cvt_firstDayNextMonth(args['year'],args['month'])})\n \"\"\"\n pipeline = cls.get_aggregate(newargs)\n if 'userid' in newargs:\n pipeline[0][\"$match\"][\"$and\"].append({\"history.0.userid\": args['userid']})\n data = cls.mongo(server,'headway','flow',pipeline)\n #db = get_mongo()['headway']['flow']\n #data = list(db.aggregate(pipeline))\n result = []\n result.append([\"項目\",\"編號\",\"委託名稱\",\"委託人\",\"執行人\",\"狀態\",\"目前站別\"])\n for item in data:\n temp = ['','','','','','','']\n temp[0]=data.index(item) + 1\n temp[1]=item[\"no\"]\n temp[2]=item[\"history\"][0][\"data\"][0][\"value\"]\n temp[3]=item[\"history\"][0][\"username\"]\n temp[6]=''\n for history in item[\"history\"][::-1]:\n if history[\"id\"]==\"ExecutorRes\":\n temp[4]=history[\"username\"]\n break\n temp[5]=cls.toStatus(item[\"status\"])\n\n result.append(temp)\n return cls.result(result)","sub_path":"app/query/report/CNT01.py","file_name":"CNT01.py","file_ext":"py","file_size_in_byte":3023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"48233587","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport argparse\nimport json\nimport queue\nimport shlex\nimport os\nimport re\nimport subprocess\nimport threading\nimport time\nimport urllib.parse\nfrom datetime import date, datetime\nfrom glob import glob\n\nimport pypandoc\n\n\nYEAR_DEATHS = \"{year}\"\nMONTH_DEATHS = \"Deaths_in_{month}_{year}\"\nDAY_DEATHS = \"{month}_{day\"\nFIND_DEATHS_SECTION = re.compile(r'^== *Deaths *==', re.MULTILINE)\nFIND_NEXT_SECTION = re.compile(r'^== *[^=]', re.MULTILINE)\nPERSON_LINK = re.compile(r'\\[\\[([^\\[]*?)\\]\\], .*?\\(b.')\n\nINFOBOX_MARKER = re.compile(r'^ *{{ *Infobox.*', re.IGNORECASE)\nINFOBOX_TOKENIZER = re.compile(r'({{ *Infobox|{{|}})', re.IGNORECASE)\nKEY_VALUE_PAIR = re.compile(r'\\| *(\\S+) *= *(.*)', re.IGNORECASE | re.MULTILINE)\nCLUTTER = re.compile(r'{{|}}|\\[\\[|\\]\\]')\nDATE_ODD = re.compile(r'(\\d{1,2}) death year and age.*?(\\d{4}).*\\| *(\\d{1,2})', re.IGNORECASE)\nDATE = re.compile(r'.*? *\\| *(\\d{4}) *\\| *(\\d+) *\\| *(\\d+).*')\nDATE2 = re.compile(r'OldStyleDate *\\| *(\\d+) *([a-z]*) *\\| *(\\d{4}).*', re.IGNORECASE)\nDATE3 = re.compile(r'(\\d+) +([a-z]*),? death year and age *\\| *(\\d{4}).*', re.IGNORECASE)\nDATE4 = re.compile(r'([a-z]*) +(\\d+),? death year and age *\\| *(\\d{4}).*', re.IGNORECASE)\nDATE_UNCERTAIN_DAY = re.compile(r'(\\d+)(?:[/-–]\\d+| or \\d+) +([a-z]+),? +(\\d{4})', re.IGNORECASE)\nDATE_UNCERTAIN_DAY2 = re.compile(r'([a-z]+) +(\\d+)(?:[/-–]\\d+| or \\d+),? +(\\d{4})', re.IGNORECASE)\nDATE_UNCERTAIN_DATE = re.compile(r'([a-z]+) +(\\d+),? or [a-z]+ +\\d+,? +(\\d{4})', re.IGNORECASE)\nDATE_YEAR_RANGE = re.compile(r'(\\d{4}) *(?:[-–]|or) *(\\d{4})', re.IGNORECASE)\nEXTRA_AGE = re.compile(r'(, *aged? *\\d*|\\(aged? *\\d*)|age *\\d+', re.IGNORECASE)\nEXTRA_REF = re.compile(r'{{ *(ref|efn|Cn).*}}')\nEXTRA_TAGS = re.compile(r'|$)||', re.IGNORECASE)\nEXTRA_PARENTHESES = re.compile(r'\\(.*\\)', re.IGNORECASE)\nEXTRA_CIRCA = re.compile(r'c\\.|circa|ca\\.', re.IGNORECASE)\nEXTRA_QUOTES = re.compile(r'[\\']', re.IGNORECASE)\nEXTRA_NOWRAP = re.compile(r'nowrap *\\|', re.IGNORECASE)\nEXTRA_BR = re.compile(r'
    ', re.IGNORECASE)\nEXTRA_ADOPTION_OF_CALENDAR = re.compile(r'Adoption *of *the *Gregorian.*', re.IGNORECASE)\n\nCHILD_ARTICLE = re.compile(r'{{ *main article *\\|(.*?)}}', re.IGNORECASE)\n\nclass Worker(threading.Thread):\n def __init__(self, thread_id, name_queue, name_queue_lock, args, function):\n super().__init__()\n self.thread_id = thread_id\n self.name_queue = name_queue\n self.name_queue_lock = name_queue_lock\n self.args = args\n self.function = function\n\n def run(self):\n while True:\n with self.name_queue_lock:\n if self.name_queue.empty():\n break\n\n name = self.name_queue.get()\n\n self.function(self.args, name)\n\n\ndef download_article(args, title, output):\n os.makedirs(os.path.dirname(output), exist_ok=True)\n if os.path.exists(output):\n if args.verbose:\n print(\" file exists: {}\".format(output))\n\n return\n\n url = \"https://en.wikipedia.org/w/api.php?action=query&titles={title}&prop=revisions&rvprop=content&format=json\" \\\n .format(title=urllib.parse.quote(title))\n command = \"curl -s -S -z {output} -o {output} {url}\".format(url=shlex.quote(url), output=shlex.quote(output))\n if args.verbose:\n print(\" running: {}\".format(command))\n\n subprocess.call(shlex.split(command))\n\n\ndef grab_year(args, year):\n if args.verbose:\n print(\"fetching year {}...\".format(year))\n\n title = YEAR_DEATHS.format(year=year)\n output = \"years/year-{year}.json\".format(year=year)\n download_article(args, title, output)\n\n\ndef grab_names_from_content(content):\n results = []\n for match in PERSON_LINK.finditer(content):\n name = match.group(1)\n chunks = name.split(\"|\")\n if len(chunks) > 1:\n name = chunks[0].strip()\n\n results.append(name)\n\n return results\n\n\ndef grab_names_from_file(input_file):\n results = []\n data = json.load(open(input_file))\n for unused_pageid, page_data in data[\"query\"][\"pages\"].items():\n if \"revisions\" not in page_data:\n continue\n\n content = page_data[\"revisions\"][0][\"*\"]\n deaths_section_match = FIND_DEATHS_SECTION.search(content)\n next_section_match = FIND_NEXT_SECTION.search(content, deaths_section_match.end())\n end = len(content)\n if next_section_match:\n end = next_section_match.start()\n\n section = content[deaths_section_match.start():end]\n\n results += grab_names_from_content(section)\n\n return results\n\n\ndef parse_date(value):\n value = value.strip(\", |\")\n if not value:\n return None\n\n value = EXTRA_PARENTHESES.sub('', value)\n value = EXTRA_CIRCA.sub('', value)\n value = EXTRA_NOWRAP.sub('', value)\n value = EXTRA_QUOTES.sub('', value)\n value = EXTRA_AGE.sub('', value)\n value = EXTRA_ADOPTION_OF_CALENDAR.sub('', value)\n value = value.replace(\"ndash\", \"-\")\n value = value.replace(\"baptized\", \"\")\n value = value.strip(\", |\")\n\n specials = {\n \"February or March 1945\": date(1945, 2, 1),\n \"February or March, 1945\": date(1945, 2, 1),\n \"1850s\": date(1850, 1, 1),\n \"1860s\": date(1860, 1, 1),\n \"Between October 4, 1919 and January 2, 1920\": date(1920, 10, 4),\n \"Unknown\": None,\n \"29 February 1900\": date(1900, 2, 18), # different calendar in eastern Europe\n \"September 13, 1922, or September 10, 1923\": date(1922, 9, 13),\n \"Death year and age|1992|1912|4|23\": date(1992, 4, 23),\n \"late 1867 or early 1868\": date(1867, 12, 31),\n \"around 1840\": date(1840, 1, 1),\n \"September 23, 1968 or 1969\": date(1968, 9, 23),\n \"FETCH_WIKIDATA\": None,\n \"Month? Day?, 1879\": date(1879, 1, 1),\n \"Unknown, 1677 and 1736 claimed\": date(1677, 1, 1),\n \"Kathryn Johanna Kuhlman\": date(1907, 5, 9),\n }\n if value in specials:\n return specials[value]\n\n # 7 {{Death year and age|df=yes|1962|1896|9}}\n # must be tried before DATE, else that matches\n mo = DATE_ODD.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%d %Y %m\")\n except ValueError:\n pass\n\n mo = DATE.match(value)\n if mo:\n return date(*map(int, mo.groups()))\n\n mo = DATE2.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%d %B %Y\")\n except ValueError:\n pass\n\n mo = DATE3.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%d %B %Y\")\n except ValueError:\n pass\n\n mo = DATE4.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%B %d %Y\")\n except ValueError:\n pass\n\n mo = DATE_YEAR_RANGE.match(value)\n if mo:\n return date(int(mo.group(1)), 1, 1)\n\n mo = DATE_UNCERTAIN_DAY.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%d %B %Y\")\n except ValueError:\n pass\n\n mo = DATE_UNCERTAIN_DAY2.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%B %d %Y\")\n except ValueError:\n pass\n\n mo = DATE_UNCERTAIN_DATE.match(value)\n if mo:\n chunk = \" \".join(mo.groups())\n try:\n return datetime.strptime(chunk, \"%B %d %Y\")\n except ValueError:\n pass\n\n formats = [\"%B %d, %Y\",\n \"%B %d , %Y\",\n \"%B %d %Y\",\n \"%d %B %Y\",\n \"%d %B, %Y\",\n \"%d %B , %Y\",\n \"%B %Y\",\n \"%Y-%m-%d\",\n \"%Y\"]\n value = EXTRA_BR.sub(\"|\", value)\n for chunk in value.split(\"|\"):\n chunk = chunk.strip()\n for date_format in formats:\n try:\n return datetime.strptime(chunk, date_format)\n except ValueError:\n pass\n\n raise Exception(\"couldn't parse this as date: {}\".format(value))\n\n\ndef parse_infobox(data, name):\n if not INFOBOX_MARKER.match(data):\n return\n\n result = {}\n for match in KEY_VALUE_PAIR.finditer(data):\n key, value = match.groups()\n if key not in (\"birth_date\", \"birth_place\",\n \"death_date\", \"death_place\", \"death_cause\",\n \"background\", \"occupation\"):\n continue\n\n value = value.replace(\"{{Greece Old Style dating}}\", \"\")\n value = EXTRA_TAGS.sub('', value)\n value = EXTRA_REF.sub('', value)\n value = CLUTTER.sub('', value)\n\n if key.endswith(\"_date\"):\n if name.startswith(\"Auguste and Louis\"):\n if key == \"birth-date\":\n value = date(1862, 10, 19)\n else:\n value = date(1954, 4, 10)\n else:\n value = parse_date(value)\n else:\n if value.startswith(\"hlist\"):\n value = list(map(lambda x: x.lower(), value.split(\"|\")[1:]))\n\n result[key] = value\n\n\n return result\n\n\ndef parse_article(filename):\n json_data = json.load(open(filename))\n page_data = list(json_data[\"query\"][\"pages\"].items())[0][1]\n if \"revisions\" not in page_data:\n return\n\n content = page_data[\"revisions\"][0][\"*\"]\n name = os.path.splitext(os.path.basename(filename))[0]\n child_articles = CHILD_ARTICLE.findall(content)\n\n result = {\n \"name\": name,\n \"article_size\": len(content),\n \"child_article_count\": len(child_articles),\n }\n\n tokens = INFOBOX_TOKENIZER.split(content)\n started = False\n level = 0\n block = \"\"\n for i, token in enumerate(tokens):\n if not started:\n if INFOBOX_MARKER.match(token):\n started = True\n level += 1\n block += token\n else:\n block += token\n if token == \"{{\":\n level += 1\n elif token == \"}}\":\n level -= 1\n if level == 0:\n break\n\n infobox = parse_infobox(block, name)\n if infobox:\n result.update(infobox)\n\n # print(result)\n return result\n\n\ndef grab_years(args, start, end):\n name_queue_lock = threading.Lock()\n name_queue = queue.Queue()\n for year in range(start, end + 1):\n name_queue.put(year)\n\n threads = []\n for i in range(args.thread_count):\n t = Worker(i, name_queue, name_queue_lock, args, grab_year)\n t.start()\n threads.append(t)\n\n while True:\n time.sleep(1)\n with name_queue_lock:\n if name_queue.empty():\n break\n\n\ndef grab_names(args, input_files, output_file):\n names = []\n for input_file in input_files:\n names += grab_names_from_file(input_file)\n\n open(args.names_file, \"a\").write(\"\\n\".join(names) + \"\\n\")\n\n\ndef download_name(args, name):\n if args.verbose:\n print(\"fetching article {}...\".format(name))\n\n output = \"article/{title}.json\".format(title=name)\n download_article(args, name, output)\n\n\ndef download_names(args, input_file):\n name_queue_lock = threading.Lock()\n name_queue = queue.Queue()\n for name in open(input_file):\n name = name.strip()\n name_queue.put(name)\n\n threads = []\n for i in range(args.thread_count):\n t = Worker(i, name_queue, name_queue_lock, args, download_name)\n t.start()\n threads.append(t)\n\n while True:\n time.sleep(1)\n with name_queue_lock:\n if name_queue.empty():\n break\n\n\ndef parse_articles(args, input_files):\n for filename in input_files:\n parse_article(filename)\n\n\ndef generate_data(args):\n for filename in glob(\"article/*.json\"):\n result = parse_article(filename)\n if not result:\n continue\n\n print(result[\"article_size\"] + result[\"child_article_count\"] * 30000, result[\"name\"])\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-v\", \"--verbose\", action=\"store_true\", default=False, help=\"verbose output\")\n parser.add_argument(\"--years\", nargs=2, metavar=(\"START\", \"END\"), type=int, help=\"grab years from START to END (inclusive)\")\n parser.add_argument(\"--names\", nargs=\"+\", metavar=\"FILE\", help=\"grab person names from given input JSON files\")\n parser.add_argument(\"-d\", \"--download-names\", action=\"store_true\", help=\"download articles for names in NAMES-FILE\")\n parser.add_argument(\"-n\", \"--names-file\", default=\"names.txt\", metavar=\"NAMES-FILE\", help=\"file to write parsed names to as well as read them from, it is appended to\")\n parser.add_argument(\"-p\", \"--parse\", nargs=\"+\", metavar=\"FILE\", help=\"parse articles\")\n parser.add_argument(\"-t\", \"--thread-count\", type=int, default=10, metavar=\"NUM\", help=\"number of threads for downloading (default: 10)\")\n parser.add_argument(\"-g\", \"--generate\", action=\"store_true\", help=\"generate data file\")\n args = parser.parse_args()\n if args.years:\n grab_years(args, args.years[0], args.years[1])\n\n if args.names:\n grab_names(args, args.names, args.names_file)\n\n if args.download_names:\n download_names(args, args.names_file)\n\n if args.parse:\n parse_articles(args, args.parse)\n\n if args.generate:\n generate_data(args)\n","sub_path":"grab_deaths.py","file_name":"grab_deaths.py","file_ext":"py","file_size_in_byte":13559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"186796076","text":"#!/usr/bin/python\n\nnumN = int(input(\"How many numbers will you enter?\"))\nsum = 0\n\nprint(\"Input num\")\n\nnum = [int(x)for x in input().split()]\n\nfor i in range(0,numN):\n\tsum = num[i] + sum\n\ntotal = sum/numN\n\nprint(\"Average : \", total)\n\n","sub_path":"py_lab/aver_num.py","file_name":"aver_num.py","file_ext":"py","file_size_in_byte":233,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"525148291","text":"# -*- coding: utf-8 -*-\n\n\nimport numpy as np\n\nfrom .base import Layer\nfrom ..activation import Tanh\nfrom ..initialization import GlorotUniform\nfrom ..initialization import Orthogonal\nfrom ..initialization import Zero\n\n\nclass SimpleRNN(Layer):\n \"\"\"Fully-connected RNN where the output is to be fed back to input.(完全连接的RNN在输出将被反馈到输入。)\n\n # Arguments\n output_dim: dimension of the internal projections and the final output.\n init: weight initialization function.\n Can be the name of an existing function (str),\n or a Theano function (see: [initializations](../initializations.md)).\n inner_init: initialization function of the inner cells.\n activation: activation function.\n Can be the name of an existing function (str),\n or a Theano function (see: [activations](../activations.md)).\n W_regularizer: instance of [WeightRegularizer](../regularizers.md)\n (eg. L1 or L2 regularization), applied to the input weights matrices.\n U_regularizer: instance of [WeightRegularizer](../regularizers.md)\n (eg. L1 or L2 regularization), applied to the recurrent weights matrices.\n b_regularizer: instance of [WeightRegularizer](../regularizers.md),\n applied to the bias.\n dropout_W: float between 0 and 1. Fraction of the input units to drop for input gates.\n dropout_U: float between 0 and 1. Fraction of the input units to drop for recurrent connections.\n\n # References\n - [A Theoretically Grounded Application of Dropout in Recurrent Neural Networks](http://arxiv.org/abs/1512.05287)\n \"\"\"\n\n def __init__(self, n_out, n_in=None, init=GlorotUniform(), inner_init=Orthogonal(), activation=Tanh(), return_sequence=False):\n self.n_out = n_out\n self.n_in = n_in\n self.init = init\n self.inner_init = inner_init\n self.activation_cls = activation.__class__\n self.activations = []\n self.return_sequence = return_sequence\n\n self.W, self.dW = None, None\n self.U, self.dU = None, None\n self.b, self.db = None, None\n self.last_outputs = None\n self.last_input = None\n self.out_shape = None\n\n def connect_to(self, prev_layer=None):\n if prev_layer is not None:\n assert len(prev_layer.out_shape) == 3\n n_in = prev_layer.out_shape[-1]\n else:\n assert self.n_in is not None\n n_in = self.n_in\n\n self.W = self.init((n_in, self.n_out))\n self.U = self.inner_init((self.n_out, self.n_out))\n self.b = Zero()((self.n_out,))\n\n if self.return_sequence:\n self.out_shape = (None, None, self.n_out)\n else:\n self.out_shape = (None, self.n_out)\n\n def forward(self, input, *args, **kwargs):\n assert np.ndim(input) == 3, 'Only support batch training.'\n\n self.last_input = input\n nb_batch, nb_timestep, nb_in = input.shape\n outputs = Zero()((nb_batch, nb_timestep, self.n_out))\n\n if len(self.activations) == 0:\n self.activations = [self.activation_cls() for _ in range(nb_timestep)]\n\n outputs[:, 0, :] = self.activations[0].forward(np.dot(input[:, 0, :], self.W) + self.b)\n\n for i in range(1, nb_timestep):\n outputs[:, i, :] = self.activations[i].forward(\n np.dot(input[:, i, :], self.W) +\n np.dot(outputs[:, i - 1, :], self.U) + self.b)\n\n self.last_outputs = outputs\n if self.return_sequence:\n return self.last_outputs\n else:\n return self.last_outputs[:, -1, :]\n\n def backward(self, pre_grad, *args, **kwargs):\n zero = Zero()\n self.dW = zero(self.W.shape)\n self.dU = zero(self.U.shape)\n self.db = zero(self.b.shape)\n\n # hiddens.shape == (nb_timesteps, nb_batch, nb_out)\n hiddens = np.transpose(self.last_outputs, (1, 0, 2))\n if self.return_sequence:\n # check shape #\n # self.outputs.shape == (nb_batch, nb_timesteps, nb_out)\n assert hiddens.shape == pre_grad.shape\n nb_timesteps = pre_grad.shape[0]\n if not self.first_layer:\n layer_grad = Zero()(pre_grad.shape)\n\n for timestep1 in np.arange(nb_timesteps)[::-1]:\n delta = pre_grad[timestep1] * self.activations[timestep1].derivative()\n for timestep2 in np.arange(timestep1)[::-1]:\n self.dU += np.dot(hiddens[timestep2].T, delta)\n self.dW += np.dot(self.last_input[:, timestep2 + 1, :].T, delta)\n self.db += np.mean(delta, axis=0)\n if not self.first_layer:\n layer_grad[timestep2 + 1] += np.dot(delta, self.W.T)\n delta = np.dot(delta, self.U.T)\n\n if timestep1 == 0 or timestep2 == 0:\n self.dW += np.dot(self.last_input[:, 0, :].T, delta)\n self.db += np.mean(delta, axis=0)\n if not self.first_layer:\n layer_grad[0] += np.dot(delta, self.W.T)\n\n else:\n nb_timesteps = self.last_outputs.shape[1]\n nb_batchs = self.last_outputs.shape[0]\n assert (nb_batchs, self.last_outputs.shape[2]) == pre_grad.shape\n if not self.first_layer:\n layer_grad = Zero()(hiddens.shape)\n\n delta = pre_grad * self.activations[nb_timesteps - 1].derivative()\n for timestep2 in np.arange(nb_timesteps - 1)[::-1]:\n self.dU += np.dot(hiddens[timestep2].T, delta)\n self.dW += np.dot(self.last_input[:, timestep2 + 1, :].T, delta)\n self.db += np.mean(delta, axis=0)\n if not self.first_layer:\n layer_grad[timestep2 + 1] += np.dot(delta, self.W.T)\n delta = np.dot(delta, self.U.T)\n\n if timestep2 == 0:\n self.dW += np.dot(self.last_input[:, timestep2 + 1, :].T, delta)\n self.db += np.mean(delta, axis=0)\n if not self.first_layer:\n layer_grad[0] += np.dot(delta, self.W.T)\n\n if not self.first_layer:\n return layer_grad\n\n @property\n def params(self):\n return self.W, self.U, self.b\n\n @property\n def grads(self):\n return self.dW, self.dU, self.db\n\n\nclass GRU(Layer):\n def __init__(self):\n pass\n\n\nclass LSTM(Layer):\n def __init__(self):\n pass\n","sub_path":"npdl/layers/reccurent.py","file_name":"reccurent.py","file_ext":"py","file_size_in_byte":6543,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"172926365","text":"#!/usr/bin/env python\n\n# Required packages\nimport argparse\nfrom configparser import ConfigParser\n\nimport requests\nfrom astropy.time import Time\nfrom astropy.io import fits\nimport numpy as np\n\n\ndef DMS_to_detector(data, detector):\n \"\"\"Transformations from Robert Jedzrejewski\n https://github.com/STScI-JWST/jwst/blob/master/jwst/refpix/reference_pixels.py#L690\n \"\"\"\n if detector == 'NRS1':\n # NRS1 is just flipped over the line X=Y\n data = np.swapaxes(data, 2, 3)\n\n if detector == 'NRS2':\n # NRS2 is flipped over the line Y=X, then rotated 180 degrees\n data = np.swapaxes(data, 2, 3)[:, :, ::-1, ::-1]\n\n if detector == 'NRCA1':\n # NRCA1 is just flipped in X\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCA2':\n # NRCA2 is just flipped in Y\n data = data[:, :, ::-1]\n\n if detector == 'NRCA3':\n # NRCA3 is just flipped in X\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCA4':\n # NRCA4 is just flipped in Y\n data = data[:, :, ::-1]\n\n if detector == 'NRCALONG':\n # NRCA3 is just flipped in X\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCB1':\n # NRCB1 is just flipped in Y\n data = data[:, :, ::-1]\n\n if detector == 'NRCB2':\n # NRCB2 is just flipped in X\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCB3':\n # NRCB1 is just flipped in Y\n data = data[:, :, ::-1]\n\n if detector == 'NRCB4':\n # NRCB4 is just flipped in X\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCBLONG':\n # NRCB1 is just flipped in Y\n data = data[:, :, ::-1]\n\n if detector == 'NIS':\n # NIRISS has a 180 degree rotation followed by a flip across the line\n # X=Y\n data = np.swapaxes(data[:, :, ::-1, ::-1], 2, 3)\n\n if detector == 'GUIDER1':\n # GUIDER1 is flipped in X and Y\n data = data[:, :, ::-1, ::-1]\n\n if detector == 'GUIDER2':\n # GUIDER2 is just flipped in X\n data = data[:, :, :, ::-1]\n\n # MIRI data doesn't need transforming\n\n return data\n\ndef detector_to_DMS(data, detector):\n if detector == 'NRS1':\n # Just flip back\n data = np.swapaxes(data, 2, 3)\n\n if detector == 'NRS2':\n # The inverse is to rotate 180 degrees, then flip over the line Y=X\n data = np.swapaxes(data[:, :, ::-1, ::-1], 2, 3)\n\n if detector == 'NRCA1':\n # Just flip back\n data = data[:, :, ::-1, ::-1]\n\n if detector == 'NRCA2':\n # Just flip back\n data = data[:, :, ::-1]\n\n if detector == 'NRCA3':\n # Just flip back\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCA4':\n # Just flip back\n data = data[:, :, ::-1]\n\n if detector == 'NRCALONG':\n # Just flip back\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCB1':\n # Just flip back\n data = data[:, :, ::-1]\n\n if detector == 'NRCB2':\n # Just flip back\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCB3':\n # Just flip back\n data = data[:, :, ::-1]\n\n if detector == 'NRCB4':\n # Just flip back\n data = data[:, :, :, ::-1]\n\n if detector == 'NRCBLONG':\n # Just flip back\n data = data[:, :, ::-1]\n\n if detector == 'NIS':\n # Just flip and rotate back\n data = np.swapaxes(data, 2, 3)[:, :, ::-1, ::-1]\n \n if detector == 'GUIDER1':\n # Just flip back\n data = data[:, :, ::-1, ::-1]\n\n if detector == 'GUIDER2':\n # Just flip back\n data = data[:, :, :, ::-1]\n\n # MIRI data doesn't need transforming\n\n return data\n\ndef main(args):\n\n config = ConfigParser()\n config.read(args.config_file)\n\n old_hdulist = fits.open(args.input_file)\n\n new_hdulist = fits.HDUList()\n new_hdulist.append(fits.PrimaryHDU())\n new_hdulist[0].header = old_hdulist[0].header\n\n # get the exposure start and end times\n start_time = Time(old_hdulist[0].header['EXPSTART'], format='mjd').isot\n end_time = Time(old_hdulist[0].header['EXPEND'], format='mjd').isot\n\n params = {'sTime' : start_time, 'eTime' : end_time}\n\n s = requests.Session()\n\n # jwdmsdevwsvm1 is for testing. The actual DB host will be different.\n url_base = 'http://jwdmsdevwsvm1.stsci.edu/JWDMSEngSpAcc_CV2CV3/TlmMnemonicDataSrv.svc/Data/'\n\n for keyword, mnemonic in config['mnemonics'].items():\n\n # get request to server.\n url = url_base + mnemonic\n\n r = s.get(url, params=params, verify=False)\n\n # Parse json\n parsed_json = r.json()\n\n # json ObsTime has format '/Date(1358619814230+0000)/' which is 1358619814.230 in UNIX time\n # isotime = Time(float(parsed_json['Data'][0]['ObsTime'][6:-7])/1000., format='unix').isot\n\n # just take the first value of the series\n new_hdulist[0].header[keyword] = (parsed_json['Data'][0]['EUValue'], mnemonic.upper())\n\n # add the Engineering Mnemonics section heading\n new_hdulist[0].header.set('', 'Engineering Mnemonics', before=config['mnemonics'].keys()[0])\n new_hdulist[0].header.set('', '', before=config['mnemonics'].keys()[0])\n new_hdulist[0].header.set('', '', before=config['mnemonics'].keys()[0])\n\n # transform from DMS to detector orientation\n pixel_data = DMS_to_detector(old_hdulist['SCI'].data, old_hdulist['PRIMARY'].header['DETECTOR'])\n\n # collapse from 4D to 3D\n nints, ngroups, nx, ny = pixel_data.shape\n\n # add reference output for MIRI\n if old_hdulist['PRIMARY'].header['INSTRUME'] == 'MIRI':\n new_hdulist[0].data = np.append(old_hdulist['SCI'].data.reshape((nints*ngroups, nx, ny)), \n old_hdulist['REFOUT'].data.reshape((nints*ngroups, 256, 1032)), axis=1)\n\n else:\n new_hdulist[0].data = pixel_data.reshape((nints*ngroups, nx, ny))\n\n new_hdulist[0].header.set('', '', before='BSCALE')\n\n # remove the NEXTEND keyword since there is only one extension now\n new_hdulist[0].header.remove('NEXTEND')\n\n # Write out\n new_hdulist.writeto(args.output_file, clobber=True)\n\nif __name__ == '__main__':\n # Command line argument handler.\n parser = argparse.ArgumentParser(\n description='Convert JWST data from DMS format to FITSWriter format',\n epilog='example: flight2ground tlm.cfg input.fits output.fits')\n parser.add_argument('config_file', help='config file with Telemetry FITS keyword/mnemonic pairs')\n parser.add_argument('input_file', help='level 1B data file to reformat')\n parser.add_argument('output_file', help='name of output file')\n args = parser.parse_args()\n main(args)","sub_path":"flight2ground/flight2ground.py","file_name":"flight2ground.py","file_ext":"py","file_size_in_byte":6602,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"401790621","text":"import FWCore.ParameterSet.Config as cms\n\nSKFlatMaker = cms.EDAnalyzer(\"SKFlatMaker\",\n\n DataYear = cms.untracked.int32(-1),\n processName = cms.untracked.string(\"HLT\"),\n DebugLevel = cms.untracked.int32(0),\n\n # -- Object Tags -- #\n Muon = cms.untracked.InputTag(\"slimmedMuons\"),\n Electron = cms.untracked.InputTag(\"slimmedElectrons\"),\n Photon = cms.untracked.InputTag(\"slimmedPhotons\"),\n Jet = cms.untracked.InputTag(\"slimmedJets\"),\n GenJet = cms.untracked.InputTag(\"slimmedGenJets\"),\n FatJet = cms.untracked.InputTag(\"slimmedJetsAK8\"),\n GenFatJet = cms.untracked.InputTag(\"slimmedGenJetsAK8\"),\n MET = cms.InputTag(\"slimmedMETs\"),\n LHEEventProduct = cms.untracked.InputTag(\"externalLHEProducer\"),\n LHERunInfoProduct = cms.untracked.InputTag(\"externalLHEProducer\"),\n GenParticle = cms.untracked.InputTag(\"genParticles\"),\n\n #### MiniIso\n pfCandsForMiniIso = cms.untracked.InputTag(\"packedPFCandidates\"),\n ## Muon\n ## https://github.com/cms-sw/cmssw/blob/f493624b3018543865bbf04bb8a48c5dae44bc82/RecoMuon/MuonIsolation/python/muonPFIsolationValues_cff.py\n miniIsoParams = cms.vdouble(0.05, 0.2, 10.0, 0.5, 0.0001, 0.01, 0.01, 0.01, 0.0),\n ## Electron\n ## https://github.com/cms-sw/cmssw/blob/09c3fce6626f70fd04223e7dacebf0b485f73f54/RecoParticleFlow/PFProducer/python/electronPFIsolationValues_cff.py\n miniIsoParamsE = cms.vdouble(0.05, 0.2, 10.0, 0.0, 0.015, 0.015, 0.08, 0.0, 0.0),\n miniIsoParamsB = cms.vdouble(0.05, 0.2, 10.0, 0.0, 0.000, 0.000, 0.00, 0.0, 0.0),\n\n # -- electron information -- #\n rho = cms.untracked.InputTag(\"fixedGridRhoFastjetAll\"),\n conversionsInputTag = cms.untracked.InputTag(\"allConversions\"),\n GsfTrack = cms.untracked.InputTag(\"electronGsfTracks\"),\n electron_EA_NHandPh_file = cms.untracked.FileInPath(\"RecoEgamma/ElectronIdentification/data/Fall17/effAreaElectrons_cone03_pfNeuHadronsAndPhotons_94X.txt\"),\n electron_IDtoSave = cms.untracked.vstring(\n\"cutBasedElectronID-Fall17-94X-V2-veto\",\n\"cutBasedElectronID-Fall17-94X-V2-loose\",\n\"cutBasedElectronID-Fall17-94X-V2-medium\",\n\"cutBasedElectronID-Fall17-94X-V2-tight\",\n'mvaEleID-Fall17-iso-V2-wp80' ,\n'mvaEleID-Fall17-iso-V2-wp90' ,\n'mvaEleID-Fall17-iso-V2-wpHZZ' ,\n'mvaEleID-Fall17-iso-V2-wpLoose',\n'mvaEleID-Fall17-noIso-V2-wp80' ,\n'mvaEleID-Fall17-noIso-V2-wp90' ,\n'mvaEleID-Fall17-noIso-V2-wpLoose',\n\"heepElectronID-HEEPV70\",\n ),\n\n #### Rochestor\n roccorPath = cms.string('SKFlatMaker/SKFlatMaker/data/roccor.Run2.v3/RoccoR2016.txt'),\n\n # -- photon information -- #\n photon_EA_CH_file = cms.untracked.FileInPath(\"RecoEgamma/PhotonIdentification/data/Fall17/effAreaPhotons_cone03_pfChargedHadrons_90percentBased_TrueVtx.txt\"),\n photon_EA_HN_file = cms.untracked.FileInPath(\"RecoEgamma/PhotonIdentification/data/Fall17/effAreaPhotons_cone03_pfNeutralHadrons_90percentBased_TrueVtx.txt\"),\n photon_EA_Ph_file = cms.untracked.FileInPath(\"RecoEgamma/PhotonIdentification/data/Fall17/effAreaPhotons_cone03_pfPhotons_90percentBased_TrueVtx.txt\"),\n\n # -- Jet information -- #\n AK4Jet_payloadName = cms.string('AK4PFchs'),\n AK8Jet_payloadName = cms.string('AK8PFPuppi'),\n AK4Jet_JER_PtRes_filepath = cms.string('SKFlatMaker/SKFlatMaker/data/JRDatabase/textFiles/Summer16_25nsV1_MC/Summer16_25nsV1_MC_PtResolution_AK4PFchs.txt'),\n AK4Jet_JER_SF_filepath = cms.string('SKFlatMaker/SKFlatMaker/data/JRDatabase/textFiles/Summer16_25nsV1_MC/Summer16_25nsV1_MC_SF_AK4PFchs.txt'),\n AK8Jet_JER_PtRes_filepath = cms.string('SKFlatMaker/SKFlatMaker/data/JRDatabase/textFiles/Summer16_25nsV1_MC/Summer16_25nsV1_MC_PtResolution_AK8PFPuppi.txt'),\n AK8Jet_JER_SF_filepath = cms.string('SKFlatMaker/SKFlatMaker/data/JRDatabase/textFiles/Summer16_25nsV1_MC/Summer16_25nsV1_MC_SF_AK8PFPuppi.txt'),\n\n # -- MET information -- #\n METFilterResults_PAT = cms.InputTag(\"TriggerResults\", \"\", \"PAT\"),\n METFilterResults_RECO = cms.InputTag(\"TriggerResults\", \"\", \"RECO\"),\n pfMET = cms.untracked.InputTag(\"pfMet\"),\n \n # -- Trigger -- #\n TriggerResults = cms.untracked.InputTag(\"TriggerResults\", \"\", \"HLT\"),\n TriggerResultsPAT = cms.untracked.InputTag(\"TriggerResults\", \"\", \"PAT\"),\n ##TriggerObject = cms.untracked.InputTag(\"selectedPatTrigger\"),\n TriggerObject = cms.untracked.InputTag(\"slimmedPatTrigger\"), \n \n # -- Else -- #\n GenEventInfo = cms.untracked.InputTag(\"generator\"),\n BeamSpot = cms.untracked.InputTag(\"offlineBeamSpot\"),\n PrimaryVertex = cms.untracked.InputTag(\"offlinePrimaryVerticesWithBS\"),\n Track = cms.untracked.InputTag(\"generalTracks\"),\n PileUpInfo = cms.untracked.InputTag(\"addPileupInfo\"),\n\n # -- Store Flags -- #\n StoreMuonFlag = cms.untracked.bool(True),\n StoreElectronFlag = cms.untracked.bool(True),\n StoreCalibElectronFlag = cms.untracked.bool(True),\n StorePhotonFlag = cms.untracked.bool(True),\n StoreJetFlag = cms.untracked.bool(True),\n StoreFatJetFlag = cms.untracked.bool(True),\n StoreMETFlag = cms.untracked.bool(True),\n StoreLHEFlag = cms.untracked.bool(True),\n StoreGENFlag = cms.untracked.bool(True),\n KeepAllGen = cms.untracked.bool(True), \n StorePriVtxFlag = cms.untracked.bool(True),\n StoreHLTReportFlag = cms.untracked.bool(True),\n StoreHLTObjectFlag = cms.untracked.bool(False),\n StoreL1PrefireFlag = cms.untracked.bool(False),\n\n # -- Filters -- #\n ApplyFilter = cms.untracked.bool(False),\n FilterType = cms.untracked.int32(0),\n\n #### PDF ID's to be save\n ScaleIDRange = cms.untracked.vint32(-999,-999),\n PDFErrorIDRange = cms.untracked.vint32(-999,-999),\n PDFAlphaSIDRange = cms.untracked.vint32(-999,-999),\n PDFAlphaSScaleValue = cms.untracked.vdouble(-999.,-999.),\n\n)\n","sub_path":"SKFlatMaker/python/SKFlatMaker_cfi.py","file_name":"SKFlatMaker_cfi.py","file_ext":"py","file_size_in_byte":5547,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"332994485","text":"from setuptools import setup, find_packages\nimport sys, os\n\nfrom pyhdhomerun.hdhr import get_hdhr\n\ntry:\n get_hdhr()\nexcept OSError as e:\n print(\"Could not load HDHomeRun library: %s\" % (e))\n sys.exit(1)\nelse:\n print(\"HDHomeRun libraries verified.\")\n \nversion = '2.3.4'\n\nsetup(name='pyhdhomerun',\n version=version,\n description=\"HDHomeRun interface library.\",\n long_description=\"\"\"\\\nHDHomeRun interface library. Supports device discovery, channel-scanning, streaming, status inquiries, channel changes, etc..\"\"\",\n classifiers=['Development Status :: 4 - Beta',\n 'License :: OSI Approved :: BSD License',\n 'Natural Language :: English',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n 'Topic :: Multimedia :: Video :: Capture'\n ],\n keywords='tv television tuner tvtuner hdhomerun',\n author='Dustin Oprea',\n author_email='myselfasunder@gmail.com',\n url='https://github.com/dsoprea/PyHdHomeRun',\n license='New BSD',\n packages=['pyhdhomerun'],\n include_package_data=True,\n zip_safe=True,\n install_requires=[\n 'setuptools',\n ],\n entry_points=\"\"\"\n # -*- Entry points: -*-\n \"\"\",\n )\n","sub_path":"pypi_install_script/pyhdhomerun-2.3.4.tar/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"250499778","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed May 13 19:33:07 2020\n\n@author: macbookariel\n\"\"\"\n# Regresión lineal simple: el primer paso será copiar el documento de data_preprocessing.\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pandas as pd\n \n# Importar el Data set\ndataset = pd.read_csv(\"Salary_Data.csv\") \n\n#En este caso, al ver la variable, tenemos que darle a format %,1 para ver 1 posicición decimal.\n# Recordemos que la columna llamada \"Index en el variable explorer NO ES UNA COLUMNA Y QUE LAS COLUMNAS EMPIEZAN\n# A NOMBRARSE EN EL CERO.\n\nX = dataset.iloc[:, :-1].values #Variable independiente = años de experiencia. Ubicada en la anteúltima posición (-1)\ny = dataset.iloc[:, 1].values # Variable dependiente = a predecir = salario. Ubicada en la columna 1\n\n\n# Dividir el data set en conjunto de entrenamiento y en conjunto de testing.\n# En este caso vamos a tomar 10 para testing (1/3) y el resto para entrenamiento (20).\n\nfrom sklearn.model_selection import train_test_split\nX_train, X_test, y_train, y_test = train_test_split(X, y,test_size = 1/3, random_state = 0)\n\n# Escalado de variables. En el caso de la regresión lineal, el modelo no requiere escalado.\n\n# from sklearn.preprocessing import StandardScaler\n# sc_X = StandardScaler()\n# X_train = sc_X.fit_transform(X_train)\n# X_test = sc_X.transform(X_test)\n\n# Crear modelo de regresión lineal simple con el conjunto de entrenamiento.\n\nfrom sklearn.linear_model import LinearRegression\n\nregression = LinearRegression()\nregression.fit(X_train, y_train) #La máquina ha aprendido, entonces, con las variables que le suministramos (igual tamaño en ambas variables!!)\n\n\n# Predecir el conjunto de test. Para ello crearemos un vector de datos con los datos de predicción para obtener la variable \n# dependiente que nos devuelve el modelo. Observemos que la variable a ser suministrada solo es la independiente (X_test) y\n# el modelo hace la predicción y la guarda en y_pred. Es decir, usando la X_test (años de experiencia) quiero que prediga el sueldo\n# y lo guarde en y_pred\n\ny_pred = regression.predict(X_test)\n\n#Visualizar los resultados de entrenamiento. Vamos a generar un scatter plot (nube de dispersión). Vamos a usar pyplot.\n\nplt.scatter(X_train, y_train, color = \"red\")\n\n# Vamos a hacer un scatter plot donde la X es el grupo de entrenamiento y la y es la predición pero sobre X_train,\n# así vemos las dos variables\n\nplt.plot(X_train, regression.predict(X_train), color = \"blue\")\nplt.title(\"Sueldo vs Años de experiencia (Conjunto de entrenamiento\")\nplt.xlabel(\"Años de experiencia\")\nplt.ylabel(\"Sueldo (en $)\")\nplt.show()\n\n# Vamos a hacer un scatter plot para ver cómo quedan los datos de test y cómo se ajusta a ellos la recta de regresión\n\nplt.scatter(X_test, y_test, color = \"red\")\n\nplt.plot(X_train, regression.predict(X_train), color = \"blue\")\nplt.title(\"Sueldo vs Años de experiencia (Conjunto de testing)\")\nplt.xlabel(\"Años de experiencia\")\nplt.ylabel(\"Sueldo (en $)\")\nplt.show()\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"datasets/Part 2 - Regression/Section 4 - Simple Linear Regression/Ariel_Regresion_Lineal_Simple.py","file_name":"Ariel_Regresion_Lineal_Simple.py","file_ext":"py","file_size_in_byte":3036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"112097525","text":"import os\nimport signal\nimport sys\n\nfrom httptask import workers\n\nclass Error(Exception): pass\n\nclass Watcher(object):\n \"\"\"See http://code.activestate.com/recipes/496735/\"\"\"\n\n def __init__(self):\n self.child = os.fork()\n if self.child == 0:\n return\n else:\n self.watch()\n\n def watch(self):\n try:\n os.wait()\n except KeyboardInterrupt:\n self.kill()\n sys.exit()\n\n def kill(self):\n try:\n os.kill(self.child, signal.SIGKILL)\n except OSError:\n pass\n\nclass Spawn(object):\n\n def __init__(self, call, url, options):\n self.call = call\n self.url = url\n self.options = options\n\n def __call__(self):\n return self.call(self.url, self.options)\n\nclass Service(object):\n\n def __init__(self, config):\n self.worker_list = []\n self.config = config\n Watcher()\n\n def get(self, id, config, name, required=False, default=None, ctype=str):\n present = name in config\n if required and not present:\n raise Error('%s is required in %s.' % (name, id))\n value = config.get(name, default)\n if present:\n del config[name]\n try:\n return ctype(value)\n except ValueError:\n raise Error('%s was type \"%s\" and should be \"%s\" in %s.' % (name, type(name), ctype, id))\n\n def run(self):\n for id, value in self.config.items():\n url = self.get(id, value, 'url', required=True)\n count = self.get(id, value, 'count', default=1, ctype=int)\n worker_type = self.get(id, value, 'type', required=True)\n spawn = None\n if worker_type == 'beanstalk':\n spawn = Spawn(workers.Beanstalk, url, value)\n if not spawn is None:\n for i in range(count):\n worker = spawn()\n worker.start()\n self.worker_list.append(worker)\n else:\n raise Error('Unknown worker type (%s)' % type)\n","sub_path":"python-httptask/httptask/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"604749482","text":"from rasa.nlu.training_data import load_data\nfrom rasa.nlu.config import RasaNLUModelConfig\nfrom rasa.nlu.model import Trainer\nfrom rasa.nlu import config\nfrom search_weather import search_weather\nimport string\nimport re\nimport random\n\nclass chatBot:\n INIT = 0\n AUTHED = 1\n CITY_CHOOSEN = 2\n LIVES_FUNCTION_CHOOSE = 3\n GOODBYE = 4\n\n interpreter = Trainer(config.load(\"resources/config_spacy.yml\")).train(load_data('resources/weather_intent.json'))\n current_information = \"\"\n weather_query = search_weather()\n cur_state = INIT\n cur_pending = None\n res_message = \"\"\n pending_intent = \"\"\n\n lives_function = [\"weather\", \"temperature\", \"windpower\", \"humidity\", \"winddirection\"]\n\n forecasts_function = [\"dayweather\", \"nightweather\", \"daytemp\", \"nighttemp\", \"daywind\", \"nightwind\", \"daypower\", \"nightpower\"]\n\n forecasts_day = [\"zr_forecasts\", \"oe_forecasts\", \"tw_forecasts\"]\n\n function_mapping = {\n \"lives_info\": \"实时信息:\",\n \"lives_weather\": \"实时天气:\",\n \"lives_temperature\": \"实时气温:\",\n \"lives_winddirection\": \"实时风向:\",\n \"lives_windpower\": \"实时风力:\",\n \"lives_humidity\": \"实时湿度:\",\n \"lives_reporttime\": \"报告时间:\",\n \"zr_forecasts_dayweather\": \"明天日间天气:\",\n \"zr_forecasts_nightweather\": \"明天夜间天气:\",\n \"zr_forecasts_daytemp\": \"明天日间气温:\",\n \"zr_forecasts_nighttemp\": \"明天夜间气温:\",\n \"zr_forecasts_daywind\": \"明天日间风向:\",\n \"zr_forecasts_nightwind\": \"明天夜间风向:\",\n \"zr_forecasts_daypower\": \"明天日间风力:\",\n \"zr_forecasts_nightpower\": \"明天夜间风力:\",\n \"oe_forecasts_dayweather\": \"后天日间天气:\",\n \"oe_forecasts_nightweather\": \"后天夜间天气:\",\n \"oe_forecasts_daytemp\": \"后天日间气温:\",\n \"oe_forecasts_nighttemp\": \"后天夜间气温:\",\n \"oe_forecasts_daywind\": \"后天日间风向:\",\n \"oe_forecasts_nightwind\": \"后天夜间风向:\",\n \"oe_forecasts_daypower\": \"后天日间风力:\",\n \"oe_forecasts_nightpower\": \"后天夜间风力:\",\n \"tw_forecasts_dayweather\": \"大后天日间天气:\",\n \"tw_forecasts_nightweather\": \"大后天夜间天气:\",\n \"tw_forecasts_daytemp\": \"大后天日间气温:\",\n \"tw_forecasts_nighttemp\": \"大后天夜间气温:\",\n \"tw_forecasts_daywind\": \"大后天日间风向:\",\n \"tw_forecasts_nightwind\": \"大后天夜间风向:\",\n \"tw_forecasts_daypower\": \"大后天日间风力:\",\n \"tw_forecasts_nightpower\": \"大后天夜间风力:\",\n }\n\n policy_rules = {\n (INIT, \"greet\"): (INIT, \"你好呀。\", None),\n (INIT, \"pending_forecasts_day\"): (INIT, \"请先登录再查询预报\", None),\n (INIT, \"pending_forecasts_function\"): (INIT, \"请先登录再查询预报\", None),\n (INIT, \"enquire_forecasts\"): (INIT, \"请先登录再查询预报\", None),\n (INIT, \"pending_function\"): (INIT, \"请先登录再查询实时信息:\", None),\n (INIT, \"city_error\"): (INIT, \"请先告诉我号码。\", None),\n (INIT, \"city_error\"): (INIT, \"城市自动设置失败,请手动设置城市:\", None),\n (INIT, \"weather_info\"): (INIT, \"请先告诉我你的号码。\", None),\n (INIT, \"number\"): (CITY_CHOOSEN, \"欢迎。当前您所在的城市已经自动设置:\", None),\n (INIT, \"city_choose\"): (INIT, \"城市已经设置,但是请先输入你的手机号码。\", None),\n (INIT, \"city_switch\"): (INIT, \"你连号码都没有设置过。\", None),\n (INIT, \"none\"): (INIT, \"你想要让我做什么?\", None),\n (AUTHED, \"pending_forecasts_day\"): (INIT, \"请先设置城市再查询预报\", None),\n (AUTHED, \"pending_forecasts_function\"): (INIT, \"请先设置城市再查询预报\", None),\n (AUTHED, \"enquire_forecasts\"): (INIT, \"请先设置城市再查询预报\", None),\n (AUTHED, \"pending_function\"): (INIT, \"请先设置城市再查询实时信息:\", None),\n (AUTHED, \"city_error\"): (AUTHED, \"城市自动设置失败,请手动设置城市:\", None),\n (AUTHED, \"city_choose\"): (CITY_CHOOSEN, \"当前城市已经切换为:\", None),\n (AUTHED, \"weather_info\"): (AUTHED, \"你还没设置过城市呢。\", None),\n (AUTHED, \"none\"): (AUTHED, \"我不清楚我应该做什么。\", None),\n (AUTHED, \"greet\"): (AUTHED, \"你已经打过招呼了鸭。\", None),\n (AUTHED, \"city_switch\"): (AUTHED, \"好的,请选择一个城市。\", None),\n (CITY_CHOOSEN, \"pending_function\"): (CITY_CHOOSEN, \"您想查询现在的什么信息:\", None),\n (CITY_CHOOSEN, \"weather_info\"): (CITY_CHOOSEN, \"帮您查询到的信息:\", None),\n (CITY_CHOOSEN, \"none\"): (CITY_CHOOSEN, \"不是很清楚\", None),\n (CITY_CHOOSEN, \"greet\"): (CITY_CHOOSEN, \"我们已经聊了会了鸭。\", None),\n (CITY_CHOOSEN, \"city_choose\"): (CITY_CHOOSEN, \"您已经设置过城市了。可以告诉我'切换'来切换城市。\", None),\n (CITY_CHOOSEN, \"goodbye\"): (INIT, \"很高兴为您服务\", None),\n (CITY_CHOOSEN, \"city_switch\"): (AUTHED, \"好的,请选择新的城市\", None),\n (CITY_CHOOSEN, \"pending_forecasts_day\"): (CITY_CHOOSEN, \"请选择你要查询的日期\", None),\n (CITY_CHOOSEN, \"pending_forecasts_function\"): (CITY_CHOOSEN, \"请选择你想查询的功能\", None),\n (CITY_CHOOSEN, \"enquire_forecasts\"): (CITY_CHOOSEN, \"准备给您进行天气预报,请问想要查询什么时候的\", None),\n (INIT, \"affirm\"): (INIT, \"你在开玩笑吧\", None),\n (AUTHED, \"affirm\"): (AUTHED, \"我也没为你查什么呀\", None),\n (CITY_CHOOSEN, \"affirm\"): (CITY_CHOOSEN, \"我很高兴得到您的夸奖,接下去想要查询什么\", None),\n (INIT, \"goodbye\"): (INIT, \"还没登录,舍不得离开\", None),\n (AUTHED, \"goodbye\"): (INIT, \"还没查询新的城市,舍不得离开\", None)\n\n }\n query_function = [\"lives_info\",\n \"lives_weather\",\n \"lives_temperature\",\n \"lives_winddirection\",\n \"lives_windpower\",\n \"lives_humidity\",\n \"lives_reporttime\",\n \"forecasts_dayweather\",\n \"forecasts_nightweather\",\n \"forecasts_daytemp\",\n \"forecasts_nighttemp\",\n \"forecasts_daywind\",\n \"forecasts_nightwind\",\n \"forecasts_daypower\",\n \"forecasts_nightpower\"]\n\n def create_intent(self, intent):\n if self.pending_intent != \"\":\n if self.pending_intent == \"lives\" and intent in self.lives_function:\n return self.pending_intent + \"_\" + intent\n if self.pending_intent in self.forecasts_day and intent in self.forecasts_function:\n return self.pending_intent + \"_\" + intent\n if self.pending_intent in self.forecasts_function and intent in self.forecasts_day:\n return intent + \"_\" + self.pending_intent\n return None\n\n\n def send_message(self, state, pending, message):\n self.res_message = \"\"\n print(\"USER : {}\".format(message))\n new_state, response, pending_state = self.respond(state, message)\n # print(\"BOT : {}\".format(response))\n self.res_message = format(response)\n if self.current_information != \"\":\n self.res_message += self.weather_query.get_city_name() + \" \" + self.current_information\n # print(self.weather_query.get_city_name() + \" \" + self.current_information)\n self.current_information = \"\"\n return new_state, pending\n\n def wether_is_forecasts(self, intent):\n if intent[0:2] in ['zr', 'oe', 'tw']:\n return True\n return False\n\n def extrect_phone_number(self, message):\n pattern = re.compile(\"[0-9]{11}\")\n match = re.search(pattern, message)\n return match\n\n def extrect_city_code(self, message):\n pattern = re.compile(\"[0-9]{6}\")\n match = re.search(pattern, message)\n return match\n\n def respond(self, state, message):\n intent = self.interpret(state, message)\n if message in self.weather_query.get_cities().keys() and intent == \"city_error\":\n intent = \"city_choose\"\n if self.extrect_city_code(message) and intent == \"city_error\":\n intent = \"city_choose\"\n if intent == \"city_choose\":\n if self.extrect_city_code(message):\n self.weather_query.set_local(self.extrect_city_code(message).group(0))\n else:\n self.weather_query.set_local(message)\n if (self.wether_is_forecasts(intent)) or (intent in self.query_function):\n if state != self.INIT:\n self.current_information = self.function_mapping[intent] + self.weather_query.get_information(intent)\n intent = \"weather_info\"\n new_state, response, pending = self.policy_rules[(state, intent)]\n if (state == self.INIT and intent == \"number\") or (state == self.AUTHED and intent == \"city_choose\"):\n response += self.weather_query.get_city_name()\n return new_state, response, pending\n\n def interpret(self, state, message):\n if self.weather_query.get_city_name() == \"\":\n return 'city_error'\n data = self.interpreter.parse(message)\n # print(data)\n intent = data[\"intent\"][\"name\"]\n # print(intent + \" \" + self.pending_intent)\n if message in self.weather_query.get_cities().keys():\n self.weather_query.set_local(self.weather_query.get_cities()[message])\n return 'city_choose'\n if self.extrect_phone_number(message):\n return 'number'\n if self.extrect_city_code(message):\n return 'city_choose'\n if \"切换\" in message:\n return 'city_switch'\n if intent == \"lives\" and self.pending_intent not in self.lives_function:\n # print(\"asdasdadsadad\")\n self.pending_intent = intent\n return \"pending_function\"\n if intent in self.forecasts_day and self.pending_intent not in self.forecasts_function:\n self.pending_intent = intent\n return \"pending_forecasts_function\"\n if intent in self.forecasts_function and self.pending_intent not in self.forecasts_day:\n self.pending_intent = intent\n return \"pending_forecasts_day\"\n if intent in self.lives_function:\n rel_intent = self.create_intent(intent)\n if rel_intent is not None:\n self.pending_intent = \"\"\n return rel_intent\n if (intent in self.forecasts_day and self.pending_intent in self.forecasts_function) or (intent in self.forecasts_function and self.pending_intent in self.forecasts_day):\n rel_intent = self.create_intent(intent)\n if rel_intent is not None:\n self.pending_intent = \"\"\n print(rel_intent)\n return rel_intent\n if ((state, intent) not in self.policy_rules.keys()) and (intent not in self.function_mapping.keys()):\n return \"none\"\n return intent\n\n def chat(self, msg):\n self.cur_state, self.cur_pending = self.send_message(self.cur_state, self.cur_pending, msg)\n return self.res_message\n","sub_path":"chatBot.py","file_name":"chatBot.py","file_ext":"py","file_size_in_byte":11580,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"255976580","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport re\nfrom register_vars import RegisterVars\n\n\nINVENTORY_FILE = './hosts'\nVARS_FILE = './group_vars/servers.yml'\n\n\nclass RegisterHostsVars(RegisterVars):\n def __init__(self):\n super(RegisterHostsVars, self).__init__()\n 'This class has no propertys.'\n\n def _generate_hostlist(self, resultlist):\n resultlist[1] = map(lambda n:n.strip().split(' mngip='),\n filter(lambda n:'mngip' in n and not re.match('\\A#', n), resultlist[0]))\n return resultlist\n\n def _refine_each_param(self, host_ip_list, param_dict):\n if not host_ip_list:\n return param_dict\n host_ip = host_ip_list.pop()\n param_dict['hosts_params'].append({'name':host_ip[0], 'ipaddr':host_ip[1]})\n return self._refine_each_param(host_ip_list, param_dict)\n \n def _refine_param(self, host_ip_list):\n host_ip_list.reverse()\n param_dict = {'hosts_params': []}\n return self._refine_each_param(host_ip_list, param_dict)\n\nif __name__ == '__main__':\n RegisterHostsVars().main(INVENTORY_FILE, VARS_FILE)\n\n","sub_path":"register_hosts_vars.py","file_name":"register_hosts_vars.py","file_ext":"py","file_size_in_byte":1134,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"354102619","text":"import json\nimport copy\n\nfile_with_old_and_new_id = 'd:/_work/_program/_python/prName_hallId_prId_format.json'\nfile_types_acbtype_refactor = 'd:/_work/_program/_python/_AKOM_types_acbtype_refactor.json'\nfile_rezult = 'd:/_work/_program/_python/_AKOM_types_acbtype_refactor_with_true_id.json'\n\ndef rezdic(doc1_element, doc2):\n result_element = []\n doc1_element_copy = copy.deepcopy(doc1_element)\n del doc1_element_copy['_id']\n for charge_hall in doc1_element_copy['chargeHalls']:\n for index, program_id in enumerate(charge_hall['programId']):\n for element in doc2:\n if charge_hall['hallId'] == element['chargeHallId'] and element['programId'] == program_id:\n charge_hall['programId'][index] = element['new_prid']\n break\n\n result_element.append(doc1_element_copy)\n return result_element\n\n\nif __name__ == '__main__':\n result_all_elements = []\n with open(file_types_acbtype_refactor) as json_data:\n doc1 = json.load(json_data)\n with open(file_with_old_and_new_id) as json_data:\n doc2 = json.load(json_data)\n for doc1_element in doc1:\n result_all_elements.extend(rezdic(doc1_element, doc2))\n\n print(result_all_elements)\n file = open(file_rezult, 'w')\n file.write(json.dumps(result_all_elements))\n file.close()","sub_path":"work with mongodb/mongodb_find_and_create5.py","file_name":"mongodb_find_and_create5.py","file_ext":"py","file_size_in_byte":1347,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"127104145","text":"import re\n\nimport pandas as pd\n\nimport astropy.units as u\nfrom universal_loader.kb.invalid_symbols import INVALID_SYMBOL_LIST\nfrom universal_loader.unstructured.text2csv import plural_to_single\n\nu.imperial.enable()\n\n\"\"\"\nNotes\n------\n\n\nThis module manages metadata knowledge base for atomic and composite units. based on astropy\n\"\"\"\n\n\ndef get_first_item(collection):\n \"\"\"\n\n helper function to help handle uncertainties\n\n Parameters\n ----------\n collection of units\n\n Returns\n -------\n first item in collection\n\n \"\"\"\n if len(collection) == 0:\n return ''\n else:\n return collection[0]\n\n\ndef emit_units_kb():\n \"\"\"\n\n Instantiate unit db from astropy\n\n Returns\n -------\n list of tuples with 5 fields - name, physical type, alias, long name and unit name\n\n Examples\n --------\n >>> kb=emit_units_kb()\n\n \"\"\"\n unit_list = []\n for att in dir(u):\n unit = getattr(u, att)\n try:\n unit_list.append([unit.name,\n unit.physical_type,\n get_first_item(unit.aliases),\n get_first_item(unit.short_names),\n get_first_item(unit.long_names),\n \"u.\" + unit.name\n ])\n except:\n continue\n\n # imperial units are used in US and UK-commonwealth\n for att in dir(u.imperial):\n unit = getattr(u.imperial, att)\n try:\n unit_list.append([unit.name, unit.physical_type,\n get_first_item(unit.aliases),\n get_first_item(unit.short_names),\n get_first_item(unit.long_names),\n \"u.imperial.\" + unit.name\n ])\n except:\n continue\n\n return unit_list\n\n\ndef get_units_as_df():\n \"\"\"\n\n Returns units as pandas dataframe\n\n Returns\n -------\n\n Examples\n --------\n >>>> get_units_as_df()\n\n \"\"\"\n unit_list = emit_units_kb()\n df = pd.DataFrame(unit_list, columns=[\"name\", \"physical_type\", \"alias\", \"short_name\", \"long_name\", \"object_type\"])\n df.drop_duplicates(inplace=True)\n return df\n\n\ndef is_unit(test_str):\n \"\"\"\n\n Check if string is atomic unit.\n\n Parameters\n ----------\n test_str: string representing unit. can be long or short version of it eg meters or m.\n\n Returns\n -------\n True/False\n\n Examples\n -------\n >>> is_unit(\"second\")\n >>> is_unit(\"parsec\")\n\n See Also\n --------\n is_composite_unit\n\n \"\"\"\n units = get_units_as_df()\n if (test_str.upper() in list(map(lambda x: x.upper(), units['short_name'].values.tolist()))) \\\n or (test_str.upper() in list(map(lambda x: x.upper(), units['long_name'].values.tolist()))) \\\n or (test_str.upper() in list(map(lambda x: x.upper(), units['alias'].values.tolist()))):\n return True\n else:\n return False\n\n\ndef is_composite_unit(test_str):\n \"\"\"\n\n Check if string can be expressed as composite units\n\n Parameters\n ----------\n\n :param test_str:\n\n Examples\n --------\n >>> is_composite_unit('miles per hour')\n >>> is_composite_unit(\"kilometers per hour\")\n >>> is_composite_unit(\"feet per second per second\")\n >>> is_composite_unit(\"kilometers per second per second\")\n >>> is_composite_unit(\"DEG/DAY\")\n >>> is_composite_unit(\"meters-per-second\")\n >>> is_composite_unit(\"meters-per-second-per-second\")\n\n Returns\n --------\n :return: True/False\n \"\"\"\n delim_found = detect_delim_for_composite_type(test_str)\n if delim_found is not None:\n unit_components = preprocess_composite_components(delim_found, test_str)\n\n all_units = list(map(lambda c: is_unit(c), unit_components))\n return all(all_units)\n\n\ndef preprocess_composite_components(delim_found, test_str):\n unit_components = test_str.split(delim_found)\n # strip the spaces\n unit_components = list(map(lambda c: c.strip(), unit_components))\n # strip any invalid symbols\n pat = re.compile(INVALID_SYMBOL_LIST)\n unit_components = list(map(lambda c: re.sub(pat, '', c), unit_components))\n # make sure each component is singular form and not plural\n unit_components = list(map(lambda c: plural_to_single(c), unit_components))\n return unit_components\n\n\ndef detect_delim_for_composite_type(test_str):\n \"\"\"\n\n Detect separator in composite type string; only two separators are supported\n / and per\n\n Parameters\n ----------\n test_str\n\n Returns\n -------\n either detected separator or None\n\n Examples\n --------\n >>> detect_delim_for_composite_type(\"miles per hour\")\n >>> \"per\"\n\n \"\"\"\n delim_found = None\n composite_delim_list = [\"per\", \"/\"]\n for delim in composite_delim_list:\n if len(test_str.split(delim)) > 1:\n delim_found = delim\n break\n return delim_found\n\n\ndef get_computable_type(test_str):\n \"\"\"\n\n Search unit database and return equivalent computing type\n\n\n Parameters\n ----------\n test_str - atomic unit\n\n Returns\n -------\n computable object or None\n\n Examples\n -----------\n >>> get_computable_type(\"mile\")\n\n\n \"\"\"\n units = get_units_as_df()\n computable_type = eval(\n units.loc[\n (units[\"name\"] == test_str.lower())\n | (units[\"long_name\"] == test_str.lower())\n | (units[\"short_name\"] == test_str.lower())\n | (units[\"alias\"] == test_str.lower())\n ][\"object_type\"].values[0])\n return computable_type\n\n\ndef get_composite_computable_type(test_str):\n \"\"\"\n\n parse string and attempt to produce composite computable type\n\n Parameters\n ----------\n test_str\n\n Returns\n -------\n computable type or none\n\n Examples\n ---------\n >>> get_composite_computable_type(\"miles per hour\")\n\n \"\"\"\n\n delim = detect_delim_for_composite_type(test_str)\n comps = preprocess_composite_components(delim, test_str)\n comp_list_types = list(map(lambda comp: get_computable_type(comp), comps))\n composite_type = comp_list_types[0] / comp_list_types[1]\n return composite_type\n","sub_path":"universal_loader/kb/units_metadata.py","file_name":"units_metadata.py","file_ext":"py","file_size_in_byte":6240,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"237570192","text":"from flask import (Blueprint, render_template, redirect, url_for, request\n , flash, abort)\n\nfrom models import User, UserRegistrationForm\nfrom application.utilities.random import random_alphanumeric\nfrom application import bcrypt\n\nusers = Blueprint('users', __name__, template_folder='templates')\n\n@users.route('/confirm-email/')\ndef confirm_email(code):\n users = User.objects(email_confirmation_code=code)\n if not users:\n abort(404)\n user = users[0]\n user.update(set__email_confirmed=True)\n\n flash('Email confirmed, thanks! You may now login.', 'success')\n return redirect(url_for('users.login'))\n\n\n@users.route('/register', methods=['GET', 'POST'])\ndef register():\n user_form = UserRegistrationForm()\n \n if request.method == 'GET':\n return render_template('register.html', user_form=user_form)\n\n if request.method == 'POST':\n if user_form.validate():\n # create the new user\n new_user = User(\n name = user_form.name.data\n , email = user_form.email.data\n , self_reported_org = user_form.organization.data\n , hashed_password = bcrypt.generate_password_hash(\n user_form.password.data, 10)\n , email_confirmation_code = random_alphanumeric(20)\n )\n new_user.save()\n new_user.send_verification_email()\n flash('We have sent you a message to confirm your email address.'\n , 'info')\n return redirect(url_for('public.root'))\n\n # validation failed\n flash('Validation failed.', 'error')\n return redirect(url_for('users.register'))\n\n\n@users.route('/login')\ndef login():\n return render_template('login.html')\n\n@users.route('/logout')\ndef logout():\n # logout ..\n return redirect(url_for('public.root'))\n\n@users.route('/members')\ndef show_all():\n users = User.objects()\n return render_template('all_users.html', users=users)\n","sub_path":"application/users/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":2000,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"392225814","text":"from os import listdir\r\nfrom nltk.stem import WordNetLemmatizer\r\nimport re\r\nimport pickle\r\nimport math\r\nfrom multiprocessing import Pool\r\nimport time\r\n#import multiprocessing\r\n\r\n\r\ndef gethumantime(sec):\r\n if sec < 60: return str(round(sec,2))+\" seconds\"\r\n else: return str(int(sec/60))+\" minutes \"+str(round(sec%60,2))+\" seconds\"\r\n\r\n\r\ndef gaussianStrength(wordGap,sentenceGap):\r\n strength = (1/wordGap)*math.exp(-sentenceGap)\r\n return(strength)\r\n\r\n\"\"\"\r\ndef gaussianStrength(wordGap):\r\n strength = math.exp(-wordGap)\r\n return(strength)\r\n\"\"\"\r\n\r\ndef stopwords_preprocessing():\r\n fileOpen = open('stopwords/unprocessed_stopwords.txt',encoding='utf-8',errors=\"ignore\")\r\n lines = fileOpen.read()\r\n fileOpen.close()\r\n fileOpen = open('stopwords/stopwords.txt','w+',encoding='utf-8',errors=\"ignore\")\r\n eachLine = lines.split('\\n')\r\n stopwords = list(set(eachLine))\r\n for i in range(len(stopwords)): stopwords[i] = re.sub('[^A-Za-z]+', '', stopwords[i].lower())\r\n stopwords = list(set(stopwords))\r\n stopwords.sort()\r\n for i in range(len(stopwords)): \r\n if i == len(stopwords)-1: fileOpen.write(stopwords[i])\r\n else: fileOpen.write(stopwords[i]+'\\n')\r\n fileOpen.close()\r\n \r\ndef get_stopwords():\r\n fileOpen = open('stopwords/stopwords.txt',encoding='utf-8',errors=\"ignore\")\r\n lines = fileOpen.read()\r\n stopwords = lines.split('\\n')\r\n return(stopwords)\r\n\r\ndef processBookInformation(fileName,count):\r\n stopwords_preprocessing()\r\n stopwords = get_stopwords()\r\n lemmatizer = WordNetLemmatizer()\r\n folder_name = \"inputBooks\"\r\n cut_size = 7\r\n #list of all books containing sentenced token words\r\n book_data = []\r\n getfile = open(folder_name+'/'+fileName,encoding='utf-8',errors=\"ignore\")\r\n lines = getfile.read()\r\n lines = re.sub('[\\\\n\\\\t\\\\r]+', ' ', lines)\r\n eachLine = lines.split('.')\r\n list_sentences =[] #list of all sentences of token words\r\n for index in range(len(eachLine)):\r\n eachLine[index] = re.sub('[^A-Za-z ]+', ' ', eachLine[index].lower())\r\n eachLine = list(filter(None, eachLine))\r\n new_eachLine = []\r\n for data in eachLine:\r\n data_token = data.split(' ')\r\n data_token = list(filter(None, data_token))\r\n chunks = [data_token[x:x+cut_size] for x in range(0, len(data_token), cut_size)]\r\n new_eachLine.extend(chunks)\r\n eachLine = new_eachLine\r\n eachLine = list(filter(None, eachLine))\r\n for index in range(len(eachLine)):\r\n token_words = eachLine[index]\r\n #token_words = list(filter(None, token_words))\r\n new_token_words = [] #list of all token words in a sentences\r\n for each_tokenWords in token_words:\r\n if each_tokenWords not in stopwords:\r\n new_token_words.append(lemmatizer.lemmatize(each_tokenWords))\r\n list_sentences.append(new_token_words)\r\n book_data = list_sentences\r\n getfile.close()\r\n print(\"Read Book \" + str(count+1))\r\n saveString = \"tokenized_books/tokenized_book\" + str(count+1)\r\n with open(saveString, 'wb') as fp: pickle.dump(book_data, fp)\r\n \r\ndef calculate_strength(count):\r\n #with open('tokenized_book', 'rb') as fp: book_data = pickle.load(fp)\r\n disp_count = count+1\r\n #print(count)\r\n edgePairs = {}\r\n max_distance = 2\r\n #strengths = []\r\n loadString = \"tokenized_books/tokenized_book\" + str(disp_count)\r\n with open(loadString, 'rb') as fp: eachBook = pickle.load(fp)\r\n #print(disp_count)\r\n for source_sentence in range(len(eachBook)):\r\n for source_word in range(len(eachBook[source_sentence])):\r\n wordGap = 0\r\n index = source_word + 1\r\n for dest_word in range(index,len(eachBook[source_sentence])):\r\n item_pair = eachBook[source_sentence][source_word] + \"\\t\" + eachBook[source_sentence][dest_word]\r\n wordGap = wordGap + 1\r\n sentenceGap = 0\r\n conceptStrength = gaussianStrength(wordGap,sentenceGap)\r\n #conceptStrength = gaussianStrength(wordGap)\r\n if item_pair in edgePairs:\r\n edgePairs[item_pair] = edgePairs[item_pair] + conceptStrength\r\n else:\r\n edgePairs[item_pair] = conceptStrength\r\n sentenceGap = 1\r\n for dest_sentence in range(source_sentence+sentenceGap,len(eachBook)):\r\n if sentenceGap>max_distance:\r\n break\r\n else:\r\n for dest_word in range(len(eachBook[dest_sentence])):\r\n item_pair = eachBook[source_sentence][source_word] + \"\\t\" + eachBook[dest_sentence][dest_word]\r\n wordGap = wordGap + 1\r\n conceptStrength = gaussianStrength(wordGap,sentenceGap)\r\n #conceptStrength = gaussianStrength(wordGap)\r\n if item_pair in edgePairs:\r\n edgePairs[item_pair] = edgePairs[item_pair] + conceptStrength\r\n else:\r\n edgePairs[item_pair] = conceptStrength\r\n sentenceGap += 1\r\n \r\n eps = []\r\n for key, value in edgePairs.items():\r\n eps.append([key,value])\r\n #eps.sort(key=lambda elem: elem[2],reverse=True)\r\n fileNameEP = \"edgepairs/edgepair_strength_\"+str(disp_count)+\".txt\"\r\n print(fileNameEP)\r\n file = open(fileNameEP,'w+')\r\n for i in range(len(edgePairs)):\r\n if i=5000:\r\n break\r\n single_esp = line.rstrip().split('\\t')\r\n value = float(single_esp[2])\r\n item_pair = single_esp[0] + \"\\t\" +single_esp[1]\r\n if item_pair in edgePairs:\r\n edgePairs[item_pair] = edgePairs[item_pair] + value\r\n else:\r\n edgePairs[item_pair] = value\r\n count +=1\r\n eps = []\r\n for key, value in edgePairs.items():\r\n eps.append([key,value])\r\n eps.sort(key=lambda elem: elem[1],reverse=True)\r\n fileNameEP = \"merged_edgeList.txt\" \r\n file = open(fileNameEP,'w+')\r\n for i in range(len(edgePairs)):\r\n if i= 3 and len(rhyme2_list) >= 2: # call function again if not enough rhymes\n break\n return rhyme1_list, rhyme2_list\n\ndef find_rhymes(rhyme_list, num_to_find):\n endings = []\n for i in range(num_to_find):\n ending = random.choice(rhyme_list)\n ending_words = set(x[0] for x in endings)\n while ending[0] in ending_words:\n ending = random.choice(rhyme_list)\n endings.append(ending)\n return endings\n\ndef count_syllables(word_key):\n phone_list = word_dict[word_key]\n stressed_list = [x for x in phone_list if x[-1].isdigit()]\n num_syllables = len(stressed_list)\n return num_syllables\n\ndef fill_line(ending, total_syllables, return_stress=False, stress=None):\n while True:\n line_words = []\n num_syllables = count_syllables(ending)\n while num_syllables < total_syllables:\n new_syllables = total_syllables\n while num_syllables + new_syllables > total_syllables:\n new_word = random.choice(word_dict.keys())\n new_syllables = count_syllables(new_word)\n line_words.append(new_word)\n num_syllables += new_syllables\n line_words.append(ending)\n if return_stress:\n line_stress = ''.join(stress_dict[x] for x in line_words)\n if line_stress[1] == '1':\n return line_words, line_stress\n if stress:\n line_stress = ''.join(stress_dict[x] for x in line_words)\n matching = 0.0\n total = len(stress)\n for i in range(total):\n if stress[i] == line_stress[i]:\n matching += 1\n if (matching / total) >= 0.7 and line_stress[1] == '1':\n break\n return line_words\n\ndef make_limerick(len1, len2):\n rhyme1, rhyme2 = choose_rhymes()\n end1, end2, end5 = find_rhymes(rhyme1, 3)\n end3, end4 = find_rhymes(rhyme2, 2)\n filled1, stress1 = fill_line(end1, len1, return_stress=True)\n filled3, stress2 = fill_line(end3, len2, return_stress=True)\n limerick_tuple = (filled1, fill_line(end2, len1, stress=stress1),\n filled3, fill_line(end4, len2, stress=stress2),\n fill_line(end5, len1, stress=stress1))\n for line in limerick_tuple:\n text_line = ' '.join(x[0] for x in line)\n print(text_line)\n\nif __name__ == '__main__':\n for i in range(3):\n make_limerick(9, 6)\n print('')\n","sub_path":"cmu_limericks.py","file_name":"cmu_limericks.py","file_ext":"py","file_size_in_byte":3957,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"472023820","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[35]:\n\n\nAsem_dirc = [\": .byte\",\": .half\", \": .word\", \": .dword\", \": .asciiz\"]\nincr = {\": .byte\":\"1\",\": .half\":\"2\",\": .word\":\"4\",\": .dword\":\"8\",\": .asciiz\":\"1\"}\n\n\n\ndef KMPSearch(minor,major): \n\tl = len(minor) \n\tL = len(major) \n\tlcmin = [0]*l \n\tj = 0 \n\tcomputeLPSArray(minor,l,lcmin) \n\ti = 0\n\twhile i < L: \n\t\tif minor[j] == major[i]: \n\t\t\ti += 1\n\t\t\tj += 1\n\t\tif j == l: \n\t\t\treturn 1\n\t\t\tj = lcmin[j-1] \n\n\t\telif i < L and minor[j] != major[i]: \n\t\t\tif j != 0: \n\t\t\t\tj = lcmin[j-1] \n\t\t\telse: \n\t\t\t\ti += 1\n\ndef computeLPSArray(minor,l,lcmin): \n\tlen = 0 \n\tlcmin[0] \n\ti = 1\n\twhile i < l: \n\t\tif minor[i]== minor[len]: \n\t\t\tlen += 1\n\t\t\tlcmin[i] = len\n\t\t\ti += 1\n\t\telse: \n\t\t\tif len != 0: \n\t\t\t\tlen = lcmin[len-1] \n\t\t\telse: \n\t\t\t\tlcmin[i] = 0\n\t\t\t\ti += 1\n\n\n\n\ndef KMPSearch2(minor, major): \n\tll=[]\n\tl = len(minor) \n\tL = len(major) \n\tlcmin = [0]*l\n\tj = 0 \n\tcomputeLPSArray(minor,l,lcmin) \n\ti = 0 \n\twhile i < L: \n\t\tif minor[j] == major[i]: \n\t\t\ti += 1\n\t\t\tj += 1\n\n\t\tif j == l: \n\t\t\tll.append ((i-j) )\n\t\t\tj = lcmin[j-1] \n\n\t\telif i < L and minor[j] != major[i]: \n\t\t\tif j != 0: \n\t\t\t\tj = lcmin[j-1] \n\t\t\telse: \n\t\t\t\ti += 1\n\n\treturn ll\n\ndef computeLPSArray(minor,l,lcmin): \n\tlen = 0 \n\tlcmin[0] \n\ti = 1 \n\twhile iiids):\n \n \n if(i==po2):\n po = po+1\n po1 = po1 +1\n break\n else:\n if(len(p.findall(list1[i]))==0):\n po = po+1\n po1 = po1 +1\n \n flag = 0\n for i in (range(len(aas))):\n if(i>iids):\n \n \n if(KMPSearch(pat, aas[i])):\n flag = 1 \n cnt = cnt+1\n break\n else:\n if(len(p.findall(list1[i]))==0):\n cnt = cnt+1\n \n cnt = cnt - po \n if(flag==0):\n print(\"Error: \"+str(pat)+\" not Defined\")\n return \n #print(pat)\n immd = cnt*2\n if(immd<0):\n immd = -1*immd\n immd_= onescomp(dec2bin2(int(immd)))\n immd_ = array2string2(immd_)\n ll = (int(immd_,2)+1 ) \n immd = ll\n \n d1 = convert((dec2bin2(int(immd))[2:8]))\n d2 = ((dec2bin2(int(immd))[0]))\n d3 = convert((dec2bin2(int(immd))[8:12]))\n d4 = ((dec2bin2(int(immd))[1]))\n reg_1 = array2string(dec2bin(int(reg1)))\n reg_2 = array2string(dec2bin(int(reg2)))\n \n output = str(d2)+str(d1)+str(reg_1)+str(reg_2)+str(funt3)+str(d3)+str(d4)+\"1100011\"\n \n return output \n\n \n\n \n \ndef Uformat(reg1,imm,b):\n reg_1 = array2string(dec2bin(int(reg1)))\n op=U[b]\n immd_ = array2string3(dec2bin3(int(imm)))\n \n if(op==\"auipc\"):\n output = str(immd_)+str(reg_1)+\"0010111\"\n else:\n output = str(immd_)+str(reg_1)+\"0110111\"\n \n return output\ndef UJformat(reg1,aas,string,po2,iids):\n hhf = string.replace(\"\\n\",\"\").split(\",\")\n pat = hhf[1]\n po=0\n pat = pat+':'\n p = re.compile('[\\w]*[:]') \n cnt=0\n cnt1 = 0\n \n po=0\n po1 = 0\n for i in (range(len(aas))):\n if(i>iids):\n \n \n if(i==po2):\n po = po+1\n po1 = po1 +1\n break\n else:\n if(len(p.findall(list1[i]))==0):\n po = po+1\n po1 = po1 +1\n \n flag = 0\n for i in (range(len(aas))):\n if(i>iids):\n \n \n if(KMPSearch(pat, aas[i])):\n flag = 1 \n cnt = cnt+1\n break\n else:\n if(len(p.findall(list1[i]))==0):\n cnt = cnt+1\n if(flag==0):\n print(\"Error: \"+str(pat)+\" not Defined\")\n return\n \n cnt = cnt - po \n \n \n immd = cnt*2\n \n if(immd<0):\n immd = -1*immd\n immd_= onescomp(dec2bin3(int(immd)))\n \n immd_ = array2string3(immd_)\n ll = (int(immd_,2)+1 ) \n immd = ll\n d4 = convert((dec2bin3(int(immd))[1:9]))\n d2 = convert((dec2bin3(int(immd))[10:20]))\n d1 = ((dec2bin3(int(immd))[0]))\n d3 = ((dec2bin3(int(immd))[9]))\n reg_1 = array2string(dec2bin(int(reg1)))\n \n \n output = str(d1)+str(d2)+str(d3)+str(d4)+str(reg_1)+\"1101111\"\n return output\n\n\n\n\n# Array of formats\n\ndef regno(string, op,aas,po,iids):\n passorder = []\n p = re.compile('[x]\\d+') \n jfj= p.findall(string)\n for z in range(len(jfj)):\n passorder.append(int(jfj[z].replace(\"x\",\"\")))\n \n funct3pass = str(R_funct3.get(str(op)))\n a,b=formatfound(string)\n var=formats[a]\n \n funct3pass = foundfunct3(a,b)\n \n if(var==\"R\"): #let vari contain the format\n ss = string\n ss = ss.split(\",\")\n ss1 =int( ss[0].split(\" \")[1].replace(\"x\",\"\"))\n ss2 = int( ss[1].replace(\"x\",\"\"))\n ss3 = int(ss[2].replace(\"x\",\"\"))\n return(Rformat(funct3pass,ss3,ss2,ss1,b))\n if(var==\"SB\"):\n if(len(passorder)==3):\n print(\"Error:Expecting one immediate value\")\n return\n if(len(passorder)==1):\n h1=(passorder[0])\n h2=h1\n else:\n h1=(passorder[0])\n h2=(passorder[1])\n \n return( SBformat(funct3pass,h2,h1,aas,string,po,iids))\n if(var==\"S\"):\n if(len(passorder)==3):\n print(\"Error:Expecting one immediate value\")\n return\n if(len(passorder)==1):\n h1=(passorder[0])\n h2=h1\n else:\n h1=(passorder[0])\n h2=(passorder[1])\n imm=immediate_found(string)\n \n return( Sformat(funct3pass,h2,h1,imm)) \n \n \n \n if(var==\"I\"):\n if(len(passorder)==3):\n print(\"Error:Expecting one immediate value\")\n return\n if(len(passorder)==1):\n h1=(passorder[0])\n h2=h1\n else:\n h1=(passorder[0])\n h2=(passorder[1])\n imm=immediate_found(string)\n return(Iformat(funct3pass,h2,h1,imm,string)) #let imm gets the immediate value\n if(var==\"U\"):\n ss = string.split(\",\")\n imm =int(ss[1],16)\n if(imm<0):\n imm = -1*imm\n immd_= onescomp(dec2bin2(int(imm)))\n immd_ = array2string2(immd_)\n ll = (int(immd_,2)+1 ) \n imm = ll\n \n \n return(Uformat((passorder[0]),imm,b))\n if(var==\"UJ\"):\n \n \n return(UJformat((passorder[0]),aas,string,po,iids))\n \nimmq = []\nimport re \ndef immediate_found(string):\n p = re.compile('[,][-/+]?[\\d]+') \n ff = str(p.findall(string))\n ff = ff.replace(\",\",\"\")\n ff = (ff.replace(\"'\",\"\"))\n ff = ff.replace(\"[\",\"\")\n ff = ff.replace(\"]\",\"\")\n \n ff = int(ff)\n return ff\n\n\n \n\n\n\ng = open(\"oo.mc.txt\",\"w+\")\np = re.compile('[+|-]?\\d+') \n\n\ndef assebly_dirc(string,PC,aas,po,iids):\n \n \n import re \n \n\n for i in range(len(Asem_dirc)):\n \n if(KMPSearch(Asem_dirc[i],string)==1):\n \n if(i!=4):\n p = re.compile('[+|-]?\\d+') \n num87 = p.findall(string)\n \n for j in range(len(num87)):\n g.write(str(hex(PC[0]))+\" \"+str(hex(int(num87[j])))+\"\\n\")\n PC[0] = PC[0] + int(incr[str(Asem_dirc[i])])\n elif(i==4):\n saa = string.split(\" \")[2]\n p = re.compile('\\w') \n daa = p.findall(saa)\n for j in range(len(daa)):\n g.write(str(hex(PC[0]))+\" \"+str(hex(ord(daa[i])-ord('a')+10))+\"\\n\")\n PC[0] = PC[0] + int(incr[str(Asem_dirc[i])])\n \n \n return\n \n a,b=formatfound(string)\n op=formats[a]\n \n \n g.write(str(hex(PC[0]))+\" \"+str(hex(int(regno(string, op,aas,po,iids),base = 2)))+\"\\n\")\n #print(regno(string, op,aas,po,iids))\n PC[0] = PC[0] + 4\ndef assebly_dirc1(string,PC,aas,po):\n \n \n import re \n \n\n for i in range(len(Asem_dirc)):\n \n if(KMPSearch(Asem_dirc[i],string)==1):\n \n \n if(i!=4):\n p = re.compile('[+|-]?\\d+')\n string = string.split(\":\")[1]\n num87 = p.findall(string)\n \n for j in range(len(num87)):\n immd = int(num87[j])\n if(immd<0):\n immd = -1*immd\n if(str(Asem_dirc[i])==\": .byte\"):\n immd_= onescomp(dec2bin8(int(immd)))\n\n immd_ = array2string8(immd_)\n if(str(Asem_dirc[i])==\": .half\"):\n immd_= onescomp(dec2bin16(int(immd)))\n\n immd_ = array2string16(immd_)\n if(str(Asem_dirc[i])==\": .word\"):\n \n immd_= onescomp(dec2bin32(int(immd)))\n\n immd_ = array2string32(immd_)\n if(str(Asem_dirc[i])==\": .dword\"):\n immd_= onescomp(dec2bin64(int(immd)))\n\n immd_ = array2string64(immd_)\n \n ll = (int(immd_,2)+1 ) \n immd = ll\n \n g.write(str(hex(PC[0]))+\" \"+str(hex(int(immd)))+\"\\n\")\n PC[0] = PC[0] + int(incr[str(Asem_dirc[i])])\n elif(i==4):\n \n saa = string.split(\" \")[2]\n p = re.compile('\\w') \n daa = p.findall(saa)\n for j in range(len(daa)):\n \n g.write(str(hex(PC[0]))+\" \"+str(hex(ord(daa[j])-ord('a')+10))+\"\\n\")\n PC[0] = PC[0] + int(incr[str(Asem_dirc[i])])\n \n \n return\n \n \n \n \nf= open(\"text.mc.txt\",\"r+\")\ng = open(\"oo.mc.txt\",\"w+\")\nlist1=f.readlines()\nPC= []\nPC.append(268435456)\np = re.compile('[\\w]*[:]') \nap = 0\nfor i in range(len(list1)):\n if(i!=0):\n if(KMPSearch(\".text\", list1[i])): \n ap =i\n break\n \n if(len(p.findall(list1[i]))!=0):\n \n \n if(len(list1[i])==1): break\n\n assebly_dirc1(list1[i],PC,list1,i)\nPC[0]=0\nfor i in range(len(list1)):\n if(i>ap):\n if(len(p.findall(list1[i]))==0):\n \n if(len(list1[i])==1): break\n\n assebly_dirc(list1[i],PC,list1,i,ap)\n \n#print(d1,d2,d3,d4,reg_1)\n#output = str(d2)+str(d4)+str(d1)+str(reg_1)+\"1101111\"\n#0000 0000 11111111 000 1 11111111 00001 1101111 \n \n\n\n\n\ng = open(\"oo.mc.txt\",\"w+\")\n\n\n\n\n\n\n\n","sub_path":"PHASE1.py","file_name":"PHASE1.py","file_ext":"py","file_size_in_byte":19939,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"646454691","text":"# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom django.urls import reverse_lazy\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom horizon import exceptions\nfrom horizon import tabs\n\nfrom gbpui import client\nfrom gbpui import column_filters as gfilters\n\nfrom gbpui.panels.application_policy import tables\n\nPolicyRulesTable = tables.PolicyRulesTable\nPolicyClassifiersTable = tables.PolicyClassifiersTable\nPolicyActionsTable = tables.PolicyActionsTable\n\n\nclass PolicyActionsTab(tabs.TableTab):\n table_classes = (PolicyActionsTable,)\n name = _(\"Policy Actions\")\n slug = \"policyactions\"\n template_name = \"horizon/common/_detail_table.html\"\n\n def get_policyactionstable_data(self):\n actions = []\n try:\n actions = client.policyaction_list(self.tab_group.request,\n tenant_id=self.tab_group.request.user.tenant_id)\n a = lambda x, y: gfilters.update_policyaction_attributes(x, y)\n actions = [a(self.request, item) for item in actions]\n except Exception as e:\n msg = _('Unable to retrieve actions list. %s') % (str(e))\n exceptions.handle(self.tab_group.request, msg)\n return actions\n\n\nclass PolicyClassifiersTab(tabs.TableTab):\n table_classes = (PolicyClassifiersTable,)\n name = _(\"Policy Classifiers\")\n slug = \"policyclassifiers\"\n template_name = \"horizon/common/_detail_table.html\"\n\n def get_policyclassifierstable_data(self):\n try:\n classifiers = client.policyclassifier_list(self.tab_group.request,\n tenant_id=self.tab_group.request.user.tenant_id)\n except Exception:\n classifiers = []\n exceptions.handle(self.tab_group.request,\n _('Unable to retrieve classifier list.'))\n else:\n classifiers = gfilters.update_classifier_attributes(classifiers)\n return classifiers\n\n\nclass PolicyRulesTab(tabs.TableTab):\n table_classes = (PolicyRulesTable,)\n name = _(\"Policy Rules\")\n slug = \"policyrules\"\n template_name = \"horizon/common/_detail_table.html\"\n\n def get_policyrulestable_data(self):\n try:\n policy_rules = client.policyrule_list(self.tab_group.request,\n tenant_id=self.tab_group.request.user.tenant_id)\n policy_rules = [gfilters.update_policyrule_attributes(\n self.request, item) for item in policy_rules]\n except Exception:\n policy_rules = []\n exceptions.handle(self.tab_group.request,\n _('Unable to retrieve policy-rule list.'))\n\n for rule in policy_rules:\n rule.set_id_as_name_if_empty()\n\n return policy_rules\n\n\nclass ApplicationPoliciesTab(tabs.TableTab):\n table_classes = (tables.ApplicationPoliciesTable,)\n name = _(\"Policy Rule Set\")\n slug = \"application_policies\"\n template_name = \"horizon/common/_detail_table.html\"\n\n def get_application_policies_table_data(self):\n policy_rule_sets = []\n try:\n policy_rule_sets = client.policy_rule_set_list(\n self.tab_group.request,\n tenant_id=self.tab_group.request.user.tenant_id)\n policy_rule_sets = [gfilters.update_pruleset_attributes(\n self.request, item) for item in policy_rule_sets]\n except Exception:\n exceptions.handle(\n self.tab_group.request,\n _('Unable to retrieve policy rule set list.'))\n\n for policy_rule_set in policy_rule_sets:\n policy_rule_set.set_id_as_name_if_empty()\n return policy_rule_sets\n\n\nclass ApplicationPoliciesTabs(tabs.TabGroup):\n slug = \"application_policies_tabs\"\n tabs = (ApplicationPoliciesTab,\n PolicyRulesTab,\n PolicyClassifiersTab,\n PolicyActionsTab)\n sticky = True\n\n\nclass PolicyRuleSetDetailsTab(tabs.Tab):\n name = _(\"Policy Rule Set Details\")\n slug = \"policy_rule_setdetails\"\n template_name = \"project/application_policy/_policy_rule_set_details.html\"\n failure_url = reverse_lazy('horizon:project:policy_rule_set:index')\n\n def get_context_data(self, request):\n cid = self.tab_group.kwargs['policy_rule_set_id']\n try:\n policy_rule_set = client.policy_rule_set_get(request, cid)\n rules = client.policyrule_list(\n request, tenant_id=request.user.tenant_id,\n policy_rule_set_id=policy_rule_set.id)\n rules = [\n item for item in rules if item.id in\n policy_rule_set.policy_rules]\n rules_with_details = []\n for rule in rules:\n r = {}\n r['name'] = rule.name\n r['id'] = rule.id\n action_list = []\n for aid in rule.policy_actions:\n action = client.policyaction_get(request, aid)\n a = {'id': action.id}\n if action.action_value:\n if action.action_type == 'redirect':\n scspec = client.get_servicechain_spec(request,\n action.action_value)\n a['name'] = \"Redirect:%s\" % scspec.name\n else:\n values = (str(action.action_type),\n str(action.action_value))\n name = \"%s:%s\" % values\n a['name'] = name\n else:\n a['name'] = str(action.action_type)\n action_list.append(a)\n r['actions'] = action_list\n r['classifier'] = client.policyclassifier_get(\n request, rule.policy_classifier_id)\n rules_with_details.append(r)\n except Exception as e:\n msg = _('Unable to retrieve policy_rule_set details.') % (str(e))\n exceptions.handle(request, msg, redirect=self.failure_url)\n return {'policy_rule_set': policy_rule_set,\n 'rules_with_details': rules_with_details}\n\n\nclass PolicyRuleSetDetailsTabs(tabs.TabGroup):\n slug = \"policy_rule_settabs\"\n tabs = (PolicyRuleSetDetailsTab,)\n\n\nclass PolicyRulesDetailsTab(tabs.Tab):\n name = _(\"Policy Rule Details\")\n slug = \"policyruledetails\"\n template_name = \"project/application_policy/_policyrules_details.html\"\n failure_url = reverse_lazy('horizon:project:policyrule:index')\n\n def get_context_data(self, request):\n ruleid = self.tab_group.kwargs['policyrule_id']\n actions = []\n classifiers = []\n try:\n policyrule = client.policyrule_get(request, ruleid)\n actions = client.policyaction_list(request,\n tenant_id=request.user.tenant_id, policyrule_id=ruleid)\n actions = [\n item for item in actions if item.id in\n policyrule.policy_actions]\n classifiers = client.policyclassifier_list(\n request, tenant_id=request.user.tenant_id,\n policyrule_id=ruleid)\n classifiers = [\n item for item in classifiers if\n item.id == policyrule.policy_classifier_id]\n except Exception:\n exceptions.handle(request,\n _('Unable to retrieve policyrule details.'),\n redirect=self.failure_url)\n return {'policyrule': policyrule,\n 'classifiers': classifiers,\n 'actions': actions}\n\n\nclass PolicyRuleDetailsTabs(tabs.TabGroup):\n slug = \"policyruletabs\"\n tabs = (PolicyRulesDetailsTab,)\n\n\nclass PolicyClassifierDetailsTab(tabs.Tab):\n name = _(\"Policy Classifier Details\")\n slug = \"policyclassifierdetails\"\n template_name = \"project/application_policy/_policyclassifier_details.html\"\n failure_url = reverse_lazy('horizon:project:policy_rule_set:index')\n\n def get_context_data(self, request):\n pcid = self.tab_group.kwargs['policyclassifier_id']\n try:\n policyclassifier = client.policyclassifier_get(request, pcid)\n policyclassifier = gfilters.update_classifier_attributes(\n policyclassifier)\n except Exception:\n exceptions.handle(request,\n _('Unable to retrieve policy_rule_set details.'),\n redirect=self.failure_url)\n return {'policyclassifier': policyclassifier}\n\n\nclass PolicyClassifierDetailsTabs(tabs.TabGroup):\n slug = \"policyclassifiertabs\"\n tabs = (PolicyClassifierDetailsTab,)\n\n\nclass PolicyActionDetailsTab(tabs.Tab):\n name = _(\"Policy Action Details\")\n slug = \"policyactiondetails\"\n template_name = \"project/application_policy/_policyaction_details.html\"\n failure_url = reverse_lazy('horizon:project:policy_rule_set:index')\n\n def get_context_data(self, request):\n paid = self.tab_group.kwargs['policyaction_id']\n try:\n policyaction = client.policyaction_get(request, paid)\n policyaction = gfilters.update_policyaction_attributes(request,\n policyaction)\n except Exception:\n exceptions.handle(request,\n _('Unable to retrieve policyaction details.'),\n redirect=self.failure_url)\n return {'policyaction': policyaction}\n\n\nclass PolicyActionDetailsTabs(tabs.TabGroup):\n slug = \"policyactiontabs\"\n tabs = (PolicyActionDetailsTab,)\n","sub_path":"gbpui/panels/application_policy/tabs.py","file_name":"tabs.py","file_ext":"py","file_size_in_byte":10061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"76765200","text":"from ..Base import Base, db\nfrom sqlalchemy.sql import func\n\nfrom flask import current_app\n\nimport re\nimport os\nimport math\nimport base64\nfrom io import BytesIO\nfrom PIL import Image as PIL_Image\nimport time\n\n\nclass Image(Base):\n __abstract__ = True\n\n file_descriptor = db.Column('file_descriptor', db.String(128), nullable = False)\n file_mime = db.Column('file_mime', db.String(32), nullable = False)\n\n def __repr__(self):\n return \"\" % (self.id, self.file_mime)\n\n def __str__(self):\n return \"%s\" % (self.id, )\n\n def serialize(self):\n obj = super(Image,self).serialize()\n \n obj['file_descriptor'] = self.file_descriptor\n obj['file_mime'] = self.file_mime\n \n return obj\n\n @staticmethod\n def save(image, filename = None, max_width = None, max_height = None):\n \n directory = os.path.join(current_app.config['UPLOADED_FILES_DEST'])\n try:\n os.stat(directory)\n except:\n os.mkdir(directory)\n \n if filename is None:\n filename = 'image_'+str(math.floor(1000*time.time()))\n\n file_descriptor = os.path.join(directory, filename)\n file_mime = image.format\n\n if file_mime == 'PNG':\n file_descriptor = file_descriptor + '.png'\n elif file_mime in ('JPEG', None,):\n image = image.convert('RGB')\n file_mime = 'JPEG'\n file_descriptor = file_descriptor + '.jpg'\n \n\n #\n # convent to jpg\n #\n\n if file_mime == 'PNG':\n #re-convert to jpeg\n image = image.convert('RGB')\n file_descriptor = re.sub(r'\\.png$', '.jpg', file_descriptor)\n file_mime = 'JPEG'\n \n \n width, height = image.size\n\n \n if max_width is not None and width > max_width:\n if file_mime == 'PNG':\n #re-convert to jpeg\n image = image.convert('RGB')\n file_descriptor = re.sub(r'\\.png$','.jpg',file_descriptor)\n file_mime = 'JPEG'\n\n height = math.floor(height*max_width/width)\n width = max_width\n image = image.resize((width, height), PIL_Image.ANTIALIAS)\n\n if max_height is not None and height > max_height:\n if file_mime == 'PNG':\n #re-convert to jpeg\n image = image.convert('RGB')\n file_descriptor = re.sub(r'\\.png$', '.jpg', file_descriptor)\n file_mime = 'JPEG'\n\n width = math.floor(width*max_height/height)\n height = max_height\n image = image.resize((width, height), PIL_Image.ANTIALIAS)\n\n\n if file_mime == 'PNG':\n image.save(file_descriptor)\n elif file_mime in ('JPEG', None,):\n image.save(file_descriptor, optimize=True, quality=95)\n else:\n print('File mimetype = %s unknown' % (file_mime,))\n return None, None\n\n return file_descriptor, file_mime\n\n\n @staticmethod\n def save_from_urlData(data, filename = None, max_width = None, max_height = None):\n\n r = re.compile('data:(.+);base64,')\n match = re.search(r, data)\n file_mime = match.group(1)\n\n image_data = bytes(re.sub(r, '', data), encoding='ascii')\n\n image = PIL_Image.open(BytesIO(base64.b64decode(image_data)))\n croped_image = image.crop((0,0,min(image.size),min(image.size)))\n\n return Image.save(croped_image,filename, max_width,max_height)\n \n\n \n","sub_path":"models/main/Image.py","file_name":"Image.py","file_ext":"py","file_size_in_byte":3575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"109398403","text":"import inspect\nimport pyspark\nimport sys\nimport subprocess\nimport shutil\nimport tempfile\nimport uuid\nimport warnings\nimport os\n\n__all__ = ['Captain']\n\nhandle_del = False\n\ntry:\n from subprocess import DEVNULL # py3k\nexcept ImportError:\n import os\n DEVNULL = open(os.devnull, 'wb')\n\n\nclass Captain(object):\n \"\"\"\n The captain of the coffee boat is used to setup packages for Spark to use.\n\n To use it run init, call the `add_pip_packages` for whatever you wish to\n add and then `launch_ship` before you create your SparkContext.\n The coffee boat captain currently works by creating a conda env and\n shipping it.\n\n \"\"\"\n def __init__(self,\n use_conda=True,\n install_local=True,\n env_name=None,\n working_dir=None,\n accept_conda_license=False,\n python_version=None,\n conda_path=None):\n \"\"\"Create a captain to captain the coffee boat and install the packages.\n\n Currently only supports conda, TODO:PEX for others.\n\n :param use_conda: Build a conda package rather than a pex package.\n :param install_local: Attempt to install packages locally as well\n :param env_name: Enviroment name to use. May squash existing enviroment\n :param working_dir: Directory for working in.\n :param accept_conda_license: If you accept the conda license. Set it\n to True to work.\n :param conda_path: Path to conda (optional). Otherwise searches system\n or self-installs.\n\n \"\"\"\n self.accept_conda_license = accept_conda_license\n self.working_dir = working_dir\n self.install_local = install_local\n self.env_name = env_name or \"auto{0}\".format(str(uuid.uuid4()))\n # Kind of hackey, but yay shells....\n self.env_name = self.env_name.replace('-', \"_\")\n self.python_version = (python_version or\n '.'.join(map(str, sys.version_info[:3])))\n if not self.working_dir:\n self.working_dir = tempfile.mkdtemp(prefix=\"coffee_boat_tmp_\")\n import atexit\n if handle_del:\n atexit.register(lambda: shutil.rmtree(self.working_dir))\n self.use_conda = use_conda\n self.conda = conda_path\n self.pip_pkgs = []\n return\n\n def add_pip_packages(self, *pkgs):\n \"\"\"Add pip packages\"\"\"\n active_context = pyspark.context.SparkContext._active_spark_context\n if self.install_local:\n args = [\"pip\", \"install\"]\n args.extend(pkgs)\n subprocess.check_call(args, stdout=DEVNULL)\n self.pip_pkgs.extend(pkgs)\n\n def launch_ship(self):\n \"\"\"Creates a relocatable environment and distributes it.\n\n .. note::\n\n This function *should* be called before you init your SparkContext, if it's\n called after we need to do some sketchy things to make it work.\n \"\"\"\n # Doing sketchy things with the gateway if we've already stopped the context\n active_context = pyspark.context.SparkContext._active_spark_context\n gateway = pyspark.context.SparkContext._gateway\n if active_context is None and gateway is not None:\n try:\n pyspark.context.SparkContext._gateway.jvm.java.lang.System.exit(0)\n except Exception:\n pass\n self._cleanup_keys()\n pyspark.context.SparkContext._gateway = None\n elif active_context is not None:\n warnings.warn(\n \"Launching on an existing SparkContext. Packages will only be available to RDDs\"\n \"created from here forward. If this makes you sad, stop the Spark context and\"\n \"re-create those RDDs you want to have access to your packages in.\")\n\n\n if self.use_conda:\n self._setup_or_find_conda()\n return self._launch_conda_ship()\n else:\n return self._launch_pex()\n\n def _launch_conda_ship(self):\n \"\"\"Create a conda enviroment, zips it up, and manipulate the environment\n variables.\n\n \"\"\"\n # Create the conda package env spec\n pkgs = [\"\"]\n pkgs.extend(map(str, self.pip_pkgs))\n pip_packages = '\\n - '.join(pkgs)\n\n # Create the package_spec\n base_package_spec = inspect.cleandoc(\"\"\"\n name: {0}\n dependencies:\n - python=={1}\n - anaconda\n - pip\n - pip:\n \"\"\").format(self.env_name, self.python_version)\n package_spec = \"{0}{1}\".format(base_package_spec, pip_packages)\n package_spec_file = tempfile.NamedTemporaryFile(dir=self.working_dir,\n delete=handle_del)\n package_spec_path = package_spec_file.name\n print(\"Writing package spec to {0}.\".format(package_spec_path))\n package_spec_file.write(package_spec)\n package_spec_file.flush()\n\n # Create the conda env\n conda_prefix = os.path.join(self.working_dir, self.env_name)\n print(\"Creating conda env\")\n if os.path.exists(conda_prefix):\n print(\"Cleaining up old prefix {0}\".format(conda_prefix))\n subprocess.check_call([\"rm\", \"-rf\", conda_prefix])\n subprocess.check_call([self.conda, \"env\", \"create\",\n \"-f\", package_spec_path,\n \"--prefix\", conda_prefix],\n stdout=DEVNULL)\n\n # Package it for distro\n zip_name = \"coffee_boat_{0}.zip\".format(self.env_name)\n zip_target = os.path.join(self.working_dir, zip_name)\n print(\"Packaging conda env\")\n subprocess.check_call([\"zip\", zip_target, \"-r\", conda_prefix],\n stdout=DEVNULL)\n relative_python_path = \".\" + conda_prefix + \"/bin/python\"\n\n # Make a self extractor script\n runner_script = inspect.cleandoc(\"\"\"#!/bin/bash\n if [ -f {0} ];\n then\n unzip {0} &>/dev/null && rm {0} &> /dev/null\n fi\n {1} \"$@\" \"\"\".format(zip_name, relative_python_path))\n script_name = \"coffee_boat_runner_{0}.sh\".format(self.env_name)\n runner_script_path = os.path.join(self.working_dir, script_name)\n with open(runner_script_path, 'w') as f:\n f.write(runner_script)\n subprocess.check_call([\"chmod\", \"a+x\", runner_script_path])\n\n # Adjust environment variables so that the env gets distributed.\n old_args = os.environ.get(\"PYSPARK_SUBMIT_ARGS\", \"pyspark-shell\")\n # Backup the old arguments\n if \"coffee_boat\" not in old_args:\n os.environ[\"BACK_PYSPARK_SUBMIT_ARGS\"] = old_args\n else:\n old_args = os.environ.get(\"BACK_PYSPARK_SUBMIT_ARGS\", \"pyspark-shell\")\n new_args = \"--files {0},{1} {2}\".format(zip_target, runner_script_path, old_args)\n print(\"using {0} as python arguments\".format(new_args))\n os.environ[\"PYSPARK_SUBMIT_ARGS\"] = new_args\n # Handle active/already running contexts.\n sc = pyspark.context.SparkContext._active_spark_context\n if sc is not None:\n print(\"Adding {0} & {1} to existing sc\".format(zip_target, runner_script_path))\n sc.addFile(zip_target)\n sc.addFile(runner_script_path)\n print(\"Updating python exec on existing sc\")\n sc.pythonExec = \"./{0}\".format(script_name)\n else:\n print(\"No active context, depending on submit args.\")\n\n if \"PYSPARK_GATEWAY_PORT\" in os.environ:\n print(\"Hey the Java process is already running, this might not work.\")\n os.environ[\"PYSPARK_PYTHON\"] = \"./{0}\".format(script_name)\n\n def _launch_pex(self):\n \"\"\" Create a pex environment.\"\"\"\n pass\n\n def _setup_or_find_conda(self):\n \"\"\"Find conda or set up a conda installation\"\"\"\n # Check if we need to setup conda or return if we already have one\n rc = subprocess.call(['which', 'conda'])\n if rc == 0:\n self.conda = \"conda\"\n return\n if self.conda is not None:\n return\n\n # Install conda if we need to\n if not self.accept_conda_license:\n raise Exception(\"Please accept the conda license by setting \"\n \"accept_conda_license\")\n python_version = sys.version_info[0]\n url = \"https://repo.continuum.io/miniconda/Miniconda%d-latest-Linux-x86_64.sh\" % python_version\n print(\"Downloading conda from %s to %s\" % (url, self.working_dir))\n mini_conda_target = \"%s/%s\" % (self.working_dir, \"miniconda.sh\")\n subprocess.check_call([\"wget\", url, \"-O\", mini_conda_target, \"-nv\"],\n shell=False,\n stdout=DEVNULL)\n print(\"Running conda setup....\")\n subprocess.check_call([\"chmod\", \"a+x\", mini_conda_target],\n shell=False,\n stdout=DEVNULL)\n conda_target = \"%s/%s\" % (self.working_dir, \"conda\")\n subprocess.check_call([mini_conda_target, \"-b\", \"-p\", conda_target],\n stdout=DEVNULL,\n stderr=DEVNULL)\n self.conda = \"%s/bin/conda\" % conda_target\n\n def _cleanup_keys(self):\n import os\n def cleanup_key(name):\n if name in os.environ:\n del os.environ[name]\n keys = [\n \"PYSPARK_PYTHON\",\n \"PYSPARK_GATEWAY_PORT\",\n \"_PYSPARK_DRIVER_CALLBACK_HOST\",\n \"_PYSPARK_DRIVER_CALLBACK_PORT\"]\n map(cleanup_key, keys)\n","sub_path":"coffee_boat/captain.py","file_name":"captain.py","file_ext":"py","file_size_in_byte":9667,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"257155949","text":"from django.urls import path\n\nfrom . import views\n\napp_name = 'app_manager'\n\nurlpatterns = [\n path('dashboard/', views.dashboard, name='dashboard'), # show companies\n\n # path('company/new', views.company_new, name='company_new'), # add new company\n # path('company/', views.company_show, name='company_show'), # show the company\n\n path('group/new', views.group_new, name='group_new'), # add new group\n path('group/modify//', views.group_modify, name='group_modify'), # modify the group\n path('group/del/', views.group_del, name='group_del'), # add new group\n\n path('group//device/new', views.device_new, name='device_new'), # add new device\n path('device/modify/', views.device_modify, name='device_show'), # modify the device\n path('device/del/', views.device_del, name='device_del'), # add new device\n\n path('user/del//', views.user_del, name='user_del'), # del user from company\n path('user/add//', views.user_add, name='user_add'), # add company_id_to_user\n\n path('logo/upload', views.logo_upload, name='logo_upload'), # show the company\n\n path('device//checklist/new', views.checklist_new, name='checklist_new'), # add new checklist\n path('device//checklist/modify/', views.checklist_new, name='checklist_new'), # modify checklist\n path('device//checklist/del/', views.checklist_new, name='checklist_new'), # del checklist\n\n path('generic//', views.list_show, name='list_show'), # add new group\n path('add//', views.list_edit, name='list_edit'), # add new group\n\n]\n","sub_path":"app_manager/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"560531582","text":"__author__ = 'PyBeaner'\n\n# You need to check the start or end of a string for specific text patterns, such as filename\n# extensions, URL schemes, and so on.\n\nfilename = \"spam.txt\"\nprint(filename.endswith(\".txt\"))\nprint(filename.startswith(\"file:\"))\n\nurl = \"http://www.python.org\"\nprint(url.startswith(\"http:\"))\n\n\nimport os\nfilenames = os.listdir(\"..\")\nprint(filenames)\nr = [name for name in filenames if name.startswith(\"Matching\")]\nprint(r)\nprint(any(name.endswith(\"String\") for name in filenames))\n\nfrom urllib.request import urlopen\n\ndef read_data(name):\n # tuple works(but not list or set)\n if name.startswith(\"http:\",\"https:\",\"ftp:\"):\n return urlopen(name).read()\n with open(name) as f:\n return f.read()\n\nchoices = [\"http\",\"ftp\"]\nurl = \"http://www.python.org\"\n# startswith first arg must be str or a tuple of str, not list\n# url.startswith(choices)\nurl.startswith(tuple(choices))","sub_path":"Chapter 2. Strings and Text/Matching Text at the Start or End of a String/example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":907,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"434976650","text":"#! /usr/bin/env python\nimport os\nimport sys\nimport unittest\nfrom irods.session import iRODSSession\nfrom irods.exception import NetworkException\nimport irods.test.config as config\n\n\nclass TestConnections(unittest.TestCase):\n\n def setUp(self):\n self.sess = iRODSSession(host=config.IRODS_SERVER_HOST,\n port=config.IRODS_SERVER_PORT, # 4444: why?\n user=config.IRODS_USER_USERNAME,\n password=config.IRODS_USER_PASSWORD,\n zone=config.IRODS_SERVER_ZONE)\n\n def tearDown(self):\n '''Close connections\n '''\n self.sess.cleanup()\n\n def test_connection(self):\n with self.sess.pool.get_connection() as conn:\n self.assertTrue(conn)\n\n def test_connection_destructor(self):\n conn = self.sess.pool.get_connection()\n conn.__del__()\n conn.release(destroy=True)\n\n def test_failed_connection(self):\n # mess with the account's port\n self.sess.pool.account.port = 6666\n\n # try connecting\n with self.assertRaises(NetworkException):\n self.sess.pool.get_connection()\n\n # set port back\n self.sess.pool.account.port = config.IRODS_SERVER_PORT\n\n def test_send_failure(self):\n with self.sess.pool.get_connection() as conn:\n # try to close connection twice, 2nd one should fail\n conn.disconnect()\n with self.assertRaises(NetworkException):\n conn.disconnect()\n\n def test_reply_failure(self):\n with self.sess.pool.get_connection() as conn:\n # close connection\n conn.disconnect()\n\n # try sending reply\n with self.assertRaises(NetworkException):\n conn.reply(0)\n\n\nif __name__ == '__main__':\n # let the tests find the parent irods lib\n sys.path.insert(0, os.path.abspath('../..'))\n unittest.main()\n","sub_path":"irods/test/connection_test.py","file_name":"connection_test.py","file_ext":"py","file_size_in_byte":1960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"492122481","text":"\"\"\"This script increases the minor version of the current project.\"\"\"\n\nimport os\n\ndef increaseVersion(fname):\n assert os.path.exists(fname)\n with open(fname) as f:\n lines = f.read().split(\"\\n\")\n for i, line in enumerate(lines):\n if line.startswith(\"__version__\"):\n oldVersionString = line.split(\"'\")[1]\n newVersion = [int(x) for x in oldVersionString.split(\".\")]\n newVersion[-1] += 1\n newVersionString = \".\".join([str(x) for x in newVersion])\n lines[i] = lines[i].replace(oldVersionString, newVersionString)\n with open(fname, 'w') as f:\n f.write(\"\\n\".join(lines))\n print(f\"Upgraded: {oldVersionString} -> {newVersionString}\")\n return\n\n\nif __name__ == \"__main__\":\n PATH_HERE = os.path.abspath(os.path.dirname(__file__))\n versionFile = os.path.abspath(PATH_HERE+\"/../../src/pyabf/__init__.py\")\n increaseVersion(versionFile)\n","sub_path":"dev/scripts/versionIncrease.py","file_name":"versionIncrease.py","file_ext":"py","file_size_in_byte":925,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"582875373","text":"from OpenGL.GL import *\nfrom OpenGL.GLUT import *\nfrom OpenGL.GLU import *\n\nfrom math import pi, cos, sin\nfrom numpy import arctan2\nfrom node import *\n\nclass Player:\n def __init__(s, parent_node, direction = 0):\n s.parent_node = parent_node #the maze node that the player is currently inside of\n s.pos = [0.0, 0.0, 0.0]\n s.keyState = {}\n s.xy_dir = direction\n s.z_dir = 0\n s.height = 0.5\n s.left_mouse_button = False\n s.mouse_position = [0, 0]\n \n def keyboardFunc(s, key, x, y):\n s.keyState[key.lower()] = True\n def keyboardUpFunc(s, key, x, y):\n s.keyState[key.lower()] = False\n def mouseFunc(s, button, state, x, y):\n if state is 0:\n s.left_mouse_button = True\n s.mouse_position = [x, y]\n else:\n s.left_mouse_button = False\n def motionFunc(s, x, y):\n if s.left_mouse_button:\n s.xy_dir += float(s.mouse_position[0] - x)/200.0\n s.z_dir += float(s.mouse_position[1] - y)/200.0\n \n s.mouse_position = [x, y]\n \n s.xy_dir = s.xy_dir%(pi*2)\n s.z_dir = min(pi/2, max(s.z_dir, -pi/2))\n \n def update(s):\n def keyDown(key):\n return key in s.keyState and s.keyState[key]\n movement = 1.0/100.0\n if keyDown('w'):\n s.pos[0] += cos(s.xy_dir)*movement\n s.pos[1] += sin(s.xy_dir)*movement\n if keyDown('s'):\n s.pos[0] -= cos(s.xy_dir)*movement\n s.pos[1] -= sin(s.xy_dir)*movement\n if keyDown('a'):\n s.pos[0] += cos(s.xy_dir + pi/2)*movement\n s.pos[1] += sin(s.xy_dir + pi/2)*movement\n if keyDown('d'):\n s.pos[0] -= cos(s.xy_dir + pi/2)*movement\n s.pos[1] -= sin(s.xy_dir + pi/2)*movement\n \n if s.parent_node is not None:\n next_node = None\n from_dir = to_dir = None\n w = float(s.parent_node.width) /2.0\n h = float(s.parent_node.height)/2.0\n conditions = {DR_LEFT: s.pos[0] < -w, DR_RIGHT: s.pos[0] > w, DR_TOP: s.pos[1] < -h, DR_BOTTOM: s.pos[1] > h}\n for dir in [DR_LEFT, DR_RIGHT, DR_TOP, DR_BOTTOM]:\n if dir in s.parent_node.doors and conditions[dir]:\n next_pair = s.parent_node.doors[dir]\n next_node = next_pair[0]\n to_dir = next_pair[1]\n from_dir = dir\n \n if next_node is not None:\n print(\"from_dir: {} to_dir: {}\".format(from_dir, to_dir))\n mag = (s.pos[0]**2 + s.pos[1]**2)**0.5\n dir = arctan2(s.pos[0], s.pos[1]) #this switches x and y apparently\n \n print('{} {}'.format(s.pos[0], s.pos[1]))\n #print('m = {}'.format([mag*cos(dir - from_dir*pi/2.0), mag*sin(dir - from_dir*pi/2.0) - 0]))\n s.pos[0:2] = [mag*sin(dir - from_dir*pi/2.0), mag*cos(dir - from_dir*pi/2.0) - 0]#3.0]\n print('{} {}'.format(s.pos[0], s.pos[1]))\n s.pos[1]+=3.0\n print('{} {}'.format(s.pos[0], s.pos[1]))\n \n mag = (s.pos[0]**2 + s.pos[1]**2)**0.5\n dir = arctan2(s.pos[0], s.pos[1])\n \n s.pos[0:2] = [mag*sin(dir + (to_dir-2)*pi/2.0), mag*cos(dir + (to_dir-2)*pi/2.0)]\n print('{} {}'.format(s.pos[0], s.pos[1]))\n \n s.xy_dir += (from_dir - to_dir + 2)*pi/2.0\n s.parent_node = next_node\n \n def displayFunc(s):\n zcomp = cos(s.z_dir)\n #print(\" {} {} {}\".format(cos(s.xy_dir)*zcomp, sin(s.xy_dir)*zcomp, cos(s.z_dir)))\n gluLookAt(0, 0, s.height, cos(s.xy_dir)*zcomp, sin(s.xy_dir)*zcomp, sin(s.z_dir) + s.height, 0,0,1)\n glTranslate(-s.pos[0], -s.pos[1], -s.pos[2])\n \n s.parent_node.displayFunc()","sub_path":"player.py","file_name":"player.py","file_ext":"py","file_size_in_byte":3968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"51386023","text":"# encoding= utf-8\n# Author: HHB\n# Data: 2022/11/09 10:54\n\n\nimport re\nfrom lxml import etree\nimport os\nimport json\nimport datetime\nimport uuid\nimport cv2\nimport time\nimport requests\nimport random\nfrom selenium.webdriver import ActionChains\nfrom selenium import webdriver\nfrom selenium.webdriver.chrome.service import Service\n\n\ndef canny(filepath, cell=7):\n img = cv2.imread(filepath, 0)\n blurred = cv2.GaussianBlur(img, (cell, cell), 0)\n return cv2.Canny(blurred, 240, 250)\n\n\ndef getPosition(img_file1, img_file2):\n img = canny(img_file1)\n img2 = img.copy()\n template = canny(img_file2, cell=5)\n w, h = template.shape[::-1]\n img = img2.copy()\n method = eval(\"cv2.TM_CCOEFF_NORMED\")\n\n res = cv2.matchTemplate(img, template, method)\n min_val, max_val, min_loc, max_loc = cv2.minMaxLoc(res)\n\n if method in [cv2.TM_SQDIFF, cv2.TM_SQDIFF_NORMED]:\n top_left = min_loc\n else:\n top_left = max_loc\n bottom_right = (top_left[0] + w, top_left[1] + h)\n\n cv2.rectangle(img, top_left, bottom_right, 255, 2)\n return top_left\n\n\ndef get_track(distance):\n v = 0\n t = 0.4\n tracks = []\n current = 0\n mid = distance * 7 / 8\n distance += 5\n while current < distance:\n if current < mid:\n a = random.randint(2, 4) # 加速运动\n else:\n a = -random.randint(1, 3) # 减速运动\n v0 = v\n s = v0 * t + 0.6 * a * (t ** 2)\n current += s\n tracks.append(round(s))\n v = v0 + a * t\n random.shuffle(tracks)\n return tracks\n\n\ndef checkCode(b, img_file1, img_file2):\n scale = 1.7\n try:\n while 1:\n t = b.find_element_by_xpath('//*[@id=\"captcha-verify-image\"]')\n t = t.get_attribute(\"src\")\n img = requests.get(t)\n f = open(img_file1, \"wb\")\n f.write(img.content)\n f.close()\n t = b.find_element_by_xpath('//*[@id=\"captcha_container\"]/div/div[2]/img[2]').get_attribute(\"src\")\n img = requests.get(t)\n f = open(img_file2, \"wb\")\n f.write(img.content)\n f.close()\n p = int(getPosition(img_file1, img_file2)[0] / scale)\n # print(p)\n button = b.find_element_by_xpath('//*[@id=\"secsdk-captcha-drag-wrapper\"]/div[2]')\n tracks = get_track(p)\n ActionChains(b).click_and_hold(button).perform()\n for x in tracks:\n ActionChains(b).move_by_offset(xoffset=x, yoffset=0).perform()\n ActionChains(b).release(button).perform()\n time.sleep(1)\n except:\n print(\"ok\")\n\n# 此脚本可以跑单个页面 也可以作为跑多个页面的线程\n# 跑单个时 修改下方要爬的url直接运行此文件\n# 作为线程时 调用get_comment(url),并传入参数即可\n\n \ndef get_review_number(b):\n # //*[@id=\"root\"]/div/div[2]/div/div/div[1]/div[1]/div[3]/div/div[2]/div[1]/div[2]/span\n # num=b.find_element_by_xpath('//*[@id=\"root\"]/div/div[2]/div/div/div[1]/div[1]/div[3]/div/div[2]/div[1]/div[2]/span').text\n # print(int(num))\n for x in range(1, 15, 5):\n time.sleep(1)\n j = x * 12\n js = 'document.documentElement.scrollTop=document.documentElement.scrollHeight* %f' % j\n b.execute_script(js)\n\n\ndef get_comment(url, id):\n # chrome_d = \"C:\\Program Files (x86)\\Google\\Chrome\\Application\\chromedriver.exe\"\n chrome_d = r\"./chromedriver.exe\"\n option = webdriver.ChromeOptions()\n option.add_argument('headless') # 添加无头模式\n b = webdriver.Chrome(executable_path=chrome_d, options=option)\n b.get(url)\n b.maximize_window()\n time.sleep(2)\n img1 = str(uuid.uuid1()) + '.jpeg'\n img2 = str(uuid.uuid1()) + '.png'\n checkCode(b, img1, img2) # 过验证码\n if os.path.exists(img1):\n os.remove(img1)\n if os.path.exists(img2):\n os.remove(img2)\n time.sleep(2)\n get_review_number(b)\n b.implicitly_wait(3)\n # Review_list = b.find_elements_by_xpath('//*[@id=\"root\"]/div/div[2]/div/div/div[1]/div[3]/div/div/div[4]/div/div')\n Review_list = b.find_element_by_xpath('//*[@id=\"root\"]/div/div[2]/div/div/div[1]/div[3]/div/div').get_attribute(\n \"outerHTML\")\n b.close()\n html = etree.HTML(Review_list)\n Review_list = html.xpath('//div[4]/div/div[@class=\"qolG5qEO\"]')\n\n review_infos = [] # 内容列表\n for i in Review_list:\n # print('1',i)\n review_html = etree.HTML(etree.tostring(i).decode())\n # print(review_html)\n review = review_html.xpath('//span[@class=\"mzZanXbP\"]/span/span/span[1]/span/text()') # 用户名和评论内容\n # print(review)\n try:\n if len(review) == 1:\n # print(1111)\n review.append('[表情]')\n if len(review[2]) != 0:\n review[1] = review[1] + review[2]\n except:\n pass\n result_like = review_html.xpath('//div[2]/div[2]/div/p/span/text()') # 点赞数\n content = etree.tostring(i).decode()\n result_time = re.findall(r'

    (.*?)

    ', content) # 评论时间\n result = re.findall(r'a href=\"//(.*?)\" class=\"yqT9PfJg\"', content) # 用户主页地址\n if len(result_time) == 0:\n result_time[0] = 0\n # print(review[0], ':', review[1])\n # print(result_time[0],' ',result[0])\n review_info = {\"用户名\": review[0], \"评论内容\": review[1], \"评论时间\": result_time[0], '点赞数': result_like[0],\n \"用户主页链接\": result[0]}\n review_infos.append(review_info)\n this = os.getcwd() # 获取当前路径\n this = this + \"\\\\tiktok_review_info\"\n ti = 'review_info%s-%s.txt' % (str(datetime.datetime.now().date()), id) # 获取时间拼接字符串作为文件名\n path = os.path.join(this, ti) # 吧两段拼成文件存储路径\n fp = open(path, 'w', encoding='utf-8')\n fp.write('[\\n')\n for i in review_infos:\n print(i)\n data = json.dumps(i, ensure_ascii=False)\n fp.write(data + ',\\n')\n fp.write(']')\n fp.close()\n\n\nif __name__ == '__main__':\n url = 'https://www.douyin.com/search/%E8%8B%8F%E5%B7%9E%E6%96%B0%E9%97%BB?aid=322ae664-96f1-430c-8161-86e1bd61ec67&publish_time=0&sort_type=0&source=normal_search&type=video'\n get_comment(url, id)","sub_path":"PoliceProject/tikTok/tilTok_test.py","file_name":"tilTok_test.py","file_ext":"py","file_size_in_byte":6327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"70402061","text":"\"\"\"\nprob: A string string of lowercase letters is given. We want to partition this string into as many parts as possible so that each letter appears in at most one part, and return a list of integers representing the size of these parts.\n\nFor example:\nInput: string = \"ababfeefhijkh\"\nOutput: [4,4,5]\n\nExplanation:\nThe partition is \"abab\", \"feef\", \"hijkh\". This is a partition so that each letter appears in at most one part.\n\nIdea: We need an array last[char] -> index of S where char occurs last. Then, let anchor and j be the start and end of the current partition. If we are at a label that occurs last at some index after j, we'll extend the partition j = last[c]. If we are at the end of the partition (i == j) then we'll append a partition size to our answer, and set the start of our new partition to i+1\nComp:\n\"\"\"\n\n\ndef partition_labels(string):\n last = {c: i for i, c in enumerate(string)}\n right, left, res = 0, 0, []\n for i, c in enumerate(string):\n right = max(right, last[c])\n if i == right:\n res.append(i - left + 1)\n left = i + 1\n return res\n\n\nstring = \"ababfeefhijkh\"\nexpected = [4, 4, 5]\nactual = partition_labels(string)\nprint(expected == actual)\n\nstring = \"ababcbacadefegdehijhklij\"\nexpected = [9, 7, 8]\nactual = partition_labels(string)\nprint(expected == actual)\n","sub_path":"other/prgcrk/array_string/pointers/partition_labels.py","file_name":"partition_labels.py","file_ext":"py","file_size_in_byte":1331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"579714111","text":"\"\"\"\n q6# Binary-Search-3\n\n## Problem1\nPow(x,n) (https://leetcode.com/problems/powx-n/)\n\nImplement pow(x, n), which calculates x raised to the power n (xn).\n\nExample 1:\n\nInput: 2.00000, 10\nOutput: 1024.00000\nExample 2:\n\nInput: 2.10000, 3\nOutput: 9.26100\nExample 3:\n\nInput: 2.00000, -2\nOutput: 0.25000\nExplanation: 2-2 = 1/22 = 1/4 = 0.25\nNote:\n\n-100.0 < x < 100.0\nn is a 32-bit signed integer, within the range [−231, 231 − 1]\n\n\nTIME - 0(LOG N)\nSPACE - O(LOG N)\n\"\"\"\n\n\nclass Solution:\n def myPow(self, x: float, n: int) -> float:\n if n == 0: # base case\n return 1\n\n y = self.myPow(x, int(n / 2)) #recursive case\n\n if n % 2 == 0: # n is even then 2 recursive calls\n return y * y\n else: # n is odd then 2 recursive calls and product with if n >0 or with 1/x if n <0\n if n > 0:\n return y * y * x\n else:\n return y * y * (1 / x)\n\n\n\n\n\n","sub_path":"powerx-n.py","file_name":"powerx-n.py","file_ext":"py","file_size_in_byte":981,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"566211311","text":"#python\nimport datetime\nimport os\n\n#libs\nfrom django.utils import timezone\nfrom django.test import TestCase, LiveServerTestCase\nfrom selenium import webdriver\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.common.by import By\n\n#local\nfrom .models import Category, Album\n\n\nclass AlbumViewsTestCase(TestCase):\n def test_index(self):\n Category.objects.create(\n name='Blues brothers',\n slug='Blues brothers',\n views=10,\n likes=20\n )\n Category.objects.create(\n name='My new category',\n slug='My new category',\n views=20,\n likes=20\n )\n\n resp = self.client.get('/albums/')\n self.assertEqual(resp.status_code, 200)\n self.assertTrue('categories' in resp.context)\n self.assertEqual([category.slug for category in\n resp.context['categories']],\n ['My new category', 'Blues brothers'])\n\n\nclass AlbumViewsTest(TestCase):\n fixtures = ['albums_views_testdata.json']\n\n def test_index(self):\n resp = self.client.get('/albums/')\n self.assertEqual(resp.status_code, 200)\n self.assertTrue('categories' in resp.context)\n self.assertEqual([category.slug for category in\n resp.context['categories']],\n ['rock', 'blues', 'pop', 'test'])\n\n category_blues = resp.context['categories'][1]\n self.assertEqual(category_blues.name, 'Blues')\n self.assertEqual(category_blues.slug, 'blues')\n self.assertEqual(category_blues.views, 0)\n self.assertEqual(category_blues.likes, 0)\n\n category_pop = resp.context['categories'][2]\n self.assertEqual(category_pop.name, 'Pop')\n self.assertEqual(category_pop.slug, 'pop')\n self.assertEqual(category_pop.views, 0)\n self.assertEqual(category_pop.likes, 0)\n\n category_rock = resp.context['categories'][0]\n self.assertEqual(category_rock.name, 'Rock')\n self.assertEqual(category_rock.slug, 'rock')\n self.assertEqual(category_rock.views, 50)\n self.assertEqual(category_rock.likes, 0)\n\n def test_albums(self):\n resp = self.client.get('/albums/category/blues/')\n self.assertEqual(resp.status_code, 200)\n self.assertTrue('albums' in resp.context)\n self.assertEqual([album.slug for album in resp.context['albums']],\n ['myalbum', 'stevealbum'])\n\n album_1 = resp.context['albums'][0]\n self.assertEqual(album_1.name, 'Myalbum')\n self.assertEqual(album_1.slug, 'myalbum')\n self.assertEqual(album_1.publication_date,\n datetime.date(year=2013, month=11, day=20))\n self.assertEqual(album_1.views, 123)\n self.assertEqual(album_1.likes, 0)\n\n album_2 = resp.context['albums'][1]\n self.assertEqual(album_2.name, 'SteveAlbum')\n self.assertEqual(album_2.slug, 'stevealbum')\n self.assertEqual(album_2.publication_date,\n datetime.date(year=2013, month=11, day=20))\n self.assertEqual(album_2.views, 0)\n self.assertEqual(album_2.likes, 0)\n\n\nclass AlbumCreate(TestCase):\n fixtures = ['albums_views_testdata.json']\n\n def create_category(self, name=\"test category\", slug=\"testcategory\"):\n return Album.objects.create(name=name,\n slug=slug,\n )\n\n def test_category_creation(self):\n cat = self.create_category()\n self.assertTrue(isinstance(cat, Category))\n self.assertEqual(cat.__unicode__(), cat.name)\n\n def create_album(self, name=\"test album\", slug=\"blah\"):\n return Album.objects.create(name=name,\n slug=slug,\n publication_date=timezone.now(),\n views=0,\n likes=0,\n )\n\n def test_album_creation(self):\n albums = self.create_album()\n self.assertTrue(isinstance(albums, Album))\n self.assertEqual(albums.__unicode__(), albums.name)\n\n\nclass CategoryTest(LiveServerTestCase):\n fixtures = ['admin_user.json']\n\n def setUp(self):\n self.browser = webdriver.Firefox()\n self.browser.implicitly_wait(3)\n\n def tearDown(self):\n self.browser.quit()\n\n def test_for_create_new_category_via_user(self):\n self.browser.get(self.live_server_url + '/albums/')\n\n login_user_link = self.browser.find_element_by_link_text('Login')\n login_user_link.click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Login', body.text)\n\n username_field = self.browser.find_element_by_name('username')\n username_field.send_keys('Squalles')\n\n password_field = self.browser.find_element_by_name('password')\n password_field.send_keys('sq321')\n password_field.send_keys(Keys.RETURN)\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('MyAlbum', body.text)\n\n edit_link = self.browser.find_element(By.CSS_SELECTOR,\n \".dropdown-toggle\")\n edit_link.click()\n edit_link = self.browser.find_element(By.ID, value=\"Create_Category\")\n edit_link.click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Create Category', body.text)\n\n name_field = self.browser.find_element_by_name('name')\n name_field.send_keys(\"My Test Category\")\n\n create_button = self.browser.find_element_by_css_selector(\n \"button[value='Save']\")\n create_button.click()\n\n\nclass CategoryModelTest(TestCase):\n def test_creating_a_new_category(self):\n category = Category()\n category.name = \"TestCategory\"\n category.content = \"Example content test\"\n category.slug = \"testcategory\"\n\n category.save()\n\n all_category_in_database = Category.objects.all()\n self.assertEqual(len(all_category_in_database), 1)\n only_category_in_database = all_category_in_database[0]\n self.assertEqual(only_category_in_database, category)\n\n self.assertEqual(only_category_in_database.name, \"TestCategory\")\n self.assertEqual(only_category_in_database.content,\n \"Example content test\")\n self.assertEqual(only_category_in_database.slug, \"testcategory\")\n\n\nclass ArtistTest(LiveServerTestCase):\n fixtures = ['admin_user.json', 'all_data.json']\n\n def setUp(self):\n self.browser = webdriver.Firefox()\n self.browser.implicitly_wait(3)\n\n def tearDown(self):\n self.browser.quit()\n\n def test_for_create_new_artist_via_user(self):\n self.browser.get(self.live_server_url + '/albums/')\n\n login_user_link = self.browser.find_element_by_link_text('Login')\n login_user_link.click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Login', body.text)\n\n username_field = self.browser.find_element_by_name('username')\n username_field.send_keys('Squalles')\n\n password_field = self.browser.find_element_by_name('password')\n password_field.send_keys('sq321')\n password_field.send_keys(Keys.RETURN)\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Genres', body.text)\n\n genres = self.browser.find_element_by_link_text('Genres')\n genres.click()\n choose_category = self.browser.find_element_by_link_text('Blues')\n choose_category.click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Edit', body.text)\n\n create_artist = self.browser.find_element_by_link_text('Edit')\n create_artist.click()\n create_artist = self.browser.find_elements_by_link_text(\n 'Create New Artist')\n create_artist[0].click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('Create Artist', body.text)\n\n self.browser.find_element_by_name('first_name').send_keys(\"John\")\n self.browser.find_element_by_name('last_name').send_keys(\"Doe\")\n self.browser.find_element_by_name('born').send_keys(\"1942-02-03\")\n self.browser.find_element_by_name('known_as').send_keys(\"Test\")\n self.browser.find_element_by_name('start_date').send_keys(\n \"1972-02-01\")\n self.browser.find_element_by_name(\"picture\").send_keys(\n os.getcwd() + \"/image.png/\")\n self.browser.find_element_by_class_name('forminput').send_keys(\"TestTag\")\n\n self.browser.find_element_by_css_selector(\n \"button[value='Add']\").click()\n\n body = self.browser.find_element_by_tag_name('body')\n self.assertIn('John Doe', body.text)\n\n\n\n\n","sub_path":"albums/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":8866,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"614551800","text":"# -*- coding: utf-8 -*-\n'''\nAnalyzer Object\n===============\n'''\n\nfrom __future__ import annotations\n\n__all__ = ('Analyzer',)\n\nimport os\nfrom analyzer.core.frequency_analyzer import FrequencyAnalyzer\nfrom analyzer.core.percent_analyzer import PercentAnalyzer\nfrom analyzer.core.tokenizer import Tokenizer\nfrom analyzer.core.word_analyzer import WordAnalyzer\nfrom analyzer.datatypes.analyzerexception import AnalyzerError\nfrom analyzer.datatypes.tokenlist import TokenList\nfrom builder.core.executer import Executer\nfrom builder.core.outputter import Outputter\nfrom builder.datatypes.outputmode import OutputMode\nfrom builder.datatypes.rawdata import RawData\nfrom builder.datatypes.resultdata import ResultData\nfrom builder.datatypes.textlist import TextList\nfrom builder.utils import assertion\nfrom builder.utils.util_file import get_content_from_text_file\nfrom builder.utils.logger import MyLogger\n\n\n# logger\nLOG = MyLogger.get_logger(__name__)\nLOG.set_file_handler()\n\n\nclass Analyzer(Executer):\n ''' Analyzer object.\n '''\n def __init__(self):\n super().__init__()\n LOG.info('ANALYZER: initialize')\n\n def execute(self, src: (str, list, TextList),\n person_names: list,\n is_debug: bool=False) -> ResultData: # pragma: no cover\n LOG.info('ANALYZER: start exec')\n is_succeeded = True\n error = None\n basesrc = None\n result = ResultData([], is_succeeded, error)\n\n if isinstance(src, str):\n basesrc = TextList(*get_content_from_text_file(src))\n elif isinstance(src, TextList):\n basesrc = src\n elif isinstance(src, (list, tuple)):\n basesrc = TextList(*src)\n else:\n msg = f'Invalid analyze source!: {src}'\n LOG.critical(msg)\n return ResultData(result, False, AnalyzerError(msg))\n\n tmp = self._rid_tag(basesrc)\n\n LOG.info('TOKENIZER: call')\n result = assertion.is_instance(Tokenizer().execute(tmp, person_names), ResultData)\n if not result.is_succeeded:\n return result\n tokens = assertion.is_instance(result.data, TokenList)\n\n LOG.info('WORD_ANALYZER: call')\n result = assertion.is_instance(WordAnalyzer().execute(tokens), ResultData)\n if not result.is_succeeded:\n return result\n word_data = assertion.is_listlike(result.data)\n\n LOG.info('PERCENT_ANALYZER: call')\n result = assertion.is_instance(PercentAnalyzer().execute(tmp), ResultData)\n if not result.is_succeeded:\n return result\n percent_data = assertion.is_listlike(result.data)\n\n LOG.info('FREQUENCY_ANALYZER: call')\n result = assertion.is_instance(FrequencyAnalyzer().execute(tokens), ResultData)\n if not result.is_succeeded:\n return result\n freq_data = assertion.is_listlike(result.data)\n\n LOG.info('Analyzer result output')\n result_data = percent_data + ['\\n---\\n'] \\\n + word_data + ['\\n---\\n'] \\\n + freq_data\n fname = 'result'\n suffix = ''\n extention = 'md'\n builddir = 'build/results'\n mode = OutputMode.CONSOLE if is_debug else OutputMode.FILE\n data = TextList(*[f'{line}\\n' for line in result_data])\n Outputter().execute(data, mode, fname, suffix, extention, builddir)\n return result\n\n #\n # private\n #\n\n def _rid_tag(self, src: TextList) -> TextList:\n LOG.info('ANALYZER: rid tags start')\n tmp = []\n for line in assertion.is_instance(src, TextList).data:\n assertion.is_str(line)\n if line.startswith('#') or line.startswith('\\n#'):\n continue\n elif line.startswith('---') or line.startswith('\\n---'):\n continue\n elif line in ('\\n', '\\n\\n'):\n continue\n else:\n tmp.append(line)\n return TextList(*tmp)\n","sub_path":"analyzer/analyzer.py","file_name":"analyzer.py","file_ext":"py","file_size_in_byte":3945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"156510914","text":"\"\"\"\n\n metadata.py\n\n\n Lukas Puehringer \n Santiago Torres \n\n\n Oct 23, 2017\n\n\n See LICENSE for licensing information.\n\n\n Provides a container class `Metablock` for signed metadata and\n functions for signing, signature verification, de-serialization and\n serialization from and to JSON.\n\n\"\"\"\n\nimport attr\nimport json\nimport canonicaljson\n\nimport securesystemslib.keys\nimport securesystemslib.formats\nimport securesystemslib.exceptions\n\nfrom in_toto.models.link import Link\nfrom in_toto.models.layout import Layout\nfrom in_toto.exceptions import SignatureVerificationError\n\n@attr.s(repr=False, init=False)\nclass Metablock(object):\n \"\"\" This object holds the in-toto metablock data structure. This includes\n the fields \"signed\" and \"signatures\", i.e., what was signed and the\n signatures. \"\"\"\n signatures = attr.ib()\n signed = attr.ib()\n\n\n def __init__(self, **kwargs):\n self.signatures = kwargs.get(\"signatures\", [])\n self.signed = kwargs.get(\"signed\")\n\n\n def __repr__(self):\n \"\"\"Returns a JSON string representation of the object.\"\"\"\n # the double {{'s is the escape sequence for an individual {. We wrap this\n # under a format string to avoid encoding to json twice (which turns a json\n # string into a string and so on...\n # FIXME:\n # We are mixing 3 JSON string formats here: The value of \"signed\" is\n # \"pretty printed canonical json\", the value of \"signatures\" is\n # \"canonical json\" and the container is just \"json\".\n # Is this really what we want?\n return '{{\"signed\": {}, \"signatures\": {}}}'.format(self.signed,\n canonicaljson.encode_canonical_json(self.signatures))\n\n\n def dump(self, filename):\n \"\"\"\n \n Write the JSON string representation of the Metablock object\n to disk.\n\n \n filename:\n The path to write the file to.\n\n \n Writing metadata file to disk\n\n \n None.\n\n \"\"\"\n with open(filename, \"wt\") as fp:\n fp.write(\"{}\".format(self))\n\n\n @staticmethod\n def load(path):\n \"\"\"\n \n Loads the JSON string representation of signed metadata from disk\n and creates a Metablock object.\n The `signed` attribute of the Metablock object is assigned a Link\n or Layout object, depending on the `_type` field in the loaded\n metadata file.\n\n \n path:\n The path to write the file to.\n\n \n Reading metadata file from disk\n\n \n None.\n\n \"\"\"\n\n with open(path, \"r\") as fp:\n data = json.load(fp)\n\n signatures = data.get(\"signatures\", [])\n signed_data = data.get(\"signed\", {})\n signed_type = signed_data.get(\"_type\")\n\n if signed_type == \"link\":\n signed = Link.read(signed_data)\n\n elif signed_type == \"layout\":\n signed = Layout.read(signed_data)\n\n else:\n raise securesystemslib.exceptions.FormatError(\"Invalid Metadata format\")\n\n return Metablock(signatures=signatures, signed=signed)\n\n\n @property\n def _type(self):\n \"\"\" Shortcut to the _type property of the contained Link or Layout object,\n should be one of \"link\" or \"layout\". \"\"\"\n return self.signed._type\n\n\n def sign(self, key):\n \"\"\"\n \n Signs the pretty printed canonical JSON representation\n (see models.common.Signable.__repr__) of the Link or Layout object\n contained in the `signed` property with the passed key and appends the\n created signature to `signatures`.\n\n \n key:\n A signing key in the format securesystemslib.formats.KEY_SCHEMA\n\n \n None.\n\n \"\"\"\n securesystemslib.formats.KEY_SCHEMA.check_match(key)\n\n signature = securesystemslib.keys.create_signature(key, repr(self.signed))\n self.signatures.append(signature)\n\n\n def verify_signatures(self, keys_dict):\n \"\"\"\n \n Verifies all signatures found in the `signatures` property using the keys\n from the passed dictionary of keys and the pretty printed canonical JSON\n representation (see models.common.Signable.__repr__) of the Link or\n Layout object contained in `signed`.\n\n Verification fails if,\n - the passed keys don't have the right format,\n - the object is not signed,\n - there is a signature for which no key was passed,\n - if any of the signatures is actually broken.\n\n Note:\n This will be revised with in-toto/in-toto#135\n\n \n keys_dict:\n Verifying keys in the format:\n securesystemslib.formats.KEYDICT_SCHEMA\n\n \n FormatError\n if the passed key dictionary is not conformant with\n securesystemslib.formats.KEYDICT_SCHEMA\n\n SignatureVerificationError\n if the Metablock is not signed\n\n if the Metablock carries a signature for which no key is found in\n the passed key dictionary, which means that multiple signatures\n have to be verified at once\n\n if any of the verified signatures is actually broken\n\n \n None.\n\n \"\"\"\n securesystemslib.formats.KEYDICT_SCHEMA.check_match(keys_dict)\n\n if not self.signatures or len(self.signatures) <= 0:\n raise SignatureVerificationError(\"No signatures found\")\n\n for signature in self.signatures:\n keyid = signature[\"keyid\"]\n try:\n key = keys_dict[keyid]\n except KeyError:\n raise SignatureVerificationError(\n \"Signature key not found, key id is '{0}'\".format(keyid))\n if not securesystemslib.keys.verify_signature(\n key, signature, repr(self.signed)):\n raise SignatureVerificationError(\"Invalid signature\")","sub_path":"in_toto/models/metadata.py","file_name":"metadata.py","file_ext":"py","file_size_in_byte":5779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"160360289","text":"from . import stock_prediction\r\nfrom flask import request, jsonify\r\nfrom webapp.models import *\r\nfrom library.datetime_function import get_offset_date\r\nfrom library.sql_function import sql_to_object,object_to_sql\r\n\r\n\r\n@stock_prediction.route('api_comprehensive_analysis', methods=['GET'])\r\ndef api_comprehensive_analysis():\r\n code = request.args.get('code')\r\n date = request.args.get('date')\r\n bar_data = Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code == code, Stock_Daily_Bar.trade_date <= date,\r\n Stock_Daily_Bar.trade_date >= get_offset_date(date, -730)).order_by(\r\n Stock_Daily_Bar.trade_date.asc()).all()\r\n basic_data = Stock_Daily_Basic.query.filter_by(ts_code=code, trade_date=date).first()\r\n # 要判断上面两个查询结果非空\r\n today_bar_data = {'open': bar_data[-1].open, 'close': bar_data[-1].close, 'high': bar_data[-1].high,\r\n 'low': bar_data[-1].low, 'change': bar_data[-1].change, 'pct_chg': bar_data[-1].pct_chg,\r\n 'pre_close': bar_data[-1].pre_close, 'vol': bar_data[-1].vol, 'amount': bar_data[-1].amount}\r\n today_basic_data = {'turnover_rate': basic_data.turnover_rate, 'pe': basic_data.pe, 'pb': basic_data.pb,\r\n 'circ_mv': basic_data.circ_mv}\r\n short_trend_data = []\r\n k_line_data = []\r\n one_year_data = []\r\n for i in bar_data[-60:]:\r\n short_trend_data.append(i.close)\r\n for i in bar_data[-250:]:\r\n one_year_data.append(i.close)\r\n for i in bar_data:\r\n trade_date = i.trade_date\r\n k_line_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n one_year_min = min(one_year_data)\r\n one_year_max = max(one_year_data)\r\n concepts = []\r\n company_data = Stock_Company_Extend.query.join(Stock_Company,\r\n Stock_Company_Extend.ts_code == Stock_Company.ts_code).add_columns(\r\n Stock_Company.province, Stock_Company.city).join(Stock_Industry_SW_3,\r\n Stock_Company_Extend.industry_sw_code == Stock_Industry_SW_3.industry_sw_3_code).add_columns(\r\n Stock_Industry_SW_3.industry_sw_3_name).join(\r\n Stock_Industry_SW_2, Stock_Industry_SW_3.belong_to == Stock_Industry_SW_2.industry_sw_2_code).add_columns(\r\n Stock_Industry_SW_2.industry_sw_2_name).join(\r\n Stock_Industry_SW_1, Stock_Industry_SW_2.belong_to == Stock_Industry_SW_1.industry_sw_1_code).add_columns(\r\n Stock_Industry_SW_1.industry_sw_1_name).filter(\r\n Stock_Company_Extend.ts_code == code).first()\r\n industry = [company_data.industry_sw_1_name, company_data.industry_sw_2_name, company_data.industry_sw_3_name]\r\n area = [company_data.province, company_data.city]\r\n concept_data = Stock_Concept_Detail.query.join(Stock_Concept_List,\r\n Stock_Concept_Detail.concept_code == Stock_Concept_List.concept_code).add_columns(\r\n Stock_Concept_List.concept_name).filter(Stock_Concept_Detail.ts_code == code,\r\n Stock_Concept_List.src == 'wind').all()\r\n for i in concept_data:\r\n concepts.append(i.concept_name)\r\n return jsonify(\r\n {'today_bar_data': today_bar_data, 'today_basic_data': today_basic_data, 'short_trend_data': short_trend_data,\r\n 'k_line_data': k_line_data, 'one_year_data': {'min': one_year_min, 'max': one_year_max},\r\n 'company_and_concept': {'area': area, 'industry': industry, 'concept': concepts}})\r\n\r\n\r\n@stock_prediction.route('api_trend_prediction', methods=['GET'])\r\ndef api_trend_prediction():\r\n code = request.args.get('code')\r\n date = request.args.get('date')\r\n bar_data = Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code == code, Stock_Daily_Bar.trade_date <= date,\r\n Stock_Daily_Bar.trade_date >= get_offset_date(date, -120)).order_by(\r\n Stock_Daily_Bar.trade_date.asc()).all()\r\n basic_data = Stock_Daily_Basic.query.filter_by(ts_code=code, trade_date=date).first()\r\n today_bar_data = {'open': bar_data[-1].open, 'close': bar_data[-1].close, 'high': bar_data[-1].high,\r\n 'low': bar_data[-1].low, 'change': bar_data[-1].change, 'pct_chg': bar_data[-1].pct_chg,\r\n 'pre_close': bar_data[-1].pre_close, 'vol': bar_data[-1].vol, 'amount': bar_data[-1].amount}\r\n if basic_data is None:\r\n today_basic_data = {'turnover_rate': 0, 'pe': 0, 'pb': 0,\r\n 'circ_mv': 0}\r\n else:\r\n today_basic_data = {'turnover_rate': basic_data.turnover_rate, 'pe': basic_data.pe, 'pb': basic_data.pb,\r\n 'circ_mv': basic_data.circ_mv}\r\n trade_point_data = Model_Trading_Point.query.filter(Model_Trading_Point.ts_code == code,\r\n Model_Trading_Point.trade_date <= date,\r\n Model_Trading_Point.trade_date >= get_offset_date(date,\r\n -120)).order_by(\r\n Model_Trading_Point.trade_date.asc()).all()\r\n aggressive_requirement = '维持现状'\r\n steady_requirement = '维持现状'\r\n trade_point_line_data = []\r\n aggressive_trade_point = []\r\n steady_trade_point = []\r\n for i in bar_data:\r\n trade_date = i.trade_date\r\n trade_point_line_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in trade_point_data:\r\n trade_date = i.trade_date\r\n if i.aggressive_buy_point is True:\r\n close = (Stock_Daily_Bar.query.filter_by(ts_code=code, trade_date=trade_date).first()).close\r\n aggressive_trade_point.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), close, '买点'])\r\n if trade_date == date:\r\n aggressive_requirement = i.aggressive_buy_requirement\r\n if i.aggressive_sell_point is True:\r\n close = (Stock_Daily_Bar.query.filter_by(ts_code=code, trade_date=trade_date).first()).close\r\n aggressive_trade_point.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), close, '卖点'])\r\n if trade_date == date:\r\n aggressive_requirement = i.aggressive_sell_requirement\r\n if i.steady_buy_point is True:\r\n close = (Stock_Daily_Bar.query.filter_by(ts_code=code, trade_date=trade_date).first()).close\r\n steady_trade_point.append([(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), close, '买点'])\r\n if trade_date == date:\r\n steady_requirement = i.steady_buy_requirement\r\n if i.steady_sell_point is True:\r\n close = (Stock_Daily_Bar.query.filter_by(ts_code=code, trade_date=trade_date).first()).close\r\n steady_trade_point.append([(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), close, '卖点'])\r\n if trade_date == date:\r\n steady_requirement = i.steady_sell_requirement\r\n trend_forecast_result = Model_Trend_Forecast.query.filter_by(ts_code=code, trade_date=date).first()\r\n first_vote = [trend_forecast_result.first_fall_vote, trend_forecast_result.first_maintain_vote,\r\n trend_forecast_result.first_rise_vote]\r\n second_vote = [trend_forecast_result.second_fall_vote, trend_forecast_result.second_maintain_vote,\r\n trend_forecast_result.second_rise_vote]\r\n third_vote = [trend_forecast_result.third_fall_vote, trend_forecast_result.third_maintain_vote,\r\n trend_forecast_result.third_rise_vote]\r\n\r\n similarity_data = []\r\n similarity_1_data = []\r\n similarity_2_data = []\r\n similarity_3_data = []\r\n similarity_match_1_data = []\r\n similarity_match_2_data = []\r\n similarity_match_3_data = []\r\n mark_line = [\r\n (bar_data[-20].trade_date[0:4] + '/' + bar_data[-20].trade_date[4:6] + '/' + bar_data[-20].trade_date[6:8]),\r\n (bar_data[-1].trade_date[0:4] + '/' + bar_data[-1].trade_date[4:6] + '/' + bar_data[-1].trade_date[6:8])]\r\n similarity_short_term_result = Model_Similarity_Short_Term.query.filter_by(ts_code=code, trade_date=date).first()\r\n similarity_1_stock = Stock_Basic.query.filter_by(ts_code=similarity_short_term_result.similarity_1_code).first()\r\n similarity_2_stock = Stock_Basic.query.filter_by(ts_code=similarity_short_term_result.similarity_2_code).first()\r\n similarity_3_stock = Stock_Basic.query.filter_by(ts_code=similarity_short_term_result.similarity_3_code).first()\r\n similarity_1_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == similarity_short_term_result.similarity_1_code,\r\n Stock_Daily_Bar.trade_date <= similarity_short_term_result.similarity_1_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n similarity_2_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == similarity_short_term_result.similarity_2_code,\r\n Stock_Daily_Bar.trade_date <= similarity_short_term_result.similarity_2_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n similarity_3_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == similarity_short_term_result.similarity_3_code,\r\n Stock_Daily_Bar.trade_date <= similarity_short_term_result.similarity_3_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n similarity_1_result.reverse()\r\n similarity_2_result.reverse()\r\n similarity_3_result.reverse()\r\n for i in bar_data[-60:]:\r\n trade_date = i.trade_date\r\n similarity_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in similarity_1_result[-30:]:\r\n factor = bar_data[-1].close / similarity_1_result[-31].close\r\n trade_date = i.trade_date\r\n similarity_1_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n for i in similarity_2_result[-30:]:\r\n factor = bar_data[-1].close / similarity_2_result[-31].close\r\n trade_date = i.trade_date\r\n similarity_2_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n for i in similarity_3_result[-30:]:\r\n factor = bar_data[-1].close / similarity_3_result[-31].close\r\n trade_date = i.trade_date\r\n similarity_3_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n similarity_1_data = similarity_data + similarity_1_data\r\n similarity_2_data = similarity_data + similarity_2_data\r\n similarity_3_data = similarity_data + similarity_3_data\r\n for i in similarity_1_result:\r\n trade_date = i.trade_date\r\n similarity_match_1_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in similarity_2_result:\r\n trade_date = i.trade_date\r\n similarity_match_2_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in similarity_3_result:\r\n trade_date = i.trade_date\r\n similarity_match_3_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n similarity_match_mark_line_1 = [(similarity_short_term_result.similarity_1_matching_start_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_1_matching_start_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_1_matching_start_time[\r\n 6:8]),\r\n (similarity_short_term_result.similarity_1_matching_end_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_1_matching_end_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_1_matching_end_time[\r\n 6:8])]\r\n similarity_match_mark_line_2 = [(similarity_short_term_result.similarity_2_matching_start_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_2_matching_start_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_2_matching_start_time[\r\n 6:8]),\r\n (similarity_short_term_result.similarity_2_matching_end_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_2_matching_end_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_2_matching_end_time[\r\n 6:8])]\r\n similarity_match_mark_line_3 = [(similarity_short_term_result.similarity_3_matching_start_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_3_matching_start_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_3_matching_start_time[\r\n 6:8]),\r\n (similarity_short_term_result.similarity_3_matching_end_time[\r\n 0:4] + '/' + similarity_short_term_result.similarity_3_matching_end_time[\r\n 4:6] + '/' + similarity_short_term_result.similarity_3_matching_end_time[\r\n 6:8])]\r\n sparkline_data = []\r\n similarity_1_sparkline_data = []\r\n similarity_2_sparkline_data = []\r\n similarity_3_sparkline_data = []\r\n for i in bar_data[-20:]:\r\n sparkline_data.append(i.pct_chg)\r\n sparkline_data.extend([None] * 30)\r\n for i in similarity_1_result[-50:]:\r\n similarity_1_sparkline_data.append(i.pct_chg)\r\n for i in similarity_2_result[-50:]:\r\n similarity_2_sparkline_data.append(i.pct_chg)\r\n for i in similarity_3_result[-50:]:\r\n similarity_3_sparkline_data.append(i.pct_chg)\r\n similarity_1_table_data = []\r\n similarity_2_table_data = []\r\n similarity_3_table_data = []\r\n for i in similarity_1_result[-30:]:\r\n similarity_1_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n for i in similarity_2_result[-30:]:\r\n similarity_2_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n for i in similarity_3_result[-30:]:\r\n similarity_3_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n similarity_short_term_data = {'similarity_1_data': similarity_1_data, 'similarity_2_data': similarity_2_data,\r\n 'similarity_3_data': similarity_3_data,\r\n 'similarity_match_1_data': similarity_match_1_data,\r\n 'similarity_match_2_data': similarity_match_2_data,\r\n 'similarity_match_3_data': similarity_match_3_data,\r\n 'mark_line': mark_line, 'similarity_match_mark_line_1': similarity_match_mark_line_1,\r\n 'similarity_match_mark_line_2': similarity_match_mark_line_2,\r\n 'similarity_match_mark_line_3': similarity_match_mark_line_3,\r\n 'sparkline_data': sparkline_data,\r\n 'similarity_1_sparkline_data': similarity_1_sparkline_data,\r\n 'similarity_2_sparkline_data': similarity_2_sparkline_data,\r\n 'similarity_3_sparkline_data': similarity_3_sparkline_data,\r\n 'similarity_1_table_data': similarity_1_table_data,\r\n 'similarity_2_table_data': similarity_2_table_data,\r\n 'similarity_3_table_data': similarity_3_table_data,\r\n 'similarity_stock': [\r\n {'code': similarity_1_stock.ts_code, 'name': similarity_1_stock.name,\r\n 'distance': similarity_short_term_result.similarity_1_distance},\r\n {'code': similarity_2_stock.ts_code, 'name': similarity_2_stock.name,\r\n 'distance': similarity_short_term_result.similarity_2_distance},\r\n {'code': similarity_3_stock.ts_code, 'name': similarity_3_stock.name,\r\n 'distance': similarity_short_term_result.similarity_3_distance}],\r\n 'trend_description': similarity_short_term_result.similarity_1_trend_description}\r\n correlation_data = []\r\n correlation_1_data = []\r\n correlation_2_data = []\r\n correlation_3_data = []\r\n correlation_match_1_data = []\r\n correlation_match_2_data = []\r\n correlation_match_3_data = []\r\n correlation_short_term_result = Model_Correlation_Short_Term.query.filter_by(ts_code=code, trade_date=date).first()\r\n correlation_1_stock = Stock_Basic.query.filter_by(ts_code=correlation_short_term_result.correlation_1_code).first()\r\n correlation_2_stock = Stock_Basic.query.filter_by(ts_code=correlation_short_term_result.correlation_2_code).first()\r\n correlation_3_stock = Stock_Basic.query.filter_by(ts_code=correlation_short_term_result.correlation_3_code).first()\r\n correlation_1_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == correlation_short_term_result.correlation_1_code,\r\n Stock_Daily_Bar.trade_date <= correlation_short_term_result.correlation_1_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n correlation_2_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == correlation_short_term_result.correlation_2_code,\r\n Stock_Daily_Bar.trade_date <= correlation_short_term_result.correlation_2_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n correlation_3_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == correlation_short_term_result.correlation_3_code,\r\n Stock_Daily_Bar.trade_date <= correlation_short_term_result.correlation_3_prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n correlation_1_result.reverse()\r\n correlation_2_result.reverse()\r\n correlation_3_result.reverse()\r\n for i in bar_data[-60:]:\r\n trade_date = i.trade_date\r\n correlation_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in correlation_1_result[-30:]:\r\n factor = bar_data[-1].close / correlation_1_result[-31].close\r\n trade_date = i.trade_date\r\n correlation_1_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n for i in correlation_2_result[-30:]:\r\n factor = bar_data[-1].close / correlation_2_result[-31].close\r\n trade_date = i.trade_date\r\n correlation_2_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n for i in correlation_3_result[-30:]:\r\n factor = bar_data[-1].close / correlation_3_result[-31].close\r\n trade_date = i.trade_date\r\n correlation_3_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open * factor, i.close * factor,\r\n i.low * factor, i.high * factor,\r\n i.amount])\r\n correlation_1_data = correlation_data + correlation_1_data\r\n correlation_2_data = correlation_data + correlation_2_data\r\n correlation_3_data = correlation_data + correlation_3_data\r\n for i in correlation_1_result:\r\n trade_date = i.trade_date\r\n correlation_match_1_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in correlation_2_result:\r\n trade_date = i.trade_date\r\n correlation_match_2_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n for i in correlation_3_result:\r\n trade_date = i.trade_date\r\n correlation_match_3_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n correlation_match_mark_line_1 = [(correlation_short_term_result.correlation_1_matching_start_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_1_matching_start_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_1_matching_start_time[\r\n 6:8]),\r\n (correlation_short_term_result.correlation_1_matching_end_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_1_matching_end_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_1_matching_end_time[\r\n 6:8])]\r\n correlation_match_mark_line_2 = [(correlation_short_term_result.correlation_2_matching_start_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_2_matching_start_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_2_matching_start_time[\r\n 6:8]),\r\n (correlation_short_term_result.correlation_2_matching_end_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_2_matching_end_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_2_matching_end_time[\r\n 6:8])]\r\n correlation_match_mark_line_3 = [(correlation_short_term_result.correlation_3_matching_start_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_3_matching_start_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_3_matching_start_time[\r\n 6:8]),\r\n (correlation_short_term_result.correlation_3_matching_end_time[\r\n 0:4] + '/' + correlation_short_term_result.correlation_3_matching_end_time[\r\n 4:6] + '/' + correlation_short_term_result.correlation_3_matching_end_time[\r\n 6:8])]\r\n correlation_1_sparkline_data = []\r\n correlation_2_sparkline_data = []\r\n correlation_3_sparkline_data = []\r\n for i in correlation_1_result[-50:]:\r\n correlation_1_sparkline_data.append(i.pct_chg)\r\n for i in correlation_2_result[-50:]:\r\n correlation_2_sparkline_data.append(i.pct_chg)\r\n for i in correlation_3_result[-50:]:\r\n correlation_3_sparkline_data.append(i.pct_chg)\r\n correlation_1_table_data = []\r\n correlation_2_table_data = []\r\n correlation_3_table_data = []\r\n for i in correlation_1_result[-30:]:\r\n correlation_1_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n for i in correlation_2_result[-30:]:\r\n correlation_2_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n for i in correlation_3_result[-30:]:\r\n correlation_3_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n correlation_short_term_data = {'correlation_1_data': correlation_1_data, 'correlation_2_data': correlation_2_data,\r\n 'correlation_3_data': correlation_3_data,\r\n 'correlation_match_1_data': correlation_match_1_data,\r\n 'correlation_match_2_data': correlation_match_2_data,\r\n 'correlation_match_3_data': correlation_match_3_data,\r\n 'mark_line': mark_line,\r\n 'correlation_match_mark_line_1': correlation_match_mark_line_1,\r\n 'correlation_match_mark_line_2': correlation_match_mark_line_2,\r\n 'correlation_match_mark_line_3': correlation_match_mark_line_3,\r\n 'sparkline_data': sparkline_data,\r\n 'correlation_1_sparkline_data': correlation_1_sparkline_data,\r\n 'correlation_2_sparkline_data': correlation_2_sparkline_data,\r\n 'correlation_3_sparkline_data': correlation_3_sparkline_data,\r\n 'correlation_1_table_data': correlation_1_table_data,\r\n 'correlation_2_table_data': correlation_2_table_data,\r\n 'correlation_3_table_data': correlation_3_table_data,\r\n 'correlation_stock': [\r\n {'code': correlation_1_stock.ts_code, 'name': correlation_1_stock.name,\r\n 'r': correlation_short_term_result.correlation_1_r},\r\n {'code': correlation_2_stock.ts_code, 'name': correlation_2_stock.name,\r\n 'r': correlation_short_term_result.correlation_2_r},\r\n {'code': correlation_3_stock.ts_code, 'name': correlation_3_stock.name,\r\n 'r': correlation_short_term_result.correlation_3_r}],\r\n 'trend_description': correlation_short_term_result.correlation_1_trend_description}\r\n\r\n similarity_history_result = Model_Similarity_History.query.filter_by(ts_code=code, trade_date=date).first()\r\n similarity_history_match_result = Stock_Daily_Bar.query.filter(\r\n Stock_Daily_Bar.ts_code == code,\r\n Stock_Daily_Bar.trade_date <= similarity_history_result.prediction_end_time, ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(90).all()\r\n similarity_history_match_result.reverse()\r\n similarity_history_match_data = []\r\n similarity_history_match_sparkline_data = []\r\n for i in similarity_history_match_result:\r\n trade_date = i.trade_date\r\n similarity_history_match_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount])\r\n similarity_history_match_sparkline_data.append(i.pct_chg)\r\n similarity_history_mark_line = [(similarity_history_result.matching_start_time[\r\n 0:4] + '/' + similarity_history_result.matching_start_time[\r\n 4:6] + '/' + similarity_history_result.matching_start_time[6:8]),\r\n (similarity_history_result.matching_end_time[\r\n 0:4] + '/' + similarity_history_result.matching_end_time[\r\n 4:6] + '/' + similarity_history_result.matching_end_time[6:8])]\r\n similarity_history_match_date = [(date[0:4] + '/' + date[4:6] + '/' + date[6:8]),\r\n (similarity_history_result.matching_end_time[\r\n 0:4] + '/' + similarity_history_result.matching_end_time[\r\n 4:6] + '/' + similarity_history_result.matching_end_time[6:8])]\r\n similarity_history_sparkline_data = []\r\n for i in bar_data[-60:]:\r\n similarity_history_sparkline_data.append(i.pct_chg)\r\n similarity_history_sparkline_data.extend([None] * 30)\r\n similarity_history_table_data = []\r\n for i in similarity_history_match_result[-30:]:\r\n similarity_history_table_data.append([i.pct_chg, i.open, i.close, i.high, i.low, i.vol])\r\n similarity_history_data = {'similarity_history_match_data': similarity_history_match_data,\r\n 'mark_line': similarity_history_mark_line, 'match_date': similarity_history_match_date,\r\n 'similarity_history_match_sparkline_data': similarity_history_match_sparkline_data,\r\n 'sparkline_data': similarity_history_sparkline_data,\r\n 'similarity_history_table_data': similarity_history_table_data}\r\n\r\n bar_data_for_state_transition_result = Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code == code,\r\n Stock_Daily_Bar.trade_date <= date,\r\n ).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(200).all()\r\n state_transition_result = Model_State_Transition.query.filter(Model_State_Transition.ts_code == code,\r\n Model_State_Transition.trade_date <= date).order_by(\r\n Model_State_Transition.trade_date.desc()).limit(200).all()\r\n bar_data_for_state_transition_result.reverse()\r\n state_transition_result.reverse()\r\n today_state_transition_data = {'s_rise_rate': state_transition_result[-1].s_rise_rate,\r\n 's_maintain_rate': state_transition_result[-1].s_maintain_rate,\r\n 's_fall_rate': state_transition_result[-1].s_fall_rate,\r\n 'l_rise_rate': state_transition_result[-1].l_rise_rate,\r\n 'l_maintain_rate': state_transition_result[-1].l_maintain_rate,\r\n 'l_fall_rate': state_transition_result[-1].l_fall_rate}\r\n s_rise_sparkline_data = []\r\n s_fall_sparkline_data = []\r\n l_rise_sparkline_data = []\r\n l_fall_sparkline_data = []\r\n state_transition_line_data = []\r\n for i in state_transition_result[-60:]:\r\n s_rise_sparkline_data.append(i.s_rise_rate)\r\n s_fall_sparkline_data.append(i.s_fall_rate)\r\n l_rise_sparkline_data.append(i.l_rise_rate)\r\n l_fall_sparkline_data.append(i.l_fall_rate)\r\n for i, j in zip(bar_data_for_state_transition_result, state_transition_result):\r\n trade_date = i.trade_date\r\n state_transition_line_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.open, i.close, i.low, i.high,\r\n i.amount, j.s_rise_rate, j.s_fall_rate, j.l_rise_rate, j.l_fall_rate])\r\n state_transition_data = {'state_transition_line_data': state_transition_line_data,\r\n 'today_state_transition_data': today_state_transition_data,\r\n 's_rise_sparkline_data': s_rise_sparkline_data,\r\n 's_fall_sparkline_data': s_fall_sparkline_data,\r\n 'l_rise_sparkline_data': l_rise_sparkline_data,\r\n 'l_fall_sparkline_data': l_fall_sparkline_data}\r\n\r\n return jsonify({'today_bar_data': today_bar_data, 'today_basic_data': today_basic_data,\r\n 'trade_point_data': {'line_data': trade_point_line_data, 'aggressive': aggressive_trade_point,\r\n 'steady': steady_trade_point, 'aggressive_requirement': aggressive_requirement,\r\n 'steady_requirement': steady_requirement},\r\n 'trend_forecast_data': {'first_vote': first_vote, 'second_vote': second_vote,\r\n 'third_vote': third_vote,\r\n 'first_change': trend_forecast_result.first_change,\r\n 'second_change': trend_forecast_result.second_change,\r\n 'third_change': trend_forecast_result.third_change},\r\n 'similarity_short_term_data': similarity_short_term_data,\r\n 'correlation_short_term_data': correlation_short_term_data,\r\n 'similarity_history_data': similarity_history_data,\r\n 'state_transition_data': state_transition_data})\r\n\r\n\r\n@stock_prediction.route('api_stock_assessment', methods=['GET'])\r\ndef api_stock_assessment():\r\n code = request.args.get('code')\r\n date = request.args.get('date')\r\n stock_daily_bar_result = Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code == code,\r\n Stock_Daily_Bar.trade_date <= date).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(500).all()\r\n stock_daily_basic_result = Stock_Daily_Basic.query.filter(Stock_Daily_Basic.ts_code == code,\r\n Stock_Daily_Basic.trade_date <= date).order_by(\r\n Stock_Daily_Basic.trade_date.desc()).limit(500).all()\r\n stock_assessment_result = Model_Stock_Assessment.query.filter(Model_Stock_Assessment.ts_code == code,\r\n Model_Stock_Assessment.trade_date <= date).order_by(\r\n Model_Stock_Assessment.trade_date.desc()).limit(500).all()\r\n stock_daily_bar_result.reverse()\r\n stock_daily_basic_result.reverse()\r\n stock_assessment_result.reverse()\r\n company_data = Stock_Company_Extend.query.join(Stock_Industry_SW_3,\r\n Stock_Company_Extend.industry_sw_code == Stock_Industry_SW_3.industry_sw_3_code).add_columns(\r\n Stock_Industry_SW_3.industry_sw_3_name).join(\r\n Stock_Industry_SW_2, Stock_Industry_SW_3.belong_to == Stock_Industry_SW_2.industry_sw_2_code).add_columns(\r\n Stock_Industry_SW_2.industry_sw_2_name).join(\r\n Stock_Industry_SW_1, Stock_Industry_SW_2.belong_to == Stock_Industry_SW_1.industry_sw_1_code).add_columns(\r\n Stock_Industry_SW_1.industry_sw_1_code, Stock_Industry_SW_1.industry_sw_1_name).join(Stock_Industry_CSRC_2,\r\n Stock_Company_Extend.industry_csrc_code== Stock_Industry_CSRC_2.industry_csrc_2_code).join(\r\n Stock_Industry_CSRC_1,\r\n Stock_Industry_CSRC_2.belong_to == Stock_Industry_CSRC_1.industry_csrc_1_code).add_columns(\r\n Stock_Industry_CSRC_1.industry_csrc_1_code, Stock_Industry_CSRC_1.industry_csrc_1_name).filter(\r\n Stock_Company_Extend.ts_code == code).first()\r\n sw_industry, sw_industry_name = company_data.industry_sw_1_code, company_data.industry_sw_1_name\r\n csrc_industry, csrc_industry_name = company_data.industry_csrc_1_code, company_data.industry_csrc_1_name\r\n sw_industry_basic_result = Stock_Industry_Basic.query.filter(Stock_Industry_Basic.industry_code == sw_industry,\r\n Stock_Industry_Basic.trade_date <= date).order_by(\r\n Stock_Industry_Basic.trade_date.desc()).limit(500).all()\r\n csrc_industry_basic_result = Stock_Industry_Basic.query.filter(Stock_Industry_Basic.industry_code == csrc_industry,\r\n Stock_Industry_Basic.trade_date <= date).order_by(\r\n Stock_Industry_Basic.trade_date.desc()).limit(500).all()\r\n sw_industry_basic_result.reverse()\r\n csrc_industry_basic_result.reverse()\r\n csrc_order_result = Stock_Daily_Basic.query.join(Stock_Basic,\r\n Stock_Daily_Basic.ts_code == Stock_Basic.ts_code).add_columns(\r\n Stock_Basic.name).join(Stock_Company_Extend, Stock_Daily_Basic.ts_code == Stock_Company_Extend.ts_code).join(\r\n Stock_Industry_CSRC_2,\r\n Stock_Company_Extend.industry_csrc_code == Stock_Industry_CSRC_2.industry_csrc_2_code).join(\r\n Stock_Industry_CSRC_1,\r\n Stock_Industry_CSRC_2.belong_to == Stock_Industry_CSRC_1.industry_csrc_1_code).filter(\r\n Stock_Daily_Basic.trade_date == date,\r\n Stock_Industry_CSRC_1.industry_csrc_1_code == csrc_industry,Stock_Daily_Basic.pe_ttm!=None).order_by(Stock_Daily_Basic.pe_ttm.asc()).all()\r\n sw_order_result = Stock_Daily_Basic.query.join(Stock_Basic,\r\n Stock_Daily_Basic.ts_code == Stock_Basic.ts_code).add_columns(\r\n Stock_Basic.name).join(Stock_Company_Extend,\r\n Stock_Daily_Basic.ts_code == Stock_Company_Extend.ts_code).join(\r\n Stock_Industry_SW_3,\r\n Stock_Company_Extend.industry_sw_code == Stock_Industry_SW_3.industry_sw_3_code).join(\r\n Stock_Industry_SW_2, Stock_Industry_SW_3.belong_to == Stock_Industry_SW_2.industry_sw_2_code).join(\r\n Stock_Industry_SW_1, Stock_Industry_SW_2.belong_to == Stock_Industry_SW_1.industry_sw_1_code).filter(\r\n Stock_Daily_Basic.trade_date == date,\r\n Stock_Industry_SW_1.industry_sw_1_code == sw_industry,Stock_Daily_Basic.pe_ttm!=None).order_by(Stock_Daily_Basic.pe_ttm.asc()).all()\r\n csrc_order_number = 0\r\n sw_order_number = 0\r\n\r\n for i in range(0, len(csrc_order_result)):\r\n if csrc_order_result[i].Stock_Daily_Basic.ts_code == code:\r\n csrc_order_number = i + 1\r\n break\r\n for i in range(0, len(sw_order_result)):\r\n if sw_order_result[i].Stock_Daily_Basic.ts_code == code:\r\n sw_order_number = i + 1\r\n break\r\n csrc_order_data = []\r\n sw_order_data = []\r\n for i in csrc_order_result[0:5]:\r\n csrc_order_data.append([i.Stock_Daily_Basic.ts_code,i.name, i.Stock_Daily_Basic.pe_ttm])\r\n for i in sw_order_result[0:5]:\r\n sw_order_data.append([i.Stock_Daily_Basic.ts_code,i.name, i.Stock_Daily_Basic.pe_ttm])\r\n close_list = []\r\n for i in stock_daily_bar_result[-15:]:\r\n trade_date = i.trade_date\r\n close_list.append([(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.close])\r\n assessment_list = []\r\n for i in stock_assessment_result:\r\n assessment_list.append([i.estimated_value, i.estimated_value_std])\r\n pe_list = []\r\n for i in stock_daily_basic_result[-20:]:\r\n trade_date = i.trade_date\r\n pe_list.append([(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.pe_ttm])\r\n csrc_pe_list = []\r\n sw_pe_list = []\r\n for i in csrc_industry_basic_result[-20:]:\r\n csrc_pe_list.append(i.pe_ttm_overall)\r\n for i in sw_industry_basic_result[-20:]:\r\n sw_pe_list.append(i.pe_ttm_overall)\r\n\r\n today_data = {'close': stock_daily_bar_result[-1].close,\r\n 'pe_ttm': stock_daily_basic_result[-1].pe_ttm,\r\n 'csrc_pe': csrc_industry_basic_result[-1].pe_ttm_overall,\r\n 'sw_pe': sw_industry_basic_result[-1].pe_ttm_overall,\r\n 'estimated_value': stock_assessment_result[-1].estimated_value,\r\n 'estimated_value_std': stock_assessment_result[-1].estimated_value_std,\r\n 'similar_pe_count': stock_assessment_result[-1].similar_pe_count,\r\n 'pe_mean': stock_assessment_result[-1].pe_mean,\r\n 'pe_std': stock_assessment_result[-1].pe_std,\r\n 'pe_min': stock_assessment_result[-1].pe_min,\r\n 'pe_max': stock_assessment_result[-1].pe_max}\r\n\r\n assessment_data = {'close': close_list, 'assessment': assessment_list}\r\n\r\n industry_compare_data = {'pe': pe_list, 'csrc_pe': csrc_pe_list, 'sw_pe': sw_pe_list}\r\n\r\n order_data = {'sw_industry': sw_industry_name, 'csrc_industry': csrc_industry_name, 'csrc_order': csrc_order_number,\r\n 'csrc_order_data': csrc_order_data, 'sw_order': sw_order_number, 'sw_order_data': sw_order_data}\r\n return jsonify(\r\n {'today_data': today_data, 'assessment_data': assessment_data, 'industry_compare_data': industry_compare_data,\r\n 'order_data': order_data})\r\n\r\n\r\n@stock_prediction.route('api_money_flow', methods=['GET'])\r\ndef api_money_flow():\r\n code = request.args.get('code')\r\n date = request.args.get('date')\r\n stock_bar_result = Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code == code,\r\n Stock_Daily_Bar.trade_date <= date).order_by(\r\n Stock_Daily_Bar.trade_date.desc()).limit(20).all()\r\n stock_daily_basic_result = Stock_Daily_Basic.query.filter(Stock_Daily_Basic.ts_code == code,\r\n Stock_Daily_Basic.trade_date <= date).order_by(\r\n Stock_Daily_Basic.trade_date.desc()).limit(20).all()\r\n money_flow_result = Market_Money_Flow.query.filter(Market_Money_Flow.ts_code == code,\r\n Market_Money_Flow.trade_date <= date).order_by(\r\n Market_Money_Flow.trade_date.desc()).limit(20).all()\r\n stock_bar_result.reverse()\r\n stock_daily_basic_result.reverse()\r\n money_flow_result.reverse()\r\n today_data = {'buy_sm': money_flow_result[-1].buy_sm_amount,\r\n 'sell_sm': money_flow_result[-1].sell_sm_amount,\r\n 'buy_md': money_flow_result[-1].buy_md_amount,\r\n 'sell_md': money_flow_result[-1].sell_md_amount,\r\n 'buy_lg': money_flow_result[-1].buy_lg_amount,\r\n 'sell_lg': money_flow_result[-1].sell_lg_amount,\r\n 'buy_elg': money_flow_result[-1].buy_elg_amount,\r\n 'sell_elg': money_flow_result[-1].sell_lg_amount,\r\n 'amount': stock_bar_result[-1].amount / 10}\r\n pct_chg_list = []\r\n trade_date_list = []\r\n for i in stock_bar_result:\r\n trade_date = i.trade_date\r\n trade_date_list.append((trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]))\r\n pct_chg_list.append(i.pct_chg)\r\n main_buy_list = []\r\n main_sell_list = []\r\n retail_buy_list = []\r\n retail_sell_list = []\r\n money_dist_data = []\r\n net_list = []\r\n for i in money_flow_result:\r\n trade_date = i.trade_date\r\n net_list.append(i.net_mf_amount)\r\n main_buy_list.append(i.buy_lg_amount + i.buy_elg_amount)\r\n main_sell_list.append(i.sell_lg_amount + i.sell_elg_amount)\r\n retail_buy_list.append(i.buy_sm_amount + i.buy_md_amount)\r\n retail_sell_list.append(i.sell_sm_amount + i.sell_md_amount)\r\n money_dist_data.append(\r\n [(trade_date[0:4] + '/' + trade_date[4:6] + '/' + trade_date[6:8]), i.buy_sm_amount + i.sell_sm_amount,\r\n i.buy_md_amount + i.sell_md_amount, i.buy_lg_amount + i.sell_lg_amount,\r\n i.buy_elg_amount + i.sell_elg_amount])\r\n corr_list = []\r\n scale_list = []\r\n for x, y, z in zip(stock_daily_basic_result, pct_chg_list, net_list):\r\n try:\r\n corr_list.append(z / (x.circ_mv / (1 + y/100) * y/100))\r\n except:\r\n corr_list.append(0)\r\n for i in money_dist_data:\r\n scale_list.append((i[3] + i[4]) / (i[1] + i[2] + i[3] + i[4]))\r\n money_flow_data = {'trade_date': trade_date_list, 'pct_chg': pct_chg_list, 'main_buy': main_buy_list,\r\n 'main_sell': main_sell_list, 'retail_buy': retail_buy_list, 'retail_sell': retail_sell_list}\r\n main_ability = {'scale': scale_list[-1],\r\n 'corr': corr_list[-1],\r\n 'scale_list': scale_list,\r\n 'corr_list': corr_list}\r\n return jsonify({'today_data': today_data, 'money_flow_data': money_flow_data, 'money_dist_data': money_dist_data,\r\n 'main_ability': main_ability})\r\n\r\n\r\n@stock_prediction.route('api_rule_statistics', methods=['GET'])\r\ndef api_rule_statistics():\r\n code = request.args.get('code')\r\n date = request.args.get('date')\r\n bar_data = Stock_Daily_Bar.query.filter_by(ts_code=code, trade_date=date).first()\r\n basic_data = Stock_Daily_Basic.query.filter_by(ts_code=code, trade_date=date).first()\r\n today_bar_data = {'open': bar_data.open, 'close': bar_data.close, 'high': bar_data.high,\r\n 'low': bar_data.low, 'change': bar_data.change, 'pct_chg': bar_data.pct_chg,\r\n 'pre_close': bar_data.pre_close, 'vol': bar_data.vol, 'amount': bar_data.amount}\r\n if basic_data is None:\r\n today_basic_data = {'turnover_rate': 0, 'pe': 0, 'pb': 0,\r\n 'circ_mv': 0}\r\n else:\r\n today_basic_data = {'turnover_rate': basic_data.turnover_rate, 'pe': basic_data.pe, 'pb': basic_data.pb,\r\n 'circ_mv': basic_data.circ_mv}\r\n rise_rise_data=[]\r\n rise_rise_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='rise_rise').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in rise_rise_result:\r\n rise_rise_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n rise_fall_data=[]\r\n rise_fall_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='rise_fall').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in rise_fall_result:\r\n rise_fall_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n rise_not_rise_data=[]\r\n rise_not_rise_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='rise_not_rise').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in rise_not_rise_result:\r\n rise_not_rise_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n rise_not_fall_data=[]\r\n rise_not_fall_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='rise_not_fall').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in rise_not_fall_result:\r\n rise_not_fall_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n fall_rise_data=[]\r\n fall_rise_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='fall_rise').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in fall_rise_result:\r\n fall_rise_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n fall_fall_data=[]\r\n fall_fall_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='fall_fall').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in fall_fall_result:\r\n fall_fall_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n fall_not_rise_data=[]\r\n fall_not_rise_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='fall_not_rise').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in fall_not_rise_result:\r\n fall_not_rise_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n fall_not_fall_data=[]\r\n fall_not_fall_result=Model_Associate_Rule.query.join(Stock_Basic,Model_Associate_Rule.associate_code==Stock_Basic.ts_code).add_columns(Stock_Basic.name).filter(Model_Associate_Rule.ts_code==code,Model_Associate_Rule.trade_date==date,Model_Associate_Rule.associate_type=='fall_not_fall').order_by(Model_Associate_Rule.probability.desc()).all()\r\n for i in fall_not_fall_result:\r\n fall_not_fall_data.append([i.name,i.Model_Associate_Rule.approval_rating,i.Model_Associate_Rule.probability,i.Model_Associate_Rule.trading_day_count,i.Model_Associate_Rule.matching_day_count,i.Model_Associate_Rule.effect_day_count])\r\n associate_rule_data={'rise_rise':rise_rise_data,'rise_fall':rise_fall_data,'rise_not_rise':rise_not_rise_data,'rise_not_fall':rise_not_fall_data,\r\n 'fall_rise':fall_rise_data,'fall_fall':fall_fall_data,'fall_not_rise':fall_not_rise_data,'fall_not_fall':fall_not_fall_data}\r\n\r\n fluctuation_statistics_result=Model_Fluctuation_Statistics.query.filter_by(ts_code=code,trade_date=date).first()\r\n fluctuation_statistics_data={'D4':fluctuation_statistics_result.D4,\r\n 'D3':fluctuation_statistics_result.D3,\r\n 'D2':fluctuation_statistics_result.D2,\r\n 'D1':fluctuation_statistics_result.D1,\r\n 'five_total_count':fluctuation_statistics_result.five_total_count,\r\n 'five_D0_list':sql_to_object(fluctuation_statistics_result.five_D0_list),\r\n 'five_appearance_count_list':sql_to_object(fluctuation_statistics_result.five_appearance_count_list),\r\n 'five_proportion_list':sql_to_object(fluctuation_statistics_result.five_proportion_list),\r\n 'four_total_count':fluctuation_statistics_result.four_total_count,\r\n 'four_D0_list':sql_to_object(fluctuation_statistics_result.four_D0_list),\r\n 'four_appearance_count_list':sql_to_object(fluctuation_statistics_result.four_appearance_count_list),\r\n 'four_proportion_list':sql_to_object(fluctuation_statistics_result.four_proportion_list),\r\n 'three_total_count':fluctuation_statistics_result.three_total_count,\r\n 'three_D0_list':sql_to_object(fluctuation_statistics_result.three_D0_list),\r\n 'three_appearance_count_list':sql_to_object(fluctuation_statistics_result.three_appearance_count_list),\r\n 'three_proportion_list':sql_to_object(fluctuation_statistics_result.three_proportion_list),\r\n 'two_total_count':fluctuation_statistics_result.two_total_count,\r\n 'two_D0_list':sql_to_object(fluctuation_statistics_result.two_D0_list),\r\n 'two_appearance_count_list':sql_to_object(fluctuation_statistics_result.two_appearance_count_list),\r\n 'two_proportion_list':sql_to_object(fluctuation_statistics_result.two_proportion_list),\r\n 'one_total_count':fluctuation_statistics_result.one_total_count,\r\n 'one_D0_list':sql_to_object(fluctuation_statistics_result.one_D0_list),\r\n 'one_appearance_count_list':sql_to_object(fluctuation_statistics_result.one_appearance_count_list),\r\n 'one_proportion_list':sql_to_object(fluctuation_statistics_result.one_proportion_list)}\r\n\r\n fluctuation_correlation_result=Model_Fluctuation_Correlation.query.filter_by(ts_code=code,trade_date=date).first()\r\n fluctuation_correlation_data={'D4':fluctuation_correlation_result.D4,\r\n 'D3':fluctuation_correlation_result.D3,\r\n 'D2':fluctuation_correlation_result.D2,\r\n 'D1':fluctuation_correlation_result.D1,\r\n 'five_sample_count':fluctuation_correlation_result.five_sample_count,\r\n 'five_total_count':fluctuation_correlation_result.five_total_count,\r\n 'five_approval_rating':fluctuation_correlation_result.five_approval_rating,\r\n 'five_D0_list':sql_to_object(fluctuation_correlation_result.five_D0_list),\r\n 'five_appearance_count_list':sql_to_object(fluctuation_correlation_result.five_appearance_count_list),\r\n 'five_proportion_list':sql_to_object(fluctuation_correlation_result.five_proportion_list),\r\n 'four_sample_count': fluctuation_correlation_result.four_sample_count,\r\n 'four_total_count': fluctuation_correlation_result.four_total_count,\r\n 'four_approval_rating': fluctuation_correlation_result.four_approval_rating,\r\n 'four_D0_list': sql_to_object(fluctuation_correlation_result.four_D0_list),\r\n 'four_appearance_count_list': sql_to_object(fluctuation_correlation_result.four_appearance_count_list),\r\n 'four_proportion_list': sql_to_object(fluctuation_correlation_result.four_proportion_list),\r\n 'three_sample_count': fluctuation_correlation_result.three_sample_count,\r\n 'three_total_count': fluctuation_correlation_result.three_total_count,\r\n 'three_approval_rating': fluctuation_correlation_result.three_approval_rating,\r\n 'three_D0_list': sql_to_object(fluctuation_correlation_result.three_D0_list),\r\n 'three_appearance_count_list': sql_to_object(fluctuation_correlation_result.three_appearance_count_list),\r\n 'three_proportion_list': sql_to_object(fluctuation_correlation_result.three_proportion_list),\r\n 'two_sample_count': fluctuation_correlation_result.two_sample_count,\r\n 'two_total_count': fluctuation_correlation_result.two_total_count,\r\n 'two_approval_rating': fluctuation_correlation_result.two_approval_rating,\r\n 'two_D0_list': sql_to_object(fluctuation_correlation_result.two_D0_list),\r\n 'two_appearance_count_list': sql_to_object(fluctuation_correlation_result.two_appearance_count_list),\r\n 'two_proportion_list': sql_to_object(fluctuation_correlation_result.two_proportion_list),\r\n }\r\n\r\n fluctuation_sequencing_result=Model_Fluctuation_Sequencing.query.filter_by(ts_code=code,trade_date='20190430').first()\r\n fluctuation_sequencing_data={'close':fluctuation_sequencing_result.close,\r\n 'high_change':fluctuation_sequencing_result.high_change,\r\n 'high_order':fluctuation_sequencing_result.high_order,\r\n 'low_change':fluctuation_sequencing_result.low_change,\r\n 'low_order':fluctuation_sequencing_result.low_order,\r\n 'order_sum':fluctuation_sequencing_result.order_sum}\r\n\r\n high_change_list=[]\r\n high_change_result=Model_Fluctuation_Sequencing.query.filter_by(trade_date='20190430').all()\r\n for i in high_change_result:\r\n high_change_list.append(int(i.high_change*100+0.5))\r\n high_change_statistics=[0]*101\r\n for i in high_change_list:\r\n high_change_statistics[100+i]+=1\r\n similarity_fluctuation_result = Model_Similarity_Fluctuation.query.filter_by(ts_code=code,trade_date=date).first()\r\n similarity_trend_result = Model_Similarity_Trend.query.filter_by(ts_code=code,trade_date=date).first()\r\n similarity_fluctuation_stock_list=sql_to_object(similarity_fluctuation_result.stock_list)\r\n similarity_fluctuation_liveness_list=sql_to_object(similarity_fluctuation_result.liveness_list)\r\n similarity_trend_stock_list = sql_to_object(similarity_trend_result.stock_list)\r\n similarity_trend_liveness_list = sql_to_object(similarity_trend_result.liveness_list)\r\n similarity_fluctuation_stock_data=[]\r\n similarity_fluctuation_pct_chg_data=[]\r\n for i in similarity_fluctuation_stock_list:\r\n stock_data=Stock_Basic.query.filter_by(ts_code=i).first()\r\n stock_bar=Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code==i,Stock_Daily_Bar.trade_date<=date).order_by(Stock_Daily_Bar.trade_date.desc()).limit(120).all()\r\n stock_bar.reverse()\r\n pct_chg_data=[]\r\n for j in stock_bar:\r\n if j.pct_chg>=2:\r\n pct_chg_data.append(1)\r\n elif j.pct_chg<=-2:\r\n pct_chg_data.append(-1)\r\n else:\r\n pct_chg_data.append(0)\r\n similarity_fluctuation_stock_data.append({'name': stock_data.name, 'symbol': stock_data.symbol})\r\n similarity_fluctuation_pct_chg_data.append(pct_chg_data)\r\n similarity_trend_stock_data=[]\r\n similarity_trend_close_data=[]\r\n for i in similarity_trend_stock_list:\r\n stock_data=Stock_Basic.query.filter_by(ts_code=i).first()\r\n stock_bar=Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code==i,Stock_Daily_Bar.trade_date<=date,Stock_Daily_Bar.trade_date>='20150101').order_by(Stock_Daily_Bar.trade_date.asc()).all()\r\n close_data=[]\r\n for j in stock_bar:\r\n close_data.append(j.close)\r\n similarity_trend_stock_data.append({'name': stock_data.name, 'symbol': stock_data.symbol})\r\n similarity_trend_close_data.append(close_data)\r\n\r\n cur_stock_pct_chg_data=[]\r\n cur_stock_close_data=[]\r\n cur_stock_bar_result=Stock_Daily_Bar.query.filter(Stock_Daily_Bar.ts_code==code,Stock_Daily_Bar.trade_date<=date,Stock_Daily_Bar.trade_date>='20150101').order_by(Stock_Daily_Bar.trade_date.asc()).all()\r\n n=0\r\n for i in cur_stock_bar_result[-120:]:\r\n if i.pct_chg >= 2:\r\n cur_stock_pct_chg_data.append(1)\r\n elif i.pct_chg <= -2:\r\n cur_stock_pct_chg_data.append(-1)\r\n else:\r\n n+=1\r\n cur_stock_pct_chg_data.append(0)\r\n cur_stock_liveness_data=(120-n)/120\r\n for i in cur_stock_bar_result:\r\n cur_stock_close_data.append(i.close)\r\n\r\n\r\n return jsonify({'associate_rule':associate_rule_data,'fluctuation_statistics':fluctuation_statistics_data,'fluctuation_correlation':fluctuation_correlation_data,\r\n 'fluctuation_sequencing':{'sequencing_data':fluctuation_sequencing_data,'high_change_statistics':high_change_statistics},\r\n 'similarity_fluctuation':{'stock':similarity_fluctuation_stock_data,'liveness':similarity_fluctuation_liveness_list,'pct_chg':similarity_fluctuation_pct_chg_data,'cur_stock_pct_chg':cur_stock_pct_chg_data,'cur_stock_liveness':cur_stock_liveness_data},\r\n 'similarity_trend':{'stock':similarity_trend_stock_data,'liveness':similarity_trend_liveness_list,'close':similarity_trend_close_data,'cur_stock_close':cur_stock_close_data,'cur_stock_liveness':cur_stock_liveness_data},\r\n 'today_bar_data':today_bar_data,'today_basic_data':today_basic_data})\r\n\r\n\r\n@stock_prediction.route('api_stock_basic', methods=['GET'])\r\ndef api_stock_basic():\r\n code = request.args.get('code')\r\n basic_data=Stock_Basic.query.filter_by(ts_code=code).first()\r\n company_data=Stock_Company.query.filter_by(ts_code=code).first()\r\n company_data_extend= Stock_Company_Extend.query.filter_by(ts_code=code).first()\r\n result={'fullname':basic_data.fullname,'website':company_data.website,'email':company_data.email,'enname':basic_data.enname,'chairman':company_data.chairman,'office':company_data.office,\r\n 'manager':company_data.manager,'secretary':company_data.secretary,'reg_capital':company_data.reg_capital,'industry':company_data_extend.industry_csrc_name,'organizationcode':company_data_extend.organizationcode,'fax':company_data_extend.fax,'setup_date':company_data.setup_date,'list_date':basic_data.list_date,\r\n 'employees':company_data.employees,'main_business':company_data.main_business,'introduction':company_data.introduction,'business_scope':company_data.business_scope}\r\n return jsonify(result)\r\n\r\n@stock_prediction.route('api_stock_basic_ipo', methods=['GET'])\r\ndef api_stock_basic_ipo():\r\n code = request.args.get('code')\r\n ipo_data=Stock_IPO_Info.query.filter_by(ts_code=code).first()\r\n result={'ipo_date':ipo_data.ipo_date,'ipo_price':ipo_data.ipo_price,'ipo_collection':ipo_data.ipo_collection,'ipo_puboffrdate':ipo_data.ipo_puboffrdate,'ipo_leadundr':ipo_data.ipo_leadundr,'ipo_nominator':ipo_data.ipo_nominator,'ipo_sponsorrepresentative':ipo_data.ipo_sponsorrepresentative,'ipo_type':ipo_data.ipo_type,\r\n 'ipo_expense':ipo_data.ipo_expense,'ipo_amount':ipo_data.ipo_amount,'ipo_weightedpe':ipo_data.ipo_weightedpe}\r\n return jsonify(result)\r\n\r\n\r\n","sub_path":"webapp/controller/stock_prediction/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":64356,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"439732715","text":"import httplib2\nimport json\n\n#google map api key\ngoogle_geocode_id = \"AIzaSyAvnFnZSiMNLeXX3eGh5kf9NEySf-zjkHY\"\n\nclass Maps:\n\t\n\t@staticmethod\n\tdef getGeocodeLocation(inputString):\n\t\tlocationString = inputString.replace(\" \",\"+\")\n\t\turl = (\"https://maps.googleapis.com/maps/api/geocode/json?address=%s&key=%s\"%(locationString, google_geocode_id))\n\t\th = httplib2.Http()\n\t\tresponse, content = h.request(url, 'GET')\n\t\tresult = json.loads(content)\n\t\treturn result['results'][0]['geometry']['location']\n\n\t#test code\n\t#print getGeocodeLocation('Chaguanas')","sub_path":"api/map.py","file_name":"map.py","file_ext":"py","file_size_in_byte":546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"72505237","text":"import abc\nfrom typing import List\nimport numpy as np\nimport attr\nfrom naturalnets.brains.i_brain import IBrain, IBrainCfg\n\n\n@attr.s(slots=True, auto_attribs=True, frozen=True, kw_only=True)\nclass ILayerBasedBrainCfg(IBrainCfg):\n # The structure of the layers\n # Each list entry translates to the size of one layer\n # The layers are in the given order\n hidden_layer_structure: List[int]\n # Whether a neuron can only use its own state form the last timestep\n diagonal_hidden_to_hidden: bool = False\n use_bias: bool = False\n\n\nclass ILayerBasedBrain(IBrain, abc.ABC):\n\n @staticmethod\n @abc.abstractmethod\n def layer_step(layer_input: np.ndarray, weight_ih, weight_hh, bias_h, hidden):\n # Compute one layer step\n pass\n\n @staticmethod\n @abc.abstractmethod\n def get_number_gates():\n # How many Gates are used in the specific network?\n # Haw many matrices are needed for each layer to calculate the next state and output value\n pass\n\n @staticmethod\n @abc.abstractmethod\n def get_number_hidden_values():\n # How many hidden values are used in one cell\n pass\n\n def __init__(self, input_size: int, output_size: int, individual: np.ndarray, configuration: ILayerBasedBrainCfg,\n brain_state: dict):\n if not type(configuration) is ILayerBasedBrainCfg:\n configuration = ILayerBasedBrainCfg(**configuration)\n hidden_layer_structure: List[int] = configuration.hidden_layer_structure\n\n # initialize weights out of individual\n\n individual_index = 0 # progress index\n # initialize empty\n self.weight_ih = [] # Weights for weighted input values\n self.weight_hh = [] # Weights for weighted stored values\n self.bias_h = [] # Biases\n self.hidden = [] # Initial values for state storage\n self.layer_output = [] # Weights from output last Layer to output nodes\n number_gates = self.get_number_gates()\n number_hidden_values = self.get_number_hidden_values()\n\n # iterate for all given layers in the structure\n for layer in range(len(hidden_layer_structure)):\n\n # Matrices for weighted input values in calculations\n layer_input_size = input_size if layer == 0 else hidden_layer_structure[layer - 1]\n number_elements = number_gates * hidden_layer_structure[layer] * layer_input_size\n self.weight_ih.append(\n np.array(\n individual[individual_index: individual_index + number_elements]\n ).reshape((number_gates, hidden_layer_structure[layer], layer_input_size))\n )\n individual_index += number_elements\n\n # Matrices for weighted state values in calculations\n if configuration.diagonal_hidden_to_hidden: # Whether each neuron can only access its own state\n self.weight_hh.append(\n [np.diag(individual[\n individual_index + k * hidden_layer_structure[layer]:\n individual_index + k * hidden_layer_structure[layer] + hidden_layer_structure[layer]\n ])\n for k in range(number_gates)\n ]\n )\n individual_index += number_gates * hidden_layer_structure[layer]\n else:\n number_elements = number_gates * hidden_layer_structure[layer] * hidden_layer_structure[layer]\n self.weight_hh.append(\n np.array(\n individual[individual_index: individual_index + number_elements]\n ).reshape((number_gates, hidden_layer_structure[layer], hidden_layer_structure[layer]))\n )\n individual_index += number_elements\n\n # initialize biases\n\n # Biases for gates\n if configuration.use_bias:\n number_elements = hidden_layer_structure[layer] * number_gates\n self.bias_h.append(\n np.array(\n individual[individual_index: individual_index + number_elements]\n ).reshape((number_gates, hidden_layer_structure[layer]))\n )\n individual_index += number_elements\n else:\n self.bias_h.append(np.zeros((number_gates, hidden_layer_structure[layer])).astype(np.float32))\n\n # initialize initial state values\n self.hidden.append(np.zeros((number_hidden_values, hidden_layer_structure[layer])))\n\n self.layer_output.append(np.zeros((hidden_layer_structure[layer])))\n # for end\n\n # Matrix for transforming output of last layer into output neurons\n number_elements = hidden_layer_structure[len(hidden_layer_structure) - 1] * output_size\n self.weight_ho = np.array(\n individual[individual_index: individual_index + number_elements]\n ).reshape((output_size, hidden_layer_structure[len(hidden_layer_structure) - 1]))\n individual_index += number_elements\n\n # Has all values been used and therefore does get_individual_size() provide the right number?\n assert individual_index == len(individual)\n\n @classmethod\n def get_free_parameter_usage(cls, input_size: int, output_size: int, configuration: dict, brain_state: dict):\n\n configuration = ILayerBasedBrainCfg(**configuration)\n number_gates = cls.get_number_gates()\n hidden_size = cls.get_number_hidden_values()\n hidden_structure = configuration.hidden_layer_structure\n individuals = {}\n\n for layer in range(len(hidden_structure)):\n layer_dict = {\n # Matrices for weighted input values\n # The first Layer don't has an output from the previous layer, but the input values\n 'input_weight_matrix': number_gates * hidden_structure[layer] * (\n input_size if layer == 0 else hidden_structure[layer - 1]),\n # Matrices for weighted state values\n 'hidden_weight_matrix': number_gates * hidden_structure[layer] * (\n 1 if configuration.diagonal_hidden_to_hidden else hidden_structure[layer])\n }\n # initialize biases\n if configuration.use_bias:\n layer_dict['bias'] = hidden_structure[layer] * number_gates\n\n individuals['layer ' + str(layer)] = layer_dict\n # for end\n\n # Matrix for transforming output of last layer into output neurons\n individuals['output_weight_matrix'] = hidden_structure[len(hidden_structure) - 1] * output_size\n return individuals\n\n def step(self, ob: np.ndarray):\n layer_input = ob\n # iterate for all given layers\n for layer in range(len(self.hidden)):\n if layer == 0:\n x = layer_input\n else:\n x = self.layer_output[layer - 1]\n # Returns a list with two elements.\n # The first element is the calculated new hidden cell state, the second is the layer output\n # Necessary for LSTM\n layer_result = self.layer_step(x, self.weight_ih[layer], self.weight_hh[layer], self.bias_h[layer],\n self.hidden[layer])\n self.hidden[layer] = layer_result[0]\n self.layer_output[layer] = layer_result[1]\n return np.dot(self.weight_ho, self.layer_output[len(self.layer_output) - 1])\n\n def reset(self):\n self.hidden = np.zeros_like(self.hidden)\n\n @classmethod\n def generate_brain_state(cls, input_size: int, output_size: int, configuration: dict):\n pass\n\n @classmethod\n def save_brain_state(cls, path, brain_state):\n pass\n","sub_path":"naturalnets/brains/i_layer_based_brain.py","file_name":"i_layer_based_brain.py","file_ext":"py","file_size_in_byte":7815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"290959533","text":"from django.conf.urls import patterns, include, url\nfrom django.contrib import admin\nfrom nordostra import views\nadmin.autodiscover()\nurlpatterns = patterns('',\n url(r'^admin/', include(admin.site.urls)),\n url(r'^session/$', 'nordostra.views.home', name='home'),\n url(r'^session/(?P\\d{1,4})/$', 'nordostra.views.percent', name='percent'),\n url(r'^accounts/', include('registration.backends.default.urls')),\n)\n\n \n","sub_path":"nordostra/nordostra/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"135680930","text":"class ApiKeySecurity:\n\n def __init__(self, key_in: str = None, name: str = None, description: str = None, **kwargs):\n self.configs = {\n 'type': 'apiKey',\n 'in': key_in,\n 'name': name,\n 'description': description,\n **kwargs\n }\n\n\nclass ApiKeySecurityIn:\n header = 'header'\n query = 'query'\n","sub_path":"nest/packages/swagger/core/api_key_security.py","file_name":"api_key_security.py","file_ext":"py","file_size_in_byte":366,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"205017750","text":"#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nimport os\nimport json\nimport base64\nimport time\nimport datetime\nimport xlrd\nfrom .task import Task\n\nimport traceback\n\nfrom models import conn, purse, api, member\n\nTEMP_DIR = os.path.dirname(os.path.realpath(__file__)) + '/temp'\n\nclass Settlement(Task):\n \n conn = None\n providerName = 'bzl-provider'\n\n def setApi(self, conf):\n Task.setApi(self, conf)\n self.conf = self.api.conf\n self.tempFile = '%s/settlementtime-%s.txt' % (TEMP_DIR, self.conf['name'])\n\n def queryUserBoardList(self, start, end, index = 0):\n params = {\n \"end_time_start\": start,\n \"end_time_end\": end,\n \"query_index\": index\n }\n # print(('query param:', params))\n data = self.api.queryUserBoard(params)\n\n if len(data) >= 30:\n newdata = self.queryUserBoardList(start, end, index = index + 30)\n data.extend(newdata)\n\n return data\n\n def settlement(self):\n try:\n # 判断是否开启\n statusResult = api.getLoginInfo(self.conn, self.conf['serviceCode'])\n # 关闭同步功能\n if not statusResult or statusResult['status'] == 0:\n return\n except Exception as e:\n traceback.print_exc()\n return\n \n now = datetime.datetime.now()\n nowStr = now.strftime('%Y-%m-%d %H:%M:%S')\n lastTimeStr = False\n\n if os.path.exists(self.tempFile):\n try:\n tempfileReader = open(self.tempFile, 'r')\n lastTimeStr = json.loads(tempfileReader.read())['lastTime']\n lastTime = datetime.datetime.strptime(lastTimeStr,'%Y-%m-%d %H:%M:%S')\n lastTimeStr = (lastTime + datetime.timedelta(minutes = -60)).strftime('%Y-%m-%d %H:%M:%S')\n except Exception as e:\n traceback.print_exc()\n if not lastTimeStr:\n lastTimeStr = (now + datetime.timedelta(days = -3)).strftime('%Y-%m-%d %H:%M:%S')\n\n try:\n dataList = self.queryUserBoardList(\n lastTimeStr,\n nowStr,\n 0\n )\n\n print(dataList)\n\n if dataList:\n print(('fetch:', len(dataList)))\n\n for record in dataList:\n self.settleRecord(record)\n else:\n print(('no records'))\n\n tempfile = open(self.tempFile, 'w+')\n tempfile.write(json.dumps({ 'lastTime': nowStr }))\n except Exception as e:\n traceback.print_exc()\n\n def getCustTimestamp(self, timeStr, seconds = 0, minutes = 0):\n t = datetime.datetime.strptime(timeStr,'%Y-%m-%d %H:%M:%S')\n t = t + datetime.timedelta(seconds = seconds, minutes = minutes)\n return str(time.mktime(t.timetuple()))\n\n def settleRecord(self, record):\n if record['pccid'] != '2525717358':\n print('user is not 2525717358')\n return\n currentTime = str(time.time())\n gameEndTime = self.getCustTimestamp(record['end_time'])\n\n print(('开始处理:', record))\n \n # 结算判断标志 pccid_roomName_clubName_buyIn_bringOut_endTime\n settleGameInfo = base64.b64encode(\n ('%s_%s_%s_%s_%s_%s' % (\n record['pccid'],\n record['room_name'],\n record['club_name'],\n record['buy_in'],\n record['bring_out'],\n record['end_time']\n )).encode('utf-8')\n )\n\n gameEndLog = {\n 'game_uid': record['pccid'],\n 'game_id': record['room_name'],\n 'board_id': '',\n 'end_game_time': gameEndTime,\n 'apply_time': currentTime,\n 'settle_game_info': settleGameInfo,\n }\n\n # 判断是否查无此人\n memberResult = purse.getPurseInfoByGameId(self.conn, record['pccid'])\n if not memberResult:\n print(('no user'))\n gameEndLog['action'] = 'no UID'\n purse.addSettleFailLog(self.conn, gameEndLog)\n return\n\n\n # 查询结算表中是否已有结算记录.如果已经存在,则抛弃\n countResult = purse.getSettleRecord(self.conn, settleGameInfo)\n if countResult['settle_count'] > 0:\n print(('already settlemented'))\n return\n\n # 查询游戏期间该用户的所有带入金额是否足够与代理接口一致,不足则不结算\n joinToken = base64.b64encode(('%s_%s' % (record['club_name'], record['room_name'])).encode('utf-8'))\n print(('join token is:', joinToken))\n\n beginTime = self.getCustTimestamp(record['end_time'], minutes = -720)\n endTime = self.getCustTimestamp(record['end_time'], minutes = 120)\n buyInAmountResult = purse.getTotoalBuyinAmount(\n self.conn,\n record['pccid'],\n beginTime,\n endTime,\n joinToken\n )\n\n print(('total buy in:',buyInAmountResult))\n\n if buyInAmountResult['totalAmount'] < record['buy_in']:\n if not buyInAmountResult['totalAmount'] or buyInAmountResult['totalAmount'] == 0:\n print(('no apply'))\n gameEndLog['action'] = 'no Buyin'\n else:\n print(('amount not match, local:', buyInAmountResult['totalAmount'], ', remote', record['buy_in']))\n gameEndLog['action'] = 'no enough, local buyin: %s, remote buyin: %s' % (buyInAmountResult['totalAmount'], record['buy_in'])\n \n purse.addSettleFailLog(self.conn, gameEndLog)\n return\n\n # 记录结算日志\n gameEndLog['action'] = 'OK'\n purse.addSettleFailLog(self.conn, gameEndLog)\n\n memberResult['settle_game_info'] = settleGameInfo\n\n # 更新钱包\n purse.updatePurse(self.conn, memberResult, record['buy_in'] + record['afterwater'])\n\n def toData(self, file):\n name2columnMap = {\n 0: 'pccname',\n 1: 'pccid',\n 6: 'username',\n 8: 'club_name',\n 10: 'room_name',\n 12: 'end_time',\n 14: 'buy_in',\n 15: 'bring_out',\n 17: 'afterwater'\n }\n data = []\n print(('begin transfer local file:', file))\n x1 = xlrd.open_workbook(file)\n sheet1 = x1.sheet_by_index(0)\n print(sheet1)\n if sheet1.nrows <= 1:\n return data\n for rn in range(1, sheet1.nrows):\n rowData = {}\n row = sheet1.row(rn)\n for cn2 in range(0, len(row)):\n if name2columnMap.has_key(cn2):\n name = name2columnMap[cn2]\n rowData[name] = row[cn2].value\n\n rowData['buy_in'] = int(float(rowData['buy_in']))\n rowData['bring_out'] = int(float(rowData['bring_out']))\n rowData['afterwater'] = int(float(rowData['afterwater']))\n\n data.append(rowData)\n\n print(('local datas:', data))\n return data\n\n def localSettlement(self):\n if not os.path.exists(self.conf['localDataPath']):\n return\n files = os.listdir(self.conf['localDataPath'])\n print(('local files:', files))\n if len(files) == 0:\n return \n\n for num in range(0, len(files)):\n if files[num] == 'failed':\n continue\n try:\n rfile = os.path.join(self.conf['localDataPath'], files[num])\n data = self.toData(rfile)\n if len(data) == 0:\n continue\n for dnum in range(0, len(data)):\n self.settleRecord(data[dnum])\n os.remove(rfile)\n except Exception as e:\n print(('local settlement fail:', rfile))\n faileddir = os.path.join(self.conf['localDataPath'], 'failed')\n if not os.path.exists(faileddir):\n os.makedirs(faileddir)\n os.rename(rfile, os.path.join(faileddir, files[num]))\n traceback.print_exc()\n\n\n def callback(self):\n try:\n self.conn = conn(self.config['db'])\n self.localSettlement()\n except Exception as e:\n traceback.print_exc()\n finally:\n self.conn.close()\n\n try:\n self.conn = conn(self.config['db'])\n self.settlement()\n except Exception as e:\n traceback.print_exc()\n finally:\n self.conn.close()\n","sub_path":"task/Settlement.py","file_name":"Settlement.py","file_ext":"py","file_size_in_byte":7510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"420397835","text":"import dash\nfrom dash.dependencies import Output, Input, State\nimport dash_core_components as dcc\nimport dash_bootstrap_components as dbc\nimport dash_html_components as html\nimport dash_daq as daq\n\nimport plotly\nimport random\nimport plotly.graph_objs as go\nfrom collections import deque\nimport sqlite3\nimport pandas as pd\nimport numpy as np\n\n#conn = sqlite3.connect('twitter.db')\n#c = conn.cursor()\n# external CSS stylesheets\nexternal_stylesheets = [\n 'https://codepen.io/chriddyp/pen/bWLwgP.css',\n {\n 'href': \"https://stackpath.bootstrapcdn.com/bootstrap/3.4.1/css/bootstrap.min.css\",\n 'rel': 'stylesheet',\n 'integrity': \"sha384-HSMxcRTRxnN+Bdg0JdbxYKrThecOKuH5zCYotlSAcp1+c8xmyTe9GYg1l9a69psu\",\n 'crossorigin': 'anonymous'\n }\n]\n# external JavaScript files\nexternal_scripts = [\n 'https://www.google-analytics.com/analytics.js',\n {'src': 'https://cdn.polyfill.io/v2/polyfill.min.js'},\n {\n 'src': 'https://cdnjs.cloudflare.com/ajax/libs/lodash.js/4.17.10/lodash.core.js',\n 'integrity': 'sha256-Qqd/EfdABZUcAxjOkMi8eGEivtdTkh3b65xCZL4qAQA=',\n 'crossorigin': 'anonymous'\n }\n]\n\napp = dash.Dash(__name__, update_title='Coletando tweets...', \n external_stylesheets=[dbc.themes.BOOTSTRAP]\n #external_scripts=external_scripts\n )\napp.layout = dbc.Container(\n [ html.H1('Bolsometro'),\n html.Hr(),\n\n dbc.NavbarSimple(\n children=[\n #dbc.NavItem(dbc.NavLink(\"Page 1\", href=\"#\")),\n dbc.DropdownMenu(\n children=[\n dbc.DropdownMenuItem(\"Redes Sociais\", header=True),\n dbc.DropdownMenuItem(\"Linkedin\", href=\"https://www.linkedin.com/in/neto-figueira/\"),\n dbc.DropdownMenuItem(\"Github\", href=\"https://github.com/netofigueira\"),\n ],\n nav=True,\n in_navbar=True,\n label=\"Contato\",\n ),\n ],\n brand=\"Tendência de Sentimento no Twitter\",\n brand_href=\"#\",\n color=\"dark\",\n dark=True,\n ),\n \n\n\n\n dbc.Col([ \n dbc.Col(dcc.Graph(id='live-graph', animate=True ) ), \n \n ]),\n###\n # dbc.Row([ \n # \n # dbc.Col(html.Div(dcc.Graph(id='my-gauge', animate=True)) ),\n\n # ]),\n dcc.Interval(\n id='graph-update',\n interval= 800,\n ),\n dcc.Interval(\n id='my-gauge-update',\n interval= 800,\n ), \n\n html.Div(html.H2('Tweets recentes ao vivo')),\n html.Hr(),\n html.Div(className='row', children=[html.Div(id=\"recent-tweets-table\")] ),\n\n\n dcc.Interval(\n id='recent-table-update',\n interval= 4*10**3,\n #n_intervals=2\n ),\n ]\n)\ndef color_select(d):\n if d > 5:\n return \"#95D2EC\"\n if d < -5:\n return \"#FF4242\"\n else:\n return \"#FFFFFF\"\n\ndef generate_table(df, max_rows=10):\n return html.Table(className=\"table\",\n children=[\n html.Thead(\n html.Tr(\n children=[\n html.Th(col.title()) for col in df.columns.values],\n \n )\n ),\n html.Tbody(\n [\n \n html.Tr(\n children=[\n html.Td(data) for data in d\n ], style={'background-color':color_select(d[1])}\n )\n for d in df.values.tolist()])\n ]\n )\n\n\n\"\"\"\n@app.callback(\n Output(\"popover\", \"is_open\"),\n [Input(\"popover-target\", \"n_clicks\")],\n [State(\"popover\", \"is_open\")],\n)\ndef toggle_popover(n, is_open):\n if n:\n return not is_open\n return is_open\n\n@app.callback(Output('my-gauge', 'figure'),\n [Input('my-gauge-update', 'n_intervals')])\n\ndef update_gauge(input_data):\n try:\n conn = sqlite3.connect('twitter.db')\n df = pd.read_sql(\"SELECT * FROM sentiment ORDER BY unix DESC LIMIT 500\", conn)\n df.sort_values('unix', inplace=True)\n s_array = df.sentiment.values\n df['sentiment'] = np.interp(s_array, (s_array.min(), s_array.max()), (-10,10) )\n\n neg = df[(df.sentiment < 0)].sentiment.mean()\n\n df['sentiment_smoothed'] = df['sentiment'].rolling(int(len(df)/100)).mean()\n\n fig = go.Indicator(\n mode = \"gauge+number\",\n value = neg,\n domain = {'x': [0, 1], 'y': [0, 1]},\n title = {'text': \"Bolsometro\"},\n #delta = {'reference': 8, 'increasing': {'color': \"Green\"}},\n gauge = {\n 'axis': {'range': [0, -10], 'tickwidth': 1, 'tickcolor': \"#EF553B\"},\n 'bar': {'color': \"#453938\"},\n 'bgcolor': \"white\",\n 'borderwidth': 2,\n 'bordercolor': \"gray\",\n 'steps': [\n {'range': [0, -5], 'color': 'white'},\n {'range': [-5, -6.5], 'color': '#ffb0a8'},\n {'range': [-6.5, -8], 'color': '#EF553B'}],\n 'threshold': {\n 'line': {'color': \"red\", 'width': 4},\n 'thickness': 0.75,\n 'value': 9}}\n )\n return {'data': [fig]}\n\n except Exception as e:\n with open('errors.txt','a') as f:\n f.write(str(e))\n f.write('\\n')\n\"\"\"\n \n@app.callback(Output('live-graph', 'figure'),\n [Input('graph-update', 'n_intervals')])\ndef update_graph_scatter(input_data):\n try:\n conn = sqlite3.connect('twitter.db')\n c = conn.cursor()\n df = pd.read_sql(\"SELECT * FROM sentiment ORDER BY unix DESC LIMIT 2000\", conn)\n df.sort_values('unix', inplace=True)\n\n s_array = df.sentiment.values\n df['sentiment'] = np.interp(s_array, (s_array.min(), s_array.max()), (-10,10) )\n\n df['sentiment_smoothed'] = df['sentiment'].rolling(int(len(df)/500)).mean()\n df.dropna(inplace=True)\n df['date'] = pd.to_datetime(df['unix'],unit='ms')\n # converting to são paulo time. \n df['date'] = df.date.dt.tz_localize('UTC').dt.tz_convert('America/Sao_Paulo')\n df.set_index('date', inplace=True)\n\n df_bolso = df[df.tweet.str.contains('bolsonaro', case=False)]\n\n df_lula = df[df.tweet.str.contains('lula', case=False)]\n\n #df = df.resample('10s').mean()\n df_bolso = df_bolso.resample('60s').mean()\n #X = df.index\n \n \n X = df_bolso.index\n\n Y= df_bolso.sentiment_smoothed.round(decimals=2)\n Y2 = df_lula.sentiment_smoothed.round(decimals=2)\n\n data = plotly.graph_objs.Scatter(\n x=X,\n y=Y,\n name='Bolsonaro',\n mode= 'lines',\n line = dict(color = 'green')\n )\n\n\n data2 = plotly.graph_objs.Scatter(\n x=X,\n y=Y2,\n name='Lula',\n mode = 'lines',\n line = dict(color = 'red')\n \n )\n\n return {'data': [data, data2],'layout' : go.Layout(xaxis=dict(range=[min(X),max(X)]),\n yaxis=dict(range=[-5,5]),)}\n\n except Exception as e:\n with open('errors.txt','a') as f:\n f.write(str(e))\n f.write('\\n')\n\n@app.callback(Output('recent-tweets-table', 'children'),\n [Input('recent-table-update', 'n_intervals')],\n ) \ndef update_recent_tweets(input_data):\n conn = sqlite3.connect('twitter.db')\n c = conn.cursor()\n\n df = pd.read_sql(\"SELECT * FROM sentiment ORDER BY unix DESC LIMIT 10\", conn)\n\n df['date'] = pd.to_datetime(df['unix'], unit='ms')\n s_array = df.sentiment.values\n df['sentiment'] = np.interp(s_array, (s_array.min(), s_array.max()), (-10,10) )\n df['sentiment'] = df.sentiment.round(2)\n df = df.drop(['unix'], axis=1)\n df = df[['date','tweet','sentiment']]\n # converting to são paulo time. \n df['date'] = df.date.dt.tz_localize('UTC').dt.tz_convert('America/Sao_Paulo')\n df.set_index('date', inplace=True)\n return generate_table(df, max_rows=10)\n\n\n#external_css = [\"https://stackpath.bootstrapcdn.com/bootstrap/3.4.1/css/bootstrap.min.css\"]\n#for css in external_css:\n# app.css.append_css({\"external_url\": css})\n\n\n\nif __name__ == '__main__':\n app.run_server(debug=True)","sub_path":"dash_bootstrap.py","file_name":"dash_bootstrap.py","file_ext":"py","file_size_in_byte":9091,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"102529090","text":"import cv2\r\nimport numpy as np\r\nimport os\r\nimport urllib.request\r\n\r\ndef getImages(link):\r\n\r\n\turls = urllib.request.urlopen(link).read().decode()\r\n\tpic_num = 1\r\n \r\n\tfor pic in urls.split('\\n'):\r\n\t\ttry:\r\n\t\t\turllib.request.urlretrieve(pic, 'people/ppl ({}).png'.format(pic_num))\r\n\t\t\tpic_num += 1\r\n\r\n\t\texcept Exception as e:\r\n\t\t\tprint(str(e))\r\n\r\n\r\nif __name__ == '__main__' :\r\n\r\n\t############## BUDYNKI\r\n\t#link = 'http://www.image-net.org/api/text/imagenet.synset.geturls?wnid=n04413969'\r\n\t############# LUDZIE\r\n\tlink = 'http://www.image-net.org/api/text/imagenet.synset.geturls?wnid=n07942152'\r\n\tgetImages(link)\r\n","sub_path":"getimg.py","file_name":"getimg.py","file_ext":"py","file_size_in_byte":610,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"49222613","text":"#!/usr/bin/env python\n\n# 2014-11-15\n# tarte.py\n# Public Domain\n\nimport time\n\nimport pigpio\n\nLED1=21\nLED2=26\nLED3=12\nLED4=6\n\n\"\"\"\n7.5 10 12 16\n\nFind the number of cycles needed for an integral switch on/off for each LED.\n\nThat's 2*7.5 which is 2 seconds worth.\n\nIn 2 seconds there will be this many cycles of on/off\n\n15 20 24 32\n\nHow many micros for each cycle?\n\n15 66666 on 66667 off = 1999995\n20 50000 on 50000 off = 2000000\n24 41666 on 41667 off = 1999992\n32 31250 on 31250 off = 2000000\n\nThere will be a slight error which will not be detectable by most means.\n\"\"\"\n\ndef wave(pi, gpio, hz, secs, on=1, offset=0):\n \"\"\"\n Generate a hz cycles per second square wave on gpio for\n secs seconds. The first transition is to level on at\n offset microseconds from the start.\n \"\"\"\n micros_left = int(secs * 1000000)\n transitions = int(2 * hz * secs)\n micros = micros_left / transitions\n\n if (offset < 0) or (offset > micros):\n print(\"Illegal offset {} for hz {}\".format(offset, hz))\n exit()\n\n wf = [] # Empty waveform.\n\n if offset:\n wf.append(pigpio.pulse(0, 0, offset))\n micros_left -= micros\n last_micros = micros - offset\n transitions -= 1\n\n for t in range(transitions, 0, -1):\n micros = micros_left / t\n if (t & 1) == (on & 1):\n wf.append(pigpio.pulse(0, 1< 0 :\n k = k + 1\n alist = aline.split('\t') #取出列表中的一项,并赋值给新的列表\n #print(alist[3],lines.index(aline))\n z = float(alist[3]) + z #统计持续时间\n \n if k != 0 :\n print(y,':共',k,'条记录,耗时','%.2f' % z,'小时')\n return [k,'%.2f' % z]\n\n\ndef tongjiFile(path_checkDateFile):\n \"\"\"\n 功能:每个维度都统计时间开销,并输出到一个文件中。\n \"\"\"\n\n with open(tongjiRecord,'at', encoding='utf-8') as fnew:\n for aweed in weedlist:\n tu_sum = sumWeed(aweed,path_checkDateFile)\n tjr = str(aweed) + '\\t' + str(tu_sum[0]) + '\\t' + str(tu_sum[1]) + '\\n'\n fnew.write(tjr)\n tu_sum = []\n print('结果已写入文件:',fnew)\n\ndef everydaySumWeed():\n \"\"\"\n 功能:按天统计数据。\n \"\"\"\n\n yourdate = str(input(\"请输入想要统计的日期\\n格式为20190121\\n\"))\n \n #查找有没有该文件\n yy=yourdate[:4]\n mm=yourdate[4:6]\n dd=yourdate[6:]\n if len(yy) < 2 :\n yy = \"0\" + str(yy)\n if len(mm) < 2 :\n mm = \"0\" + str(mm)\n if len(dd) < 2 :\n dd = \"0\" + str(dd) \n newfilename = str(yy)+str(mm)+str(dd)\n pathNew = path_Clear + 'everyday' + newfilename + '.txt'\n print(pathNew)\n if os.path.exists(pathNew):\n tongjiFile(pathNew)\n else:\n print('未查找该日期的数据')\n\nmain()\n","sub_path":"timebill.py","file_name":"timebill.py","file_ext":"py","file_size_in_byte":8557,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"149940740","text":"from helpers import *\n\nfrom mobject.tex_mobject import TexMobject\nfrom mobject import Mobject\nfrom mobject.image_mobject import ImageMobject\nfrom mobject.vectorized_mobject import *\n\nfrom animation.animation import Animation\nfrom animation.transform import *\nfrom animation.simple_animations import *\nfrom animation.playground import *\nfrom topics.geometry import *\nfrom topics.characters import *\nfrom topics.functions import *\nfrom topics.fractals import *\nfrom topics.number_line import *\nfrom topics.combinatorics import *\nfrom topics.numerals import *\nfrom topics.three_dimensions import *\nfrom topics.objects import *\nfrom scene import Scene\nfrom scene.zoomed_scene import ZoomedScene\nfrom scene.reconfigurable_scene import ReconfigurableScene\nfrom camera import Camera\nfrom mobject.svg_mobject import *\nfrom mobject.tex_mobject import *\n\nclass ConfettiSpiril(Animation):\n CONFIG = {\n \"x_start\" : 0,\n \"spiril_radius\" : 1,\n \"num_spirils\" : 4,\n \"run_time\" : 10,\n \"rate_func\" : None,\n }\n def __init__(self, mobject, **kwargs):\n digest_config(self, kwargs)\n mobject.next_to(self.x_start*RIGHT + SPACE_HEIGHT*UP, UP)\n self.total_vert_shift = \\\n 2*SPACE_HEIGHT + mobject.get_height() + 2*MED_SMALL_BUFF\n \n Animation.__init__(self, mobject, **kwargs)\n\n def update_submobject(self, submobject, starting_submobject, alpha):\n submobject.points = np.array(starting_submobject.points)\n\n def update_mobject(self, alpha):\n Animation.update_mobject(self, alpha)\n angle = alpha*self.num_spirils*2*np.pi\n vert_shift = alpha*self.total_vert_shift\n\n start_center = self.mobject.get_center()\n self.mobject.shift(self.spiril_radius*OUT)\n self.mobject.rotate(angle, axis = UP, about_point = start_center)\n self.mobject.shift(vert_shift*DOWN)\n\nclass Anniversary(TeacherStudentsScene):\n CONFIG = {\n \"num_confetti_squares\" : 50,\n }\n def construct(self):\n self.celebrate()\n self.complain()\n\n def celebrate(self):\n title = TextMobject(\"2 year Anniversary!\")\n title.scale(1.5)\n title.to_edge(UP)\n\n first_video = Rectangle(\n height = 2, width = 2*(16.0/9),\n stroke_color = WHITE,\n fill_color = \"#111111\",\n fill_opacity = 0.75,\n )\n first_video.next_to(self.get_teacher(), UP+LEFT)\n first_video.shift(RIGHT)\n formula = TexMobject(\"e^{\\\\pi i} = -1\")\n formula.move_to(first_video)\n first_video.add(formula)\n\n hats = self.get_party_hats()\n confetti_spirils = self.get_confetti_animations()\n self.play(\n Write(title, run_time = 2),\n *[\n ApplyMethod(pi.change_mode, \"hooray\")\n for pi in self.get_pi_creatures()\n ]\n )\n self.play(\n DrawBorderThenFill(\n hats,\n submobject_mode = \"lagged_start\",\n rate_func = None,\n run_time = 2,\n ),\n *confetti_spirils + [\n Succession(\n Animation(pi, run_time = 2),\n ApplyMethod(pi.look, UP+LEFT),\n ApplyMethod(pi.look, UP+RIGHT),\n Animation(pi),\n ApplyMethod(pi.look_at, first_video),\n rate_func = None\n )\n for pi in self.get_students()\n ] + [\n Succession(\n Animation(self.get_teacher(), run_time = 2),\n Blink(self.get_teacher()),\n Animation(self.get_teacher(), run_time = 2),\n ApplyMethod(self.get_teacher().change_mode, \"raise_right_hand\"),\n rate_func = None\n ),\n DrawBorderThenFill(\n first_video, \n run_time = 10,\n rate_func = squish_rate_func(smooth, 0.5, 0.7)\n )\n ]\n )\n self.change_student_modes(*[\"confused\"]*3)\n\n def complain(self):\n self.student_says(\n \"Why are you \\\\\\\\ talking so fast?\",\n student_index = 0,\n target_mode = \"sassy\",\n )\n self.change_student_modes(*[\"sassy\"]*3)\n self.play(self.get_teacher().change_mode, \"guilty\")\n self.dither(2)\n\n def get_party_hats(self):\n hats = VGroup(*[\n PartyHat(\n pi_creature = pi,\n height = 0.5*pi.get_height()\n )\n for pi in self.get_pi_creatures()\n ])\n max_angle = np.pi/6\n for hat in hats:\n hat.rotate(\n random.random()*2*max_angle - max_angle,\n about_point = hat.get_bottom()\n )\n return hats\n\n def get_confetti_animations(self):\n colors = [RED, YELLOW, GREEN, BLUE, PURPLE, RED]\n confetti_squares = [\n Square(\n side_length = 0.2,\n stroke_width = 0,\n fill_opacity = 0.5,\n fill_color = random.choice(colors),\n )\n for x in range(self.num_confetti_squares)\n ]\n confetti_spirils = [\n ConfettiSpiril(\n square,\n x_start = 2*random.random()*SPACE_WIDTH - SPACE_WIDTH,\n rate_func = squish_rate_func(lambda t : t, a, a+0.5)\n )\n for a, square in zip(\n np.linspace(0, 0.5, self.num_confetti_squares),\n confetti_squares\n )\n ]\n return confetti_spirils\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"efvgt.py","file_name":"efvgt.py","file_ext":"py","file_size_in_byte":5727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"485682946","text":"#\n# Chess knight problem\n# lets try 8 x 8 sq\n#\n# X\n# 0 1 2 3 4 5 6 7\n# 8 9 10 11 12 13 14 15\n# 16 17 18 19 20 21 22 23\n# Y 24 25 26 _27_ 28 29 30 31\n# 32 33 34 35 36 37 38 39\n# 40 41 42 43 44 45 46 47\n# 48 49 50 51 52 53 54 55\n# 56 57 58 59 60 61 62 63\n\n\ndef pprint (M):\n for i in range(len(M)):\n if i%8 == 0:\n print (\"\")\n print (\"{:3}\".format(M[i]),end='')\n\n\ndef get_moves (pos, M):\n \"\"\" return possible moves \"\"\"\n y,x = divmod(pos, 8)\n moves=[-17, -15, -10, -6, +6, +10, +15, +17]\n list=[]\n # 1: if pos + move is > 63 or < 0, don't allow\n for i in moves:\n # 2: check for moves across a board boundary (not allowed)\n if (pos + i >= 0) and (pos+i <= 63):\n ynew,xnew=divmod(pos+i, 8)\n if abs(xnew-x) <= 2:\n # 3: check that the proposed position has not yet been occupied\n if M[pos+i] == 0:\n list.append (pos+i)\n list.sort()\n list.reverse()\n return (list)\n\n\ndef ulist(M, pos,v):\n \"\"\" append a value to M and return \"\"\"\n import copy\n list= copy.deepcopy(M)\n list[pos]=v\n return list\n \n\ndef solve (M, cpos, move):\n \"\"\" solve the question recursively \"\"\" \n if move == 64:\n print (\"\\n\\nmove: \", move)\n print (\"sum: \", sum(M))\n pprint (M)\n #exit()\n for next in get_moves(cpos, M):\n solve(ulist(M, next, move+1), next, move+1)\n\n\nif __name__ == '__main__':\n M=[0]*64\n M[2]=1\n solve(M, cpos=2, move=1)\n\n","sub_path":"chess.py","file_name":"chess.py","file_ext":"py","file_size_in_byte":1432,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"582133235","text":"#!/usr/bin/env python\n# Determine what databases to backup and do it\n\nimport os, sys\nimport re\nimport subprocess\nimport MySQLdb\nimport optparse\n\nparser = optparse.OptionParser()\nparser.add_option('-v', '--verbose', action=\"store_true\",\n help = \"Enable verbose output\",\n dest = \"verbose\",\n default = False)\nparser.add_option('-s', '--secrets-file', action=\"store\",\n help = \"Secrets file containing mysql login (%default)\",\n dest = \"secret\",\n default = \"/root/.my.cnf\")\nparser.add_option('-e', '--exclude-file', action=\"store\",\n help = \"File of databases to exclude from backup, one per line (%default)\",\n dest = \"exclude\",\n default = \"/etc/copy-db.exclude\")\nparser.add_option('-d', '--backup-dir', action=\"store\",\n help = \"Directory to store backup (%default)\",\n dest = \"backup_dir\",\n default = \"/var/lib/mysql-backup\")\nparser.add_option('-n', '--no-hot-copy', action=\"store_true\",\n help = \"Never backup using mysqlhotcopy\",\n dest = \"disable_hotcopy\",\n default = False)\n\ndo_verbose = False\n(options,args) = parser.parse_args()\n\ndo_verbose = options.verbose\nsecret_file = options.secret\nexclude_file = options.exclude\nbackup_dir = options.backup_dir\ndisable_hotcopy = options.disable_hotcopy\n\ndef verbose(s):\n if do_verbose:\n print >>sys.stderr, s\n\n# we connect multiple times to avoid leaving an idle connection open\n# while we do long dump operations; it might get timed out\ndef connect_to_db():\n return MySQLdb.connect(host=\"localhost\",\n user=\"root\",\n read_default_file=secret_file)\n\ndbs = [] # Databases on the machine, got from MySQL\nuidbs = {} # Databases not to be backed up, read from copy-db.exclude\n\n# Get available DBs list\nconn = connect_to_db()\nconn.set_character_set(\"utf8\")\ncursor = conn.cursor()\ncursor.execute(\"SHOW DATABASES\")\nfor fields in cursor.fetchall():\n dbs.append(unicode(fields[0], \"utf8\"))\ncursor.close()\nconn.close()\n\n# Get not-to-be-backed-up list\ntry:\n list = open(exclude_file)\nexcept IOError:\n list = []\nfor line in list:\n if not line.startswith ('#'):\n dbname, who, when = line.strip ().split ()\n uidbs[dbname] = (who, when)\n\n# Spit warnings and remove not-to-be-backed-up databases from the list\nfor i in uidbs:\n if i not in dbs:\n sys.stderr.write('WARNING: redundant entry for database %s in %s\\n\\n' % (i, exclude_file))\n else:\n verbose ('database %s not being backed up (request by %s on %s)' % (i, uidbs[i][0], uidbs[i][1]))\n dbs.remove (i)\n\n# Turn a database name into a filename. What we consider\n# filename-safe is the same MySQL, but the encoding of non-safe\n# characters differs. MySQL has tables for some non-ASCII unicode -\n# e.g. U+00C0 LATIN CHARACTER CAPITAL LETTER A WITH ACUTE is @0G\n# then it uses @xxxx for the rest. We use @xxxx for everything.\n# We don't actually need a match with what MySQL does, just\n# something that won't contain meta-characters like '/', but matching\n# up for ASCII names like 'db_backup' is slightly useful\ndef encode_as_filename(s):\n return re.sub('[^A-Za-z0-9]', escape_match, s)\n\ndef escape_match(m):\n o = ord(m.group(0))\n if o < 0x10000:\n return \"@%04x\" % o\n else:\n return \"@%04x@%04x\" % (0xd800 + (o / 1024), 0xdc00 + (o % 1024))\n\n# Backup!\nfor db in dbs:\n # mysqlhotcopy only works for MyISAM and ARCHIVE tables. If a database has\n # only tables of those types, then we use mysqlhotcopy.\n #\n # For InnoDB tables we can use mysqldump --single-transaction to get a\n # consistent snapshot of the database.\n #\n # For tables with a mixture of InnoDB and MyISAM tables, neither of the\n # above methods will work and give a consistent snapshot. We could\n # use 'mysqldump --lock-tables', but that would keep the entire database\n # locked for the entire length of the dump. Instead we assume that in\n # this case, the application doesn't care much about the consistentcy\n # of the MyISAM tables and use --single-transaction anyways. (This is the\n # right thing to do for bugzilla where everything but the bugs_fulltext\n # table is InnoDB. bugs_fulltext is just a mirror of the other tables for\n # searching purposes.)\n #\n # Note that mysqlhotcopy is not necessarily faster than mysqldump - the\n # compressed dump will typically be much smaller and faster to write to\n # disk than the copy. The hot copy, on the other hand, may be more rsync\n # friendly when we rsync the databases to the backup machine (This theory\n # is untested.)\n #\n # Future enhancement would be to extent copy-db.exclude to allow specifying\n # per-database backup methods.\n\n can_hotcopy = True\n\n db_filename = encode_as_filename(db)\n if db_filename != db:\n # mysqlhotcopy doesn't understand encoded database names\n can_hotcopy = False\n if disable_hotcopy:\n can_hotcopy = False\n\n # Figure out what types of tables the database has\n conn = connect_to_db()\n conn.set_character_set(\"utf8\")\n conn.select_db(db.encode(\"utf8\"))\n cursor = conn.cursor()\n cursor.execute(\"SHOW TABLE STATUS\")\n for fields in cursor.fetchall():\n engine = fields[1]\n if engine != 'MyISAM' and engine != 'ARCHIVE':\n can_hotcopy = False\n cursor.close()\n conn.close()\n\n if can_hotcopy:\n verbose(\"Backing up %s via mysqlhotcopy\"% db)\n hotcopy = subprocess.Popen(['mysqlhotcopy', '--quiet', '--allowold', db, backup_dir])\n hotcopy.wait()\n else:\n verbose(\"Backing up %s via mysqldump\" % db)\n outfilename = os.path.join(backup_dir, db_filename + \".dump.gz\")\n outfilename_tmp = outfilename + \".tmp\"\n\n # Add a bit of error checking before freaking out\n if not os.path.exists(backup_dir):\n sys.stderr.write(\"ERROR: '%s' does not exist to backup files into\\n\" % backup_dir)\n sys.exit(1)\n if not os.access(backup_dir, os.W_OK):\n sys.stderr.write(\"ERROR: '%s' is not writable\\n\" % backup_dir)\n sys.exit(1)\n\n outfile = open(outfilename_tmp, \"w\")\n dump = subprocess.Popen(['mysqldump',\n '--single-transaction',\n '--default-character-set=utf8',\n db.encode(\"utf8\")],\n stdout=subprocess.PIPE)\n gzip = subprocess.Popen(['gzip', '-c'],\n stdin=dump.stdout, stdout=outfile)\n dump.wait()\n gzip.wait()\n outfile.close()\n if dump.returncode == 0 and gzip.returncode == 0:\n os.rename(outfilename_tmp, outfilename)\n else:\n print >>sys.stderr, \"Failed to back up %s, leaving old backup\" % db\n os.remove(outfilename_tmp)\n","sub_path":"copy-db.py","file_name":"copy-db.py","file_ext":"py","file_size_in_byte":7023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"26865490","text":"from .voc_converter import VocConverter\nfrom .voc_to_sa_pixel import voc_instance_segmentation_to_sa_pixel\nfrom .voc_to_sa_vector import voc_object_detection_to_sa_vector\n\n\nclass VocObjectDetectionStrategy(VocConverter):\n name = \"ObjectDetection converter\"\n\n def __init__(\n self, dataset_name, export_root, project_type, output_dir, task,\n direction\n ):\n self.direction = direction\n super().__init__(\n dataset_name, export_root, project_type, output_dir, task\n )\n\n self.__setup_conversion_algorithm()\n\n def __setup_conversion_algorithm(self):\n if self.direction == \"to\":\n raise NotImplementedError(\"Doesn't support yet\")\n else:\n if self.project_type == \"Vector\":\n if self.task == \"object_detection\":\n self.conversion_algorithm = voc_object_detection_to_sa_vector\n elif self.task == \"instance_segmentation\":\n raise NotImplementedError(\"Doesn't support yet\")\n elif self.project_type == \"Pixel\":\n if self.task == \"object_detection\":\n raise NotImplementedError(\"Doesn't support yet\")\n elif self.task == \"instance_segmentation\":\n self.conversion_algorithm = voc_instance_segmentation_to_sa_pixel\n\n def __str__(self):\n return '{} object'.format(self.name)\n\n def from_sa_format(self):\n pass\n\n def to_sa_format(self):\n loader = self.conversion_algorithm(self.export_root, self.output_dir)\n","sub_path":"superannotate/input_converters/converters/voc_converters/voc_strategies.py","file_name":"voc_strategies.py","file_ext":"py","file_size_in_byte":1559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"225026666","text":"_payload = [(1, 28), (2, 14), (3, 9), (4, 7), (5, 5), (7, 4), (9, 3), (14, 2), (28, 1)]\n\n\ndef cnt_bits(x):\n i = 0\n while (x >> i) > 0:\n i += 1\n return i\n\n\ndef simple9_encode_lst(lst):\n res = []\n i = 0\n while i < len(lst):\n j = i\n m = cnt_bits(lst[j])\n while j < len(lst) and (j - i + 1) * m <= 28:\n j += 1\n if j < len(lst):\n m = max(m, cnt_bits(lst[j]))\n num = simple9_encode(lst[i:j])\n res.append(num)\n i = j\n return res\n\n\ndef simple9_encode(x):\n n = len(x)\n x.reverse()\n res = 0\n res += (n << 28)\n m = 28 // n\n for i in range(n):\n res += (x[i] << (i * m + 28 % n))\n return res\n\n\ndef simple9_decode_lst(x):\n res = []\n for num in x:\n res += simple9_decode(num)\n return res\n\n\ndef simple9_decode(x):\n res = []\n n = x >> 28\n m = 28 // n\n for i in range(n):\n ll = (28 - i * m)\n cur = (x & ((1 << ll) - 1)) >> (ll - m)\n res.append(cur)\n return res\n\n\ndef x_to_str(x):\n s = ''\n for i in range(32):\n if (x & (1 << i)) > 0:\n s += '1'\n else:\n s += '0'\n i += 1\n return s[::-1]\n","sub_path":"indexes/simple9.py","file_name":"simple9.py","file_ext":"py","file_size_in_byte":1205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"226556395","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport glob\nimport re\n\ndef count_key_words(key_words):\n contents = []\n \n for file_name in glob.glob('*.txt'):\n with open(file_name, 'r') as file:\n contents.append(file.read())\n \n str = ''.join(contents).lower()\n return dict(map(lambda x:(x, len(re.findall(x, str))), key_words))\n \nif __name__ == '__main__':\n print(count_key_words(['byte', 'type']))","sub_path":"006/count_key_words.py","file_name":"count_key_words.py","file_ext":"py","file_size_in_byte":447,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"329098198","text":"import os\nimport random\nimport numpy as np\nimport tensorflow as tf\nimport matplotlib.pyplot as plt\n\nFEATURESMAP = {\n #1: ['weight0', 'height0', 'waistline0','hipline0','chest0','thigh0','arm0'],\n 1: ['weight0', 'height0', 'waistline0'],\n 2: ['weight1', 'height0', 'waistline1'],\n 3: ['weight2', 'height0', 'waistline2'],\n 4: ['weight3', 'height0', 'waistline3']\n}\n\ndef processData(filePath):\n\n dataList = []\n\n with open(filePath, 'r') as f:\n records = f.readlines()\n\n # 读第一行获取每一列的列名\n # strip() 去掉行尾的换行符\n keys = records[0].strip().split(',')\n\n for i, record in enumerate(records):\n if i > 0:\n dic = {}\n values = record.strip().split(',')\n for index, key in enumerate(keys):\n dic[key] = float(values[index])\n\n dataList.append(dic)\n\n return dataList\n\ndef generateDataAndLabel(type, metaDatas, week):\n LABELMAP = {}\n if type == 'weight':\n LABELMAP = {\n 1: 'deltaWeightAll',\n 2: 'deltaWeight2',\n 3: 'deltaWeight3',\n 4: 'deltaWeight4'\n }\n elif type == 'waist':\n LABELMAP = {\n 1: 'deltaWaistAll',\n 2: 'deltaWaist2',\n 3: 'deltaWaist3',\n 4: 'deltaWaist4'\n }\n trainDatas = []\n trainLabels = []\n\n useableKeys = FEATURESMAP[week]\n\n for metaData in metaDatas:\n trainData = []\n for key in useableKeys:\n trainData.append(metaData[key])\n deltaWeight = metaData[LABELMAP[week]]\n\n\n trainDatas.append(trainData)\n trainLabels.append(deltaWeight)\n\n return (trainDatas, trainLabels)\n\n\n# Normalize by column (min-max norm)\ndef normalize_cols(m):\n col_max = m.max(axis=0)\n col_min = m.min(axis=0)\n return (m-col_min) / (col_max - col_min)\n\n\n#[{ all values of a person },{}...]\ntrainDataList = processData(filePath='/home/shen/Trying/Predict/up/t1/more_dimension/more_dimension_data/train.csv')\n\n(trainDatas, trainLabels) = generateDataAndLabel(type='weight', metaDatas=trainDataList, week=1)\n\ntrainDatas = np.array(trainDatas)\ntrainLabels = np.array(trainLabels)\ntrainLabels /= 10\n\nx_vals_train = np.nan_to_num(normalize_cols(trainDatas))\n\n\n\n\ntestDataList = processData(filePath='/home/shen/Trying/Predict/up/t1/more_dimension/more_dimension_data/test.csv')\n\n(testDatas, testLabels) = generateDataAndLabel(type='weight', metaDatas=testDataList, week=1)\ntestDatas = np.array(testDatas)\ntestLabels = np.array(testLabels)\ntestLabels /= 10\n\nx_vals_test = np.nan_to_num(normalize_cols(testDatas))\n\n# Create graph session\nsess = tf.Session()\n\n\n\n# Declare batch size\n#batch_size = 3000\nbatch_size = 4500\n\n# Define Variable Functions (weights and bias)\ndef init_weight(shape, st_dev):\n weight = tf.Variable(tf.random_normal(shape, stddev=st_dev))\n return weight\n\n\ndef init_bias(shape, st_dev):\n bias = tf.Variable(tf.random_normal(shape, stddev=st_dev))\n return bias\n\n# Create a fully connected layer:\ndef fully_connected(input_layer, weights, biases):\n layer = tf.add(tf.matmul(input_layer, weights), biases)\n return tf.nn.relu(layer)\n\n# Initialize placeholders\n#x_data = tf.placeholder(shape=[None, 3], dtype=tf.float32)\nx_data = tf.placeholder(shape=[None, 1], dtype=tf.float32)\ny_target = tf.placeholder(shape=[None, 1], dtype=tf.float32)\n\n\nweight_1 = init_weight(shape=[1, 10], st_dev=1.0)\nbias_1 = init_bias(shape=[10], st_dev=1.0)\nlayer_1 = fully_connected(x_data, weight_1, bias_1)\n\n\nweight_2 = init_weight(shape=[10, 10], st_dev=1.0)\nbias_2 = init_bias(shape=[10], st_dev=1.0)\nlayer_2 = fully_connected(layer_1, weight_2, bias_2)\n\n\n\nweight_3 = init_weight(shape=[10, 10], st_dev=1.0)\nbias_3 = init_bias(shape=[10], st_dev=1.0)\nlayer_3 = fully_connected(layer_2, weight_3, bias_3)\n\n\n\nweight_4 = init_weight(shape=[10, 8], st_dev=1.0)\nbias_4 = init_bias(shape=[8], st_dev=1.0)\nlayer_4 = fully_connected(layer_3, weight_4, bias_4)\n\nweight_5 = init_weight(shape=[8,1], st_dev=1.0)\nbias_5 = init_bias(shape=[1], st_dev=1.0)\n\nfinal_output = tf.sigmoid(tf.add(tf.matmul(layer_4, weight_5), bias_5))\n\n# Declare loss function (MSE)\nloss = tf.reduce_mean(tf.square(y_target - final_output))\n\n\n# This is caculate the accuracy\nTemp =tf.abs( tf.subtract(final_output ,y_target))\naccuracy = tf.reduce_mean( tf.cast( tf.less(Temp ,0.12) ,tf.float32))\n\n\n# Declare optimizer\n# my_opt = tf.train.GradientDescentOptimizer(0.005)\n# train_step = my_opt.minimize(loss)\n\nglobal_step = tf.Variable(0, trainable=False)\nstarter_learning_rate = 0.005\nlearning_rate = tf.train.exponential_decay(starter_learning_rate, global_step,40000, 0.96, staircase=True)\noptimizer = tf.train.GradientDescentOptimizer(learning_rate)\ntrain_step = optimizer.minimize(loss, global_step=global_step)\n\n\n# Initialize variables\ninit = tf.global_variables_initializer()\nsess.run(init)\n\n# # Training loop\n\nsaver = tf.train.Saver() # defaults to saving all variables - in this case w and b\n\nloss_vec = []\ntest_loss = []\nfor i in range(80000):\n rand_index = np.random.choice(len(x_vals_train), size=batch_size, replace=False)\n rand_x = x_vals_train[rand_index]\n rand_y = np.transpose([trainLabels[rand_index]])\n sess.run(train_step, feed_dict={x_data: rand_x, y_target: rand_y})\n\n temp_loss = sess.run(loss, feed_dict={x_data: rand_x, y_target: rand_y})\n loss_vec.append(np.sqrt(temp_loss))\n\n\n if (i + 1) % 100 == 0:\n print('Generation: ' + str(i+1) + '. Loss = ' + str(temp_loss))\n\n\n# save_dir ='/home/shen/Trying/Predict/up/t1/more_dimension/Net_save/'\n# saver.save(sess, save_dir + 'model.ckpt', global_step=i+1)\n\n\nx = x_vals_test\ny = np.transpose([testLabels])\nprint(\"Testing Accuracy:\", sess.run(accuracy*100, feed_dict={x_data: x,y_target: y,}\n ))\n\n# Plot loss (MSE) over time\nplt.plot(loss_vec, 'k-', label='Train Loss')\nplt.title('Loss (MSE) per Generation')\nplt.legend(loc='upper right')\nplt.xlabel('Generation')\nplt.ylabel('Loss')\nplt.show()\n","sub_path":"git/NeuralNet_predict_weight/t1/more_dimension/NerualNet_final.py","file_name":"NerualNet_final.py","file_ext":"py","file_size_in_byte":6046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"14364491","text":"from datetime import datetime\n\nfrom nba_data.data.game import Game\nfrom nba_data.data.matchup import Matchup\nfrom nba_data.data.outcome import Outcome\nfrom nba_data.data.season import Season\nfrom nba_data.data.season_type import SeasonType\n\n\nclass TeamGameLogDeserializer:\n game_date_format = \"%b %d, %Y\"\n\n result_set_index = 0\n game_id_index = 1\n game_date_index = 2\n matchup_index = 3\n home_team_outcome_index = 4\n\n def __init__(self):\n pass\n\n @staticmethod\n def deserialize_team_game_log(team_game_log_json):\n deserialized_results = []\n results = team_game_log_json[\"resultSets\"][TeamGameLogDeserializer.result_set_index][\"rowSet\"]\n season = Season.get_season(team_game_log_json[\"parameters\"][Season.get_query_parameter_name()])\n season_type = SeasonType.get_season_type(team_game_log_json[\"parameters\"][SeasonType.get_query_parameter_name()])\n for result in results:\n matchup = TeamGameLogDeserializer.parse_matchup(result[TeamGameLogDeserializer.matchup_index])\n home_team_outcome = Outcome.get_outcome_from_abbreviation(result[TeamGameLogDeserializer.home_team_outcome_index])\n deserialized_results.append(\n Game(nba_id=str(result[TeamGameLogDeserializer.game_id_index]),\n matchup=matchup,\n date=TeamGameLogDeserializer.parse_date(result[TeamGameLogDeserializer.game_date_index]),\n season=season,\n season_type=season_type,\n home_team_outcome=home_team_outcome))\n return deserialized_results\n\n @staticmethod\n def parse_matchup(matchup):\n\n if \" vs. \" in matchup:\n teams = matchup.split(\" vs. \")\n return Matchup.create(home_team_abbreviation=str(teams[0]),\n away_team_abbreviation=str(teams[1]))\n\n elif \" @ \" in matchup:\n teams = matchup.split(\" @ \")\n return Matchup.create(home_team_abbreviation=str(teams[1]),\n away_team_abbreviation=str(teams[0]))\n\n else:\n raise RuntimeError(\"Unexpected matchup: %s\", matchup)\n\n @staticmethod\n def parse_date(date_string):\n return datetime.strptime(date_string, TeamGameLogDeserializer.game_date_format).date()","sub_path":"nba_data/deserializers/team_game_log_deserializer.py","file_name":"team_game_log_deserializer.py","file_ext":"py","file_size_in_byte":2335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"539236976","text":"import time\n\nimport Moana.Core.Logger.Logger as mLogger\nimport Moana.Core.Ruler.Ruler as mRuler\n\nclass BaseWorkerManager:\n def __init__(self):\n self._logger = mLogger.Logger.logger()\n self._ruler = mRuler.Ruler.ruler()\n\n self._workers = {} # Processes\n self._instanceWorkers = {} # Processes\n self._handlers = {} # interface class\n self._roles = {} # name\n\n self._rWorkers = {} # Running workers\n self._currentWorker = \"Core\"\n\n self._logger.debug(\"Generated {0}\".format(self))\n\n def __repr__(self):\n return type(self).__name__\n\n def _addHandler(self, name, handler):\n self._handlers[name] = handler\n self._roles[name] = [name]\n\n def addWorker(self, name, worker, handler):\n workerType = worker.__class__.__name__\n self._logger.debug(\n \"Add {0} worker: {1}\".format(workerType, name))\n\n if name == \"Core\":\n self._addHandler(name, handler)\n if workerType == \"Process\":\n self._workers[name] = worker\n self._addHandler(name, handler)\n else:\n self._logger.critical(\n \"Wrong worker type: {0}\".format(workerType))\n\n def setRole(self, role, worker):\n self._roles[worker].append(role)\n self._logger.debug(\n \"Append role {0} to {1}\".format(role, worker))\n\n def getRoles(self):\n return self._roles\n\n def getHandler(self, role):\n handler = None\n\n for k, v in self._roles.items():\n if role in v:\n handler = k\n break\n\n if not handler:\n self._logger.critical(\n \"Wrong role: {0}\".format(role))\n\n return self._handlers[handler]\n\n def _runWorker(self, worker, name):\n if name in self._rWorkers:\n self._logger.debug(\"Already running: {0}\".format(name))\n return\n\n worker.start()\n self._rWorkers[name] = worker\n self._logger.debug(\"Run: {0}\".format(name))\n\n def runMessenger(self):\n messengers = None\n\n exe = self._ruler.getConfig(\"exe\")\n if exe == \"Moana\":\n messengers = (\"Publisher\", \"Subscriber\")\n elif exe == \"Boat\":\n messengers = (\"Requester\", \"Replyer\")\n else:\n self._logger.critical(\n \"Wrong exe type: {0}\".format(exe))\n\n for messenger in messengers:\n if messenger in self._workers:\n worker = self._workers[messenger]\n self._runWorker(worker, messenger)\n\n time.sleep(1)\n\n def runWorkers(self):\n self.runMessenger()\n\n for name in self._workers.keys():\n worker = self._workers[name]\n self._runWorker(worker, name)\n\n def runInstanceWorker(self, name, worker, handler):\n workerType = worker.__class__.__name__\n self._logger.debug(\n \"Add {0} instance worker: {1}\".format(workerType, name))\n\n if workerType == \"Process\":\n self._instanceWorkers[name] = worker\n self._addHandler(name, handler)\n else:\n self._logger.critical(\n \"Wrong worker type: {0}\".format(workerType))\n\n self._runWorker(worker, name)\n def stop(self):\n for workerName, worker in self._rWorkers.items():\n worker.terminate()\n self._logger.debug(\"Terminate: {0}\".format(workerName))\n\n def getWorkersList(self):\n return self._workers\n\n def deleteRunningWorker(self, name):\n del self._rWorkers[name]\n\n self._logger.debug(\n \"Worker deleted: {0}\".format(name))\n\n def getHandlersList(self):\n return self._handlers\n\n def getWorkerType(self):\n types = []\n for k, v in self._workers.items():\n types.append(type(v))\n\n return types\n\n def getHandlerType(self):\n types = []\n for k, v in self._handlers.items():\n types.append(type(v))\n\n return types\n\n def getCurrentWorker(self):\n return self._currentWorker\n\n def setCurrentWorker(self, worker):\n self._currentWorker = worker\n\n def getRunningWorkersList(self):\n return self._rWorkers\n\nclass WorkerManager(BaseWorkerManager):\n _instance = None\n\n @classmethod\n def workerManager(cls, *args, **kwargs):\n if not cls._instance:\n cls._instance = cls(*args, **kwargs)\n\n return cls._instance","sub_path":"Moana/Core/WorkerManager/WorkerManager.py","file_name":"WorkerManager.py","file_ext":"py","file_size_in_byte":4445,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"577321623","text":"import sys\r\nif sys.platform == 'linux':\r\n STATUSMARIO = 0x569f4699\r\n MOEDAS = 0x569f543f\r\n YOSHICOINS = 0x569f5aa2\r\nelse:\r\n STATUSMARIO = 0x72B481\r\n MOEDAS = 0x72C227\r\n YOSHICOINS = 0x72C888\r\n\r\nMARIO_PEQUENO = 0\r\nMARIO_GRANDE = 1\r\nMARIO_PENINHA = 2\r\nMARIO_FLOR_DE_FOGO = 3","sub_path":"Enderecos.py","file_name":"Enderecos.py","file_ext":"py","file_size_in_byte":290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"634695304","text":"\"\"\"Contains the `Enduse` Class. This is the most important class\nwhere the change in enduse specific energy demand is simulated\ndepending on scenaric assumptions.\"\"\"\nimport logging\nimport math\nimport numpy as np\nfrom energy_demand.profiles import load_profile as lp\nfrom energy_demand.profiles import load_factors as lf\nfrom energy_demand.technologies import diffusion_technologies\nfrom energy_demand.technologies import fuel_service_switch\nfrom energy_demand.technologies import tech_related\nfrom energy_demand.basic import lookup_tables\n\nclass Enduse(object):\n \"\"\"Enduse Class for all endueses in each SubModel\n\n For every region and sector, a different instance\n is generated. In this class, first the change in\n energy demand is calculated on a annual temporal scale.\n Calculations are performed in a cascade (e.g. first\n reducing climate change induced savings, then substracting\n further behavioral savings etc.). After annual calculations,\n the demand is converted to hourly demand.\n\n Also within this function, the fuel inputs are converted\n to energy service (short: service) and converted back to\n fuels (e.g. electricit).\n\n Arguments\n ----------\n submodel : str\n Submodel\n region : str\n Region name\n scenario_data : dict\n Scenario data\n assumptions : dict\n Assumptions\n non_regional_lp_stock : dict\n Load profile stock\n base_yr : int\n Base year\n curr_yr : int\n Current year\n enduse : str\n Enduse name\n sector : str\n Sector name\n fuel : array\n Yearly fuel data for different fueltypes\n tech_stock : object\n Technology stock of region\n heating_factor_y : array\n Distribution of fuel within year to days (yd) (directly correlates with HDD)\n cooling_factor_y : array\n Distribution of fuel within year to days (yd) (directly correlates with CDD)\n fuel_fueltype_tech_p_by : dict\n Fuel tech assumtions in base year\n sig_param_tech : dict\n Sigmoid parameters\n enduse_overall_change : dict\n Assumptions related to overal change in endyear\n regional_lp_stock : object\n Load profile stock\n dw_stock : object,default=False\n Dwelling stock\n reg_scen_drivers : bool,default=None\n Scenario drivers per enduse\n flat_profile_crit : bool,default=False\n Criteria of enduse has a flat shape or not\n\n Note\n ----\n - Load profiles are assigned independently of the fueltype, i.e.\n the same profiles are assumed to hold true across different fueltypes\n\n - ``self.fuel_y`` is always overwritten\n in the cascade of calculations\n\n Warning\n -------\n Not all enduses have technologies assigned. Load peaks are derived\n from techstock in case technologies are defined. Otherwise enduse load\n profiles are used.\n \"\"\"\n def __init__(\n self,\n submodel,\n region,\n scenario_data,\n assumptions,\n regional_lp_stock,\n non_regional_lp_stock,\n base_yr,\n curr_yr,\n enduse,\n sector,\n fuel,\n tech_stock,\n heating_factor_y,\n cooling_factor_y,\n fuel_fueltype_tech_p_by,\n sig_param_tech,\n enduse_overall_change,\n criterias,\n strategy_variables,\n fueltypes_nr,\n fueltypes,\n model_yeardays_nrs,\n dw_stock=False,\n reg_scen_drivers=None,\n flat_profile_crit=False\n ):\n \"\"\"Enduse class constructor\n \"\"\"\n #logging.info(\" =====Enduse: {} Sector: {}\".format(enduse, sector))\n self.region = region\n self.enduse = enduse\n self.fuel_y = fuel\n self.flat_profile_crit = flat_profile_crit\n\n self.techs_fuel_yh = None\n\n if np.sum(fuel) == 0:\n #If enduse has no fuel return empty shapes\n self.flat_profile_crit = True\n self.fuel_y = fuel\n self.fuel_yh = 0\n self.enduse_techs = []\n else:\n # Get correct parameters depending on model configuration\n load_profiles = get_lp_stock(\n enduse,\n non_regional_lp_stock,\n regional_lp_stock)\n\n # Get technologies of enduse\n self.enduse_techs = get_enduse_techs(fuel_fueltype_tech_p_by)\n\n # -------------------------------\n # Cascade of calculations on a yearly scale\n # --------------------------------\n # --Change fuel consumption based on climate change induced temperature differences\n _fuel_new_y = apply_climate_change(\n enduse,\n self.fuel_y,\n cooling_factor_y,\n heating_factor_y,\n assumptions.enduse_space_heating,\n assumptions.ss_enduse_space_cooling)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train B: \" + str(np.sum(self.fuel_y)))\n\n # --Change fuel consumption based on smart meter induced general savings\n _fuel_new_y = apply_smart_metering(\n enduse,\n self.fuel_y,\n assumptions.smart_meter_assump,\n strategy_variables,\n base_yr,\n curr_yr)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train C: \" + str(np.sum(self.fuel_y)))\n\n # --Enduse specific fuel consumption change in %\n _fuel_new_y = apply_specific_change(\n enduse,\n self.fuel_y,\n enduse_overall_change,\n strategy_variables,\n base_yr,\n curr_yr)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train D: \" + str(np.sum(self.fuel_y)))\n\n # Calculate new fuel demands after scenario drivers\n _fuel_new_y = apply_scenario_drivers(\n submodel,\n enduse,\n sector,\n self.fuel_y,\n dw_stock,\n region,\n scenario_data['gva'],\n scenario_data['population'],\n scenario_data['industry_gva'],\n reg_scen_drivers,\n base_yr,\n curr_yr)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train E: \" + str(np.sum(self.fuel_y)))\n\n # Apply cooling scenario variable\n _fuel_new_y = apply_cooling(\n enduse,\n self.fuel_y,\n strategy_variables,\n assumptions.cooled_ss_floorarea_by,\n enduse_overall_change['other_enduse_mode_info'],\n base_yr,\n curr_yr)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train E1: \" + str(np.sum(self.fuel_y)))\n\n # Industry related change\n _fuel_new_y = industry_enduse_changes(\n enduse,\n sector,\n base_yr,\n curr_yr,\n strategy_variables,\n self.fuel_y,\n enduse_overall_change['other_enduse_mode_info'],\n assumptions)\n self.fuel_y = _fuel_new_y\n #logging.debug(\"... Fuel train E2: \" + str(np.sum(self.fuel_y)))\n\n # ----------------------------------\n # Hourly Disaggregation\n # ----------------------------------\n if self.enduse_techs == []:\n \"\"\"If no technologies are defined for an enduse, the load profiles\n are read from dummy shape, which show the load profiles of the whole enduse.\n No switches can be implemented and only overall change of enduse.\n\n Note: for heating, technologies need to be assigned.\n \"\"\"\n if flat_profile_crit:\n self.fuel_y = self.fuel_y * model_yeardays_nrs / 365.0\n else:\n self.fuel_yh = assign_lp_no_techs(\n enduse,\n sector,\n load_profiles,\n self.fuel_y)\n else:\n \"\"\"If technologies are defined for an enduse\n \"\"\"\n # ----\n # Get enduse specific configurations\n # ----\n mode_constrained = get_enduse_configuration(\n criterias['mode_constrained'],\n enduse,\n assumptions.enduse_space_heating)\n\n # ------------------------------------\n # Calculate regional energy service\n # ------------------------------------\n s_tot_y_cy, s_tech_y_by = fuel_to_service(\n enduse,\n self.fuel_y,\n fuel_fueltype_tech_p_by,\n tech_stock,\n fueltypes,\n mode_constrained)\n\n # ------------------------------------\n # Reduction of service because of heat recovery\n # ------------------------------------\n s_tot_y_cy, s_tech_y_cy = apply_heat_recovery(\n enduse,\n strategy_variables,\n assumptions.enduse_overall_change,\n s_tot_y_cy,\n s_tech_y_by,\n base_yr,\n curr_yr)\n\n # ------------------------------------\n # Reduction of service because of improvement in air leakeage\n # ------------------------------------\n s_tot_y_cy, s_tech_y_cy = apply_air_leakage(\n enduse,\n strategy_variables,\n assumptions.enduse_overall_change,\n s_tot_y_cy,\n s_tech_y_cy,\n base_yr,\n curr_yr)\n\n # --------------------------------\n # Switches\n # Calculate services per technology for cy based on fitted parameters\n # --------------------------------\n s_tech_y_cy = calc_service_switch(\n enduse,\n s_tech_y_cy,\n self.enduse_techs,\n sig_param_tech,\n curr_yr,\n base_yr,\n sector,\n assumptions.crit_switch_happening)\n\n # -------------------------------------------\n # Convert annual service to fuel per fueltype\n # -------------------------------------------\n self.fuel_y, fuel_tech_y = service_to_fuel(\n enduse,\n s_tech_y_cy,\n tech_stock,\n fueltypes_nr,\n fueltypes,\n mode_constrained)\n #logging.debug(\"... Fuel train Post service: \" + str(np.sum(self.fuel_y)))\n\n # Delete all technologies with no fuel assigned\n for tech, fuel_tech in fuel_tech_y.items():\n if fuel_tech == 0:\n self.enduse_techs.remove(tech)\n\n # ------------------------------------------\n # Assign load profiles\n # ------------------------------------------\n if self.flat_profile_crit:\n #logging.info(\"flat profile\")\n self.fuel_y = calc_fuel_tech_y(\n enduse,\n tech_stock,\n fuel_tech_y,\n fueltypes_nr,\n fueltypes,\n mode_constrained)\n else:\n fuel_yh = calc_fuel_tech_yh(\n enduse,\n sector,\n self.enduse_techs,\n fuel_tech_y,\n load_profiles,\n fueltypes_nr,\n fueltypes,\n model_yeardays_nrs,\n mode_constrained)\n\n # --------------------------------------\n # Demand Management (peak shaving)\n # ---------------------------------------\n if mode_constrained:\n self.techs_fuel_yh = {}\n for tech in fuel_yh:\n self.techs_fuel_yh[tech] = demand_management(\n enduse,\n base_yr,\n curr_yr,\n strategy_variables,\n fuel_yh[tech],\n [tech],\n sector,\n fuel_tech_y,\n tech_stock,\n load_profiles,\n mode_constrained=True)\n\n self.fuel_yh = None\n else: # (not specific for technologies)\n\n # Demand management for heating related technologies\n self.fuel_yh = demand_management(\n enduse,\n base_yr,\n curr_yr,\n strategy_variables,\n fuel_yh,\n self.enduse_techs,\n sector,\n fuel_tech_y,\n tech_stock,\n load_profiles,\n mode_constrained=False)\n\ndef demand_management(\n enduse,\n base_yr,\n curr_yr,\n strategy_variables,\n fuel_yh,\n enduse_techs,\n sector,\n fuel_tech_y,\n tech_stock,\n load_profiles,\n mode_constrained\n ):\n \"\"\"Demand management. This function shifts peak per of this enduse\n depending on peak shifting factors. So far only inter day load shifting\n\n Arguments\n ----------\n enduse : str\n Enduse\n base_yr : int\n Base year\n curr_yr : int\n Current year\n strategy_variables : dict\n Assumptions of strategy variables\n fuel_yh : array\n Fuel per hours\n enduse_techs : list\n Enduse specfic technologies\n sector : str\n Sector\n fuel_tech_y : dict\n Annual fuel per technology\n tech_stock : obj\n Technology stock\n load_profiles : obj\n Load profiles\n mode_constrained : bool\n Running mode\n If mode_constrained, always only one technology imported\n\n Returns\n -------\n fuel_yh : array\n Fuel of yh\n \"\"\"\n # ------------------------------\n # Test if peak is shifted or not\n # ------------------------------\n try:\n # Get assumed load shift\n param_name = 'demand_management_improvement__{}'.format(enduse)\n\n if strategy_variables[param_name]['scenario_value'] == 0:\n\n # no load management\n peak_shift_crit = False\n else:\n # load management\n peak_shift_crit = True\n except KeyError:\n\n # no load management\n peak_shift_crit = False\n\n # ------------------------------\n # If peak shifting implemented, calculate new lp\n # ------------------------------\n if peak_shift_crit:\n\n # Calculate average for every day\n if mode_constrained:\n average_fuel_yd = np.average(fuel_yh, axis=1)\n else:\n average_fuel_yd = np.average(fuel_yh, axis=2)\n\n # Calculate load factors (only inter_day load shifting as for now)\n loadfactor_yd_cy = lf.calc_lf_d(\n fuel_yh, average_fuel_yd, mode_constrained)\n\n # Calculate current year load factors\n lf_improved_cy = calc_lf_improvement(\n strategy_variables[param_name]['scenario_value'],\n base_yr,\n curr_yr,\n loadfactor_yd_cy,\n strategy_variables['demand_management_yr_until_changed']['scenario_value'])\n\n fuel_yh = lf.peak_shaving_max_min(\n lf_improved_cy, average_fuel_yd, fuel_yh, mode_constrained)\n\n else: # no peak shifting\n pass\n\n return fuel_yh\n\ndef calc_lf_improvement(\n lf_improvement_ey,\n base_yr,\n curr_yr,\n loadfactor_yd_cy,\n yr_until_changed\n ):\n \"\"\"Calculate load factor improvement depending on linear diffusion\n over time.\n\n Arguments\n ---------\n lf_improvement_ey : dict\n Load factor improvement until end year\n base_yr : int\n Base year\n curr_yr : int\n Current year\n loadfactor_yd_cy : float\n Yd Load factor of current year\n yr_until_changed : int\n Year until fully changed\n\n Returns\n -------\n lf_improved_cy : str\n Improved load factor of current year\n peak_shift_crit : bool\n True: Peak is shifted, False: Peak isn't shifed\n \"\"\"\n # Calculate linear diffusion of improvement of load management\n lin_diff_factor = diffusion_technologies.linear_diff(\n base_yr, curr_yr, 0, 1, yr_until_changed)\n\n # Current year load factor improvement\n lf_improvement_cy = lf_improvement_ey * lin_diff_factor\n\n # Add load factor improvement to current year load factor\n lf_improved_cy = loadfactor_yd_cy + lf_improvement_cy\n\n # Where load factor larger than zero, set to 1\n lf_improved_cy[lf_improved_cy > 1] = 1\n\n return lf_improved_cy\n\ndef assign_lp_no_techs(enduse, sector, load_profiles, fuel_y):\n \"\"\"Assign load profiles for an enduse which has no technologies defined\n\n Arguments\n ---------\n enduse : str\n Enduse\n sector : str\n Enduse\n load_profiles : obj\n Load profiles\n fuel_y : array\n Fuels\n\n Returns\n -------\n fuel_yh : array\n Fuel yh\n \"\"\"\n fuel = fuel_y[:, np.newaxis, np.newaxis]\n\n fuel_yh = load_profiles.get_lp(\n enduse, sector, 'placeholder_tech', 'shape_yh') * fuel\n\n return fuel_yh\n\ndef get_lp_stock(enduse, non_regional_lp_stock, regional_lp_stock):\n \"\"\"Defines the load profile stock depending on `enduse`.\n (Get regional or non-regional load profile data)\n\n Arguments\n ----------\n enduse : str\n Enduse\n non_regional_lp_stock : object\n Non regional dependent load profiles\n regional_lp_stock : object\n Regional dependent load profiles\n\n Returns\n -------\n load_profiles : object\n Load profile\n\n Note\n -----\n Because for some enduses the load profiles depend on the region\n they are stored in the `WeatherRegion` Class. One such example is\n heating. If the enduse is not dependent on the region, the same\n load profile can be used for all regions\n\n If the enduse depends on regional factors, `regional_lp_stock`\n is returned. Otherwise, non-regional load profiles which can\n be applied for all regions is used (`non_regional_lp_stock`)\n \"\"\"\n if enduse in non_regional_lp_stock.stock_enduses:\n return non_regional_lp_stock\n else:\n return regional_lp_stock\n\ndef get_running_mode(enduse, mode_constrained, enduse_space_heating):\n \"\"\"Checks which mode needs to be run for an enduse.\n\n Arguments\n -----------\n mode_constrained : bool\n Criteria of running mode\n enduse_space_heating : dict\n All heating enduses across all models\n\n Returns\n -------\n bool : bool\n The return value\n\n Note\n ----\n If 'crit_mode' == True, then overall heat is provided to\n the supply model not specified for technologies. Otherwise,\n heat demand is supplied per technology\n \"\"\"\n if mode_constrained:\n return True\n elif not mode_constrained and enduse in enduse_space_heating:\n return False\n elif not mode_constrained and enduse not in enduse_space_heating:\n # All other not constrained enduses where technologies are defined\n # are run in 'constrained' mode (e.g. lighting)\n return True\n\ndef get_enduse_configuration(\n mode_constrained,\n enduse,\n enduse_space_heating,\n ):\n \"\"\"Get enduse specific configuration\n\n Arguments\n ---------\n mode_constrained : bool\n Constrained mode criteria\n enduse : str\n Enduse\n enduse_space_heating : list\n All endueses classified as space heating\n base_yr, curr_yr : int\n Base, current, year\n \"\"\"\n mode_constrained = get_running_mode(\n enduse,\n mode_constrained,\n enduse_space_heating)\n\n return mode_constrained\n\ndef get_peak_day_all_fueltypes(fuel_yh):\n \"\"\"Iterate yh and get day with highes fuel (across all fueltypes).\n The day with most fuel across all fueltypes is considered to\n be the peak day. Over the simulation period,\n the peak day may change date in a year.\n\n Arguments\n ---------\n fuel_yh : array (fueltype, 365, 24)\n Fuel for every yh (fueltypes, yh) \n\n Return\n ------\n peak_day_nr : int\n Day with most fuel or service across all fueltypes\n \"\"\"\n\n # Sum all fuel across all fueltypes for every hour in a year\n all_fueltypes_tot_h = np.sum(fuel_yh, axis=0)\n\n if np.sum(all_fueltypes_tot_h) == 0:\n logging.warning(\"No peak can be found because no fuel assigned\")\n return 0\n else:\n # Sum fuel within every hour for every day and get day with maximum fuel\n peak_day_nr = np.argmax(np.sum(all_fueltypes_tot_h, axis=1))\n\n return peak_day_nr\n\ndef get_peak_day_single_fueltype(fuel_yh):\n \"\"\"Iterate yh and get day with highes fuel for a single fueltype\n The day with most fuel is considered to\n be the peak day. Over the simulation period,\n the peak day may change date in a year. If no fuel is\n provided, the program is crashed\n\n Arguments\n ---------\n fuel_yh : array (365, 24)\n Fuel for every yh (yh)\n\n Return\n ------\n peak_day_nr : int\n Day with most fuel or service\n \"\"\"\n if np.sum(fuel_yh) == 0:\n logging.info(\"No peak can be found because no fuel assigned\")\n # Return first entry of element (which is zero)\n return 0\n else:\n # Sum fuel within every hour for every day and get day with maximum fuel\n peak_day_nr = np.argmax(np.sum(fuel_yh, axis=1))\n return peak_day_nr\n\ndef calc_peak_tech_dh(\n enduse,\n sector,\n enduse_techs,\n enduse_fuel_tech,\n fuel_yh,\n tech_stock,\n load_profile,\n mode_constrained\n ):\n \"\"\"Iterate technologies in enduse and calculate peak demand\n\n Arguments\n ----------\n enduse : str\n Enduse\n sector : str\n Sector\n enduse_techs : list\n Enduse technologies\n enduse_fuel_tech : array\n Fuel per enduse and technology\n fuel_yh : array\n Fuel per hours\n tech_stock : data\n Technology stock\n load_profile : object\n Load profile\n mode_constrained : bool\n Constrained mode criteria\n\n Returns\n -------\n fuels_peak_dh : array\n Peak values for peak day for every fueltype\n\n Note\n ----\n - This function gets the hourly values of the peak day for every fueltype.\n The daily fuel is converted to dh for each technology.\n\n - For some technology types (heat_pump)\n the dh peak day profile is not read in from technology\n stock but from shape_yh of peak day.\n \"\"\"\n if mode_constrained:\n fuels_peak_dh = {}\n else:\n fuels_peak_dh = np.zeros((24), dtype=float)\n\n for tech in enduse_techs:\n\n tech_type = tech_stock.get_tech_attr(\n enduse, tech, 'tech_type')\n\n if tech_type == 'heat_pump':\n \"\"\"Read fuel from peak day\n \"\"\"\n # Get day with most fuel\n if isinstance(fuel_yh, dict):\n peak_day_nr = get_peak_day_single_fueltype(fuel_yh[tech])\n else:\n peak_day_nr = get_peak_day_single_fueltype(fuel_yh)\n\n # Calculate absolute fuel values for yd (multiply fuel with yd_shape)\n fuel_tech_yd = enduse_fuel_tech[tech] * load_profile.get_lp(\n enduse, sector, tech, 'shape_yd')\n\n # Calculate fuel for peak day\n fuel_tech_peak_d = fuel_tech_yd[peak_day_nr]\n\n # The 'shape_peak_dh'is not defined in technology stock because\n # in the 'Region' the peak day is not yet known\n # Therefore, the shape_yh is read in and with help of information on peak day\n tech_peak_dh = load_profile.get_lp(\n enduse, sector, tech, 'shape_y_dh')[peak_day_nr]\n else:\n \"\"\"Calculate fuel with peak factor\n \"\"\"\n f_peak_yd = load_profile.get_lp(\n enduse, sector, tech, 'f_peak_yd')\n\n # Calculate fuel for peak day\n fuel_tech_peak_d = enduse_fuel_tech[tech] * f_peak_yd\n\n # Assign Peak shape of a peak day of a technology\n tech_peak_dh = load_profile.get_shape_peak_dh(\n enduse, sector, tech)\n\n # Multiply absolute d fuels with dh peak fuel shape\n fuel_tech_peak_dh = tech_peak_dh * fuel_tech_peak_d\n\n if mode_constrained:\n fuels_peak_dh[tech] = fuel_tech_peak_dh\n else:\n # Peak day fuel shape * fueltype distribution for peak day\n fuels_peak_dh += fuel_tech_peak_dh\n\n return fuels_peak_dh\n\ndef get_enduse_techs(fuel_fueltype_tech_p_by):\n \"\"\"Get all defined technologies of an enduse\n\n Arguments\n ----------\n fuel_fueltype_tech_p_by : dict\n Percentage of fuel per enduse per technology\n\n Return\n ------\n enduse_techs : list\n All technologies\n\n Note\n ----\n All technologies are read out, including those which\n are potentially defined in fuel or service switches.\n\n If for an enduse a dummy technology is defined,\n the technologies of an enduse are set to an empty\n list.\n\n Warning\n -------\n For every enduse technologes must either be defined\n for no fueltype or for all fueltypes\n \"\"\"\n enduse_techs = []\n\n for tech_fueltype in fuel_fueltype_tech_p_by.values():\n if 'placeholder_tech' in tech_fueltype.keys():\n return []\n else:\n enduse_techs += tech_fueltype.keys()\n\n return list(set(enduse_techs))\n\ndef calc_fuel_tech_yh(\n enduse,\n sector,\n enduse_techs,\n enduse_fuel_tech,\n load_profiles,\n fueltypes_nr,\n fueltypes,\n model_yeardays_nrs,\n mode_constrained\n ):\n \"\"\"Iterate fuels for each technology and assign shape yd and yh shape\n\n Arguments\n ----------\n enduse_fuel_tech : dict\n Fuel per technology in enduse\n tech_stock : object\n Technologies\n load_profiles : object\n Load profiles\n fueltypes_nr : dict\n Nr of fueltypes\n fueltypes : dict\n Fueltypes lookup\n mode_constrained : bool\n Mode criteria\n model_yeardays_nrs : int\n Number of modelled yeardays\n\n Return\n ------\n fuels_yh : array\n Fueltype storing hourly fuel for every fueltype (fueltype, model_yeardays_nrs, 24)\n \"\"\"\n if mode_constrained:\n\n fuels_yh = {}\n for tech in enduse_techs:\n\n load_profile = load_profiles.get_lp(\n enduse, sector, tech, 'shape_yh')\n\n if model_yeardays_nrs != 365:\n load_profile = lp.abs_to_rel(load_profile)\n\n fuel_tech_yh = enduse_fuel_tech[tech] * load_profile\n\n fuels_yh[tech] = fuel_tech_yh\n else:\n # --\n # Unconstrained mode, i.e. not technolog specific.\n # Store according to fueltype and heat\n # --\n fuels_yh = np.zeros((fueltypes_nr, model_yeardays_nrs, 24), dtype=float)\n\n for tech in enduse_techs:\n\n load_profile = load_profiles.get_lp(\n enduse, sector, tech, 'shape_yh')\n\n if model_yeardays_nrs != 365:\n load_profile = lp.abs_to_rel(load_profile)\n\n # If no fuel for this tech and not defined in enduse\n fuel_tech_yh = enduse_fuel_tech[tech] * load_profile\n\n fuels_yh[fueltypes['heat']] += fuel_tech_yh\n\n return fuels_yh\n\ndef calc_fuel_tech_y(\n enduse,\n tech_stock,\n fuel_tech_y,\n fueltypes_nr,\n fueltypes,\n mode_constrained\n ):\n \"\"\"Calculate yearly fuel per technology (no load profile assigned).\n\n Arguments\n -----------\n enduse : str\n Enduse\n tech_stock : object\n Technology stock\n fuel_tech_y : dict\n Fuel per technology per year\n lookups : dict\n look-up\n fueltype : dict\n Integer of fueltypes\n mode_constrained : bool\n Running mode\n\n Returns\n -------\n fuel_y : array\n Fuel per year per fueltype\n\n Note\n ----\n This function can be run in two different modes\n \"\"\"\n fuel_y = np.zeros((fueltypes_nr), dtype=float)\n\n for tech, fuel_tech_y in fuel_tech_y.items():\n if mode_constrained:\n fueltype_int = tech_stock.get_tech_attr(\n enduse, tech, 'fueltype_int')\n\n fuel_y[fueltype_int] += np.sum(fuel_tech_y)\n else:\n # Assign all to heat fueltype\n fuel_y[fueltypes['heat']] += np.sum(fuel_tech_y)\n\n return fuel_y\n\ndef service_to_fuel(\n enduse,\n service_tech,\n tech_stock,\n fueltypes_nr,\n fueltypes,\n mode_constrained\n ):\n \"\"\"Convert yearly energy service to yearly fuel demand.\n For every technology the service is taken and converted\n to fuel based on efficiency of current year\n\n Arguments\n ------\n enduse : str\n Enduse\n service_tech : dict\n Service per fueltype and technology\n tech_stock : object\n Technological stock\n fueltypes_nr : int\n Number of fueltypes\n fueltypes : dict\n Fueltypes\n mode_constrained : bool\n Mode running criteria\n\n Returns\n -------\n fuel_y : array\n Fuel per fueltype\n fuel_per_tech : dict\n Fuel per technology\n\n Note\n -----\n - Fuel = Energy service / efficiency\n \"\"\"\n fuel_tech_y = {}\n fuel_y = np.zeros((fueltypes_nr), dtype=float)\n\n if mode_constrained:\n for tech, service in service_tech.items():\n\n tech_eff = tech_stock.get_tech_attr(\n enduse, tech, 'eff_cy')\n fueltype_int = tech_stock.get_tech_attr(\n enduse, tech, 'fueltype_int')\n\n # Convert to fuel\n fuel = service / tech_eff\n\n # Add fuel\n fuel_tech_y[tech] = fuel\n fuel_y[fueltype_int] += fuel\n else:\n for tech, fuel_tech in service_tech.items():\n fuel_y[fueltypes['heat']] += fuel_tech\n fuel_tech_y[tech] = fuel_tech\n\n return fuel_y, fuel_tech_y\n\ndef fuel_to_service(\n enduse,\n fuel_y,\n fuel_fueltype_tech_p_by,\n tech_stock,\n fueltypes,\n mode_constrained\n ):\n \"\"\"Converts fuel to energy service. Calculate energy service\n of each technology based on assumptions about base year fuel\n shares of an enduse (`fuel_fueltype_tech_p_by`).\n\n Arguments\n ----------\n enduse : str\n Enduse\n fuel_y : array\n Fuel per fueltype\n fuel_fueltype_tech_p_by : dict\n Fuel composition of base year for every fueltype for each\n enduse (assumtions for national scale)\n tech_stock : object\n Technology stock of region\n fueltypes : dict\n Fueltype look-up\n mode_constrained : bool\n Criteria about mode\n\n Return\n ------\n tot_s_y : array\n Total annual energy service per technology\n s_tech_y : dict\n Total annual energy service per technology\n\n Note\n -----\n - Efficiency changes of technologis are considered.\n - Energy service = fuel * efficiency\n - This function can be run in two modes, depending on `mode_constrained`\n - The base year efficiency is taken because the actual service can\n only be calculated with base year.\n Efficiencies are only considered if converting back to fuel\n The 'self.fuel_y' is taken because the actual\n service was reduced e.g. due to smart meters or temperatur changes\n \"\"\"\n s_tech_y = {}\n s_tot_y = 0\n\n # Calculate share of service\n for fueltype_int, tech_list in fuel_fueltype_tech_p_by.items():\n\n # Get technologies to iterate\n if tech_list == {} and fuel_y[fueltype_int] == 0: # No technology or fuel defined\n techs_with_fuel = {}\n elif tech_list == {} and fuel_y[fueltype_int] > 0: # Fuel defined but no technologies\n fueltype_str = tech_related.get_fueltype_str(fueltypes, fueltype_int)\n placeholder_tech = 'placeholder_tech__{}'.format(fueltype_str)\n techs_with_fuel = {placeholder_tech: 1.0}\n else:\n techs_with_fuel = tech_list\n\n for tech, fuel_share in techs_with_fuel.items():\n\n if mode_constrained:\n \"\"\"Constrained version\n \"\"\"\n tech_eff = tech_stock.get_tech_attr(enduse, tech, 'eff_by')\n\n # Get fuel share and convert fuel to service per technology\n s_tech = fuel_y[fueltype_int] * fuel_share * tech_eff\n\n s_tech_y[tech] = s_tech\n\n # Sum total yearly service\n s_tot_y += s_tech #(y)\n else:\n \"\"\"Unconstrained version\n efficiencies are not considered, because not technology\n specific service calculation\n \"\"\"\n # Calculate fuel share\n fuel_tech = fuel_y[fueltype_int] * fuel_share\n\n s_tech_y[tech] = fuel_tech\n\n # Sum total yearly service\n s_tot_y += fuel_tech\n\n return s_tot_y, s_tech_y\n\ndef apply_heat_recovery(\n enduse,\n strategy_variables,\n enduse_overall_change,\n service,\n service_techs,\n base_yr,\n curr_yr\n ):\n \"\"\"Reduce heating demand according to assumption on heat reuse\n\n Arguments\n ----------\n enduse : str\n Enduse\n strategy_variables : dict\n Strategy variables\n enduse_overall_change : dict\n Sigmoid diffusion info\n service : dict or array\n Service of current year\n crit_dict : str\n Criteria to run function differently\n base_yr : int\n Base year\n curr_yr : int\n Current year\n\n Returns\n -------\n service_reduced : dict or array\n Reduced service after assumption on reuse\n\n Note\n ----\n A standard sigmoid diffusion is assumed from base year to end year\n \"\"\"\n try:\n # Fraction of heat recovered until end year\n heat_recovered_p = strategy_variables[\"heat_recoved__{}\".format(enduse)]['scenario_value']\n\n if heat_recovered_p == 0:\n return service, service_techs\n else:\n # Fraction of heat recovered in current year\n sig_diff_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['heat_recovered_yr_until_changed']['scenario_value'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_midpoint'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_steepness'])\n\n heat_recovered_p_cy = sig_diff_factor * heat_recovered_p\n\n # Apply to technologies each stored in dictionary\n service_reduced_techs = {}\n for tech, service_tech in service_techs.items():\n service_reduced_techs[tech] = service_tech * (1.0 - heat_recovered_p_cy)\n\n # Apply to array\n service_reduced = service * (1.0 - heat_recovered_p_cy)\n\n return service_reduced, service_reduced_techs\n except KeyError:\n # no recycling defined\n return service, service_techs\n\ndef apply_air_leakage(\n enduse,\n strategy_variables,\n enduse_overall_change,\n service,\n service_techs,\n base_yr,\n curr_yr\n ):\n \"\"\"Reduce heating demand according to assumption on\n improvements in air leaking\n\n Arguments\n ----------\n enduse : str\n Enduse\n strategy_variables : dict\n Strategy variables\n enduse_overall_change : dict\n Sigmoid diffusion info\n service : dict or array\n Service of current year\n crit_dict : str\n Criteria to run function differently\n base_yr : int\n Base year\n curr_yr : int\n Current year\n\n Returns\n -------\n service_reduced : dict or array\n Service after assumptions on air leaking improvements\n\n Note\n ----\n A standard sigmoid diffusion is assumed from base year to end year\n \"\"\"\n try:\n # Fraction of heat recovered until end year\n air_leakage_improvement = strategy_variables[\"air_leakage__{}\".format(enduse)]['scenario_value']\n\n if air_leakage_improvement == 0:\n return service, service_techs\n else:\n air_leakage_by = 1\n\n # Fraction of heat recovered in current year\n sig_diff_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['air_leakage_yr_until_changed']['scenario_value'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_midpoint'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_steepness'])\n\n air_leakage_improvement_cy = sig_diff_factor * air_leakage_improvement\n air_leakage_cy = 1 - air_leakage_improvement_cy\n\n f_improvement = air_leakage_cy / air_leakage_by\n\n # Apply to technologies each stored in dictionary or array\n service_reduced_techs = {}\n for tech, service_tech in service_techs.items():\n service_reduced_techs[tech] = service_tech * f_improvement\n\n service_reduced = service * f_improvement\n\n return service_reduced, service_reduced_techs\n except KeyError:\n return service, service_techs\n\ndef apply_scenario_drivers(\n submodel,\n enduse,\n sector,\n fuel_y,\n dw_stock,\n region,\n gva,\n population,\n industry_gva,\n reg_scen_drivers,\n base_yr,\n curr_yr\n ):\n \"\"\"The fuel data for every end use are multiplied with respective\n scenario drivers. If no dwelling specific scenario driver is found,\n the identical fuel is returned.\n\n Arguments\n ----------\n enduse: str\n Enduse\n fuel_y : array\n Yearly fuel per fueltype\n dw_stock : object\n Dwelling stock\n region : str\n Region name\n gva : dict\n GVA\n population : dict\n Population\n reg_scen_drivers : dict\n Scenario drivers per enduse\n base_yr : int\n Base year\n curr_yr : int\n Current year\n\n Returns\n -------\n fuel_y : array\n Changed yearly fuel per fueltype\n \"\"\"\n if reg_scen_drivers is None:\n reg_scen_drivers = {}\n\n if not dw_stock:\n \"\"\"Calculate non-dwelling related scenario drivers, if no dwelling stock\n Info: No dwelling stock is defined for this submodel\n \"\"\"\n scenario_drivers = reg_scen_drivers[enduse]\n\n by_driver, cy_driver = 1, 1 #not 0\n\n for scenario_driver in scenario_drivers:\n\n # Get correct data depending on driver\n if scenario_driver == 'gva':\n by_driver_data = gva[base_yr][region]\n cy_driver_data = gva[curr_yr][region]\n\n '''if submodel == 'is_submodel':\n\n # Map enduse to SIC letter\n lu_industry_sic = lookup_tables.industrydemand_name_sic2007()\n sic_letter = lu_industry_sic[sector][sic_2007_letter]\n\n by_driver_data = industry_gva[base_yr][region][sic_lettersector]\n cy_driver_data = industry_gva[curr_yr][region][sic_letter]\n else:\n\n # Calculate overall GVA for all sectors TODO\n\n by_driver_data = gva[base_yr][region]\n cy_driver_data = gva[curr_yr][region]'''\n\n elif scenario_driver == 'population':\n by_driver_data = population[base_yr][region]\n cy_driver_data = population[curr_yr][region]\n #TODO :ADD OTHER ENDSES\n\n if math.isnan(by_driver_data):\n logging.warning(\"INF ERROR\")\n by_driver_data = 1\n if math.isnan(cy_driver_data):\n logging.warning(\"INF ERROR\")\n cy_driver_data = 1\n\n # Multiply drivers\n by_driver *= by_driver_data\n cy_driver *= cy_driver_data\n\n try:\n factor_driver = cy_driver / by_driver # FROZEN (as in chapter 3.1.2 EQ E-2)\n except ZeroDivisionError:\n factor_driver = 1\n\n if math.isnan(factor_driver):\n raise Exception(\"Error xcx\")\n\n fuel_y = fuel_y * factor_driver\n else:\n \"\"\"Scenario driver calculation based on dwelling stock\n \"\"\"\n # Test if enduse has a dwelling related scenario driver\n if hasattr(dw_stock[base_yr], enduse) and curr_yr != base_yr:\n\n # Scenariodriver of dwelling stock base year and new stock\n by_driver = getattr(dw_stock[base_yr], enduse)\n cy_driver = getattr(dw_stock[curr_yr], enduse)\n #assert by_driver != 'nan' and assert cy_driver != 'nan'\n\n # base year / current (checked)\n try:\n factor_driver = cy_driver / by_driver\n except ZeroDivisionError:\n factor_driver = 1\n\n # Check if float('nan')\n if math.isnan(factor_driver):\n logging.warning(\"Something went wrong wtih scenario\")\n factor_driver = 1\n\n #logging.debug(\"... Scenario drivers: {} {} {}\".format(\n # by_driver, cy_driver, factor_driver))\n\n fuel_y = fuel_y * factor_driver\n else:\n pass #enduse not define with scenario drivers\n\n assert math.isnan(np.sum(fuel_y)) != 'nan' #SPEED ESTING\n\n return fuel_y\n\ndef apply_specific_change(\n enduse,\n fuel_y,\n enduse_overall_change,\n strategy_variables,\n base_yr,\n curr_yr\n ):\n \"\"\"Calculates fuel based on assumed overall enduse specific\n fuel consumption changes.\n\n The changes are assumed across all fueltypes.\n Because for enduses where no technologies are defined, a linear\n diffusion is suggested to best represent multiple sigmoid efficiency\n improvements of individual technologies.\n\n Either a sigmoid standard diffusion or linear diffusion can be\n implemented. Linear is suggested.\n\n Arguments\n ----------\n enduse : str\n Enduse\n fuel_y : array\n Yearly fuel per fueltype\n enduse_overall_change : dict\n Info about how the enduse is overall changed (e.g. diff method)\n strategy_variables : dict\n Change in overall enduse for every enduse (percent ey)\n base_yr : int\n Base year\n curr_yr : int\n Current year\n\n Returns\n -------\n fuel_y : array\n Yearly new fuels\n \"\"\"\n # Fuel consumption shares in base and end year\n percent_by = 1.0\n\n percent_ey = percent_by + strategy_variables['enduse_change__{}'.format(enduse)]['scenario_value']\n\n # Share of fuel consumption difference\n diff_fuel_consump = percent_ey - percent_by\n diffusion_choice = enduse_overall_change['other_enduse_mode_info']['diff_method']\n\n if diff_fuel_consump != 0: # If change in fuel consumption\n\n # Lineare diffusion up to cy\n if diffusion_choice == 'linear':\n lin_diff_factor = diffusion_technologies.linear_diff(\n base_yr,\n curr_yr,\n percent_by,\n percent_ey,\n strategy_variables['enduse_specific_change_yr_until_changed']['scenario_value'])\n change_cy = lin_diff_factor\n\n # Sigmoid diffusion up to cy\n elif diffusion_choice == 'sigmoid':\n sig_diff_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['enduse_specific_change_yr_until_changed']['scenario_value'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_midpoint'],\n enduse_overall_change['other_enduse_mode_info']['sigmoid']['sig_steepness'])\n change_cy = diff_fuel_consump * sig_diff_factor\n\n return fuel_y * change_cy\n else:\n return fuel_y\n\ndef apply_climate_change(\n enduse,\n fuel_y,\n cooling_factor_y,\n heating_factor_y,\n enduse_space_heating,\n enduse_space_cooling\n ):\n \"\"\"Change fuel demand for heat and cooling service\n depending on changes in HDD and CDD within a region\n (e.g. climate change induced)\n\n Arguments\n ----------\n enduse : str\n Enduse\n fuel_y : array\n Yearly fuel per fueltype\n cooling_factor_y : array\n Distribution of fuel within year to days (yd)\n heating_factor_y : array\n Distribution of fuel within year to days (yd)\n enduse_space_heating : list\n Enduses defined as space heating\n enduse_space_cooling : list\n Enduses defined as space cooling\n\n Return\n ------\n fuel_y : array\n Changed yearly fuel per fueltype\n\n Note\n ----\n - `cooling_factor_y` and `heating_factor_y` are based on the sum\n over the year. Therefore it is assumed that fuel correlates\n directly with HDD or CDD.\n \"\"\"\n if enduse in enduse_space_heating:\n fuel_y = fuel_y * heating_factor_y\n elif enduse in enduse_space_cooling:\n fuel_y = fuel_y * cooling_factor_y\n\n return fuel_y\n\ndef apply_smart_metering(\n enduse,\n fuel_y,\n sm_assump,\n strategy_variables,\n base_yr,\n curr_yr\n ):\n \"\"\"Calculate fuel savings depending on smart meter penetration\n\n Arguments\n ----------\n enduse : str\n Enduse\n fuel_y : array\n Yearly fuel per fueltype\n sm_assump : dict\n smart meter assumptions\n strategy_variables : dict\n Base simulation parameters\n base_yr, curr_yr : int\n years\n\n Returns\n -------\n fuel_y : array\n New fuel per year\n\n Note\n -----\n - The smart meter penetration is assumed with a sigmoid diffusion.\n\n - In the assumptions the maximum penetration and also the\n generally fuel savings for each enduse can be defined.\n \"\"\"\n try:\n\n enduse_savings = strategy_variables['smart_meter_improvement_{}'.format(enduse)]['scenario_value']\n\n # Sigmoid diffusion up to current year\n sigm_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['smart_meter_yr_until_changed']['scenario_value'],\n sm_assump['smart_meter_diff_params']['sig_midpoint'],\n sm_assump['smart_meter_diff_params']['sig_steepness'])\n\n # Check if float\n assert isinstance(sigm_factor, float)\n\n # Improvement of smart meter penetration\n penetration_improvement = strategy_variables['smart_meter_improvement_p']['scenario_value']\n\n # Smart Meter penetration (percentage of people having smart meters)\n penetration_by = sm_assump['smart_meter_p_by']\n penetration_cy = sm_assump['smart_meter_p_by'] + sigm_factor * penetration_improvement\n\n saved_fuel = fuel_y * (penetration_cy - penetration_by) * enduse_savings\n fuel_y = fuel_y - saved_fuel\n\n return fuel_y\n\n except KeyError:\n # not defined for this enduse\n return fuel_y\n\ndef convert_service_to_p(tot_s_y, s_fueltype_tech):\n \"\"\"Calculate fraction of service for every technology\n of total service\n\n Arguments\n ----------\n tot_s_y : float\n Total yearly service\n s_fueltype_tech : dict\n Service per technology and fueltype\n\n Returns\n -------\n s_tech_p : dict\n All tecnology services are\n provided as a fraction of total service\n\n Note\n ----\n Iterate over values in dict and apply calculations\n \"\"\"\n if tot_s_y == 0:\n _total_service = 0\n else:\n _total_service = 1 / tot_s_y\n\n # Iterate all technologies and calculate fraction of total service\n s_tech_p = {}\n for tech_services in s_fueltype_tech.values():\n for tech, service_tech in tech_services.items():\n s_tech_p[tech] = _total_service * service_tech\n\n return s_tech_p\n\ndef get_service_diffusion(sig_param_tech, curr_yr):\n \"\"\"Calculate energy service fraction of technologies with increased service\n for current year based on sigmoid diffusion\n\n Arguments\n ----------\n sig_param_tech : dict\n Sigmoid diffusion parameters per technology\n curr_yr : dict\n Current year\n\n Returns\n -------\n s_tech_p : dict\n Share of service per technology of current year\n \"\"\"\n if sig_param_tech['l_parameter'] is None:\n s_tech_p = 0\n elif sig_param_tech['l_parameter'] == 'linear':\n s_tech_p = 'identical'\n else:\n s_tech_p = diffusion_technologies.sigmoid_function(\n curr_yr,\n sig_param_tech['l_parameter'],\n sig_param_tech['midpoint'],\n sig_param_tech['steepness'])\n\n return s_tech_p\n\ndef calc_service_switch(\n enduse,\n s_tech_y_cy,\n all_technologies,\n sig_param_tech,\n curr_yr,\n base_yr,\n sector,\n crit_switch_happening\n ):\n \"\"\"Apply change in service depending on defined service switches.\n\n The service which is fulfilled by new technologies as defined\n in the service switches is substracted of the replaced\n technologies proportionally to the base year distribution\n of these technologies.\n\n Arguments\n ---------\n tot_s_yh_cy : array\n Hourly service of all technologies\n all_technologies : dict\n Technologies to iterate\n sig_param_tech : dict\n Sigmoid diffusion parameters\n curr_yr : int\n Current year\n\n Returns\n -------\n switched_s_tech_y_cy : dict\n Service per technology in current year after switch in a year\n \"\"\"\n # ----------------------------------------\n # Test wheter switch is defined or not\n # ----------------------------------------\n crit_switch_service = fuel_service_switch.get_switch_criteria(\n enduse,\n sector,\n crit_switch_happening,\n base_yr,\n curr_yr)\n\n # ----------------------------------------\n # Calculate switch\n # ----------------------------------------\n if crit_switch_service:\n\n switched_s_tech_y_cy = {}\n\n # Service of all technologies\n service_all_techs = sum(s_tech_y_cy.values())\n\n for tech in all_technologies:\n\n # Calculated service share per tech for cy with sigmoid parameters\n s_tech_cy_p = get_service_diffusion(\n sig_param_tech[tech], curr_yr)\n\n if s_tech_cy_p == 'identical':\n switched_s_tech_y_cy[tech] = s_tech_y_cy[tech]\n else:\n switched_s_tech_y_cy[tech] = service_all_techs * s_tech_cy_p\n\n assert switched_s_tech_y_cy[tech] >= 0\n\n return switched_s_tech_y_cy\n else:\n return s_tech_y_cy\n\ndef apply_cooling(\n enduse,\n fuel_y,\n strategy_variables,\n cooled_floorarea_p_by,\n other_enduse_mode_info,\n base_yr,\n curr_yr):\n \"\"\"Apply changes for cooling enduses depending\n on assumption of how much of the floor area in percent\n is cooled\n\n It is aassumption a linear correlation between the\n percentage of cooled floor space (area) and energy demand.\n\n Arguments\n ---------\n enduse : str\n Enduse\n fuel_y : array\n Annual fuel demand\n strategy_variables : dict\n Strategy variables\n cooled_floorarea_p_by : dict\n Assumption about cooling floor area in base year\n other_enduse_mode_info : dict\n diffusion parameters\n base_yr : int\n Base year\n curr_yr : int\n Current year\n\n Returns\n -------\n fuel_y : array\n Fuel array (either changed fuel depending on cooling percentage)\n of identical array\n \"\"\"\n try:\n\n # Floor area share cooled in end year\n cooled_floorearea_p_ey = cooled_floorarea_p_by + strategy_variables[\"cooled_floorarea__{}\".format(enduse)]['scenario_value']\n\n # Fraction of heat recovered up to current year\n sig_diff_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['cooled_floorarea_yr_until_changed']['scenario_value'],\n other_enduse_mode_info['sigmoid']['sig_midpoint'],\n other_enduse_mode_info['sigmoid']['sig_steepness'])\n\n # Additionall floor area\n additional_floor_area_p = sig_diff_factor * (cooled_floorearea_p_ey - cooled_floorarea_p_by)\n\n cooled_floorarea_p_cy = cooled_floorarea_p_by + additional_floor_area_p\n\n # Calculate factor\n floorare_cooling_factor = cooled_floorarea_p_cy / cooled_floorarea_p_by\n\n # Apply factor\n fuel_y = fuel_y * floorare_cooling_factor\n return fuel_y\n\n except KeyError:\n # no cooling defined for enduse\n return fuel_y\n\ndef industry_enduse_changes(\n enduse,\n sector,\n base_yr,\n curr_yr,\n strategy_variables,\n fuels,\n other_enduse_mode_info,\n assumptions\n ):\n \"\"\"This function changes the demand if the enduse\n is a an industrial enduse depending on assumed\n industry related scenario paramters\n\n Arguments\n ---------\n enduse : str\n Enduse\n sector : str\n Sector\n curr_yr : int\n Current year\n strategy_variables : dict\n All strategy variables\n fuels : array\n Annual fuels\n\n Returns\n --------\n fuels : np.array\n Changed fuels depending on scenario\n\n Info\n ----\n OLD MODEL TODO\n\n \"\"\"\n factor = 1\n\n if enduse == \"is_low_temp_process\":\n\n # Diffusion of policy\n #cy_factor = by_value / cy_value / by_value\n #Multiply fuels\n #fuels = fuels * cy_factor\n\n '''\n Theoretical maximal potential for every sector\n --> improvement in % of every sector?\n\n\n '''\n pass\n elif enduse == 'is_high_temp_process':\n\n\n if sector == 'basic_metals':\n\n # Calculate factor depending on fraction of hot and cold steel rolling process\n factor = hot_cold_process(\n base_yr,\n curr_yr,\n strategy_variables,\n other_enduse_mode_info,\n assumptions)\n\n #elif sector == 'non_metallic_mineral_products':\n\n # # Calculate factor depending on cement processes\n\n else:\n pass\n\n fuels_out = fuels * factor\n\n return fuels_out\n\ndef hot_cold_process(\n base_yr,\n curr_yr,\n strategy_variables,\n other_enduse_mode_info,\n assumptions\n ):\n \"\"\"Calculate factor based on the fraction of hot\n and cold rolling processes in steel manufacturing.\n The fraction of either process is calculated based on\n the scenario input of the future share of cold rollling\n processes. A sigmoid diffusion towards this fucture defined\n fraction is implemented.\n\n Arguments\n ----------\n base_yr : int\n Base year\n curr_yr : int\n Current year\n strategy_variables : dict\n Strategy variables\n other_enduse_mode_info : dict\n Sigmoid diffusion parameters\n assumptions : dict\n Assumptions including efficiencies of either process\n and the base year share\n\n Returns\n -------\n factor : float\n Factor to change energy demand\n \"\"\"\n\n # Reduce demand depending on fraction of hot and cold steel rolling process\n p_cold_rolling_by = assumptions.p_cold_rolling_steel_by\n p_hot_rolling_by = 1.0 - p_cold_rolling_by\n\n # Get sigmoid transition for share in rolling\n sig_diff_factor = diffusion_technologies.sigmoid_diffusion(\n base_yr,\n curr_yr,\n strategy_variables['hot_cold_rolling_yr_until_changed']['scenario_value'],\n other_enduse_mode_info['sigmoid']['sig_midpoint'],\n other_enduse_mode_info['sigmoid']['sig_steepness'])\n\n # Difference p cold rolling\n diff_cold_rolling = strategy_variables['p_cold_rolling_steel']['scenario_value'] - p_cold_rolling_by\n\n # Difference until cy\n diff_cold_rolling_cy = sig_diff_factor * diff_cold_rolling\n\n # Calculate cy p\n p_cold_rolling_cy = p_cold_rolling_by + diff_cold_rolling_cy\n p_hot_rolling_cy = 1 - p_cold_rolling_cy\n\n # Calculate factor\n eff_cold = assumptions.eff_cold_rolling_process\n eff_hot = assumptions.eff_hot_rolling_process\n\n p_by = p_cold_rolling_by * eff_cold + p_hot_rolling_by * eff_hot\n p_cy = p_cold_rolling_cy * eff_cold + p_hot_rolling_cy * eff_hot\n\n factor = p_cy / p_by\n\n return factor\n","sub_path":"energy_demand/enduse_func.py","file_name":"enduse_func.py","file_ext":"py","file_size_in_byte":57900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"140602583","text":"#Created By: Logan Fillo\n#Created On: 2019-03-11\n\n\"\"\"\nThis module contains the functionality\nfor creating test cases\n\"\"\"\n\nimport sys\n\nimport constants as const\nimport vehicle\nimport scene\nimport object\n\n\nclass SimTestCase:\n \"\"\"\n A class representing a simulation test case\n \"\"\"\n def __init__(self, name):\n \"\"\"\n Constructs a SimTestCase\n \"\"\"\n self.name = name\n self.is_dynamic = const.DEFAULT_IS_DYNAMIC\n self.timeout = const.DEFAULT_TIMEOUT\n self.vehicle = None\n self.scene = None\n self.objects = None\n\n\nclass SimTestBuilder:\n \"\"\"\n A helper class for building simulation tests\n \"\"\"\n def __init__(self):\n \"\"\"\n Construct a SimTestBuilder\n \"\"\"\n self._curr_test = None\n self._curr_vhcl = None\n self._curr_scene = None\n self._curr_objs = None\n\n # For storing forall data\n self._glob_timeout = None\n self._glob_is_dynamic = None\n self._glob_vhcl_data = dict()\n self._glob_scene_data = dict()\n self._glob_objs = []\n\n def build_test_case(self, name):\n \"\"\"\n Constructs the objects needed for a test\n and sets their global data, if any\n\n :param name: test case name\n \"\"\"\n self._curr_test = SimTestCase(name)\n self._curr_vhcl = vehicle.Vehicle()\n self._curr_scene = scene.Scene()\n self._curr_objs = []\n self._set_globals()\n\n def get_result(self):\n \"\"\"\n Returns the current test case configured with\n the current vehicle, scene, and objects\n \"\"\"\n self._curr_test.vehicle = self._curr_vhcl\n self._curr_test.scene = self._curr_scene\n self._curr_test.objects = self._curr_objs\n return self._curr_test\n\n def add_object(self, name, model,\n x_pos, y_pos, z_pos,\n r_rot, p_rot, y_rot,\n is_global):\n sim_object = object.SimObject(name, model,\n x_pos, y_pos, z_pos,\n r_rot, p_rot, y_rot)\n if is_global:\n self._glob_objs.append(sim_object)\n else:\n self._curr_objs.append(sim_object)\n\n def set_wave_scale(self, scale, is_global):\n self._set_scene_elem(const.SCENE_WAVE_SCALE,\n 1*(10**(-scale)), is_global)\n\n def set_timeout(self, timeout, is_global):\n if is_global:\n self._glob_timeout = timeout\n else:\n self._curr_test.timeout = timeout\n\n def use_dynamics(self, is_dynamic, is_global):\n if is_global:\n self._glob_is_dynamic = is_dynamic\n else:\n self._curr_test.is_dynamic = is_dynamic\n\n def set_vehicle_position(self, x, y, z, is_global):\n if z < 0:\n sys.exit(\"Error: vehicle z position cannot be less than 0 (above water)\")\n self._set_vhcl_elem(const.VEHICLE_X_POS, x, is_global)\n self._set_vhcl_elem(const.VEHICLE_Y_POS, y, is_global)\n self._set_vhcl_elem(const.VEHICLE_Z_POS, z, is_global)\n\n def set_vehicle_rotation(self, r, p, y, is_global):\n self._set_vhcl_elem(const.VEHICLE_R_ROT, r, is_global)\n self._set_vhcl_elem(const.VEHICLE_P_ROT, p, is_global)\n self._set_vhcl_elem(const.VEHICLE_Y_ROT, y, is_global)\n\n def _set_globals(self):\n if self._glob_timeout is not None:\n self._curr_test.timeout = self._glob_timeout\n if self._glob_is_dynamic is not None:\n self._curr_test.is_dynamic = self._glob_is_dynamic\n if not len(self._glob_objs) == 0:\n self._curr_objs = self._glob_objs[:]\n for elem in self._glob_vhcl_data:\n self._curr_vhcl.data[elem] = self._glob_vhcl_data[elem]\n for elem in self._glob_scene_data:\n self._curr_scene.data[elem] = self._glob_scene_data[elem]\n\n def _set_vhcl_elem(self, elem, val, is_global):\n if is_global:\n self._glob_vhcl_data[elem] = val\n else:\n self._curr_vhcl.data[elem] = val\n\n def _set_scene_elem(self, elem, val, is_global):\n if is_global:\n self._glob_scene_data[elem] = val\n else:\n self._curr_scene.data[elem] = val\n","sub_path":"src/simulator/pysimtest/_pysimtest/case.py","file_name":"case.py","file_ext":"py","file_size_in_byte":4351,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"62407831","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 24 14:30:05 2019\n\n@author: eo\n\"\"\"\n\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Add local path\n\nimport os\nimport sys\n\ndef find_path_to_local(target_folder = \"local\"):\n \n # Skip path finding if we successfully import the dummy file\n try:\n from local.dummy import dummy_func; dummy_func(); return\n except ImportError:\n print(\"\", \"Couldn't find local directory!\", \"Searching for path...\", sep=\"\\n\")\n \n # Figure out where this file is located so we can work backwards to find the target folder\n file_directory = os.path.dirname(os.path.abspath(__file__))\n path_check = []\n \n # Check parent directories to see if we hit the main project directory containing the target folder\n prev_working_path = working_path = file_directory\n while True:\n \n # If we find the target folder in the given directory, add it to the python path (if it's not already there)\n if target_folder in os.listdir(working_path):\n if working_path not in sys.path:\n tilde_swarm = \"~\"*(4 + len(working_path))\n print(\"\\n{}\\nPython path updated:\\n {}\\n{}\".format(tilde_swarm, working_path, tilde_swarm))\n sys.path.append(working_path)\n break\n \n # Stop if we hit the filesystem root directory (parent directory isn't changing)\n prev_working_path, working_path = working_path, os.path.dirname(working_path)\n path_check.append(prev_working_path)\n if prev_working_path == working_path:\n print(\"\\nTried paths:\", *path_check, \"\", sep=\"\\n \")\n raise ImportError(\"Can't find '{}' directory!\".format(target_folder))\n \nfind_path_to_local()\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Imports\n\nimport cv2\nimport numpy as np\n\nfrom collections import deque\n\nfrom local.lib.ui_utils.local_ui.drawing import Entity_Drawer\n\nfrom local.eolib.utils.cli_tools import Color\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Define Classes\n\n\nclass Simple_Window:\n \n # ................................................................................................................. \n \n def __init__(self, window_name,\n frame_wh = None,\n provide_mouse_xy = False,\n create_on_startup = True):\n \n # Get window name so we can continue to refer to this window!\n self.window_name = window_name\n \n # Allocate variables for (potential) mouse-xy feedback\n self.enable_mouse_feedback = provide_mouse_xy\n self._mouse_feedback = None\n \n # Variables for recording the window position\n self.x_px = None\n self.y_px = None\n \n # Variables used to record the size of the displayed image\n self.window_wh_is_set = False\n self.width = None\n self.height = None\n if frame_wh is not None:\n self.set_window_wh(*frame_wh)\n \n # Create the display, if needed\n if create_on_startup:\n self.create_window()\n \n # ................................................................................................................. \n \n def __repr__(self):\n return \"{} ({})\".format(self.class_name, self.window_name)\n \n # ................................................................................................................. \n \n @property\n def class_name(self):\n return self.__class__.__name__\n \n # ................................................................................................................. \n \n @property\n def mouse_xy(self):\n return self._mouse_feedback.xy if self.enable_mouse_feedback else None\n \n # ................................................................................................................. \n \n def set_window_wh(self, window_width, window_height): \n self.width = window_width\n self.height = window_height\n self.window_wh_is_set = True\n \n # ................................................................................................................. \n \n def get_window_name(self):\n return self.window_name\n \n # .................................................................................................................\n \n def get_window_wh(self):\n return self.width, self.height\n \n # ................................................................................................................. \n \n def imshow(self, display_frame):\n \n # Check if the window exists and whether the input data is valid\n window_exists = self.exists()\n valid_frame_data = (display_frame is not None)\n \n # Only update showing if the window exists & a valid image is supplied\n if window_exists and valid_frame_data:\n cv2.imshow(self.window_name, display_frame)\n \n return window_exists\n \n # ................................................................................................................. \n \n def imshow_blank(self, blank_wh = None):\n \n # Set blank size if needed\n if self.width is None and self.height is None:\n blank_wh = (500, 30) if blank_wh is None else blank_wh\n else:\n blank_wh = (self.width, self.height)\n \n # Only update showing if the window exists\n window_exists = self.exists()\n if window_exists:\n blank_image = np.zeros((blank_wh[1], blank_wh[0], 3), dtype=np.uint8)\n cv2.imshow(self.window_name, blank_image)\n \n return window_exists\n \n # ................................................................................................................. \n \n def move_corner_pixels(self, x_pixels, y_pixels, create_if_missing = True):\n \n ''' Move the window corner to a screen position, specified in pixels '''\n \n # Make sure the window exists before we move it around\n self._create_window_if_missing(create_if_missing)\n \n # Force inputs to integers, since floats aren't accepted\n self.x_px = int(round(x_pixels))\n self.y_px = int(round(y_pixels))\n \n cv2.moveWindow(self.window_name, self.x_px, self.y_px)\n \n return self\n \n # ................................................................................................................. \n \n def move_center_pixels(self, x_pixels, y_pixels, frame_width = None, frame_height = None,\n create_if_missing = True):\n \n '''\n Move the window center to a screen position, specified in pixels\n '''\n \n # Make sure the window exists before we move it around\n self._create_window_if_missing(create_if_missing)\n \n # Update frame width/height if needed\n self.width = frame_width if frame_width is not None else self.width\n self.height = frame_height if frame_height is not None else self.height\n \n # Get the frame half sizing for centering\n try:\n half_frame_width = self.width / 2\n half_frame_height = self.height / 2\n except TypeError:\n raise AttributeError(\"Can't move the window without knowing it's frame width/height!\")\n \n # Find window corner location, so that the frame center lands at the target screen position\n window_corner_x_px = x_pixels - half_frame_width\n window_corner_y_px = y_pixels - half_frame_height\n \n return self.move_corner_pixels(window_corner_x_px, window_corner_y_px)\n \n # .................................................................................................................\n \n def exists(self):\n return cv2.getWindowProperty(self.window_name, cv2.WND_PROP_AUTOSIZE) > 0\n \n # ................................................................................................................. \n \n def close(self):\n if self.exists(): \n cv2.destroyWindow(self.window_name)\n \n # ................................................................................................................. \n \n def attach_callback(self, mouse_callback, callback_data = {}, create_if_missing = True):\n \n self._create_window_if_missing(create_if_missing)\n cv2.setMouseCallback(self.window_name, mouse_callback, callback_data)\n \n return self\n \n # ................................................................................................................. \n \n def add_trackbar(self, label, initial_value, max_value): \n cv2.createTrackbar(label, self.window_name, initial_value, max_value, lambda x: None)\n \n # ................................................................................................................. \n \n def set_trackbar(self, label, new_value):\n cv2.setTrackbarPos(label, self.window_name, new_value)\n \n # ................................................................................................................. \n \n def read_trackbar(self, label):\n '''Returns current trackbar value (integer)'''\n return cv2.getTrackbarPos(label, self.window_name)\n \n # ................................................................................................................. \n \n def create_window(self):\n \n # Create window\n cv2.namedWindow(self.window_name)\n self.imshow_blank()\n self.move_corner_pixels(x_pixels = 50, y_pixels = 50, create_if_missing = False)\n \n # Enable mouse xy reporting, if needed\n if self.enable_mouse_feedback:\n self._mouse_feedback = Mouse_Follower()\n cv2.setMouseCallback(self.window_name, self._mouse_feedback)\n \n return self\n \n # ................................................................................................................. \n \n def _create_window_if_missing(self, create_if_missing = True):\n \n ''' Function which checks that the window exists, and if not, will create it '''\n \n window_is_missing = (not self.exists())\n if window_is_missing and create_if_missing:\n self.create_window()\n \n return\n \n # .................................................................................................................\n # .................................................................................................................\n\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n \nclass Control_Window(Simple_Window):\n \n # .................................................................................................................\n \n def __init__(self, window_name, control_list_json, frame_wh = (500, 30),\n create_on_startup = True):\n \n # Inherit from simple window\n provide_mouse_xy = False\n super().__init__(window_name, frame_wh, provide_mouse_xy, create_on_startup)\n \n # Set width/height of control window (which is normally just a small blackout area)\n self.width, self.height = frame_wh\n \n # Store controls for future reference if needed\n self.control_list_json = control_list_json\n \n # Draw initial blank frame\n self._blank_frame = np.zeros((self.height, self.width, 3), dtype=np.uint8)\n self.imshow(self._blank_frame) \n \n # Allocate space to store trackbar info\n self.tooltip_dict = {}\n self.units_dict = {}\n self.menu_labels_dict = {}\n self.variable_to_label_lut = []\n self.trackbar_position_dict = {}\n self.trackbar_minimums_dict = {}\n self.variable_name_list = []\n \n # Allocate space for mapping functions\n self._map_to_raw_func_dict = {}\n self._raw_to_map_func_dict = {}\n \n # Create all the trackbars and read their initial values\n self._build_trackbars(self.control_list_json)\n \n # .................................................................................................................\n \n def print_info(self, return_string = True):\n \n # Loop through every variable and construct a nice info printout\n print_str_list = []\n for each_variable in self.variable_name_list:\n \n # Pull out control info\n control_label = self.variable_to_label_lut[each_variable]\n control_units = self.units_dict[each_variable]\n control_tooltip = self.tooltip_dict[each_variable]\n control_menu_labels = self.menu_labels_dict[each_variable]\n \n # Bail if no tooltip is present. If a tooltip comes in as a list, treat it as separate lines to print out\n if control_tooltip is None:\n continue\n elif control_tooltip == \"\":\n tooltip_str = \" No tooltip!\"\n elif type(control_tooltip) in {tuple, list}:\n tooltip_str = \"\\n\".join([\" {}\".format(each_line) for each_line in control_tooltip])\n else:\n tooltip_str = \" {}\".format(control_tooltip)\n \n # Set up the heading title string for each variable print out\n heading_str = control_label\n if control_units:\n heading_str += \" ({}):\".format(control_units)\n elif len(control_menu_labels) > 1:\n menu_labels_str = \" / \".join(control_menu_labels)\n heading_str += \" ({}):\".format(menu_labels_str)\n else:\n heading_str += \":\"\n \n # Combine the label heading and tooltip strings\n new_print_str = [\"\",\n Color(heading_str).bold.str,\n tooltip_str]\n \n # Add to the list of variable info printouts\n print_str_list += new_print_str\n \n # Finally, return or print out all the control info\n print_out_str = \"\\n\".join(print_str_list)\n if return_string:\n return print_out_str\n else:\n print(print_out_str)\n \n # .................................................................................................................\n \n def set_trackbar(self, variable_name, mapped_value):\n \n map_to_raw_func = self._map_to_raw_func_dict[variable_name]\n raw_value = map_to_raw_func(mapped_value)\n \n control_label = self.variable_to_label_lut[variable_name]\n self._set_trackbar_raw(control_label, raw_value)\n \n # .................................................................................................................\n \n def read_trackbar(self, variable_name, force_minimum = True):\n \n # Get the function needed to map from raw (trackbar position) values to mapped values\n raw_to_map_func = self._raw_to_map_func_dict[variable_name]\n \n # Get trackbar position\n control_label = self.variable_to_label_lut[variable_name]\n raw_value = self._read_trackbar_raw(control_label)\n \n # Stop the user from going below some minimum trackbar position\n if force_minimum:\n tb_min = self.trackbar_minimums_dict[control_label]\n if raw_value < tb_min:\n raw_value = tb_min\n self._set_trackbar_raw(control_label, raw_value)\n \n # Check if the trackbar position changed\n value_changed = (raw_value != self.trackbar_position_dict[control_label])\n if value_changed:\n self.trackbar_position_dict[control_label] = raw_value\n \n # Finally, return the variable in it's proper representation\n map_value = raw_to_map_func(raw_value)\n \n return value_changed, map_value\n \n # .................................................................................................................\n \n def read_trackbar_changes(self):\n \n # Loop through all trackbar, reading values\n # Need to check values against some recorded value, and report them if they changed...\n # Report back using variable_name and new value\n value_changes_dict = {}\n for each_variable in self.variable_name_list:\n value_changed, map_value = self.read_trackbar(each_variable)\n if value_changed:\n value_changes_dict.update({each_variable: map_value})\n \n return value_changes_dict\n \n # .................................................................................................................\n \n def _set_trackbar_raw(self, label, raw_value): \n if self.exists():\n cv2.setTrackbarPos(label, self.window_name, raw_value)\n \n # .................................................................................................................\n \n def _read_trackbar_raw(self, label):\n return cv2.getTrackbarPos(label, self.window_name) if self.exists() else self.trackbar_position_dict[label]\n \n # .................................................................................................................\n \n def _build_trackbars(self, control_list):\n \n # Allocate space to store trackbar info\n self.tooltip_dict = {}\n self.units_dict = {}\n self.menu_labels_dict = {}\n self.variable_to_label_lut = {}\n self.trackbar_position_dict = {}\n self.trackbar_minimums_dict = {}\n self.variable_name_list = []\n \n for each_entry in control_list:\n \n '''\n print(\"\")\n print(\"({})\".format(os.path.basename(__file__)))\n print(\"CONTROL ENTRY:\", each_entry)\n print(\"Control_type\", control_type)\n '''\n \n # Get important identifying info\n control_label = each_entry[\"label\"]\n variable_name = each_entry[\"variable_name\"]\n control_type = each_entry[\"control_type\"]\n visible = each_entry.get(\"visible\", True)\n \n # Skip over any non-visible controls\n if not visible:\n continue\n \n # Configure each control and figure out the trackbar (initial) settings\n config_function = self._control_type_lookup(control_type)\n tb_minimum, tb_maximum, tb_initial = config_function(each_entry) \n self.add_trackbar(control_label, tb_initial, tb_maximum)\n \n # Store the units and tooltip info so we can print it out later\n self.tooltip_dict[variable_name] = each_entry.get(\"tooltip\", \"\")\n self.units_dict[variable_name] = each_entry.get(\"units\", \"\")\n self.menu_labels_dict[variable_name], _ = zip(*each_entry.get(\"option_label_value_list\", [(\"\", \"\")]))\n \n # Store data we'll need for reading the trackbars later\n self.variable_to_label_lut[variable_name] = control_label\n self.trackbar_position_dict[control_label] = tb_initial\n self.trackbar_minimums_dict[control_label] = tb_minimum\n self.variable_name_list.append(variable_name)\n \n # .................................................................................................................\n \n def _control_type_lookup(self, control_type):\n\n # Use dictionary as a simple lookup for matching control types with configuration functions\n ctrl_type_lut = {\"toggle\": self._toggle_config,\n \"slider\": self._slider_config,\n \"numentry\": self._numentry_config,\n \"menu\": self._menu_config,\n \"button\": self._button_config} \n \n return ctrl_type_lut[control_type]\n \n # .................................................................................................................\n \n def _toggle_config(self, config_data):\n \n # Expects\n '''\n variable_name, label\n default_value,\n tooltip, visible\n '''\n \n # Pull out some relevant data for convenience\n variable_name = config_data[\"variable_name\"]\n default_value = config_data.get(\"default_value\", 0)\n \n # Get the mapping functions based on the config data\n raw_to_map_func, map_to_raw_func = bool_to_int()\n \n # Store the mapping functions so we can use them when reading/setting the trackbar\n self._map_to_raw_func_dict[variable_name] = map_to_raw_func\n self._raw_to_map_func_dict[variable_name] = raw_to_map_func\n \n # Get the default and maximum trackbar values\n trackbar_initial = map_to_raw_func(default_value)\n trackbar_minimum = 0\n trackbar_maximum = 1\n \n return trackbar_minimum, trackbar_maximum, trackbar_initial\n \n # ................................................................................................................. \n \n def _slider_config(self, config_data):\n \n # Expects\n '''\n variable_name, label\n default_value\n min_value, max_value, step_size \n units, return_type, zero_referenced,\n tooltip, visible\n '''\n \n # Pull out some relevant data for convenience\n variable_name = config_data[\"variable_name\"]\n default_value = config_data.get(\"default_value\", 0)\n min_value = config_data[\"min_value\"]\n max_value = config_data[\"max_value\"]\n step_size = config_data.get(\"step_size\", 1)\n return_type = return_type_strings_to_functions(config_data.get(\"return_type\", None))\n zero_referenced = config_data.get(\"zero_referenced\", False)\n \n # Get the mapping functions based on the config data\n if zero_referenced:\n raw_to_map_func, map_to_raw_func = minceil_affine(min_value, max_value, step_size, return_type)\n else:\n raw_to_map_func, map_to_raw_func = simple_affine(min_value, max_value, step_size, return_type)\n \n # Store the mapping functions so we can use them when reading/setting the trackbar\n self._map_to_raw_func_dict[variable_name] = map_to_raw_func\n self._raw_to_map_func_dict[variable_name] = raw_to_map_func\n \n # Get the default and maximum trackbar values\n trackbar_initial = map_to_raw_func(default_value)\n trackbar_minimum = map_to_raw_func(min_value)\n trackbar_maximum = map_to_raw_func(max_value)\n \n return trackbar_minimum, trackbar_maximum, trackbar_initial\n \n # .................................................................................................................\n \n def _numentry_config(self, config_data):\n \n # Expects\n '''\n variable_name,\n label,\n default_value,\n min_value,\n max_value,\n step_size = 1,\n units = None,\n return_type = float,\n zero_referenced = False,\n force_min = True,\n force_max = True,\n force_step = True,\n tooltip = \"\",\n visible = True\n '''\n \n # Pull outdata that allows us to createa regular slider \n # (numerical entry doesn't behave differently from a slider when using the local ui!)\n grab_slider_keys = [\"variable_name\", \"label\", \"default_value\",\n \"min_value\", \"max_value\", \"step_size\",\n \"units\", \"return_type\", \"zero_referenced\",\n \"tooltip\", \"visible\"]\n slider_config_data = {each_key: config_data[each_key] for each_key in grab_slider_keys}\n \n return self._slider_config(slider_config_data)\n \n # .................................................................................................................\n \n def _menu_config(self, config_data):\n \n # Expects\n '''\n variable_name,\n label,\n default_value,\n option_label_value_dict,\n tooltip = \"\",\n visible = True\n '''\n \n # Pull out some relevant data for convenience\n variable_name = config_data[\"variable_name\"]\n default_value = config_data.get(\"default_value\", 0)\n option_label_value_list = config_data[\"option_label_value_list\"]\n \n # Separate the labels and values, since we only need the values for local usage\n label_list, value_list = list(zip(*option_label_value_list))\n \n # Get the mapping functions based on the config data\n raw_to_map_func, map_to_raw_func = value_list_lookup(value_list)\n \n # Store the mapping functions so we can use them when reading/setting the trackbar\n self._map_to_raw_func_dict[variable_name] = map_to_raw_func\n self._raw_to_map_func_dict[variable_name] = raw_to_map_func\n \n # Get the default and maximum trackbar values\n trackbar_initial = map_to_raw_func(default_value)\n trackbar_minimum = 0\n trackbar_maximum = len(value_list) - 1\n \n return trackbar_minimum, trackbar_maximum, trackbar_initial\n \n # .................................................................................................................\n \n def _button_config(self, config_data):\n \n # Expects\n '''\n variable_name,\n label,\n default_value,\n return_type = bool,\n tooltip = \"\",\n visible = True\n '''\n \n # Pull out some relevant data for convenience\n variable_name = config_data[\"variable_name\"]\n default_value = config_data.get(\"default_value\", 0)\n \n # Get the mapping functions based on the config data\n set_trackbar_func = self._set_trackbar_raw\n raw_to_map_func, map_to_raw_func = button_map(variable_name, set_trackbar_func)\n \n # Store the mapping functions so we can use them when reading/setting the trackbar\n self._map_to_raw_func_dict[variable_name] = map_to_raw_func\n self._raw_to_map_func_dict[variable_name] = raw_to_map_func\n \n # Get the default and maximum trackbar values\n trackbar_initial = map_to_raw_func(default_value)\n trackbar_minimum = 0\n trackbar_maximum = 1\n \n return trackbar_minimum, trackbar_maximum, trackbar_initial\n \n # .................................................................................................................\n \n # .................................................................................................................\n # .................................................................................................................\n\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n\n\nclass Slideshow_Window(Simple_Window):\n \n # .................................................................................................................\n \n def __init__(self, window_name,\n frame_wh = None,\n missing_image_test = \"No image...\",\n max_storage = 10):\n \n # Inherit from parent class\n provide_mouse_xy = False\n super().__init__(window_name, frame_wh, provide_mouse_xy)\n \n # Initialize storage variables\n self.frame_deck = self._initialize_empty_frame_deck(missing_image_test, max_storage)\n self.current_select = 0\n self._update_enabled = True\n \n # Add a trackbar to control access to selecting which image to display\n self._trackbar_enable_label = \"Enable Updates\"\n self._trackbar_select_label = \"Image select\"\n self.add_trackbar(self._trackbar_enable_label, 1, 1)\n self.add_trackbar(self._trackbar_select_label, self.current_select, max_storage - 1)\n \n # Draw initial image\n self.imshow_by_index()\n \n # .................................................................................................................\n \n def _initialize_empty_frame_deck(self, missing_image_test, deque_size, default_blank_wh = (360, 240)):\n \n # Create new deque for storing 'slideshow' images\n new_deque = deque([], maxlen = deque_size)\n \n # Create blank frame\n frame_width = self.width if self.width else default_blank_wh[0]\n frame_height = self.height if self.height else default_blank_wh[1]\n \n # Draw an empty frame with some text indicating that no image is available\n blank_frame = np.zeros((frame_height, frame_width, 3), dtype=np.uint8)\n cv2.putText(blank_frame, missing_image_test,\n (10, 35), cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 255, 255), 1, cv2.LINE_AA)\n \n # Fill in frame deque with blank frames\n for k in range(deque_size):\n new_deque.append(blank_frame.copy())\n \n return new_deque\n \n # .................................................................................................................\n \n def imshow(self, display_frame):\n \n # Only update if the window exists\n window_exists = self.exists()\n if window_exists:\n \n # Only update the frame deck (and display) if the slideshow updates are still enabled\n if self._update_enabled:\n self.frame_deck.appendleft(display_frame)\n self.imshow_by_index()\n \n return window_exists\n \n # .................................................................................................................\n \n def imshow_by_index(self, index_select = None):\n \n # Automatically use the current index if one isn't provided\n if index_select is None:\n index_select = self.current_select\n \n # Only update if the window exists\n window_exists = self.exists()\n if window_exists:\n cv2.imshow(self.window_name, self.frame_deck[index_select])\n \n return window_exists\n \n # .................................................................................................................\n \n def read_trackbars(self):\n \n # Determine if updates are enabled\n self._update_enabled = self.read_trackbar(self._trackbar_enable_label)\n \n # Determine if we need to update the displayed index\n new_select = self.read_trackbar(self._trackbar_select_label)\n if new_select != self.current_select:\n self.current_select = new_select\n self.imshow_by_index(new_select)\n \n # .................................................................................................................\n # .................................................................................................................\n\n\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n\n\nclass Max_WH_Window(Simple_Window):\n \n # .................................................................................................................\n \n def __init__(self, window_name,\n frame_wh = None,\n max_wh = None,\n interpolation_type = cv2.INTER_NEAREST,\n provide_mouse_xy = False,\n create_on_startup = True):\n \n # Inherit from parent class\n super().__init__(window_name, frame_wh, provide_mouse_xy, create_on_startup)\n \n # Variables for limiting frame size\n self.interpolation_type = interpolation_type\n self.max_width = None\n self.max_height = None\n self._check_resize = False\n if max_wh is not None:\n self._check_resize = True\n self.max_width, self.max_height = max_wh\n\n # ................................................................................................................. \n \n def imshow(self, display_frame):\n \n # Check if the window exists (by looking for window properties)\n window_exists = self.exists()\n \n # Don't do anything if a valid frame isn't supplied\n if display_frame is None:\n return self.exists()\n \n # Only update showing if the window exists\n if window_exists:\n cv2.imshow(self.window_name, self._scale_to_max_wh(display_frame))\n \n return window_exists\n \n # ................................................................................................................. \n \n def _scale_to_max_wh(self, display_frame):\n \n # Don't do anything if we're not checking for resizing\n if not self._check_resize:\n return display_frame\n \n # Check if we need to resize the displayed frame\n display_height, display_width = display_frame.shape[0:2]\n needs_resize = (display_width > self.max_width) or (display_height > self.max_height)\n if not needs_resize:\n return display_frame\n \n width_scale = display_width / self.max_width\n height_scale = display_height / self.max_height\n max_scale = max(width_scale, height_scale)\n \n # Figure out scaled width/height values and apply resizing!\n scaled_width = int(display_width / max_scale)\n scaled_height = int(display_height / max_scale)\n return cv2.resize(display_frame, dsize = (scaled_width, scaled_height),\n interpolation = self.interpolation_type)\n \n # .................................................................................................................\n # .................................................................................................................\n\n\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n# /////////////////////////////////////////////////////////////////////////////////////////////////////////////////////\n\n\nclass Drawing_Window(Simple_Window):\n \n # .................................................................................................................\n \n def __init__(self, window_name, frame_wh, drawing_json,\n border_size_px = 60, create_on_startup = True):\n \n # Don't pass a real frame size on initialization, we want to make sure the drawing\n # uses the actual displayed frame size, so everything scales properly...\n # but we don't know this sizing until we get our first .imshow() call! So figure out sizing there\n initial_frame_wh = None\n \n # Inherit from parent class\n drawing_name = \"{} (Drawing)\".format(window_name)\n provide_mouse_xy = False\n super().__init__(drawing_name, initial_frame_wh, provide_mouse_xy, create_on_startup = False)\n \n # Convert drawing json data to entity drawing inputs\n self.drawing_variable_name = drawing_json[\"variable_name\"]\n min_max_entities = drawing_json[\"min_max_entities\"]\n min_max_points = drawing_json[\"min_max_points\"]\n real_border_size = (border_size_px if drawing_json[\"out_of_bounds\"] else 0)\n default_entities = drawing_json.get(\"default_value\", [[]])\n \n # Handle None entries\n min_entities = None if (min_max_entities is None) else min_max_entities[0]\n max_entities = None if (min_max_entities is None) else min_max_entities[1]\n min_points = 3 if (min_max_points is None) else min_max_points[0]\n max_points = None if (min_max_points is None) else min_max_points[1]\n \n # Bundle config needed by entity drawing object\n drawer_config = {\"minimum_entities\": min_entities,\n \"maximum_entities\": max_entities,\n \"minimum_points\": min_points,\n \"maximum_points\": max_points,\n \"border_size_px\": real_border_size}\n \n # Set up object to keep tracking of drawing\n self.drawer = Entity_Drawer(frame_wh, **drawer_config)\n self.drawer.initialize_entities(default_entities)\n \n # Create the display, if needed\n if create_on_startup:\n self.create_window()\n \n # ................................................................................................................. \n \n def print_info(self):\n \n # Print out info for each window\n header_str = \"Drawing Controls\"\n max_len = 60\n full_spacer_len = max(0, max_len - len(header_str))\n half_spacer_len = int(full_spacer_len / 2)\n end_spacer_len = max(0, max_len - len(header_str) - 2*half_spacer_len)\n \n # Build components for printing control title blocks, then print control info!\n title_spacer = (\" \" * half_spacer_len)\n end_spacer = (\" \" * end_spacer_len)\n full_heading_str = \"\".join([title_spacer, header_str, title_spacer, end_spacer])\n \n # Create key highlight function\n keycolor = Color().bold.italic\n key_text = lambda key_code, info: \" [{}] {}\".format(keycolor(key_code), info)\n \n # Big printout to explain how to do drawing stuff\n print(\"\",\n \"\",\n \"\",\n Color(full_heading_str.upper()).bold.invert,\n \"\",\n Color(\"Hover Mode:\").bold.underline,\n \"\",\n key_text(\"left-click\", \"to move points\"),\n key_text(\"shift + left-click\", \"to enter drawing mode\"),\n key_text(\"ctrl + left-click\", \"to insert points into an existing shape\"),\n key_text(\"right-click\", \"to delete a single point\"),\n key_text(\"ctrl + right-click\", \"to delete an entire shape\"),\n key_text(\"ctrl + z\", \"to undo recent actions\"),\n key_text(\"arrow keys\", \"to nudge points (hold shift for a larger effect)\"),\n key_text(\"b key\", \"to snap points to nearby borders\"),\n \"\",\n Color(\"Drawing Mode:\").bold.underline,\n \"\",\n key_text(\"shift + left-click\", \"to add more points to a shape-in-progress\"),\n key_text(\"double left-click\", \"to complete a shape\"),\n key_text(\"right-click\", \"to cancel a shape\"),\n key_text(\"ctrl + z\", \"to undo last point\"),\n \"\",\n sep=\"\\n\")\n \n # ................................................................................................................. \n \n def initialize_drawing(self, initial_settings_dict):\n \n # Load existing initial data, if present\n variable_in_initial_settings = (self.drawing_variable_name in initial_settings_dict)\n if variable_in_initial_settings:\n initial_entities = initial_settings_dict[self.drawing_variable_name]\n self.drawer.initialize_entities(initial_entities)\n \n return\n \n # ................................................................................................................. \n \n def update_control(self):\n \n # Get changes in zone data\n variables_changed_dict = {}\n if self.drawer.on_change():\n variables_changed_dict.update({self.drawing_variable_name: self.drawer.get_entities_list()})\n \n return variables_changed_dict\n \n # ................................................................................................................. \n \n def keypress(self, key_code, modifier_code):\n self.drawer.keypress_callback(key_code, modifier_code)\n \n # ................................................................................................................. \n \n def imshow(self, display_frame):\n \n # Check if the window exists (by looking for window properties)\n window_exists = self.exists()\n \n # Don't do anything if a valid frame isn't supplied\n if display_frame is None:\n return self.exists()\n \n # Make sure we're using the right frame size, since the drawing depends on it!\n if not self.window_wh_is_set:\n display_height, display_width = display_frame.shape[0:2]\n self.set_window_wh(display_width, display_height)\n self.drawer.update_frame_wh(display_width, display_height)\n \n # Only update showing if the window exists\n if window_exists:\n drawn_frame = self.drawer.annotate(display_frame)\n cv2.imshow(self.window_name, drawn_frame)\n \n return window_exists\n \n # ................................................................................................................. \n \n def create_window(self):\n \n # Create window\n cv2.namedWindow(self.window_name)\n self.imshow_blank()\n self.move_corner_pixels(x_pixels = 50, y_pixels = 50, create_if_missing = False)\n \n # Add drawing callback\n cv2.setMouseCallback(self.window_name, self.drawer)\n \n return self\n \n # ................................................................................................................. \n # ................................................................................................................. \n\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Define callback handlers\n\nclass Mouse_Follower:\n \n # .................................................................................................................\n \n def __init__(self):\n \n # Allocate storage for mouse position and whether following is enabled or not\n self.mouse_xy = np.array((0, 0), dtype=np.int32)\n self.follow_state = True\n \n # .................................................................................................................\n \n def __call__(self, *args, **kwargs):\n ''' Convenience wrapper. Allows object to be used as a callback function directly '''\n return self.callback(*args, **kwargs)\n \n # .................................................................................................................\n \n def callback(self, event, mx, my, flags, param):\n \n # Record mouse xy position\n if self.follow_state:\n self.mouse_xy = np.int32((mx, my))\n \n # Toggle following state on left click\n if event == cv2.EVENT_LBUTTONDOWN:\n self.follow_state = (not self.follow_state)\n \n # .................................................................................................................\n \n def draw_mouse_xy(self, display_frame, point_radius = 5, point_color = (255, 0, 255)):\n \n ''' Function to help with debugging. Displays a point at the mouse location, along with x/y co-ordinates '''\n \n xy_tuple = tuple(self.xy)\n text_xy = (xy_tuple[0] + point_radius + 2, xy_tuple[1] + 5)\n \n drawn_frame = display_frame.copy()\n cv2.circle(drawn_frame, xy_tuple, point_radius, point_color, -1, cv2.LINE_AA)\n cv2.putText(drawn_frame,\n \"({:.0f}, {:.0f})\".format(*xy_tuple),\n text_xy,\n cv2.FONT_HERSHEY_SIMPLEX,\n 0.5,\n (255, 255, 255),\n 1,\n cv2.LINE_AA)\n \n return drawn_frame\n \n # .................................................................................................................\n \n @property\n def xy(self):\n return self.mouse_xy\n \n # .................................................................................................................\n # .................................................................................................................\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Define functions\n\n# .....................................................................................................................\n \ndef simple_affine(min_value, max_value, step_size = 1, return_type = None):\n \n # Using y = mx + b (y -> mapped value, x -> raw/trackbar value)\n # Where y = min when x = 0\n # y = max_value when x = (max_value - min_value) / step_size\n # So:\n # min = 0 + b\n # max = m * (1 / step) * (max - min) + b\n \n # Therefore\n # b = min\n # m = step * (max - b) * (1 / max - min)\n #\n # b = min, m = step\n # y = step * x + min\n # x = (y - min) / step\n \n def raw_to_map_func(raw_value):\n map_value = step_size * raw_value + min_value\n return return_type(map_value) if return_type else map_value\n \n def map_to_raw_func(map_value): \n raw_value = (map_value - min_value) / step_size\n return (int(round(raw_value)))\n \n return raw_to_map_func, map_to_raw_func\n\n# .....................................................................................................................\n \ndef minceil_affine(min_value, max_value, step_size = 1, return_type = None):\n \n # Same as an affine mapping, but won't allow the value to go below the minimum value\n # Intended for cases where the trackbar is pinned to 0 for display purposes.\n \n # Pre-calculate/generate some useful variables\n simple_raw_to_map_func, simple_map_to_raw_func = simple_affine(min_value, max_value, step_size, return_type)\n min_raw_value = min_value / step_size\n min_offset = (min_value / step_size)\n \n def raw_to_map_func(raw_value): \n ceil_raw_value = max(min_raw_value, raw_value) - min_offset\n map_value = simple_raw_to_map_func(ceil_raw_value)\n return map_value\n \n def map_to_raw_func(map_value):\n raw_value = simple_map_to_raw_func(map_value) + min_offset\n return int(round(raw_value))\n \n return raw_to_map_func, map_to_raw_func\n\n# .....................................................................................................................\n \ndef value_list_lookup(value_list):\n \n def raw_to_map_func(raw_value):\n # For menus, the raw value is the trackbar location,\n # which selects a value from the value list as a simple (list) index\n return value_list[raw_value]\n \n def map_to_raw_func(map_value):\n # For menus, the mapped value is an entry in the value list\n # The raw value is the trackbar location, which is also just the index of the value in the list\n return value_list.index(map_value)\n \n return raw_to_map_func, map_to_raw_func\n\n# .....................................................................................................................\n \ndef bool_to_int():\n \n def raw_to_map_func(raw_value):\n return bool(raw_value)\n \n def map_to_raw_func(map_value):\n return int(map_value)\n \n return raw_to_map_func, map_to_raw_func\n\n# .....................................................................................................................\n \ndef button_map(control_label, set_trackbar_func):\n \n raise NotImplementedError\n def raw_to_map_func(raw_value):\n button_state = bool(raw_value)\n set_trackbar_func(control_label, 0) # Does this work?\n return button_state\n \n def map_to_raw_func(map_value):\n return int(map_value)\n \n return raw_to_map_func, map_to_raw_func\n\n# .....................................................................................................................\n \ndef return_type_strings_to_functions(return_type_str):\n \n ret_str_lut = {None: None,\n \"string\": str,\n \"integer\": int,\n \"float\": float,\n \"bool\": bool,\n \"list\": list,\n \"tuple\": tuple}\n \n return ret_str_lut[return_type_str]\n\n# .....................................................................................................................\n# .....................................................................................................................\n\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Demo\n\nif __name__ == \"__main__\":\n \n # Set display parameters\n frame_width, frame_height = 600, 300\n blank_frame = np.full((frame_height, frame_width, 3), (33, 166, 83), dtype=np.uint8)\n frame_wh = (frame_width, frame_height)\n \n # Set up example mouse follower\n follower = Mouse_Follower()\n \n # Window creation & callback assignment\n window_name = \"FOLLOWER EXAMPLE\"\n cv2.namedWindow(window_name) \n cv2.setMouseCallback(window_name, follower)\n \n while True:\n \n # Get a clean copy of the video\n display_frame = blank_frame.copy()\n \n # Draw mouse location as an example\n drawn_frame = follower.draw_mouse_xy(display_frame)\n cv2.imshow(window_name, drawn_frame)\n \n # Get keypress\n keypress = cv2.waitKey(40)\n esc_key_press = (keypress == 27)\n q_key_pressed = (keypress == 113)\n if esc_key_press or q_key_pressed:\n break\n \n # Clean up windows\n cv2.destroyAllWindows()\n\n# ---------------------------------------------------------------------------------------------------------------------\n#%% Scrap\n\n\n","sub_path":"local/lib/ui_utils/local_ui/windows_base.py","file_name":"windows_base.py","file_ext":"py","file_size_in_byte":51074,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"459894856","text":"import pymysql\r\nimport random\r\nconnection = pymysql.connect(\"46.4.115.158\",\"beo\", \"beo@123\",\"testdb\")\r\ntable_name=str(input(\"Enter the Table name:\"))\r\naction = connection.cursor()\r\nrows=int(input(\"How many rows want to add:\"))\r\nsql=\"\"\"desc \"\"\"+table_name\r\naction.execute(sql)\r\nresult=action.fetchall()\r\naa=int(action.execute(sql))\r\n#print(aa)\r\nsql=\"\"\"insert into \"\"\"+table_name+\" values\"\r\nfor i in range(1,rows+1):\r\n count = 1\r\n sql = sql + \"(\"\r\n print(\"ROW: \", i)\r\n for j in result:\r\n print(\"Enter\",j[0],end='')\r\n value=input()\r\n sql=sql+\"'\"+value+\"'\"\r\n if(count 0.75 and SPE > 0.75:\n print('SEN:%f' % (SEN))\n print('SPE:%f' % (SPE))\n print(\"Acc:%f\" % acc)\n print('cutoff:%f' % (cutoff))\n print('PPV:%f' % (TP / (TP + FP)))\n print('NPV:%f' % (TN / (TN + FN)))\n print(\"tp:%d tn:%d fp:%d fn:%d\" % (TP, TN, FP, FN))\n shrink = True\n return cutoff, shrink, SEN, SPE, acc, TP, TN, FP, FN\n\n\ndef evaluate_on_testset(fpr, tpr, threshold, cutoff, pos_num, neg_num):\n '''find tpr and fpr'''\n temp_pos = 0\n for cont, thre in enumerate(threshold):\n if thre < cutoff:\n temp_pos = cont\n break\n if temp_pos == 0:\n temp_pos = 1\n print('temp_pos is 0')\n # print(temp_pos)\n proportion = (threshold[temp_pos - 1] - cutoff) / (threshold[temp_pos - 1] - threshold[temp_pos])\n # print(proportion)\n sen = tpr[temp_pos - 1] + (tpr[temp_pos] - tpr[temp_pos - 1]) * proportion\n spe = 1 - (fpr[temp_pos - 1] + (fpr[temp_pos] - fpr[temp_pos - 1]) * proportion)\n TP = sen * pos_num\n TN = spe * neg_num\n FP = neg_num - TN\n FN = pos_num - TP\n acc = (TP + TN) / (FP + FN + TP + TN)\n shrink = False\n if sen > 0.75 and spe > 0.7:\n print('SEN:%f' % (sen))\n print('SPE:%f' % (spe))\n print(\"Acc:%f\" % acc)\n print('PPV:%f' % (TP / (TP + FP)))\n print('NPV:%f' % (TN / (TN + FN)))\n print(\"tp:%d tn:%d fp:%d fn:%d\" % (TP, TN, FP, FN))\n shrink = True\n return shrink, sen, spe, acc, TP, TN, FP, FN\n\nif __name__ == '__main__':\n path = \"E:\\\\img_breast_dataset\"\n print(\"get data set\")\n train_neg_set, train_pos_set, test_pos_set, test_neg_set = get_data_set(path=path, times=0)\n total_train_set = train_neg_set + train_pos_set\n total_test_set = test_neg_set + test_pos_set\n print(\"get data set completed\")\n print(\"get vgg features\")\n vgg_feature_selector = vgg_models.vgg_16_bn(num_class=2, pretrained=True, pool_out=True, first_fc_out=False)\n num_features = get_vgg_feature(vgg_feature_selector, total_train_set, total_test_set, batchsize=16)\n print(\"get vgg features completed\")\n print(num_features)\n # Classifer model\n model = pool_feature_model.pool_feature_model_2(num_features=num_features, hidden=1800, num_class=2)\n # Learning rate and momentum\n lr = 0.1\n classifier_optimizer = optim.SGD(model.parameters(), lr=lr, momentum=0.9)\n\n length = len(train_neg_set)\n i = 0\n t = 0\n con = False\n\n TR_sen, TR_spe, TR_acc, TR_tp, TR_tn, TR_fp, TR_fn = 0, 0, 0, 0, 0, 0, 0\n TE_sen, TE_spe, TE_acc, TE_tp, TE_tn, TE_fp, TE_fn = 0, 0, 0, 0, 0, 0, 0\n roc_auc_TE, roc_auc_TR = 0, 0\n G_y_t1, G_y_p1, G_y_t2, G_y_p2 = 0, 0, 0, 0\n\n for epoch in range(1, 2000):\n if (epoch + 1) % 20 == 0:\n for param_group in classifier_optimizer.param_groups:\n param_group['lr'] = lr / math.sqrt(epoch)\n print(\"training\")\n pos_set, i = get_batch(i=i, train_pos_set=train_pos_set, length=length)\n\n train_set = train_neg_set + pos_set\n random.shuffle(train_set)\n\n train_batch(epoch, train_set, optimizer=classifier_optimizer, model=model, batchsize=16)\n\n print(\"Testing trainset\")\n y_t1, y_p1 = test(test_set=total_train_set, model=model)\n fpr, tpr, threshold = roc_curve(y_t1, y_p1, pos_label=1)\n cutoff, shrink_tr, tr_sen, tr_spe, tr_acc, tr_tp, tr_tn, tr_fp, tr_fn = evaluate_on_trainset(fpr=fpr, tpr=tpr, threshold=threshold,\n pos_num=len(train_pos_set),\n neg_num=len(train_neg_set))\n roc_auc_tr = auc(fpr, tpr)\n print(\"auc: %f\" % roc_auc_tr)\n\n print(\"Testing testset\")\n y_t2, y_p2 = test(test_set=total_test_set, model=model)\n fpr, tpr, threshold = roc_curve(y_t2, y_p2, pos_label=1)\n shrink_te, te_sen, te_spe, te_acc, te_tp, te_tn, te_fp, te_fn = evaluate_on_testset(fpr=fpr, tpr=tpr, threshold=threshold, cutoff=cutoff,\n pos_num=len(test_pos_set),\n neg_num=len(test_neg_set))\n roc_auc_te = auc(fpr, tpr)\n print(\"auc: %f\" % roc_auc_te)\n\n if t < 10:\n if shrink_tr and shrink_te:\n if roc_auc_te > roc_auc_TE and roc_auc_tr > roc_auc_TR:\n t = 0\n con = True\n # update\n roc_auc_TE = roc_auc_te\n roc_auc_TR = roc_auc_tr\n TR_sen, TR_spe, TR_acc, TR_tp, TR_tn, TR_fp, TR_fn = tr_sen, tr_spe, tr_acc, tr_tp, tr_tn, tr_fp, tr_fn\n TE_sen, TE_spe, TE_acc, TE_tp, TE_tn, TE_fp, TE_fn = te_sen, te_spe, te_acc, te_tp, te_tn, te_fp, te_fn\n G_y_t1, G_y_p1, G_y_t2, G_y_p2 = y_t1, y_p1, y_t2, y_p2\n lr = lr / 2\n elif con == True:\n t += 1\n \n else:\n print(\"Best result:\")\n print(\"Acc:%f\" % roc_auc_TR)\n print('SEN:%f' % (TR_sen))\n print('SPE:%f' % (TR_spe))\n print(\"Acc:%f\" % TR_acc)\n print('PPV:%f' % (TR_tp / (TR_tp + TR_fp)))\n print('NPV:%f' % (TR_tn / (TR_tn + TR_fn)))\n print(\"tp:%d tn:%d fp:%d fn:%d\" % (TR_tp, TR_tn, TR_fp, TR_fn))\n print(\"\\n\")\n print(\"Acc:%f\" % roc_auc_TE)\n print('SEN:%f' % (TE_sen))\n print('SPE:%f' % (TE_spe))\n print(\"Acc:%f\" % TE_acc)\n print('PPV:%f' % (TE_tp / (TE_tp + TE_fp)))\n print('NPV:%f' % (TE_tn / (TE_tn + TE_fn)))\n print(\"tp:%d tn:%d fp:%d fn:%d\" % (TE_tp, TE_tn, TE_fp, TE_fn))\n p_plt(G_y_t1, G_y_p1, G_y_t2, G_y_p2)\n print(time.asctime(time.localtime(time.time())))\n auc_cal(G_y_t1, G_y_p1)\n auc_cal(G_y_t2, G_y_p2)\n plt.show()\n print(time.asctime(time.localtime(time.time())))","sub_path":"transfer_find_best.py","file_name":"transfer_find_best.py","file_ext":"py","file_size_in_byte":7044,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"200561110","text":"import os\nimport bpy\nfrom bpy.props import BoolProperty, IntProperty\n\nfrom .functions.global_settings import ProjectSettings, Extensions\nfrom .functions.file_management import *\nfrom .functions.sequences import find_empty_channel\n\n\n# TODO: Fix img imported from subfolder -\nclass ImportLocalFootage(bpy.types.Operator):\n bl_idname = \"power_sequencer.import_local_footage\"\n bl_label = \"PS.Import local footage\"\n bl_description = \"Import video and audio from the project \\\n folder to VSE strips\"\n bl_options = {'REGISTER', 'UNDO'}\n\n import_all = BoolProperty(\n name=\"Always Reimport\",\n description=\"If true, always import all local files to new strips. \\\n If False, only import new files (check if footage has \\\n already been imported to the VSE).\",\n default=False)\n keep_audio = BoolProperty(\n name=\"Keep audio from video files\",\n description=\"If False, the audio that comes with video files \\\n will not be imported\",\n default=True)\n\n img_length = IntProperty(\n name=\"Image strip length\",\n description=\"Controls the duration of the imported image strip\",\n default=96,\n min=1)\n img_padding = IntProperty(\n name=\"Image strip padding\",\n description=\"Padding added between imported image strips in frames\",\n default=24,\n min=1)\n\n @classmethod\n def poll(cls, context):\n return True\n\n def execute(self, context):\n if not bpy.data.is_saved:\n self.report(\n {\"ERROR_INVALID_INPUT\"},\n \"You need to save your project first. Import cancelled.\")\n return {\"CANCELLED\"}\n\n sequencer = bpy.ops.sequencer\n context = bpy.context\n frame_current = bpy.context.scene.frame_current\n empty_channel = find_empty_channel()\n\n bpy.ops.screen.animation_cancel(restore_frame=True)\n\n # Store reference to the Sequencer area to import files to\n for window in bpy.context.window_manager.windows:\n screen = window.screen\n for area in screen.areas:\n if area.type == 'SEQUENCE_EDITOR':\n SEQUENCER_AREA = {'window': window,\n 'screen': screen,\n 'area': area,\n 'scene': bpy.context.scene}\n\n\n # Find folders for audio, img and video strips\n directory = get_working_directory()\n folders, files, files_dict = {}, {}, {}\n file_types = \"AUDIO\", \"IMG\", \"VIDEO\"\n\n for folder in os.listdir(path=directory):\n folder_upper = folder.upper()\n if folder_upper in file_types:\n folders[folder_upper] = os.path.join(directory, folder)\n\n for file_type in file_types:\n if file_type not in folders.keys():\n continue\n files[file_type] = find_files(folders[file_type],\n Extensions.DICT[file_type],\n recursive=file_type == \"IMG\")\n\n # TODO: walk the project dir tree and collect all files that have a supported Extension\n #\n # files, files_dict = {}, {}\n # file_types = \"AUDIO\", \"IMG\", \"VIDEO\"\n # for file_type in file_types:\n # files[file_type] = find_files_temp(get_working_directory(),\n # Extensions.DICT[file_type])\n # filepaths = []\n # for dirpath, dirname, filenames in os.walk(directory, topdown=True):\n # for filename in filenames:\n # files.append(os.path.join(dirparth, filename)\n\n # Find or create new text files to keep track of imported material\n TEXT_FILE_PREFIX = 'IMPORT_'\n texts = bpy.data.texts\n import_files = {}\n for file_type in file_types:\n if texts.get(TEXT_FILE_PREFIX + file_type):\n import_files[file_type] = texts[TEXT_FILE_PREFIX + file_type]\n\n if not import_files:\n for name in file_types:\n import_files[name] = create_text_file(TEXT_FILE_PREFIX + name)\n assert len(import_files) == 3\n\n # Write new imported paths to the text files and import new strips\n channel_offset = 0\n new_sequences, new_video_sequences = [], []\n for name in file_types:\n if name not in folders.keys():\n continue\n\n text_file_content = [\n line.body\n for line in bpy.data.texts[TEXT_FILE_PREFIX + name].lines\n ]\n new_paths = [path\n for path in files[name]\n if path not in text_file_content]\n for line in new_paths:\n bpy.data.texts[TEXT_FILE_PREFIX + name].write(line + \"\\n\")\n\n if not new_paths:\n continue\n\n # Import new strips if new files were found\n import_channel = empty_channel + channel_offset\n folder = folders[name]\n files_dict = files_to_dict(new_paths, folder)\n\n if name == \"VIDEO\":\n import_channel += 1 if self.keep_audio else 0\n sequencer.movie_strip_add(SEQUENCER_AREA,\n filepath=folder,\n files=files_dict,\n frame_start=frame_current,\n channel=import_channel,\n sound=self.keep_audio)\n new_sequences.extend(bpy.context.selected_sequences)\n # Blender places audio tracks on top, we want them under video\n new_video_sequences.extend(bpy.context.selected_sequences)\n elif name == \"AUDIO\":\n sequencer.sound_strip_add(\n SEQUENCER_AREA,\n filepath=folder,\n files=files_dict,\n frame_start=frame_current,\n channel=import_channel)\n new_sequences.extend(bpy.context.selected_sequences)\n elif name == \"IMG\":\n img_frame = frame_current\n for img in files_dict:\n path = os.path.join(folder, img['subfolder'])\n # FIXME: temp hack so images import properly\n file = [{'name': img['name'].replace(\"img\\\\\", \"\")}]\n sequencer.image_strip_add(\n SEQUENCER_AREA,\n directory=path,\n files=file,\n frame_start=img_frame,\n frame_end=img_frame + self.img_length,\n channel=import_channel)\n new_sequences.extend(bpy.context.selected_sequences)\n img_frame += self.img_length + self.img_padding\n channel_offset += 1\n\n # Swap channels for audio and video tracks\n if not new_video_sequences:\n return {\"FINISHED\"}\n\n # Reorder the sequences so all MOVIE strips are on top\n sequencer.select_all(action='DESELECT')\n for s in new_video_sequences:\n s.select = True\n sequencer.meta_make()\n sequencer.meta_toggle()\n videos_in_meta = [s for s in bpy.context.selected_sequences if s.type == 'MOVIE']\n for s in videos_in_meta:\n s.channel += 2\n for s in new_video_sequences:\n s.channel -= 1\n sequencer.meta_toggle()\n sequencer.meta_separate()\n\n # Set the strips to use proxies based if set in the addon preferences\n prefs = context.user_preferences.addons[__package__].preferences\n if prefs.auto_render_proxies:\n bpy.ops.power_sequencer.set_video_proxies()\n\n # Show audio waveforms\n for s in [strip for strip in new_sequences if strip.type == 'SOUND']:\n s.show_waveform = True\n\n for s in new_sequences:\n s.select = True\n return {\"FINISHED\"}\n\n\n# TODO: Ignore the blender proxy folders\n# TODO: Detect img sequences\ndef find_files(directory,\n file_extensions,\n recursive=False,\n ignore_folders=('_proxy', 'BL_proxy')):\n \"\"\"\n Walks through a folder and returns a list of filepaths\n that match the extensions.\n Args:\n - file_extensions is a tuple of extensions with the form \"*.ext\".\n Use the Extensions helper class in .functions.global_settings.\n It gives default extensions to check the files against.\n Returns a list of file paths, or [] if nothing was found\n \"\"\"\n if not directory and file_extensions:\n return None\n\n files = []\n\n from glob import glob\n from os.path import basename\n\n # TODO: Folder containing img files = img sequence?\n for ext in file_extensions:\n source_pattern = directory + \"/\"\n pattern = source_pattern + ext\n files.extend(glob(pattern))\n if not recursive:\n continue\n pattern = source_pattern + \"**/\" + ext\n files.extend(glob(pattern))\n\n if basename(directory) == \"IMG\":\n psd_names = [f for f in glob(directory + \"/*.psd\")]\n for i, name in enumerate(psd_names):\n psd_names[i] = name[len(directory):-4]\n\n psd_folders = (f for f in os.listdir(directory) if f in psd_names)\n for f in psd_folders:\n for ext in file_extensions:\n files.extend(glob(directory + \"/\" + f + \"/\" + ext))\n return files\n\n\n# TODO: issue with img vs other strip types: img have separate filepath and filename slots\n# but video/audio only have direct filepath e.g. audio/file.wav\ndef files_to_dict(files, folder_path):\n \"\"\"Converts a list of files to Blender's dictionary format for import\n Returns a list of dictionaries with the\n {'name': filename, 'subfolder': subfolder} format\n If the provided files are placed at the root of the import folders,\n subfolder will be an empty string\n Args:\n - files: a list or a tuple of files\n - folder_path: a string of the path to the files' containing folder\"\"\"\n if not files and folder_path:\n return None\n\n dictionary = []\n for f in files:\n filepath_tail = f[len(folder_path) + 1:]\n head, tail = os.path.split(filepath_tail)\n\n project_path, subfolder_name = os.path.split(folder_path)\n dict_form = {'name': os.path.join(subfolder_name, tail), 'subfolder': head}\n dictionary.append(dict_form)\n return dictionary\n","sub_path":"load_files.py","file_name":"load_files.py","file_ext":"py","file_size_in_byte":10680,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"365434998","text":"# encoding: utf-8\nimport pandas as pd\nfrom collections import Counter\nimport time\nimport sys\nimport re\nimport pymorphy2\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom datetime import datetime\nfrom sklearn.base import TransformerMixin\nfrom sklearn.cluster import AgglomerativeClustering, DBSCAN\nfrom sklearn.pipeline import Pipeline\nfrom functools import wraps\n\n\nclass DenseTransformer(TransformerMixin):\n\n def fit(self, X, y=None, **fit_params):\n return self\n\n def transform(self, X, y=None, **fit_params):\n return X.todense()\n\n\ndef timethis(func):\n \"\"\"\n Decorator that reports the execution time.\n \"\"\"\n\n @wraps(func)\n def wrapper(*args, **kwargs):\n start = time.time()\n print(\"Executing {}...\".format(func.__name__))\n result = func(*args, **kwargs)\n end = time.time()\n print('{} took {}'.format(func.__name__, end - start))\n return result\n\n return wrapper\n\n\nmorph = pymorphy2.MorphAnalyzer()\ncount = Counter()\nnrows_to_load = None\nrandom_sample = None\nnum_clasters_for_kMeans = 20\nengine_for_pd = 'c'\n# engine_for_pd = 'python'\ndata_file_name = r'D:\\data\\oscar'\n\n\ndef text_cleaner(text):\n text = str(text).lower()\n text = re.sub(r'[\\W]+', ' ', text) # удаление лишних символов\n\n text = ' '.join(list(map(lambda x: morph.parse(x)[0].normal_form, text.split())))\n stw = ['в', 'по', 'на', 'из', 'и', 'или', 'не', 'но', 'за', 'над', 'под', 'то', 'для', \"как\",\n 'a', 'at', 'on', 'of', 'and', 'or', 'in', 'for', 'at']\n remove = r'\\b(' + '|'.join(stw) + ')\\b'\n text = re.sub(remove, ' ', text)\n\n text = re.sub(r'\\b\\w\\b', ' ', text) # удаление отдельно стоящих букв\n\n # text = re.sub(r'\\b\\d+\\b', ' digit ', text) # замена цифр\n return text\n\n\n@timethis\ndef load_from_csv(file):\n df = pd.read_csv(file, '\\t', parse_dates=['datetime'], index_col='datetime',\n converters={'normal_query': str},\n nrows=nrows_to_load,\n engine=engine_for_pd,\n )\n # df.info()\n print(\"считано: {}\".format(df.shape))\n return df\n\n\n@timethis\ndef load_data_and_lemmatize(file='oscar1'):\n df = pd.read_csv(file, '\\t', parse_dates=['datetime'], index_col='datetime',\n converters={'normal_query': text_cleaner},\n nrows=nrows_to_load,\n engine=engine_for_pd,\n )\n # df.info()\n return df\n\n\n@timethis\ndef save_data(data, file='oscar2'):\n data.to_csv(file, sep='\\t')\n\n\n@timethis\ndef learn_and_predict_DBSCAN(data, *args):\n text_clstz = Pipeline([\n ('tfidf', TfidfVectorizer(binary=True)),\n # ('svd', TruncatedSVD(n_components=100, random_state=123)),\n # ('to_dense', DenseTransformer()),\n ('DBSCAN', DBSCAN(eps=0.2, min_samples=5, metric='euclidean'))\n ])\n data['tag'] = text_clstz.fit_predict(data['normal_query'])\n print(\"{} clusters\".format(len(set(data['tag'].tolist()))))\n return data\n\n\n@timethis\ndef learn_and_predict_AgglomerativeClustering(data, num_clasters_for_kMeans):\n text_clstz = Pipeline([\n ('tfidf', TfidfVectorizer(binary=True)),\n # ('svd', TruncatedSVD(n_components=100, random_state=123)),\n ('to_dense', DenseTransformer()),\n ('afp', AgglomerativeClustering(n_clusters=num_clasters_for_kMeans,\n affinity='cosine',\n linkage='complete'))\n ])\n data['tag'] = text_clstz.fit_predict(data['normal_query'])\n print(\"{} clusters\".format(len(set(data['tag'].tolist()))))\n return data\n\n\ndef commons(text):\n count.clear()\n count.update(filter(lambda x: len(x) > 2, text.split()))\n text = str(count.most_common(10))\n return text\n\n\ndef count_words(text):\n count.update(filter(lambda x: len(x) > 2, text.split())) # take words more than 2 symbols\n return text\n\n\n@timethis\ndef save_file(dt, excel_file_to):\n writer = pd.ExcelWriter(excel_file_to, engine='xlsxwriter')\n dt.to_excel(writer, 'Sheet1')\n writer.save()\n\n\n@timethis\ndef filter_data_oscar(data):\n return data[data['normal_query'].str.contains('oscar|оскар')].sort_index()\n\n\n@timethis\ndef main():\n data = load_from_csv(data_file_name) # load all data\n only_oscars = filter_data_oscar(data) # filter only that contains oscars\n save_data(only_oscars, data_file_name + '_oscar') # save obly oscar to file\n data = '' # free memory from trash\n only_oscars = load_from_csv(data_file_name + '_oscar') # load again only oscars from file\n only_oscars['normal_query'] = only_oscars['normal_query'].map(\n lambda x: text_cleaner(x)) # lemmatize only oscars\n\n save_data(only_oscars, data_file_name + '_oscar_normal') # save lemmatized oscars\n only_oscars = only_oscars.sort_index()\n # on MOW time oscar was from 03/00 25/02/19 till 07/00 25/02/2019\n # lets assume that time in this table is MOW\n delimited = [only_oscars[:datetime(2019, 2, 25, 2, 59, 59)].copy(deep=True), # before\n only_oscars[datetime(2019, 2, 25, 3, 0, 0):datetime(2019, 2, 25, 6, 59, 59)].copy(\n deep=True), # during\n only_oscars[datetime(2019, 2, 25, 7, 0, 0):].copy(deep=True)] # after\n\n names_of_periods = {0: 'BEFORE\\n\\n', 1: 'DURING\\n\\n', 2: 'AFTER\\n\\n'}\n # print info about these datasets\n for (num, d) in enumerate(delimited):\n print(names_of_periods[num])\n d.info()\n print(d.head())\n\n for (num, d) in enumerate(delimited):\n print(names_of_periods[num])\n count.clear()\n fn = names_of_periods[num].strip().lower()\n d['normal_query'].map(lambda x: count_words(x))\n save_file(pd.DataFrame(count.most_common()),\n r'd:\\data\\oscar_freq_{}.xlsx'.format(fn)) # save frequency of words to excel\n\n # lets try to make Hierarchical clustering\n data = learn_and_predict_AgglomerativeClustering(d, num_clasters_for_kMeans)\n save_data(data.sort_values('tag'), r'D:\\data\\oscarAgglomerativeClustering_{}'.format(fn))\n # most common words in clusters\n df = data.groupby('tag')['normal_query'].apply(lambda words: ' '.join(words))\n df = pd.DataFrame(df)\n df['normal_query'] = df['normal_query'].apply(commons)\n\n save_data(df, data_file_name + 'Agglomerative_Group_{}'.format(fn))\n\n #also lets try DBSCAN clustering\n data = learn_and_predict_DBSCAN(d, num_clasters_for_kMeans)\n save_data(data.sort_values('tag'), r'D:\\data\\oscarDBSCANClustering_{}'.format(fn))\n # most common words in clusters\n df = data.groupby('tag')['normal_query'].apply(lambda words: ' '.join(words))\n df = pd.DataFrame(df)\n df['normal_query'] = df['normal_query'].apply(commons)\n\n save_data(df, data_file_name + 'DBSCAN_Group_{}'.format(fn))\n\n\nif __name__ == '__main__':\n sys.exit(main())\n","sub_path":"tsts/clasterizer-text/tst_oscar_new.py","file_name":"tst_oscar_new.py","file_ext":"py","file_size_in_byte":7044,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"341861071","text":"# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages\n\nwith open('README.rst') as f:\n readme = f.read()\n\nsetup(\n name='pytemplate',\n version='0.0.1',\n description='Generic template for python projects',\n long_description=readme,\n author='Quadyster Cloud Devs',\n author_email='',\n url='https://github.com/raghava-aparna/python-project-template',\n packages=find_packages(exclude=('tests', 'docs'))\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"34113920","text":"from __future__ import print_function\nfrom app import mongo, db\nfrom datetime import datetime\nfrom bson.objectid import ObjectId\nfrom random import randint\nfrom rtw import *\nimport sys\n\n#subCategoryBelongsTo(identifications): dict | None\n#pickRandomCategory(): dict | None\n#pickRandomSubCategory(identifications): dict | None\n\n#createUser(identifications): ObjectId\n#updateUser(identifications, updates): bool\n#getUser(identifications): dict | None\n#isUserOnline(identifications): bool\n#changeUserStatus(identifications, online): bool\n#updateScore(identifications, score): bool\n\n#createGame(identifications): ObjectId\n#updateGame(identifications, updates): bool\n#getGame(identifications): dict | None\n#findWaitingGame(identifications, user): dict | None\n#joinGame(identifications, user): bool\n#isGameReady(identifications): bool\n#startGame(identifications): bool\n#checkGameStatus(identifications): int\n#userFromGame(identifications, number): ObjectId | None\n#finishGame(identifications): bool\n\ndef subCategoryBelongsTo(identifications):\n\t#cursor = mongo.db.subcategories.find(identifications)\n\tcursor = db.subcategories.find(identifications)\n\tif cursor.count() > 0:\n\t\tidentifications = dict()\n\t\tidentifications[\"_id\"] = cursor[0][\"category\"]\n\t\t#cursor = mongo.db.categories.find(identifications)\n\t\tcursor = db.categories.find(identifications)\n\t\tif cursor.count() > 0:\n\t\t\treturn cursor[0]\n\t\treturn None\n\treturn None\n\ndef pickRandomCategory():\n\t#cursor = mongo.db.categories.find({})\n\tcursor = db.categories.find({})\n\tif cursor.count() == 0:\n\t\treturn None\n\trand = randint(0, cursor.count() - 1)\n\treturn cursor[rand]\t\n\ndef pickRandomSubCategory(identifications):\n\t#cursor = mongo.db.subcategories.find(identifications)\n\tcursor = db.subcategories.find(identifications)\n\tif cursor.count() == 0:\n\t\treturn None\n\trand = randint(0, cursor.count() - 1)\n\treturn cursor[rand]\n\ndef createUser(identifications):\n\tidentifications[\"online\"] = False\n\tidentifications[\"score\"] = 0\n\tregDate = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\t\n\tidentifications[\"regDate\"] = regDate\n\t#_id = mongo.db.users.insert_one(identifications)\n\t_id = db.users.insert_one(identifications)\n\treturn _id.inserted_id\n\ndef updateUser(identifications, updates):\n\t#cursor = mongo.db.users.find(identifications)\n\tcursor = db.users.find(identifications)\n\tif cursor.count() > 0:\n\t\t#mongo.db.users.update_one(identifications, {\"$set\": updates})\n\t\tdb.users.update_one(identifications, {\"$set\": updates})\n\t\treturn True\n\treturn False\n\ndef getUser(identifications):\n\t#cursor = mongo.db.users.find(identifications)\n\tcursor = db.users.find(identifications)\n\tif cursor.count() > 0:\n\t\treturn cursor[0]\n\treturn None\n\ndef isUserOnline(identifications):\n\tuser = getUser(identifications)\n\treturn user != None and user[\"online\"]\n\ndef changeUserStatus(identifications, online):\n\t#cursor = mongo.db.users.find(identifications)\n\tcursor = db.users.find(identifications)\n\tupdates = dict()\n\tupdates[\"online\"] = online\n\treturn updateUser(identifications, updates)\n\ndef updateScore(identifications, score):\n\t#cursor = mongo.db.users.find(identifications)\n\tcursor = db.users.find(identifications)\n\tif cursor.count() == 0:\n\t\treturn False\n\tupdates = dict()\n\tupdates[\"score\"] = cursor[0][\"score\"] + score\n\treturn updateUser(identifications, updates)\n\n#initial identifications: gameType, theme\ndef createGame(identifications):\n\tcreatedTime = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n\tidentifications[\"createdTime\"] = createdTime\n\tidentifications[\"finished\"] = False\n\tidentifications[\"user1\"] = None\n\tidentifications[\"data1\"] = None\n\tidentifications[\"score1\"] = 0\n\tif not \"data2\" in identifications.keys():\n\t\tidentifications[\"data2\"] = None \t\t\t#for game type 2, data2 will be the collection of hints\n\tidentifications[\"score2\"] = 0 \t\t\t\t\t#for game type 2, score2 will store the number of hints needed for the user to give the correct answer\n\tif identifications[\"gameType\"] != 2 and identifications[\"gameType\"] != 4:\n\t\tidentifications[\"user2\"] = None\n\tidentifications[\"start\"] = None\n\tidentifications[\"finish\"] = None\n\tidentifications[\"status\"] = 0\n\t#_id = mongo.db.games.insert_one(identifications)\n\t_id = db.games.insert_one(identifications)\n\treturn _id.inserted_id\n\ndef updateGame(identifications, updates):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0:\n\t\t#mongo.db.games.update_one(identifications, {\"$set\": updates})\n\t\tdb.games.update_one(identifications, {\"$set\": updates})\n\t\treturn True\n\treturn False\n\ndef getGame(identifications):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0:\n\t\treturn cursor[0]\n\treturn None\n\n#initial identifications: gameType\n#STATUS:\n#0: esperando jogador\n#1: encerrado\n#2: sendo jogado\n#3: parcialmente encerrado\n#4: jogo nao criado\ndef findWaitingGame(identifications, user):\n\tidentifications[\"status\"] = 0\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tfor i in range(cursor.count()):\n\t\tif cursor[i][\"user1\"] != user:\n\t\t\treturn cursor[i]\n\treturn None\n\ndef joinGame(identifications, user):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0:\n\t\tupdates = dict()\n\t\tif cursor[0][\"user1\"] == None:\n\t\t\tupdates[\"user1\"] = user\n\t\t\treturn updateGame(identifications, updates)\n\t\telif \"user2\" in cursor[0].keys() and cursor[0][\"user2\"] == None:\n\t\t\tupdates[\"user2\"] = user\n\t\t\treturn updateGame(identifications, updates)\n\t\treturn False\n\treturn False\n\ndef isGameReady(identifications):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0:\n\t\tif \"user2\" in cursor[0].keys():\n\t\t\treturn cursor[0][\"user1\"] != None and cursor[0][\"user2\"] != None\n\t\telse:\n\t\t\treturn cursor[0][\"user1\"] != None\n\treturn False\n\ndef startGame(identifications):\n\tstart = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n\tupdates = dict()\n\tupdates[\"status\"] = 2\n\tupdates[\"start\"] = start\n\treturn updateGame(identifications, updates) \n\ndef checkGameStatus(identifications):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tfor c in cursor:\n\t\treturn int(c[\"status\"])\n\treturn 4\n\ndef userFromGame(identifications, number):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0 and \"user\" + str(number) in cursor[0].keys():\n\t\treturn cursor[0][\"user\" + str(number)]\n\treturn None\t\n\ndef pickRandomGame(identifications):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() == 0:\n\t\treturn None\n\trand = randint(0, cursor.count() - 1)\n\treturn cursor[rand]\n\ndef finishGame(identifications):\n\t#cursor = mongo.db.games.find(identifications)\n\tcursor = db.games.find(identifications)\n\tif cursor.count() > 0:\n\t\tif cursor[0][\"data1\"] != None and cursor[0][\"data2\"] != None and not cursor[0][\"finished\"]:\n\t\t\tupdateGame(identifications, {\"finished\": True})\n\t\t\tsubIdentifications = dict()\n\t\t\tsubIdentifications[\"name\"] = str(cursor[0][\"theme\"])\n\t\t\tif int(cursor[0][\"gameType\"]) == 1:\n\t\t\t\tcategory = subCategoryBelongsTo(subIdentifications)\n\t\t\t\tif category == None: \n\t\t\t\t\treturn False\n\t\t\telif int(cursor[0][\"gameType\"]) == 2:\n\t\t\t\t#cur = mongo.db.categories.find(subIdentifications)\n\t\t\t\tcur = db.categories.find(subIdentifications)\n\t\t\t\tif cur.count() == 0:\n\t\t\t\t\treturn False\n\t\t\t\telse:\n\t\t\t\t\tcategory = cur[0]\n\t\t\tdata1 = str.split(str(cursor[0][\"data1\"]).lower(), \"||\")\n\t\t\tdata2 = str.split(str(cursor[0][\"data2\"]).lower(), \"||\")\n\t\t\tgameType = int(cursor[0][\"gameType\"])\n\t\t\tupdates = dict()\n\t\t\tif gameType == 1:\n\t\t\t\tscore1, score2 = calculateScores(data1, data2, category[\"name\"], 1)\n\t\t\t\tupdates[\"score1\"] = score1\n\t\t\t\tidUser = dict()\n\t\t\t\tidUser[\"_id\"] = cursor[0][\"user1\"]\n\t\t\t\tupdateScore(idUser, score1)\n\t\t\t\tif \"score2\" in cursor[0].keys():\n\t\t\t\t\tupdates[\"score2\"] = score2\n\t\t\t\t\tidUser[\"_id\"] = cursor[0][\"user2\"]\n\t\t\t\t\tupdateScore(idUser, score2)\n\t\t\telif gameType == 2:\n\t\t\t\tscore1 = int(cursor[0][\"score1\"])\n\t\t\t\tidUser = dict()\n\t\t\t\tidUser[\"_id\"] = cursor[0][\"user1\"]\n\t\t\t\tupdateScore(idUser, score1)\n\t\t\t\tfor i in range(cursor[0][\"score2\"]):\n\t\t\t\t\tfor category in data1:\n\t\t\t\t\t\tentity = data2[i]\n\t\t\t\t\t\t#exists, score = existsInNell(entity, category)\n\t\t\t\t\t\texists, score = None, -1\n\t\t\t\t\t\tfbIdent = dict()\n\t\t\t\t\t\tfbUpdates = dict()\n\t\t\t\t\t\tfbIdent[\"entity\"] = entity\n\t\t\t\t\t\tfbIdent[\"category\"] = category\n\t\t\t\t\t\tfbUpdates[\"score\"] = score\n\t\t\t\t\t\tfbUpdates[\"count\"] = 1\n\t\t\t\t\t\tfbUpdates[\"lazy\"] = True\n\t\t\t\t\t\taddFeedback(fbIdent, fbUpdates, 2)\n\t\t\telif gameType == 3:\n\t\t\t\tscore1, score2 = int(cursor[0][\"score1\"]), int(cursor[0][\"score2\"])\n\t\t\t\tidUser = dict()\n\t\t\t\tidUser[\"_id\"] = cursor[0][\"user1\"]\n\t\t\t\tupdateScore(idUser, score1)\n\t\t\t\tidUser = dict()\n\t\t\t\tidUser[\"_id\"] = cursor[0][\"user2\"]\n\t\t\t\tupdateScore(idUser, score2)\n\t\t\t\tfor category in data1:\n\t\t\t\t\tentity = subIdentifications[\"name\"].split(\"||\")[0]\t\t\t\t\t\n\t\t\t\t\t#exists, score = existsInNell(entity, category)\n\t\t\t\t\texists, score = None, -1\n\t\t\t\t\tfbIdent, fbUpdates = dict(), dict()\n\t\t\t\t\tfbIdent[\"entity\"] = entity\n\t\t\t\t\tfbIdent[\"category\"] = category\n\t\t\t\t\tfbUpdates[\"score\"] = score\n\t\t\t\t\tfbUpdates[\"count\"] = 1\n\t\t\t\t\tfbUpdates[\"lazy\"] = True\n\t\t\t\t\taddFeedback(fbIdent, fbUpdates, 3)\n\t\t\t\tfor category in data2:\n\t\t\t\t\tentity = subIdentifications[\"name\"].split(\"||\")[1]\t\t\t\t\t\n\t\t\t\t\t#exists, score = existsInNell(entity, category)\n\t\t\t\t\texists, score = None, -1\n\t\t\t\t\tfbIdent, fbUpdates = dict(), dict()\n\t\t\t\t\tfbIdent[\"entity\"] = entity\n\t\t\t\t\tfbIdent[\"category\"] = category\n\t\t\t\t\tfbUpdates[\"score\"] = score\n\t\t\t\t\tfbUpdates[\"count\"] = 1\n\t\t\t\t\tfbUpdates[\"lazy\"] = True\n\t\t\t\t\taddFeedback(fbIdent, fbUpdates, 3)\n\t\t\telse:\n\t\t\t\tscore1 = int(cursor[0][\"score1\"])\n\t\t\t\tidUser = dict()\n\t\t\t\tidUser[\"_id\"] = cursor[0][\"user1\"]\n\t\t\t\tupdateScore(idUser, score1)\n\t\t\t\tfor i in range(cursor[0][\"score2\"]):\n\t\t\t\t\tfor entity in data1:\n\t\t\t\t\t\tcategory = data2[i]\n\t\t\t\t\t\t#exists, score = existsInNell(entity, category)\n\t\t\t\t\t\texists, score = None, -1\n\t\t\t\t\t\tfbIdent = dict()\n\t\t\t\t\t\tfbUpdates = dict()\n\t\t\t\t\t\tfbIdent[\"entity\"] = entity\n\t\t\t\t\t\tfbIdent[\"category\"] = category\n\t\t\t\t\t\tfbUpdates[\"score\"] = score\n\t\t\t\t\t\tfbUpdates[\"count\"] = 1\n\t\t\t\t\t\tfbUpdates[\"lazy\"] = True\n\t\t\t\t\t\taddFeedback(fbIdent, fbUpdates, 4)\n\t\t\tfinish = datetime.now().strftime('%Y-%m-%d %H:%M:%S')\n\t\t\tupdates[\"status\"] = 1\n\t\t\tupdates[\"finish\"] = finish\n\t\t\treturn updateGame(identifications, updates)\n\t\telse:\n\t\t\tupdates = dict()\n\t\t\tupdates[\"status\"] = 3\n\t\t\treturn updateGame(identifications, updates)\t\t\n\treturn False\n\n","sub_path":"query.py","file_name":"query.py","file_ext":"py","file_size_in_byte":10510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"565018273","text":"from app.api import bp\nfrom flask import jsonify, request, url_for, g, abort\nfrom app.models import User\nfrom app import db\nfrom app.api.errors import bad_request\nfrom app.api.auth import token_auth\n\n\n@bp.route('/users/', methods=['GET'])\n@token_auth.login_required\ndef get_user(id):\n\tresponse = jsonify(User.query.get_or_404(id).to_dict())\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n@bp.route('/users/', methods=['GET'])\n@token_auth.login_required\ndef get_user_username(username):\n\tresponse = jsonify(User.query.filter_by(username=username).first().to_dict())\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n\n@bp.route('/users', methods=['GET'])\n@token_auth.login_required\ndef get_users():\n\tpage = request.args.get('page', 1, type=int)\n\tper_page = min(request.args.get('per_page', 10, type=int), 100)\n\tdata = User.to_collection_dict(User.query, page, per_page, 'api.get_users')\n\tresponse = jsonify(data)\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n\n@bp.route('/users//followers', methods=['GET'])\n@token_auth.login_required\ndef get_followers(id):\n\tuser = User.query.get_or_404(id)\n\tpage = request.args.get('page', 1, type=int)\n\tper_page = min(request.args.get('per_page', 10, type=int), 100)\n\tdata = User.to_collection_dict(user.followers, page, per_page,'api.get_followers', id=id)\n\tresponse = jsonify(data)\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n@bp.route('/users//followed', methods=['GET'])\n@token_auth.login_required\ndef get_followed(id):\n\tuser = User.query.get_or_404(id)\n\tpage = request.args.get('page', 1, type=int)\n\tper_page = min(request.args.get('per_page', 10, type=int), 100)\n\tdata = User.to_collection_dict(user.followed, page, per_page,'api.get_followed', id=id)\n\tresponse = jsonify(data)\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n\n@bp.route('/users', methods=['POST'])\ndef create_user():\n\tdata = request.get_json() or {}\n\tif 'username' not in data or 'email' not in data or 'password' not in data:\n\t\treturn bad_request('must include username, email and password fields')\n\tif User.query.filter_by(username=data['username']).first():\n\t\treturn bad_request('please use a different username')\n\tif User.query.filter_by(email=data['email']).first():\n\t\treturn bad_request('please use a different email address')\n\tuser = User()\n\tuser.from_dict(data, new_user=True)\n\tdb.session.add(user)\n\tdb.session.commit()\n\tresponse = jsonify(user.to_dict())\n\tresponse.status_code = 201\n\tresponse.headers['Location'] = url_for('api.get_user', id=user.id)\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n\n\n@bp.route('/users/', methods=['PUT'])\n@token_auth.login_required\ndef update_user(id):\n\tprint(id)\n\tif g.current_user.id != id:\n\t\tprint(\"not current user\")\n\t\tabort(403)\n\tuser = User.query.get_or_404(id)\n\tdata = request.get_json() or {}\n\tprint(data)\n\tif 'username' in data and data['username'] != user.username and User.query.filter_by(username=data['username']).first():\n\t\treturn bad_request('please use a different username')\n\tif 'email' in data and data['email'] != user.email and User.query.filter_by(email=data['email']).first():\n\t\treturn bad_request('please use a different email address')\n\tuser.from_dict(data, new_user=False)\n\tdb.session.commit()\n\tresponse = jsonify(user.to_dict())\n\tresponse.headers.add('Access-Control-Allow-Origin', '*')\n\treturn response\n","sub_path":"app/api/users.py","file_name":"users.py","file_ext":"py","file_size_in_byte":3501,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"95911217","text":"from django.contrib import admin\nfrom django.urls import path\n\nfrom .views import (\n home, \n ProfilePageView, \n submittedView,\n about,\n volunteer,\n TranslateFormView,\n NarrateFormView,\n NarrationStatus,\n TranslationStatus,\n NarrateUpdate,\n TranslateUpdate,\n)\n\nurlpatterns = [\n path('', home, name='home'),\n path('about/', about, name='about'),\n path('volunteers/', volunteer, name='volunteers'),\n path('profile/', ProfilePageView.as_view(), name='profile'),\n path('submitted-translation/', TranslateFormView.as_view(), name='translate'),\n path('submitted-narration/', NarrateFormView.as_view(), name='narrate'),\n path('narration-status/', NarrationStatus, name='narrate-status'),\n path('narration//update/', NarrateUpdate.as_view(), name='edit-narration'),\n path('translation//update/', TranslateUpdate.as_view(), name='edit-translation'),\n path('translation-status/', TranslationStatus, name='translate-status'),\n]\n","sub_path":"pages/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":995,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"623515476","text":"import datetime\nfrom typing import List, Optional\n\nfrom pymongo import ReplaceOne, DESCENDING\nfrom pymongo.database import Database\n\nfrom pepy.domain.model import Project, ProjectDownloads, ProjectName, Downloads\nfrom pepy.domain.repository import ProjectRepository\n\n\nclass MongoProjectRepository(ProjectRepository):\n def __init__(self, client: Database):\n self._client = client\n self._client.projects.create_index([(\"name\", DESCENDING)])\n\n def get(self, project_name: str) -> Optional[Project]:\n project_data = self._client.projects.find_one({\"name\": project_name.strip().lower()})\n if project_data is None:\n return None\n project = Project(ProjectName(project_data[\"name\"]), Downloads(project_data[\"total_downloads\"]))\n downloads = sorted(project_data[\"downloads\"].items(), key=lambda x: x[0])\n for date, version_downloads in downloads:\n for r in version_downloads:\n project.add_downloads(datetime.date.fromisoformat(date), r[0], Downloads(r[1]))\n # Don't count the downloads twice\n project.total_downloads -= Downloads(r[1])\n return project\n\n def save(self, project: Project):\n data = self._convert_to_raw(project)\n query = {\"name\": project.name.name}\n self._client.projects.replace_one(query, data, upsert=True)\n\n def _convert_to_raw(self, project):\n data = {\n \"name\": project.name.name,\n \"total_downloads\": project.total_downloads.value,\n \"downloads\": {\n date.isoformat(): [(version, x.value) for version, x in list(versions.items())]\n for date, versions in project._latest_downloads.items()\n },\n }\n return data\n\n def save_projects(self, projects: List[Project]):\n requests = []\n for project in projects:\n requests.append(ReplaceOne({\"name\": project.name.name}, self._convert_to_raw(project), upsert=True))\n self._client.projects.bulk_write(requests)\n\n def update_downloads(self, projects_downloads: List[ProjectDownloads]):\n pass\n\n def save_day_downloads(self, project_downloads: List[ProjectDownloads]):\n pass\n","sub_path":"pepy/infrastructure/db_repository.py","file_name":"db_repository.py","file_ext":"py","file_size_in_byte":2222,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"12714397","text":"import numpy as np\nimport pandas as pd\n\nimport pyqtgraph as pg\nfrom pyqtgraph.Qt import QtCore, QtGui\n\nfrom ..spikesorter import SpikeSorter\nfrom .traceviewer import TraceViewer\nfrom .lists import PeakList, ClusterList\nfrom .ndscatter import NDScatter\n\nimport itertools\n\nclass SpikeSortingWindow(QtGui.QMainWindow):\n def __init__(self, spikesorter):\n QtGui.QMainWindow.__init__(self)\n \n self.spikesorter = spikesorter\n \n self.traceviewer = TraceViewer(spikesorter = spikesorter)\n self.peaklist = PeakList(spikesorter = spikesorter)\n self.clusterlist = ClusterList(spikesorter = spikesorter)\n self.ndscatter = NDScatter(spikesorter = spikesorter)\n \n all = [self.traceviewer, self.peaklist, self.ndscatter]\n \n for w1, w2 in itertools.combinations(all,2):\n w1.peak_selection_changed.connect(w2.on_peak_selection_changed)\n w2.peak_selection_changed.connect(w1.on_peak_selection_changed)\n \n docks = {}\n docks['traceviewer'] = QtGui.QDockWidget('traceviewer',self)\n docks['traceviewer'].setWidget(self.traceviewer)\n self.addDockWidget(QtCore.Qt.RightDockWidgetArea, docks['traceviewer'])\n docks['peaklist'] = QtGui.QDockWidget('peaklist',self)\n docks['peaklist'].setWidget(self.peaklist)\n self.addDockWidget(QtCore.Qt.LeftDockWidgetArea, docks['peaklist'])\n docks['ndscatter'] = QtGui.QDockWidget('ndscatter',self)\n docks['ndscatter'].setWidget(self.ndscatter)\n self.addDockWidget(QtCore.Qt.LeftDockWidgetArea, docks['ndscatter'])\n\n docks['clusterlist'] = QtGui.QDockWidget('clusterlist',self)\n docks['clusterlist'].setWidget(self.clusterlist)\n self.splitDockWidget(docks['peaklist'], docks['clusterlist'], QtCore.Qt.Horizontal)\n \n self.spikesorter.refresh_colors()\n\n \n @classmethod\n def from_classes(cls, dataio, peakdetector, waveformextractor, clustering):\n spikesorter = SpikeSorter(dataio = dataio)\n \n spikesorter.all_peaks = pd.DataFrame(np.zeros(peakdetector.peak_index.size, dtype = 'int32'), columns = ['label'], index = peakdetector.peak_index)\n spikesorter.all_peaks['label'] = clustering.labels\n spikesorter.all_peaks['selected'] = False\n spikesorter.all_waveforms = waveformextractor.get_ajusted_waveforms()\n spikesorter.clustering = clustering\n \n spikesorter.refresh_colors()\n \n return SpikeSortingWindow(spikesorter)\n \n ","sub_path":"tridesclous/gui/mainwindow.py","file_name":"mainwindow.py","file_ext":"py","file_size_in_byte":2551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"546575699","text":"for i in range(10):\n for j in range(i):\n print(i)\n j += 1\nprint(\"\\n\")\ni += 1\nprint(\" \")\nprint(\"--\")\nprint(\"-\")\nn = int(input(\"entrer un nobre\"))\ncpt = 0\nfor i in range (1,n):\n if( n%i == 0 ):\n cpt = cpt + i\nif(cpt == n):\n print(\"le nombre \",n,\"est parfait\")\nelse:\n print(\"le nombre \",n,\"n'est pas parfait\")","sub_path":"Exo2S.py","file_name":"Exo2S.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"219108305","text":"import numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\nfrom matplotlib.colors import LinearSegmentedColormap\n\nmina = -50.0\nmaxa = 1000.0\n\ncdict1 = {'red': ((0.0, 0.0, 0.0),\n (0.5, 0.0, 0.1),\n (1.0, 1.0, 1.0)),\n\n 'green': ((0.0, 0.0, 0.0),\n (1.0, 0.0, 0.0)),\n\n 'blue': ((0.0, 0.0, 1.0),\n (0.5, 0.1, 0.0),\n (1.0, 0.0, 0.0))\n }\nblue_red1 = LinearSegmentedColormap('BlueRed1', cdict1)\nplt.register_cmap(cmap=blue_red1)\n\n\ndef viz_file(fname, c='k'):\n xx = []\n yy = []\n cc = []\n\n for idx, line in enumerate(open(fname)):\n #if idx == 0: continue\n\n x, y, z = line.rstrip(' \\n').split(' ')\n x = float(x)\n y = float(y)\n z = float(z)\n if not (90 < y < 135):\n continue\n\n xx.append(x)\n yy.append(y)\n\n if z < mina: z = mina\n if z > maxa: z = maxa\n #z = (z - mina) / (maxa - mina)\n cc.append(z)\n\n sc = plt.scatter(yy, xx, s=10, c=cc, cmap=blue_red1, vmin=mina, vmax=maxa, edgecolors='none', alpha=0.5)\n plt.grid(b=True, which='both')\n plt.colorbar(sc)\n plt.subplots_adjust(left=0.05, right=0.97, top=0.95, bottom=0.05)\n\n\n\ndef compare_viz():\n plt.figure(1, figsize=(12,8))\n plt.subplot(2,2,1)\n plt.ylim([0,50])\n plt.xlim([90,130])\n viz_file('pprand_alg_1e4.csv')\n\n plt.subplot(2,2,2)\n plt.ylim([0,50])\n plt.xlim([90,130])\n viz_file('rand_alt_2000.csv')\n\n plt.subplot(2,2,3)\n plt.ylim([0,50])\n plt.xlim([90,130])\n viz_file('stratified_alt_2000.csv')\n\n plt.subplot(2,2,4)\n plt.ylim([0,50])\n plt.xlim([90,130])\n viz_file('mcs_alt_2000.csv')\n\n plt.show()\n\n\ndef viz_single():\n viz_file('rand_alt_1e6.csv', c = 'k')\n #viz_file('pp_mcs_1e4.csv', c = 'k')\n plt.xlim([80, 135])\n plt.ylim([0, 60])\n plt.show()\n\n\nif __name__ == \"__main__\":\n viz_single()\n","sub_path":"dataset/geolife/viz.py","file_name":"viz.py","file_ext":"py","file_size_in_byte":1957,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"373858535","text":"from django.contrib import admin\nfrom django.contrib.auth.admin import UserAdmin\n\nfrom app.models import CustomUser, Community, Post, Comment, Bookmark, DisLikePost, LikePost, ActivityStream, \\\n CommunityCategory\nfrom .forms import CustomUserCreationForm, CustomUserChangeForm\n\n\n# Register your models here.\n\nclass CustomUserAdmin(UserAdmin):\n add_form = CustomUserCreationForm\n form = CustomUserChangeForm\n model = CustomUser\n list_display = ['username', 'email']\n\n # For customizing the form in Abstract User\n\n add_fieldsets = (\n (None, {\n 'classes': ('wide',),\n 'fields': ('username', 'password1', 'password2', 'community','image')}\n ),\n )\n\n\nadmin.site.register(CustomUser, CustomUserAdmin)\n\n\nclass PostAdmin(admin.ModelAdmin):\n prepopulated_fields = {'post_slug': ('title',)}\n\n\nclass CommunityAdmin(admin.ModelAdmin):\n prepopulated_fields = {'name_slug': ('name',)}\n\n\n# Update the registration to include this customised interface\n\n# register community\nadmin.site.register(Community, CommunityAdmin)\nadmin.site.register(Post, PostAdmin)\nadmin.site.register(Comment)\nadmin.site.register(Bookmark)\nadmin.site.register(LikePost)\nadmin.site.register(DisLikePost)\nadmin.site.register(ActivityStream)\nadmin.site.register(CommunityCategory)\n","sub_path":"app/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"102634251","text":"# coding=utf-8\n\nimport json\nimport os\nimport sys\n\nbase_path = os.path.abspath(os.path.join(os.getcwd(), \"..\"))\nsys.path.append(base_path)\n\n\ndef read_json(file_name=None):\n if file_name == None:\n file_path = base_path+\"/data/api_data.json\"\n else:\n file_path = base_path+file_name\n\n with open(file_path,encoding='UTF-8') as f:\n data = json.load(f)\n return data\n\n\ndef get_value(key, file_name=None):\n data = read_json(file_name)\n return data.get(key)\n\n\ndef write_value(data, file_name=None):\n data_value = json.dumps(data)\n if file_name == None:\n path = base_path+\"/Config/cookie.json\"\n else:\n path = base_path+file_name\n with open(path, \"w\") as f:\n f.write(data_value)\n","sub_path":"common/get_data.py","file_name":"get_data.py","file_ext":"py","file_size_in_byte":741,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"492918988","text":"#!/usr/local/bin/python\n# encoding: utf-8\n# -*- coding: utf-8 -*-\n\nimport sys\nfrom jinja2 import Environment, FileSystemLoader\n\ndef create_article(name):\n _tmpArticle = open('./tmp/tmp.html')\n _title=None\n _sub_title=None\n _content=[]\n for i, line in enumerate(_tmpArticle):\n if i == 0:\n _title = line.decode('utf-8')\n elif i == 1:\n _sub_title = line.decode('utf-8')\n else:\n _content.append(line.decode('utf-8'))\n\n _tmpArticle.close()\n env = Environment(loader=FileSystemLoader('./templates'))\n template = env.get_template('article-template.html')\n _html = template.render(title=_title, sub_title=_sub_title, content=''.join(_content))\n _newArticle = open('articles-html/%s.html' % name, 'w+')\n _newArticle.write(_html.encode('utf-8'))\n _newArticle.close()\n\nif __name__=='__main__':\n create_article(sys.argv[1])\n","sub_path":"tools/generate_html.py","file_name":"generate_html.py","file_ext":"py","file_size_in_byte":905,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"477035162","text":"import pandas as pd\n\ndef convert_dates(x):\n x['businessDate']=pd.to_datetime(x['businessDate'])\n x['month']=x['businessDate'].dt.month\n x['is_month_start']=x['businessDate'].dt.is_month_start\n x['is_month_end']=x['businessDate'].dt.is_month_end\n x['year']=x['businessDate'].dt.year\n x['dayofweek']=x['businessDate'].dt.dayofweek\n x['quarter'] = x['businessDate'].apply(lambda x: x.quarter)\n x['week_of_year'] = x['businessDate'].apply(lambda x: x.weekofyear)\n x['day_of_year'] = x['businessDate'].apply(lambda x: x.dayofyear)\n x['Is_Mon'] = (x.dayofweek == 0) *1\n x['Is_Tue'] = (x.dayofweek == 1) *1\n x['Is_Wed'] = (x.dayofweek == 2) *1\n x['Is_Thu'] = (x.dayofweek == 3) *1\n x['Is_Fri'] = (x.dayofweek == 4) *1\n x['Is_Sat'] = (x.dayofweek == 5) *1\n x['Is_Sun'] = (x.dayofweek == 6) *1\n x['Is_wknd'] = x.dayofweek // 4\n x.pop('businessDate')\n # x.pop('year')\n return x\n\ndef merge(x,y,col,col_name):\n x =pd.merge(x, y, how='left', on=None, left_on=col, right_on=col,\n left_index=False, right_index=False, sort=True,\n copy=True, indicator=False,validate=None)\n x=x.rename(columns={'sales':col_name})\n return x","sub_path":"util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":1198,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"64001656","text":"import json\nimport pandas as pd\nimport re\nimport random\nimport argparse\n\n\ndef get_post_titles(inp):\n \"\"\"\n (file) --> (list)\n this function takes as input a file with Reddit post collected with the reddit API\n returns a list of all the titles of each of the posts.\n \"\"\"\n list_of_titles = []\n file_in = open(inp,'r')\n for line in file_in:\n \n data = json.loads(line)\n list_of_titles.append(data['data']['title'])\n return list_of_titles\n \n\n\ndef get_titles_by_candidate(list_of_titles, candidate):\n \"\"\"\n (list, string (biden or trump in lower case)) --> list\n \n This takes as input a list of post titles and returns a list of post titles containing \n the name of the candidate.\n \"\"\"\n if candidate != 'trump' and candidate != 'biden':\n raise ValueError ('candidate must be equal to \"trump\" or \"biden\" ') \n \n titles_containing_the_candidate = []\n for title in list_of_titles:\n lower_title = title.lower()\n if re.search(f\"[^0-9a-zA-Z]{candidate}[^0-9a-zA-Z]\", lower_title) or re.search(f\"{candidate}[^0-9a-zA-Z]\", lower_title):\n titles_containing_the_candidate.append(title)\n \n return titles_containing_the_candidate\n\n \ndef choose_random_line(list_of_post, num_post):\n \"\"\"\n (list, int) --> list\n This function takes as input a list of titles posts and a number. \n It returns a list of list of length of that number containing radomly selected posts\n from list_of_posts.\n \"\"\"\n \n random_post = []\n \n while(len(random_post) < num_post):\n mytitle = list_of_post.pop(random.randint(0,len(list_of_post)-1))\n random_post.append(mytitle)\n return random_post\n\n\n \n \n \ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('input_file_d1', help = 'this is the file for one of your reddit post collection')\n parser.add_argument('input_file_d2')\n parser.add_argument('input_file_d3')\n parser.add_argument('-c','--candidate')\n parser.add_argument('-o','--output_file')\n args = parser.parse_args()\n \n ## We now get the list of titles for the 3 days we collected the reddit data from\n titles_day1 = get_post_titles(args.input_file_d1)\n titles_day2 = get_post_titles(args.input_file_d2)\n titles_day3 = get_post_titles(args.input_file_d3)\n \n # Now from the lists of titles from the 3 days we get the ones containing the name of \n # of the candidate of our choice\n candidate_titles_day1 = get_titles_by_candidate(titles_day1, args.candidate)\n candidate_titles_day2 = get_titles_by_candidate(titles_day2, args.candidate)\n candidate_titles_day3 = get_titles_by_candidate(titles_day3, args.candidate)\n \n \n # We now chose randomly for the three list of titles containign \n shortlist_day1 = choose_random_line(candidate_titles_day1,66)\n shortlist_day2 = choose_random_line(candidate_titles_day2,66)\n shortlist_day3 = choose_random_line(candidate_titles_day3,66)\n \n \n \n sample_titles = []\n larger_list = [shortlist_day1,shortlist_day2]#,shortlist_day3]\n for sublist in larger_list:\n for title in sublist:\n sample_titles.append(title)\n \n \n \n posts = {'titles': sample_titles}\n \n df = pd.DataFrame(posts,columns = ['titles'])\n \n df.to_csv(f'{args.output_file}.csv', index = False, encoding = 'utf-8')\n \nif __name__ == '__main__':\n main()\n","sub_path":"filtering_the_post.py","file_name":"filtering_the_post.py","file_ext":"py","file_size_in_byte":3446,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"590748294","text":"# -*- coding: utf-8 -*-\n\"\"\"\n\n@author: Carbognin Alberto\n\"\"\"\n\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport os\nimport json\nimport random\n\nenumDict = {'FuelStation':0,\n 'ParkingArea':1,\n 'RailwayStation':2,\n 'BusStation':3,\n 'CarSharingPark':4,\n 'BikeSharingPark':5,\n 'Campsite':6}\n\nif __name__ == \"__main__\":\n print(\"Loading dataSet...\")\n\n print('Start making datasets...')\n Facility = pd.DataFrame(columns=['FacilityID','RouteID','Price','Contact','FacilityEnum','Address','Calendar'])\n Price =pd.DataFrame(columns=['FacilityID','TicketID','AutonomousTransportID','Cost','CurrencyType'])\n Contact = pd.DataFrame(columns=['AgencyID','FacilityID','Phone','Email','Website'])\n Address = pd.DataFrame(columns=['FacilityID','RouteID','Province','City','Street','Number','Cap','Location'])\n Location = pd.DataFrame(columns=['AddressID','StopId','Latitude','Longitude','Altitude'])\n Calendar = pd.DataFrame(columns=['FacilityID','StopTimesID','ServiceID','Monday','Tuesday','Wednesday',\n 'Thursday','Friday','Saturday','Sunday','StartDate','EndDate','Exceptions'])\n CalendarDates = pd.DataFrame(columns=['CalendarId','ServiceId','Date','ExceptionType'])\n\n\n #Load the file\n\n json_file_path = '../../dataset/Formal Modeling/data/bikesharing_assured.json'\n\n with open(json_file_path, 'r') as j:\n bikesharing_areas = json.loads(j.read())\n #fuel_pumps = json.loads('../../dataset/Formal Modeling/data/fuel_pumps_assured.json')\n\n print(json.loads(bikesharing_areas[0]))\n\n print(\"Found: {} records in the dataset.\".format(len(bikesharing_areas)))\n \n for i in range(len(bikesharing_areas)):\n bikesharing_area = bikesharing_areas[i]\n \n if i%300 == 0:\n print(\"Processing {}/{} record.\".format(i, len(bikesharing_areas))) \n record = json.loads(bikesharing_area)\n\n address = record['address']\n #print(\"\\taddress:\", address)\n\n location = address['location']\n contact = record['contact']\n\n # ['AgencyID','FacilityID','Phone','Email','Website']\n Contact.loc[i] = [None, i, contact['phone'], contact['email'], contact['website']]\n\n # ['AddressID','StopId','Latitude','Longitude','Altitude']\n Location.loc[i] = [i, None, location['latitude'], location['longitude'], location['altitude']]\n\n # ['FacilityID','RouteID','Province','City','Street','Number','Cap','Location']\n Address.loc[i] = [None, None, address['province'], address['city'], address['street'], address['number'], address['cap'], i]\n\n # ['FacilityID','RouteID','Price','Contact','FacilityEnum','Address','Calendar']\n Facility.loc[i] = [i, None, None, i, 5, i, None]\n\n\n\n\n\n \n print('Start exporting datasets...')\n exportPath = '../../dataset/Data Integration/data/BikeSharingArea/'\n os.mkdir(exportPath)\n Facility.to_csv(exportPath+'Facility.csv')\n Price.to_csv(exportPath+'Price.csv')\n Contact.to_csv(exportPath+'Contact.csv')\n Address.to_csv(exportPath+'Address.csv')\n Location.to_csv(exportPath+'Location.csv')\n Calendar.to_csv(exportPath+'Calendar.csv')\n CalendarDates.to_csv(exportPath+'CalendarDates.csv')\n print('export done.')\n \n","sub_path":"code/Code/DataIntegration/BikeSharingKarma.py","file_name":"BikeSharingKarma.py","file_ext":"py","file_size_in_byte":3504,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"23226823","text":"in_parametrs = ''\nnum = \"{:{align}{width}.{precision}f}\"\nwhile in_parametrs != 'F':\n\n a = int(input(\"Введите результат первого дня - \"))\n b = int(input(\"введите общий километраж - \"))\n\n day = 0\n while a < b:\n day = day + 1\n if day == 1:\n a = a\n else:\n a = a * 1.1\n #print(num.format(a, align='<', width=8, precision=2))\n aa = num.format(a, align='<', width=8, precision=2)\n print(f'{day} день пройдено {aa}')\n print(f'Потребовалось {day} дней')\n\n in_parametrs = input('Введите F д��я окончания программы')\n","sub_path":"Урок1-6.py","file_name":"Урок1-6.py","file_ext":"py","file_size_in_byte":692,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"260931266","text":"from hytest import *\nfrom lib.webapi import apimgr\n\nclass c2:\n name = '添加客户 - API-0152'\n\n #清除方法\n def teardown(self):\n apimgr.customer_del(self.addedCustomerId)\n\n def teststeps(self):\n\n STEP(1,'先列出客户')\n r = apimgr.customer_list()\n listRet1 = r.json()\n customerlist1 = listRet1[\"retlist\"]\n\n\n\n STEP(2, '添加一个客户')\n r = apimgr.customer_add('南京市鼓楼医院',\n '13345679934',\n \"南京市鼓楼北路\")\n\n addRet = r.json()\n\n self.addedCustomerId = addRet['id']\n\n CHECK_POINT('返回的ret值=0',\n addRet['ret'] == 0)\n\n\n STEP(3, '再次列出客户')\n\n r = apimgr.customer_list(11)\n\n listRet = r.json()\n\n expected = {\n \"ret\": 0,\n \"retlist\": [\n {\n \"address\": \"南京市鼓楼北路\",\n \"id\": addRet['id'],\n \"name\": \"南京市鼓楼医院\",\n \"phonenumber\": \"13345679934\"\n }\n ] + customerlist1,\n 'total': 11\n }\n\n CHECK_POINT('返回的消息体数据正确',\n expected == listRet)\n\n","sub_path":"hytest/autotest_bysms_03/cases/数据环境-空白/数据环境-10个客户/客户API/添加客户.py","file_name":"添加客户.py","file_ext":"py","file_size_in_byte":1281,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"643738200","text":"\"\"\"\nBy listing the first six prime numbers: 2, 3, 5, 7, 11, and 13, we can see that the 6th prime is 13.\n\nWhat is the 10001st prime number?\n\"\"\"\nfrom itertools import count\n\n\ndef main():\n generator = prime_generator(10001)\n p\n for p in generator:\n print(p)\n\n\ndef is_prime(n, primes):\n for i in primes:\n if n % i == 0:\n return False\n return True\n\n\ndef prime_generator(n):\n primes = []\n for i in count(2):\n if len(primes) >= n:\n break\n if is_prime(i, primes):\n primes.append(i)\n yield i\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"src/project-euler/problem7.py","file_name":"problem7.py","file_ext":"py","file_size_in_byte":617,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"215543263","text":"import numpy\n\n# This file create the Hamiltonian trajectory for LH2 using the structure\n# from the 1kzu pdb structure file included.\n# To run the code use python 3.7 or newer and type:\n# python GenHam.py\n\nE0=12255 # cm-1 The B850 Chromophore average gap\nE1=240 # cm-1 The extra energy for the B800 Chromophore gap\nMU0=4.481 # Debye The transition dipole before the application of a scale factor\nsigma=320 # cm-1 The width of the energy distribution for B850\nsigma1=141 # cm-1 The width of the energy distribution for B800\ntau=150 # fs The correlation time for the overdamped Brownian oscillators\nfcs=1.25; # Factor to scale couplings (resulting in an effective dipole of 5.001 Debye)\ndt=3 # fs Time between generated snapshots\nunits=27 # The number of chromophores\nsteps=200000 # Number of timesteps\n\n# Set op constants for generating fluctuations\nalpha=numpy.exp(-dt/tau)\nbeta=numpy.sqrt(1-alpha**2)\ndsiatoicm=(0.393430307**2)*219474.631370515*(0.529177249**3) # Conv from deb^2/a^3 to cm-1\n\n# Initialize arrays \n# Positions of Mg, NB, and ND\nx=numpy.zeros((units,3))\ny=numpy.zeros((units,3))\nz=numpy.zeros((units,3))\n# Dipole moment\nmu=numpy.zeros((units,3))\n# Dye position\nr=numpy.zeros((units,3))\n# Hamiltonian\nH=numpy.zeros(int(units*(units+1)/2))\n# Dipole moments\nmu4bin=numpy.zeros((units*3))\n# Positions\npos4bin=numpy.zeros((units*3))\n# Helping arrays for Hamiltonian construction\nB800=numpy.zeros((units))\nssigma=numpy.ones((units))*sigma\n\n# Open files\nfile_input=open(\"1kzu.pdb\",\"r\")\nfile_H=open(\"Energy.bin\",\"wb\")\nfile_x=open(\"Positions.txt\",\"w\")\nfile_mu=open(\"Dipole.bin\",\"wb\")\nfile_pos=open(\"Positions.bin\",\"wb\")\nfile_HH=open(\"Ham.txt\",\"w\")\nfile_dp=open(\"Dipole.txt\",\"w\")\n\n# Symmetry constants to recover C3 symmetry operations\nsa=-0.5\nsb=0.866025\n\n# Read structure from pdb file\nindex=0\nwhile True:\n data=file_input.readline()\n if not data: break\n words=data.split()\n if len(words)>2:\n if words[2]==\"MG\":\n x[index][0]=float(words[6])\n y[index][0]=float(words[7])\n z[index][0]=float(words[8])\n x[index+9][0]=sa*x[index][0]-sb*y[index][0]\n y[index+9][0]=sb*x[index][0]+sa*y[index][0]\n z[index+9][0]=z[index][0]\n x[index+18][0]=sa*x[index][0]+sb*y[index][0]\n y[index+18][0]=-sb*x[index][0]+sa*y[index][0]\n z[index+18][0]=z[index][0]\n if words[2]==\"NB\":\n x[index][1]=float(words[6])\n y[index][1]=float(words[7])\n z[index][1]=float(words[8])\n x[index+9][1]=sa*x[index][1]-sb*y[index][1]\n y[index+9][1]=sb*x[index][1]+sa*y[index][1]\n z[index+9][1]=z[index][1]\n x[index+18][1]=sa*x[index][1]+sb*y[index][1]\n y[index+18][1]=-sb*x[index][1]+sa*y[index][1]\n z[index+18][1]=z[index][1]\n if words[2]==\"ND\":\n x[index][2]=float(words[6])\n y[index][2]=float(words[7])\n z[index][2]=float(words[8])\n x[index+9][2]=sa*x[index][2]-sb*y[index][2]\n y[index+9][2]=sb*x[index][2]+sa*y[index][2]\n z[index+9][2]=z[index][2]\n x[index+18][2]=sa*x[index][2]+sb*y[index][2]\n y[index+18][2]=-sb*x[index][2]+sa*y[index][2]\n z[index+18][2]=z[index][2]\n index=index+1\n\n# Verify that the correct number (9) of unique dye atoms were read from file\nprint(index)\n\n# Construct arrays with transition dipole moments and positions\nfor atom in range(27):\n # Write positions to human readable file\n file_x.write(str(x[atom][0]) + \" \" + str(y[atom][0]) + \" \" + str(z[atom][0]) + \"\\n\")\n mu[atom][0]=x[atom][2]-x[atom][1]\n mu[atom][1]=y[atom][2]-y[atom][1]\n mu[atom][2]=z[atom][2]-z[atom][1] \n mum=numpy.sqrt(mu[atom][0]**2+mu[atom][1]**2+mu[atom][2]**2)\n mu[atom][0]=mu[atom][0]/mum\n mu[atom][1]=mu[atom][1]/mum\n mu[atom][2]=mu[atom][2]/mum\n # Write transition dipole moments to human readable file\n file_dp.write(str(mu[atom][0]) + \" \" + str(mu[atom][1]) + \" \" + str(mu[atom][2]) + \"\\n\")\n r[atom][0]=x[atom][0]\n r[atom][1]=y[atom][0]\n r[atom][2]=z[atom][0]\n # Store data in arrays for saving in binary files\n mu4bin[atom]=mu[atom][0]\n mu4bin[units+atom]=mu[atom][1]\n mu4bin[2*units+atom]=mu[atom][2]\n pos4bin[atom]=r[atom][0]\n pos4bin[units+atom]=r[atom][1]\n pos4bin[2*units+atom]=r[atom][2]\n\n# Generate helping arrays for Hamiltonian construction.\n# Accounting for difference between B850 and B800 chromophores\nfor atom in range(9):\n B800[3*atom+1]=E1\n ssigma[3*atom+1]=sigma1\n\n# Create Hamiltonian first off-diagonal part\nfor ai in range(27):\n for aj in range(ai):\n if ai!=aj:\n rr=r[ai,:]-r[aj,:]\n rd=numpy.sqrt(rr[0]**2+rr[1]**2+rr[2]**2)\n # Equation for transition-dipole coupling\n J=sum(mu[ai,:]*mu[aj,:])/(rd**3)-3*sum(mu[ai,:]*rr[:])*sum(mu[aj,:]*rr[:])/(rd**5)\n # Convert to cm-1\n J=J*dsiatoicm*MU0*MU0*fcs\n # Do indexing for tridiagonal matrix\n ind=int(ai+units*aj-(aj*(aj+1)/2))\n # Print information to screen\n print(ai)\n print(aj)\n print(ind)\n H[ind]=J\n print(J)\n\n# Create initial random numbers\ndiag=numpy.random.randn(27)\n# Create diagonal elements\nfor st in range(steps):\n for ai in range(27):\n ind=int(ai+units*ai-(ai*(ai+1)/2))\n # Find energy gap including shift for B800 chromophores\n H[ind]=diag[ai]*ssigma[ai]+E0+B800[ai]\n # Update random numbers according to J. Chem. Phys. 127:084507 (2007) \n diag=diag*alpha+numpy.random.randn(27)*beta\n # Save Hamiltonian, dipoles, and positions to binary files\n Hf=numpy.array(H,'float32')\n step=numpy.array([0],'float32')\n step.tofile(file_H)\n Hf.tofile(file_H)\n step.tofile(file_mu)\n muf=numpy.array(mu4bin,'float32')\n muf.tofile(file_mu)\n step.tofile(file_pos)\n puf=numpy.array(pos4bin,'float32')\n puf.tofile(file_pos)\n\n# Write square Hamiltonian to hunam readable file\nfor ai in range(27):\n for aj in range(27):\n if ajai:\n ind=int(aj+units*ai-(ai*(ai+1)/2))\n if ai==aj:\n file_HH.write(str(E0+B800[ai]) + \" \")\n if ai!=aj:\n file_HH.write(str(H[ind]) + \" \")\n file_HH.write(\"\\n\")\n\n# Close all files\nfile_input.close\nfile_H.close\nfile_x.close\nfile_mu.close\nfile_pos.close\nfile_dp.close\n","sub_path":"GenHam.py","file_name":"GenHam.py","file_ext":"py","file_size_in_byte":6287,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"489456130","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom sklearn.cluster import DBSCAN\nfrom collections import Counter\nimport glob\nimport re\nimport pandas as pd\nimport sys\nimport pyfits\nfrom astropy.table import Table\n\nCCD = sys.argv[1]\ndir_in_CCD = '/home/student01/out_team1/'+CCD\ndir_in_CCD_image = '/home/student01/out_team1/'+CCD\ndir_out_CCD = '/home/student02/dic/' + CCD +'object'\ndir_out_CCD_moving = '/home/student02/dic/' + CCD +'moving'\n\ndef get_column_name(filename):\n column_names = []\n with open(filename) as f:\n for line in f:\n if line[0] == '#':\n column_names.append(line.split()[2])\n else:\n break\n \n return column_names\n\ndef file_number(file_name):\n #_index = [m.start() for m in re.finditer('_', file_name)][-1]\n #ext_index = [m.start() for m in re.finditer('\\.', file_name)][-1]\n pieces = file_name.split('/')[-1]\n #print pieces\n numbers = pieces.split('_')[3]\n #print numbers\n return int(numbers)\n #return int(file_name[_index+1:ext_index])\n\n\nfiles = glob.glob(dir_in_CCD+\"*.fits.cat\")\n#print files\nfiles = sorted(files, key=lambda f: file_number(f))\n\n#print files\n#images = glob.glob(dir_in_CCD_image+\"*.fz\")\n\n#print files\n#print images\n#columns = [0,55,56,174,11,10,180]\ncolumn_names = get_column_name(files[0])\nData = None\n#start = 0\n#data_by_epochs = []\n\ninfo_name = ['NUMBER','X_WORLD','Y_WORLD','FLAGS','FLUXERR_AUTO','FLUX_AUTO']\nfor i, file_ in enumerate(files):\n \n #f = np.loadtxt(file_)[:,columns]\n #hdulist = pyfits.open(images[i])\n #epoc_time = hdulist[0].header['MJD-OBS']i\n #epoc_time = i\n #f = np.loadtxt(file_)\n #f = pd.DataFrame(data=f, columns = column_names)\n t = Table.read(file_, table_id=0)\n f = pd.DataFrame(np.array(t))\n f = f[info_name].values\n #print f.shape\n #f = np.hstack((f,np.zeros(f.shape[0], dtype=float).reshape((f.shape[0],1))+epoc_time))\n f = np.hstack((f,np.zeros(f.shape[0], dtype=int).reshape((f.shape[0],1))+i))\n f = np.hstack((f,np.zeros(f.shape[0], dtype=int).reshape((f.shape[0],1))+file_number(file_)))\n #f = np.hstack((f,np.array(f.shape[0]*[file_]).reshape((f.shape[0],1))))\n\n #print f[0,:]\n if Data is None:\n Data = f\n else:\n Data = np.vstack((Data,f))\n\nflag_threshold = 0\nfilt = np.ones(Data.shape[0], dtype=bool)\nfilt[Data[:,3]>flag_threshold] = False\nData = Data[filt,:]\n\nsnr_threshold = 0.15\nfilt2 = np.ones(Data.shape[0], dtype=bool)\nsnr = Data[:,4]/Data[:,5]\nfilt2[snr>snr_threshold] = False\nData = Data[filt2,:]\n\nX = Data[:,[1,2]]\n\ndb = DBSCAN(eps=0.0003, min_samples=3).fit(X)\n\ncore_samples_mask = np.zeros_like(db.labels_, dtype=bool)\ncore_samples_mask[db.core_sample_indices_] = True\nlabels = np.asarray(db.labels_, dtype=int)\n\nn_clusters_ = len(set(labels)) - (1 if -1 in labels else 0)\n\nunique_labels = set(labels)\n\nn_stat = labels[labels>-1].size\n\nstatic = np.zeros(n_clusters_*len(files)).reshape((n_clusters_,len(files))) - 1\n#print Data.shape\n#print len(labels)\n\nfor i in xrange(len(labels)):\n#for i in xrange(len(labels)):\n if labels[i]>-1:\n static[labels[i],Data[i,6]] = Data[i, 0]\n#print labels\npdstatic= pd.DataFrame(data=static, columns = files)\n\nmoving = Data[:, [0,1,2,7]]\n\nmoving = moving[labels==-1,:]\n\n#for i, file_ in enumerate(files):\n#\tData[Data[:,-1]==i,-1] = file_\n\npdstatic.to_pickle(dir_out_CCD)\n\npdmoving = pd.DataFrame(data=moving, columns = ['NUMBER','X','Y','EPOCH'])\npdmoving.to_pickle(dir_out_CCD_moving)\n","sub_path":"code group 2/backup/step1.py","file_name":"step1.py","file_ext":"py","file_size_in_byte":3494,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"459933414","text":"# Time Complexity : O(n + m) \n# Space Complexity : O(1) (We are running algorithm in place)\n# Did this code successfully run on Leetcode : Yes\n# Three line explanation of solution in plain english:\n# - This is similer to merging two array in divide and conqure for sorting but in reverse.\n \nclass Solution:\n def merge(self, nums1: List[int], m: int, nums2: List[int], n: int) -> None:\n \"\"\"\n Do not return anything, modify nums1 in-place instead.\n \"\"\"\n# Initialize m, n and k. m and n is already given, we have to just reduce it by 1 because array starts from 0. K is sum of m and n or length of first array which also has exact space for merging second array.\n m = m - 1\n n = n - 1\n k = len(nums1) - 1\n# We will run this loop until m or n reach less than zero.\n while(n > -1 and m > -1):\n# If number in nums1 is greater than number in nums2we will append it to k index and reduce m index\n if nums1[m] > nums2[n]:\n nums1[k] = nums1[m]\n m -= 1\n# If number in nums2 is greater than number in nums1 we will append it to k index and reduce n index\n elif nums1[m] < nums2[n]:\n nums1[k] = nums2[n]\n n -= 1\n# If both numer are same we can append both of them to the k. M and n will be reduced once but k needs to reduce twice.\n else:\n nums1[k] = nums1[m]\n k -= 1\n m -= 1\n nums1[k] = nums2[n]\n n -= 1\n# Every time we are reducing k\n k -= 1\n \n# because we are stopping our while loop if m or n any one of them reach to -1. The other one might not have reached to -1. \n# Checking if m reached to -1.(If all element from nums1 are appended or not)\n if m > -1:\n# If some elements remained in nums1 we will append it to k\n while( m > -1):\n nums1[k] = nums1[m]\n k -= 1\n m -= 1\n \n# Checking if n reached to -1. \n if n > -1:\n# If some elements remained in nums2 we will append it to k\n while (n > -1):\n nums1[k] = nums2[n]\n k -= 1\n n -= 1\n","sub_path":"Problem2.py","file_name":"Problem2.py","file_ext":"py","file_size_in_byte":2292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"411620130","text":"\"\"\"\nThis code is to implement the IndRNN (only the recurrent part). The code is based on the implementation from \nhttps://github.com/StefOe/indrnn-pytorch/blob/master/indrnn.py.\nSince this only contains the recurrent part of IndRNN, fully connected layers or convolutional layers are needed before it.\nPlease cite the following paper if you find it useful.\nShuai Li, Wanqing Li, Chris Cook, Ce Zhu, and Yanbo Gao. \"Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN,\" \nIn Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5457-5466. 2018.\n@inproceedings{li2018independently,\n title={Independently recurrent neural network (indrnn): Building A longer and deeper RNN},\n author={Li, Shuai and Li, Wanqing and Cook, Chris and Zhu, Ce and Gao, Yanbo},\n booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},\n pages={5457--5466},\n year={2018}\n}\n\"\"\"\n\n\nimport torch\nfrom torch.nn import Parameter\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torch.autograd import Variable\nimport math\n\n\nclass IndRNNCell_onlyrecurrent(nn.Module):\n r\"\"\"An IndRNN cell with ReLU non-linearity. This is only the recurrent part where the input is already processed with w_{ih} * x + b_{ih}.\n\n .. math::\n input=w_{ih} * x + b_{ih}\n h' = \\relu(input + w_{hh} (*) h)\n With (*) being element-wise vector multiplication.\n\n Args:\n hidden_size: The number of features in the hidden state h\n\n Inputs: input, hidden\n - **input** (batch, input_size): tensor containing input features\n - **hidden** (batch, hidden_size): tensor containing the initial hidden\n state for each element in the batch.\n\n Outputs: h'\n - **h'** (batch, hidden_size): tensor containing the next hidden state\n for each element in the batch\n \"\"\"\n\n def __init__(self, hidden_size, \n hidden_max_abs=None, recurrent_init=None):\n super(IndRNNCell_onlyrecurrent, self).__init__()\n self.hidden_size = hidden_size\n self.recurrent_init = recurrent_init\n self.weight_hh = Parameter(torch.Tensor(hidden_size)) \n self.reset_parameters()\n\n def reset_parameters(self):\n for name, weight in self.named_parameters():\n if \"weight_hh\" in name:\n if self.recurrent_init is None:\n nn.init.uniform(weight, a=0, b=1)\n else:\n self.recurrent_init(weight)\n\n def forward(self, input, hx):\n return F.relu(input + hx * self.weight_hh.unsqueeze(0).expand(hx.size(0), len(self.weight_hh)))\n\n\nclass IndRNN_onlyrecurrent(nn.Module):\n r\"\"\"Applies an IndRNN with `ReLU` non-linearity to an input sequence. \n This is only the recurrent part where the input is already processed with w_{ih} * x + b_{ih}.\n\n\n For each element in the input sequence, each layer computes the following\n function:\n\n .. math::\n\n h_t = \\relu(input_t + w_{hh} (*) h_{(t-1)})\n\n where :math:`h_t` is the hidden state at time `t`, and :math:`input_t`\n is the input at time `t`. (*) is element-wise multiplication.\n\n Args:\n hidden_size: The number of features in the hidden state `h` \n\n Inputs: input, h_0\n - **input** of shape `(seq_len, batch, input_size)`: tensor containing the features\n of the input sequence. The input can also be a packed variable length\n sequence. See :func:`torch.nn.utils.rnn.pack_padded_sequence`\n or :func:`torch.nn.utils.rnn.pack_sequence`\n for details.\n - **h_0** of shape `( batch, hidden_size)`: tensor\n containing the initial hidden state for each element in the batch.\n Defaults to zero if not provided.\n\n Outputs: output \n - **output** of shape `(seq_len, batch, hidden_size)`\n \"\"\"\n\n def __init__(self, hidden_size,recurrent_init=None, **kwargs):\n super(IndRNN_onlyrecurrent, self).__init__()\n self.hidden_size = hidden_size\n self.indrnn_cell=IndRNNCell_onlyrecurrent(hidden_size, **kwargs)\n\n if recurrent_init is not None:\n kwargs[\"recurrent_init\"] = recurrent_init\n self.recurrent_init=recurrent_init\n # h0 = torch.zeros(hidden_size * num_directions)\n # self.register_buffer('h0', torch.autograd.Variable(h0))\n self.reset_parameters()\n\n def reset_parameters(self):\n for name, weight in self.named_parameters():\n if \"weight_hh\" in name:\n if self.recurrent_init is None:\n nn.init.uniform(weight, a=0, b=1)\n else:\n self.recurrent_init(weight)\n\n def forward(self, input, h0=None):\n assert input.dim() == 2 or input.dim() == 3 \n if h0 is None:\n h0 = input.data.new(input.size(-2),input.size(-1)).zero_().contiguous()\n elif (h0.size(-1)!=input.size(-1)) or (h0.size(-2)!=input.size(-2)):\n raise RuntimeError(\n 'The initial hidden size must be equal to input_size. Expected {}, got {}'.format(\n h0.size(), input.size()))\n outputs=[]\n hx_cell=h0\n for input_t in input:\n hx_cell = self.indrnn_cell(input_t, hx_cell)\n outputs.append(hx_cell)\n out_put = torch.stack(outputs, 0)\n return out_put\n","sub_path":"IndRNN_onlyrecurrent.py","file_name":"IndRNN_onlyrecurrent.py","file_ext":"py","file_size_in_byte":5388,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"378389856","text":"# https://github.com/agdelma/IntroCompPhysics/blob/master/Notebooks/16_SimpleHarmonicMotion.ipynb\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy.constants import pi as π\nfrom scipy.constants import g\n\nℓ = 0.25\nΔt = 0.01\n\nt = np.arange(0.0, 4.0, Δt)\nθ, ω = np.zeros_like(t), np.zeros_like(t)\nθ[0] = π/12.0\n\nbetterθ, betterω = np.zeros_like(t), np.zeros_like(t)\nbetterθ[0] = π/4\n\nfor n in range(t.size - 1):\n# θ[n + 1] = θ[n] + ω[n] * Δt\n# ω[n + 1] = ω[n] - (g / ℓ) * np.sin(θ[n]) * Δt\n ω[n + 1] = ω[n] - (g / ℓ) * np.sin(θ[n]) * Δt\n θ[n + 1] = θ[n] + ω[n + 1] * Δt\n\ndef nonlinearθ(ℓ, θ0, t):\n '''Special function for non-linear pendulum'''\n from scipy import special\n k = np.sin(θ0 / 2)\n K = special.ellipk(k*k)\n (sn, cn, dn, ph) = special.ellipj(K - np.sqrt(g/l) * t, k * k)\n return 2 * np.arcsin(k * sn)\n\n# Small angle solution\n#plt.plot(t, θ[0]*np.cos(np.sqrt(g/ℓ)*t), label='Small angle solution')\nplt.plot(t, nonlinearθ(ℓ, θ[0], t), label = \"Exact\")\n\n# the Euler method\nplt.plot(t,θ, label='Euler method')\n\nplt.legend(loc='lower left')\n\nplt.xlabel('Time [s]')\nplt.ylabel('θ(t) [rad]')\n\nplt.show()\n","sub_path":"InClass/10-14-16.py","file_name":"10-14-16.py","file_ext":"py","file_size_in_byte":1194,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"79947433","text":"from sensor.co2 import mhz19b\nimport argparse\nimport subprocess\n\n\nclass Mhz19bCtrl:\n SERIAL_DEV = '/dev/ttyS0'\n SERIAL_START = 'sudo systemctl start serial-getty@ttyS0.service'\n SERIAL_STOP = 'sudo systemctl stop serial-getty@ttyS0.service'\n\n def __init__(self):\n self.p = subprocess.call(self.SERIAL_STOP, stdout=subprocess.PIPE, shell=True)\n self.sensor = mhz19b.Mhz19b(self.SERIAL_DEV)\n\n def __del__(self):\n self.p = subprocess.call(self.SERIAL_START, stdout=subprocess.PIPE, shell=True)\n\n def calibration(self, status):\n if status == 'on':\n return self.sensor.ABC_logic_ON()\n elif status == 'off':\n return self.sensor.ABC_logic_OFF()\n elif status == 'zero':\n return self.sensor.zero_calibration()\n\n def read_co2(self):\n return self.sensor.read_co2()\n","sub_path":"homeiot/sensor/co2/mhz19bCtrl.py","file_name":"mhz19bCtrl.py","file_ext":"py","file_size_in_byte":797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"182701663","text":"# -*- coding: utf-8 -*-\n\nimport os\nimport time\nimport datetime\nimport json\nimport linecache\nimport collections\nimport markdown2html\n\nimport sys\nreload(sys) \nsys.setdefaultencoding('utf8')\n\nauthor='Jet Wang'\njson_file=r''\nmarkdown_file=r''\n\ndef getFileModifyTime(filename): \n '''\n time.localtime(os.stat(file).st_ctime) #判断文件的创建时间\n time.localtime(os.stat(file).st_mtime) #判断文件的最后修改时间\n '''\n filemt = time.localtime(os.stat(filename).st_mtime) #判断文件的最后修改时间\n filetime = datetime.datetime(filemt[0] , filemt[1] , filemt[2], filemt[3], filemt[4], filemt[5])\n return filetime\n\ndef convertDateFormat(dt):\n date_str = dt.strftime(\"%Y-%m-%d %H:%M:%S\")\n return date_str[:10]+'T'+date_str[11:]+'.000Z'\n \ndef buildJsonFromMarkdownFile(filename, fileid):\n '''\n jekyll博客markdown文件的格式例子\n \n ---\n layout: post\n title: \"文章标题\"\n description: \"\"\n category: 技术\n tags: [Linux]\n ---\n 正文部分\n \n '''\n \n posts = collections.OrderedDict()\n \n posts['id'] = str(fileid)\n posts['uuid'] = str(uuidGen())\n \n #获取post中文标题,去掉引号\n posts['title'] = linecache.getline(filename, 3)[len('title:'):].strip()[1:-1]\n #获取post链接地址,根据markdown文件名拼接,去掉.md后缀\n posts['slug'] = filename.split('\\\\')[-1][:-3]\n #获取post正文内容\n text = []\n lines = open(filename, 'r').readlines()[9:]\n for line in lines:\n# text.append(line.strip('\\n').decode('utf-8'))\n text.append(line)\n post_content = ''.join(text)\n posts['markdown'] = post_content\n \n posts['mobiledoc'] = 'null'\n posts['html'] = markdown2html.convertMarkdownText2HtmlText(filename)\n posts['image'] = 'null'\n posts['featured'] = '0'\n posts['page'] = '0'\n posts['status'] = 'published'\n posts['language'] = 'en_US'\n posts['visibility'] = 'public'\n posts['meta_title'] = posts['title']\n posts['meta_description'] = 'null'\n posts['author_id'] = 1\n posts['created_at'] = convertDateFormat(getFileModifyTime(filename))\n posts['created_by'] = 1\n posts['updated_at'] = convertDateFormat(getFileModifyTime(filename))\n posts['updated_by'] = 1\n posts['published_at'] = convertDateFormat(getFileModifyTime(filename))\n posts['published_by'] = 1\n\n return posts\n \ndef uuidGen():\n import uuid\n return uuid.uuid1()\n\ndef dict2JsonFile(dict, filename):\n j = json.dumps(dict,ensure_ascii=False,indent=4)\n f = open(filename, 'w+')\n print >> f, j\n f.close()\n \ndef getContentFromJsonFile(filename):\n '''\n Ghost博客导出的post格式\n \n \"posts\": [{\n \"id\": 2,\n \"uuid\": \"4150734c-4fd6-437c-9a05-d4d28d3bb986\",\n \"title\": \"Hello Ghost\",\n \"slug\": \"hello-ghost\",\n \"markdown\": \"This is the begining\",\n \"mobiledoc\": null,\n \"html\": \"

    This is the begining

    \",\n \"image\": null,\n \"featured\": 0,\n \"page\": 0,\n \"status\": \"published\",\n \"language\": \"en_US\",\n \"visibility\": \"public\",\n \"meta_title\": null,\n \"meta_description\": null,\n \"author_id\": 1,\n \"created_at\": \"2016-07-06T08:30:55.000Z\",\n \"created_by\": 1,\n \"updated_at\": \"2016-07-06T08:31:28.000Z\",\n \"updated_by\": 1,\n \"published_at\": \"2016-07-06T08:31:28.000Z\",\n \"published_by\": 1\n },\n '''\n \n json_post = {}\n \n with open(filename) as json_data:\n d = json.load(json_data)\n for d1 in d['db']:\n for d2 in d1['data']['posts']:\n id = d2['id']\n uuid = d2['uuid']\n title = d2['title']\n slug = d2['slug']\n markdown = d2['markdown']\n mobiledoc = d2['mobiledoc']\n html = d2['html']\n image = d2['image']\n featured = d2['featured']\n page = d2['page']\n status = d2['status']\n language = d2['language']\n visibility = d2['visibility']\n meta_title = d2['meta_title']\n meta_description = d2['meta_description']\n author_id = d2['author_id']\n created_at = d2['created_at']\n created_by = d2['created_by']\n updated_at = d2['updated_at']\n updated_by = d2['updated_by']\n published_at = d2['published_at']\n published_by = d2['published_by']\n\n json_post[uuid] = [id,title,slug,markdown,mobiledoc,html,image,featured,\n page,status,language,visibility,meta_title,meta_description,\n author_id,created_at,created_by,updated_at,updated_by,published_at,\n published_by]\n \n return json_post\n\ndef setContentToJsonFile():\n '''\n 提取jekyll博客markdown文件的内容,按照Ghost博客的格式构造出可以供导入的json文件\n '''\n i = 3\n targetFile = r'C:\\temp\\ghost\\2016-08-15-all-post-contents.json'\n folder = r'D:\\code\\github\\hexo\\myhexoblog\\source\\_posts'\n posts = []\n for dirpath, dirnames, filenames in os.walk(folder):\n for filename in filenames:\n i = i + 1\n filename = os.path.join(dirpath, filename)\n post = buildJsonFromMarkdownFile(filename, i)\n posts.append(post)\n \n dict2JsonFile(posts, targetFile)\n \nif __name__ == '__main__':\n setContentToJsonFile()\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":5782,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"35698425","text":"# -*- coding: utf-8 -*-\n\nfrom odoo import models\nfrom odoo.tools.translate import _\nfrom odoo.tools.misc import formatLang, format_date\n\nimport logging\n_logger = logging.getLogger(__name__)\n\n\nLINE_FILLER = '*'\nINV_LINES_PER_STUB = 9\n\nclass report_print_check(models.Model):\n _inherit = 'account.payment'\n\n def make_stub_line(self, invoice):\n result = super(report_print_check,self).make_stub_line(invoice)\n discount = 0.0\n for line in invoice.invoice_line_ids:\n discount += line.price_unit *(line.discount or 0.00) / 100.0\n result['discount'] = formatLang(self.env, discount, currency_obj=invoice.currency_id)\n _logger.info('========dis==%r',discount)\n return result","sub_path":"prime_doors_check_printing_update/models/print_check.py","file_name":"print_check.py","file_ext":"py","file_size_in_byte":725,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"581267285","text":"# determine how long it will take you to save enough money to make housing down payment\n\n# annual salary\nannual_salary = int(input(\"Enter your annual salary: \"))\n# dedicate a certain amount of your salary each month to saving for the down payment\nportion_saved = float(input(\"Enter the percent of your salary to save, as a decimal: \"))\n# cost of dream house\ntotal_cost = int(input(\"Enter the cost of your dream house: \"))\n\n\ndef calc():\n # portion of cost needed for down payment\n portion_down_payment = .25\n # monthly return on investment rate\n r = 0.04 / 12\n # down payment for dream house\n down_payment = portion_down_payment * total_cost\n monthly_savings = annual_salary / 12 * portion_saved\n current_savings = 0\n return_on_invest = 0\n month_count = 0\n\n while current_savings != down_payment:\n month_count += 1\n return_on_invest = current_savings * r\n current_savings = current_savings + monthly_savings + return_on_invest\n # print(month_count, current_savings, monthly_savings, return_on_invest)\n\n if current_savings > down_payment:\n print(\"Number of months: {}\".format(month_count))\n break\n\n\ncalc()\n","sub_path":"MIT_Python/ps1a.py","file_name":"ps1a.py","file_ext":"py","file_size_in_byte":1194,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"349618597","text":"# Copyright (c) 2001-2016, Canal TP and/or its affiliates. All rights reserved.\n#\n# This file is part of Navitia,\n# the software to build cool stuff with public transport.\n#\n# Hope you'll enjoy and contribute to this project,\n# powered by Canal TP (www.canaltp.fr).\n# Help us simplify mobility and open public transport:\n# a non ending quest to the responsive locomotion way of traveling!\n#\n# LICENCE: This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n# Stay tuned using\n# twitter @navitia\n# IRC #navitia on freenode\n# https://groups.google.com/d/forum/navitia\n# www.navitia.io\nfrom flask_restful import abort\nimport flask_restful\nfrom pymongo.errors import PyMongoError\nfrom tartare import app\nfrom tartare.core import models\nimport logging\nfrom tartare.interfaces import schema\nfrom marshmallow import ValidationError\nfrom flask import request\n\n\nclass Coverage(flask_restful.Resource):\n def post(self):\n coverage_schema = schema.CoverageSchema(strict=True)\n try:\n coverage = coverage_schema.load(request.json).data\n except ValidationError as err:\n return {'error': err.messages}, 400\n\n try:\n coverage.save()\n except PyMongoError as e:\n logging.getLogger(__name__).exception('impossible to add coverage {}'.format(coverage))\n return {'error': str(e)}, 500\n\n return {'coverage': coverage_schema.dump(coverage).data}, 201\n\n def get(self, coverage_id=None):\n if coverage_id:\n c = models.Coverage.get(coverage_id)\n if c is None:\n abort(404)\n\n result = schema.CoverageSchema().dump(c)\n return {'coverage': result.data}, 200\n\n coverages = models.Coverage.all()\n\n return {'coverages': schema.CoverageSchema(many=True).dump(coverages).data}, 200\n\n def delete(self, coverage_id):\n c = models.Coverage.delete(coverage_id)\n if c == 0:\n abort(404)\n return {'coverage': None}, 204\n\n def patch(self, coverage_id):\n coverage = models.Coverage.get(coverage_id)\n if coverage is None:\n abort(404)\n if 'id' in request.json and coverage.id != request.json['id']:\n return {'error': 'The modification of the id is not possible'}, 400\n coverage_schema = schema.CoverageSchema(partial=True)\n errors = coverage_schema.validate(request.json, partial=True)\n if errors:\n return {'error': errors}, 400\n\n logging.debug(request.json)\n try:\n coverage = models.Coverage.update(coverage_id, request.json)\n except PyMongoError as e:\n logging.getLogger(__name__).exception('impossible to update coverage with dataset {}'.format(request.json))\n return {'error': str(e)}, 500\n\n return {'coverage': schema.CoverageSchema().dump(coverage).data}, 200\n","sub_path":"tartare/interfaces/coverages.py","file_name":"coverages.py","file_ext":"py","file_size_in_byte":3483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"586633630","text":"from __future__ import division, print_function, absolute_import\n\nimport codecs\nimport json\nimport logging\nimport os\nimport urllib\n\nimport numpy as np\nimport pandas as pd\nimport requests\n\n\nfrom config import pathDict, api_call\n\nlogging.basicConfig(level=logging.DEBUG, filename=\"logfile.log\", filemode=\"w\",\n format=\"%(asctime)-15s %(levelname)-8s %(message)s\")\n\n\n\nmap_size = [400,400]\nzoom_lvl = 40\nmetaURL_head = 'https://maps.googleapis.com/maps/api/geocode/json?address='\naerialURL_head = 'https://maps.googleapis.com/maps/api/staticmap?center='\nmetaURL_tail = '&key=%s'%(api_call['google_meta_key'])\naerialURL_tail = '&maptype=satellite&key=%s'%(api_call['google_aerial_key'])\nreader = codecs.getreader(\"utf-8\")\n\n# Statistic and Image Dump paths\ngoogle_aaerial_stats_path = pathDict['google_aerial_stats_path']\ngoogle_house_dump_path = os.path.join(pathDict['google_aerial_image_path'], 'house')\ngoogle_land_dump_path = os.path.join(pathDict['google_aerial_image_path'], 'land')\ngoogle_unknown_dump_path = os.path.join(pathDict['google_aerial_image_path'], 'unknown')\n\nfor dir in [google_aaerial_stats_path, google_house_dump_path, google_land_dump_path, google_unknown_dump_path]:\n if not os.path.exists(dir):\n os.makedirs(dir)\n\n\ndef metadata_prep(metadata):\n metadata.columns = ['row_id', 'removed', 'property_id', 'state', 'county_name', 'pin',\n 'address_line1', 'address_line2', 'address_city', 'address_zip',\n 'zoning', 'improvement_level', 'type', 'exterior',\n 'last_reviewed_timestamp', 'gone_timestamp', 'indicator',\n 'assessor_photo']\n\n metadata['state'] = metadata['state'].astype('str')\n metadata['county_name'] = metadata['county_name'].astype('str')\n metadata['pin'] = metadata['pin'].astype('str')\n metadata['address_line1'] = metadata['address_line1'].astype('str')\n metadata['address_line2'] = metadata['address_line2'].astype('str')\n metadata['address_city'] = metadata['address_city'].astype('str')\n metadata['address_zip'] = metadata['address_zip'].astype('str')\n metadata['zoning'] = metadata['zoning'].astype('str')\n metadata['improvement_level'] = metadata['improvement_level'].astype('str')\n metadata['type'] = metadata['type'].astype('str')\n metadata['exterior'] = metadata['exterior'].astype('str')\n metadata['last_reviewed_timestamp'] = metadata['last_reviewed_timestamp'].astype('str') # .astype('str')\n metadata['gone_timestamp'] = metadata['gone_timestamp'].astype('str')\n metadata['indicator'] = metadata['indicator'].astype('str')\n metadata['assessor_photo'] = metadata['assessor_photo'].astype('str')\n \n return metadata\n\n\n\nclass GoogleFetch_AerialMap():\n def __init__(self, params):\n pass\n \n @staticmethod\n def get_latlon_locationtype(address_line, city=None, state=None):\n '''\n :param address_line : '555E, 33rd Place'\n :param city: 'chicago'\n :param state: 'IL'\n :return:\n lat: latitude of the property\n lon: longitude of the property\n location_type = ROOFTOP\n url : The URL used to fetch the meta data information\n \n '''\n address_string = '+'.join([add for add in address_line.split(' ')])\n if city:\n address_string = address_string + '+' + '+'.join([add for add in city.split(' ')])\n if state:\n address_string = address_string + '+' + state\n \n # print (address_string)\n url = metaURL_head + address_string + metaURL_tail\n r = urllib.request.urlopen(url)\n res_body = r.read()\n content = json.loads(res_body.decode(\"utf-8\"))\n\n if content['status'] == 'OK':\n # try:\n lat = content['results'][0]['geometry']['location']['lat']\n lon = content['results'][0]['geometry']['location']['lng']\n location_type = content['results'][0]['geometry']['location_type']\n return lat, lon, location_type, url\n elif content['status'] == 'OVER_QUERY_LIMIT':\n return 'EXCEED', 'EXCEED', 'EXCEED', 'EXCEED'\n else:\n # except KeyError:\n logging.info('GET_LATLON: Content lat lon not found')\n return None, None, None, None\n \n @staticmethod\n def get_aerial_image_given_latlon(lat, lon, zoom=19, map_size='400x400'):\n '''\n :param lat: The input latitude\n :param lon: The input Longitude\n :param zoom: The input zoom level\n :param map_size: The input mapSize\n :return:\n : image_data : The image to be saved\n : location_url: The url used to fetch the image\n '''\n location_url = aerialURL_head + str(lat) + ' ' + str(lon) + '&zoom=' + str(zoom) + '&size=' + map_size + \\\n aerialURL_tail\n try:\n img_data = requests.get(location_url).content\n return img_data, location_url\n\n except:\n logging.info('GET_AERIAL_IMAGE: Response error')\n return None, None, None\n\n\n\ndef fetch_google_aerial_images(dataIN, batch_size, get_stats=False):\n data_arr = np.array(dataIN[['pin', 'address_line1', 'address_city', 'indicator']], dtype='str')\n\n statistics = []\n prev = 0\n state = 'IL'\n zoom = 19\n map_size = '400x400'\n for num, (pin, add1, city, indicator) in enumerate(data_arr):\n # if num\n lat = 'nan'\n lon = 'nan'\n meta_url = 'nan'\n img_url = 'nan'\n location_type = 'nan'\n if str(add1) != 'nan':\n\n lat, lon, location_type, meta_url = GoogleFetch_AerialMap.get_latlon_locationtype(address_line=add1,\n city=city, state=state)\n if lat ==None or lon == None or meta_url == None:\n lat = 'nan'\n lon = 'nan'\n meta_url = 'nan'\n elif lat == 'EXCEED':\n logging.info('Total extraction quota for today EXCEEDS the Free Quota LIMIT')\n else:\n image_data, img_url = GoogleFetch_AerialMap.get_aerial_image_given_latlon(lat=lat, lon=lon,\n zoom=zoom, map_size=map_size)\n \n if indicator == \"Likely House\":\n with open(os.path.join(google_house_dump_path, '%s.jpg' % str(pin)), 'wb') as handler:\n handler.write(image_data)\n elif indicator == 'Likely Land':\n with open(os.path.join(google_land_dump_path, '%s.jpg' % str(pin)),'wb') as handler:\n handler.write(image_data)\n else:\n with open(os.path.join(google_unknown_dump_path, '%s.jpg' % str(pin)), 'wb') as handler:\n handler.write(image_data)\n \n b = \"TOTAL RECORDS PARSED: IMAGES DONE ======== %s\"\n print(b % (num), end=\"\\r\")\n \n if get_stats:\n statistics.append([pin, add1, city, lat, lon, location_type, indicator, meta_url, img_url, ])\n \n if ((num + 1) % batch_size) == 0 or num == len(data_arr) - 1:\n if get_stats:\n file_path = os.path.join(google_aaerial_stats_path, '%s_%s.csv' % (prev, num))\n statistics = pd.DataFrame(statistics,\n columns=['pin', 'address', 'city', 'lat','lon','loc_type',\n 'indicator','meta_url','img_url'])\n statistics.to_csv(file_path, index=None)\n prev = num + 1\n \n statistics = []\n \n \n \n\n\n\ndebugg1 = False\nif debugg1:\n lat, lon, location_type, url = GoogleFetch_AerialMap.get_latlon_locationtype(address_line='555E 33rd place',\n city='chicago', state='IL')\n img, location_url = GoogleFetch_AerialMap.get_aerial_image_given_latlon(lat=lat, lon=lon, zoom=19,\n map_size='400x400')\n\ndebugg = False\nif debugg:\n input_path = os.path.join(pathDict['parent_path'], 'house_metadata_nw.csv')\n print (input_path)\n metadata = pd.read_csv(input_path)\n logging.info('Metadata shape: %s', str(metadata.shape))\n metadata = metadata_prep(metadata)\n \n # Remove Test Data set\n metadata = metadata[metadata['removed'] == 0]\n logging.info('Metadata after removing test data set shape: %s', str(metadata.shape))\n\n # Remove data where the last_reviewed_timestamp column doesn't have a valid timestamp\n metadata = metadata[metadata['last_reviewed_timestamp'] != 'nan']\n logging.info('Metadata after retaining last_reviewed_timestamp: %s', str(metadata.shape))\n logging.info('Metadata Head: \\n %s', str(metadata.head()))\n\n metadata = pd.concat([metadata[metadata['indicator'] == 'Likely Land'].head(50),\n metadata[metadata['indicator'] == 'Likely House'].head(50)])\n\n fetch_google_aerial_images(metadata, batch_size=50, get_stats=True)\n","sub_path":"external_data/google_aerial.py","file_name":"google_aerial.py","file_ext":"py","file_size_in_byte":9262,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"253678339","text":"\"\"\"\n------------------------------------------------------------------------\n_THREAD.PY\n\nAUTHOR(S): Peter Walker pwalker@csumb.edu\n\nPURPOSE- This object will represent a single thread, or pipe, in a network\n speed test. It will hold an array of Measurement objects, and has\n some basic object information\n------------------------------------------------------------------------\n\"\"\"\nif __name__==\"__main__\":\n raise SystemExit\n\n# IMPORTS\nfrom __Base import Formatting\nfrom _Measurement import Measurement as Msmt\nfrom _Measurement import Final_Measurement as FMsmt\n#END IMPORTS\n\n\nclass Thread(Formatting):\n\n \"\"\"\n An abstract Thread class that takes care of most parsing\n\n ATTRIBUTES\n ThreadNumber Integer, the number identifier associated with this thread\n DataDirection String, the direction of data travel ('UP' or 'DOWN')\n LocalIP String, the IP address of the device conducting the test\n LocalPort Integer, the port this test is using\n ServerIP String, the IP address of the server this device is connected to\n ServerPort Integer, the port this device is sending information to\n Measurements List of Measurement objects\n FinalMsmt Measurement object, the final summation measurement\n \"\"\"\n\n def __init__(self, dataArr=None, threadNum=0, direction=\"UP\", units=(\"KBytes\", \"Kbits/sec\")):\n \"\"\"\n Used to initialize an object of this class\n ARGS:\n dataArr List of Strings, each String is a measurement that will be parsed and stored in this object\n ThreadNum Integer, the number that this thread is (generally between 3 and 6)\n direction String, the direction of the data in this thread (UP or DOWN)\n units Tuple of two Strings, the units being used by the measurements\n \"\"\"\n #Setting up the whitespace padding that this class will need\n Formatting.__init__(self)\n self.StringPadding = self.StringPadding*3\n #Class variables\n self.Measurements = []\n self.ThreadNumber = threadNum\n self.DataDirection = direction\n #This function assumes that the array of strings (dataArr) is not in order\n #This takes the given data String and parses the object information\n for line in dataArr:\n if \"connected with\" in line:\n line = line.split(\"local\", 1)[1].strip()\n self.LocalIP = line.split(\"port\")[0].strip()\n line = line.split(\"port\", 1)[1].strip()\n self.LocalPort = line.split(\"connected\")[0].strip()\n line = line.split(\"connected with\", 1)[1].strip()\n self.ServerIP = line.split(\"port\", 1)[0].strip()\n line = line.split(\"port\", 1)[1].strip()\n self.ServerPort = line.split(\"\\n\")[0].strip()\n break\n #END IF\n #END FOR\n #Removing the line from the array of pings that contains the connection info\n # and then creating all of the pings from the remaining strings\n allMeasurements = [dataLine for dataLine in dataArr\n if \"connected with\" not in dataLine]\n self.FinalMsmt = None\n for line in allMeasurements:\n #We do a quick check for the string stored in units[1]. If that string is\n # present in a line, then it must be a measurement that we want to parse\n if (units[1] in line) and (\"%\" not in line):\n #Make a measurement object out of the line.\n newMsmt = Msmt(data=line, units=units)\n #If the measurement's start time is one second behind it's end time, then\n # we can assume that this is one of interval measurements. Otherwise, it is\n # the final summary measurement, and we put the object in self.FinalMsmt\n if (newMsmt.TimeStart == newMsmt.TimeEnd-1):\n #This is for the UDP 1 second tests, where this only 1 regular\n # measurement, and then a final measurement. We delete the old object,\n # and create a new one of type Final_Measurement\n if (newMsmt.TimeStart == 0) and (len(self.Measurements) == 1):\n FinalMsmt = FMsmt(data=line, units=units)\n self.FinalMsmt = FinalMsmt\n break\n else:\n self.Measurements.insert(int(newMsmt.TimeStart), newMsmt)\n #END IF/ELSE\n else:\n FinalMsmt = FMsmt(data=line, units=units)\n self.FinalMsmt = FinalMsmt\n #END IF/ELSE\n #END IF\n #END FOR\n #END DEF\n\n def arrayOfMsmts(self, attribute=\"Speed\"):\n \"\"\"\n Will return an array of the Measurements in self.Measurements as an array\n of Numbers. Can be given the attribute of the measurement that needs to be arraytized\n ARGS:\n attribute String, can be \"speed\" or \"size\" (attribute of Measurment)\n RETURNS:\n list, containing all of the values in the Measurement object of myMeasurements\n \"\"\"\n if attribute not in [\"Speed\", \"Size\"]:\n print(\"The attribute specified must be either 'Speed' or 'Size'. Using 'Speed'\")\n attribute = \"Speed\"\n #END IF\n #This uses list comprehension to return a list of all of the measurement's attribute,\n # specified by the variable attribute\n return [msmt.__dict__[attribute] for msmt in self.Measurements]\n #END DEF\n\n\n# STRING PRINTOUT --------------------------------------------------------------\n\n def __str__(self):\n \"\"\"Returns a string representation of the object\"\"\"\n string = (self.StringPadding +\n \"Thread Number: {}\\n\".format(self.ThreadNumber) +\n self.StringPadding +\n \"Data Direction: {}\\n\".format(self.DataDirection) +\n self.StringPadding +\n \"Local: {}:{}\\n\".format(self.LocalIP,self.LocalPort) +\n self.StringPadding +\n \"Server: {}:{}\\n\".format(self.ServerIP,self.ServerPort)\n )\n for msmt in self.Measurements:\n string += str(msmt) + \"\\n\"\n string += str(self.FinalMsmt) + \"\\n\"\n return string\n #END DEF\n#END CLASS\n","sub_path":"PyFiles/FileParser/_Thread.py","file_name":"_Thread.py","file_ext":"py","file_size_in_byte":6525,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"362270007","text":"from django.db import transaction\nfrom django.utils import timezone\nfrom rest_framework import viewsets, permissions, status\nfrom rest_framework.decorators import action\nfrom rest_framework.response import Response\nfrom island.exceptions import PermissionDenied, ServiceUnavailable, APIException\nfrom twitter_image.models import Task, TweetData, ImageData, TaskTweet\nfrom twitter_image.serializers import TaskSerializer, TaskTweetSerializer, ImageDataSerializer\nfrom aiohttp.client_exceptions import ClientError\nfrom asyncio import TimeoutError\nimport logging\n\nlogger = logging.getLogger('twitter_image')\n\n\nclass TaskViewSet(viewsets.ReadOnlyModelViewSet):\n serializer_class = TaskSerializer\n permission_classes = [permissions.IsAuthenticated]\n filterset_fields = ['username', 'tag']\n search_fields = ['tag']\n ordering_fields = ['last_update']\n ordering = ['-last_update']\n\n def get_queryset(self):\n return Task.objects.filter(owner=self.request.user)\n\n @action(detail=True, methods=['patch'])\n def refresh(self, request, pk=None):\n try:\n with transaction.atomic():\n task = Task.objects.select_for_update(skip_locked=True).get(pk=pk)\n if task.owner != request.user:\n raise PermissionDenied()\n with timezone.override(None):\n task.update()\n task.last_update = timezone.now()\n task.save()\n serializer = TaskSerializer(task, context={'request': request})\n return Response(serializer.data, status=status.HTTP_201_CREATED)\n except (ClientError, TimeoutError) as e:\n logger.error('[Proxy Down]', exc_info=e)\n raise ServiceUnavailable()\n except BaseException as e:\n logger.fatal('[Unknow Error]', exc_info=e)\n raise APIException()\n\n\nclass TaskTweetViewSet(viewsets.ReadOnlyModelViewSet):\n serializer_class = TaskTweetSerializer\n permission_classes = [permissions.IsAuthenticated]\n filterset_fields = ['new', 'task']\n search_fields = ['tweet__tweet']\n ordering_fields = ['tweet__time']\n ordering = ['-tweet__time']\n\n def get_queryset(self):\n return TaskTweet.objects.filter(task__owner=self.request.user)\n","sub_path":"twitter_image/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2273,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"199323041","text":"# -*- coding: utf-8 -*-\nimport os\nimport sys\nimport traceback\nimport json\nfrom pprint import pprint\n\nfrom django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.forms.formsets import formset_factory\n\nfrom kerotan.forms import AddressForm\n\nfrom .ekitan.ekitan_api import Ekitan\nfrom .gmaps_geocoder.gmaps_geocoder import GoogleMapsGeocoder\nfrom .bing_news.bing_api import Bing\nfrom .wikipedia.wikipedia_api import get_overview\n\nAPI_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)), u'..', u'..', u'API')\nsys.path.append(API_DIR)\nfrom get_API_key import get_API_key\n\n\ndef display_google_map(request):\n if request.method == 'POST':\n form = AddressForm(request.POST)\n company_address = {\n \"TIS\": \"東京都新宿区 西新宿8丁目17−1\",\n \"幕張メッセ\": \"〒261-0023 千葉市美浜区中瀬2-1\",\n \"新宿駅\": \"〒160-0023 東京都新宿区西新宿1丁目1−3\",\n \"東京駅\": \"〒100-0005 東京都千代田区丸の内1 丁目 呑んき1丸の内北口店 9\",\n \"大阪駅\": \"大阪府大阪市北区梅田3丁目1−1\",\n \"立命館大学\": \"〒525-8577 滋賀県草津市野路東1丁目1-1\",\n }\n company_overview = {\n \"TIS\": \"TIS株式会社(ティーアイエス)は国内大手システムインテグレーター。傘下にインテック、アグレックス、クオリカ、AJSなどを擁するTISインテックグループの中核企業である。\",\n \"幕張メッセ\": \"幕張メッセ(まくはりメッセ、Makuhari Messe)は、千葉県千葉市美浜区にある大型の会議・展示施設である。また、株式会社幕張メッセは、これを運営する企業である。\",\n \"新宿駅\": \"新宿駅(しんじゅくえき)は、東京都新宿区・渋谷区にある、東日本旅客鉄道(JR東日本)・京王電鉄・小田急電鉄・東京地下鉄(東京メトロ)・東京都交通局(都営地下鉄)の駅である。\",\n \"東京駅\": \"東京駅(とうきょうえき)は、東京都千代田区丸の内一丁目にある、東日本旅客鉄道(JR東日本)・東海旅客鉄道(JR東海)・東京地下鉄(東京メトロ)の駅である。\",\n \"大阪駅\": \"大阪駅(おおさかえき)は、大阪府大阪市北区梅田三丁目にある、西日本旅客鉄道(JR西日本)の駅である。\",\n \"立命館大学\": \"立命館大学(りつめいかんだいがく、英語: Ritsumeikan University)は、京都府京都市中京区西ノ京朱雀町1に本部を置く日本の私立大学である。1922年に設置された。大学の略称は立命、立命館、立命大。近畿圏では立大も使用される[1]。\"\n }\n# \t\tcompany_overview={\n# \t\t\t\"TIS\":\"TIS株式会社(初代)は、1971年(昭和46年)4月、三和銀行および三和グループを中心に大阪市東区(現在の大阪市中央区)に\\\n# 株式会社東洋情報システム(資本金6億円)として設立された。現在の法人は、2016年(平成28年)7月にITホールディングス株式会社がTIS株式会社(初代)を\\\n# 吸収合併し、商号変更したものである。三和グループに属する企業で構成されるみどり会の会員企業でもある。\\\n# JCBを中心としたクレジットカード会社の基幹システムなどに強みを持ち、国内シェアは5割弱で首位。\",\n# \t\t\t\"幕張メッセ\":\"幕張メッセのメッセとは、ドイツ語の'見本市'の意味を指す'Messe'に由来する。1989年(平成元年)10月9日に開業した。\\\n# 現在は、東京国際展示場(東京ビッグサイト、東西合計)に次ぐ国内2番目の規模となっている。運営会社は、1986年(昭和61年)4月30日に設立され、\\\n# 2005年(平成17年)7月1日に社名を株式会社日本コンベンションセンター(Nippon Convention Center)から株式会社幕張メッセに変更した。\\\n# 当館の整備費は増設分を合わせると約558億円を投じており、年間の維持管理費が約20億円掛かるとされており、千葉県と千葉市の公費による一部負担が続いている。\\\n# しかし、都心から近い東京国際展示場開業の影響を受け、2011年(平成23年)に東京モーターショーが東京ビッグサイトへ移るなど利用は伸び悩み、\\\n# 2013年(平成23年)度の稼働率は約40%となり、横浜市のパシフィコ横浜の稼働率の約70%を大きく下回っている。\",\n# \t\t\t\"新宿駅\":\"東京の副都心・新宿に位置するターミナル駅である。1885年(明治18年)に日本鉄道により現在の山手線が開業したのが当駅の始まりである。\\\n# 4年後の1889年(明治22年)には南豊島郡淀橋町となる。開業時から新宿を副都心にする計画が発��されるまでは当駅周辺はまだ街の外れであり利用客は少ないものだったが、\\\n# 大正期に入り次第に市街地が拡大するにつれ、多くの私鉄が乗り入れるようになる。ターミナルとなって周辺が発展するにつれて利用客は増え続け、1931年には私鉄や国鉄などを合わせた\\\n# 利用者数で日本一になった[1]。そして、1966年(昭和41年)の乗車人数では、国鉄池袋駅の41万67人を抜いて、当駅が41万69人と日本一になっている。\\\n# さらに1960年代から当駅西側一帯で進められた新宿副都心計画によって、70年代には多くの超高層ビルが建てられ利用者の増加に拍車がかかった。\\\n# 現在ではJR・私鉄・地下鉄の多くの路線が周辺地域のベッドタウンとを結んでおり、多くのビジネス客が利用する。さらに、当駅周辺は日本最大の繁華街・歓楽街となっており、\\\n# 昼夜を問わず人の流れが絶えない。JRの駅を中心に東・西・南口、周辺の各地下鉄駅、商業施設などが通路や地下街などで広範囲に連絡している。\\\n# 一日平均乗降者数は約335万人(2013年)[3]と世界一(ギネス世界記録認定)多い駅であり、地下道などで接続する西武新宿駅まで含めると約358万人(2013年)となり、\\\n# この数字は横浜市の人口に匹敵する。年間の乗降客数に直すと約13億人となりインドの人口をも上回る規模となる。\",\n# \t\t\t\"東京駅\":\"東京の表玄関とも言うべきターミナル駅で、特に東海道新幹線と東北新幹線の起点となっており、全国の新幹線網における最大の拠点となっている。\\\n# また、東海道本線や東北本線など主要幹線の起点駅でもある。当駅から乗り換えなしで実に33都道府県[1]と結んでおり、1日当たりの列車発着本数は約3000本という日本を代表する\\\n# ターミナル駅の一つである。プラットホームの数は日本一多く、在来線が地上5面10線と地下4面8線の合計9面18線、新幹線が地上5面10線、地下鉄は地下1面2線を有しており、\\\n# 面積は東京ドーム約3.6個分に相当する。赤レンガ造りの丸の内口駅舎は辰野金吾らが設計したもので、1914年に竣工、2003年に国の重要文化財に指定されている。\\\n# 「関東の駅百選」認定駅でもある。\",\n# \t\t\t\"大阪駅\":\"大阪府の代表駅(府庁所在地駅)として第1回近畿の駅百選にも選定されている西日本最大の駅。駅長が配置された直営駅であり、\\\n# 管理駅として東海道本線の塚本駅を管轄している。JRの特定都区市内制度における「大阪市内」に属する駅であり、運賃計算の中心駅となる。また、アーバンネットワークの運行の要衝となる駅で、\\\n# 運行系統の軸をなしている。大阪市街の北玄関である梅田に位置し、駅前や駅の東側・南側を中心に繁華街が広がっている。\\\n# 東京・山陽・九州方面への長距離列車については、1964年開業の新大阪駅を発着する東海道・山陽新幹線に地位を譲ったものの、当駅は現在でも北陸方面との特急の始発・終着駅であり、\\\n# 新快速を始めとする京阪神の都市間連絡列車や、北近畿・山陰方面との特急、東京駅発着の寝台特急などの在来線特急も発着している。\\\n# かつては東北・北海道方面に向かう夜行列車も発着していたが、2015年3月のダイヤ改正で寝台特急トワイライトエクスプレスが廃止されたことで東北・北海道方面を行き来する夜行列車は全て消滅した。\\\n# これによって大阪駅を起点、終点とする夜行列車は全て消滅した。ただし、トワイライトエクスプレスは同年5月16日(土)に山陽方面のツアー列車として復活した。\\\n# 貨物列車は北方貨物線および梅田貨物線(いずれも通称)を利用するため大阪駅を通過しない。\"}\n\n if form.is_valid():\n print(\"form.is_valid() ok\")\n # # 出発フォームの入力内容が、辞書に登録されているかどうか確認。\n # try:\n # # 登録されていれば、辞書から住所を取得し、出発住所とする\n # start_company = company_address[form.cleaned_data[\"start_address\"]]\n # except:\n # # 入力内容が、辞書に登録されていなければ、入力内容を出発住所とする。\n # start_company = form.cleaned_data[\"start_address\"]\n\n # # 到着フォームでも同様の処理。\n # try:\n # arriv_company = company_address[form.cleaned_data[\"arriv_address\"]]\n # # 到着の方が辞書に含まれるなら、概要とニュースと取得する。ので、フラグを立てる。\n # FLAG_getOverviewNews = 1\n # except:\n # # 入力内容が、辞書に登録されていなかった。\n # arriv_company = form.cleaned_data[\"arriv_address\"]\n # FLAG_getOverviewNews = 0\n\n # ---------------------------------------------------------------------------------\n # ---------------------------------------------------------------------------------\n # geocode,会社概要、ニュースを取得する。\n try:\n # Get geocode by Google Maps API.\n try:\n print(\"try get_geocode : \", end=\"\")\n # gmg = GoogleMapsGeocoder()\n geocode = {}\n # geocode.update({\"start\": gmg.get_geocode(start_company)[\"location\"]})\n # geocode.update({\"arriv\": gmg.get_geocode(arriv_company)[\"location\"]})\n start_location_latitude = float(form.cleaned_data[\"start_location\"][1:-1].split(\", \")[0])\n start_location_latitude = round(start_location_latitude, 9)\n start_location_longitude = float(form.cleaned_data[\"start_location\"][1:-1].split(\", \")[1])\n start_location_longitude = round(start_location_longitude, 8)\n start_location = {\"lat\": start_location_latitude, \"lng\": start_location_longitude}\n print(form.cleaned_data[\"start_location\"])\n # print(start_location_latitude)\n # print(start_location_longitude)\n print(start_location)\n arriv_location_latitude = float(form.cleaned_data[\"arriv_location\"][1:-1].split(\", \")[0])\n arriv_location_latitude = round(arriv_location_latitude, 9)\n arriv_location_longitude = float(form.cleaned_data[\"arriv_location\"][1:-1].split(\", \")[1])\n arriv_location_longitude = round(arriv_location_longitude, 8)\n arriv_location = {\"lat\": arriv_location_latitude, \"lng\": arriv_location_longitude}\n print(form.cleaned_data[\"arriv_location\"])\n # print(arriv_location_latitude)\n # print(arriv_location_longitude)\n print(arriv_location)\n \n geocode.update({\"start\": start_location})\n geocode.update({\"arriv\": arriv_location})\n\n # geocode.update({\"start\": form.cleaned_data[\"start_location\"]})\n # geocode.update({\"arriv\": form.cleaned_data[\"arriv_location\"]})\n\n print(\"finished.\")\n except Exception as e:\n # 入力された住所から、geocodeを特定できない。\n # 再入力させるために、入力画面に戻す。\n print(\"--------------------------------------------\")\n print(type(e))\n print(e)\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n raise\n except:\n # 謎のエラー発生時\n print(\"--------------------------------------------\")\n print(\"謎Error in GoogleMaps.\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n raise\n\n # ---------------------------------------------------------------------------------\n # 会社概要を取得\n # if FLAG_getOverviewNews == 1:\n try:\n # wikipedia APIから取得\n print(\"try company_overview : \", end=\"\")\n # overview = company_overview[form.cleaned_data[\"arriv_address\"]]\n overview = get_overview(form.cleaned_data[\"arriv_address\"])\n image_company_chart = \"/static/img/TIS_chart.png\"\n image_company_building = \"/static/img/TIS_building.png\"\n print(\"finished.\")\n except:\n # エラー発生時.\n print(\"--------------------------------------------\")\n print(\"Error in 会社概要.\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n raise\n # elif FLAG_getOverviewNews == 0:\n # 概要はとってこない\n # print(\"getting overview was passed.\")\n # overview = \"\"\n # image_company_chart = \"\"\n # image_company_building = \"\"\n\n # ---------------------------------------------------------------------------------\n # ニュースを取得\n try:\n # bing search APIで、関連ニュースをとってくる\n print(\"try bing.web_search : \", end=\"\")\n bing = Bing()\n keys = [\"Title\", \"Url\", \"Source\", \"Description\", \"Date\"]\n query = form.cleaned_data[\"arriv_address\"]\n news = bing.web_search(query, 5, keys)\n print(\"finished.\")\n print(\"news\", json.dumps(news, indent=2))\n except (TypeError, ConnectionAbortedError, ConnectionResetError, MaxRetryError, requests.packages.urllib3.exceptions.MaxRetryError, AttributeError, TransportError, googlemaps.exceptions.TransportError, ConnectionError, requests.exceptions.ConnectionError, NewConnectionError):\n # 一分間に連続してリクエスト送ると、回数制限に引っかかってエラー。たぶん。その場合、空のニュースを返すことにする。\n # raiseはしない。\n print(\"--------------------------------------------\")\n print(\"BingSearchAPI回数制限に引っかかりました。時間を置きましょう。\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n news = []\n except FileNotFoundError:\n # 検索結果のファイル出力時のエラー。\n # raiseはしない。\n print(\"--------------------------------------------\")\n print(\"BingSearchAPIの結果出力時のエラー。\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n news = []\n except:\n # 謎のエラー発生時。\n print(\"--------------------------------------------\")\n print(\"謎Error in ニュース.\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n # raise #Exception補足できないから、もうraiseしない。\n news = []\n\n\n # raiseしたら、Top画面に戻す。\n except:\n print(\"raised.\")\n return render_to_response('kerotan/main_page.html', {'form': form}, RequestContext(request))\n\n\n # ---------------------------------------------------------------------------------\n # ---------------------------------------------------------------------------------\n # Get route infomation by Ekitan API.\n try:\n print(\"try ekitan search : \", end=\"\")\n ekitan = Ekitan()\n _, results_filtered = ekitan.norikae_search(s_ido=geocode[\"start\"][\"lat\"], s_keido=geocode[\"start\"][\"lng\"], t_ido=geocode[\"arriv\"][\"lat\"], t_keido=geocode[\"arriv\"][\"lng\"], )\n print(\"finished.\")\n # print(\"results_filtered\",results_filtered)\n return render_to_response('kerotan/main_page.html', {\n 'form': form, 'route': results_filtered, \\\n 'start_latitude': geocode[\"start\"][\"lat\"], 'start_longitude': geocode[\"start\"][\"lng\"],\\\n 'arriv_latitude': geocode[\"arriv\"][\"lat\"], 'arriv_longitude': geocode[\"arriv\"][\"lng\"],\\\n 'news': news, 'overview': overview, 'image_company_chart': image_company_chart, 'image_company_building': image_company_building\\\n }, RequestContext(request))\n\n except FileNotFoundError:\n # APIキーが記述されたファイルの読み込みエラー\n print(\"--------------------------------------------\")\n print(\"APIキーが記述されたファイルの読み込みエラー\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n # raise\n except:\n print(\"--------------------------------------------\")\n print(traceback.print_exc())\n print(\"--------------------------------------------\")\n return render_to_response('kerotan/main_page.html', {'form': form}, RequestContext(request))\n # return render_to_response('kerotan/main_page.html', {\n # 'form': form, 'route': results_filtered, \\\n # 'start_latitude': geocode[\"start\"][\"lat\"], 'start_longitude': geocode[\"start\"][\"lng\"],\\\n # 'arriv_latitude': geocode[\"arriv\"][\"lat\"], 'arriv_longitude': geocode[\"arriv\"][\"lng\"],\\\n # 'news': news, 'overview': overview, 'image_company_chart': image_company_chart, 'image_company_building': image_company_building\\\n # }, RequestContext(request))\n\n\n # form.is_valid()を満たさない場合\n else:\n return render_to_response('kerotan/main_page.html', {'form': form}, RequestContext(request))\n # return render_to_response('kerotan/main_page.html', {'formset':formset}, RequestContext(request))\n\n # request.methodがPOSTじゃない場合\n else:\n form = AddressForm()\n return render_to_response('kerotan/main_page.html', {'form': form}, RequestContext(request))\n # return renderrender(RequestContext(request), 'kerotan/main_page.html')\n","sub_path":"kerotan/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":20827,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"311544446","text":"from flask import Flask, request, jsonify\nimport json\nimport os\nimport urllib.request\nfrom flask_cors import CORS\nimport pymysql\n\n# Constant For BASE URL\nBASE_URL = \"/api/v1/\"\nAUTH_CODE = \"Sanketnaik@1999\"\n\n# Connect to MySQL database\ndb = pymysql.connect(\"remotemysql.com\", \"a0EjWgNg3d\", \"xgrm2qldV8\", \"a0EjWgNg3d\", port=3306, autocommit=True)\ncursor = db.cursor()\n\n# Initialize Flask App\napp = Flask(__name__)\nCORS(app, resources={r\"/api/v1/*\": {\"origins\": \"*\"}})\n\n#\n# CREATE DATABASE\n#\n@app.route(BASE_URL + 'create_db', methods=['POST'])\ndef create_db():\n data = request.form\n\n authentication = data['auth']\n\n if authentication == AUTH_CODE:\n try:\n cursor.execute(\"CREATE DATABASE IF NOT EXISTS personal_diary;\")\n cursor.execute('USE personal_diary')\n return jsonify({\"result\": \"SUCCESS\"})\n except:\n return jsonify({\"result\": \"ERROR\"})\n else:\n return jsonify({\"result\": \"ERROR\"})\n\n\n#\n# START OF USER ACTIONS\n#\napp.route(BASE_URL + 'add-user', methods=['POST'])\ndef add_user():\n data = request.form\n\n name = data['display_name']\n email = data['email']\n photoURL = data['photoURL']\n uid = data['uid']\n try:\n cursor.execute(\n \"CREATE TABLE IF NOT EXISTS users (ID int NOT NULL AUTO_INCREMENT ,display_name VARCHAR(100), email VARCHAR(200), photoURL VARCHAR(1000), uid VARCHAR(100), PRIMARY KEY (ID));\")\n\n cursor.execute(f'INSERT INTO users (display_name, email, photoURL, uid) VALUES (\"{name}\", \"{email}\", \"{photoURL}\", \"{uid}\");');\n return jsonify({\"result\": \"SUCCESS\"})\n\n except:\n return jsonify({\"result\": \"ERROR\"})\n\n\n@app.route(BASE_URL + 'get-user-data', methods=['POST'])\ndef get_user_data():\n\n data = request.form\n email = data['email']\n\n try:\n cursor.execute(f\"select * from users where `email` = \\\"{email}\\\"\")\n data = cursor.fetchone()\n return jsonify({\"result\": \"SUCCESS\", \"data\": data})\n except:\n return jsonify({\"result\": \"ERROR\"})\n\n#\n# END OF USER ACTIONS\n#\n\n#\n# START OF DIARY ACTIONS\n#\n@app.route(BASE_URL + 'add-entry', methods=['POST'])\ndef add_entry():\n\n formData = request.form\n uid = formData[\"uid\"]\n id = formData[\"id\"]\n data = formData[\"data\"]\n month = formData[\"month\"]\n year = formData[\"year\"]\n date = formData[\"date\"]\n day = formData[\"day\"]\n\n try:\n cursor.execute(f\"CREATE TABLE IF NOT EXISTS {uid}_diary (UID int NOT NULL AUTO_INCREMENT, id VARCHAR(20), date VARCHAR(10), month VARCHAR(20), year VARCHAR(10), day VARCHAR(10), data VARCHAR(2000), PRIMARY KEY (UID));\")\n cursor.execute(f\"insert into {uid}_diary (id, date, month, year, day, data) VALUES (\\\"{id}\\\", \\\"{date}\\\", \\\"{month}\\\", \\\"{year}\\\", \\\"{day}\\\", \\\"{data}\\\");\")\n\n return jsonify({\"result\": \"SUCCESS\"})\n except:\n return jsonify({\"result\": \"ERROR\"})\n\n\n@app.route(BASE_URL + 'get-data', methods=['POST'])\ndef get_diary_data():\n\n data = request.form\n uid = data['uid']\n\n try:\n cursor.execute(f\"select * from {uid}_diary ORDER BY id DESC;\")\n result = cursor.fetchall()\n return jsonify({\"result\": \"SUCCESS\", \"data\": result})\n except:\n return jsonify({\"result\": \"ERROR\"})\n\n\n@app.route(BASE_URL + 'update-entry', methods=['POST'])\ndef update_entry():\n\n formData = request.form\n uid = formData['uid']\n data = formData['data']\n id = formData['id']\n\n try:\n cursor.execute(f'update {uid}_diary SET `data` = \"{data}\" where id = \"{id}\";')\n return jsonify({\"result\": \"SUCCESS\"})\n except:\n return jsonify({\"result\": \"ERROR\"})\n\n#\n# END OF DIARY ACTIONS\n#\n\n# Run Flask App\nif __name__ == \"__main__\":\n app.run(debug=True)\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"143613713","text":"\n# Change log\n# 20200308 Soumya Added feed for space.com\n# 20200308 Soumya Added BeautifulSoup to scrap summary from space.com \n\nfrom django.shortcuts import render\nfrom django.http import JsonResponse,HttpResponse\nimport feedparser\nimport re\nimport requests \nfrom bs4 import BeautifulSoup\n# Import rest_framework\nfrom rest_framework.response import Response\nfrom rest_framework.views import APIView\n\nfrom .serializers import PostSerializer, NASASerializer, SPACEDOTCOMSerializer\nfrom .models import Post, NASA, SPACEDOTCOM\n\n\nclass TestView(APIView):\n def get(self, request, *args, **kwargs):\n qs = Post.objects.all()\n serializer = PostSerializer(qs, many=True)\n return Response(serializer.data)\n\n def post(self, request, *args, **kwargs):\n serializer = PostSerializer(data=request.data)\n if serializer.is_valid():\n serializer.save()\n return Response(serializer.data)\n else:\n return Response(serializer.errors)\n\n\nclass NASAView(APIView):\n def get(self, request, *args, **kwargs):\n feed = feedparser.parse('https://www.nasa.gov/rss/dyn/breaking_news.rss')\n for entry in feed.entries:\n obj, created = NASA.objects.get_or_create(\n link=entry['link'],\n title=entry['title'],\n author=feed['feed']['author'],\n summary=entry['summary'],\n published=entry['published'],\n article_id=entry['dc_identifier'],\n author_img_url='',\n article_img_url=entry['links'][1]['href']\n )\n\n qs = NASA.objects.all()\n serializer = NASASerializer(qs, many=True)\n return Response(serializer.data) \n\n\nclass SPACEDOTCOMView(APIView):\n def get(self, request, *args, **kwargs):\n article_img_url = ''\n feed = feedparser.parse('https://www.space.com/feeds/all')\n for entry in feed.entries:\n for link in entry['links']:\n if ('image' in link.type):\n article_img_url = link.href\n # Space.com doesnt provide the summary content\n # So, use beautiful soup to scrap the article\n # We are fetching the 'complete' sentences within 500char limit\n r = requests.get(entry['link'])\n htmlContent = r.content\n soup = BeautifulSoup(htmlContent, 'html.parser')\n summary = str(soup.find(id=\"article-body\").get_text())[0:500].split(\".\")\n summary.pop()\n summary = \".\".join(summary)\n\n\n obj, created = SPACEDOTCOM.objects.get_or_create(\n link=entry['link'],\n title=entry['title'],\n author=entry['author'],\n summary=summary,\n published=entry['published'],\n published_parsed=entry['published_parsed'],\n article_id='',\n author_img_url='',\n article_img_url=article_img_url\n )\n\n qs = SPACEDOTCOM.objects.all()\n serializer = SPACEDOTCOMSerializer(qs, many=True)\n return Response(serializer.data) \n\n\ndef dummy(request):\n return JsonResponse({\n 'author': 'Soumya',\n 'NASA': '/NASA',\n 'SPACEDOTCOM': '/SPACEDOTCOM'\n }, safe=False)\n","sub_path":"core/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":3312,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"174338359","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy.stats import lognorm\n\nx,y,r = np.loadtxt('dist2', unpack = True)\n\n# Do some stats\nmean = np.average(r)\nstd = np.nanstd(r)\nr = r[np.nonzero(r)]\n\nprint(f'mean: {mean}, std/mean: {std/mean}')\n\n# Create circles\ncircles = (plt.Circle((xi,yi),ri,fill=False) for xi, yi, ri in zip(x, y, r))\n\n# Create figure and axis, then add circles\nfig, ax = plt.subplots()\nfor circle in circles:\n ax.add_patch(circle)\n\n# Neaten up\nplt.axis('scaled')\nplt.xlabel('x (nm)')\nplt.ylabel('y (nm)')\nplt.show()\n","sub_path":"data_analysis/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"526903375","text":"from django.conf import settings\nfrom django.contrib import admin\nfrom django.urls import include, path\n\nurlpatterns = [\n path('', include('apps.landing.urls')),\n path('api/v1/ingredients/', include('apps.ingredients.urls')),\n path('api/v1/recipes/', include('apps.recipes.urls')),\n path('accounts/', include('apps.accounts.urls')),\n path('admin/', admin.site.urls),\n path('api-auth/', include('rest_framework.urls'))\n]\n\nif settings.DEBUG:\n import debug_toolbar\n urlpatterns = [\n path('__debug__/', include(debug_toolbar.urls)),\n ] + urlpatterns","sub_path":"project/core/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"651537675","text":"from .base import *\nfrom .mgr import CoreManager as Mgr\n\n\nclass CreationPhaseManager(object):\n\n _id_generator = id_generator()\n\n def __init__(self, obj_type, has_color=False, add_to_hist=False):\n\n self._obj = None\n self._obj_type = obj_type\n self._has_color = has_color\n self._add_to_hist = add_to_hist\n self._custom_obj_name = \"\"\n\n self._origin_pos = Point3()\n self._creation_handlers = []\n self._current_creation_phase = 0\n\n if has_color:\n self.set_next_object_color()\n else:\n GlobalData[\"next_%s_color\" % obj_type] = None\n\n Mgr.expose(\"custom_%s_name\" % obj_type, lambda: self._custom_obj_name)\n Mgr.accept(\"set_custom_%s_name\" % obj_type, self.__set_custom_object_name)\n\n def setup(self, creation_phases, status_text):\n\n creation_status = {}\n mode_text = \"Create %s\" % status_text[\"obj_type\"]\n info_text = \"LMB-drag to start creation\"\n creation_status[\"idle\"] = {\"mode\": mode_text, \"info\": info_text}\n\n add_state = Mgr.add_state\n bind = Mgr.bind_state\n state_persistence = -12\n\n for i, phase_data in enumerate(creation_phases):\n\n main_starter, main_handler = phase_data\n\n if i == 0:\n creation_starter = self.__get_creation_starter(main_starter)\n Mgr.accept(\"start_%s_creation\" % self._obj_type, creation_starter)\n on_enter_state = None\n else:\n on_enter_state = self.__get_creation_phase_starter(main_starter)\n\n state_id = \"%s_creation_phase_%s\" % (self._obj_type, i + 1)\n add_state(state_id, state_persistence, on_enter_state)\n\n self._creation_handlers.append(self.__get_creation_phase_handler(main_handler))\n\n binding_id = \"quit %s creation\" % self._obj_type\n bind(state_id, binding_id, \"escape\", self.__end_creation)\n binding_id = \"cancel %s creation\" % self._obj_type\n bind(state_id, binding_id, \"mouse3-up\", self.__end_creation)\n\n info_text = \"move mouse to %s;\" % status_text[\"phase%s\" % (i + 1)]\n get_command = lambda state_id: lambda: Mgr.enter_state(state_id)\n\n if i == len(creation_phases) - 1:\n binding_id = \"finalize %s creation\" % self._obj_type\n bind(state_id, binding_id, \"mouse1-up\",\n lambda: self.__end_creation(cancel=False))\n info_text += \" release LMB to finalize;\"\n else:\n binding_id = \"start %s creation phase %s\" % (self._obj_type, i + 2)\n next_state_id = \"%s_creation_phase_%s\" % (self._obj_type, i + 2)\n bind(state_id, binding_id, \"mouse1-up\", get_command(next_state_id))\n info_text += \" release LMB to set;\"\n\n info_text += \" RMB to cancel\"\n creation_status[\"phase%s\" % (i + 1)] = {\"mode\": mode_text, \"info\": info_text}\n\n status_data = GlobalData[\"status_data\"][\"create\"]\n status_data[self._obj_type] = creation_status\n\n return True\n\n def __get_creation_starter(self, main_creation_func):\n\n def start_creation(origin_pos):\n\n self._origin_pos = origin_pos\n main_creation_func()\n\n Mgr.enter_state(\"%s_creation_phase_1\" % self._obj_type)\n Mgr.add_task(self._creation_handlers[0], \"draw_object\", sort=3)\n Mgr.update_app(\"status\", \"create\", self._obj_type, \"phase1\")\n\n return start_creation\n\n def __get_creation_phase_starter(self, main_start_func):\n\n def start_creation_phase(prev_state_id, is_active):\n\n Mgr.remove_task(\"draw_object\")\n main_start_func()\n self._current_creation_phase += 1\n creation_handler = self._creation_handlers[self._current_creation_phase]\n Mgr.add_task(creation_handler, \"draw_object\", sort=3)\n phase_id = self._current_creation_phase + 1\n Mgr.update_app(\"status\", \"create\", self._obj_type, \"phase%s\" % phase_id)\n\n return start_creation_phase\n\n def __get_creation_phase_handler(self, main_handler_func):\n\n def handle_creation_phase(task):\n\n main_handler_func()\n\n return task.cont\n\n return handle_creation_phase\n\n def __set_custom_object_name(self, custom_name):\n\n self._custom_obj_name = custom_name\n\n def init_object(self, obj):\n\n self._obj = obj\n\n def get_object(self):\n\n return self._obj\n\n def get_object_type(self):\n\n return self._obj_type\n\n def generate_object_id(self):\n\n obj_id = (self._obj_type,) + self._id_generator.next()\n\n return obj_id\n\n def set_next_object_color(self):\n\n color_values = tuple(random.random() * .4 + .5 for i in range(3))\n GlobalData[\"next_%s_color\" % self._obj_type] = color_values\n\n def get_next_object_color(self):\n\n r, g, b = GlobalData[\"next_%s_color\" % self._obj_type]\n color = VBase4(r, g, b, 1.)\n\n return color\n\n def get_origin_pos(self):\n\n return self._origin_pos\n\n def add_history(self, toplevel_obj):\n\n Mgr.do(\"update_history_time\")\n name = toplevel_obj.get_name()\n event_descr = 'Create \"%s\"' % name\n obj_id = toplevel_obj.get_id()\n obj_data = {obj_id: toplevel_obj.get_data_to_store(\"creation\")}\n event_data = {\"objects\": obj_data}\n event_data[\"object_ids\"] = set(Mgr.get(\"object_ids\"))\n Mgr.do(\"add_history\", event_descr, event_data, update_time_id=False)\n\n def __end_creation(self, cancel=True):\n\n Mgr.remove_task(\"draw_object\")\n process = None\n\n if cancel or not self._obj.is_valid():\n\n self._obj.destroy()\n\n else:\n\n finalization = self._obj.finalize()\n\n if finalization:\n\n def finalize():\n\n finalization.next()\n\n for step in finalization:\n yield True\n\n obj_type = self._obj_type\n name = Mgr.get(\"next_obj_name\", obj_type)\n Mgr.update_remotely(\"next_obj_name\", name)\n\n if self._add_to_hist:\n self.add_history(self._obj.get_toplevel_object())\n\n yield False\n\n process = finalize()\n\n else:\n\n obj_type = self._obj_type\n name = Mgr.get(\"next_obj_name\", obj_type)\n Mgr.update_remotely(\"next_obj_name\", name)\n\n if self._has_color:\n self.set_next_object_color()\n\n if self._add_to_hist:\n self.add_history(self._obj.get_toplevel_object())\n\n self._obj = None\n self._current_creation_phase = 0\n\n Mgr.notify(\"creation_ended\")\n Mgr.enter_state(\"creation_mode\")\n\n if process and process.next():\n Mgr.show_screenshot()\n descr = \"Creating %s...\" % self._obj_type\n Mgr.do_gradually(process, \"creation\", descr, cancellable=True)\n","sub_path":"src/core/base/creation_mgr.py","file_name":"creation_mgr.py","file_ext":"py","file_size_in_byte":7095,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"203839692","text":"def rickroll(cycle, message, regex):\n cycle.log(cycle.bot.sendMessage(message.chatID, \"Tsk. Rickrolls are so outdated.\", replyID=message.messageID))\n\ndef reddit(cycle, message, regex):\n cycle.log(cycle.bot.sendMessage(message.chatID, 'Pfft, redditfag. Try voat.co or just remove yourself.', replyID=message.messageID))\n\nplugin={\n 'name': 'anti-rickroll',\n 'functions': {\n r'dQw4w9WgXcQ': rickroll,\n r'reddit': reddit\n },\n 'description':\n \"\"\"Detects and snuffs rickrolls and other things\"\"\",\n 'summary':\n None #undocumented\n }\n","sub_path":"plugins/rickroll.py","file_name":"rickroll.py","file_ext":"py","file_size_in_byte":582,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"114854232","text":"from PyQt4.QtCore import *\nfrom PyQt4.QtGui import *\nimport subprocess \nimport os\nimport time\nimport shlex\n\nclass MakeProcess(QObject):\n\tdef __init__(self, server, directory):\n\t\tsuper(type(self),self).__init__()\n\t\tself.notice=SIGNAL(\"progress(QString)\")\n\t\tself.server=server\n\t\tself.settings=QSettings(\"Gogo, Inc.\", \"Batsview\")\n\t\tself.directory=directory\n\n\tdef make(self,target):\n\t\tself.target=target\n\n\tdef lru(self, lru):\n\t\tself.lru = lru\n\n\tdef work(self):\n\t\tif self.lru:\n\t\t\tmakecmd='make %s BORG=%s'%(self.target, self.lru)\n\t\telse:\n\t\t\tmakecmd='make %s'%(self.target)\n\t\tuser=str(self.settings.value(\"username\").toString())\n\t\tif len(user)>0:\n\t\t\tuser=user+\"@\"\n\t\tshell=str(self.settings.value(\"shell\").toString())\n\t\tif shell==\"xterm\":\n\t\t\tcmd='xterm -T \"%s\" -e \"ssh %s%s \\'bash -c \\\\\"cd %s; if ! %s; then echo FAIL; sleep 3000; fi\\\\\"\\'\"'%(makecmd,user,self.server,self.directory, makecmd)\n\t\telse:\n\t\t\tcmd='gnome-terminal -e \"ssh %s%s \\'bash -c \\\\\"cd %s; if ! %s; then echo FAIL; sleep 3000; fi\\\\\"\\'\"'%(user,self.server,self.directory, makecmd)\n\t\tos.system(cmd)\n\t\tself.emit(self.notice, \"DONE\")\n\n\tdef old(self):\n\t\tcmd='xterm -e \"ssh %s \\'bash -c \\\\\"cd %s; if ! make %s BORG=%s; then echo FAIL; sleep 3000; fi\\\\\"\\'\"'%(self.server,self.directory, self.target, self.lru)\n\t\tproc=subprocess.Popen(shlex.split(cmd),stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n\n\t\t\n\t\tline=\" \"\n\t\twhile line != '':\n\t\t\tline=proc.stdout.readline()\n\t\t\tself.emit(self.notice, line[0:-1])\n\t\t\t\n\t\t\ttime.sleep(1)\n\t\tresult=proc.wait()\n\t\tself.emit(self.notice, \"DONE\")\n\nclass Monitor(QObject):\n\tdef __init__(self, server, directory):\n\t\tsuper(type(self),self).__init__()\n\t\tself.proc=MakeProcess(server,directory)\n\t\tself.thr=QThread(self)\n\t\tself.proc.moveToThread(self.thr)\n\n\t\tself.connect(self.thr, SIGNAL(\"started()\"),self.proc.work)\n\t\tself.connect(self.proc, SIGNAL(\"progress(QString)\"),self.progress)\n\n\tdef run(self, cmd=\"help\", lru=None):\n\t\tself.proc.make(cmd)\n\t\tself.proc.lru(lru)\n\t\tself.start_time=time.time()\n\t\tself.thr.start()\n\t\n\tdef progress(self,title):\n\t\tif title==\"DONE\":\n\t\t\tself.stop_time=time.time()\n\t\t\tself.thr.quit()\n\n\tdef wait(self):\n\t\tself.thr.wait()\n\tdef duration(self):\n\t\tif self.stop_time and self.start_time:\n\t\t\treturn self.stop_time - self.start_time\n\n\t\nif __name__ == \"__main__\":\n\timport sys\n\timport signal\n\tsignal.signal(signal.SIGINT, signal.SIG_DFL)\n\tapp=QApplication(sys.argv)\n\twin=Monitor()\n\twin.run(\"rfs\")\n\twin.wait()\n\tapp.exec_()\t\n","sub_path":"build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":2429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"297063968","text":"\"\"\"Contains the scoring algorithms used in the model.\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom .models import OPC\nfrom .utils import k_kohler, ri_eff\nfrom .mie import cscat\n\n\ndef compute_bin_assessment(opc, refr, kappa, rh_values=[0., 35., 95.]):\n \"\"\"Assess the ability of an OPC to assign particles to their correct bin.\n\n Parameters\n ----------\n opc: opcsim.OPC\n refr: complex\n The complex refractive index of the material to assess\n kappa: float\n The kappa value to use for hygroscopic growth\n rh_values: list-like\n A list of relative humidities to assess the OPC at.\n\n Returns\n -------\n rv: pd.DataFrame\n A dataframe containing the results with self-explanatory columns.\n\n Examples\n --------\n\n \"\"\"\n assert(isinstance(opc, OPC)), \"opc must be an instance of the opcsim.OPC class\"\n\n # init the dataframe to hold our results\n rv = pd.DataFrame()\n\n for rh in rh_values:\n for i, _bins in enumerate(opc.bins):\n # compute the wet diameter\n wet_diam_lo = k_kohler(diam_dry=_bins[0], kappa=kappa, rh=rh)\n wet_diam_hi = k_kohler(diam_dry=_bins[-1], kappa=kappa, rh=rh)\n\n # compute the pct_dry\n pct_dry = (_bins[0]**3) / (wet_diam_lo**3)\n\n # compute the effective RI\n ri = ri_eff(species=[refr, complex(1.333, 0)], weights=[pct_dry, 1-pct_dry])\n\n # compute the scattering cross-section\n cscat_lo_exp = cscat(\n dp=_bins[0], wl=opc.wl, refr=refr, theta1=opc.theta[0], theta2=opc.theta[1])\n cscat_hi_exp = cscat(\n dp=_bins[-1], wl=opc.wl, refr=refr, theta1=opc.theta[0], theta2=opc.theta[1])\n\n cscat_lo = cscat(\n dp=wet_diam_lo, wl=opc.wl, refr=ri, theta1=opc.theta[0], theta2=opc.theta[1])\n cscat_hi = cscat(\n dp=wet_diam_hi, wl=opc.wl, refr=ri, theta1=opc.theta[0], theta2=opc.theta[1])\n\n # assign bins\n bin_assign_lo = opc.calibration_function(values=[cscat_lo])\n bin_assign_hi = opc.calibration_function(values=[cscat_hi])\n\n # add results to the dataframe\n rv = rv.append({\n \"bin_true\": i,\n \"bin_lo\": bin_assign_lo[0] if len(bin_assign_lo) > 0 else -99,\n \"bin_hi\": bin_assign_hi[0] if len(bin_assign_hi) > 0 else -99,\n \"refr_eff\": ri,\n \"rh\": rh,\n \"cscat_hi_ratio\": cscat_hi / cscat_hi_exp,\n \"cscat_lo_ratio\": cscat_lo / cscat_lo_exp,\n }, ignore_index=True)\n \n # force datatypes to be correct\n rv[\"bin_true\"] = rv[\"bin_true\"].astype(int)\n rv[\"bin_lo\"] = rv[\"bin_lo\"].astype(int)\n rv[\"bin_hi\"] = rv[\"bin_hi\"].astype(int)\n rv[\"rh\"] = rv[\"rh\"].astype(float)\n rv[\"cscat_hi_ratio\"] = rv[\"cscat_hi_ratio\"].astype(float)\n rv[\"cscat_lo_ratio\"] = rv[\"cscat_lo_ratio\"].astype(float)\n\n return rv\n\n# def nv_score(model, distribution, dmin=0.0, dmax=2.5, **kwargs):\n# \"\"\"Calculate and return the number-to-volume ratio.\n\n# The total number of particles is calculated by calculating the total number\n# of particles in each individual bin, and then summing them. The total volume\n# in the distribution is calculated by integrating the Volume-weighted CDF\n# between 0 and `dmax` microns.\n\n# Parameters\n# ----------\n# model : OPC\n# A valid OPC model describing an OPC that can be evaluated.\n# distribution : AerosolDistribution\n# A valid AerosolDistribution instance that can be evaluated.\n# dmin : float\n# The minimum particle size to integrate the CDF under. Default is 0.0\n# microns.\n# dmax : float\n# The maximum particle size to integrate the CDF under. Default is 2.5\n# microns.\n\n# Returns\n# -------\n# N/V : float\n# Returns the number-to-volume ratio as a single float.\n\n# Examples\n# --------\n\n# Compute the number-to-volume ratio for a 2-bin OPC on the Urban distribution\n\n# >>> opc = opcsim.OPC(n_bins=2)\n# >>> urban = opcsim.load_distribution(\"Urban\")\n# >>> n_v = opcsim.metrics.nv_score(opc, urban)\n\n# \"\"\"\n# # evaluate the total number of particles in each bin (then sum)\n# total_number = model.number(distribution, **kwargs).sum()\n\n# # evaluate the total volume in the distribution < dmax\n# total_volume = distribution.cdf(weight='volume', dmax=dmax)\n\n# return total_number / total_volume\n\n# def vv_score(model, distribution, dmin=0.0, dmax=2.5, **kwargs):\n# \"\"\"Calculate and return the volume-to-volume ratio.\n\n# The total volume of particles per the OPC is calculated by calculating the\n# total number of particles in each individual bin, and then multiplying each\n# bin by a 'volume-factor'. The sum of individual bin volumes is then used.\n# The total volume in the distribution is calculated by integrating the\n# Volume-weighted CDF between 0 and `dmax` microns.\n\n# Parameters\n# ----------\n# model : OPC\n# A valid OPC model describing an OPC that can be evaluated.\n# distribution : AerosolDistribution\n# A valid AerosolDistribution instance that can be evaluated.\n# dmin : float\n# The minimum particle size to integrate the CDF under. Default is 0.0\n# microns.\n# dmax : float\n# The maximum particle size to integrate the CDF under. Default is 2.5\n# microns.\n\n# Returns\n# -------\n# V/V : float\n# Returns the volume-to-volume ratio as a single float.\n\n# Examples\n# --------\n\n# Compute the number-to-volume ratio for a 2-bin OPC on the Urban distribution\n\n# >>> opc = opcsim.OPC(n_bins=2)\n# >>> urban = opcsim.load_distribution(\"Urban\")\n# >>> v_v = opcsim.metrics.vv_score(opc, urban)\n\n# \"\"\"\n# # evaluate the total number of particles in each bin (then sum)\n# measured_volume = model.volume(distribution, **kwargs).sum()\n\n# # evaluate the total volume in the distribution < dmax\n# total_volume = distribution.cdf(weight='volume', dmin=dmin, dmax=dmax)\n\n# return measured_volume / total_volume\n","sub_path":"opcsim/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":6188,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"321371836","text":"import discord\r\nimport asyncio\r\nfrom discord.ext.commands import Bot\r\nimport platform\r\nimport random\r\nfrom egg_assets import greet_txt, tulku_memes, bw_text, fortune_list\r\nimport re\r\n\r\ncommands = dict()\r\nvars = dict()\r\nqueue = []\r\n\r\n'''\r\nscript = 'eggscript\\n' \\\r\n '0 for 0:3\\n' \\\r\n '1 print hello\\n' \\\r\n '2 end\\n' \\\r\n '3 number i 0\\n' \\\r\n '4 number j 2 \\n' \\\r\n '5 print true\\n' \\\r\n '6 ++ i\\n' \\\r\n '7 if < $i $j\\n' \\\r\n '8 print $i\\n' \\\r\n '9 goto 4\\n' \\\r\n '10 end\\n' \\\r\n '11 print $i\\n' \\\r\n '12 -- i\\n' \\\r\n '13 print $i'\r\n'''\r\ndef for_loop(start, end):\r\n param = commands[start][4:]\r\n x = int(param[0:param.find(':')])\r\n y = int(param[param.find(':') + 1:])\r\n\r\n for i in range(x, y):\r\n for j, cmd in enumerate(range(start + 1, end)):\r\n queue.insert(j, cmd)\r\n\r\nasync def prnt(start, client, channel):\r\n param = commands[start][6:]\r\n if param.startswith('$'):\r\n print(vars.keys())\r\n var_name = param[1:]\r\n await client.send_message(channel, str(vars[var_name]))\r\n else:\r\n await client.send_message(channel, commands[start][6:])\r\n\r\ndef number(start):\r\n param = commands[start][7:]\r\n var_name = param[0:param.find(' ')]\r\n var_val = int(param[param.find(' ') + 1:])\r\n\r\n vars[var_name] = var_val\r\n\r\ndef incr(start):\r\n param = commands[start][3:]\r\n new_value = vars[param] + 1\r\n vars[param] = new_value\r\n\r\ndef decr(start):\r\n param = commands[start][3:]\r\n new_value = vars[param] - 1\r\n vars[param] = new_value\r\n\r\ndef boolean_operation(boolean_op, x, y):\r\n if boolean_op == '>':\r\n return x > y\r\n elif boolean_op == '<':\r\n return x < y\r\n elif boolean_op == '=':\r\n return x == y\r\n\r\n#TODO: work on this\r\ndef if_statement(start, end):\r\n param = commands[start][3:]\r\n boolean_op = param[0:1]\r\n values = param[2:].split(' ')\r\n x = int(vars[values[0][1:]] if values[0].startswith('$') else values[0])\r\n y = int(vars[values[1][1:]] if values[1].startswith('$') else values[1])\r\n if boolean_operation(boolean_op, x, y):\r\n for j, cmd in enumerate(range(start + 1, end)):\r\n queue.insert(j, cmd)\r\n else:\r\n for j, cmd in enumerate(range(start, end)):\r\n queue.pop(0)\r\n\r\n\r\ndef goto(start):\r\n param = int(commands[start][5:])\r\n for j, cmd in enumerate(range(param, start + 1)):\r\n queue.insert(j, cmd)\r\n\r\nasync def parse(client, channel):\r\n while (len(queue) > 0):\r\n i = queue.pop(0)\r\n cmd = commands[i]\r\n print('current line: ' + str(i))\r\n print('current cmd: ' + cmd)\r\n print('current queue: ' + str(queue))\r\n\r\n if cmd.startswith('STOP'):\r\n while (len(queue) > 0):\r\n queue.pop()\r\n elif cmd.startswith('for'):\r\n end_for = -1\r\n for j in range(i, len(commands.keys())):\r\n if commands[j].startswith('end'):\r\n end_for = j\r\n break\r\n for_loop(i, end_for)\r\n elif cmd.startswith('print'):\r\n await prnt(i, client, channel)\r\n elif cmd.startswith('number'):\r\n number(i)\r\n elif cmd.startswith('++'):\r\n incr(i)\r\n elif cmd.startswith('--'):\r\n decr(i)\r\n elif cmd.startswith('if'):\r\n end_if = -1\r\n for j in range(i, len(commands.keys())):\r\n if commands[j].startswith('end'):\r\n end_if = j\r\n break\r\n if_statement(i, end_if)\r\n elif cmd.startswith('goto'):\r\n goto(i)\r\n\r\nasync def go(script, client, channel):\r\n for line in script.split(';'):\r\n if line.startswith('eggscript;') or len(line) == 0:\r\n continue\r\n lineInd = int(line[0:line.find(' ')])\r\n lineCommand = line[line.find(' ') + 1:]\r\n\r\n commands[lineInd] = lineCommand\r\n\r\n queue.append(lineInd)\r\n\r\n await parse(client, channel)\r\n\r\n","sub_path":"eggscript.py","file_name":"eggscript.py","file_ext":"py","file_size_in_byte":4023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"453871149","text":"# coding: utf8\n\"\"\" Tests:\n\n- :class:`MarginalProbaDescentProcesses` check that the marginal probability of selecting any integer is indeed given by the ``_bernoulli_param`` attribute\n\"\"\"\n\nimport unittest\n\nimport numpy as np\n\nimport sys\nsys.path.append('..')\n\nfrom dppy.exotic_dpps import CarriesProcess, DescentProcess, VirtualDescentProcess\n\n\nclass MarginalProbaDescentProcesses(unittest.TestCase):\n \"\"\" Check that the marginal probability of selecting any integer is indeed given by the ``_bernoulli_param`` attribute\n \"\"\"\n\n size = 10000\n tol = 1e-2\n\n def test_carries_process(self):\n\n cp = CarriesProcess(base=10)\n cp.sample(size=self.size)\n\n estim = len(cp.list_of_samples[-1]) / self.size\n\n self.assertTrue(np.abs(estim - cp._bernoulli_param) < self.tol)\n\n def test_descent_process(self):\n\n dp = DescentProcess()\n dp.sample(size=self.size)\n\n estim = len(dp.list_of_samples[-1]) / self.size\n\n self.assertTrue(np.abs(estim - dp._bernoulli_param) < self.tol)\n\n def test_virtual_descent_process(self):\n\n vdp = VirtualDescentProcess(x_0=0.5)\n vdp.sample(size=self.size)\n\n estim = len(vdp.list_of_samples[-1]) / self.size\n\n self.assertTrue(np.abs(estim - vdp._bernoulli_param) < self.tol)\n\n\ndef main():\n\n unittest.main()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"tests/test_descent_processes.py","file_name":"test_descent_processes.py","file_ext":"py","file_size_in_byte":1368,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"292870974","text":"import numpy as np\n\nfrom scipy.optimize import curve_fit\nfrom scipy.special import erf\n\n#=================================================================================================\n\ndef extractLineProfile( img, line=[ 1., 0. ] ):\n '''\n extractLineProfile( img, line ):\n Returns a line profile of a given image 'img' along the line=[ m, c ]\n where m is the slope and c is the y-intercept.\n\n Params: \n img: 2D Numpy array\n line: list containing slope and y-intercept\n\n Returns:\n intens: 1D Numpy array of intensities along the line\n '''\n x, y = np.meshgrid( np.arange( img.shape[0] ), np.arange( img.shape[1] ) )\n y = np.flipud( y )\n here = np.where( np.absolute( y - ( line[0]*x + line[1] ) ) < 1. )\n return img[ here ], here\n\n#=================================================================================================\n\ndef myEdge( x, mu, sig, amp=1. ):\n '''\n myEdge: \n Models a rising edge by the following function:\n f(x) = \\frac{A}{2} \\left[ 1 + \\text{erf}\\left(\\frac{x-\\mu}{\\sigma}\\right) \\right ]\n '''\n return( amp / 2. ) * ( 1. + erf( ( x - mu ) / sig ) )\n\n#=================================================================================================\n\nif __name__=='__main__':\n import spatialResolution as sr\n from argparse import Namespace\n import scipy.io as sio\n\n filename = '/home/smaddali/ANL/BeamRuns/Feb2018/reconstructions/stdSample_solution.mat'\n calfilename = '/home/smaddali/ANL/Manuscripts/HEBCDI/data/bases_FINAL.mat'\n dat = Namespace( **sio.loadmat( filename ) )\n cal = Namespace( **sio.loadmat( calfilename ) )\n print( 'Array shape = ', dat.rho.shape )\n intens, here = sr.extractLineProfile( \n np.absolute( dat.rho[:,:,33] ), \n line=[ 1., 0. ]\n )\n\n # fitting edge\n data = intens[50:66] \n fspace_steps = cal.real_stdSample[:,0].reshape(-1,1)@here[0].reshape(1,-1) + cal.real_stdSample[:,1].reshape(-1,1)@here[1].reshape(1,-1)\n fspace_steps = fspace_steps - fspace_steps[:,0].reshape( -1, 1 ).repeat( fspace_steps.shape[1], axis=1 )\n my_x = np.sqrt( ( fspace_steps**2 ).sum( axis=0 ) )[50:66]\n seg = np.array(\n [ \n [ here[1][50], here[1][66] ], \n [ here[0][50], here[0][66] ]\n ]\n )\n popt, pcov = curve_fit( \n myEdge, \n my_x, data, \n p0=[ 1750., 100., 0.09 ]\n )\n \n # plotting\n plt.figure( 1 )\n plt.clf()\n plt.imshow( np.absolute( dat.rho[:,:,33] ) )\n plt.xlim( [ 44, 84 ] )\n plt.ylim( [ 44, 84 ] )\n plt.colorbar()\n plt.plot( seg[0], seg[1], 'c' )\n plt.set_cmap( 'inferno' )\n plt.xticks( [] )\n plt.yticks( [] )\n \n plt.figure( 2 )\n plt.clf()\n plt.plot( my_x, data, '-o', label='Line profile' )\n plt.plot( my_x, myEdge( my_x, popt[0], popt[1], popt[2] ), label='Fitted edge' )\n plt.grid()\n plt.legend( loc='best' )\n plt.xlabel( 'nm', fontsize=18, fontweight='bold' )\n plt.ylabel( '$\\\\left|\\\\rho\\\\right|$', fontsize=18, fontweight='bold' )\n","sub_path":"Python/spatialResolution.py","file_name":"spatialResolution.py","file_ext":"py","file_size_in_byte":3050,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"605488469","text":"from re import split\n\nfrom django.contrib.auth import get_user_model\nfrom django.contrib.auth.tokens import default_token_generator\nfrom django.db.models.aggregates import Avg\nfrom rest_framework import permissions\nfrom rest_framework.decorators import action, api_view, permission_classes\nfrom rest_framework.exceptions import ValidationError\nfrom rest_framework.filters import SearchFilter\nfrom rest_framework.generics import get_object_or_404\nfrom rest_framework.mixins import (CreateModelMixin, DestroyModelMixin,\n ListModelMixin)\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.permissions import AllowAny, IsAuthenticated\nfrom rest_framework.views import Response, status\nfrom rest_framework.viewsets import GenericViewSet, ModelViewSet\nfrom rest_framework_simplejwt.tokens import AccessToken\n\nfrom .filters import TitleFilter\nfrom .models import Category, Genre, Review, Title\nfrom .permissions import IsAdmin, IsModerator, IsOwnerOrReadOnly, ReadOnly\nfrom .serializers import (CategorySerializer, CommentsSerializer,\n ConfirmTokenSerializer, CreateUserSerializer,\n GenreSerializer, ReviewsSerializer,\n TitleSerializerRead, TitleSerializerWrite,\n UserSerializer)\n\nUser = get_user_model()\n\n\nclass CreateListDestroyView(CreateModelMixin, ListModelMixin,\n DestroyModelMixin, GenericViewSet):\n pass\n\n\n@api_view(['POST'])\n@permission_classes([AllowAny])\ndef create_user(request):\n serialized = CreateUserSerializer(data=request.data)\n serialized.is_valid()\n username = split(r'_', serialized.data['email'])[0]\n user, created = User.objects.get_or_create(\n email=serialized.data['email'],\n defaults={'username': username})\n if created:\n user.set_unusable_password()\n user.save()\n confirmation_code = default_token_generator.make_token(user)\n user.email_user(\n subject='Confirmation code',\n message='Код подтверждения - {}'.format(confirmation_code)\n )\n return Response(serialized.data, status=status.HTTP_201_CREATED)\n\n\n@api_view(['POST'])\n@permission_classes([AllowAny])\ndef get_token(request):\n serialized = ConfirmTokenSerializer(data=request.data)\n serialized.is_valid(raise_exception=True)\n user = User.objects.get(email=serialized.validated_data['email'])\n if not default_token_generator.check_token(\n user,\n serialized.validated_data['confirmation_code']\n ):\n raise ValidationError('Data is not valid')\n access = AccessToken.for_user(user)\n return Response(\n {\"token\": str(access)}\n )\n\n\nclass UsersListCreateViewSet(ModelViewSet):\n lookup_field = 'username'\n queryset = User.objects.all()\n serializer_class = UserSerializer\n permission_classes = [IsAuthenticated, IsAdmin]\n filter_backends = [SearchFilter]\n search_fields = ('username',)\n\n @action(\n methods=['GET', 'PATCH'],\n detail=False,\n permission_classes=[IsAuthenticated]\n )\n def me(self, request, pk=None):\n user = self.request.user\n if request.method == 'PATCH':\n serializer = self.get_serializer(\n user,\n data=request.data,\n partial=True)\n serializer.is_valid(raise_exception=True)\n serializer.save(role=user.role)\n serializer = self.get_serializer(user)\n return Response(serializer.data)\n\n\nclass ReviewsViewSet(ModelViewSet):\n serializer_class = ReviewsSerializer\n permission_classes = [IsOwnerOrReadOnly | IsAdmin | IsModerator]\n\n def get_queryset(self):\n title = get_object_or_404(\n Title,\n id=self.kwargs['title_id']\n )\n return title.reviews.all()\n\n def perform_create(self, serializer):\n title = get_object_or_404(Title, id=self.kwargs.get('title_id'))\n serializer.save(author=self.request.user, title=title)\n\n\nclass CommentsViewSet(ModelViewSet):\n serializer_class = CommentsSerializer\n permission_classes = [IsOwnerOrReadOnly | IsAdmin | IsModerator]\n pagination_class = PageNumberPagination\n\n def get_queryset(self):\n review = get_object_or_404(\n Review,\n title__id=self.kwargs['title_id'],\n id=self.kwargs['review_id']\n )\n return review.comments.all()\n\n def perform_create(self, serializer):\n review = get_object_or_404(\n Review,\n title__id=self.kwargs['title_id'],\n id=self.kwargs.get('review_id')\n )\n serializer.save(author=self.request.user, review=review)\n\n\nclass CategoryViewSet(CreateListDestroyView):\n queryset = Category.objects.all()\n permission_classes = [ReadOnly | IsAdmin]\n serializer_class = CategorySerializer\n lookup_field = 'slug'\n filter_backends = [SearchFilter]\n search_fields = ['=name']\n\n\nclass GenreViewSet(CreateListDestroyView):\n queryset = Genre.objects.all()\n permission_classes = [ReadOnly | IsAdmin]\n serializer_class = GenreSerializer\n lookup_field = 'slug'\n filter_backends = [SearchFilter]\n search_fields = ['=name']\n\n\nclass TitleViewSet(ModelViewSet):\n queryset = Title.objects.annotate(\n rating=Avg('reviews__score')\n ).order_by('-rating')\n permission_classes = [ReadOnly | IsAdmin]\n filterset_class = TitleFilter\n\n def get_serializer_class(self):\n if self.request.method in permissions.SAFE_METHODS:\n return TitleSerializerRead\n return TitleSerializerWrite\n","sub_path":"api_v1/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5630,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"207503841","text":"#Archivo: Tuplas.py\n#Descripcion: Ejemplo en el que se muestra la utilizacion de las tuplas\n#\n\n#Declaracion de una tupla\nconjunto = (1,2,\"tres\",\"cuatro\")\n\n#Recorrido de una tupla como secuencia\ni = 0\nfor elemento in conjunto:\n print(f'conjunto[{i}]={elemento}')\n i+=1\n\n#Modificacion de un elemento de una tupla - No inmutable\n#conjunto[0] = 0\n#print('Despues de modificar, la lista quedo',conjunto)\n\n#Eliminacion de un elemento de una tupla - No inmutable\n#del(conjunto[0])\n\n#Agregado de un elemento de una lista\n#conjunto.append(100)\n#print('Despues de agregar el elemento al final, la lista quedo',conjunto)\n\n#Verificacion de a pertenencia\n\nif 'cuatro' in conjunto:\n print(\"El cuatro esta en el conjunto\")\n\n#Eliminacion de la tupla entera\ndel conjunto\n\n#Genera una excepcion - Manejo de excepciones\ntry:\n print('Despues de eliminar la el conjunto quedo',conjunto)\nexcept:\n print('Ya no existe el conjunto')\n\n","sub_path":"Tuplas.py","file_name":"Tuplas.py","file_ext":"py","file_size_in_byte":924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"62715222","text":"import uiScriptLocale\n\nROOT = \"d:/ymir work/ui/minimap/\"\n\nwindow = {\n\t\"name\" : \"AtlasWindow\",\n\t\"style\" : (\"movable\", \"float\",),\n\n\t\"x\" : SCREEN_WIDTH - 136 - 256 - 10,\n\t\"y\" : 0,\n\n\t\"width\" : 256 + 15,\n\t\"height\" : 256 + 38+50,\n\n\t\"children\" :\n\t(\n\t\t{\n\t\t\t\"name\" : \"TitleBar\",\n\t\t\t\"type\" : \"roofbar\",\n\t\t\t\"style\" : (\"attach\",),\n\n\t\t\t\"x\" : -8,\n\t\t\t\"y\" : 7,\n\n\t\t\t\"width\" : 256+30+15,\n\t\t\t\"color\" : \"red\",\n\t\t\t\n\t\t\t\"children\" : (\n\t\t\t\t{\n\t\t\t\t\t\"name\" : \"enableWarpWindowButton\",\n\t\t\t\t\t\"type\" : \"button\",\n\t\t\t\t\t\n\t\t\t\t\t\"x\" : 35,\n\t\t\t\t\t\"y\" : 20,\n\t\t\t\t\t\n\t\t\t\t\t\"default_image\" : \"yamato_button/button_small_n.tga\",\n\t\t\t\t\t\"over_image\" : \"yamato_button/button_small_h.tga\",\n\t\t\t\t\t\"down_image\" : \"yamato_button/button_small_p.tga\",\t\t\n\t\t\t\t\t\"text\" : \"GM: Warp\",\n\t\t\t\t\n\t\t\t\t},\n\t\t\t\n\t\t\t),\n\n\t\t},\n\t\t## BOARD\n\t\t{\n\t\t\t\"name\" : \"board\",\n\t\t\t\"type\" : \"board\",\n\t\t\t\n\t\t\t\"style\" : (\"attach\",),\n\n\t\t\t\"x\" : 0,\n\t\t\t\"y\" : 50,\n\n\t\t\t\"width\" : 256 + 15,\n\t\t\t\"height\" : 256 + 38,\n\t\t\t\n\t\t\t# \"children\" : (\n\t\t\t\t# {\n\t\t\t\t\t# \"name\" : \"warpWindow\",\n\t\t\t\t\t# \"type\" : \"window\",\n\t\t\t\t\t# \"style\" : (\"attach\",),\n\t\t\t\t\t\n\t\t\t\t\t# \"x\" : 0,\n\t\t\t\t\t# \"y\" : 50,\n\n\t\t\t\t\t# \"width\" : 256 + 15,\n\t\t\t\t\t# \"height\" : 256 + 38,\n\t\t\t\t# },\n\t\t\t\n\t\t\t# ),\n\n\t\t},\n\t),\n}\n\n# import uiScriptLocale\n\n# ROOT = \"d:/ymir work/ui/minimap/\"\n\n# window = {\n\t# \"name\" : \"AtlasWindow\",\n\t# \"style\" : (\"movable\", \"float\",),\n\n\t# \"x\" : SCREEN_WIDTH - 136 - 256 - 10,\n\t# \"y\" : 0,\n\n\t# \"width\" : 256 + 15,\n\t# \"height\" : 256 + 38,\n\n\t# \"children\" :\n\t# (\n\t\t# ## BOARD\n\t\t# {\n\t\t\t# \"name\" : \"board\",\n\t\t\t# \"type\" : \"board_with_titlebar\",\n\n\t\t\t# \"x\" : 0,\n\t\t\t# \"y\" : 0,\n\n\t\t\t# \"width\" : 256 + 15,\n\t\t\t# \"height\" : 256 + 38,\n\n\t\t\t# \"title\" : uiScriptLocale.ZONE_MAP,\n\t\t# },\n\t# ),\n# }\n","sub_path":"uiscript/uiscript/atlaswindow.py","file_name":"atlaswindow.py","file_ext":"py","file_size_in_byte":1638,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"359707694","text":"from selenium.webdriver.common.by import By\nfrom .base_page import BasePage\nfrom .locators import MainLocators\n\nclass MainPage(BasePage):\n\n def start_bnt_click(self):\n self.browser.execute_script(\"window.scrollBy(0, 600);\")\n start_button = self.browser.find_element(*MainLocators.START_GAME_BNT)\n start_button.click()\n self.browser.execute_script(\"window.scrollBy(0, -300);\")\n\n def close_cookies(self):\n close_button = self.browser.find_element(*MainLocators.CLOSE_COOKIES_BTN)\n close_button.click()\n","sub_path":"pages/main_page.py","file_name":"main_page.py","file_ext":"py","file_size_in_byte":548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"472934245","text":"\nimport csv\nfrom os.path import expanduser\nhome = expanduser(\"~\")\n\nvars = [\"ACEDEPRS\", \"ACEDRINK\", \"ACEDRUGS\", \"ACEPRISN\", \"ACEDIVRC\", \"ACEPUNCH\", \"ACEHURT\", \"ACESWEAR\", \"ACETOUCH\", \"ACETTHEM\", \"ACEHVSEX\"]\n\nparsed_file_path = home + \"/Downloads/ICGE_course/project/project_ace_python/ace_2012_project/\"\nparsed_info_file = \"parsed_ace_data_2012.csv\"\nstate_age_csv_file = \"state_age_distribution_ace_data_2012.csv\"\nstate_age_ace_csv_file = \"state_age_ace_statistic_2012.csv\"\n\nvariables_nbr = 359\nage_variable_idx = 50\nace_var_start_idx, ace_var_end_idx = 229, 239\n\nfieldnames = [i for i in range(1, variables_nbr + 1)]\n\nwith open(state_age_csv_file, 'w') as csvfile1, open(state_age_ace_csv_file, 'w') as csvfile2:\n writer_age = csv.writer(csvfile1, delimiter=',')\n writer_ace = csv.writer(csvfile2, delimiter=',')\n\n age_categories = [\"state_id\", \"people\", \"Don't know/Not sure\", \"Refused\", \"18-24\", \"25-34\", \"35-44\", \"45-54\",\n \"55-64\", \"65-99\"]\n writer_age.writerow(age_categories)\n\n states = [i for i in range(1, 57)] # no states: 52 43 14 7 3\n states.append(66) # Guam\n states.append(72) # Puerto Rico\n other_states = set()\n\n for state_id in states:\n\n state_content = []\n with open(parsed_file_path + parsed_info_file, \"r\") as content:\n\n for idx, line in enumerate(content):\n if idx == 0:\n continue\n\n line = line.split(\",\")\n if line[variables_nbr - 1] != \"nan\":\n line[variables_nbr - 1] = line[variables_nbr - 1][:len(line[variables_nbr - 1]) - 1]\n\n if int(line[0]) == state_id: # state data\n state_content.append(line)\n\n if 56 < int(line[0]) < 1:\n other_states.add(int(line[0]))\n\n#####################################################################################################\n # getting state and age based distribution for 2012 dataset\n age_distribution = {i: 0 for i in age_categories[2:]}\n\n # getting ace statistics by states and ages for 2012 dataset\n general_age_ace_data = {}\n\n for age_group_idx in age_distribution:\n general_ace_distributions = []\n\n # 22.1-4 and 22.5 -- combined with \"Parents not married\", idx = 8\n ace_answers1 = [\"yes\", \"no\", \"Don’t know/Not Sure\", \"Parents not married\", \"Refused\", \"Not asked or Missing\"]\n ace_answers_values1 = [\"1\", \"2\", \"7\", \"8\", \"9\", \"nan\"]\n for j in range(0, 5): # five of ace questions with these set of answers\n ace_set1_distribution = {i: 0 for i in ace_answers_values1}\n general_ace_distributions.append(ace_set1_distribution)\n # writer_ace.writerow(age_categories)\n\n # 22.6-11\n ace_answers2 = [\"Never\", \"Once\", \"More than once\", \"Don’t know/Not Sure\", \"Refused\", \"Not asked or Missing\"]\n ace_answers_values2 = [\"1\", \"2\", \"3\", \"7\", \"9\", \"nan\"]\n for j in range(0, 6): # six of ace questions with these set of answers\n ace_set2_distribution = {i: 0 for i in ace_answers_values2}\n general_ace_distributions.append(ace_set2_distribution)\n\n general_age_ace_data[age_group_idx] = general_ace_distributions\n\n ace_columns = [\"state_id\", \"state sample population\", \"age category\", \"age category population\",\n \"ace question\", \"\"]\n for i in range(0, len(ace_answers1)):\n ace_columns.append(ace_answers1[i] + \" / \" + ace_answers2[i] + \" -- number of responses\")\n writer_ace.writerow(ace_columns)\n\n for idx, line in enumerate(state_content):\n if idx == 0:\n continue\n\n age_nbr = int(line[age_variable_idx])\n\n ace_values = [line[i] for i in range(ace_var_start_idx, ace_var_end_idx + 1)]\n # age_category = age_categories[2]\n # for var_idx in range(len(ace_values)):\n # # age_ace_data = general_age_ace_data[age_category]\n # # ace_data_distribution = age_ace_data[age_category][var_idx]\n # # ace_data_distribution[ace_values[var_idx]] += 1\n # general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n\n if age_nbr == 7:\n age_distribution[age_categories[2]] += 1\n age_category = age_categories[2]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif age_nbr == 9:\n age_distribution[age_categories[3]] += 1\n age_category = age_categories[3]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 25 > age_nbr > 17:\n age_distribution[age_categories[4]] += 1\n age_category = age_categories[4]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 35 > age_nbr > 24:\n age_distribution[age_categories[5]] += 1\n age_category = age_categories[5]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 45 > age_nbr > 34:\n age_distribution[age_categories[6]] += 1\n age_category = age_categories[6]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 55 > age_nbr > 44:\n age_distribution[age_categories[7]] += 1\n age_category = age_categories[7]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 65 > age_nbr > 54:\n age_distribution[age_categories[8]] += 1\n age_category = age_categories[8]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n elif 100 > age_nbr > 64:\n age_distribution[age_categories[9]] += 1\n age_category = age_categories[9]\n for var_idx in range(len(ace_values)):\n general_age_ace_data[age_category][var_idx][ace_values[var_idx]] += 1\n\n#####################################################################################################\n # outputting rows into csv file\n\n print(state_id)\n age_data = [state_id, len(state_content)]\n [age_data.append(j) for i, j in age_distribution.items()]\n writer_age.writerow(age_data)\n\n no_ace_status = 0\n age_ace_line = [state_id, len(state_content)]\n writer_ace.writerow(age_ace_line)\n\n age_idx = 0\n for i, j in age_distribution.items():\n age_ace_line = [\"\\t\", str(i), str(j)]\n writer_ace.writerow(age_ace_line)\n\n ace_data = general_age_ace_data[i]\n for var_id in range(len(vars)):\n age_ace_line = [\"\\t\", \"\\t\", vars[var_id]]\n writer_ace.writerow(age_ace_line)\n\n age_ace_line = [\"\\t\", \"\\t\", \"\\t\"]\n responses = ace_data[var_id]\n [age_ace_line.append(str(k) + \" -- \" + str(l)) for k, l in responses.items()]\n writer_ace.writerow(age_ace_line)\n\n cc = 0\n for k, l in responses.items():\n if cc == 5:\n continue\n no_ace_status += l\n cc += 1\n\n age_idx += 1\n if no_ace_status != 0:\n print(\"Stat exists in :\" + str(state_id))\n print(\"other states: \")\n print(other_states)\n\n\n#####################################################################################################\n","sub_path":"ace_2012_project/ace_need_data_2012.py","file_name":"ace_need_data_2012.py","file_ext":"py","file_size_in_byte":8108,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"540766730","text":"\"\"\"\nSplit each videos frame by frame.\n\"\"\"\n\nimport os\nimport imageio\nfrom scipy import misc\nfrom tqdm import tqdm\n\nnames = {\n '1_1': (0, 624),\n '1_2': (625, 1452),\n '2_1': (1453, 2001),\n '2_2': (2002, 2686),\n '2_3': (2687, 3454),\n '2_4': (3455, 4033),\n '2_5': (4034, 4928),\n '2_6': (4929, 5595),\n '3_1': (5596, 6253),\n '3_2': (6254, 6930),\n '3_3': (6931, 7738)\n}\n\nif not os.path.exists('UMN_frames'):\n os.makedirs('UMN_frames')\nv = imageio.get_reader('UMN/Crowd-Activity-All.avi', 'ffmpeg')\nnb_frames = v.get_meta_data()['nframes']\nfor f in tqdm(range(nb_frames)):\n try:\n f_data = v.get_data(f)\n misc.imsave('UMN_frames/frame_{}.png'.format(f), misc.imresize(f_data, (224, 224)))\n except RuntimeError:\n pass\nfor k in tqdm(list(names.keys())):\n for i in range(names[k][0], names[k][1] + 1):\n os.rename('UMN_frames/frame_{}.png'.format(i),\n 'UMN_frames/frame_{}-{}.png'.format(k, i))\n","sub_path":"data/UMN_split_to_frames.py","file_name":"UMN_split_to_frames.py","file_ext":"py","file_size_in_byte":971,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"5050313","text":"# -*- coding: utf-8 -*-\n\nfrom backoff import expo\nfrom backoff import on_exception\nfrom google.cloud import error_reporting\nfrom google.cloud import logging\nfrom logging import basicConfig\nfrom logging import getLogger\nfrom logging import NOTSET\n\n# The format for local logs.\nLOGS_FORMAT = (\"%(asctime)s \"\n \"%(name)s \"\n \"%(process)d \"\n \"%(thread)d \"\n \"%(levelname)s \"\n \"%(message)s\")\n\n# The path to the log file for local logging.\nLOG_FILE = \"/trump2cash/temp/trump2cash.log\"\n\n\nclass Logs:\n \"\"\"A helper for logging locally or in the cloud.\"\"\"\n\n def __init__(self, name, to_cloud=True):\n self.to_cloud = to_cloud\n\n # Initialize the local file logger.\n self.local_logger = getLogger(name)\n basicConfig(format=LOGS_FORMAT, level=NOTSET, filename=LOG_FILE)\n\n # If requested, also initialize the Stackdriver logging and error\n # reporting clients.\n if self.to_cloud:\n self.cloud_logger = logging.Client().logger(name)\n self.error_client = error_reporting.Client()\n\n def debug(self, text):\n \"\"\"Logs at the DEBUG level.\"\"\"\n\n if self.to_cloud:\n self.safe_cloud_log_text(text, severity=\"DEBUG\")\n else:\n self.local_logger.debug(text)\n\n def info(self, text):\n \"\"\"Logs at the INFO level.\"\"\"\n\n if self.to_cloud:\n self.safe_cloud_log_text(text, severity=\"INFO\")\n else:\n self.local_logger.info(text)\n\n def warn(self, text):\n \"\"\"Logs at the WARNING level.\"\"\"\n\n if self.to_cloud:\n self.safe_cloud_log_text(text, severity=\"WARNING\")\n else:\n self.local_logger.warning(text)\n\n def error(self, text):\n \"\"\"Logs at the ERROR level.\"\"\"\n\n if self.to_cloud:\n self.safe_cloud_log_text(text, severity=\"ERROR\")\n else:\n self.local_logger.error(text)\n\n def catch(self, exception):\n \"\"\"Logs an exception.\"\"\"\n\n if self.to_cloud:\n self.safe_report_exception()\n self.safe_cloud_log_text(str(exception), severity=\"CRITICAL\")\n else:\n self.local_logger.critical(str(exception))\n\n def safe_cloud_log_text(self, text, severity):\n \"\"\"Logs to the cloud, retries if necessary, and eventually fails over\n to local logs.\n \"\"\"\n\n try:\n self.retry_cloud_log_text(text, severity)\n except BaseException as exception:\n self.local_logger.error(\"Failed to log to cloud: %s %s %s\" %\n (exception, severity, text))\n\n @on_exception(expo, BaseException, max_tries=7)\n def retry_cloud_log_text(self, text, severity):\n \"\"\"Logs to the cloud and retries up to 7 times with exponential backoff\n if the upload fails.\n \"\"\"\n\n self.cloud_logger.log_text(text, severity=severity)\n\n def safe_report_exception(self):\n \"\"\"Reports the latest exception, retries if necessary, and eventually\n fails over to local logs.\n \"\"\"\n\n try:\n self.retry_report_exception()\n except BaseException as exception:\n self.local_logger.error(\"Failed to report exception: %s\" %\n exception)\n\n @on_exception(expo, BaseException, max_tries=7)\n def retry_report_exception(self):\n \"\"\"Reports the latest exception and retries up to 7 times with\n exponential backoff if the upload fails.\n \"\"\"\n\n self.error_client.report_exception()\n","sub_path":"logs.py","file_name":"logs.py","file_ext":"py","file_size_in_byte":3577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"215772360","text":"import os\nimport numpy as np\nimport cvxpy as cp\nimport scipy.io as sio\nimport scipy.misc\n\nfrom utils import setup_logger\n\nimport argparse\n\nparser = argparse.ArgumentParser(description='Test LADMM with synthetic data')\nparser.add_argument('-c', '--cols', type=int, default=0, help='number of columns in A to be replaced')\nparser.add_argument('-p', '--p', type=float, default=0.2, help='p in the Bernoulli distribution')\nparser.add_argument('-m', '--mu', type=float, default=0.0, help='mu of Gaussian dist')\nparser.add_argument('-s', '--sigma', type=float, default=2.0, help='sigma of Gaussian dist')\nparser.add_argument('--data-type', type=str, default='gaussian', help='data type')\nparser.add_argument('-a', '--alpha', type=float, default=0.01, help='hyper-param in the objective')\nparser.add_argument('--split', type=str, default='test', help='calculate train or test split')\nparser.add_argument('--batch-size', type=int, default=20, help='batch size')\n\ndef loss_l1(X):\n return cp.sum(cp.abs(X))\n\ndef objective_fn(Z, X, A, alpha):\n residual_l1 = loss_l1(cp.matmul(A, Z) - X)\n regularizer_l1 = loss_l1(Z)\n return (residual_l1 + alpha * regularizer_l1) / X.shape[1]\n\nif __name__ == '__main__':\n args = parser.parse_args()\n\n alpha = args.alpha\n split = args.split\n batch_size = args.batch_size\n\n # test data file\n test_file = 'syn_data'\n test_file += '_cols{}'.format(args.cols) if args.cols > 0 else ''\n test_file += '_p{}_mu{}_s{}'.format(args.p, args.mu, args.sigma)\n test_file += '_{}'.format(args.data_type) if args.data_type != 'gaussian' else ''\n test_file += '.mat'\n print('using testing data file {}'.format(test_file))\n\n # logger file\n if not os.path.isdir('cvx-solutions'):\n os.makedirs('cvx-solutions')\n if not os.path.isdir('cvx-solutions/logs'):\n os.makedirs('cvx-solutions/logs')\n save_file = os.path.join('cvx-solutions', '{}-alpha{}-{}.npy'.format(test_file[:-4], alpha, split))\n log_file = os.path.join('cvx-solutions/logs', '{}-alpha{}-{}.log'.format(test_file[:-4], alpha, split))\n print = setup_logger(log_file)\n\n syn_data = sio.loadmat(test_file)\n A = syn_data['A'].astype(np.float32)\n m, n = A.shape\n\n X = syn_data[split + '_x'].astype(np.float32).T # (m, #samples)\n Z = syn_data[split + '_z'].astype(np.float32).T # (n, #samples)\n E = syn_data[split + '_e'].astype(np.float32).T # (m, #samples)\n n_samples = X.shape[1]\n\n Z_var = cp.Variable((n,batch_size))\n X_param = cp.Parameter((m,batch_size))\n objective = cp.Minimize(objective_fn(Z_var, X_param, A, alpha))\n problem = cp.Problem(objective)\n\n Z_sol = np.zeros(Z.shape, dtype=np.float32)\n\n for i in range(n_samples // batch_size):\n\n X_param.value = X[:, i*batch_size:(i+1)*batch_size]\n out = problem.solve()\n print('[{:2d}/{:2d}]\\t{}'.format(i+1, n_samples//batch_size, out))\n Z_sol[:, i*batch_size:(i+1)*batch_size] = Z_var.value\n\n np.save(save_file, Z_sol)\n print('Solutions saved to file {}'.format(save_file))\n\n","sub_path":"compute_cvx_solutions.py","file_name":"compute_cvx_solutions.py","file_ext":"py","file_size_in_byte":3040,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"74164527","text":"'''\n Maan Qraitem \n Hateful Meme Classification \n'''\n\nfrom __future__ import print_function, division\nimport os\nimport os.path as osp \nimport torch\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\nimport json \nfrom vocab import Vocab\nfrom tqdm import tqdm\nimport pickle\n#import cv2\n\nclass FbDataset(Dataset): \n def __init__(self, dataset, root_dir, vocab, mode='train'):\n self.root_dir = root_dir\n self.meme_data = pd.read_json(osp.join(root_dir, dataset+'.jsonl'), lines=True) \n self.img_features = np.load(osp.join(root_dir, 'img_features_%s.npy'%dataset))\n self.mode = mode \n with open(osp.join(root_dir, 'img2idx_%s.pkl'%dataset), 'rb') as handle:\n self.img2idx = pickle.load(handle)\n \n self.entries = [] \n self.vocab = vocab\n self.max_length = 15\n \n\n self.load_data() \n self.tokenize() \n\n def add_entry(self, meme_entry):\n image_id = meme_entry['img'].split(\".\")[0].split(\"/\")[1]\n meme_text = meme_entry['text'] \n meme_img_feature = self.img_features[self.img2idx[image_id]] \n meme_id = meme_entry['id'] \n\n entry = { \n 'id':meme_id,\n 'imag_id': image_id,\n 'text': meme_text, \n 'img_feature': meme_img_feature\n }\n \n if self.mode == 'train': \n meme_label = np.array(meme_entry['label']).astype(np.float)\n entry['label'] = meme_label \n\n self.entries.append(entry)\n\n def load_data(self): \n for index, meme_entry in tqdm(self.meme_data.iterrows(), total=self.meme_data.shape[0]):\n self.add_entry(meme_entry) \n\n \n def tokenize(self): \n for entry in self.entries: \n meme_text = entry['text']\n meme_tokens = self.vocab.tokenize(meme_text)\n meme_tokens = [self.vocab.word2idx[token] for token in meme_tokens] \n meme_tokens = meme_tokens[:self.max_length] \n if len(meme_tokens) < self.max_length: \n padding = [self.vocab.padding_word_idx()] * (self.max_length - len(meme_tokens)) \n meme_tokens = meme_tokens + padding\n \n assert len(meme_tokens) == self.max_length, \"meme text size is not %d\"%self.max_length\n entry['text_tokens'] = np.array(meme_tokens) \n\n\n def __len__(self): \n return len(self.entries) \n\n def __getitem__(self, index):\n entry = self.entries[index] \n entry_id = str(entry['id'])\n entry_tokens = entry['text_tokens']\n entry_text = entry['text'] \n entry_img_feature = entry['img_feature'] \n \n sample = { \n 'id' : entry_id, \n 'tokens': entry_tokens, \n 'text' : entry_text, \n 'img_feature' : entry_img_feature, \n } \n\n if self.mode == 'train': \n entry_label = entry['label'] \n sample['label'] = entry_label \n\n return sample \n\ndef main():\n root_path = '../../data'\n vocab = Vocab(root_path) \n vocab.loadfiles('vocab_data.pkl') \n fbdataset = FbDataset('train.jsonl', root_path, vocab) \n for i in range(len(fbdataset)): \n sample = fbdataset[i]\n #print(sample['id'])\n #print(sample['tokens']) \n #print(sample['text']) \n #cv2.imshow('image',sample['img'])\n #cv2.waitKey(0)\n #cv2.destroyAllWindows() \n\nif __name__ == \"__main__\":\n main()\n \n\n","sub_path":"fbhm_text_only/dataset.py","file_name":"dataset.py","file_ext":"py","file_size_in_byte":3245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"90287916","text":"\r\n#Basketball Court Code adapted from Savvas Tjortjoglou\r\nfrom matplotlib.patches import Circle, Rectangle, Arc\r\nimport matplotlib.pyplot as plt\r\nimport seaborn as sns\r\nimport pandas\r\nimport math \r\n\r\nCATEGORIES = ['GameID', 'Game Date', 'Team ID', 'Team Name', 'Shot Period', 'Shot Number', 'Made/Missed', \r\n 'LeftValue', 'TopValue', 'Shot Info', 'Player Name', 'Player ID', 'ycoord', 'xcoord']\r\n\r\nDATAFRAME = pandas.ExcelFile('C:/Users/Gautam Goel/Documents/Documents/Projects/NCAA Project/New Code/dataframe18.xlsx')\r\nDATAFRAME.sheet_names\r\n\r\nDATA = DATAFRAME.parse('Sheet1')\r\nDATALIST = DATA.values.tolist()\r\nPLAYERDATAMAKE, PLAYERDATAMISS = [], []\r\ncurrTeam = 'duke-blue-devils'\r\ncurrPlayer = 'trae young'\r\n\r\n#Collect player shot data\r\ndef collect_shots():\r\n for data in DATALIST: \r\n if data[11] == currPlayer and data[7] == 'made': PLAYERDATAMAKE.append(data[1::])\r\n elif data[11] == currPlayer: PLAYERDATAMISS.append(data[1::])\r\n #if data[4] == currTeam: TEAMDATA.append(data[1::])\r\n \r\n for val in PLAYERDATAMAKE:\r\n left_value, top_value = val[7], val[8]\r\n \r\n if left_value < 50: val.append(float(left_value * 0.94))\r\n else: val.append(float((100 - left_value) * 0.94))\r\n \r\n if left_value <= 47: val.append(float(top_value * 0.5))\r\n else: val.append(float(50 - (top_value * 0.5)))\r\n \r\n val[13] = (val[13] - 25) * 10\r\n val[12] = val[12] * 10\r\n \r\n for val in PLAYERDATAMISS:\r\n left_value, top_value = val[7], val[8]\r\n \r\n if left_value < 50: val.append(float(left_value * 0.94))\r\n else: val.append(float((100 - left_value) * 0.94))\r\n \r\n if left_value <= 47: val.append(float(top_value * 0.5))\r\n else: val.append(float(50 - (top_value * 0.5)))\r\n \r\n val[13] = (val[13] - 25) * 10\r\n val[12] = val[12] * 10\r\n\r\n#Create NCAA basketball court \r\ndef create_court(ax = None, color = 'black', lw = 1):\r\n if ax is None: ax = plt.gca()\r\n \r\n hoop = Circle((0, 52.5), radius = 7.5, linewidth=lw, color=color, fill=False)\r\n backboard = Rectangle((-30, 40), 60, -1, linewidth=lw, color=color)\r\n paint = Rectangle((-60, 0), 120, 190, linewidth=lw, color=color, fill=False)\r\n \r\n free_throw_arc = Arc((0, 190), 120, 120, theta1=0, theta2=180, linewidth=lw, color=color, fill=False)\r\n three_point_arc = Arc((0, 0), 415, 500, theta1=0, theta2=180, linewidth=lw, color=color)\r\n center_arc = Arc((0, 470), 120, 120, theta1=180, theta2=0, linewidth=lw, color=color)\r\n \r\n line1 = Rectangle((-250, 100), 500, -1, linewidth=lw, color=color)\r\n line2 = Rectangle((-250, 250), 500, -1, linewidth=lw, color=color)\r\n\r\n court_elements = [hoop, backboard, paint, free_throw_arc, three_point_arc, center_arc, line1, line2]\r\n \r\n for element in court_elements: ax.add_patch(element)\r\n return ax\r\n \r\ndef main():\r\n #collect_shots()\r\n playermake_dataframe = pandas.DataFrame(PLAYERDATAMAKE, columns = CATEGORIES)\r\n playermiss_dataframe = pandas.DataFrame(PLAYERDATAMISS, columns = CATEGORIES)\r\n \r\n sns.set_style(\"white\")\r\n sns.set_color_codes()\r\n \r\n plt.figure(figsize=(12,11))\r\n create_court()\r\n plt.scatter(playermake_dataframe.xcoord, playermake_dataframe.ycoord, c='blue')\r\n plt.scatter(playermiss_dataframe.xcoord, playermiss_dataframe.ycoord, c='red')\r\n\r\n plt.xlim(-250, 250)\r\n plt.ylim(0, 500)\r\n \r\n \"\"\"\r\n cmap = plt.cm.YlOrRd_r\r\n cmap = plt.cm.gist_heat_r\r\n joint_chart = sns.jointplot(player_dataframe.xcoord, player_dataframe.ycoord, stat_func=None, \r\n kind='scatter', space=0, alpha=0.5)\r\n joint_chart = sns.jointplot(player_dataframe.xcoord, player_dataframe.ycoord, stat_func=None, \r\n kind='kde', space=0, color=cmap(0.1), cmap=cmap, n_levels=50)\r\n joint_chart = sns.jointplot(player_dataframe.xcoord, player_dataframe.ycoord, stat_func=None, \r\n kind='hex', space=0, color=cmap(.2), cmap=cmap)\r\n joint_chart.fig.set_size_inches(12, 11)\r\n \r\n ax = joint_chart.ax_joint\r\n create_court(ax)\r\n ax.set_xlim(-250, 250)\r\n ax.set_ylim(500, 0) \r\n ax.set_xlabel('')\r\n ax.set_ylabel('')\r\n ax.tick_params(labelbottom='off', labelleft='off')\r\n ax.set_title('Trae Young FGA 2017-2018 Reg. Season', y=1.2, fontsize=16)\r\n ax.text(-250, 500, 'Data Source ESPN.com', fontsize=12) \r\n \"\"\"\r\n \r\n plt.show()\r\n\r\nmain()\r\n","sub_path":"Shot Chart Visualization.py","file_name":"Shot Chart Visualization.py","file_ext":"py","file_size_in_byte":4531,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"423621569","text":"#!/usr/bin/env python\n\"\"\"Strip frame types from an i3 file.\"\"\"\n\nimport sys\nimport os\n\ntry:\n from icecube import icetray,dataio\nexcept ImportError:\n print('You must be inside an IceCube metaproject environment.')\n sys.exit(1)\n\ndef files(args):\n \"\"\"A frame generator that can continue over multiple files\"\"\"\n for a in args:\n for f in dataio.I3File(a):\n yield f\n\ndef main():\n from optparse import OptionParser,OptionGroup\n \n usage = 'usage: %prog [options] input_file [input_file2 ...] output_file\\n\\n'\n usage += 'Strip frame types from an i3 file.'\n parser = OptionParser(usage=usage)\n parser.add_option('-g','--geometry',default=False,action='store_true',\n help='Strip Geometry frames')\n parser.add_option('-c','--calibration',default=False,action='store_true',\n help='Strip Calibration frames')\n parser.add_option('-d','--detector',default=False,action='store_true',\n help='Strip Detector frames')\n parser.add_option('--no-trayinfo',default=True,action='store_false',\n dest='trayinfo',help='Do not strip TrayInfo frames')\n \n (options,args) = parser.parse_args()\n \n if len(args) < 2:\n print('ERROR: require an input and output file')\n print('')\n parser.print_help()\n \n elif os.path.exists(args[-1]):\n print('ERROR: output file',args[-1],'already exists!')\n print('')\n parser.print_help()\n \n else:\n count = 0\n outfile = dataio.I3File(args[-1],'w')\n try:\n for frame in files(args[:-1]):\n if options.geometry and frame.Stop == icetray.I3Frame.Geometry:\n continue\n elif options.calibration and frame.Stop == icetray.I3Frame.Calibration:\n continue\n elif options.detector and frame.Stop == icetray.I3Frame.Detector:\n continue\n elif options.trayinfo and frame.Stop == icetray.I3Frame.TrayInfo:\n continue\n outfile.push(frame)\n finally:\n outfile.close()\n\nif __name__ == '__main__':\n main()\n","sub_path":"dataio/resources/examples/strip.py","file_name":"strip.py","file_ext":"py","file_size_in_byte":2194,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"81834460","text":"import FWCore.ParameterSet.Config as cms\nfrom RecoLocalCalo.HGCalRecProducers.HGCalRecHit_cfi import dEdX_weights as dEdX\n\nHGCalPhotonIDValueMap = cms.EDProducer(\"HGCalPhotonIDValueMapProducer\",\n photons = cms.InputTag(\"photonsFromMultiCl\"),\n pcaRadius = cms.double(3.),\n EERecHits = cms.InputTag('HGCalRecHit:HGCEERecHits'),\n FHRecHits = cms.InputTag('HGCalRecHit:HGCHEFRecHits'),\n BHRecHits = cms.InputTag('HGCalRecHit:HGCHEBRecHits'),\n dEdXWeights = dEdX,\n)\n","sub_path":"EgammaAnalysis/python/HGCalPhotonIDValueMap_cfi.py","file_name":"HGCalPhotonIDValueMap_cfi.py","file_ext":"py","file_size_in_byte":479,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"285735845","text":"str=\"Mr John Smith \"\nnewstr=str.strip()\nn=len(newstr)\nres=[0]*n\nprint(res)\nfor i in range(0,n):\n if ' ' in newstr[i] :\n res[i]='%20'\n else:\n res[i] = newstr[i]\nprint(''.join(res))\n","sub_path":"Strings/replaceWith%20.py","file_name":"replaceWith%20.py","file_ext":"py","file_size_in_byte":202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"152646669","text":"import click\nfrom cryptography.hazmat.backends import default_backend\nfrom cryptography.hazmat.primitives.ciphers import (\n Cipher, algorithms, modes\n)\nfrom passlib.hash import bcrypt\n\n\ndef encrypt(key, plaintext, associated_data):\n # Generate a random 96-bit IV.\n iv = b'HQ\\xd9\\xb3Kz\\n\\xcc\\xb224Q\\xdb\\xc7u\\xb7' # os.urandom(12)\n\n # Construct an AES-GCM Cipher object with the given key and a\n # randomly generated IV.\n encryptor = Cipher(\n algorithms.AES(key),\n modes.GCM(iv),\n backend=default_backend()\n ).encryptor()\n\n # associated_data will be authenticated but not encrypted,\n # it must also be passed in on decryption.\n encryptor.authenticate_additional_data(associated_data)\n\n # Encrypt the plaintext and get the associated ciphertext.\n # GCM does not require padding.\n ciphertext = encryptor.update(plaintext) + encryptor.finalize()\n\n return iv, ciphertext, encryptor.tag\n\n\ndef decrypt(key, associated_data, iv, ciphertext, tag):\n # Construct a Cipher object, with the key, iv, and additionally the\n # GCM tag used for authenticating the message.\n decryptor = Cipher(\n algorithms.AES(key),\n modes.GCM(iv, tag),\n backend=default_backend()\n ).decryptor()\n\n # We put associated_data back in or the tag will fail to verify\n # when we finalize the decryptor.\n decryptor.authenticate_additional_data(associated_data)\n\n # Decryption gets us the authenticated plaintext.\n # If the tag does not match an InvalidTag exception will be raised.\n return decryptor.update(ciphertext) + decryptor.finalize()\n\n\ndef hash(value, salt):\n if salt == \"\":\n raise Exception(\"You need to specify salt (at least 1 character).\")\n return bcrypt.using(rounds=12, salt=salt.ljust(22, \"x\")).hash(value)\n\n\ndef correctness_hash(*strings, fake=False):\n if not fake:\n return bcrypt.using(rounds=12).hash(''.join(map(str, strings)))\n return bcrypt.using(rounds=12).hash(''.join(map(str, strings)) + \"1\")\n\n\ndef check_correctness_hash(query_result, *keys):\n for item in query_result:\n secret = \"\".join(str(item[key]) for key in keys)\n if not bcrypt.verify(secret, item[\"correctness_hash\"]):\n click.echo(f\"{item} failed correctness hash test!\")\n","sub_path":"client/crypto_utils.py","file_name":"crypto_utils.py","file_ext":"py","file_size_in_byte":2283,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"288149996","text":"\"\"\"Module with psapoc views and html classes.\n\"\"\"\n\nfrom django import forms\nfrom django.shortcuts import render, get_object_or_404, get_list_or_404\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.core.urlresolvers import reverse\nfrom django.views import generic\nfrom django.forms import ModelForm, BaseForm\nfrom django.forms.formsets import formset_factory\nfrom django_tables2 import RequestConfig\nfrom django.db.models import Avg, Max, Min, Count, Sum\nimport django_tables2 as tables\nimport datetime\nfrom django.contrib.auth.decorators import login_required\nfrom psapoc.tables import *\nfrom psapoc.forms import *\nfrom psapoc.models import *\nfrom django.contrib import messages\n\n\n@login_required\ndef projectlist(request):\n\t\"\"\"View to generate a project list table and render to html with template.\"\"\"\n\tq = Project.objects.all()\n\ttable = ProjectTable(q)\n\ttable.exclude = 'alternative_billto','alternative_payer','quote_currency','dream_package','project_manager'\n\ttable.sequence = 'id','view','manage','name','customer','delivery_region','quoted_amount','_revenue','_margin'\n\t\n\t# Enable pagination on table\n\tRequestConfig(request, paginate={\"per_page\": 25}).configure(table)\n\treturn render(request, 'psapoc/projectlist.html', {'table': table})\n\ndef home(request):\n\t\"\"\"View that redirects to the /psapoc/ root.\"\"\"\n\treturn HttpResponseRedirect('/psapoc/') \n\n@login_required\ndef projectfinancials(request, projectid):\n\t\"\"\"View that collect and renders project details with billable items and effort costing with template.\n\t\n\tinput : primary key for project (projectid)\n\toutput : template-rendered with 4 tables summarising project info\n\t\t- financial breakdown of baseline, revenue, cost, and resulting margin\n\t\t- list of fixed price items on the project, with aggregated revenue target\n\t\t- list of time-bade price items on the project, with aggregate revenue targets\n\t\t- all tasks associated to the project, with total time and cost booked on it\t\n\t\n\t\"\"\"\n\tproject = Project.objects.get(pk=projectid)\n\tprojectlist = Project.objects.filter(pk=projectid)\n \n\tfixbillselection = FixBillItem.objects.filter(project=projectid)\n\ttimedbillselection = TimedBillItem.objects.filter(project=projectid)\n\ttaskselection = Task.objects.filter(project=projectid)\n\n\t# convert to renderable tables\n\tfixtable1 = FixBillItemTable(fixbillselection, prefix='1-')\n\ttimedtable2 = TimedBillItemWithTimeTable(timedbillselection, prefix='2-')\n\ttasktable3 = TaskTable(taskselection,prefix='3-')\n\tprojectresultstable = ProjectResultsTable(projectlist,prefix='4-')\n\n\t# set database columns NOT to display\n\tfixtable1.exclude = 'project','completion','planned_date','completed_date','amount','currency','billed_date'\n\ttimedtable2.exclude = 'project','completion','planned_date','completed_date','billed_date'\n\ttasktable3.exclude = 'project','completion','planned_start_date','planned_end_date'\n\t\n\t# set the order of columns to display (db columns and calculated fields from table class)\n\tfixtable1.sequence = 'id','product','_revenue','invoice_text','billed'\n\ttimedtable2.sequence = 'id','product','_revenue','invoice_text','billed','task','_hours','_cost','billing_rate','_hourly_billing_rate','revised_hours'\n\ttasktable3.sequence = 'id','name','resource','_hours','_cost'\n\n\t# Set pagination on each table\n\tRequestConfig(request, paginate={\"per_page\": 10}).configure(fixtable1)\n\tRequestConfig(request, paginate={\"per_page\": 10}).configure(timedtable2)\n\tRequestConfig(request, paginate={\"per_page\": 10}).configure(tasktable3)\n#\tRequestConfig(request).configure(projecttable)\n\n\t# render with the template and pass tables with contents\n\treturn render(request, 'psapoc/projectfinancials.html', {'project':project,'table1': fixtable1,'table2':timedtable2,'table3':tasktable3,'projecttable':projectresultstable})\n\n\n\t# Timesheet form for current week, or week having given date\n@login_required\t\t\t\t\t\ndef TimeSheet(request, resourceid=0, daterange=None):\n\t\"\"\"View to generate timesheet to collect timesheet form input.\n\t\n\tUses currently logged in user for task selection\n\tPass last date for timeinput, defaults to today\n\t\"\"\"\n\t# Initialise parameters\n\t# resource ID of specified or currently logged user\n\tif resourceid == 0:\n\t\tresourceid=request.user.id\n\ttry:\n\t\tSelectedResource = Resource.objects.get(id=resourceid)\n\texcept:\n\t\t# error: specified resource does not exist. redirect to homepage with error message\n\t\treturn HttpResponseRedirect('/psapoc/')\n\t\n\t# if this is submitted timesheet, need to process the form data\n\tif request.method == 'POST':\n\t\tprint('got a POST')\n\t\tfor key, value in request.POST.items():\n\t\t\tprint('looping through post items')\n\t\t\tif ('task' in key) and ('day' in key) and (float(value) != 0): # processing task_$_day_$ field names only, skipping zeros\n\t\t\t\tprint('found a non zero value with task and day in the field label that needs saving')\n\t\t\t\tparsedkey = key.split('_')\n\t\t\t\tnewentry = TimeEntry()\n\t\t\t\tnewentry.task = Task.objects.get(id=parsedkey[1])\n\t\t\t\tnewentry.date = datetime.datetime.strptime(parsedkey[3], '%Y-%m-%d').date()\n\t\t\t\tnewentry.hours = float(value)\n\t\t\t\tnewentry.resource = Resource.objects.get(id=resourceid)\n\t\t\t\tnewentry.project = Task.objects.get(id=parsedkey[1]).project\n\t\t\t\tnewentry.save() # commit new record to DB\n\t\t\t\tmessages.success(request, 'Timesheet data for ' + key + ' added.')\n\t\treturn HttpResponseRedirect(request.path)\n\t# if this is not submission, then create a blank timesheet form for all tasks assigned to the resource\n\telse:\n\t\tformset = [] # dict where each row in the form gets separate item\n\t\t\n\t\theader = TaskTimeSheetForm(daterange)\n\t\theader.createcolumnheaders()# generates row with column headers\n\t\tpreviousm = header.previousmonday() # date passed in the template for a toggle to previous week\n\t\tnextm = header.nextmonday() # date passed in the template for a toggle to next week\n\t\tformset.append(header) # first row in the set\n\t\t\n\t\tfor atask in Task.objects.filter(resource=resourceid, completion__lt=100): # generate new row with fields for each task\n\t\t\ttaskline = TaskTimeSheetForm(daterange)\n\t\t\ttaskline.createtaskrow(atask)\n\t\t\tformset.append(taskline)\n\t\t\n\t\tsummary = TaskTimeSheetForm(daterange)\n\t\tsummary.createsummaryrow(resourceid)\n\t\tformset.append(summary)\n\t\t\n\t\treturn render(request, 'psapoc/timesheet.html',{'formset':formset,'resource':SelectedResource,'previous':previousm.strftime('%Y-%m-%d'),'next':nextm.strftime('%Y-%m-%d')})\n\t\n# unused view, attempt at creqting a model-based form\n@login_required\ndef dosomething(request, project_id):\n\t\"\"\"View to create a model-based form\n\tobsolete; not used in this project\n\t\n\tprocess form input, and if absent, generate new blank form from Project db model\"\"\"\n\t# if this is a POST request we need to process the form data\n\tif request.method == 'POST':\n\t\t# create a form instance and populate it with data from the request:\n\t\tform = ProjectForm(request.POST)\n\t\t# check whether it's valid:\n\t\tif form.is_valid():\n\t\t\tform.save()\n\n\t# if a GET (or any other method) we'll create a blank form\n\telse:\n\t\tform = ProjectForm()\n\treturn render(request, 'psapoc/dosomething.html',{'form':form})\n","sub_path":"Proof_of_concept/psapoc/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":7092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"137167742","text":"import base64\nfrom Crypto.Cipher import AES\n\nKEY = '1234567890123456'\n\n\n# AES加密算法的作用是替代原先的DES加密算法,该加密算法是对称加密算法,即加解密双方持有的KEY是相同的。\n# 密钥长度的支持为128、192、256(分别对应16字节、24字节、32字节),分组长度128位(16字节)。\n# 密钥可随机生成,这里作为演示的KEY为16字节,但是实际生产中,最好使用32字节的KEY。\n\nclass AESCipher(object):\n '''\n AES加密算法是密码学中的高级加密标准,又称Rijndael加密法。是一种区块加密技术,已然成为对称密钥加密中最流行的算法之一。\n\n 该示例采用 AES-ECB 模式(其余还有CBC、CFB、OFB三种加密模式),PKCS7 补码方法,进行加解密。\n '''\n\n def __init__(self, key):\n self.bs = 16 # 分组长度固定为16字节,若为其他数值,则加解密必然失败。\n self.key = key\n\n def aes_encrypt(self, raw):\n '''\n 加密算法\n\n :param raw:\n :return:\n '''\n raw = self._pad(raw) # 先进行补位,使加密数据的长度为16字节的整数倍。\n iv = b''\n cipher = AES.new(self.key, AES.MODE_ECB, iv)\n return base64.b64encode(iv + cipher.encrypt(raw))\n\n def aes_decrypt(self, enc):\n '''\n 解密算法\n\n :param enc:\n :return:\n '''\n enc = base64.b64decode(enc)\n iv = b''\n cipher = AES.new(self.key, AES.MODE_ECB, iv)\n return self._unpad(cipher.decrypt(enc)) # 去掉补位,还原数据\n\n def _pad(self, s, charset='ascii'):\n '''\n 补位算法\n\n :param s:\n :param charset:\n :return:\n '''\n if isinstance(s, str):\n return s + (self.bs - len(s) % self.bs) * chr(self.bs - len(s) % self.bs)\n if isinstance(s, (bytes, bytearray)):\n return s + bytes((self.bs - len(s) % self.bs) * chr(self.bs - len(s) % self.bs), encoding=charset)\n\n def _unpad(self, s):\n '''\n 去补位算法\n\n :param s:\n :return:\n '''\n return s[:-ord(s[len(s) - 1:])]\n\n\nif __name__ == \"__main__\":\n aes = AESCipher(KEY)\n\n data = '阿萨德' * 200\n # data = b'123456789' * 200\n en_str = aes.aes_encrypt(data) # 进行加密\n de_str = aes.aes_decrypt(en_str) # 进行解密\n\n if isinstance(data, str):\n print(de_str.decode('utf-8') == data)\n if isinstance(data, (bytes, bytearray)):\n print(de_str == data)\n","sub_path":"new_p6_network_security/AES加密与解密.py","file_name":"AES加密与解密.py","file_ext":"py","file_size_in_byte":2557,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"95332639","text":"import os, sys\nsys.path.append(\"..\")\nsys.path.append('../DSB2017')\nfrom DSB2017.main import inference\n\nfrom app.base import blueprint\nfrom app.base.forms import AnonymousForm\nfrom flask import flash, render_template, redirect, request, url_for, current_app\nfrom werkzeug.utils import secure_filename\n\n\n# default page\n@blueprint.route('/')\ndef route_default():\n return redirect(url_for('base_blueprint.home'))\n\n\n@blueprint.route('/home')\ndef home():\n return render_template('homepage/home.html', title='Home')\n\n\n@blueprint.route('/about')\ndef about():\n return render_template('homepage/about.html', title='About')\n\n\n@blueprint.route('/contact')\ndef contact():\n return render_template('homepage/contact.html', title='Contact')\n\n\n@blueprint.route('/upload', methods=['GET', 'POST'])\ndef upload():\n form = AnonymousForm()\n\n if form.cancel.data:\n return redirect(request.url)\n\n if form.submit.data and form.validate_on_submit():\n raw_file = form.raw_file.data\n mhd_file = form.mhd_file.data\n\n if raw_file and mhd_file:\n raw_file_name = secure_filename(raw_file.filename)\n mhd_file_name = secure_filename(mhd_file.filename)\n\n raw_path = os.path.join(current_app.config['UPLOAD_FOLDER'], raw_file_name)\n mhd_path = os.path.join(current_app.config['UPLOAD_FOLDER'], mhd_file_name)\n\n raw_file.save(raw_path)\n mhd_file.save(mhd_path)\n\n import time\n # time.sleep(5)\n\n inference(mhd_path)\n\n return redirect(url_for('base_blueprint.result', raw_file_name=raw_file_name, mhd_file_name=mhd_file_name))\n\n return render_template('homepage/upload.html', title=\"Upload\", form=form)\n\n\n@blueprint.route('/result//', methods=['GET', 'POST'])\ndef result(raw_file_name, mhd_file_name):\n raw = url_for('static', filename=f'uploaded_ct_scan/{raw_file_name}')\n mhd = url_for('static', filename=f'uploaded_ct_scan/{mhd_file_name}')\n\n return render_template('homepage/result.html', title=\"Upload\", ct_scan_files=[raw, mhd])\n","sub_path":"lung/app/base/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":2106,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"83702830","text":"import logging\nlogger = logging.getLogger(__name__)\n\n\nclass ProcessBot:\n\n # def __init__(self, content):\n @property\n def __init__(self):\n\n __version__ = '0.1.0'\n __name__ = 'ProcessBot'\n\n logger.info(\" {0}: {1} \".format(__name__, __version__))\n\n # self.content = content\n\n @property\n def processSearch(self, content):\n for result in content:\n print(result)\n","sub_path":"kollekt/process/ProcessBot.py","file_name":"ProcessBot.py","file_ext":"py","file_size_in_byte":418,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"629846456","text":"#!/usr/bin/env python\nimport sys\nimport time\n\nfrom colorama import Fore\nfrom tf.transformations import euler_from_quaternion\nimport rospy\nimport tf2_ros\nimport geometry_msgs\nfrom geometry_msgs.msg import Twist\nimport math\nimport tf_conversions\nfrom geometry_msgs.msg import PoseStamped\nfrom gazebo_msgs.msg import ModelStates\nfrom sensor_msgs.msg import Image\nfrom cv_bridge import CvBridge\nimport cv2\nimport numpy as np\nfrom sensor_msgs.msg import LaserScan\nimport sys\n\n\nclass Player:\n\n def __init__(self):\n\n rospy.init_node('driver', anonymous=False)\n name = rospy.get_name().strip('/')\n rospy.sleep(0.2) # make sure the rospy time works\n\n self.pub = rospy.Publisher(name + '/cmd_vel', Twist, queue_size=10)\n rospy.Subscriber(name + \"/camera/rgb/image_raw\", Image, self.ImageCallback)\n rospy.Subscriber(name + \"/scan\", LaserScan, self.lidar_callback)\n\n\n #-------------------------------------------------------------------\n #---------------------Variable Initialization-----------------------\n #-------------------------------------------------------------------\n\n # make sure that the code only works it there is image\n self.exist_image = 0\n\n # connectivity between pixels\n self.connectivity = 4\n\n self.my_team = None\n self.index_color = {'blue': 0, 'green': 1, 'red': 2}\n\n self.blue_limits = {'B': {'max': 255, 'min': 100}, 'G': {'max': 50, 'min': 0}, 'R': {'max': 50, 'min': 0}}\n self.red_limits = {'B': {'max': 50, 'min': 0}, 'G': {'max': 50, 'min': 0}, 'R': {'max': 255, 'min': 100}}\n self.green_limits = {'B': {'max': 50, 'min': 0}, 'G': {'max': 255, 'min': 100}, 'R': {'max': 50, 'min': 0}}\n\n self.size_bool = False\n self.list_centroid=[]\n self.state='wait'\n self.substate=None\n self.rate = rospy.Rate(10) # 10hz\n\n # -------------------------------------------------------------------\n # ---------------------Read Parameters-------------------------------\n # -------------------------------------------------------------------\n\n names_red = rospy.get_param('/red_players')\n names_green = rospy.get_param('/green_players')\n names_blue = rospy.get_param('/blue_players')\n\n\n\n\n\n\n if name in names_red:\n self.my_team = 'red'\n self.hunter_team = 'blue'\n self.prey_team = 'green'\n self.my_team_players = names_red\n self.prey_team_players = names_green\n self.hunter_team_players = names_blue\n elif name in names_green:\n self.my_team = 'green'\n self.hunter_team = 'red'\n self.prey_team = 'blue'\n self.my_team_players = names_green\n self.prey_team_players = names_blue\n self.hunter_team_players = names_red\n elif name in names_blue:\n self.my_team = 'blue'\n self.hunter_team = 'green'\n self.prey_team = 'red'\n self.my_team_players = names_blue\n self.prey_team_players = names_red\n self.hunter_team_players = names_green\n else:\n rospy.logfatal('Something is wrong \\n You have been suspended from your own team')\n exit(0)\n\n print('My name is ' + name + ' I am team ' + self.my_team + Fore.GREEN + ' I am hunting ' + Fore.RESET + str(\n self.prey_team_players)\n + Fore.RED + ' and fleeing from ' + Fore.RESET + str(self.hunter_team_players))\n\n\n\n # -------------------------------------------------------------------\n # ---------------------Callbacks Functions---------------------------\n # -------------------------------------------------------------------\n\n\n\n rospy.Timer(rospy.Duration(0.1), self.print_state, oneshot=False)\n\n\n\n def ImageCallback(self,message):\n\n bridge = CvBridge()\n self.cv_image = bridge.imgmsg_to_cv2(message, desired_encoding='bgr8')\n\n self.blue_mask = cv2.inRange(self.cv_image, (self.blue_limits['B']['min'], self.blue_limits['G']['min'], self.blue_limits['R']['min']),\n (self.blue_limits['B']['max'], self.blue_limits['G']['max'], self.blue_limits['R']['max']))\n\n self.red_mask = cv2.inRange(self.cv_image, (self.red_limits['B']['min'], self.red_limits['G']['min'], self.red_limits['R']['min']),\n (self.red_limits['B']['max'], self.red_limits['G']['max'], self.red_limits['R']['max']))\n\n self.green_mask = cv2.inRange(self.cv_image, (self.green_limits['B']['min'], self.green_limits['G']['min'], self.green_limits['R']['min']),\n (self.green_limits['B']['max'], self.green_limits['G']['max'], self.green_limits['R']['max']))\n\n self.exist_image=1\n\n\n def lidar_callback(self,message):\n\n angle = message.angle_min\n\n #print(message.ranges[0])\n\n if message.ranges[0]<1.3:\n self.substate=\"escape_wall\"\n else:\n self.substate=None\n\n def get_centroid(self):\n\n if self.exist_image:\n\n if self.size_bool == False:\n self.height = self.cv_image.shape[0]\n self.width = self.cv_image.shape[1]\n self.size_bool = True\n\n\n mask_list = [self.blue_mask,self.green_mask,self.red_mask]\n\n for_index=0\n list_biggest_target=[]\n self.list_centroid=[]\n\n\n for mask in mask_list:\n\n output = cv2.connectedComponentsWithStats(mask, self.connectivity, cv2.CV_32S)\n num_labels = output[0] # integer with the number of object in the image\n labels = output[1] # in labels we have an image, and each element has a value equivalent to its label\n stats = output[2] # in stats we have all data for each object\n centroids = output[3] # in centroids we have all centroids coordinates for each object\n\n # finding the object with bigger area\n anyone = True\n maximum_area = 0\n object_index = 1\n # if num_labels == 1 means that there is no object, so we cannot paint!\n if num_labels == 1:\n anyone = False\n for i in range(1, num_labels):\n\n object_area = stats[i, cv2.CC_STAT_AREA]\n\n if object_area > maximum_area:\n maximum_area = object_area\n object_index = i\n\n # if maximum_area <500 the object is too small, so its possible that it is not the phone but noise instead\n if maximum_area < 20:\n anyone = False\n # extracting biggest object from segmentation limits\n biggest_target = (labels == object_index)\n biggest_target = biggest_target.astype(np.uint8) * 255\n list_biggest_target.append(biggest_target)\n\n if anyone:\n centroid_coord = centroids[object_index, :].astype(np.uint)\n centroid_coord = tuple(centroid_coord)\n self.list_centroid.append(centroid_coord)\n else:\n self.list_centroid.append(None)\n\n for_index+=1\n\n\n #Algoritmo de decisao\n\n self.DecisionMaking()\n\n\n def DecisionMaking(self):\n\n\n\n if self.list_centroid[self.index_color[self.hunter_team]]!=None and self.list_centroid[self.index_color[self.prey_team]]==None:\n # Caso eu veja um atacante meu e nao veja presas, FUGIR\n self.state='flee'\n elif self.list_centroid[self.index_color[self.hunter_team]]!=None and self.list_centroid[self.index_color[self.prey_team]]!=None:\n #Se eu vir os 2, tenho de tomar uma decisao dependendo da distancia entre eles\n\n\n\n distance_prey_hunter=abs(float(self.list_centroid[self.index_color[self.hunter_team]][0])-float(self.list_centroid[self.index_color[self.prey_team]][0]))\n #print(distance_prey_hunter)\n if distance_prey_hunter > self.width/3:\n self.state='atack'\n else:\n self.state='flee'\n elif self.list_centroid[self.index_color[self.hunter_team]]==None and self.list_centroid[self.index_color[self.prey_team]]!=None:\n self.state = 'atack'\n elif self.list_centroid[self.index_color[self.hunter_team]]==None and self.list_centroid[self.index_color[self.prey_team]]==None:\n self.state = 'wait'\n\n self.take_action()\n\n\n def take_action(self):\n\n twist = Twist()\n\n if self.state=='atack':\n\n if self.list_centroid[self.index_color[self.prey_team]]!=None:\n\n horizontal_distance=self.find_direction(self.list_centroid[self.index_color[self.prey_team]])\n #print(horizontal_distance)\n\n twist.linear.x = 1.0\n twist.angular.z = horizontal_distance/500\n\n elif self.state=='flee':\n\n horizontal_distance = self.find_direction(self.list_centroid[self.index_color[self.hunter_team]])\n #print(horizontal_distance)\n\n if horizontal_distance>0:\n signal = -1\n else:\n signal = 1\n\n twist = Twist()\n\n twist.linear.x = 1.5\n twist.angular.z = signal*1.5\n\n\n\n else:\n\n if self.substate==\"escape_wall\":\n #fugir da parede\n twist.linear.x = 0.2\n twist.angular.z = 2\n\n else:\n twist.linear.x = 0.6\n twist.angular.z = 0.3\n\n self.pub.publish(twist)\n\n\n def find_direction(self,centroid_coord):\n\n #if > 0 turn left\n #if < 0 turn right\n return (self.width/2) - centroid_coord[0]\n\n def print_state(self,event):\n\n if self.state == 'flee':\n print(Fore.RED + self.state + ' from ' + self.hunter_team + Fore.RESET)\n elif self.state == 'atack':\n print(Fore.GREEN + self.state + ' and kill ' + self.prey_team + Fore.RESET)\n else:\n if self.substate==None:\n print(Fore.BLUE + self.state + ' for my next prey ' + Fore.RESET)\n else:\n print(Fore.BLUE +'Too close of a wall, better turn around' + Fore.RESET)\n\n\n\n\n\ndef main():\n\n player = Player()\n\n\n while not rospy.is_shutdown():\n\n player.get_centroid()\n\n player.rate.sleep()\n\n\n\n # ---------------\n # program's end\n # ---------------\n cv2.destroyAllWindows()\n\nif __name__ == '__main__':\n main()\n","sub_path":"p_g5_core/src/driver_camera_lidar.py","file_name":"driver_camera_lidar.py","file_ext":"py","file_size_in_byte":10670,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"143761664","text":"from django.contrib.auth.decorators import login_required\nfrom django.urls import path\nfrom .views import contact, ContactCreate, ContactUpdate, ContactDelete, ContactList, ContactDetail\n\napp_name='contact'\nurlpatterns = [\n\tpath('', contact, name = 'contact'),\n\n\tpath('messages/', login_required(ContactList.as_view()), name = 'contact-messages'),\n\tpath('messages//', login_required(ContactDetail.as_view()), name = 'contact-update'),\n\tpath('messages/add/', login_required(ContactCreate.as_view()), name = 'contact-create'),\n\tpath('messages//update/', login_required(ContactUpdate.as_view()), name = 'contact-update'),\n\tpath('messages//delete/', login_required(ContactDelete.as_view()), name = 'contact-delete'),\n]\n","sub_path":"contact/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"398841044","text":"import json\nfrom enum import Enum\n\nimport marshmallow\nfrom marshmallow import fields as mf\n\nfrom typing import Any, Dict, Final, List, NoReturn, Union\n\n\nActorId = int\n\nVERSION = (0, 0, 2)\nPORT_BROADCAST = 54300\nPORT_DATA = 54301\nINVENTORY_SIZE: int = 20\nSERIALIZATION_TYPE_FIELD = \"_type\"\nUNASSIGNED_ACTOR_ID: Final[ActorId] = -1\n\n\nclass Hand(Enum):\n LEFT = 0\n RIGHT = 1\n\n def other(self):\n if self == Hand.LEFT:\n return Hand.RIGHT\n else:\n return Hand.LEFT\n\n\nclass DamageVariant(Enum):\n HIT = \"hit\"\n CHOP = \"chop\"\n SMASH = \"smash\"\n ATTACK = \"attack\"\n\n\nclass UpdateVariant(Enum):\n \"\"\"Generic enum describing type of item stack update.\"\"\"\n\n SWAP = 0\n MERGE = 1\n\n\nclass Attachement(Enum):\n LEFT_ITEM = \"left_item\"\n RIGHT_ITEM = \"right_item\"\n\n\nclass Stats:\n class Schema(marshmallow.Schema):\n hunger = mf.Float()\n max_hunger = mf.Float()\n\n @marshmallow.post_load\n def make(self, data, **kwargs):\n return Stats(**data)\n\n def __init__(self, hunger: float, max_hunger: float) -> None:\n self.hunger = hunger\n self.max_hunger = max_hunger\n\n\nclass Debugable:\n DEBUG_FIELDS: List[str] = list()\n\n def __str__(self) -> str:\n fields = {field: str(getattr(self, field)) for field in self.DEBUG_FIELDS}\n return f\"{type(self).__name__}{fields}\"\n\n\nclass Serializable:\n SERIALIZATION_NAME: str = \"___\"\n\n class Schema(marshmallow.Schema):\n pass\n\n def to_dict(self) -> Dict[str, Any]:\n result = self.Schema().dump(self)\n result[SERIALIZATION_TYPE_FIELD] = self.SERIALIZATION_NAME\n return result\n\n def to_json(self) -> str:\n return json.dumps(self.to_dict())\n\n\ndef assert_exhaustive(x: NoReturn) -> NoReturn:\n assert False, \"Unhandled type: {}\".format(type(x).__name__)\n","sub_path":"python/edgin_around_api/defs.py","file_name":"defs.py","file_ext":"py","file_size_in_byte":1849,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"138486123","text":"#!/usr/bin/env python\n#***************************************************\n# * Title: UUV Simulator\n# * Author: The UUV Simulator Authors\n# * Date: 2016\n# * Availability: https://uuvsimulator.github.io/\n#***************************************************\n\nimport rospy\nimport numpy as np\nfrom control_interfaces import DPPIDControllerBase\n\n\nclass ROV_PIDController(DPPIDControllerBase):\n \"\"\"PID controller for the dynamic positioning of ROVs.\"\"\"\n\n _LABEL = 'PID'\n def __init__(self):\n self._tau = np.zeros(6)\n DPPIDControllerBase.__init__(self, False)\n self._is_init = True\n\n def update_controller(self):\n if not self._is_init:\n return False\n # Update PID control action\n self._tau = self.update_pid()\n self.publish_control_wrench(self._tau)\n return True\n\nif __name__ == '__main__':\n print('Starting PID')\n rospy.init_node('rov_pid_controller')\n\n try:\n node = ROV_PIDController()\n rospy.spin()\n except rospy.ROSInterruptException:\n print('caught exception')\n print('exiting')\n","sub_path":"control/control/trajectory_control/scripts/rov_pid_controller.py","file_name":"rov_pid_controller.py","file_ext":"py","file_size_in_byte":1092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"571197014","text":"import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport scipy.signal\nfrom scipy.fftpack import fft, fftshift\n\ndef main():\n create_sma_plots()\n create_window_comparison_plots()\n\ndef create_sma_plots():\n # Equations from https://tttapa.github.io/Pages/Mathematics/Systems-and-Control-Theory/Digital-filters/Simple%20Moving%20Average/Simple-Moving-Average.html#:~:text=The%20cutoff%20frequency%20is%20defined,)%20%E2%89%88%20%E2%88%92%203.01%20d%20B%20.\n\n # Function for calculating the cut-off frequency of a moving average filter\n def get_sma_cutoff(N, **kwargs):\n func = lambda w: np.sin(N*w/2) - N/np.sqrt(2) * np.sin(w/2) # |H(e^jω)| = √2/2\n deriv = lambda w: np.cos(N*w/2) * N/2 - N/np.sqrt(2) * np.cos(w/2) / 2 # dfunc/dx\n omega_0 = np.pi/N # Starting condition: halfway the first period of sin(Nω/2)\n return scipy.optimize.newton(func, omega_0, deriv, **kwargs)\n\n N = 10 # Window size (number of samples)\n fs_Hz = 1000 # Sampling frequency\n f = np.linspace(0, fs_Hz/2, 1000)\n w = 2*np.pi*(f/fs_Hz)\n with np.errstate(divide='ignore', invalid='ignore'):\n freq_response = (1/(N**2))*((np.sin(w*N/2)**2)/(np.sin(w/2)**2))\n\n freq_response_dB = 10*np.log10(freq_response)\n\n # SMA coefficients\n b = np.ones(N)\n a = np.array([N] + [0]*(N-1))\n w, h = scipy.signal.freqz(b, a, worN=4096)\n f = (w*fs_Hz)/(2*np.pi) # Convert from rad/sample to Hz\n freq_response_dB = 20*np.log10(abs(h))\n phase_response = np.angle(h)*(180/np.pi)\n\n w_c = get_sma_cutoff(N)\n f_c_Hz = w_c * fs_Hz / (2 * np.pi)\n # print(f'f_c_Hz={f_c_Hz}')\n\n fig, axes = plt.subplots(1, 1, figsize=(10, 7), squeeze=False)\n ax = axes[0][0]\n ax.plot(f, freq_response_dB, label='Frequency response')\n ax.axvline(fs_Hz/N, color='C1', linestyle='--', label='Window frequency')\n ax.axvline(f_c_Hz, color='C2', linestyle='--', label='Cutoff frequency')\n ax.set_xlabel('Frequency (Hz)')\n ax.set_ylabel('Magnitude (dB)')\n ax.set_ylim(-50, 10)\n ax.set_title('Magnitude Response Of SMA')\n ax.legend()\n plt.tight_layout()\n plt.savefig('frequency-response-of-sma-magnitude.png')\n\n fig, axes = plt.subplots(1, 1, figsize=(10, 7), squeeze=False)\n ax = axes[0][0]\n ax.plot(f, phase_response, label='Phase response')\n ax.axvline(fs_Hz/N, color='C1', linestyle='--', label='Window frequency')\n ax.axvline(f_c_Hz, color='C2', linestyle='--', label='Cutoff frequency')\n ax.set_xlabel('Frequency (Hz)')\n ax.set_ylabel('Phase (°)')\n ax.set_ylim(-180, 90)\n ax.set_yticks([-180, -135, -90, -45, 0, 45, 90])\n ax.set_title('Phase Response Of SMA')\n ax.legend()\n plt.tight_layout()\n plt.savefig('frequency-response-of-sma-phase.png')\n\ndef create_window_comparison_plots():\n N = 51\n\n window_data = [\n {\n 'name': 'Boxcar',\n 'values': scipy.signal.boxcar(N),\n },\n {\n 'name': 'Exponential',\n 'values': scipy.signal.windows.exponential(N, tau=3.0),\n },\n {\n 'name': 'Gaussian',\n 'values': scipy.signal.windows.gaussian(N, std=7),\n },\n {\n 'name': 'Blackman',\n 'values': scipy.signal.blackman(N),\n },\n ]\n\n def plot_window(window, ax, label):\n ax.plot(window, label=label)\n\n def plot_freq_response(window, ax, label):\n with np.errstate(divide='ignore', invalid='ignore'):\n # fft() does not center the DC component, need to use fftshift() later\n # to do that\n # 2048 significantly larger than window size, so 0 padding will occur\n A = fft(window, 2048) / (len(window)/2.0)\n freq = np.linspace(-0.5, 0.5, len(A)) # This is normalized frequency (w.r.t sampling frequency)\n response = 20 * np.log10(np.abs(fftshift(A / abs(A).max())))\n ax.plot(freq, response, label=label)\n\n fig, axes = plt.subplots(1, 1, figsize=(10, 7), squeeze=False)\n ax = axes[0][0]\n for window in window_data:\n plot_window(window['values'], ax, window['name'])\n ax.set_title(\"Popular window shapes, N=51\")\n ax.set_xlabel('Sample')\n ax.set_ylabel('Weight')\n ax.legend()\n plt.tight_layout()\n plt.savefig('window-comparison-shapes.png')\n\n fig, axes = plt.subplots(1, 1, figsize=(10, 7), squeeze=False)\n ax = axes[0][0]\n for window in window_data:\n plot_freq_response(window['values'], ax, window['name'])\n\n ax.axis([0, 0.5, -120, 0])\n ax.set_title(\"Frequency response of popular windows, N=51\")\n ax.set_ylabel(\"Normalized magnitude [dB]\")\n ax.set_xlabel(\"Normalized frequency [cycles per sample]\")\n ax.legend()\n plt.tight_layout()\n plt.savefig('window-comparison-frequency-response.png')\n\nif __name__ == '__main__':\n main()","sub_path":"content/programming/signal-processing/digital-filters/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4832,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"31597131","text":"\"\"\"SNGAN ResNet for conditional generation of ImageNet\"\"\"\n\n# no ACGAN, 1\n# NoLabelConcatInG, 1\n# DECAY, 1\n# N_CRITIC = 5\n# biases=True\n\nimport os\nimport sys\n\nsys.path.append(os.getcwd())\n\nimport numpy as np\nimport tensorflow as tf\n\nimport time\nimport functools\nimport locale\n\nimport common.misc\nimport common.data.cifar10\nimport common.inception.inception_score\n\nimport common as lib\nimport common.ops.linear\nimport common.ops.conv2d\nimport common.ops.embedding\nimport common.ops.normalization\nimport common.plot\n\nfrom common.data import ILSVRC2012\n\n# Download CIFAR-10 (Python version) at\n# https://www.cs.toronto.edu/~kriz/cifar.html and fill in the path to the extracted files here!\nDATA_DIR = '/media/newhd/data/ILSVRC2012/train/'\nif len(DATA_DIR) == 0:\n raise Exception('Please specify path to data directory in gan_cifar.py!')\n\nBATCH_SIZE = 32 # Critic batch size\nGEN_BS_MULTIPLE = 2 # Generator batch size, as a multiple of BATCH_SIZE\nITERS = 450000 # How many iterations to train for\nDIM_G = 128 # Generator dimensionality\nDIM_D = 128 # Critic dimensionality\nNORMALIZATION_G = True # Use batchnorm in generator?\nNORMALIZATION_D = False # Use batchnorm (or layernorm) in critic?\nOUTPUT_DIM = 49152 # Number of pixels in CIFAR10 (128*128*3)\nLR = 0.0002 # 2e-4 # Initial learning rate\nDECAY = True # Whether to decay LR over learning\nN_CRITIC = 5 # 5 # Critic steps per generator steps\nINCEPTION_FREQUENCY = 1000 # How frequently to calculate Inception score\n\nCONDITIONAL = True # Whether to train a conditional or unconditional model\nACGAN = False # If CONDITIONAL, whether to use ACGAN or \"vanilla\" conditioning\nACGAN_SCALE = 1. # How to scale the critic's ACGAN loss relative to WGAN loss\nACGAN_SCALE_G = 0.1 # How to scale generator's ACGAN loss relative to WGAN loss\n\n# SPECTRAL_NORM_UPDATE_OPS = \"spectral_norm_update_ops\"\nWORD2VEC_FILE = None\nVOCAB_SIZE = 1000\nEMBEDDING_DIM = 300 # 620\nCHECKPOINT_DIR = os.path.join(DATA_DIR, 'checkpoint')\nLOSS_TYPE = 'HINGE' # 'Goodfellow', 'HINGE', 'WGAN', 'WGAN-GP'\nSOFT_PLUS = False\nRESTORE = False\n\nif CONDITIONAL and (not ACGAN) and (not NORMALIZATION_D):\n print(\"WARNING! Conditional model without normalization in D might be effectively unconditional!\")\n\nN_GPUS = 1\nif N_GPUS not in [1, 2]:\n raise Exception('Only 1 or 2 GPUs supported!')\nDEVICES = ['/gpu:{}'.format(i) for i in range(N_GPUS)]\nif len(DEVICES) == 1: # Hack because the code assumes 2 GPUs\n DEVICES = [DEVICES[0], DEVICES[0]]\n\nlib.print_model_settings(locals().copy())\n\n\ndef nonlinearity(x, activation_fn='relu', leakiness=0.2):\n if activation_fn == 'relu':\n return tf.nn.relu(x)\n if activation_fn == 'lrelu':\n assert 0 < leakiness <= 1, \"leakiness must be <= 1\"\n return tf.maximum(x, leakiness * x)\n\n\ndef Normalize(name, inputs, labels=None):\n \"\"\"This is messy, but basically it chooses between batchnorm, layernorm,\n their conditional variants, or nothing, depending on the value of `name` and\n the global hyperparam flags.\"\"\"\n\n with tf.variable_scope(name):\n if not CONDITIONAL:\n labels = None\n if CONDITIONAL and ACGAN and ('D.' in name):\n labels = None\n\n if ('D.' in name) and NORMALIZATION_D:\n return lib.ops.normalization.layer_norm(name, [1, 2, 3], inputs)\n elif ('G.' in name) and NORMALIZATION_G:\n if labels is not None:\n # inputs_ = tf.transpose(inputs, [0, 3, 1, 2], name='NHWC_to_NCHW')\n outputs = lib.ops.normalization.cond_batchnorm(name, [0, 1, 2], inputs, labels=labels, n_labels=1000)\n # return tf.transpose(outputs, [0, 2, 3, 1], name='NCHW_to_NHWC')\n return outputs\n else:\n # inputs_ = tf.transpose(inputs, [0, 3, 1, 2], name='NHWC_to_NCHW')\n outputs = lib.ops.normalization.batch_norm(inputs, fused=True)\n # return tf.transpose(outputs, [0, 2, 3, 1], name='NCHW_to_NHWC')\n return outputs\n else:\n return inputs\n\n\ndef ConvMeanPool(inputs, output_dim, filter_size=3, stride=1, name=None,\n spectral_normed=False, update_collection=None, inputs_norm=False,\n he_init=True, biases=True):\n output = lib.ops.conv2d.Conv2D(inputs, inputs.shape.as_list()[-1], output_dim, filter_size, stride, name,\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=he_init, biases=biases)\n # output = tf.nn.avg_pool(inputs, [1, 2, 2, 1], [1, 2, 2, 1], padding='VALID')\n output = tf.add_n(\n [output[:, ::2, ::2, :], output[:, 1::2, ::2, :], output[:, ::2, 1::2, :], output[:, 1::2, 1::2, :]]) / 4.\n return output\n\n\ndef MeanPoolConv(inputs, output_dim, filter_size=3, stride=1, name=None,\n spectral_normed=False, update_collection=None, inputs_norm=False,\n he_init=True, biases=True):\n output = inputs\n output = tf.add_n(\n [output[:, ::2, ::2, :], output[:, 1::2, ::2, :], output[:, ::2, 1::2, :], output[:, 1::2, 1::2, :]]) / 4.\n # output = tf.nn.avg_pool(inputs, [1, 2, 2, 1], [1, 2, 2, 1], padding='VALID')\n output = lib.ops.conv2d.Conv2D(output, output.shape.as_list()[-1], output_dim, filter_size, stride, name,\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=he_init, biases=biases)\n\n return output\n\n\ndef UpsampleConv(inputs, output_dim, filter_size=3, stride=1, name=None,\n spectral_normed=False, update_collection=None, inputs_norm=False,\n he_init=True, biases=True):\n output = inputs\n output = tf.concat([output, output, output, output], axis=3)\n output = tf.depth_to_space(output, 2)\n # w, h = inputs.shape.as_list()[1], inputs.shape.as_list()[2]\n # output = tf.image.resize_images(inputs, [w * 2, h * 2])\n output = lib.ops.conv2d.Conv2D(output, output.shape.as_list()[-1], output_dim, filter_size, stride, name,\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=he_init, biases=biases)\n\n return output\n\n\ndef ResidualBlock(inputs, input_dim, output_dim, filter_size, name,\n spectral_normed=False, update_collection=None, inputs_norm=False,\n resample=None, labels=None, biases=True):\n \"\"\"resample: None, 'down', or 'up'.\n \"\"\"\n if resample == 'down':\n conv_1 = functools.partial(lib.ops.conv2d.Conv2D, input_dim=input_dim, output_dim=input_dim)\n conv_2 = functools.partial(ConvMeanPool, output_dim=output_dim)\n conv_shortcut = ConvMeanPool\n elif resample == 'up':\n conv_1 = functools.partial(UpsampleConv, output_dim=output_dim)\n conv_shortcut = UpsampleConv\n conv_2 = functools.partial(lib.ops.conv2d.Conv2D, input_dim=output_dim, output_dim=output_dim)\n elif resample is None:\n conv_shortcut = functools.partial(lib.ops.conv2d.Conv2D, input_dim=input_dim)\n conv_1 = functools.partial(lib.ops.conv2d.Conv2D, input_dim=input_dim, output_dim=output_dim)\n conv_2 = functools.partial(lib.ops.conv2d.Conv2D, input_dim=output_dim, output_dim=output_dim)\n else:\n raise Exception('invalid resample value')\n\n if output_dim == input_dim and resample is None:\n shortcut = inputs # Identity skip-connection\n else:\n shortcut = conv_shortcut(inputs=inputs, output_dim=output_dim, filter_size=1, name=name + '.Shortcut',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=False, biases=biases)\n\n output = inputs\n output = Normalize(name + '.N1', output, labels=labels)\n output = nonlinearity(output)\n # if resample == 'up':\n # output = nonlinearity(output)\n # else:\n # output = lrelu(output, leakiness=0.2)\n\n output = conv_1(inputs=output, filter_size=filter_size, name=name + '.Conv1',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=True, biases=biases)\n\n output = Normalize(name + '.N2', output, labels=labels)\n output = nonlinearity(output)\n # if resample == 'up':\n # output = nonlinearity(output)\n # else:\n # output = lrelu(output, leakiness=0.2)\n\n output = conv_2(inputs=output, filter_size=filter_size, name=name + '.Conv2',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=True, biases=biases)\n\n return shortcut + output\n\n\ndef OptimizedResBlockDisc1(inputs,\n spectral_normed=False, update_collection=None, inputs_norm=False,\n biases=True):\n conv_1 = functools.partial(lib.ops.conv2d.Conv2D, input_dim=3, output_dim=DIM_D // 2)\n conv_2 = functools.partial(ConvMeanPool, output_dim=DIM_D // 2)\n conv_shortcut = MeanPoolConv\n shortcut = conv_shortcut(inputs=inputs, output_dim=DIM_D // 2, filter_size=1, name='D.Block.1.Shortcut',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=False, biases=biases)\n\n output = inputs\n output = conv_1(inputs=output, filter_size=3, name='D.Block.1.Conv1',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=True, biases=biases)\n output = nonlinearity(output)\n # output = lrelu(output, leakiness=0.2)\n output = conv_2(inputs=output, filter_size=3, name='D.Block.1.Conv2',\n spectral_normed=spectral_normed,\n update_collection=update_collection,\n he_init=True, biases=biases)\n return shortcut + output\n\n\ndef Generator(n_samples_, labels, noise=None, reuse=False):\n with tf.variable_scope(\"Generator\", reuse=reuse):\n if noise is None:\n noise = tf.random_normal([n_samples_, 128])\n\n output = lib.ops.linear.Linear(noise, 128, 4 * 4 * DIM_G * 8, 'G.Input')\n output = tf.reshape(output, [-1, 4, 4, DIM_G * 8])\n # 1024\n output = ResidualBlock(output, DIM_G * 8, DIM_G * 8, 3, 'G.Block.1', resample='up', labels=labels, biases=True)\n print('G.1: {}'.format(output.shape.as_list()))\n # 512\n output = ResidualBlock(output, DIM_G * 8, DIM_G * 4, 3, 'G.Block.2', resample='up', labels=labels, biases=True)\n print('G.2: {}'.format(output.shape.as_list()))\n # 256\n output = ResidualBlock(output, DIM_G * 4, DIM_G * 2, 3, 'G.Block.3', resample='up', labels=labels, biases=True)\n print('G.3: {}'.format(output.shape.as_list()))\n # 128\n output = ResidualBlock(output, DIM_G * 2, DIM_G, 3, 'G.Block.4', resample='up', labels=labels, biases=True)\n print('G.4: {}'.format(output.shape.as_list()))\n # 64\n output = ResidualBlock(output, DIM_G, DIM_G // 2, 3, 'G.Block.5', resample='up', labels=labels, biases=True)\n print('G.5: {}'.format(output.shape.as_list()))\n output = Normalize('G.OutputNorm', output, labels)\n output = nonlinearity(output)\n\n output = lib.ops.conv2d.Conv2D(output, DIM_G // 2, 3, 3, 1, 'G.Output', he_init=False)\n output = tf.tanh(output)\n print('G.output.shape: {}'.format(output.shape.as_list()))\n return tf.reshape(output, [-1, OUTPUT_DIM])\n # return tf.reshape(tf.transpose(output, [0, 3, 1, 2], name='NHWC_to_NCHW'), [-1, OUTPUT_DIM])\n\n\ndef Discriminator(inputs, labels, update_collection=None, reuse=False):\n with tf.variable_scope(\"Discriminator\", reuse=reuse):\n output = tf.reshape(inputs, [-1, 128, 128, 3])\n # output = tf.transpose(output, [0, 2, 3, 1], name='NCHW_to_NHWC')\n output = OptimizedResBlockDisc1(output,\n spectral_normed=True,\n update_collection=update_collection,\n biases=True)\n\n # output = ResidualBlock(output, 3, DIM_D // 2, 3, 'Discriminator.1',\n # spectral_normed=True,\n # update_collection=update_collection,\n # resample='down', labels=labels, biases=True)\n output = ResidualBlock(output, DIM_D // 2, DIM_D, 3, 'D.Block.2',\n spectral_normed=True,\n update_collection=update_collection,\n resample='down', labels=labels, biases=True)\n\n output = ResidualBlock(output, DIM_D, DIM_D * 2, 3, 'D.Block.3',\n spectral_normed=True,\n update_collection=update_collection,\n resample='down', labels=labels, biases=True)\n\n # embedding labels, and concatenate to 'output'.\n # (N, EMBEDDING_DIM)\n embedding_y = lib.ops.embedding.embed_y(labels, VOCAB_SIZE, EMBEDDING_DIM, word2vec_file=WORD2VEC_FILE)\n embedding_y = lib.ops.linear.Linear(embedding_y, EMBEDDING_DIM, DIM_D, 'D.Embedding_y',\n spectral_normed=True,\n update_collection=update_collection,\n biases=True) # (N, DIM_D)\n\n embedding_y = tf.expand_dims(tf.expand_dims(embedding_y, axis=1), axis=1)\n embedding_y = tf.tile(embedding_y, multiples=[1, output.shape.as_list()[1], output.shape.as_list()[2], 1])\n output = tf.concat(values=[output, embedding_y], axis=3)\n\n output = ResidualBlock(output, DIM_D * 3, DIM_D * 4, 3, 'D.Block.4',\n spectral_normed=True,\n update_collection=update_collection,\n resample='down', labels=labels, biases=True)\n output = ResidualBlock(output, DIM_D * 4, DIM_D * 8, 3, 'D.Block.5',\n spectral_normed=True,\n update_collection=update_collection,\n resample='down', labels=labels, biases=True)\n output = ResidualBlock(output, DIM_D * 8, DIM_D * 8, 3, 'D.Block.6',\n spectral_normed=True,\n update_collection=update_collection,\n resample=None, labels=labels, biases=True)\n output = nonlinearity(output)\n # output = lrelu(output, leakiness=0.2)\n output = tf.reduce_mean(output, axis=[1, 2])\n output_wgan = lib.ops.linear.Linear(output, DIM_D * 8, 1, 'D.Output',\n spectral_normed=True,\n update_collection=update_collection)\n output_wgan = tf.reshape(output_wgan, [-1])\n if CONDITIONAL and ACGAN:\n output_acgan = lib.ops.linear.Linear(output, DIM_D, 10, 'D.ACGANOutput',\n spectral_normed=True,\n update_collection=update_collection,\n biases=True)\n return output_wgan, output_acgan\n else:\n return output_wgan, None\n\n\n# with tf.Graph().as_default() as g:\nconfig = tf.ConfigProto(allow_soft_placement=True)\nconfig.gpu_options.allow_growth = True\nwith tf.Session(config=config) as session:\n _iteration = tf.placeholder(tf.int32, shape=None)\n all_real_data_int = tf.placeholder(tf.int32, shape=[BATCH_SIZE, OUTPUT_DIM])\n all_real_labels = tf.placeholder(tf.int32, shape=[BATCH_SIZE])\n\n labels_splits = tf.split(all_real_labels, len(DEVICES), axis=0)\n\n fake_data_splits = []\n for i, device in enumerate(DEVICES):\n with tf.device(device):\n if i > 0:\n fake_data_splits.append(Generator(int(BATCH_SIZE / len(DEVICES)), labels_splits[i], reuse=True))\n else:\n fake_data_splits.append(Generator(int(BATCH_SIZE / len(DEVICES)), labels_splits[i]))\n\n all_real_data = tf.reshape(2 * ((tf.cast(all_real_data_int, tf.float32) / 256.) - .5), [BATCH_SIZE, OUTPUT_DIM])\n all_real_data += tf.random_uniform(shape=[BATCH_SIZE, OUTPUT_DIM], minval=0., maxval=1. / 128) # dequantize\n all_real_data_splits = tf.split(all_real_data, len(DEVICES), axis=0)\n\n DEVICES_A = DEVICES[int(len(DEVICES) / 2):]\n # DEVICES_B = DEVICES[:int(len(DEVICES) / 2)]\n\n disc_costs = []\n disc_acgan_costs = []\n disc_acgan_accs = []\n disc_acgan_fake_accs = []\n for i, device in enumerate(DEVICES_A):\n with tf.device(device):\n real_and_fake_data = tf.concat(values=[\n all_real_data_splits[i],\n all_real_data_splits[len(DEVICES_A) + i],\n fake_data_splits[i],\n fake_data_splits[len(DEVICES_A) + i]\n ], axis=0)\n real_and_fake_labels = tf.concat(values=[\n labels_splits[i],\n labels_splits[len(DEVICES_A) + i],\n labels_splits[i],\n labels_splits[len(DEVICES_A) + i]\n ], axis=0)\n disc_all, disc_all_acgan = Discriminator(real_and_fake_data, real_and_fake_labels, update_collection=None)\n disc_real = disc_all[:int(BATCH_SIZE / len(DEVICES_A))]\n disc_fake = disc_all[int(BATCH_SIZE / len(DEVICES_A)):]\n if LOSS_TYPE == 'Goodfellow':\n if SOFT_PLUS:\n disc_real_l = -tf.reduce_mean(tf.nn.softplus(tf.log(tf.nn.sigmoid(disc_real))))\n disc_fake_l = -tf.reduce_mean(tf.nn.softplus(tf.log(1 - tf.nn.sigmoid(disc_fake))))\n else:\n disc_real_l = -tf.reduce_mean(tf.log(tf.nn.sigmoid(disc_real)))\n disc_fake_l = -tf.reduce_mean(tf.log(1 - tf.nn.sigmoid(disc_fake)))\n disc_costs.append(disc_real_l + disc_fake_l)\n elif LOSS_TYPE == 'HINGE':\n if SOFT_PLUS:\n disc_real_l = tf.reduce_mean(tf.nn.softplus(-tf.minimum(0., -1 + disc_real)))\n disc_fake_l = tf.reduce_mean(tf.nn.softplus(-tf.minimum(0., -1 - disc_fake)))\n else:\n # disc_real_l = -tf.reduce_mean(tf.minimum(0., -1 + disc_real))\n # disc_fake_l = -tf.reduce_mean(tf.minimum(0., -1 - disc_fake))\n disc_real_l = tf.reduce_mean(tf.nn.relu(1. - disc_real))\n disc_fake_l = tf.reduce_mean(tf.nn.relu(1. + disc_fake))\n disc_costs.append(disc_real_l + disc_fake_l)\n elif LOSS_TYPE == 'WGAN':\n if SOFT_PLUS:\n disc_costs.append(\n tf.reduce_mean(tf.nn.softplus(disc_fake)) + tf.reduce_mean(tf.nn.softplus(-disc_real)))\n else:\n disc_costs.append(tf.reduce_mean(disc_fake) - tf.reduce_mean(disc_real))\n\n if CONDITIONAL and ACGAN:\n disc_acgan_costs.append(tf.reduce_mean(\n tf.nn.sparse_softmax_cross_entropy_with_logits(\n logits=disc_all_acgan[:int(BATCH_SIZE / len(DEVICES_A))],\n labels=real_and_fake_labels[:int(BATCH_SIZE / len(DEVICES_A))])\n ))\n disc_acgan_accs.append(tf.reduce_mean(\n tf.cast(\n tf.equal(\n tf.to_int32(tf.argmax(disc_all_acgan[:int(BATCH_SIZE / len(DEVICES_A))], axis=1)),\n real_and_fake_labels[:int(BATCH_SIZE / len(DEVICES_A))]\n ),\n tf.float32\n )\n ))\n disc_acgan_fake_accs.append(tf.reduce_mean(\n tf.cast(\n tf.equal(\n tf.to_int32(tf.argmax(disc_all_acgan[int(BATCH_SIZE / len(DEVICES_A)):], axis=1)),\n real_and_fake_labels[int(BATCH_SIZE / len(DEVICES_A)):]\n ),\n tf.float32\n )\n ))\n\n # gradient_penalty, not included\n # if LOSS_TYPE == 'WGAN-GP'\n # for i, device in enumerate(DEVICES_B):\n # with tf.device(device):\n # real_data = tf.concat([all_real_data_splits[i], all_real_data_splits[len(DEVICES_A) + i]], axis=0)\n # fake_data = tf.concat([fake_data_splits[i], fake_data_splits[len(DEVICES_A) + i]], axis=0)\n # labels = tf.concat([\n # labels_splits[i],\n # labels_splits[len(DEVICES_A) + i],\n # ], axis=0)\n # alpha = tf.random_uniform(\n # shape=[int(BATCH_SIZE / len(DEVICES_A)), 1],\n # minval=0.,\n # maxval=1.\n # )\n # differences = fake_data - real_data\n # interpolates = real_data + (alpha * differences)\n # gradients = tf.gradients(Discriminator(interpolates, labels)[0], [interpolates])[0]\n # slopes = tf.sqrt(tf.reduce_sum(tf.square(gradients), reduction_indices=[1]))\n # gradient_penalty = 10 * tf.reduce_mean((slopes - 1.) ** 2)\n # disc_costs.append(gradient_penalty)\n\n disc_wgan = tf.add_n(disc_costs) / len(DEVICES_A)\n # tf.summary.scalar('D_wgan_cost', disc_wgan)\n if CONDITIONAL and ACGAN:\n disc_acgan = tf.add_n(disc_acgan_costs) / len(DEVICES_A)\n disc_acgan_acc = tf.add_n(disc_acgan_accs) / len(DEVICES_A)\n disc_acgan_fake_acc = tf.add_n(disc_acgan_fake_accs) / len(DEVICES_A)\n disc_cost = disc_wgan + (ACGAN_SCALE * disc_acgan)\n\n tf.summary.scalar('D_acgan_cost', disc_acgan)\n tf.summary.scalar('D_acgan_accuracy', disc_acgan_acc)\n tf.summary.scalar('D_acgan_fake_accuracy', disc_acgan_fake_acc)\n tf.summary.scalar('D_cost', disc_cost)\n else:\n disc_acgan = tf.constant(0.)\n disc_acgan_acc = tf.constant(0.)\n disc_acgan_fake_acc = tf.constant(0.)\n disc_cost = disc_wgan\n\n if DECAY:\n decay = tf.where(\n tf.less(_iteration, 400000),\n 1.0, tf.maximum(0., 1. - (tf.cast(_iteration, tf.float32) / 450000)))\n else:\n decay = 1.\n tf.summary.scalar('lr', LR * decay)\n\n gen_costs = []\n gen_acgan_costs = []\n for device in DEVICES:\n with tf.device(device):\n n_samples = GEN_BS_MULTIPLE * int(BATCH_SIZE / len(DEVICES))\n fake_labels = tf.cast(tf.random_uniform([n_samples]) * 1000, tf.int32)\n if CONDITIONAL and ACGAN:\n disc_fake, disc_fake_acgan = Discriminator(Generator(n_samples, fake_labels, reuse=True),\n fake_labels,\n update_collection=\"NO_OPS\",\n reuse=True)\n gen_costs.append(-tf.reduce_mean(tf.nn.softplus(disc_fake)))\n # gen_costs.append(-tf.reduce_mean(disc_fake))\n gen_acgan_costs.append(tf.reduce_mean(\n tf.nn.sparse_softmax_cross_entropy_with_logits(logits=disc_fake_acgan, labels=fake_labels)\n ))\n else:\n disc_fake, _ = Discriminator(Generator(n_samples, fake_labels, reuse=True),\n fake_labels,\n update_collection=\"NO_OPS\",\n reuse=True)\n if LOSS_TYPE == 'Goodfellow':\n if SOFT_PLUS:\n gen_costs.append(tf.reduce_mean(tf.nn.softplus(-tf.log(tf.nn.sigmoid(disc_fake)))))\n else:\n gen_costs.append(-tf.reduce_mean(tf.log(tf.nn.sigmoid(disc_fake))))\n elif LOSS_TYPE == 'HINGE':\n if SOFT_PLUS:\n gen_costs.append(tf.reduce_mean(tf.nn.softplus(-disc_fake)))\n else:\n gen_costs.append(-tf.reduce_mean(disc_fake))\n elif LOSS_TYPE == 'WGAN':\n if SOFT_PLUS:\n gen_costs.append(tf.reduce_mean(tf.nn.softplus(-disc_fake)))\n else:\n gen_costs.append(-tf.reduce_mean(disc_fake))\n gen_cost = (tf.add_n(gen_costs) / len(DEVICES))\n # tf.summary.scalar('G_wgan_cost', gen_cost)\n if CONDITIONAL and ACGAN:\n gen_cost += (ACGAN_SCALE_G * (tf.add_n(gen_acgan_costs) / len(DEVICES)))\n tf.summary.scalar('G_acgan_costs', tf.add_n(gen_acgan_costs) / len(DEVICES))\n tf.summary.scalar('G_cost', gen_cost)\n\n # gen_params = lib.params_with_name('Generator')\n # disc_params = lib.params_with_name('D.')\n gen_params = [var for var in tf.trainable_variables() if 'Generator' in var.name]\n print('\\ngen_params:')\n for var in gen_params:\n print(var.name)\n\n disc_params = [var for var in tf.trainable_variables() if 'Discriminator' in var.name]\n print('\\ndisc_params:')\n for var in disc_params:\n print(var.name)\n\n print('\\ntrainable_variables.name:')\n for var in tf.trainable_variables():\n print(var.name)\n\n gen_opt = tf.train.AdamOptimizer(learning_rate=LR * decay, beta1=0., beta2=0.9)\n disc_opt = tf.train.AdamOptimizer(learning_rate=LR * decay, beta1=0., beta2=0.9)\n gen_gv = gen_opt.compute_gradients(gen_cost, var_list=gen_params)\n disc_gv = disc_opt.compute_gradients(disc_cost, var_list=disc_params)\n gen_train_op = gen_opt.apply_gradients(gen_gv)\n disc_train_op = disc_opt.apply_gradients(disc_gv)\n\n # Function for generating samples\n frame_i = [0]\n fixed_noise = tf.constant(np.random.normal(size=(25, 128)).astype('float32'))\n # tiger shark(3), electric locomotive(547), mountain bike(671), submarine(833)\n # gray whale(147), Welsh springer spaniel(218), Persian cat(283), tiger(292),\n # chiffonier(493), fire truck(555), mosque(668), palace(698),\n # schooner(780), daisy(985), sandbar(977), pizza(963)\n sample_labels = np.array([3, 547, 671, 833, 147, 218, 283, 292, 493, 555, 668, 698, 780, 985, 977, 963],\n dtype='int32')\n # sample_labels = np.repeat(sample_labels, 25)\n fixed_labels = tf.constant(sample_labels)\n samples_prob = tf.multinomial(tf.log([[0.6] * 16]), 1)\n category = tf.cast(samples_prob[0][0], tf.int32)\n samples_label = fixed_labels[category]\n samples_label = tf.expand_dims(samples_label, axis=0)\n samples_label = tf.tile(samples_label, [25])\n fixed_noise_samples = Generator(25, samples_label, noise=fixed_noise, reuse=True)\n\n\n def generate_image(frame):\n samples = session.run(fixed_noise_samples)\n samples_label_ = session.run(fixed_labels[category])\n samples = ((samples + 1.) * (255. / 2)).astype('int32')\n # samples = np.split(samples, 16, 0)\n # for sample in samples:\n samples = np.reshape(samples, (25, 128, 128, 3))\n common.misc.save_images(samples, 'samples_{}_{}.png'.format(frame, samples_label_))\n\n\n # Function for calculating inception score\n fake_labels_100 = tf.cast(tf.random_uniform([100]) * 1000, tf.int32)\n samples_100 = Generator(100, fake_labels_100, reuse=True)\n\n\n def get_inception_score(n):\n all_samples = []\n for i in range(int(n / 100)):\n all_samples.append(session.run(samples_100))\n all_samples = np.concatenate(all_samples, axis=0)\n all_samples = ((all_samples + 1.) * (255.99 / 2)).astype('int32')\n all_samples = all_samples.reshape((-1, 128, 128, 3))\n return common.inception.inception_score.get_inception_score(list(all_samples))\n\n\n # Function for reading data\n # train_gen, dev_gen = lib.cifar10.load(BATCH_SIZE, DATA_DIR)\n #\n #\n # def inf_train_gen():\n # while True:\n # for images_, labels_ in train_gen():\n # yield images_, labels_\n #\n #\n # gen = inf_train_gen()\n\n for name, grads_and_vars in [('G', gen_gv), ('D', disc_gv)]:\n print(\"{} Params:\".format(name))\n total_param_count = 0\n for g, v in grads_and_vars:\n shape = v.get_shape()\n shape_str = \",\".join([str(x) for x in v.get_shape()])\n\n param_count = 1\n for dim in shape:\n param_count *= int(dim)\n total_param_count += param_count\n\n if g is None:\n print(\"\\t{} ({}) [no grad!]\".format(v.name, shape_str))\n else:\n print(\"\\t{} ({})\".format(v.name, shape_str))\n print(\"Total param count: {}\".format(locale.format(\"%d\", total_param_count, grouping=True)))\n\n summaries_op = tf.summary.merge_all()\n saver = tf.train.Saver(max_to_keep=5)\n summary_writer = tf.summary.FileWriter(CHECKPOINT_DIR, graph=session.graph)\n session.run(tf.global_variables_initializer())\n\n if RESTORE:\n ckpt = tf.train.latest_checkpoint(CHECKPOINT_DIR)\n if ckpt:\n print('restore model from: {}...'.format(ckpt))\n saver.restore(session, ckpt)\n\n filenames, labels = ILSVRC2012.get_filenames_labels(DATA_DIR)\n data_, labels_ = ILSVRC2012.input_fn(filenames, labels, BATCH_SIZE, 21)\n for iteration in range(ITERS):\n start_time = time.time()\n\n if 0 < iteration:\n _ = session.run([gen_train_op], feed_dict={_iteration: iteration})\n\n for i in range(N_CRITIC):\n # _data, _labels = next(gen)\n # data_, labels_ = ILSVRC2012.input_fn(filenames, labels, BATCH_SIZE, 21)\n _data, _labels = session.run([data_, labels_])\n\n # print('image_resized.shape: {}'.format(_data.shape)) # (N, 128, 128, 3)\n # _data = np.transpose(_data, axes=[0, 3, 1, 2]) # 'NHWC_to_NCHW'\n # print('image_transposed.shape: {}'.format(_data.shape)) # (N, 3, 128, 128)\n _data = np.reshape(_data, [_data.shape[0], -1])\n # print('image_flatten.shape: {}'.format(_data.shape)) # (N, 3*128*128)\n # print('_labels.shape: {}'.format(_labels.shape)) # (N,)\n # print('_data: {}'.format(_data))\n # print('_labels: {}'.format(_labels))\n\n if CONDITIONAL and ACGAN:\n _disc_cost, _disc_wgan, _gen_cost, _disc_acgan, _disc_acgan_acc, \\\n _disc_acgan_fake_acc, _, summaries = session.run(\n [disc_cost, disc_wgan, gen_cost, disc_acgan, disc_acgan_acc,\n disc_acgan_fake_acc, disc_train_op, summaries_op],\n feed_dict={all_real_data_int: _data,\n all_real_labels: _labels,\n _iteration: iteration})\n else:\n _disc_cost, _disc_wgan, _gen_cost, _, summaries = session.run(\n [disc_cost, disc_wgan, gen_cost, disc_train_op, summaries_op],\n feed_dict={all_real_data_int: _data,\n all_real_labels: _labels,\n _iteration: iteration})\n\n summary_writer.add_summary(summaries, global_step=iteration)\n\n # lib.plot.plot('cost', _disc_cost)\n lib.plot.plot('d_cost', _disc_wgan)\n lib.plot.plot('g_cost', _gen_cost)\n if CONDITIONAL and ACGAN:\n lib.plot.plot('disc_wgan', _disc_wgan)\n lib.plot.plot('acgan', _disc_acgan)\n lib.plot.plot('acc_real', _disc_acgan_acc)\n lib.plot.plot('acc_fake', _disc_acgan_fake_acc)\n # lib.plot.plot('time', time.time() - start_time)\n\n if iteration % INCEPTION_FREQUENCY == INCEPTION_FREQUENCY - 1:\n inception_score = get_inception_score(50000)\n lib.plot.plot('inception_50k', inception_score[0])\n lib.plot.plot('inception_50k_std', inception_score[1])\n\n # Calculate dev loss and generate samples every 100 iters\n if iteration % 100 == 99:\n # dev_disc_costs = []\n # for images, _labels in dev_gen():\n # _dev_disc_cost = session.run([disc_cost],\n # feed_dict={all_real_data_int: images,\n # all_real_labels: _labels})\n # dev_disc_costs.append(_dev_disc_cost)\n # lib.plot.plot('dev_cost', np.mean(dev_disc_costs))\n\n generate_image(iteration)\n\n if (iteration < 500) or (iteration % 1000 == 999):\n lib.plot.flush()\n\n if not os.path.exists(CHECKPOINT_DIR):\n os.mkdir(CHECKPOINT_DIR)\n saver.save(session, os.path.join(CHECKPOINT_DIR, 'model.ckpt'), global_step=iteration)\n\n lib.plot.tick()\n\n summary_writer.flush()\n summary_writer.close()\n","sub_path":"SNGAN/gan_imagNet_resnet.py","file_name":"gan_imagNet_resnet.py","file_ext":"py","file_size_in_byte":33201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"601545219","text":"import tensorflow as tf\nfrom larq import utils\nimport numpy as np\n\ntry:\n from tensorflow.keras.metrics import Metric\nexcept: # TensorFlow 1.13 doesn't export this as a public API\n from tensorflow.python.keras.metrics import Metric\n\n\nclass FlipRatio(Metric):\n \"\"\"Computes the mean ration of changed values in a given tensor.\n\n !!! example\n ```python\n m = metrics.FlipRatio(values_shape=(2,))\n m.update_state((1, 1)) # result: 0\n m.update_state((2, 2)) # result: 1\n m.update_state((1, 2)) # result: 0.75\n print('Final result: ', m.result().numpy()) # Final result: 0.75\n ```\n\n # Arguments\n values_shape: Shape of the tensor for which to track changes.\n values_dtype: Data type of the tensor for which to track changes.\n name: Name of the metric.\n dtype: Data type of the moving mean.\n \"\"\"\n\n def __init__(\n self, values_shape=(), values_dtype=\"int8\", name=\"flip_ratio\", dtype=None\n ):\n super().__init__(name=name, dtype=dtype)\n self.values_dtype = tf.as_dtype(values_dtype)\n self.values_shape = tf.TensorShape(values_shape).as_list()\n with tf.init_scope():\n self._previous_values = self.add_weight(\n \"previous_values\",\n shape=values_shape,\n dtype=self.values_dtype,\n initializer=tf.keras.initializers.zeros,\n )\n self.total = self.add_weight(\n \"total\", initializer=tf.keras.initializers.zeros\n )\n self.count = self.add_weight(\n \"count\", initializer=tf.keras.initializers.zeros\n )\n self._size = np.prod(self.values_shape)\n\n def update_state(self, values, sample_weight=None):\n values = tf.cast(values, self.values_dtype)\n changed_values = tf.math.count_nonzero(tf.equal(self._previous_values, values))\n flip_ratio = 1 - (tf.cast(changed_values, self.dtype) / self._size)\n\n update_total_op = self.total.assign_add(flip_ratio * tf.sign(self.count))\n with tf.control_dependencies([update_total_op]):\n update_count_op = self.count.assign_add(1)\n with tf.control_dependencies([update_count_op]):\n return self._previous_values.assign(values)\n\n def result(self):\n return tf.compat.v1.div_no_nan(self.total, self.count - 1)\n\n def reset_states(self):\n tf.keras.backend.batch_set_value(\n [(v, 0) for v in self.variables if v != self._previous_values]\n )\n\n def get_config(self):\n return {\n **super().get_config(),\n \"values_shape\": self.values_shape,\n \"values_dtype\": self.values_dtype.name,\n }\n\n def add_weight(\n self,\n name,\n shape=(),\n aggregation=tf.VariableAggregation.SUM,\n synchronization=tf.VariableSynchronization.ON_READ,\n initializer=None,\n dtype=None,\n ):\n if utils.tf_1_14_or_newer():\n return super().add_weight(\n name=name,\n shape=shape,\n aggregation=aggregation,\n synchronization=synchronization,\n initializer=initializer,\n dtype=dtype,\n )\n else:\n # Call explicitely tf.keras.layers.Layer.add_weight because TF 1.13\n # doesn't support setting a custom dtype\n return tf.keras.layers.Layer.add_weight(\n self,\n name=name,\n shape=shape,\n dtype=self._dtype if dtype is None else dtype,\n trainable=False,\n initializer=initializer,\n collections=[],\n synchronization=synchronization,\n aggregation=aggregation,\n )\n","sub_path":"larq/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":3814,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"468899756","text":"import math\nimport numpy as np\nimport pylab\nimport matplotlib.pyplot as plt\nx=np.arange(-3,3,0.01)\nfunk = str(input())\nylist = []\nfor h in range(len(x)):\n ylist.append(eval(str(funk.replace(\"x\",str(x[h])))))\nplt.plot(x,ylist)\nplt.axis('equal')\nplt.grid(True)\nplt.title(r'$Your func$')\nplt.show()\n","sub_path":"3.py","file_name":"3.py","file_ext":"py","file_size_in_byte":299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"61036064","text":"#coding=utf-8\nimport ConfigParser\nimport os\nimport smtplib\nimport datetime\nimport time\nfrom email.mime.text import MIMEText\nfrom email.header import Header\nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.image import MIMEImage\nfrom email.utils import parseaddr, formataddr\n\ndef get_config(section, key):\n\tconfig = ConfigParser.ConfigParser()\n\tpath = os.path.split(os.path.realpath(__file__))[0] + '/info.conf'\n\tconfig.read(path)\n\treturn config.get(section, key)\n\nsender = get_config('address', 'sender')\nkey = get_config('address', 'key')\nhost = get_config('address', 'host')\nreceiver = get_config('address', 'receiver') \n\n# 三个参数:第一个为文本内容,第二个 plain 设置文本格式,第三个 utf-8 设置编码\nmessage = MIMEMultipart('related') \nmessage['From'] = Header(\"章先生~\", 'utf-8').encode()#内容中显示的发件人\nmessage['To'] = Header(\"收件人哦~\", 'utf-8').encode()#内容中显示的收件人\nmessage['Subject'] = Header('I Love You~', 'utf-8').encode()#邮件的题目\n\nmsgAlternative = MIMEMultipart('alternative')\nmessage.attach(msgAlternative)\nmail_msg = \"\"\"\n

    i love you测试...

    \n

    图片演示:

    \n

    \n\"\"\"\nmsgAlternative.attach(MIMEText(mail_msg, 'html', 'utf-8'))\n\nfilename = get_config('image', 'path') + get_config('image', 'name') + '.jpg'\nfp = open(filename, 'rb')\nimage = MIMEImage(fp.read());\nfp.close()\nimage.add_header('Content-ID', '')\nmessage.attach(image)\n\n\nwhile True:\n\thour = get_config('time', 'hour')\n\tminute = get_config('time', 'minute') \n\tsecond = get_config('time', 'second') \n\tcurrent_time = time.localtime(time.time()) \n\tif ((current_time.tm_hour == int(hour)) and (current_time.tm_min == int(minute)) and (current_time.tm_sec == int(second))):\n\t\ttry:\n\t\t\tsmtpObj = smtplib.SMTP_SSL()#这个点要注意\n\t\t\tsmtpObj.connect(host)\n\t\t\tsmtpObj.login(sender, key) #邮箱登录\n\t\t\tsmtpObj.sendmail(sender, receiver, message.as_string())\n\t\t\tprint (\"邮件发送成功\")\n\t\texcept smtplib.SMTPException as e:\n\t\t\tprint (\"Error: 发送邮件产生错误\")\n\t\t\tprint(e)\n\ttime.sleep(1)\nsmtpObj.close()\n","sub_path":"i_love_you.py","file_name":"i_love_you.py","file_ext":"py","file_size_in_byte":2125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"116728447","text":"from astropy.io import fits\nimport matplotlib.pyplot as plt\nimport numpy as np\n\ndef mean_fits(files):\n if len(files) == 0:\n return []\n\n datasets = []\n\n for file in files:\n hdulist = fits.open(file)\n datasets.append(hdulist[0].data)\n hdulist.close()\n\n result = np.mean(datasets, axis=0)\n return result\n\ndef run_test(files):\n data = mean_fits(files)\n \n if len(data) > 0:\n print(data[100, 100])\n\n plt.imshow(data.T, cmap=plt.cm.viridis)\n plt.colorbar()\n plt.show()\n\nif __name__ == '__main__':\n # Test Case 1\n print(\"Test Case #1\")\n run_test(['image0.fits', 'image1.fits', 'image2.fits'])\n \n # Test Case 2\n print(\"Test Case #2\")\n run_test(['image0.fits', 'image1.fits', 'image3.fits'])\n \n # Test Case 3\n print(\"Test Case #3\")\n run_test(['image0.fits', 'image1.fits', 'image2.fits', 'image3.fits', 'image4.fits'])\n\n # Additional Test Case 1\n print(\"Additional Test Case #1\")\n run_test([])","sub_path":"week1/1b/4_mean_of_a_set_of_fits_files/program.py","file_name":"program.py","file_ext":"py","file_size_in_byte":1005,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"2233026","text":"# use open cv to show new images from AirSim \r\n\r\nfrom PythonClient import *\r\nimport cv2\r\nimport time\r\nimport sys\r\n\r\nclient = AirSimClient('127.0.0.1')\r\n\r\nhelp = False\r\n\r\nfontFace = cv2.FONT_HERSHEY_SIMPLEX\r\nfontScale = 0.5\r\nthickness = 2\r\ntextSize, baseline = cv2.getTextSize(\"FPS\", fontFace, fontScale, thickness)\r\nprint (textSize)\r\ntextOrg = (10, 10 + textSize[1])\r\nframeCount = 0\r\nstartTime=time.clock()\r\nfps = 0\r\n\r\nwhile True:\r\n # because this method returns std::vector, msgpack decides to encode it as a string unfortunately.\r\n result = client.getImageForCamera(0, AirSimImageType.Depth)\r\n if (result == \"\\0\"):\r\n if (not help):\r\n help = True\r\n print(\"Please press '1' in the AirSim view to enable the Depth camera view\")\r\n else:\r\n rawImage = np.fromstring(result, np.int8)\r\n png = cv2.imdecode(rawImage, cv2.IMREAD_UNCHANGED)\r\n \r\n cv2.putText(png,'FPS ' + str(fps),textOrg, fontFace, fontScale,(255,0,255),thickness)\r\n cv2.imshow(\"Depth\", png)\r\n\r\n frameCount = frameCount + 1\r\n endTime=time.clock()\r\n diff = endTime - startTime\r\n if (diff > 1):\r\n fps = frameCount\r\n frameCount = 0\r\n startTime = endTime\r\n \r\n key = cv2.waitKey(1) & 0xFF;\r\n if (key == 27 or key == ord('q') or key == ord('x')):\r\n break;\r\n","sub_path":"PythonClient/camera.py","file_name":"camera.py","file_ext":"py","file_size_in_byte":1343,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"239463923","text":"import os\n\nfrom cs50 import SQL\nfrom flask import Flask, flash, jsonify, redirect, render_template, request, session\nfrom flask_session import Session\nfrom tempfile import mkdtemp\nfrom werkzeug.exceptions import default_exceptions, HTTPException, InternalServerError\nfrom werkzeug.security import check_password_hash, generate_password_hash\n\nfrom helpers import apology, login_required, lookup, usd\nfrom itertools import chain\n\n# Configure application\napp = Flask(__name__)\n\n# Ensure templates are auto-reloaded\napp.config[\"TEMPLATES_AUTO_RELOAD\"] = True\n\n# Ensure responses aren't cached\n@app.after_request\ndef after_request(response):\n response.headers[\"Cache-Control\"] = \"no-cache, no-store, must-revalidate\"\n response.headers[\"Expires\"] = 0\n response.headers[\"Pragma\"] = \"no-cache\"\n return response\n\n# Custom filter\napp.jinja_env.filters[\"usd\"] = usd\n\n# Configure session to use filesystem (instead of signed cookies)\napp.config[\"SESSION_FILE_DIR\"] = mkdtemp()\napp.config[\"SESSION_PERMANENT\"] = False\napp.config[\"SESSION_TYPE\"] = \"filesystem\"\nSession(app)\n\n# Configure CS50 Library to use SQLite database\ndb = SQL(\"sqlite:///finance.db\")\n\n\n@app.route(\"/\")\n@login_required\ndef index():\n\n # arrange the data into group and name it stocks\n stocks = db.execute(\"SELECT Symbol,Name,Shares FROM portfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n\n # obtaining the cash form the database table name users\n result = db.execute(\"SELECT cash FROM users WHERE id=:id\",\n id=session[\"user_id\"])\n cash = result[0]\n y = cash[\"cash\"]\n cashs= float(y)\n\n grandtotal = cashs\n\n #delete the data from newportfolio.\n delete = db.execute(\"DELETE FROM newportfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n\n #obtaining the current price of every stock\n for stock in stocks:\n\n symbol = str(stock[\"Symbol\"])\n name = str(stock[\"Name\"])\n shares = int(stock[\"Shares\"])\n quote = lookup(name)\n price = float(quote[\"price\"])\n Total = float(price * shares)\n grandtotal += Total\n share = str(shares)\n\n\n #add all the new products and their price to the newportfolio table\n newportfolio = db.execute(\"INSERT INTO newportfolio(name,symbol,shares,price,total,id)VALUES(:name,:symbol,:shares,:price,:Total,:id)\",\n symbol=symbol,name=name,shares=share,price=price,Total=Total,\n id=session[\"user_id\"])\n\n #select the products from the table newportfolio to display through index.html\n newportfolio2 = db.execute(\"SELECT Name,Symbol,Shares,price,Total FROM newportfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n\n return render_template(\"index.html\",newportfolio=newportfolio2,total=usd(grandtotal),cash=usd(cashs))\n\n\n\n\n@app.route(\"/buy\", methods=[\"GET\", \"POST\"])\n@login_required\ndef buy():\n #if user reached via post\n if request.method == \"POST\":\n #make variables to facilitate it to me\n stock = request.form.get(\"symbol\")\n shares = request.form.get(\"shares\")\n int_shares = int(shares)\n\n # Ensure stock is submitted\n if not stock:\n return apology(\"must write a name of a stock\",400)\n\n # Ensure number of shares is submittes\n elif not shares :\n return apology(\"must provide how many do you want to purchase\",400)\n\n #Ensure that the number is positive\n elif int_shares < 1 :\n return apology(\"must be a positive number\",400)\n\n\n #select their money from the database\n list = db.execute(\"SELECT cash FROM users WHERE id=:id\",\n id=session[\"user_id\"])\n t = list[0]\n y = t['cash']\n r = int(y)\n\n #obtain the price,name, and symbol of the product form lookup function\n dict ={}\n dict = lookup(stock)\n price = dict['price']\n symbol = dict['symbol']\n name = dict['name']\n total = price*int_shares\n\n #obtain the Names from the database and make list of them\n Names = db.execute(\"SELECT Name FROM portfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n #making a list of Names from the database\n list = []\n for x in Names:\n Name = x[\"Name\"]\n if (stock == Name):\n list.append(Name)\n stocks = list\n\n\n #check that the user has enough money and if the stock is in stocks\n if (total<=r and stock in stocks ):\n\n newshare = db.execute(\"UPDATE portfolio set Shares = Shares+:int_shares , Total=:total+Total , datetime=datetime('now') WHERE Name=:stock AND id=:id \",\n stock=stock,int_shares=int_shares,total=total,id=session[\"user_id\"])\n\n newshare2 =db.execute(\"UPDATE newportfolio set Shares = Shares+:int_shares , Total=:total+Total WHERE Name=:stock AND id=:id\",\n stock=stock,int_shares=int_shares,total=total,id=session[\"user_id\"])\n\n newportfolio3 = db.execute(\"SELECT Name,Symbol,Shares,price,Total FROM newportfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n #update user cash\n update = db.execute(\"UPDATE users SET cash = :r - :total WHERE id=:id\",\n r=r,total=total,id=session[\"user_id\"])\n\n\n elif (total <= r and stock not in stocks) :\n\n\n #add the stock name and the username and the price to the portfolio database\n portfolio = db.execute(\"INSERT INTO portfolio(symbol,name,shares,price,total,id,Situation,datetime)VALUES(:name,:symbol,:int_shares,:price,:total,:id,'Bought',datetime('now'))\",\n symbol=symbol,name=name,int_shares=int_shares,price=price,total=total,id=session[\"user_id\"])\n\n\n\n #add the stock name and the username and the price to the newportfolio database\n newportfolio = db.execute(\"INSERT INTO newportfolio(symbol,name,shares,price,total,id)VALUES(:name,:symbol,:int_shares,:price,:total,:id)\",\n symbol=symbol,name=name,int_shares=int_shares,price=price,total=total,id=session[\"user_id\"])\n\n #update user cash\n update = db.execute('UPDATE users SET cash = :r - :total WHERE id=:id',\n r = r, total=total,id=session[\"user_id\"])\n\n #select form newportfolio\n newportfolio3 = db.execute(\"SELECT Name,Symbol,Shares,price,Total FROM newportfolio WHERE Name=Name AND id=:id GROUP BY Name,Symbol,Shares,Price,Total\",\n id=session[\"user_id\"])\n if (total <=r ):\n\n\n\n #bring cash from users table\n result = db.execute(\"SELECT cash FROM users WHERE id=:id\",\n id=session[\"user_id\"])\n cash = result[0]\n y = cash[\"cash\"]\n cashs = float(y)\n\n #add all the total form newportfolio table\n total0 = db.execute(\"SELECT SUM(Total) FROM newportfolio WHERE id=:id\",\n id=session[\"user_id\"])\n total2 = total0[0]\n x = total2[\"SUM(Total)\"]\n total3 = float(x)\n grandtotal = total3 + cashs\n\n #insert data into the history table\n Table = db.execute(\" SELECT datetime FROM portfolio WHERE Name=:stock AND id=:id\",\n stock=stock,id=session[\"user_id\"])\n date = Table[0]\n date2= date[\"datetime\"]\n\n\n History = db.execute(\"INSERT INTO History(Name,Price,Shares,Total,id,Situation,datetime)VALUES(:stock,:price,:int_shares,:total,:id,'bought',:date2)\",\n stock=stock,price=price,int_shares=int_shares,total=total,date2=date2,id=session[\"user_id\"])\n\n return render_template(\"index.html\",newportfolio=newportfolio3,cash=usd(cashs),total=usd(grandtotal))\n else:\n return apology(\"not enough money\",403)\n\n\n\n #if user via get means without clicking submit\n else:\n return render_template(\"buy.html\")\n\n\n@app.route(\"/check\", methods=[\"GET\"])\ndef check():\n\n username = request.args.get(\"username\")\n\n #getting all the usersname from the database\n users = db.execute(\"SELECT username FROM users WHERE username=:username\",username=username)\n\n\n\n #check if both existed\n if len(username)>0 and not users:\n return jsonify(\"true\")\n elif users and username:\n return jsonify(\"false\")\n\n\n\n\n #return redirect(\"/\")\n\n@app.route(\"/history\")\n@login_required\ndef history():\n\n\n History = db.execute(\"SELECT * FROM History WHERE id=:id\",\n id=session[\"user_id\"])\n return render_template('history.html',History=History)\n\n\n@app.route(\"/login\", methods=[\"GET\", \"POST\"])\ndef login():\n\n \"\"\"Log user in\"\"\"\n\n # Forget any user_id\n session.clear()\n\n # User reached route via POST (as by submitting a form via POST)\n if request.method == \"POST\":\n\n # Ensure username was submitted\n if not request.form.get(\"username\"):\n return apology(\"must provide username\", 403)\n\n # Ensure password was submitted\n elif not request.form.get(\"password\"):\n return apology(\"must provide password\", 403)\n\n # Query database for username\n rows = db.execute(\"SELECT * FROM users WHERE username = :username\",\n username=request.form.get(\"username\"))\n\n # Ensure username exists and password is correct\n if len(rows) != 1 or not check_password_hash(rows[0][\"hash\"], request.form.get(\"password\")):\n return apology(\"invalid username and/or password\", 403)\n\n # Remember which user has logged in\n session[\"user_id\"] = rows[0][\"id\"]\n\n # Redirect user to home page\n return redirect(\"/\")\n\n # User reached route via GET (as by clicking a link or via redirect)\n else:\n return render_template(\"login.html\")\n\n\n@app.route(\"/logout\")\ndef logout():\n \"\"\"Log user out\"\"\"\n\n # Forget any user_id\n session.clear()\n\n # Redirect user to login form\n return redirect(\"/\")\n\n\n@app.route(\"/quote\", methods=[\"GET\", \"POST\"])\n#@login_required\ndef quote():\n if request.method == \"POST\":\n if not request.form.get(\"symbol\"):\n return apology(\"choose a stock\", 400)\n\n symbol = request.form.get(\"symbol\")\n dict={}\n dict = lookup(symbol)\n if not dict:\n return apology(\"This stock is not available\",400)\n price = dict['price']\n symbol = dict['symbol']\n name = dict['name']\n\n\n\n\n\n\n\n\n return render_template(\"stock.html\",name=name,price=usd(price),symbol=symbol)\n else:\n\n return render_template(\"quote.html\")\n\n\n\n@app.route(\"/register\", methods=[\"GET\", \"POST\"])\ndef register():\n\n \"\"\"Register user\"\"\"\n if request.method == \"POST\":\n\n if not request.form.get(\"username\"):\n return apology(\"must provide username\", 400)\n elif not request.form.get(\"password\"):\n return apology(\"must provide password\", 400)\n elif not request.form.get(\"confirmation\"):\n return apology(\"must reconfirm password\", 400)\n elif request.form.get(\"password\") != request.form.get(\"confirmation\"):\n return apology(\"password must be the same \",400)\n\n hash=generate_password_hash(request.form.get(\"password\"))\n\n result = db.execute(\n \"INSERT INTO users (username,hash)VALUES(:username,:hash)\",\n username = request.form.get(\"username\"), hash =hash)\n\n if not result:\n return apology(\"Try another username\",400)\n\n session[\"user_id\"] = result\n\n return redirect(\"/\")\n\n else:\n\n return render_template(\"register.html\")\n\n\n@app.route(\"/sell\", methods=[\"GET\", \"POST\"])\n@login_required\ndef sell():\n #if user via post\n if request.method == \"POST\":\n\n #get a stock\n stock = request.form.get(\"stock\")\n\n #check that stock is submitted\n if not stock:\n return apology(\"Must enter a stock name\",403)\n\n\n #bring all the Names from the database\n stocks = db.execute(\"SELECT Name FROM newportfolio WHERE id=:id\",\n id=session[\"user_id\"])\n\n #obtaining the Names without hashing\n list = []\n for x in stocks:\n stock2 = x[\"Name\"]\n list.append(stock2)\n Names = list\n\n #iterate to see if what the user submitted is available\n if stock not in Names:\n return apology(\"not available\",403)\n\n #know how many shares\n shares = request.form.get(\"shares\")\n\n #translate shares into integer\n int_shares = int(shares)\n\n #obtain the price of the stock\n quote = lookup(stock)\n price = float(quote[\"price\"])\n total = float(price*int_shares)\n\n #check shares\n if not shares:\n return apology(\"Must enter shares\",403)\n\n #Check shares is positive\n elif int_shares<1:\n return apology(\"Must be positive\",403)\n\n #select the shares of the stock\n shares = db.execute(\"SELECT Shares FROM portfolio WHERE Name = :stock AND id=:id \",\n stock=stock,id=session[\"user_id\"])\n\n #remove the hash from the shares\n share = shares[0]\n share2 = share[\"Shares\"]\n share3 = int(share2)\n\n\n #ensure that there is enough shares\n if int_shares > share3:\n return apology(\"There is not enough shares\",403)\n\n #update shares and datetime in the portfolio\n updateshare = db.execute(\"UPDATE portfolio set Shares=Shares-:int_shares,Total=Total-:total,datetime=datetime('now') WHERE Name = :stock AND id=:id\",\n stock=stock,int_shares=int_shares,total=total,id=session[\"user_id\"])\n\n #select the updateshare from the portfolio\n updateshare2 = db.execute(\"SELECT Shares FROM portfolio WHERE Name = :stock AND id=:id \",\n stock=stock,id=session[\"user_id\"])\n\n #remove the hash\n updateshare3 = updateshare2[0]\n updateshare4 = updateshare3[\"Shares\"]\n updateshare5 = int(updateshare4)\n\n #update cash\n updatecash = db.execute(\"UPDATE users SET cash=cash+:total WHERE id=:id\",\n total=total,id=session[\"user_id\"])\n\n #see if the updateshare is 0 delete\n if updateshare5 == 0 :\n Delete = db.execute(\"DELETE FROM portfolio WHERE Name=:stock AND id=:id \",\n stock=stock,id=session[\"user_id\"])\n\n #obtain datetime from portfolio\n date = db.execute(\"SELECT datetime FROM portfolio WHERE Name=:stock AND id=:id\",\n stock=stock,id=session[\"user_id\"])\n date2 = date[0]\n date3 = date2[\"datetime\"]\n\n #insert into history table\n History = db.execute(\"INSERT INTO History (Name,Price,Shares,Total,id,Situation,datetime)VALUES(:stock,:price,:int_shares,:total,:id,'sold',:date3)\",\n stock=stock,price=price,int_shares=int_shares,total=total,date3=date3,id=session[\"user_id\"])\n\n History2 = db.execute(\"SELECT * FROM History WHERE id=:id\",\n id=session[\"user_id\"])\n\n return render_template(\"history.html\",History=History2)\n #if user via get\n else:\n return render_template(\"sell.html\")\n\ndef errorhandler(e):\n \"\"\"Handle error\"\"\"\n if not isinstance(e, HTTPException):\n e = InternalServerError()\n return apology(e.name, e.code)\n\n\n# Listen for errors\nfor code in default_exceptions:\n app.errorhandler(code)(errorhandler)\n","sub_path":"environment/finance/.~c9_invoke_9wziLU.py","file_name":".~c9_invoke_9wziLU.py","file_ext":"py","file_size_in_byte":15162,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"84257480","text":"from django.shortcuts import render,redirect\nfrom django.template.context_processors import csrf\nfrom django.template import loader\nfrom django.http import HttpResponse\nfrom django.http import JsonResponse\n\nfrom django.template.context_processors import csrf\nfrom system.models import issued,Books,Faculty,Requests,issued,StudentIssued\nfrom datetime import datetime,date\n\nfrom django.core import mail\nfrom django.core.mail import EmailMessage,send_mail\nfrom django.template.loader import render_to_string\nfrom django.utils.html import strip_tags\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.files import File\n\n# Create your views here.\n\ndef BaseLayout(request):\n return render(request,'administrator/base.html')\n\n@login_required(login_url='/login/')\ndef AllFaculty(request):\n all_faculty=Faculty.objects.all()\n req_count = Requests.objects.all()\n print(req_count)\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n allissued = issued.objects.filter(return_date__lt=mydate)\n context={\n 'faculties':all_faculty,\n 'req_count':req_count,\n 'overdue':allissued\n }\n return render(request,'administrator/Member.html',context=context)\n\n@login_required(login_url='/login/')\ndef AllBooks(request):\n req_count = Requests.objects.all()\n all_books= Books.objects.all()\n #print(all_books)\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n allissued = issued.objects.filter(return_date__lt=mydate)\n context={\n 'books':all_books,\n 'req_count':req_count,\n 'overdue':allissued\n }\n return render(request,'administrator/books.html',context=context)\n #return render_to_response('administrator/books.html', context=context)\n\n\n@login_required(login_url='/login/')\ndef Add(request):\n if request.method == \"POST\":\n id = request.POST.get('id','')\n name= request.POST.get('name','')\n email= request.POST.get('email','')\n phone_no = request.POST.get('phone_no','')\n if request.POST.get('ea','') == \"Add\":\n print(phone_no)\n password= User.objects.make_random_password(length=8) #Generate password randomly\n print(password) #Print Randomly generated password \n a=User.objects.create_user(id,email,password)\n a.save()\n faculty= Faculty(id=id,name=name,email=email,phone_no=phone_no,password=password,date_joined=datetime.now())\n faculty.save()\n # Receiver email\n to=email\n # body and subject of mail \n body=\"Hey %s ! \\n \\n Password for your Ce-Department Library account is %s \\n\"%(name,password)\n #Composing email and sending mail\n email=EmailMessage('CE-Department',body,to=[to])\n email.send()\n else:\n faculty = Faculty.objects.get(id=id)\n faculty.name = name\n faculty.email = email\n faculty.phone_no = phone_no\n faculty.save()\n all_faculty=Faculty.objects.all()\n return redirect('/administrator/Faculties/')\n\n@login_required(login_url='/login/')\ndef AddBook(request):\n if request.method == \"POST\":\n last_book = Books.objects.last()\n #print(last_book.id)\n last_id = last_book.id.split('-')\n #print(last_id)\n id = int(last_id[1]) + 1\n Title = request.POST.get('Title','')\n id= \"CE-\" + str(id)\n Publisher= request.POST.get('Publisher','')\n Author= request.POST.get('Author','')\n a=Books(sr_no=int(last_book.sr_no)+1,id=id,title=Title,publisher=Publisher,author=Author,available=True)\n a.save()\n return redirect('/administrator/Books/')\n\n\n\n@login_required(login_url='/login/')\ndef Editdata(request):\n if request.method == \"GET\":\n #print(request.GET.get('id'))\n fac_data = Faculty.objects.get(id=request.GET.get('id'))\n #print(fac_data.id)\n data={\n 'id':fac_data.id,\n 'name':fac_data.name,\n 'email':fac_data.email,\n 'phone_no':fac_data.phone_no,\n }\n return JsonResponse(data)\n\n@login_required(login_url='/login/')\ndef DeleteFac(request):\n if request.method == \"GET\":\n fac_data = Faculty.objects.get(id=request.GET.get('id'))\n fac_data.delete()\n data={\n 'cond':True\n }\n return JsonResponse(data)\n\n@login_required(login_url='/login/') \ndef InputCSV(request):\n if request.method == \"POST\":\n if(request.POST.get('type') == \"fac\"):\n csv_file = request.FILES[\"csv_file\"]\n file_data = csv_file.read().decode(\"utf-8\")\t\n lines = file_data.split(\"\\n\")\n for line in lines:\t\t\t\n fields = line.split(',')\n print(fields)\n faculty= Faculty(id=fields[0],name=fields[1],email=fields[2],phone_no=fields[3],password=fields[4],date_joined=datetime.now())\n faculty.save()\n print(line)\n return redirect('/administrator/Faculties/')\n elif(request.POST.get('type') == \"Book\"):\n return redirect('/administrator/Faculties/')\n else:\n return redirect('/administrator/Faculties/')\n else:\n return render(request,'administrator/upload.html')\n\n\"\"\"def UploadDatabase(request):\n if request.method == \"POST\":\n csv_file = request.FILES[\"csv_file\"]\n file_data = csv_file.read().decode(\"utf-8\")\t\n lines = file_data.split(\"\\n\")\n for line in lines:\t\t\t\t\t\t\n\t\t\tfields = line.split(\",\")\n faculty= Faculty(id=fields[0],name=fields[1],email=fields[2],phone_no=fields[3],password=fields[4],date_joined=datetime.now())\n faculty.save()\n return HttpResponse(\"saved\")\"\"\"\n\n@login_required(login_url='/login/')\ndef ChangebookStatus(request):\n if request.method == \"GET\":\n book = Books.objects.get(id=request.GET.get('bookid'))\n if request.GET.get('cond') == 'add':\n book.available = True\n book.save()\n else:\n book.available = False\n book.save()\n return JsonResponse({\"successful\":True})\n#faculty\n@login_required(login_url='/login/')\ndef BookRequests(request):\n due = 6\n with open('./system/Due.txt','r') as f:\n f.readline()\n f.readline()\n due = int(f.readline())\n print(due)\n\n if request.method == \"POST\":\n req = Requests.objects.get(id = request.POST.get('req_id'))\n if (req.date.month+due) >= 12:\n new_month = (req.date.month+due) % 12\n new_year=req.date.year+1\n if new_month==0:\n new_month = 12\n new_year = req.date.year\n else:\n new_month = (req.date.month+due)\n new_year = req.date.year\n new_date=date(new_year,new_month,req.date.day)\n issue=issued(book_id=req.book_id,faculty_id=req.faculty_id,issue_date=date.today(),return_date=new_date)\n req.delete()\n issue.save()\n same_book = Requests.objects.filter(book_id=req.book_id)\n for b in same_book:\n b.delete()\n return redirect('/administrator/BookRequest/')\n else:\n req_count = Requests.objects.all()\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n allissued = issued.objects.filter(return_date__lt=mydate)\n context={\n 'all_requests':Requests.objects.all(),\n 'req_count':req_count,\n 'overdue':allissued\n }\n return render(request,'administrator/request.html',context=context)\n\n@login_required(login_url='/login/')\ndef BookIssued(request):\n if request.method == \"GET\":\n req_count = Requests.objects.all()\n allissued = issued.objects.all().order_by('return_date')\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n #allissued = issued.objects.all()\n context={\n 'allissued':allissued,\n 'req_count':req_count,\n 'overdue':issued.objects.filter(return_date__lt=mydate)\n }\n return render(request,'administrator/issued.html',context=context)\n if request.method == \"POST\":\n if request.POST.get('status') == \"return\":\n myissue = issued.objects.get(id=request.POST.get('issue_id'))\n myissue.delete()\n return redirect('/administrator/BookIssued/')\n if request.POST.get('status') == \"renew\":\n myissue = issued.objects.get(id=request.POST.get('issue_id'))\n if (myissue.return_date.month+6) >= 12:\n new_month = (myissue.return_date.month+6) % 12\n new_year=myissue.return_date.year+1\n if new_month==0:\n new_month = 12\n new_year = myissue.return_date.year\n else:\n new_month = (myissue.return_date.month+6)\n new_year = myissue.return_date.year\n new_date=date(new_year,new_month,myissue.return_date.day)\n myissue.return_date = new_date\n myissue.save()\n return redirect('/administrator/BookIssued/')\n\n@login_required(login_url='/login/')\ndef Notify(request):\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%m/%d/%Y')\n subject = 'DeadLine Of Book'\n context={\n 'end' : mydate,\n }\n html_message = render_to_string('administrator/mail.html', context)\n plain_message = strip_tags(html_message)\n from_email = 'From '\n to = 'gdthumar.code@gmail.com'\n mail.send_mail(subject, plain_message, from_email, [to], html_message=html_message)\n return HttpResponse('Sent')\n\n@login_required(login_url='/login/')\ndef OverDue(request):\n if request.method == \"GET\":\n req_count = Requests.objects.all()\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n allissued = issued.objects.filter(return_date__lt=mydate)\n context={\n 'allissued':allissued,\n 'req_count':req_count,\n 'today':date.today()\n }\n return render(request,'administrator/overdue.html',context=context)\n\n@login_required(login_url='/login/')\ndef Send_Notification(request):\n t = datetime(date.today().year, date.today().month, date.today().day, 0, 0)\n mydate=t.strftime('%Y-%m-%d')\n print(mydate)\n allissued = issued.objects.filter(return_date__lt=mydate)\n from_email = 'From '\n subject = 'DeadLine Of Book'\n to=[]\n for issue in allissued:\n context={\n 'end' : issue.return_date,\n }\n to.append(issue.faculty_id.email)\n html_message = render_to_string('administrator/mail.html', context)\n plain_message = strip_tags(html_message)\n to = issue.faculty_id.email\n mail.send_mail(subject, plain_message, from_email, [to], html_message=html_message)\n return JsonResponse({\"successful\":mydate})\n\n@login_required(login_url='/login/')\ndef ChangeSetting(request):\n if request.method == \"GET\":\n f=open(\"./system/Due.txt\", \"r\")\n templines=f.readlines()\n lines = []\n for line in templines:\n lines.append(line)\n print(lines)\n studentduedate=lines[0]\n studentduecharge=lines[1]\n facultyduedate=lines[2]\n facultyduecharge=lines[3]\n #t = date(date.today().year, date.today().month, date.today().day)\n #print(t)\n data={\n 'studentduedate':studentduedate,\n 'studentduecharge':studentduecharge,\n 'facultyduedate':facultyduedate,\n 'facultyduecharge':facultyduecharge,\n }\n return JsonResponse(data)\n else:\n studentduedate=request.POST.get('studentduedate')\n studentduecharge=request.POST.get('studentduecharge')\n facultyduedate=request.POST.get('facultyduedate')\n facultyduecharge=request.POST.get('facultyduecharge')\n # 1-> student date 2->charge 3->faculty date 4->charge\n print(studentduedate,studentduecharge)\n print(facultyduedate,facultyduecharge)\n with open('./system/Due.txt','w+') as f:\n f.write(studentduedate+'\\n'+ studentduecharge+'\\n'+ facultyduedate+'\\n'+ facultyduecharge+ '\\n')\n\n return redirect('/administrator/Books/')\n\ndef AboutUs(request):\n return render(request,'administrator/aboutus.html')\n\n\n \n\n\n\n\n","sub_path":"library/system/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":12742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"2086946","text":"from __future__ import absolute_import, division, print_function, unicode_literals\n\nimport copy\nfrom collections import namedtuple\n\nfrom echomesh.base import Args\nfrom echomesh.base import CommandFile\nfrom echomesh.base import GetPrefix\nfrom echomesh.base import Leafs\nfrom echomesh.base import Merge\nfrom echomesh.base import Yaml\n\n_ARGUMENT_ERROR = \"\"\"\nERROR: Didn't understand arguments to echomesh: \"%s\".\n\nechomesh needs to be called with arguments looking like \"name=value\".\n\nExamples:\n echomesh\n echomesh debug=true\n echomesh audio.input.enable=false light.enable=false\n\"\"\"\n\n_ASSIGNMENT_ERROR = \"\"\"\nERROR: couldn't assign a variable from: \"%s\".\n\nVariable assignments look like \"name=value\" and you can have more than one\nper line.\n\nExamples:\n debug=true\n audio.input.enable=false light.enable=false\n\"\"\"\n\nFileConfig = namedtuple('FileConfig', 'file base edits changes')\n\nclass MergeConfig(object):\n def __init__(self, args):\n self.args = args\n self.read()\n\n def read(self):\n self._read_file_configs()\n self.arg_config = self._assignment_to_config(self.args, _ARGUMENT_ERROR)\n return self.recalculate()\n\n def recalculate(self):\n self.config = None\n self.changed = {}\n for _, configs in self.file_configs:\n self.config = Merge.merge(self.config, *configs)\n self.changed = Merge.merge(self.changed, *configs[2:])\n\n arg = copy.deepcopy(self.arg_config)\n clean_arg = Merge.difference_strict(arg, self.changed)\n self.config = Merge.merge_for_config(self.config, clean_arg)\n\n return self.config\n\n def has_changes(self):\n return any(configs[2] for (_, configs) in self.file_configs)\n\n def get_changes(self):\n return [(f, c[2]) for (f, c) in self.file_configs if c[2]]\n\n def assign(self, args, index=2): # default is 'master'\n configs = self.file_configs[index][1]\n\n while len(configs) < 3:\n configs.append({})\n assignments = self._assignment_to_config(args, _ASSIGNMENT_ERROR)\n configs[2] = Merge.merge(configs[2], assignments)\n self.recalculate()\n return assignments\n\n def save(self):\n saved_files = []\n for f, configs in self.file_configs:\n if len(configs) > 2 and configs[2]:\n saved_files.append(f)\n configs[1] = Merge.merge(*configs[1:])\n while len(configs) > 2:\n configs.pop()\n with open(f, 'r') as fo:\n data = fo.read().split(Yaml.SEPARATOR)[0]\n\n with open(f, 'wb') as fw:\n fw.write(data)\n fw.write(Yaml.SEPARATOR)\n fw.write(Yaml.encode_one(configs[1]))\n\n self.arg_config = Merge.difference_strict(self.arg_config, self.changed)\n self.recalculate()\n return saved_files\n\n def assignments(self, index=2):\n assigned = self.file_configs[index][1]\n return (len(assigned) > 2 and Leafs.leafs(assigned[2])) or {}\n\n def _read_file_configs(self):\n self.file_configs = []\n base_config = None\n\n for f in reversed(CommandFile.expand('config.yml')):\n configs = Yaml.read(f, 'config')\n for c in configs:\n if base_config:\n base_config = Merge.merge_for_config(base_config, c)\n else:\n base_config = copy.deepcopy(c)\n while len(configs) < 3:\n configs.append({})\n self.file_configs.append([f, configs])\n\n def _assignment_to_config(self, args, error):\n args = ' '.join(args)\n config = {}\n base_config = self.file_configs[0][1][0]\n assert isinstance(base_config, dict)\n try:\n split_args = Args.split(args)\n except Exception as e:\n e.arg = '%s %s' % (error, args)\n raise\n\n for addr, value in split_args:\n try:\n GetPrefix.set_assignment(addr, value, base_config, config,\n unmapped_names=Merge.CONFIG_EXCEPTIONS)\n except GetPrefix.PrefixException:\n raise Exception('Can\\'t understand configuration address \"%s\"' % addr)\n except Exception:\n raise Exception('Can\\'t understand configuration value \"%s\" in %s=%s' %\n (value, addr, value))\n return config\n\n","sub_path":"code/python/echomesh/base/MergeConfig.py","file_name":"MergeConfig.py","file_ext":"py","file_size_in_byte":4026,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"400530774","text":"import sys, os\r\nDAFAPP_DIR = os.environ.get(\"DAFAPPSERVER_ROOTDIR\")\r\nsys.path.append(DAFAPP_DIR + 'ibank/accounting/script_modules')\r\nimport accountingapi\r\n\r\ndef CreateJournal(config, cnnumber, cndate, branch_code, user_id_commit):\r\n journal = config.CreatePObject('Journal')\r\n datevalue = accountingapi.GetActiveAccDay(config).DateValue\r\n datevalue = config.ModDateTime.EncodeDate(datevalue[0],datevalue[1],datevalue[2])\r\n journal.journal_date = datevalue\r\n journal.description = 'Jurnal CN %s' % (cnnumber)\r\n journal.branch_code = branch_code\r\n journalapi = accountingapi.Journal(journal)\r\n journal_type = 'GL' # Jurnal umum\r\n journalapi.SetNewInstance(journal_type)\r\n\r\n journal.is_posted = 'F'\r\n journal.is_partlychecked = 'T'\r\n journal.userid_create = user_id_commit\r\n \r\n return journal\r\n\r\ndef CreateJournalItem(config, strDescription, journal, oAccountInstance, debit, credit, user_id_commit):\r\n item = config.CreatePObject('JournalItem')\r\n item.description = strDescription\r\n itemapi = accountingapi.JournalItem(item)\r\n itemapi.SetNewInstance(journal, oAccountInstance, debit, credit)\r\n item.JournalItemStatus = 'C'\r\n item.userid_create = user_id_commit\r\n item.userid_check = user_id_commit\r\n\r\ndef DAFScriptMain(config, parameter, returnpacket):\r\n # config: ISysConfig object\r\n # parameter: TPClassUIDataPacket\r\n # returnpacket: TPClassUIDataPacket (undefined structure)\r\n\r\n user_id_commit = parameter.FirstRecord.user_id_commit\r\n cnvalue = parameter.FirstRecord.cnvalue\r\n cash = parameter.FirstRecord.cash\r\n todeposit = parameter.FirstRecord.todeposit\r\n cnnumber = parameter.FirstRecord.cnnumber\r\n cndate = parameter.FirstRecord.cndate\r\n branch_code = parameter.FirstRecord.branch_code\r\n currency_code = parameter.FirstRecord.currency_code\r\n\r\n oAccInstCNSales = accountingapi.GetAccModuleIntfInstance(branch_code, currency_code, 'cnsales', config)\r\n oAccInstCNCash = accountingapi.GetAccModuleIntfInstance(branch_code, currency_code, 'cncash', config)\r\n #oAccInstCNCustDeposit = accountingapi.GetAccModuleIntfInstance(branch_code, currency_code, 'cncustdeposit', config)\r\n oAccInstCNPybDeposit = accountingapi.GetAccModuleIntfInstance(branch_code, currency_code, 'cnpybdeposit', config)\r\n #oAccInstCNRcvSales = accountingapi.GetAccModuleIntfInstance(branch_code, currency_code, 'cnrcvsales', config)\r\n\r\n config.BeginTransaction()\r\n try:\r\n journal = CreateJournal(config, cnnumber, cndate, branch_code, user_id_commit)\r\n\r\n # journal item cn sales\r\n strDescription = 'Pengurangan nilai penjualan akibat CN '+ cnnumber\r\n CreateJournalItem(config, strDescription, journal, oAccInstCNSales, cnvalue, 0.0, user_id_commit)\r\n\r\n # sementara tidak ada cn ke cash, semua cn didepositkan\r\n # journal item cash\r\n #strDescription = 'Pembayaran tunai untuk CN '+ cnnumber\r\n #CreateJournalItem(config, strDescription, journal, oAccInstCNCash, 0.0, cash, user_id_commit)\r\n\r\n # tidak dianggap mengurangi piutang karena tidak terikat dengan invoice tertentu\r\n # journal item piutang\r\n #strDescription = 'Pengurangan piutang usaha akibat CN '+ cnnumber\r\n #CreateJournalItem(config, strDescription, journal, oAccInstCNRcvSales, 0.0, todeposit, user_id_commit)\r\n\r\n # journal item hutang\r\n strDescription = 'Penambahan hutang usaha akibat CN '+ cnnumber\r\n CreateJournalItem(config, strDescription, journal, oAccInstCNPybDeposit, 0.0, cnvalue, user_id_commit)\r\n\r\n #CheckDebitCreditBal(config, journal.journal_no)\r\n\r\n config.Commit()\r\n isSucceed = 1\r\n except:\r\n config.Rollback()\r\n isSucceed = 0\r\n raise\r\n \r\n returnpacket.CreateValues(['isSucceed',isSucceed])\r\n\r\n return 1\r\n\r\n","sub_path":"scripts/sales/cn journal.py","file_name":"cn journal.py","file_ext":"py","file_size_in_byte":3666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"} +{"seq_id":"230024583","text":"# -*- mode:python;coding:utf-8 -*-\nimport pygst\npygst.require('0.10')\nimport gst\nimport gobject, sys\n\n\nclass GPlayer:\n def __init__(self):\n self.player = gst.Pipeline(\"esono\")\n\n self.source = gst.element_factory_make('gnomevfssrc', 'file-source')\n self.source = gst.element_factory_make(\"audiotestsrc\", \"audio\")\n \n self.volume = gst.element_factory_make('volume','volume')\n self.volume.set_property('volume', 0.1)\n self.sink = gst.element_factory_make(\"jackaudiosink\", \"sink\")\n self.player.add(self.source, self.volume, self.sink )\n gst.element_link_many(self.source, self.volume, self.sink )\n\n def play(self, path):\n # self.source.set_property('location', 'file://' + path )\n self.player.set_state(gst.STATE_PLAYING)\n \n\n\n'''\ndef play_uri(uri):\n \" play an uri like file:///home/foo/bar.mp3 \"\n\n mainloop = gobject.MainLoop()\n player = gst.element_factory_make(\"playbin\", \"player\")\n \n player.set_property('uri', uri)\n player.set_state(gst.STATE_PLAYING)\n\n mainloop.run()\n\nplay_uri(\"file:///data/audio/sc/marco_bernabe/terje_paulsen__the_abundant_emptiness_between_cold_and_heat.flac\")\n'''\nmainloop = gobject.MainLoop()\np = GPlayer()\np.play(\"/data/audio/sc/marco_bernabe/terje_paulsen__the_abundant_emptiness_between_cold_and_heat.flac\")\np.play(\"/usr/lib/pd/doc/sound/voice2.wav\")\nmainloop.run()\n","sub_path":"esonoclaste.app/Contents/Resources/esono/tests/zz/gstreamer.py","file_name":"gstreamer.py","file_ext":"py","file_size_in_byte":1402,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"50"}