diff --git "a/1712.jsonl" "b/1712.jsonl"
new file mode 100644--- /dev/null
+++ "b/1712.jsonl"
@@ -0,0 +1,669 @@
+{"seq_id":"97765736","text":"inStore = 5\n\nx = int(input(\"How many candies do you want? \"))\n\ni = 1 # start distribution\nwhile i <= x:\n if i > inStore:\n print(\"Ran out of all candies!\")\n break # come out of the loop\n print(\"Here is \" + str(i))\n i += 1\n\nfor i in range(1, 30):\n if i % 3 == 0:\n print(\"NO\")\n continue # skip the remaining statements\n print(i)\n","sub_path":"Introduction/Continue.py","file_name":"Continue.py","file_ext":"py","file_size_in_byte":369,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"96721301","text":"# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\nimport os\n\nfrom azure.cli.testsdk import (\n ScenarioTest, JMESPathCheck, JMESPathCheckExists\n)\nfrom azure.cli.testsdk.base import execute\nfrom azure.cli.testsdk.preparers import AbstractPreparer\nimport azure.cli.core.azlogging as azlogging\n\nlogger = azlogging.get_az_logger(__name__)\n\n\nclass SelectNoSecClusterPreparer(AbstractPreparer):\n def __init__(self, parameter_name=\"endpoint\",\n endpoint=\"http://127.0.0.1:10550\",\n env_variable_name=\"AZURE_CLI_SF_ENDPOINT\"):\n # Name randomization unnecessary\n super(SelectNoSecClusterPreparer, self).__init__(\"test\", 10)\n self.endpoint = endpoint\n self.parameter_name = parameter_name\n self.env_variable_name = env_variable_name\n\n def create_resource(self, _, **kwargs):\n # Omit name here since there is no randomization required\n endpoint = os.environ.get(self.env_variable_name, self.endpoint)\n logger.debug(\"endpoint %s\", endpoint)\n template = \"az sf cluster select --endpoint {}\"\n execute(template.format(endpoint))\n return {self.parameter_name: endpoint}\n\n\nclass ServiceFabricTests(ScenarioTest):\n\n package_path = \"/media/share/EchoServerApplication3\"\n package_name = \"EchoServerApplication3\"\n application_type_name = \"EchoServerApp\"\n application_type_version = \"3.0\"\n application_name = \"fabric:/app1\"\n application_id = \"app1\"\n\n # Application tests\n\n @SelectNoSecClusterPreparer()\n def sf_test_application_lifecycle(self):\n self.cmd(\"az sf application upload --path {}\".format(\n self.package_path\n ))\n\n self.cmd(\n \"az sf application provision \"\n \"--application-type-build-path {}\".format(\n self.package_name\n )\n )\n\n self.cmd(\n \"az sf application type\",\n checks=[\n JMESPathCheck(\n \"items[0].name\",\n self.application_type_name\n ),\n JMESPathCheck(\n \"items[0].version\",\n self.application_type_version\n )\n ]\n )\n\n self.cmd(\n \"az sf application create \"\n \"--app-type {} --version {} --name {}\".format(\n self.application_type_name,\n self.application_type_version,\n self.application_name\n )\n )\n\n self.cmd(\n \"az sf application list\",\n checks=[\n JMESPathCheck(\"items[0].id\", self.application_id)\n ]\n )\n\n self.cmd(\n \"az sf application health \"\n \"--application-id {}\".format(self.application_id),\n checks=[\n JMESPathCheck(\"name\", self.application_name),\n JMESPathCheckExists(\"aggregatedHealthState\")\n ]\n )\n\n self.cmd(\n \"az sf application delete --application-id {}\".format(\n self.application_id\n )\n )\n\n self.cmd(\n \"az sf application unprovision \"\n \"--application-type-name {} \"\n \"--application-type-version {}\".format(\n self.application_type_name, self.application_type_version\n )\n )\n","sub_path":"src/command_modules/azure-cli-sf/tests/manual_sf_commands.py","file_name":"manual_sf_commands.py","file_ext":"py","file_size_in_byte":3644,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"451499539","text":"import matplotlib.pyplot as plt\n\ndays=int(input(\"How many days do you have?\"))\nno_of_subjects=int(input(\"How many subjects to study?\"))\n\nsubjects={}\ntotal_hours=0\n\nfor i in range(no_of_subjects):\n\tname=input(\" Subject name\")\n\thours=int(input(\"Hours to study\"))\n\tsubjects[name]=hours\n\t\n\ttotal_hours=total_hours+hours\n\ndataset=[]\nxaxis=[]\n\nfor i in range(days+1,0,-1):\n\txaxis.append(str(i))\n\tif i <=2:\n\t\thours = total_hours/i\n\t\tdataset.append(hours)\n\telse:\n\t\thours = total_hours/(i)\n\t\tdataset.append(hours)\n\nif days <= 2 :\n\tstudyhours = total_hours/(days)\n\nelse:\n\tstudyhours = total_hours/(days)\n\t\t\n\thours = int(studyhours)\n\tminutes = float(studyhours*60) % 60\n\n\n\nprint(\"You need to study daily for {} hours {} minutes\".format(hours,float(minutes)))\n\nchoice=int(input(\"Press 1 to show graph\"))\n\nif choice==1:\n\n\tfor i in range(len(xaxis)):\n\t\tx=xaxis[i]\n\t\ty=dataset[i]\n\n\t\tprint(\"If you start on day {}, you need {} hours daily\".format(x,float(y)))\n\n\tplt.plot(xaxis,dataset,color=\"g\")\n\tplt.title('You have to study for:')\n\tplt.xlabel('days')\n\tplt.ylabel('Hours to study')\n\n\tplt.show()\n\n","sub_path":"daily_study_hours_linegraph.py","file_name":"daily_study_hours_linegraph.py","file_ext":"py","file_size_in_byte":1081,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"321368281","text":"from matplotlib import pyplot as plt\nimport z3\nfrom z3 import Real, Optimize, If\nfrom fractions import Fraction\n\n\nEPS = 1e-2\n\n\ndef plotline(ax, b, e, *args):\n ax.plot([b[0], e[0]], [b[1], e[1]], *args)\n\n\ndef add_cond_on_segment(optimizer, b, e, x, y):\n optimizer.add((e[0]-b[0])*(y-b[1])-(e[1]-b[1])*(x-b[0]) == 0)\n optimizer.add(min(b[0], e[0]) <= x)\n optimizer.add(max(b[0], e[0]) >= x)\n optimizer.add(min(b[1], e[1]) <= y)\n optimizer.add(max(b[1], e[1]) >= y)\n\n\ndef add_cond_orthogonal(optimizer, lx, ly, cx, cy, rx, ry):\n optimizer.add((lx-cx)*(rx-cx)+(ly-cy)*(ry-cy) ==0)\n\n\ndef convert_py_type(ratnumtype):\n return float(ratnumtype.as_fraction())\n\n\ndef z3abs(x):\n return If(x >= 0, x, -x)\n\n\ndef search_rectangle(b1, e1, b2, e2, b3, e3, b4, e4):\n x1 = Real('x1')\n x2 = Real('x2')\n x3 = Real('x3')\n x4 = Real('x4')\n\n y1 = Real('y1')\n y2 = Real('y2')\n y3 = Real('y3')\n y4 = Real('y4')\n\n optimizer = Optimize()\n\n add_cond_on_segment(optimizer, b1, e1, x1, y1)\n add_cond_on_segment(optimizer, b2, e2, x2, y2)\n add_cond_on_segment(optimizer, b3, e3, x3, y3)\n add_cond_on_segment(optimizer, b4, e4, x4, y4)\n\n add_cond_orthogonal(optimizer, x1, y1, x2, y2, x3, y3)\n add_cond_orthogonal(optimizer, x2, y2, x3, y3, x4, y4)\n add_cond_orthogonal(optimizer, x3, y3, x4, y4, x1, y1)\n add_cond_orthogonal(optimizer, x4, y4, x1, y1, x2, y2)\n\n # equal_distance\n #optimizer.add_soft((x2-x1)**2+(y2-y1)**2==(x4-x3)**2+(y4-y3)**2)\n #optimizer.add_soft((x3-x2)**2+(y3-y2)**2==(x4-x1)**2+(y4-y1)**2)\n\n #optimizer.maximize(z3abs((x2-x1)*(y4-y1)-(x4-x1)*(y2-y1)))\n\n result = optimizer.check()\n\n print(result)\n fig = plt.figure()\n ax = fig.add_subplot(111)\n plotline(ax, b1, e1)\n plotline(ax, b2, e2)\n plotline(ax, b3, e3)\n plotline(ax, b4, e4)\n\n if result != z3.unsat:\n print(optimizer.model())\n model = optimizer.model()\n x1 = convert_py_type(model[x1])\n x2 = convert_py_type(model[x2])\n x3 = convert_py_type(model[x3])\n x4 = convert_py_type(model[x4])\n y1 = convert_py_type(model[y1])\n y2 = convert_py_type(model[y2])\n y3 = convert_py_type(model[y3])\n y4 = convert_py_type(model[y4])\n\n plotline(ax, [x1, y1], [x2, y2], 'k')\n plotline(ax, [x2, y2], [x3, y3], 'k')\n plotline(ax, [x3, y3], [x4, y4], 'k')\n plotline(ax, [x4, y4], [x1, y1], 'k')\n\n else:\n print('no soltion')\n plt.gca().set_aspect('equal', adjustable='box')\n plt.show()\n\n\ndef getexample1():\n b1 = [1, -1]\n e1 = [-1, 1]\n\n b2 = [-1, 2]\n e2 = [1, 4]\n\n b3 = [2, 4]\n e3 = [4, 2]\n\n b4 = [4, 1]\n e4 = [2, -1]\n search_rectangle(b1, e1, b2, e2, b3, e3, b4, e4)\n\n\ndef getexample2():\n b1 = [-1, 0]\n e1 = [1, 0]\n\n b2 = [-2, 1]\n e2 = [-2, 3]\n\n b3 = [-1, 4]\n e3 = [1, 4]\n\n b4 = [2, 1]\n e4 = [2, 3]\n search_rectangle(b1, e1, b2, e2, b3, e3, b4, e4)\n\n\ndef main():\n getexample2()\nif __name__ == '__main__':\n main()\n","sub_path":"optimizer/geoetryModel/inner_rect.py","file_name":"inner_rect.py","file_ext":"py","file_size_in_byte":3033,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"180766163","text":"from pyramid.response import Response\nfrom pyramid.view import view_config\nimport json\nimport sys\nimport traceback\n\n\ndef format_traceback():\n etype, value, tb = sys.exc_info()\n ret = ''.join(traceback.format_exception(etype, value, tb))\n return '
%s
' % ret\n\n\nERROR_PAGE = \"\"\"\nAn error occured
\n\n HOME\n
\n\n%(error)s\n\"\"\"\n\n\n@view_config(context=Exception)\ndef internal_server_error(request):\n tb = format_traceback()\n if not request.is_xhr:\n return Response(ERROR_PAGE % {'error': tb})\n from cone.app.browser.ajax import (\n AjaxContinue,\n AjaxMessage,\n )\n continuation = AjaxContinue([AjaxMessage(tb, 'error', None)]).definitions\n ret = {\n 'mode': 'NONE',\n 'selector': 'NONE',\n 'payload': '',\n 'continuation': continuation,\n }\n response = Response(json.dumps(ret))\n response.content_type = 'application/json'\n return response\n","sub_path":"src/cone/app/browser/exception.py","file_name":"exception.py","file_ext":"py","file_size_in_byte":959,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"447758455","text":"from os import error\nfrom typing import Sized\nfrom sqlalchemy.sql.functions import percent_rank\nfrom elasticsearch import Elasticsearch\nimport numpy as np\nimport pandas as pd\nfrom datetime import datetime\nimport re\n\n\nES_HOST = '192.168.1.7'\nPORT = 9200\n\ndef connect_elasticsearch():\n _es_obj = Elasticsearch(hosts=ES_HOST, port=PORT, timeout=200)\n if _es_obj.ping():\n print('Connect success')\n else:\n print('Not connected!!!')\n return _es_obj\n\n\n\ndef validate_iso8601(str_val):\n is_iso8601 = re.compile('^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}(\\.[0-9]{1,3})?$').match\n try: \n if is_iso8601( str_val ) is not None:\n return True\n except:\n pass\n return False\n\ndef searchall(kwards): \n format_date = ['%d/%m/%Y %H:%M:%S', '%d-%m-%Y %H:%M:%S', '%d-%m-%Y %H%M%S','%d-%m-%Y %H-%M-%S',\\\n '%d%m%Y %H:%M:%S', '%d/%m/%Y %H/%M/%S', '%d/%m/%Y %H%M%S',\\\n '%Y/%m/%d %H:%M:%S','%Y-%m-%d %H:%M:%S','%Y%m%d %H:%M:%S']\n \n _es_obj = connect_elasticsearch()\n field = {}\n #_start = datetime.strptime(kwards['startTime'], '%d/%m/%Y %H:%M:%S').isoformat()\n #_stop = datetime.strptime(kwards['stopTime'], '%d/%m/%Y %H:%M:%S').isoformat()\n _start = None\n _stop = None\n if validate_iso8601(kwards['startTime']):\n _start = kwards['startTime']\n _stop = kwards['stopTime']\n \n else:\n for formater in format_date:\n try:\n datetime.strptime(kwards['startTime'], formater)\n except ValueError as e:\n continue\n else:\n _start = datetime.strptime(kwards['startTime'], formater).isoformat()\n _stop = datetime.strptime(kwards['stopTime'], formater).isoformat() \n \n if not _start or not _stop:\n return f\"{kwards['startTime']} or {kwards['stopTime']} is not reconized\"\n query = {\n \"query\":{\n \"range\":{\n \"startTime\":{\n \"gte\":_start,\n \"lte\":_stop\n }\n }\n }\n }\n \n try:\n data = _es_obj.search(index=kwards['index'],body=query, size=5000)\n except TypeError:\n return pd.DataFrame({})\n else:\n df = query_to_df(data)\n return df\n\n\ndef searchOne(kwards):\n _es_obj = connect_elasticsearch()\n field = {}\n \n format_date = ['%d/%m/%Y %H:%M:%S', '%d-%m-%Y %H:%M:%S', '%d%m%Y %H:%M:%S', \\\n '%Y/%m/%d %H:%M:%S','%Y-%m-%d %H:%M:%S','%Y%m%d %H:%M:%S']\n \n is_iso8601 = re.compile('^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}(\\.[0-9]{1,3})?$').match\n _number = kwards['number']\n \n if validate_iso8601(kwards['startTime']):\n _start = kwards['startTime']\n _stop = kwards['stopTime']\n else:\n for formater in format_date:\n try:\n datetime.strptime(kwards['startTime'], formater)\n except ValueError as e:\n continue\n else:\n _start = datetime.strptime(kwards['startTime'], formater).isoformat()\n \n _stop = datetime.strptime(kwards['stopTime'], formater).isoformat()\n \n query ={\n \"query\": {\n \"bool\": {\n \"must\": [{\n \"term\": {\n \"cust\":_number \n }\n },\n {\n \"range\": {\n \"startTime\": {\n \"gte\": _start,\n \"lt\": _stop\n }\n }\n }\n ]\n }\n }\n }\n try:\n data = _es_obj.search(index=kwards['index'],body=query, size=5000)\n except TypeError:\n return pd.DataFrame({})\n else:\n df = query_to_df(data)\n return df\n\ndef query_to_df(data):\n field = {}\n list_df = []\n for doc in data['hits']['hits']:\n source_data = doc['_source']\n for key, valu in source_data.items():\n try:\n field[key] = np.append(field[key], valu)\n except KeyError as e:\n field[key] = np.array([valu])\n \n df = pd.DataFrame(field)\n df['nBytesUp+nBytesDn'] = df['nBytesUp']+df['nBytesDn']\n print(df['nBytesUp+nBytesDn'])\n return df","sub_path":"backend/src/elasticsearch/elastic.py","file_name":"elastic.py","file_ext":"py","file_size_in_byte":4388,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"167126911","text":"import collections\nimport discord\nfrom plumeria.event import bus\nfrom plumeria.message import ProxyMessage, Response\nfrom plumeria.command import commands, Command, Context, CommandError, Mapping, channel_only\nfrom plumeria.perms import server_admins_only\nfrom plumeria.storage import Session\n\n\nclass AliasManager:\n def __init__(self):\n self.aliases = collections.defaultdict(lambda: {})\n\n def load(self):\n self.aliases.clear()\n session = Session()\n try:\n for row in session.execute(\"SELECT server_id, alias, command FROM alias\").fetchall():\n server_id, alias, command = row\n self.aliases[server_id][alias] = command\n finally:\n session.close()\n\n def create(self, server, alias, command):\n alias = alias.lower()\n session = Session()\n try:\n session.execute(\"REPLACE INTO alias (server_id, alias, command) VALUES (%s, %s, %s)\",\n [server.id, alias, command])\n session.commit()\n finally:\n session.rollback()\n self.aliases[server.id][alias] = command\n\n def delete(self, server, alias):\n alias = alias.lower()\n session = Session()\n try:\n session.execute(\"DELETE FROM alias WHERE server_id = %s AND alias = %s\",\n [server.id, alias])\n session.commit()\n finally:\n session.rollback()\n if alias in self.aliases[server.id]:\n del self.aliases[server.id][alias]\n\n def match_command(self, message, command):\n if command in self.aliases[message.channel.server.id]:\n return self.aliases[message.channel.server.id][command]\n return None\n\n def get_mappings(self, server_id):\n if server_id in self.aliases:\n mappings = []\n for alias, command in self.aliases[server_id].items():\n mappings.append(\n Mapping([alias], Command(None, category=\"(Server-Specific)\", description=command, help=command)))\n return mappings\n else:\n return []\n\n\naliases = AliasManager()\n\n\n@bus.event('init')\nasync def init():\n aliases.load()\n\n\n@commands.register('echo', cost=0.2, category=\"Utility\")\nasync def echo(message):\n \"\"\"\n Simply returns the input string.\n\n Can be used to create an alias::\n\n alias website echo Our website is http://example.com\n \"\"\"\n return Response(message.content)\n\n\n@commands.register('alias', cost=4, category='Alias')\n@channel_only\n@server_admins_only\nasync def alias(message):\n \"\"\"\n Creates a new command alias.\n\n Example::\n\n alias serverinfo a2squery example.com:27015\n\n If commands need to be piped, escape the pipe symbol with a ^::\n\n alias example echo flight simulator ^| drawtext\n \"\"\"\n parts = message.content.split(\" \", 1)\n if len(parts) == 2:\n aliases.create(message.channel.server, parts[0], parts[1])\n await message.respond(\"Created the command alias *{}*.\".format(parts[0]))\n else:\n raise CommandError(\" \")\n\n\n@commands.register('deletealias', cost=4, category='Alias')\n@channel_only\n@server_admins_only\nasync def deletealias(message):\n \"\"\"\n Deletes an alias by name.\n \"\"\"\n if len(message.content):\n aliases.delete(message.channel.server, message.content)\n await message.respond(\"Deleted the command alias *{}*.\".format(message.content))\n else:\n raise CommandError(\"\")\n\n\n@commands.enumerator\nasync def alias_enumerator(server_id):\n if server_id:\n return aliases.get_mappings(server_id)\n else:\n return []\n\n\n@commands.intercept\nasync def alias_listener(message, value, depth):\n if not message.channel.is_private: # public channels only\n value = aliases.match_command(message, value)\n if value:\n message = ProxyMessage(message)\n message.content = value\n return await commands.execute(message, Context(), expect_prefix=False)\n return False\n","sub_path":"plumeria/plugins/alias.py","file_name":"alias.py","file_ext":"py","file_size_in_byte":4070,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"86057008","text":"# -*-coding:utf-8 -*\n\nimport os\n\ndef init_tab():\n\ttab = []\n\tfor l in range(0, 3):\n\t\ttab.append([])\n\t\ttab[l] = [0 for i in range(0, 3)]\n\treturn tab\n\ndef ft_get_val_tab(tab):\n\tls = [0,0,0,0,0,0,0,0]\n\tfor x in range(0, 3):\n\t\tfor y in range(0, 3):\n\t\t\tls[x] += tab[x][y]\n\t\t\tls[x+3] += tab[y][x]\n\t\tls[6] += tab[x][x]\n\t\tls[7] += tab[2-x][x]\n\treturn ls\n\ndef ft_game_is_over(tab, nb_tour):\n\tcount1 = 0\n\tcount2 = 0\n\tif nb_tour > 4:\n\t\tval = ft_get_val_tab(tab)\n\t\tfor i, nb in enumerate(val):\n\t\t\tcount1 += nb - 1 if nb > 1 else 0\n\t\t\tcount2 += (-nb) - 1 if nb < -1 else 0\n\t\tprint(count1, count2, nb_tour)\n\t\tif count1 > 1:\n\t\t\treturn 1\n\t\tif count2 > 1:\n\t\t\treturn 2\n\t\tif nb_tour > 7:\n\t\t\treturn -1\n\treturn 0\n\ndef get_input_pos():\n\tpos = [-1,-1]\n\tfor i,entier in enumerate(pos):\n\t\terror = 1\n\t\twhile error == 1:\n\t\t\terror = 0\n\t\t\tr = input(\"entrer la {} sur laquel vous-voulez jouer\\n>>>\"\\\n\t\t\t\t.format(\"ligne\" if i == 0 else \"colonne\"))\n\t\t\ttry:\n\t\t\t\tpos[i] = int(r)\n\t\t\t\tif pos[i] < 1 or pos[i] > 3:\n\t\t\t\t\traise ValueError(\"le nb de la {} doit etre compris entre 1 et 3\"\\\n\t\t\t\t\t\t.format(\"ligne\" if i == 0 else \"colonne\"))\n\t\t\texcept Exception as e:\n\t\t\t\tprint(e)\n\t\t\t\terror = 1\n\treturn pos[0] - 1, pos[1] - 1\n\ndef ft_get_pos(tab):\n\tfree = 0\n\twhile free != 1:\n\t\tx, y = get_input_pos()\n\t\tif tab[x][y] != 0:\n\t\t\tprint(\"les coordonnée {} sont dejas prise.\\nVellez les re-saisir\"\\\n\t\t\t\t.format((x + 1, y + 1)))\n\t\telse:\n\t\t\tfree = 1\n\treturn x, y\n\ndef ft_place_coor(tab, pos, joueur1):\n\tx, y = pos\n\ttab[x][y] = 1 if joueur1 == True else -1\n\ndef ft_dysplay_tab(tab):\n\tsrc = \"\"\n\tfor i,line in enumerate(tab):\n\t\tsrc += \" - - -\\n|\"\n\t\tfor y, nb in enumerate(line):\n\t\t\tif nb == 0:\n\t\t\t\tsrc += \" |\"\n\t\t\telse :\n\t\t\t\tsrc += (\"X\" if nb == 1 else \"O\") + \"|\"\n\t\tsrc += \"\\n\"\n\tsrc += \" - - -\\n\"\n\tprint(src)\n\ndef main():\n\ttab = init_tab()\n\tjoueur1 = False\n\tgame = 0\n\tnb_tour = 0\n\twhile game == 0:\n\t\tnb_tour += 1\n\t\tjoueur1 = False if joueur1 == True else True\n\t\tprint(\"au tour du joueur {} de jouer\".format(1 if joueur1 == True else 2))\n\t\tft_dysplay_tab(tab)\n\t\tx, y = ft_get_pos(tab)\n\t\tft_place_coor(tab, (x, y), joueur1)\n\t\tgame = ft_game_is_over(tab, nb_tour)\n\tif game < 0:\n\t\tprint(\"égalité\")\n\telse:\n\t\tprint(\"le jouer {} à gagne\".format(game))\n\tos.system(\"pause\")\n\nif __name__ == '__main__':\n\tmain()","sub_path":"projet/morpion/morpion.py","file_name":"morpion.py","file_ext":"py","file_size_in_byte":2246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"466434564","text":"\"\"\"primer_proyecto_django URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/3.2/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.urls import path\nfrom . import views\n\nurlpatterns = [\n path('', views.root, name='index'),\n path(\"blogs\", views.index, name=\"blogs\"),\n path(\"blogs/new\", views.new, name=\"new\"),\n path(\"blogs/create\", views.create, name=\"create\"),\n path(\"blogs/\", views.show, name=\"show\"),\n path(\"blogs//edit\", views.edit, name=\"edit\"),\n path(\"blogs//delete\", views.destroy, name=\"delete\"),\n path(\"blogs/json\", views.json, name=\"json\"),\n\n]\n","sub_path":"core/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"457262966","text":"import time\nimport datetime\nimport json\n\n\nclass DBObject(object):\n \"\"\"This class defines how objects in the mongoDB should be represented.\"\"\"\n\n def __init__(\n self,\n classes=[],\n type='',\n meta={}\n ):\n self.created = datetime.datetime.now()\n self.updated = datetime.datetime.now()\n self.classes = classes\n self.type = type\n self.meta = meta\n self.structure = '{}{}'.format('#', self.__class__.__name__)\n\n\n def export(self):\n \"\"\"This function exports the object as a dict. This is used when\n putting an object in the database.\"\"\"\n\n return self.__dict__\n\nclass Media(DBObject):\n def __init__(\n self,\n path=None,\n *args,\n **kwargs\n ):\n DBObject.__init__(self, *args, **kwargs)\n self.path = path\n\nclass Option(DBObject):\n def __init__(\n self,\n key=None,\n value=None,\n *args,\n **kwargs\n ):\n DBObject.__init__(self, *args, **kwargs)\n self.key = key\n self.value = value\n\nclass User(DBObject):\n def __init__(\n self,\n username=None,\n email=None,\n password=None,\n media_id=None,\n description=None,\n *args,\n **kwargs\n ):\n DBObject.__init__(self, *args, **kwargs)\n self.username = username\n self.email = email\n self.password = password\n self.media_id = media_id\n self.description = description\n\nclass Location(DBObject):\n def __init__(\n self,\n city=None,\n name=None,\n *args,\n **kwargs\n ):\n DBObject.__init__(self, *args, **kwargs)\n self.city = city\n self.name = name\n\nclass Event(DBObject):\n def __init__(\n self,\n user_id=None,\n location_id=None,\n users=None,\n max_users=1,\n hour=None,\n minute=None,\n *args,\n **kwargs\n ):\n DBObject.__init__(self, *args, **kwargs)\n self.user_id = user_id\n self.location_id = location_id\n self.users = users\n self.max_users = max_users\n self.hour = hour\n self.minute = minute\n","sub_path":"meetndrink/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":2381,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"415753339","text":"import logging\nimport ujson\nfrom uuid import UUID\n\nfrom django.http import HttpRequest, HttpResponse, HttpResponseBadRequest, HttpResponseNotFound\n\n# Create your views here.\nfrom gameapi.games_manager import DoesNotExist, game_manager\nfrom gameapi.models import Game, Token, Card\n\nlogger = logging.getLogger(__name__)\n\n\ndef game_auth(fn):\n def game_auth_wrapper(request, game_id, *args, **kwargs):\n try:\n game = check_game_id(game_id)\n except DoesNotExist:\n return HttpResponseNotFound('Game does not exist')\n try:\n token_str = request.GET['token']\n except KeyError:\n return HttpResponseBadRequest('No token provided')\n try:\n token = check_token_exists(token_str)\n except Token.DoesNotExist:\n return HttpResponseBadRequest('Token invalid')\n\n if not check_token_in_game(game, token.token):\n logger.warning('Token is not part of this game. (%s not in %s)', token, game.players)\n return HttpResponseBadRequest('Token is not part of this game.')\n return fn(request, *args, game=game, token=token, **kwargs)\n\n return game_auth_wrapper\n\n\ndef check_game_id(game_id: UUID):\n return game_manager.get_game(game_id)\n\n\ndef check_token_exists(token: str):\n return Token.objects.get(token=token)\n\n\ndef check_token_in_game(game: Game, token: Token):\n return token in game.players\n\n\n@game_auth\ndef get_state(request: HttpRequest, game: Game, token: Token):\n logger.debug('/get_state handler')\n state = game.get_state(token.token)\n logger.debug(state)\n return HttpResponse(\n content=ujson.dumps(state),\n )\n\n\n@game_auth\ndef take_action(request: HttpRequest, game: Game, token: Token):\n logger.debug('/take_action handler')\n if 'action' not in request.GET:\n return HttpResponseBadRequest('No action provided')\n action_str = request.GET['action']\n card_str = request.GET.get(key='card', default=None)\n\n uuid = token.token\n action = Game.Action(action_str)\n card = Card.from_string(card_str) if card_str else None\n action_accepted = True\n try:\n game.take_action(uuid, action, card)\n except Game.ActionNotAllowed:\n logger.warning(\n 'Action not allowed %s %s',\n action, token,\n )\n action_accepted = False\n except Game.ActionInvalid:\n logger.warning(\n 'Action invalid %s %s',\n action, token,\n )\n action_accepted = False\n\n new_state = game.get_state(uuid)\n new_state.update({\n 'action_accepted': action_accepted,\n })\n logger.debug(new_state)\n return HttpResponse(content=ujson.dumps(new_state))\n","sub_path":"gameserver/gameapi/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"494392049","text":"import re\nimport os\n\n\ndef unformat(test_string, format_string):\n x = re.search('^' + format_string.format('(.*)') + '$', test_string,flags=re.DOTALL)\n return None if x is None else x.group(1)\n\ndef clear():\n os.system('cls' if os.name == 'nt' else 'clear')\n\nclass _Getch:\n \"\"\"\n Gets a single character from standard input. \n \n Does not echo to the screen.\n \"\"\"\n\n def __init__(self):\n try:\n self.impl = _GetchWindows()\n except ImportError:\n self.impl = _GetchUnix()\n\n def __call__(self): \n char = self.impl()\n if char == '\\x03':\n raise KeyboardInterrupt\n elif char == '\\x04':\n raise EOFError\n return char\n\n\n\nclass _GetchUnix:\n def __init__(self):\n import tty, sys\n\n def __call__(self):\n import sys, tty, termios\n fd = sys.stdin.fileno()\n old_settings = termios.tcgetattr(fd)\n try:\n tty.setraw(sys.stdin.fileno())\n ch = sys.stdin.read(1)\n finally:\n termios.tcsetattr(fd, termios.TCSADRAIN, old_settings)\n return ch\n\n\nclass _GetchWindows:\n def __init__(self):\n import msvcrt\n\n def __call__(self):\n import msvcrt\n return msvcrt.getch()\n\ngetch = _Getch()\n","sub_path":"utilities.py","file_name":"utilities.py","file_ext":"py","file_size_in_byte":1281,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"488837108","text":"import os\r\nfrom Games import TicTacToe\r\nfrom File_storage import *\r\n\r\nEPSILON = 0.1\r\nSANITY_CHECK = True\r\ncwd = os.getcwd()\r\ncwd = cwd + '\\\\tensorflow_logs'\r\n\r\ndef init():\r\n # create game\r\n global game\r\n game = TicTacToe()\r\n\r\n # initialize Monte Carlo tree\r\n global mct\r\n mct = load_mct()\r\n if mct == []:\r\n mct.append(Node(game))\r\n \r\n","sub_path":"const.py","file_name":"const.py","file_ext":"py","file_size_in_byte":365,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"610274554","text":"import sys\nfrom PyQt4 import QtGui, QtCore\nfrom PyQt4.QtGui import *\nimport socket\n\n\nclass Window(QtGui.QMainWindow):\n\n\tdef __init__(self):\n\t\tsuper(Window, self).__init__()\n\t\tself.setGeometry(50, 50, 100, 115)\n\t\tself.setWindowTitle(\"Command server!\")\n\t\tself.setWindowIcon(QtGui.QIcon('pythonlogo.png'))\n\n\n\t\tself.my_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\t\tself.my_socket.connect(('127.0.0.1', 1729))\n\n\t\textractAction = QtGui.QAction(\"&Exit the app\", self)\n\t\textractAction.setShortcut(\"Ctrl+Q\")\n\t\textractAction.setStatusTip('Leave The App')\n\t\textractAction.triggered.connect(self.close_application)\n\n\t\tself.home()\n\n\tdef home(self):\n\t\tbtn = QtGui.QPushButton(\"Quit\", self)\n\t\tbtn.clicked.connect(self.close_application)\n\t\tbtn.resize(btn.minimumSizeHint())\n\t\tbtn.move(0, 80)\n\n\t\tprintScrnAction = QtGui.QAction(QtGui.QIcon('images/camera.png'), 'Print the screen', self)\n\t\tprintScrnAction.triggered.connect(self.print_screen)\n\t\tself.toolBar = self.addToolBar(\"Take a screencap\")\n\t\tself.toolBar.addAction(printScrnAction)\n\n\t\tdirAction = QtGui.QAction(QtGui.QIcon('images/folder.png'), 'Print contents of directory', self)\n\t\tdirAction.triggered.connect(self.directory)\n\t\tself.toolBar = self.addToolBar(\"Print contents of directory\")\n\t\tself.toolBar.addAction(dirAction)\n\n\t\tdeleteAction = QtGui.QAction(QtGui.QIcon('images/delete.png'), 'Deletes a file', self)\n\t\tdeleteAction.triggered.connect(self.delete)\n\t\tself.toolBar = self.addToolBar(\"Delete a file\")\n\t\tself.toolBar.addAction(deleteAction)\n\n\t\tcopyAction = QtGui.QAction(QtGui.QIcon('images/copy.png'), 'Copies a file', self)\n\t\tcopyAction.triggered.connect(self.copy)\n\t\tself.toolBar = self.addToolBar(\"Copy a file\")\n\t\tself.toolBar.addAction(copyAction)\n\n\t\texecAction = QtGui.QAction(QtGui.QIcon('images/execute.png'), 'Executes a file', self)\n\t\texecAction.triggered.connect(self.execute)\n\t\tself.toolBar = self.addToolBar(\"Execute a file\")\n\t\tself.toolBar.addAction(execAction)\n\n\t\tcolor = QtGui.QColor(0, 0, 0)\n\n\t\tfontColor = QtGui.QAction('Font bg Color', self)\n\t\tfontColor.triggered.connect(self.color_picker)\n\t\tself.toolBar.addAction(fontColor)\n\n\t\tself.styleChoice = QtGui.QLabel(\"\", self)\n\t\tself.styleChoice.move(0, 0)\n\t\tself.styleChoice.setGeometry(1, 1, 800, 400)\n\t\tcomboBox = QtGui.QComboBox(self)\n\t\tcomboBox.addItem(\"motif\")\n\t\tcomboBox.addItem(\"Windows\")\n\t\tcomboBox.addItem(\"cde\")\n\t\tcomboBox.addItem(\"Plastique\")\n\t\tcomboBox.addItem(\"Cleanlooks\")\n\n\t\tcomboBox.move(0, 50)\n\t\tself.styleChoice.move(0, 150)\n\t\tcomboBox.activated[str].connect(self.style_choice)\n\n\t\tself.show()\n\n\tdef color_picker(self):\n\t\tcolor = QtGui.QColorDialog.getColor()\n\t\tself.styleChoice.setStyleSheet(\"QWidget { background-color: %s}\" % color.name())\n\n\tdef style_choice(self, text):\n\t\tself.styleChoice.setText(text)\n\t\tQtGui.QApplication.setStyle(QtGui.QStyleFactory.create(text))\n\n\tdef close_application(self):\n\t\tchoice = QtGui.QMessageBox.question(self, 'Qutting!', \"Are you sure you want to exit the app?\", QtGui.QMessageBox.Yes | QtGui.QMessageBox.No)\n\t\tif choice == QtGui.QMessageBox.Yes:\n\t\t\tsys.exit()\n\t\telse:\n\t\t\tpass\n\n\tdef print_screen(self):\n\t\tself.send_request_to_server(self.my_socket, \"TAKE_SCREENSHOT\")\n\t\tself.send_request_to_server(self.my_socket, \"SEND_FILE B:/newimage.jpg\")\n\t\tnewfile = open('B:\\\\newimage.jpg', 'wb')\n\t\twhile True:\n\t\t\tdata = self.my_socket.recv(1024)\n\t\t\tnewfile.write(data)\n\t\t\tif \"Image sent!\" in data:\n\t\t\t\tbreak\n\t\t\tpopup = QtGui.QMessageBox.information(self, 'Image recieved!', 'The image has been recevied.')\n\n\tdef directory(self):\n\t\ttext, ok = QInputDialog.getText(self, 'Enter your desired directory', 'Please entire your desired directory:')\n\t\tif ok and text != '':\n\t\t\ttext = 'DIR ' + text\n\t\t\tself.send_request_to_server(self.my_socket, str(text))\n\t\t\tpopup = QtGui.QMessageBox.information(self, 'Contents:', str(self.my_socket.recv(1024)))\n\n\tdef delete(self):\n\t\ttext, ok = QInputDialog.getText(self, 'Enter your desired file', 'Please entire your desired file:')\n\t\tif ok and text != '':\n\t\t\ttext = 'DELETE ' + text\n\t\t\tself.send_request_to_server(self.my_socket, str(text))\n\t\t\tpopup = QtGui.QMessageBox.information(self, 'Deleted!', 'File has been deleted!')\n\n\tdef copy(self):\n\t\ttext1, ok1 = QInputDialog.getText(self, 'Enter your desired file', 'Please entire your desired file:')\n\t\ttext2, ok2 = QInputDialog.getText(self, 'Enter your desired destination', 'Please entire your desired destination:')\n\t\tif ok1 and ok2 and text1 != '' and text2 != '':\n\t\t\ttext = 'COPY ' + text1 + \" \" + text2\n\t\t\tself.send_request_to_server(self.my_socket, str(text))\n\t\t\tpopup = QtGui.QMessageBox.information(self, 'Copied!', 'File has been copied!')\n\n\tdef execute(self):\n\t\ttext, ok = QInputDialog.getText(self, 'Enter your desired file', 'Please entire your desired file:')\n\t\tif ok and text != '':\n\t\t\ttext = \"EXECUTE \" + text\n\t\t\tself.send_request_to_server(self.my_socket, str(text))\n\t\t\tpopup = QtGui.QMessageBox.information(self, 'Executed!', 'File has been executed!')\n\n\t@staticmethod\n\tdef send_request_to_server(my_socket, request):\n\t\tmy_socket.send(request)\n\napp = QtGui.QApplication(sys.argv)\nGUI = Window()\nsys.exit(app.exec_())\n","sub_path":"gui.py","file_name":"gui.py","file_ext":"py","file_size_in_byte":5092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"137488035","text":"\"\"\"\nProject 1 - Problem 2 (P1)\n\"\"\"\n\n########################################################\n######## X-MAC: Trade_off Energy with Delay using GT\n########################################################\n# Radio subsystem varaible definition\n\nP = 32. # Payload [byte]\nR = 31.25 # CC2420 Radio Rate [kbyte/s = Byte/ms]\nD = 8 # number of levels\nC = 5 # neighbors size (connectivity)\nN = C*D**2 # number of nodes\n#### BE CAREFUL: Times are in milliseconds (ms)\nLmax = 5000. # Maximal allowed Delay (ms)\nEmax = 1. # MAximal Energy Budjet (J)\n\nL_pbl = 4. # preamble length [byte]\nL_hdr = 9. + L_pbl # header length [byte]\nL_ack = 9. + L_pbl # ACK length [byte]\nL_ps = 5. + L_pbl # preamble strobe length [byte]\n\nTal = 0.95 # ack listen period [ms]\nThdr = L_hdr/R # header transmission duration [ms]\nTack = L_ack/R # ACK transmission duration [ms]\nTps = L_ps/R # preamble strobe transmission duration [ms]\nTcw = 15*0.62 # Contention window size [ms]\nTcs = 2.60 # Time [ms] to turn the radio into TX and probe the channel (carrier sense)\nTdata = Thdr + P/R + Tack # data packet transmission duration [ms]\n\n### Sampling frequency\n# Fs = 1.0/(60*30*1000) # e.g. Min traffic rate 1 pkt/half_hour = 1/(60*30*1000) pk/ms\nFs = 1./(5*60*1000) #default\n\n# Sleep period: Parameter Bounds\nTw_max = 500. # Maximum Duration of Tw in ms\nTw_min = 100. # Minimum Duration of Tw in ms\n\n\ndef Nd(d):\n if d == 0:\n return 1\n return (2*d - 1)*C\n\ndef Id(d):\n if d == 0:\n return C\n elif d == D:\n return 0\n return (2*d + 1)/(2*d - 1)\n # return Nd(d+1)/Nd(d)\n\ndef Fout(d):\n if d == D:\n return Fs\n return Fs*(D**2 - d**2 + 2*d - 1)/(2*d - 1)\n # return FI(d)+Fs\n\ndef FI(d):\n if d == 0:\n return Fs*C*D**2\n return Fs*(D**2 - d**2)/(2*d - 1)\n # return Id(d)*Fout(d+1)\n\ndef FB(d):\n return (C - Id(d))*Fout(d)\n\ndef Etx(d, Tw):\n return (Tcs + Tal + Ttx(Tw))*Fout(d)\n\ndef Ttx(Tw):\n return (Tw/2 + Tack + Tdata)\n\n# d = 1 # first ring\n# α1 = Tcs + Tal + (3/2)*Tps*((Tps + Tal)/2 + Tack + Tdata)*FB(d)\n# α2 = Fout(d)/2\n# α3 = ((Tps + Tal)/2 + Tcs + Tal + Tack + Tdata)*Fout(d) + ((3/2)*Tps + Tack + Tdata)*FI(d) + (3/4)*Tps*FB(d)\n\n# d = D # last ring\n# β1 = sum(1/2 for i in range(1, d+1))\n# β2 = sum(Tcw/2 + Tdata for i in range(1, d+1))\n\n# # print(α1, α2, α3, β1, β2)\n\n# E = lambda Tw: α1/Tw + α2*Tw + α3\n# L = lambda Tw: β1*Tw + β2\n\n#########\n# Plots #\n#########\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport random\n\n# (a) The energy as a function of Tw, where Tw=[Twmin, Twmax] = [100, 500] ms\n# Tw = np.linspace(Tw_min/1000,Tw_max/1000,100)\n# Tw = np.linspace(Tw_min,Tw_max,100)\nTw_linsp = np.linspace(70,500,100)\n# Tw_linsp = np.linspace(70/1000,500/1000,100)\n\nfig = plt.figure()\nax = fig.add_subplot(1, 1, 1)\nax.spines['left'].set_position('center')\nax.spines['bottom'].set_position('zero')\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\n\n# plt.plot(Tw, E(Tw), 'blue', label='Energy E(Tw)')\n# plt.legend(loc='lower left')\n# plt.xlabel(\"Tw\")\n# plt.show()\n\n# minute = [1,5,10,15,20,25,30] #minutes\n# Fs = 1./(minute[1]*60*1000)\n\n# d = 1 # first ring\n# α1 = Tcs + Tal + (3/2)*Tps*((Tps + Tal)/2 + Tack + Tdata)*FB(d)\n# α2 = Fout(d)/2\n# α3 = ((Tps + Tal)/2 + Tcs + Tal + Tack + Tdata)*Fout(d) + ((3/2)*Tps + Tack + Tdata)*FI(d) + (3/4)*Tps*FB(d)\n\n# d = D # last ring\n# β1 = sum(1/2 for i in range(1, d+1))\n# β2 = sum(Tcw/2 + Tdata for i in range(1, d+1))\n\n# E = lambda Tw: α1/Tw + α2*Tw + α3\n# L = lambda Tw: β1*Tw + β2\n\n\n\n# %%\n################\n# Optimization #\n################\n\n# def P1():\nfrom gpkit import Variable, VectorVariable, Model\nfrom gpkit.nomials import Monomial, Posynomial, PosynomialInequality\n\nLmax_range = [1000,2000,3000,4000,5000]\n\nfor minute in [5,10,15,20,25,30]: #minutes\n Fs = 1./(minute*60*1000)\n\n d = 1 # first ring\n α1 = Tcs + Tal + (3/2)*Tps*((Tps + Tal)/2 + Tack + Tdata)*FB(d)\n α2 = Fout(d)/2\n α3 = ((Tps + Tal)/2 + Tcs + Tal + Tack + Tdata)*Fout(d) + ((3/2)*Tps + Tack + Tdata)*FI(d) + (3/4)*Tps*FB(d)\n\n d = D # last ring\n β1 = sum(1/2 for i in range(1, d+1))\n β2 = sum(Tcw/2 + Tdata for i in range(1, d+1))\n\n E = lambda Tw: α1/Tw + α2*Tw + α3\n L = lambda Tw: β1*Tw + β2\n\n Tw = Variable(\"Tw\")\n\n objective = α1/Tw + α2*Tw + α3\n constraints = [ β1*Tw + β2 <= 5000,\n Tw >= Tw_min,\n Id(0)*Etx(1,Tw) <= 1/4\n ]\n m = Model(objective, constraints)\n sol = m.solve(verbosity=0)\n print(str(minute) + 'min:')\n print(sol.table())\n # print(sol[\"cost\"])\n # print(sol[\"variables\"][Tw])\n\n # Plot Energy(Tw)\n plt.plot(Tw_linsp, E(Tw_linsp), random.choice(\"bgrcmyk\"), label='Energy E(Tw) for Fs(' + str(minute) + 'min) = ' + str(Fs)[:4] + str(Fs)[-4:])\n\n # Plot Energy(Tw) optimal point\n ax.scatter(sol[\"variables\"][Tw], sol[\"cost\"], color='red')\n\nplt.title('Lmax = ' + str(5000))\nplt.xlabel(\"Tw [ms]\")\nplt.legend(loc='upper right')\nplt.show()\n\n# %%\n","sub_path":"ex2_p1.py","file_name":"ex2_p1.py","file_ext":"py","file_size_in_byte":5280,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"43390372","text":"import pyrebase\n\nconfig = {\n \"apiKey\": \"AIzaSyDKstoSby1YdpTfy7xqAiDPt5Ta50PoOIw\",\n \"authDomain\": \"nomad-e1934.firebaseapp.com\",\n \"databaseURL\": \"https://nomad-e1934.firebaseio.com\",\n \"projectId\": \"nomad-e1934\",\n \"storageBucket\": \"nomad-e1934.appspot.com\",\n}\n\nfirebase = pyrebase.initialize_app(config)\n\n# Database Variable\ndb = firebase.database()\n\ncurrentState = ''\n\ndef getState():\n state = db.child(\"PiMove\").get()\n currentState = state\n\ndef stream_handler(message):\n print(message[\"event\"])\n print(message[\"path\"]) \n print(message[\"data\"])\n\n\nmy_stream = db.child(\"PiMove\").child(\"Movement\").stream(stream_handler)","sub_path":"nomadBackend/python/nomadBrain.py","file_name":"nomadBrain.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"264556647","text":"#!/usr/bin/python\n\nimport csv\n \nencabezado = True\n\nwith open('archivo.csv') as File:\n reader = csv.reader(File, delimiter=',', quotechar=',',quoting=csv.QUOTE_MINIMAL)\n for row in reader:\n if (encabezado == False):\n if float(row[4])>1200 :\n print(row[1])\n encabezado = False","sub_path":"computacion2/exercises/ejercicio3.py","file_name":"ejercicio3.py","file_ext":"py","file_size_in_byte":320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"352247953","text":"from django.conf.urls import patterns, url\nfrom rmc import views\n\nurlpatterns = patterns('',\n\turl(r'^$', views.home, name='home'),\n\turl(r'^chef/', views.chef, name='chef'),\n\turl(r'^events/', views.events, name='events'),\n\turl(r'^menus/', views.menus, name='menus'),\n\turl(r'^inquiry/', views.inquiry, name='inquiry'),\n)","sub_path":"rmc/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"138655545","text":"import tweepy\n\n# consumer key and secret\nconsumer_key = 'Lfz4zIl81lxWH2JsnjYRjfSSy'\nconsumer_secret = 'UJOKsHAGOh44V8rwp7oRH5ljeXhR2ph2vNd85FpKH0jxLJm1OE'\n\n# access token and secret\naccess_token = '746872621-vGxyYKMQvIMVC2qyNnexlGllJZGJ1gS9C0GynFKc'\naccess_token_secret = 'TgNBfyCs92r1NMDxyR1rL00dqY8FAXTwYV8yFXJFhzmMQ'\n\n# connecting for authentication\nauth = tweepy.OAuthHandler(consumer_key, consumer_secret)\n# OAuthHandler equipped with an access token\nauth.set_access_token(access_token, access_token_secret)\n\napi = tweepy.API(auth)\n\nget_data = api.search('@Abhishekkumbhaj')\n\nfor i in get_data:\n print(i.text)\n","sub_path":"TextBlob__Tweepy/tweepy_demo.py","file_name":"tweepy_demo.py","file_ext":"py","file_size_in_byte":618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"463323691","text":"import csv\nimport qrcode\nimport base64\nfrom io import BytesIO\n\n\ndef save_iterable_to_a_file(iterable, filename):\n f = open(filename, \"w+\")\n [f.write(f\"{x}\\n\") for x in iterable]\n f.close()\n\n\ndef read_data(file_name):\n \"\"\"\nfrom applications.core.utils import read_data\nread_data(\"fixtures/info.csv\")\n :param file_name:\n :return:\n \"\"\"\n info_file = open(file_name)\n reader = csv.reader(info_file)\n all_rows = list()\n for r in reader:\n all_rows.append(r)\n\n return all_rows\n\n\ndef extract_categories(file_name):\n \"\"\"\nfrom applications.core.utils import extract_categories\nextract_categories(\"fixtures/info.csv\")\n :param file_name:\n :return:\n \"\"\"\n all_rows = read_data(file_name)\n categories = {r[3].lower() for r in all_rows}\n subcategories = {r[4].lower() for r in all_rows}\n categories_attached = {f\"{r[3].lower()} - {r[4].lower()}\" for r in all_rows}\n save_iterable_to_a_file(categories, \"fixtures/categories.txt\")\n save_iterable_to_a_file(subcategories, \"fixtures/subcategories.txt\")\n save_iterable_to_a_file(categories_attached, \"fixtures/categories_attached.txt\")\n\n\ndef get_base64_qr_code(data):\n qr_image = qrcode.QRCode(\n version=1,\n error_correction=qrcode.constants.ERROR_CORRECT_L,\n box_size=10,\n border=4,\n )\n qr_image.add_data(data)\n qr_image.make(fit=True)\n\n img = qr_image.make_image()\n\n buffered_image = BytesIO()\n img.save(buffered_image, format=\"JPEG\")\n return base64.b64encode(buffered_image.getvalue()).decode()","sub_path":"applications/core/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"135274231","text":"import sys\nimport os\nimport cv2\nimport os.path as osp\nimport numpy as np\nimport time\nimport pdb\nfrom collections import OrderedDict\n\nimport torch\nimport torch.nn as nn\nfrom torch.utils import data\nfrom torch.autograd import Variable\nimport torch.optim as optim\nimport torch.backends.cudnn as cudnn\n\n# from deeplab.model_dplv3_plus_meta_v3 import Res_Dplv3_Decoder \nfrom deeplab.model_dplv3_plus_meta_v2 import Res_Dplv3_Decoder \nfrom deeplab.datasets import DavisTestDataset\n# from new_loss import class_balanced_cross_entropy_loss as balanced_bce_loss \n\nfrom mvos_utils import *\nimport pdb\n\n# gpu setting \nos.environ['CUDA_VISIBLE_DEVICES'] = '0'\ncudnn.enabled = True\ncudnn.benchmark = True\n\n# meta learning setting\nlayer5_bn_name = ['layer5.assp.0.1', 'layer5.assp.1.1', 'layer5.assp.2.1', 'layer5.assp.3.1', \n 'layer5.image_pool.2', 'layer5.encode_fea.1', 'layer5.low_level_fea.1',\n 'layer5.decode_1.1', 'layer5.decode_2.1']\ntrain_mode = 'block_4_5'\nmeta_restore_from = './snapshots/meta_davis17.pth'\nnormal_restore_from = './snapshots/base_seg.pth'\nn_meta_init = 50\nname_suffix = ''\n\n\n# dataset setting\ncfg = {}\ncfg['img_root'] = './dataset/davis/JPEGImages/480p'\ncfg['gt_root'] = './dataset/davis/Annotations/480p_split'\ncfg['list_root'] = './dataset/davis/ImageSets/2017/val_w_imgs_no.txt'\ncfg['crop_size'] = (480, 854)\ncfg['img_mean'] = np.array((104.00698793,116.66876762,122.67891434), dtype=np.float32)\ncfg['random_scale'] = False\ncfg['random_mirror'] = False\ncfg['ignore_label'] = 0\ncfg['num_steps'] = n_meta_init\ncfg['batch_size'] = 1\ncfg['max_label'] = 1.0\n\n\n# get sequences\nimg_ids = [i_id.strip() for i_id in open(cfg['list_root'])]\nseq_numbers = len(img_ids)\n\n\n# create network\nmodel = Res_Dplv3_Decoder(num_classes=1, output_stride=16)\ninterp = nn.Upsample(size=cfg['crop_size'], mode='bilinear')\ncriterion = nn.BCEWithLogitsLoss().cuda()\nsigmoid = nn.Sigmoid()\n\n\n# testing save root\nmeta_model_name = meta_restore_from.split('/')[-1][:-4]\nsave_root_prefix = '_'.join(['./results/davis-17-val/' + meta_model_name, str(n_meta_init), name_suffix])\n\ntotal_runtime = 0.\n# process sequence\nfor i in range(seq_numbers):\n cur_seq_name = img_ids[i].split()[0]\n cur_seq_range = int(img_ids[i].split()[1])\n cur_seq_image_path = osp.join(cfg['img_root'], cur_seq_name)\n cur_seq_label_path = osp.join(cfg['gt_root'], cur_seq_name)\n\n # set learnable parameters\n train_all_features = set_trainable_feas(train_mode)\n model.set_learnable_params(train_all_features, layer5_bn_name)\n\n # load non-meta-trained model\n saved_state_dict = torch.load(normal_restore_from)\n model.load_state_dict(saved_state_dict)\n bn_params = get_bn_params(saved_state_dict, train_all_features, layer5_bn_name)\n\n # load meta-trained model\n saved_state_dict = torch.load(meta_restore_from)\n meta_init = saved_state_dict['meta_init']\n meta_alpha = saved_state_dict['meta_alpha']\n model.copy_meta_weights(meta_init)\n model.train()\n model.cuda()\n\n # meta-initialize the model\n model_weights = OrderedDict()\n for name, p in model.named_parameters():\n if name.split('.')[0] in train_all_features and name not in bn_params:\n model_weights[name] = p\n\n # frozen bn or not\n for name, m in model.named_modules():\n if isinstance(m, nn.BatchNorm2d):\n m.eval()\n m.weight.requires_grad = False\n m.bias.requires_grad = False\n\n ol_trainloader = data.DataLoader(DavisTestDataset(cfg, cur_seq_name), batch_size=cfg['batch_size'],\n shuffle=True, num_workers=0, pin_memory=True)\n\n # meta-finetune on first frame\n tic = time.time()\n print (\"Meta-Training on %s...\" % cur_seq_name)\n for i_iter, batch in enumerate(ol_trainloader):\n image, label = batch\n image = Variable(image).cuda()\n label = Variable(label).cuda()\n pred = interp(model.forward(image, model_weights, train_mode, bn_params))\n\n loss = criterion(pred, label)\n grads = torch.autograd.grad(loss, model_weights.values(), retain_graph=False, create_graph=True)\n model_weights = OrderedDict((name, param - torch.mul(meta_alpha,grad)) for \n ((name, param), (_, meta_alpha), grad) in\n zip(model_weights.items(), meta_alpha.items(), grads))\n #print ('loss is %f' % loss.data.cpu().numpy())\n \n #if i_iter == n_meta_init:\n # break\n\n print ('Meta fine-tuning time is : %.2f' % (time.time() - tic))\n\n # update model parameters\n model.copy_meta_weights(model_weights)\n\n\n # online testing\n model.eval()\n print (\"Evaluate on %s...\" % cur_seq_name)\n cur_save_root = save_root_prefix + '/' + cur_seq_name + '/'\n if not osp.exists(cur_save_root):\n os.makedirs(cur_save_root)\n\n output = label.data[0, 0].cpu().numpy()\n output = np.array(output*255, dtype=np.uint8)\n cv2.imwrite(cur_save_root + '00000.png', output)\n for j in range(1, cur_seq_range):\n # tic = time.time()\n img_file = osp.join(cur_seq_image_path, '%05d' % j + '.jpg')\n img, ori_img_size = load_image_label_davis17(img_file, cfg['crop_size'])\n output = model(Variable(img, volatile = True).cuda())\n output = sigmoid(interp(output)).cpu().data[0,0].numpy()\n # print ('Per frame forward time is : %.2f' % (time.time() - tic))\n if not (ori_img_size==cfg['crop_size']):\n #output = cv2.resize(output, ori_img_size[::-1], interpolation = cv2.INTER_NEAREST)\n output = cv2.resize(output, ori_img_size[::-1], interpolation = cv2.INTER_LINEAR)\n output = np.array(output*255, dtype=np.uint8)\n cv2.imwrite(cur_save_root + '/%05d' % j + '.png', output)\n\n total_runtime += (time.time() - tic) / (cur_seq_range - 1)\n\nprint ('Average per-frame runtime is : {}'.format(total_runtime/seq_numbers))\n","sub_path":"demo_mvos_davis17.py","file_name":"demo_mvos_davis17.py","file_ext":"py","file_size_in_byte":5937,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"587329811","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom itertools import *\nfrom operator import itemgetter\nfrom fnmatch import fnmatch\nimport sys, csv\nimport time\n\ndef group_wc_in_full(d_wc, merge_full):\n reslist = []\n res_no_matches = []\n for wc in d_wc:\n matches = 0\n for full in merge_full:\n if match(full[1], wc[1]):\n matches += 1\n if matches == 1:\n cont = 0\n for full in merge_full:\n if match(full[1], wc[1]):\n new_key = full[0] + \", (\" + wc[0] + \")\"\n merge_full[cont] = (new_key, full[1])\n cont += 1\n if matches > 1:\n cont = 0\n for full in merge_full:\n if match(full[1], wc[1]):\n new_key = full[0] + \", [\" + wc[0] + \"]\"\n merge_full[cont] = (new_key, full[1])\n cont += 1\n if matches == 0:\n res_no_matches.append(wc)\n return merge_full, res_no_matches\n\n\ndef group_no_wc(mergelist):\n reslist = []\n total_keys = []\n for i in range(len(mergelist)):\n isMatch = False\n compared = False\n keys = []\n anyval = None\n keys.append(mergelist[i][0])\n if mergelist[i][0] in total_keys:\n continue\n for j in range(i + 1, len(mergelist)):\n if mergelist[j][0] in total_keys:\n continue\n compared = True\n if mergelist[i][1] == mergelist[j][1]:\n isMatch = True\n keys.append(mergelist[j][0])\n total_keys.append(mergelist[j][0])\n anyval = mergelist[j][1]\n if compared:\n if isMatch:\n str_keys = \", \".join(keys)\n new_item = (str_keys, anyval)\n reslist.append(new_item)\n else:\n reslist.append(mergelist[i])\n return reslist\n\n\ndef match(v1, v2):\n if not \"-\" in v1 and not \"-\" in v2:\n return v1 == v2\n v1_reg = v1.replace(\"-\", \"?\")\n cont = 0\n for s in v2:\n if s == \"-\":\n v1_reg = v1_reg[:cont] + \"?\" + v1_reg[cont + 1:]\n cont += 1\n return fnmatch(v2, v1_reg)\n\n\ndef group_nonmatched(nonmatched):\n pivot = nonmatched[0]\n res = []\n res.append(pivot)\n for i in range(1, len(nonmatched)):\n count = 0\n keys = []\n keys.append(nonmatched[i][0])\n for j in range(len(res)):\n try:\n a = res[j]\n except IndexError:\n continue\n if match(nonmatched[i][1], res[j][1]):\n count += 1\n keys.append(str(res[j][0]))\n del (res[j])\n if count == 0:\n res.append(nonmatched[i])\n else:\n new_tuple = (\", \".join(keys), nonmatched[i][1])\n res.append(new_tuple)\n for i in range(len(res)):\n res[i] = (\"{\" + res[i][0] + \"}\", res[i][1])\n return res\n\n\ndef merger(input_file_name, output_file_name):\n t = time.process_time()\n input_file = open(input_file_name, 'r', encoding='utf8')\n input_file_csv = csv.reader(input_file, delimiter=',')\n input_file_list = list(input_file_csv)\n d = {}\n print('start', len(input_file_list[1:]))\n for line in input_file_list[1:]:\n str_line = \"\".join(line[3:len(line)])\n d[line[0]] = str_line\n input_file.close()\n # d_full = list((key, value) for key, value in d.iteritems())\n d_no_wc = list((key, value) for key, value in d.items() if not \"-\" in value)\n d_wc = list((key, value) for key, value in d.items() if \"-\" in value)\n\n # 1- merge_full <= MERGE d_no_wc\n merge_full = group_no_wc(d_no_wc)\n\n # 2- merge_full_wc <= MERGE d_wc in merge_full\n merge_full_wc, wc_no_match = group_wc_in_full(d_wc, merge_full)\n\n # 3 merge nonmatched\n if len(wc_no_match) > 0:\n merged_wc_no_match = group_nonmatched(wc_no_match)\n else:\n merged_wc_no_match = []\n\n final = merged_wc_no_match + merge_full_wc\n idx = 0\n for row in final:\n positions = {}\n ids = row[0]\n ids = ids.replace(\"(\", \"\").replace(\")\", \"\")\n ids = ids.replace(\"[\", \"\").replace(\"]\", \"\")\n ids = ids.replace(\"{\", \"\").replace(\"}\", \"\")\n ids = ids.replace(\" \", \"\")\n ids = ids.split(',')\n for id in ids:\n id_clean = id.replace(\" \", \"\")\n for line in input_file_list:\n if line[0] == id_clean:\n positions.setdefault(line[1], []).append((id_clean, line[2]))\n # get most common element in list (dict keys)\n # if position:\n if len(positions) > 0:\n chromosome = max(set(positions.keys()), key=list(positions.keys()).count)\n _id, position = min(positions[chromosome], key=lambda x: x[1])\n final[idx] += (chromosome, position, _id,)\n else:\n final[idx] += ('', '', '',)\n idx += 1\n output_file = open(output_file_name, 'w')\n for item in final:\n row = list()\n row.append('\"' + item[0] + '\"') # ids\n row.append(item[2]) # chromosome\n row.append(item[3]) # position\n for c in item[1]:\n row.append(c)\n output_file.write(\",\".join(row) + \"\\n\")\n print('end', len(final))\n elapsed_time = time.process_time() - t\n print('in seconds: ', elapsed_time)\n output_file.close()\n\n\nif __name__ == \"__main__\":\n import argparse\n\n parser = argparse.ArgumentParser() # pylint: disable=invalid-name\n parser.add_argument(\"-i\", \"--input\", help=\"Input file\")\n parser.add_argument(\"-o\", \"--output\", help=\"Output file\")\n args = parser.parse_args() # pylint: disable=invalid-name\n merger(args.input, args.output)\n","sub_path":"merger.py","file_name":"merger.py","file_ext":"py","file_size_in_byte":5743,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"533665542","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# ilp - a tag based file indexer\n#\n# Author: slowpoke \n#\n# This program is Free Software under the non-terms\n# of the Anti-License. Do whatever the fuck you want.\n\nimport os\n\nfrom .utils import funcmap, funcset\n\n\nclass Index:\n\n \"\"\"A collection of files and their associated tags.\n\n This object is immutable and changes no other state. All methods which\n would cause a state change return a new index object instead, with the\n changes applied.\n\n The index contains three major data structures:\n\n - a mapping of files to their content's hashes\n - a mapping of hashes to all filenames they identify\n - a mapping of all tags names to a set of the filenames they tag\n\n \"\"\"\n\n files = property(lambda self: self._files)\n tags = property(lambda self: self._tags)\n hashes = property(lambda self: self._hashes)\n\n def __init__(self, files=None, tags=None, hashes=None):\n self._files = funcmap() if files is None else funcmap(files)\n self._tags = funcmap() if tags is None else funcmap(tags)\n self._hashes = funcmap() if hashes is None else funcmap(hashes)\n\n @classmethod\n def new(cls, files=None, tags=None, hashes=None):\n return cls(files, tags, hashes)\n\n def add_file(self, path, hashstring):\n \"\"\"Add a new file to the index.\"\"\"\n if path in self.files and self.files.get(path) != hashstring:\n raise ValueError(\n \"Path exists in the database, but has a different hash.\")\n new_files = self.files.set(path, hashstring)\n pathlist = self.hashes.get(hashstring, funcset())\n new_hashes = self.hashes.set(\n hashstring,\n pathlist.add(path))\n return self.new(new_files, self.tags, new_hashes)\n\n def remove_file(self, path):\n \"\"\"Remove a path from the index.\n\n NOTE: Even if you remove the last duplicate of a file from the index,\n it will still store the hash, albeit with an empty list of paths\n associated with it.\n\n \"\"\"\n if not path in self.files:\n raise KeyError(\"Path not in the index: {}\".format(path))\n hashstring = self.files.get(path)\n new_files = self.files.remove(path)\n pathlist = self.hashes.get(hashstring, funcset())\n new_hashes = self.hashes.set(\n hashstring,\n pathlist.discard(path))\n return self.new(new_files, self.tags, new_hashes)\n\n def add_tag(self, name):\n \"\"\"Add a new tag to the index.\n\n This method will silently return if the tag already exists.\n\n \"\"\"\n if name in self.tags:\n raise KeyError(\"Tag already in the index: {}\".format(name))\n return self.new(\n self.files,\n self.tags.set(name, funcset()),\n self.hashes)\n\n def remove_tag(self, name):\n \"\"\"Remove a tag from the index.\"\"\"\n if name not in self.tags:\n raise KeyError(\"Tag not in the index: {}\".format(name))\n return self.new(\n self.files,\n self.tags.remove(name),\n self.hashes)\n\n def tag_file(self, path, tag_name):\n \"\"\"Tag a file.\"\"\"\n if tag_name not in self.tags:\n raise KeyError(\"Tag not in the index: {}\".format(tag_name))\n if path not in self.files:\n raise KeyError(\"Path not in the index: {}\".format(path))\n hashstring = self.files.get(path)\n new_tag = self.tags.get(tag_name).add(hashstring)\n return self.new(\n self.files,\n self.tags.set(tag_name, new_tag),\n self.hashes)\n\n def untag_file(self, path, tag_name):\n \"\"\"Untag a file.\"\"\"\n if tag_name not in self.tags:\n raise KeyError(\"Tag not in the index: {}\".format(tag_name))\n if path not in self.files:\n raise KeyError(\"Path not in the index: {}\".format(path))\n hashstring = self.files.get(path)\n new_tag = self.tags.get(tag_name).discard(hashstring)\n return self.new(\n self.files,\n self.tags.set(tag_name, new_tag),\n self.hashes)\n\n def file_has_tag(self, path, tag_name):\n \"\"\"Check if a file is tagged by a tag.\"\"\"\n if tag_name not in self.tags:\n # if the tag doesn't exist, no path can be tagged by it.\n return False\n tag = self.tags.get(tag_name)\n hashstring = self.files.get(path)\n return hashstring in tag\n\n def tags_of_file(self, path):\n \"\"\"Return the names of all tags which are tagging a file.\"\"\"\n hashstring = self.files.get(path)\n return funcset(tag_name\n for tag_name, tag in self.tags.items()\n if hashstring in tag)\n","sub_path":"ilp/index.py","file_name":"index.py","file_ext":"py","file_size_in_byte":4787,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"91336256","text":"import urllib.request\nimport os\n\n# Define a function so you can rip the html from a target domain and print to filepath\n\ndef html_rip(targeturl, filepath):\n with urllib.request.urlopen(targeturl) as response:\n html = str(response.read())\n tempfile = open(filepath, \"w\")\n tempfile.write(html)\n tempfile.close\n\n\n# Specify the target domain\n\nhtml_rip(input(\"Whats the complete url of the target? \"), input(\"What's the absolute path of the filepath you want to create? \"))\n","sub_path":"network-survival-tool-master/ClientBrowser.py","file_name":"ClientBrowser.py","file_ext":"py","file_size_in_byte":492,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"183593803","text":"import os\nimport pyperclip\n\npyperclip.copy('Err')\ntabs = 0\ntabPath = \"\"\ntabsmod = 0\ntabsPlus = 0\nrootDir = \"D:/\"\nsaveFile = r\"C:\\Users\\Sharvin\\Desktop\"\nmasterFilePath = os.path.join(saveFile, '__myList.txt')\nwith open(masterFilePath, 'w', encoding='utf-8') as f:\n for dirpaths, dirnames, filenames in os.walk(rootDir):\n\n _, root = os.path.split(dirpaths)\n\n if root == \"\":\n tabs = 0\n else:\n tabPath = dirpaths.replace('E:/', '')\n tabs = len(tabPath.split('\\\\'))\n tabsmod = tabs - 1\n tabsPlus = tabsmod + 1\n\n if tabsmod == 0:\n f.write('\\n')\n\n f.write('\\t' * tabsmod + root + '\\n')\n\n if tabs == 0 and root == \"\":\n for file in filenames:\n f.write('\\t' * tabs + file + '\\n')\n else:\n for file in filenames:\n f.write('\\t' * tabsPlus + file + '\\n')\n\nwith open(masterFilePath, 'r', encoding='utf-8') as f:\n copiedText = f.read()\n pyperclip.copy(copiedText)\n\nprint('='.center(40, '='))\nprint('COMPLETED'.center(40, '='))\nprint('='.center(40, '='))","sub_path":"ListFolderContent_MODIFIED_append_to_file_in_desktop.py","file_name":"ListFolderContent_MODIFIED_append_to_file_in_desktop.py","file_ext":"py","file_size_in_byte":1113,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"353194300","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n'''\n@File : selenium_test.py\n@Author: Eveline Xue\n@Date : 2019/4/28 13:50\n@Desc : 使用selenium自动化测试工具模拟浏览器点击\n'''\nimport random\nimport time\n\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\n\nfrom referer_proxy import dynamic_load\nfrom utils.useragents import useragent\n\nEXECUTE_PATH = r'E:\\Codes\\xuecode\\handsomeboy\\chromeDriver\\chromedriver74.exe'\n\n\n# 90\ndef search(query_url):\n browser, wait = dynamic_load()\n browser.get(query_url)\n browser.implicitly_wait(10)\n try:\n wait.until(\n EC.presence_of_all_elements_located((By.LINK_TEXT, '普通下载'))\n )\n submit = wait.until(EC.element_to_be_clickable((By.XPATH, '//*[@id=\"go\"]/a[4]/span')))\n except:\n submit = wait.until(EC.element_to_be_clickable((By.XPATH, '//*[@id=\"tt_6\"]/a[1]')))\n submit.click()\n time.sleep(20)\n print('click over')\n # url = 'https://httpbin.org/headers'\n # browser.execute_script('window.location.href = \"{}\";'.format(url))\n # wait.until(lambda driver: driver.current_url == url)\n # # print(browser.page_source)\n browser.quit()\n\n\ndef get_exploer():\n # 加表头\n options = webdriver.ChromeOptions()\n options.add_argument(\n 'user-agent={}'.format(random.choice(useragent)))\n browser = webdriver.Chrome(executable_path=EXECUTE_PATH, options=options)\n wait = WebDriverWait(browser, 10)\n return browser, wait\n\n\ndef main():\n search('http://www.90pan.com/b1173540')\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"code/selenium_test.py","file_name":"selenium_test.py","file_ext":"py","file_size_in_byte":1707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"595153280","text":"import inspect\nfrom flask import abort\n\n\n__all__ = ['validate']\n\n\ndef validate(*__validators__):\n for __validator__ in __validators__:\n if not inspect.isclass(__validator__):\n raise TypeError('The validate decorator only accepts class.')\n method = getattr(__validator__, 'validate', None)\n if method is None or not callable(method):\n raise AttributeError('The validator should contains a callable method \"validate(self)\"')\n\n def decorator(func):\n\n def func_wrapper(*args, **kwargs):\n for __validator__ in __validators__:\n if not __validator__().validate():\n abort(400)\n return func(*args, **kwargs)\n\n return func_wrapper\n\n return decorator\n","sub_path":"flask_validate/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":762,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"621690059","text":"import sys\n\ndef main():\n\tfor line in sys.stdin:\n\t\t\n\t\tinput = [int(num) for num in line.split()]\n\t\t\n\t\ta=input[0]\n\t\tb=input[1]\n\t\tc=input[2]\n\t\t\n\t\tdiscriminant = getDiscriminant(a, b, c)\n\t\tif discriminant < 0:\n\t\t\tprint (\"None\")\n\t\t\tcontinue\n\t\t\n\t\troots = []\n\t\troots.append(((-b)+((discriminant)**0.5))/(2*a))\n\t\troots.append(((-b)-((discriminant)**0.5))/(2*a))\n\t\t\n\t\tprint (\"r1 = {}, r2 = {}\".format(roots[0], roots[1]))\n\t\t\t\ndef getDiscriminant(a, b, c):\n\treturn (b**2) - (4*a*c)\n\t\nif __name__ == \"__main__\":\n\tmain()","sub_path":"62/roots_62.py","file_name":"roots_62.py","file_ext":"py","file_size_in_byte":508,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"548064377","text":"from numpy import *\nfrom time import time\n# 二乗誤差 (a : 真の行列, b : 近似行列)\ndef difcost(a,b):\n\tdif = 0\n\tfor i in range(shape(a)[0]):\n\t\tfor j in range(shape(a)[1]):\n\t\t\tdif += pow(a[i,j]-b[i,j],2)\n\treturn dif\n\n# v : 分解対象の行列 , pc : 特徴の数 , iter : 実行回数\ndef factorize(v,pc=10,t = 20,eps=1e-12):\n\tic = shape(v)[0] # 重み行列の行数(データ項目数)\n\tfc = shape(v)[1] # 特徴行列の列数(特徴項目数)\n\n\t#初期化\n\tw = matrix([[random.random() for i in range(pc)] for j in range(ic)]) \n\th = matrix([[random.random() for i in range(fc)] for j in range(pc)])\n\n\tstart = time()\n\twhile True:\n\t\twh = w*h\n\t\tcost = difcost(v,wh)\n\t\t#print(cost)\n\t\tnow = time()\n\t\tif (now - start) > t or cost < eps:\n\t\t\tbreak\n\t\t# 特徴行列の更新\n\t\thn = (transpose(w)*v)\n\t\thd = (transpose(w)*w*h)\n\t\th = matrix(array(h)*array(hn)/array(hd))\n\n\t\t# 重み行列の更新\n\t\twn = (v*transpose(h))\n\t\twd = (w*h*transpose(h))\n\n\t\tw = matrix(array(w)*array(wn)/array(wd))\n\n\treturn w,h\n\n#テスト実行\n\n\n\n\n\n\n","sub_path":"nmf.py","file_name":"nmf.py","file_ext":"py","file_size_in_byte":1032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"620662464","text":"import sys\nsys.path.append('.')\n\nimport os\nimport argparse\nimport math\nimport numpy as np\nfrom skimage import io, filters, util\nimport json\nfrom tqdm import tqdm\nfrom util import interpolate\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='Applies blur with random sigma to images and saves the amount to pose file.')\n parser.add_argument('path_in', help='Path to NeRF dataset.')\n parser.add_argument('path_out', help='Path to save to.')\n parser.add_argument('--subsets', nargs='+', default=['train'], help='Subsets to process.')\n parser.add_argument('--max_sigma', type=float, default=0, help='Max sigma value.')\n parser.add_argument('--dataset_size_increase', type=int, default=1, help='Integer factor by which to increase the dataset.')\n parser.add_argument('--p', type=float, default=3, help='Exponent stearing the sigma distribution.')\n args = parser.parse_args()\n\n # Create output dir\n os.makedirs(args.path_out)\n\n # Loop over images\n for subset in args.subsets:\n print('Processing {} subset.'.format(subset))\n\n # Get image filenames\n imgs_path = os.path.join(args.path_in, subset)\n img_names = sorted(os.listdir(imgs_path))\n n_imgs = len(img_names)\n n_imgs_out = n_imgs * args.dataset_size_increase\n\n # Get pose file path\n path_pose = os.path.join(args.path_in, 'transforms_' + subset + '.json')\n with open(path_pose) as pose_file:\n pose_dict = json.load(pose_file)\n\n # Get image output file path\n path_out_subset = os.path.join(args.path_out, subset)\n os.makedirs(path_out_subset)\n\n # Inverse cdf of exponential distribution for sampling\n def inv_cdf(x, p):\n if -1e-4 < p and p < 1e-4:\n return x\n else:\n return -np.log(1 - x * (1 - np.exp(-p))) / p\n\n # Sample random sigma values\n np.random.seed(0)\n p = 3\n samples = inv_cdf(np.random.rand(n_imgs_out), p)\n sigma = (samples * args.max_sigma).tolist()\n\n # Get format string for image names\n min_chars = math.ceil(np.log10(n_imgs_out))\n format_str = '{:0' + str(min_chars) + 'd}'\n\n # Define map\n def blur_img(inp):\n idx, img_name, sigma = inp\n format = os.path.splitext(img_name)[-1]\n\n if format == '.png':\n # Load image\n img = io.imread(os.path.join(imgs_path, img_name))\n img = util.img_as_float(img)\n img[:,:,:3] = (img[:,:,:3] ** 2.2 * (img[:,:,3:]))\n img = filters.gaussian(img, sigma=sigma, mode='constant', multichannel=True)\n img[:,:,:3] = (img[:,:,:3] / (img[:,:,3:] + 1e-5)) ** (1 / 2.2)\n\n # Save image\n img_prefix = img_name.split('_')[0]\n img_name_out = img_prefix + '_' + format_str.format(idx) + '.png'\n\n io.imsave(os.path.join(path_out_subset, img_name_out), util.img_as_ubyte(np.clip(img,0,1)), check_contrast=False)\n elif format == '.exr':\n import pyexr as exr\n\n # Load image\n img = exr.read(os.path.join(imgs_path, img_name))\n\n # Apply blur\n img = interpolate.filtered_downsample(img, 1, sigma).numpy()\n\n # Save image\n img_prefix = img_name.split('_')[0]\n img_name_out = img_prefix + '_' + format_str.format(idx) + '.exr'\n\n exr.write(os.path.join(path_out_subset, img_name_out), img)\n else:\n raise ValueError('Unknown filetype.')\n\n with tqdm(total=n_imgs_out) as pbar:\n for inp in zip(range(n_imgs_out), img_names * args.dataset_size_increase, sigma):\n blur_img(inp)\n pbar.update()\n\n # Get output pose dict\n pose_dict_out = {}\n pose_dict_out.update({'camera_angle_x': pose_dict['camera_angle_x'], 'frames': []})\n\n # Add sigma values to pose dict\n for i in range(n_imgs_out):\n pose_dict_out['frames'].append(dict(pose_dict['frames'][i % n_imgs]))\n\n img_path = pose_dict_out['frames'][i]['file_path'].split('_')[0]\n img_path_out = img_path + '_' + format_str.format(i)\n pose_dict_out['frames'][i]['file_path'] = img_path_out\n\n updated_parameters = {'Blur': sigma[i]}\n updated_parameters.update(pose_dict_out['frames'][i]['driver_parameters'])\n pose_dict_out['frames'][i]['driver_parameters'] = updated_parameters\n\n # Save pose dict to new location\n path_out_pose = os.path.join(args.path_out, 'transforms_' + subset + '.json')\n with open(path_out_pose, 'w+') as pose_file:\n json.dump(pose_dict_out, pose_file, sort_keys=False, indent=4)","sub_path":"data/blur.py","file_name":"blur.py","file_ext":"py","file_size_in_byte":4848,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"246801716","text":"#! /usr/bin/python\n# -*- coding:utf-8 -*-\n'''\nPour utiliser le gestionnaire décommentez les lignes de test\nrepéré avec #ligne de test\n'''\n\nimport os\nimport Data, time, random\nfrom time import localtime, strftime\nfrom datetime import datetime\n\nclass Ligne(object):\n '''Objet contenant une ligne qui sera instancier plusieurs fois'''\n def __init__(self, ligne):\n self.num_ligne = ligne\n self.villes_aller = self.init_ville(self.num_ligne, 'aller')\n self.villes_retour = self.init_ville(self.num_ligne, 'retour')\n self.aller = Data.lignes[ligne]['aller']\n self.retour = Data.lignes[ligne]['retour']\n self.ligne = Data.lignes[ligne]\n\n def init_ville(self, num_ligne, Sens):\n '''Initialise la liste des villes'''\n villes = []\n periode = ['scol', 'vac_ete', 'autres_vac']\n\n for sens in Data.lignes[num_ligne]:\n if sens == Sens:\n for row in Data.lignes[num_ligne][sens]:\n for i in range(len(row)):\n if row[i][0] not in villes and row[i][0] not in periode:\n villes.append(row[i][0])\n\n return villes\n\nclass Insert(object):\n '''Objet d'insertion dans la BD'''\n def __init__(self):\n self.dossier = os.listdir('sources')\n\n def lecture_fichier(self, path):\n '''Lit un csv ou un txt et le renvoi sous forme de tableau.'''\n \"\"\"Ouvre et lit un fichier\"\"\"\n with open(path, 'r') as fichier:\n read_fichier = fichier.read()\n\n tab_fichier = []\n saut_de_ligne = True\n\n \"\"\"Insert le contenue du fichier dans tableau Séparateur: saut de ligne\"\"\"\n while saut_de_ligne:\n if '\\n' in read_fichier:\n index = read_fichier.index('\\n')\n tab_fichier.append([read_fichier[:index]])\n read_fichier = read_fichier[index+1:]\n else:\n saut_de_ligne = False\n\n \"\"\"Découpe chaque ligne du tableau. Séparateur: ; \"\"\"\n for j in range(len(tab_fichier)):\n ligne_tab = ';'.join(tab_fichier[j])\n ligne_tab = ligne_tab.split(';')\n tab_fichier[j] = ligne_tab\n\n return tab_fichier\n\n def suppr_inutile(self, tableau):\n '''Supprime les colonnes inutile (contenant ND)'''\n index = None\n IsND = True\n while IsND:\n if 'ND' not in tableau[0]:\n IsND = False\n break\n for row in tableau:\n if len(row) == len(tableau[0]) and len(row) <= 8:\n break\n if row == tableau[0] and 'ND' in row:\n index = tableau[0].index('ND')\n tableau[0].pop(index)\n else:\n row.pop(index)\n return tableau\n\n def inscription(self, lignes, dico_lignes):\n '''Inscrit les données dans la BD'''\n nonInscrit = True\n while nonInscrit:\n try:\n with open('Data.py', 'a') as bd:\n bd.write(lignes)\n bd.write(dico_lignes)\n nonInscrit = False\n except NameError:\n print('Echec de lecture du fichier : '+NameError)\n continue\n\n def insert(self):\n '''Insert les données du dossier \"sources\" dans la BD'''\n lignes = []\n dico_lignes = [\"lignes = {\\n\"]\n\n \"\"\"Liste les sous dossier de /sources\"\"\"\n for i in range(len(self.dossier)):\n paths = []\n path_dossier = 'sources/' + self.dossier[i]\n fichier = os.listdir(path_dossier)\n\n \"\"\"Liste les fichiers contenus dans chaque sous dossier\"\"\"\n for j in range(len(fichier)):\n path_fichier = path_dossier + '/' + fichier[j]\n paths.append(path_fichier)\n\n dico_bd = []\n for path in paths:\n '''\n Créer de trois tableaux pour chaque \"catégories\"\n scol = periode scolaire...\n '''\n fichier = self.lecture_fichier(path)\n\n scol = [fichier[1]] + fichier[4:]\n scol = self.suppr_inutile(scol)\n\n fichier = self.lecture_fichier(path)\n\n vac_ete = [fichier[2]] + fichier[4:]\n vac_ete = self.suppr_inutile(vac_ete)\n\n fichier = self.lecture_fichier(path)\n\n autre_vac = [fichier[3]] + fichier[4:]\n autre_vac = self.suppr_inutile(autre_vac)\n\n if dico_bd == []:\n dico_bd.append(fichier[0][0])\n dico_bd.append(fichier[0][1])\n dico_bd.append(fichier[0][2])\n\n else:\n dico_bd.append(fichier[0][2])\n\n '''inscrit les tableaux dans une var'''\n lignes.append(fichier[0][2]+' = [\\n'+str(scol)+',\\n'+str(autre_vac)+',\\n'+str(vac_ete)+']\\n')\n\n if i == len(self.dossier)-1:\n dico_lignes.append(\" '\"+dico_bd[1]+\"': {'aller': \"+dico_bd[2]+\", 'retour': \"+dico_bd[3]+\"}\\n}\")\n else:\n dico_lignes.append(\" '\"+dico_bd[1]+\"': {'aller': \"+dico_bd[2]+\", 'retour': \"+dico_bd[3]+\"},\\n\")\n lignes = ''.join(lignes)\n dico_lignes = ''.join(dico_lignes)\n self.inscription(lignes, dico_lignes)\n\nclass Temps(object):\n '''Objet de gestion de la date et de l'heure '''\n def __init__(self):\n self.date = self.init_date()\n self.heure = self.init_heure()\n self.numJourAn = self.init_numjouran()\n self.numJourSem = self.init_numjoursem()\n self.jour_feries = ['15/8/2015','1/11/2015','11/11/2015','25/12/2015','1/1/2016','28/3/2016','1/5/2016','5/5/2016','16/5/2016','14/7/2016']\n self.jour = self.init_jour()\n self.vacances = {'vac_ete': ['5/07/2015', '1/09/2015'], 'toussaint': ['17/10/2015', '2/11/2015'], 'noel': ['19/12/2015', '4/01/2016'], 'hiver': ['13/02/2016', '29/02/2016'], 'printemps': ['9/04/2016', '25/04/2016']}\n self.periode = self.init_periode()\n\n def init_date(self):\n '''Mise en forme de la date'''\n objet_date = datetime.now()\n if len(str(objet_date.month)) == 1:\n mois = '0' + str(objet_date.month)\n else:\n mois = str(objet_date.month)\n date = str(objet_date.day)+'/'+mois+'/'+str(objet_date.year)\n return date\n\n def init_heure(self):\n '''Initialisation de l'heure'''\n heure = strftime(\"%H:%M\", localtime())\n heure = heure.split(':')\n heure = int(str(heure[0])+str(heure[1]))\n return heure\n\n def init_numjouran(self):\n '''Donne le numéro du jour dans l'année'''\n date = self.date.split('/')\n jour, mois, annee = date\n jour, mois, annee = int(jour), int(mois), int(annee)\n if ((annee % 4 == 0 and annee % 100 != 0) or annee % 400 == 0): # bissextile?\n return (0,31,60,91,121,152,182,213,244,274,305,335,366)[mois-1] + jour\n else:\n return (0,31,59,90,120,151,181,212,243,273,304,334,365)[mois-1] + jour\n\n def init_numjoursem(self):\n '''Donne le numéro du jour de la semaine'''\n date = self.date.split('/')\n annee = int(date[2])-1\n jour = (annee+(annee//4)-(annee//100)+(annee//400)+self.numJourAn) % 7\n if jour == 0: jour = 7\n return jour\n\n def init_jour(self):\n if self.date in self.jour_feries:\n return 'Ferie'\n return ['Lundi', 'Mardi', 'mercredi', 'Jeudi', 'Vendredi', 'Samedi', 'Dimanche'][self.numJourSem-1]\n\n def date_anterieur(self, premiere_date, seconde_date):\n '''Dit si une premiere date est antèrieure ou égale à une seconde date'''\n premiere_date = premiere_date.split('/')\n seconde_date = seconde_date.split('/')\n if premiere_date == seconde_date:\n return 'identique'\n else:\n if int(seconde_date[2]) > int(premiere_date[2]):\n return True\n elif int(seconde_date[2]) == int(premiere_date[2]):\n if int(seconde_date[1]) > int(premiere_date[1]):\n return True\n elif int(seconde_date[1]) == int(premiere_date[1]):\n if int(seconde_date[0]) >= int(premiere_date[0]):\n return True\n return False\n\n def date_posterieur(self, premiere_date, seconde_date):\n '''Dit si une premiere date est postérieure ou égale à une seconde date'''\n premiere_date = premiere_date.split('/')\n seconde_date = seconde_date.split('/')\n if premiere_date == seconde_date:\n return 'identique'\n else:\n if int(seconde_date[2]) < int(premiere_date[2]):\n return True\n elif int(seconde_date[2]) == int(premiere_date[2]):\n if int(seconde_date[1]) < int(premiere_date[1]):\n return True\n elif int(seconde_date[1]) == int(premiere_date[1]):\n if int(seconde_date[0]) <= int(premiere_date[0]):\n return True\n return False\n\n def init_periode(self):\n '''Initialise la période de l'année'''\n for vacance in self.vacances:\n posterieur = self.date_posterieur(self.date, self.vacances[vacance][0])\n anterieur = self.date_anterieur(self.date, self.vacances[vacance][1])\n\n if posterieur == True and anterieur == True:\n if vacance == 'vac_ete':\n retour = 2\n break\n else:\n retour = 1\n break\n elif posterieur == False or anterieur == False:\n retour = 0\n return retour\n\nclass Select(object):\n ''''Objet selection des horaires'''\n def __init__(self, ville_depart, arret_depart, ville_arriver, arret_arriver):\n self.lignes = self.init_lignes()\n self.ville_depart = ville_depart\n self.arret_depart = arret_depart\n self.ville_arriver = ville_arriver\n self.arret_arriver = arret_arriver\n self.date = Temps()\n self.num_ligne = self.init_num_ligne()\n self.sens = self.init_sens()\n self.fiche = self.init_fiche()\n self.index_jour =self.init_index_jour()\n self.index_heure = None\n self.depart, self.arriver = self.init_depart_arriver()\n self.depart = self.select_arret(self.depart)\n self.arriver = self.select_arret(self.arriver)\n self.depart = self.select_horaire(self.depart)\n self.arriver = self.select_horaire(self.arriver)\n self.retour = self.mise_en_forme(self.depart, self.arriver)\n\n def init_lignes(self):\n '''Initialise le dictionnaire contenant les objets ligne'''\n lignes = {}\n for num_ligne in Data.lignes:\n lignes[str(num_ligne)] = Ligne(num_ligne)\n return lignes\n\n def init_num_ligne(self):\n '''Détermine le numéro de la ligne'''\n for ligne in self.lignes:\n if self.ville_depart in self.lignes[ligne].villes_aller and self.ville_arriver in self.lignes[ligne].villes_aller:\n return ligne\n\n def init_sens(self):\n '''Détermine le sens grace à la position des villes dans la liste'''\n if self.lignes[self.num_ligne].villes_aller.index(self.ville_depart) < self.lignes[self.num_ligne].villes_aller.index(self.ville_arriver):\n return 'aller'\n else:\n return 'retour'\n\n def init_fiche(self):\n '''Renvoi la fiche voule grace au sens'''\n if self.sens == 'aller': return self.lignes[self.num_ligne].aller[self.date.periode]\n elif self.sens == 'retour': return self.lignes[self.num_ligne].retour[self.date.periode]\n\n def init_index_jour(self):\n '''Détermine l'index des colonnes voulue (jour) pour selectionner les bons horaires'''\n index_jour = []\n aujourdhui = self.date.jour\n aujourdhui = aujourdhui[0]\n for jour in self.fiche[0]:\n if jour != self.fiche[0][0] and jour != self.fiche[0][1]:\n if aujourdhui in jour:\n index_jour.append(self.fiche[0].index(jour))\n index_jour.reverse()\n index_jour.append(1)\n index_jour.append(0)\n index_jour.reverse()\n return index_jour\n\n def init_depart_arriver(self):\n '''Sélectionne les lignes contenant les villes de depart et d'arriver'''\n depart = []\n arriver = []\n for horaire in self.fiche:\n buffer_depart = []\n buffer_arriver = []\n if horaire[0] == self.ville_depart:\n if horaire[:3] == ['LE CHEYLARD', 'Collège St Louis', '0']:continue\n for index in self.index_jour:\n buffer_depart.append(horaire[index])\n elif horaire[0] == self.ville_arriver:\n for index in self.index_jour:\n buffer_arriver.append(horaire[index])\n if buffer_arriver != []:\n arriver.append(buffer_arriver)\n if buffer_depart != []:\n depart.append(buffer_depart)\n self.index_heure = self.init_index_heure(depart)\n return depart, arriver\n\n def init_index_heure(self, depart):\n '''Détermine l'index des colonnes voulue (heure) pour selectionner les bons horaires'''\n index_heure = []\n for DepArr in depart:\n if depart.index(DepArr) == 1: break\n for heure in DepArr:\n print(heure)\n try:\n decoupe = heure.split(':')\n decoupe = int(str(decoupe[0])+str(decoupe[1]))\n if decoupe > self.date.heure:\n if str(DepArr.index(heure)) not in index_heure:\n index_heure.append(DepArr.index(heure))\n except:\n index_heure.append(DepArr.index(heure))\n return index_heure\n\n def select_arret(self, ligne):\n if ligne[0][0] == self.ville_depart:\n tab_retour = []\n for arret in ligne:\n if self.arret_depart == arret[1]:\n tab_retour = arret\n break\n else:\n continue\n return tab_retour\n else:\n tab_retour = []\n for arret in ligne:\n if self.arret_arriver == arret[1]:\n tab_retour = arret\n else:\n continue\n return tab_retour\n\n def select_horaire(self, arret):\n '''Renvoi la ville, l'arret et les horaire en fonction des index des heures voulue'''\n tab_retour = []\n for heure in arret:\n if arret.index(heure) in self.index_heure:\n tab_retour.append(heure)\n else:\n continue\n return tab_retour\n\n def mise_en_forme(self, depart, arriver):\n '''Met en forme le dictionnaire de retour'''\n dico_retour = {}\n '''Met en forme le départ'''\n dico_retour[self.ville_depart] = {}\n depart = depart[1:]\n for i in range(len(depart)):\n if i == 0:continue\n dico_retour[self.ville_depart]['bus'+str(i)] = [depart[0], depart[i]]\n\n '''Met en forme l'arriver'''\n dico_retour[self.ville_arriver] = {}\n arriver = arriver[1:]\n for i in range(len(arriver)):\n if i == 0:continue\n dico_retour[self.ville_arriver]['bus'+str(i)] = [arriver[0], arriver[i]]\n\n return dico_retour\n\n################################################\n'''Section fonctions app'''\n\ndef select_villes(lettre_choisi):\n '''Recupère la liste des villes et la renvoi sous forme de tableau'''\n villes_aller = []\n periode = ['scol', 'vac_ete', 'autres_vac']\n for ligne in Data.lignes:\n for sens in Data.lignes[ligne]:\n if sens == 'aller':\n for row in Data.lignes[ligne][sens]:\n for i in range(len(row)):\n if row[i][0] not in villes_aller and row[i][0] not in periode:\n if row[i][0][0] == lettre_choisi:\n villes_aller.append(row[i][0])\n\n villes_aller.sort()\n return villes_aller\n\ndef select_arrets(ville_selectionne):\n periode = ['scol', 'vac_ete', 'autres_vac']\n arrets = []\n for ligne in Data.lignes:\n for sens in Data.lignes[ligne]:\n if sens == 'aller':\n for row in Data.lignes[ligne][sens]:\n for i in range(len(row)):\n if row[i][0] not in arrets and row[i][0] not in periode:\n if row[i][0] == ville_selectionne and row[i][1] not in arrets:\n arrets.append(row[i][1])\n arrets_retour = arrets\n arrets_retour.sort()\n return arrets_retour\n\ndef select_seconde_villes(ville_selectionne, lettre_choisi):\n '''Recupère la liste des villes et la renvoi sous forme de tableau'''\n periode = ['scol', 'vac_ete', 'autres_vac']\n\n villes = []\n for ligne in Data.lignes:\n for sens in Data.lignes[ligne]:\n if sens == 'aller':\n for row in Data.lignes[ligne][sens]:\n for i in range(len(row)):\n if row[i][0] not in villes and row[i][0] not in periode:\n if row[i][0][0] == lettre_choisi and row[i][0] != ville_selectionne:\n villes.append(row[i][0])\n villes_retour = villes\n villes_retour.sort()\n return villes_retour\n\ndef select_horaire(depart, arret_depart, arriver, arret_arriver):\n '''\n Le select général\n Les valeurs demandées sont sous cette forme:\n depart = 'LE CHEYLARD'\n arriver = 'CHARMES'\n '''\n if depart != None or arriver != None:\n selection = Select(depart, arret_depart, arriver, arret_arriver)\n return selection.retour\n else:\n return 'ERREUR'\n\n################################################\n\nif __name__ == '__main__':#lignes de test\n #print('1-select 2-insert')\n #choix = input()\n choix = '1'\n if choix == '1':\n print(select_horaire('LE CHEYLARD', 'Gendarmerie', 'CHARMES', 'Centre'))\n #test_select = Select('LE CHEYLARD', 'CHARMES')\n #for row in test_select.retour: print(row, test_select.retour[row])\n #for row in test_select.depart: print(row)\n #for row in test_select.arriver: print(row)\n\n #main()\n elif choix == '2':\n insertion = Insert()\n insertion.insert()","sub_path":".buildozer/android/app/gestion_bd.py","file_name":"gestion_bd.py","file_ext":"py","file_size_in_byte":18693,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"342787510","text":"class Solution:\n def numDistinct_simple(self, s, t):\n \"\"\"\n :type s: str\n :type t: str\n :rtype: int\n \"\"\"\n dp = [1] + [0] * (len(t))\n for c in s:\n for i in range(len(t) - 1, -1, -1):\n if c == t[i]:\n dp[i + 1] += dp[i]\n print(dp)\n return dp[-1]\n\n def numDistinct(self, s, t):\n \"\"\"\n :type s: str\n :type t: str\n :rtype: int\n \"\"\"\n n, m = len(s), len(t)\n dp = [[0] * (n+1) for _ in range(m+1)]\n dp[0][0] = 1\n # All i can create the empty target sequence, just skip everything\n for i in range(n+1):\n dp[0][i] = 1\n\n for i in range(1, m+1):\n for j in range(1, n+1):\n if s[j-1] != t[i-1]:\n dp[i][j] = dp[i][j-1]\n else:\n dp[i][j] = dp[i][j-1] + dp[i-1][j-1]\n print(dp[i])\n return dp[m][n]\n\nsol = Solution()\nprint(sol.numDistinct(\"rabbbit\", \"rabbit\"))\n# print(sol.numDistinct(\"rabbb\", \"rabb\"))\n","sub_path":"distinct_strs.py","file_name":"distinct_strs.py","file_ext":"py","file_size_in_byte":1084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"383906471","text":"import com.ihsan.foundation.pobjecthelper as phelper\nimport sys\n\ndef TampilNama(config, parameter, returnpacket):\n status = returnpacket.CreateValues(\n ['IsErr',0],\n ['ErrMessage',''],\n ['NamaBeasiswa',''],\n ['PhoneNumber',''],\n ['alamat',''],\n ['kota',''],\n ['kodepos',''],\n )\n try:\n param = parameter.FirstRecord\n helper = phelper.PObjectHelper(config)\n\n oCust = helper.GetObject('Mustahiq', param.idmustahiq)\n oBeasiswa = helper.GetObject('Customer', oCust.CustomerId)\n\n status.NamaBeasiswa = oBeasiswa.CustomerName\n status.PhoneNumber = oBeasiswa.PhoneNumber\n status.alamat = oBeasiswa.AddressStreet\n status.kota = oBeasiswa.AddressCity\n status.kodepos = oBeasiswa.AddressPostalCode\n\n except:\n status.IsErr = 1\n status.ErrMessage = str(sys.exc_info()[1])\n raise\n\ndef Simpan(config, parameter, returns):\n status = returns.CreateValues(\n ['Is_Err',0],['Err_Message',''],['Pesan','']\n )\n recParam = parameter.FirstRecord\n helper = phelper.PObjectHelper(config)\n ProductId = recParam.ProductId\n config.BeginTransaction()\n param = {'tipe':'Beasiswa'}\n try:\n\n assProd = helper.CreatePObject('MustahiqProduct', param)\n assProd.ProductId = ProductId\n assProd.MustahiqId = recParam.GetFieldByName('LMustahiqProduct.MustahiqId')\n\n UpdateProduk = helper.GetObject('Product',ProductId )\n SisaQuota = UpdateProduk.CurrentQuotaPosition + 1\n UpdateProduk.CurrentQuotaPosition = SisaQuota\n if SisaQuota >= UpdateProduk.MustahiqQuota :\n UpdateProduk.IsQuotaFull = 'T'\n\n config.Commit()\n status.Pesan='Data Beasiswa Berhasil diregistrasi..'\n except:\n config.Rollback()\n status.Is_Err = 1\n status.Err_Message = str(sys.exc_info()[1])\n raise\n\n","sub_path":"dialogs/Beasiswa/fAddDataBeasiswaNasabah_data.py","file_name":"fAddDataBeasiswaNasabah_data.py","file_ext":"py","file_size_in_byte":1820,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"102967698","text":"\nfrom tkinter import *\n\n\nclass Application(Frame):\n \"\"\"GUI app with 3 buttons\"\"\"\n\n def __init__(self, master) :\n \"\"\"\"init the frame\"\"\"\n Frame.__init__(self,master)\n self.grid()\n self.create_widgets()\n\n def create_widgets (self):\n \"\"\"create 3 buttons \"\"\"\n #create the game title\n self.title = Label(self, text = \"Majority Rules\\nThe game of guessing what the world thinks\\n\")\n self.title.grid(sticky=N)\n\n self.instruction = Label(self, text = \"Enter a search term:\")\n self.instruction.grid(sticky=N)\n\n self.searchterm = Entry(self)\n self.searchterm.grid()\n \n #googleit button\n self.button1 = Button(self, text = \"Google it.\")\n self.button1.grid()\n\n #randomsearchterm button\n self.button2 = Button(self)\n self.button2.grid()\n self.button2.configure(text = \"I'm feeling lucky!\")\n\n self.QUIT = Button(self, text=\"Quit\", command=root.destroy)\n self.QUIT.grid()\n\n \nroot = Tk()\nroot.title(\"Majority Rules\")\nroot.geometry()\n\napp = Application(root)\n\nroot.mainloop()\n","sub_path":"tkinter_enter_searchterm.py","file_name":"tkinter_enter_searchterm.py","file_ext":"py","file_size_in_byte":1132,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"137583251","text":"\r\n\r\nimport pandas as pd\r\ndiabetes_data = pd.read_csv(\"E:/360digiTMG assignment/Data Science/Ensemble technique/Datasets_ET/Diabeted_Ensemble.csv\")\r\n\r\n# Input and Output Split\r\npredictors = diabetes_data.loc[:, diabetes_data.columns!=\" Class variable\"]\r\ntype(predictors)\r\n\r\ntarget = diabetes_data[\" Class variable\"]\r\ntype(target)\r\n\r\n# Train Test partition of the data\r\nfrom sklearn.model_selection import train_test_split\r\nx_train, x_test, y_train, y_test = train_test_split(predictors, target, test_size = 0.2, random_state=0)\r\n\r\n# GridSearchCV\r\n\r\nfrom sklearn.model_selection import GridSearchCV\r\nfrom sklearn.ensemble import RandomForestClassifier\r\n\r\nrf_clf_grid = RandomForestClassifier(n_estimators=500, n_jobs=1, random_state=42)\r\n\r\nparam_grid = {\"max_features\": [4, 5, 6, 7, 8, 9, 10], \"min_samples_split\": [2, 3, 10]}\r\n\r\ngrid_search = GridSearchCV(rf_clf_grid, param_grid, n_jobs = -1, cv = 5, scoring = 'accuracy')\r\n\r\ngrid_search.fit(x_train, y_train)\r\n\r\ngrid_search.best_params_\r\n\r\ncv_rf_clf_grid = grid_search.best_estimator_\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\nconfusion_matrix(y_test, cv_rf_clf_grid.predict(x_test))\r\naccuracy_score(y_test, cv_rf_clf_grid.predict(x_test))\r\n\r\n#Bagging technique\r\nfrom sklearn import tree\r\nclftree = tree.DecisionTreeClassifier()\r\nfrom sklearn.ensemble import BaggingClassifier\r\n\r\nbag_clf = BaggingClassifier(base_estimator = clftree, n_estimators = 500,\r\n bootstrap = True, n_jobs = 1, random_state = 42)\r\n\r\nbag_clf.fit(x_train, y_train)\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\n# Evaluation on Testing Data\r\nconfusion_matrix(y_test, bag_clf.predict(x_test))\r\naccuracy_score(y_test, bag_clf.predict(x_test))\r\n\r\n# Evaluation on Training Data\r\nconfusion_matrix(y_train, bag_clf.predict(x_train))\r\naccuracy_score(y_train, bag_clf.predict(x_train))\r\n\r\n#AdaBoosting technique\r\nfrom sklearn.ensemble import AdaBoostClassifier\r\n\r\nada_clf = AdaBoostClassifier(learning_rate = 0.02, n_estimators = 5000)\r\n\r\nada_clf.fit(x_train, y_train)\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\n# Evaluation on Testing Data\r\nconfusion_matrix(y_test, ada_clf.predict(x_test))\r\naccuracy_score(y_test, ada_clf.predict(x_test))\r\n\r\n# Evaluation on Training Data\r\nconfusion_matrix(y_train, ada_clf.predict(x_train))\r\naccuracy_score(y_train, ada_clf.predict(x_train))\r\n\r\n#Grddient boosting\r\nfrom sklearn.ensemble import GradientBoostingClassifier\r\n\r\nboost_clf = GradientBoostingClassifier()\r\n\r\nboost_clf.fit(x_train, y_train)\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\n# Evaluation on Testing Data\r\nconfusion_matrix(y_test, boost_clf.predict(x_test))\r\naccuracy_score(y_test, boost_clf.predict(x_test))\r\n\r\n# Evaluation on Training Data\r\nconfusion_matrix(y_train, boost_clf.predict(x_train))\r\naccuracy_score(y_train, boost_clf.predict(x_train))\r\n\r\n# Hyperparameters\r\nboost_clf2 = GradientBoostingClassifier(learning_rate = 0.02, n_estimators = 1000, max_depth = 1)\r\nboost_clf2.fit(x_train, y_train)\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\n# Evaluation on Testing Data\r\nconfusion_matrix(y_test, boost_clf2.predict(x_test))\r\naccuracy_score(y_test, boost_clf2.predict(x_test))\r\n\r\n# Evaluation on Training Data\r\nconfusion_matrix(y_train, boost_clf2.predict(x_train))\r\naccuracy_score(y_train, boost_clf2.predict(x_train))\r\n\r\n#XGBoosting\r\nimport xgboost as xgb\r\n\r\nxgb_clf = xgb.XGBClassifier(max_depths = 5, n_estimators = 10000, learning_rate = 0.3, n_jobs = -1)\r\n\r\nxgb_clf.fit(x_train, y_train)\r\n\r\nfrom sklearn.metrics import accuracy_score, confusion_matrix\r\n\r\n# Evaluation on Testing Data\r\nconfusion_matrix(y_test, xgb_clf.predict(x_test))\r\naccuracy_score(y_test, xgb_clf.predict(x_test))\r\n\r\n# Evaluation on Training Data\r\nconfusion_matrix(y_train, xgb_clf.predict(x_train))\r\naccuracy_score(y_train, xgb_clf.predict(x_train))\r\n\r\n#Voting technique\r\n# Import the required libraries\r\nfrom sklearn import linear_model, svm, neighbors, naive_bayes\r\nfrom sklearn.ensemble import VotingClassifier\r\nfrom sklearn.metrics import accuracy_score\r\n\r\n# Instantiate the learners (classifiers)\r\nlearner_1 = neighbors.KNeighborsClassifier(n_neighbors=5)\r\nlearner_2 = linear_model.Perceptron(tol=1e-2, random_state=0)\r\nlearner_3 = svm.SVC(gamma=0.001)\r\n\r\n# Instantiate the voting classifier\r\nvoting = VotingClassifier([('KNN', learner_1),\r\n ('Prc', learner_2),\r\n ('SVM', learner_3)])\r\n\r\n# Fit classifier with the training data\r\nvoting.fit(x_train, y_train)\r\n\r\n# Predict the most voted class\r\nhard_predictions = voting.predict(x_test)\r\n\r\n# Accuracy of hard voting\r\nprint('Hard Voting:', accuracy_score(y_test, hard_predictions))\r\n\r\n# Soft Voting\r\n# Instantiate the learners (classifiers)\r\nlearner_4 = neighbors.KNeighborsClassifier(n_neighbors = 5)\r\nlearner_5 = naive_bayes.GaussianNB()\r\nlearner_6 = svm.SVC(gamma = 0.001, probability = True)\r\n\r\n# Instantiate the voting classifier\r\nvoting = VotingClassifier([('KNN', learner_4),\r\n ('NB', learner_5),\r\n ('SVM', learner_6)],\r\n voting = 'soft')\r\n\r\n# Fit classifier with the training data\r\nvoting.fit(x_train, y_train)\r\nlearner_4.fit(x_train, y_train)\r\nlearner_5.fit(x_train, y_train)\r\nlearner_6.fit(x_train, y_train)\r\n\r\n# Predict the most probable class\r\nsoft_predictions = voting.predict(x_test)\r\n\r\n# Get the base learner predictions\r\npredictions_4 = learner_4.predict(x_test)\r\npredictions_5 = learner_5.predict(x_test)\r\npredictions_6 = learner_6.predict(x_test)\r\n\r\n# Accuracies of base learners\r\nprint('L4:', accuracy_score(y_test, predictions_4))\r\nprint('L5:', accuracy_score(y_test, predictions_5))\r\nprint('L6:', accuracy_score(y_test, predictions_6))\r\n\r\n# Accuracy of Soft voting\r\nprint('Soft Voting:', accuracy_score(y_test, soft_predictions))\r\n","sub_path":"diabetes.py","file_name":"diabetes.py","file_ext":"py","file_size_in_byte":5880,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"291588959","text":"# ispred lika crta more baklji po tlu\n\nimport time\nfrom mc import * # import api-ja\nfrom crtanje import * # tu je funkcija koju zovem\n\nmc = Minecraft() # inicijalizacija sustava za rad sa Minecraftom\nimport time\n\nzaObradu = [GRASS.id, SANDSTONE.id, SAND.id, STONE.id, DIRT.id, GRAVEL.id, COBBLESTONE.id, CLAY.id, GOLD_ORE.id,\n IRON_ORE.id, COAL_ORE.id, DIAMOND_ORE.id, OBSIDIAN.id, REDSTONE_ORE.id, LAPIS_LAZULI_ORE.id,\n 129] # 129 emerald\n\npopis = dict()\n\n\ndef stoneCutter(orMj, orSm, dimenzije=5, visina=5):\n a = 1\n for dY in range(visina, -1, -1):\n mc.postToChat(\"Level: %s \" % dY)\n for dZ in range(-3 - dY, 3 + dY):\n for dX in range(1, dimenzije + 1 + dY):\n a += 1\n\n gdje = rel2abs((int(orMj[0]), int(orMj[1]), int(orMj[2])), (dX, dZ, dY), orSm)\n # id = mc.spawnEntity('Minecart', int (gdje [0]) ,int (gdje [1]) ,int (gdje [2]) , \"{Type:0}\" )\n # block = getBlockWithData( int (gdje [0]) ,int (gdje [1]) ,int (gdje [2]))\n myBlock = mc.getBlockWithData(int(gdje[0]), int(gdje[1]), int(gdje[2]))\n myBlock = mc.getBlockWithData(int(gdje[0]), int(gdje[1]), int(gdje[2]))\n\n if myBlock.id in (10, 11): # makni lavu\n mc.setBlock(int(gdje[0]), int(gdje[1]), int(gdje[2]), AIR.id, 0)\n if myBlock.id in zaObradu:\n a = a + 1\n # time.sleep ( 0.5 )\n mc.setBlock(int(gdje[0]), int(gdje[1]), int(gdje[2]), AIR.id, 0)\n\n if popis.has_key((myBlock.id, myBlock.data)):\n popis[(myBlock.id, myBlock.data)] += 1\n else:\n popis[(myBlock.id, myBlock.data)] = 1\n\n\n\n\n\n # time.sleep ( 0.5 )\n # myId = mc.spawnEntity('Item', int (gdje [0]) ,int (gdje [1]) ,int (gdje [2] ) , sto )\n\n for bla in popis.keys():\n blok = bla[0]\n modifikacija = bla[1]\n # prijevodi:\n # diamond\n # 56 : 264,\n if bla[0] == 56:\n blok = 264\n # redstone\n # 73 : 331 ,\n if bla[0] == 73:\n blok = 331\n # lapis\n # 21 : 351 , 4\n if bla[0] == 21:\n blok = 351\n modifikacija = 4\n # emerald\n # 129 : 388\n if bla[0] == 129:\n blok = 388\n # coal COAL_ORE.id 263\n if bla[0] == COAL_ORE.id:\n blok = 263\n modifikacija = 0\n\n mc.postToChat(\"Key: %s %s \" % (bla[0], bla[1]))\n mc.postToChat(\"Value: %s \" % popis[bla])\n while popis[bla] > 0:\n if popis[bla] > 64:\n sto = ('{Item:{id:%s,Count:%s,Damage:%s}}' % (blok, 64, modifikacija))\n else:\n sto = ('{Item:{id:%s,Count:%s,Damage:%s}}' % (blok, popis[bla], modifikacija))\n\n mc.postToChat(\"XXX: %s \" % (sto))\n gdje = rel2abs(orMj, (0, 0, 3), orSm)\n time.sleep(6)\n myId = mc.spawnEntity('Item', int(gdje[0]), int(gdje[1]), int(gdje[2]), sto)\n popis[bla] -= 64\n\n mc.postToChat(\"Kraj : XXXXXXXXXXXX\")\n return 1\n\n\nif __name__ == \"__main__\": # direktan poziv\n orMj = gdjeSam()\n orSm = gdjeGledam()\n stoneCutter(orMj, orSm, dimenzije=35, visina=5)\n # bakljada (dimenzije = 200 , visina = 80)\n","sub_path":"dugiStoneCutter.py","file_name":"dugiStoneCutter.py","file_ext":"py","file_size_in_byte":3403,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"202494647","text":"from office365.runtime.client_value import ClientValue\n\n\nclass Folder(ClientValue):\n\n def __init__(self, childCount=None, view=None):\n \"\"\"\n\n :param int childCount:\n :param office365.onedrive.folderView.FolderView view:\n \"\"\"\n super(Folder, self).__init__()\n self.childCount = childCount\n self.view = view\n","sub_path":"office365/onedrive/folder.py","file_name":"folder.py","file_ext":"py","file_size_in_byte":356,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"480263024","text":"# coding=utf8\n\nimport argparse\nimport numpy as np\nfrom tensorflow.python import pywrap_tensorflow\n\n\ndef fetch(checkpoint_path, tensor_name, save_path=None):\n reader = pywrap_tensorflow.NewCheckpointReader(checkpoint_path)\n var_to_shape_map = reader.get_variable_to_shape_map()\n target_tensor = None\n for key in var_to_shape_map:\n if tensor_name == key:\n target_tensor = reader.get_tensor(key)\n break\n print(target_tensor)\n if save_path and isinstance(target_tensor, np.ndarray):\n np.save(save_path, target_tensor)\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n\n parser.add_argument(\"--checkpoint\", help=\"Checkpoint path\", required=True)\n parser.add_argument(\"--tensor\", help=\"Tensor name\", required=True)\n parser.add_argument(\"--save\", help=\"Save path\", required=False)\n args = parser.parse_args()\n\n fetch(args.checkpoint, args.tensor, args.save)\n","sub_path":"baseline_with_case/tools/fetch_embedding_from_checkpoint.py","file_name":"fetch_embedding_from_checkpoint.py","file_ext":"py","file_size_in_byte":938,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"363733558","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 12 12:02:57 2017\n\n@author: fmonnery\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D\n\ndatas = pd.read_csv('D:\\WorkSpace\\_Machine_Learning\\house_data.csv')\n\ndata = datas.dropna()\nZ = data.isnull().sum().sum() \n \nt=data['surface']\nx=data['price']\ny=data['arrondissement']\n\nprint(\"----------------------Calcul des coefficients---------\")\nXi=np.matrix([np.ones(data.shape[0]),data['surface'].as_matrix(),y.as_matrix()]).T\nYi=np.matrix(data['price']).T\n\nThetus = np.linalg.inv(Xi.T.dot(Xi)).dot(Xi.T).dot(Yi)\n\nAbscisse=[0,400]\nOrdonnee=[Thetus.item(0)+Abscisse[0]*Thetus.item(1),Thetus.item(0)+Abscisse[1]*Thetus.item(1)]\nDim_Z = [Thetus.item(0)+Abscisse[0]*Thetus.item(2),Thetus.item(0)+Abscisse[1]*Thetus.item(2)]\n\nprint(\"Origine : \",Thetus.item(0))\nprint(\"Coefficient n° 1 : \",Thetus.item(1))\nprint(\"Coefficient n° 2 : \",Thetus.item(2))\n\nprint(\"----------------------Représentation en 3 dimensions---------\")\nax = plt.subplot(111, projection='3d')\nax.scatter(t,x,c=y)\nax = plt.plot(Abscisse,Ordonnee,Dim_Z,linestyle='--',c='#000000')\n\nplt.legend(y)\nplt.show()\n\nprint(\"----------------------Représentation sur 2 dimensions---------\")\nasc = plt.scatter(t,x,c=y)\nasc = plt.plot(Abscisse,Ordonnee,linestyle='--',c='#000000')\nplt.show()\n\nprint(\"----------------------Calcul de la qualité du modèle---------\")\nMat = np.matrix([t.as_matrix(),y.as_matrix(),x.as_matrix()])\ni=0\nprediction = np.empty(data.shape[0])\nerreur = np.empty(data.shape[0])\n\nwhile i < data.shape[0]:\n prediction[i] = Thetus.item(0) + Mat[0].item(i)*Thetus.item(1) + Mat[1].item(i)*Thetus.item(2)\n erreur[i] = np.sqrt((prediction[i] - Mat[2].item(i))**2)\n i += 1 \n\nresidu = erreur.mean() \n\nprint (\"Erreur de prédiction moyenne : \",residu) \nprint (\"Erreur de prédiction cumulée : \",erreur.sum()) \n\ndonnees = np.asarray(Yi).reshape(-1)\n\nCompare = np.matrix([donnees,prediction,erreur])\n\nprint(\"----------------------Prédiction sur données---------\")\nprint(\"Prix du loyer pour une surface de 50 m² : \",Thetus.item(0) + 50*Thetus.item(1) + 10*Thetus.item(2))\n\n\n\n\n\n\n \n \n \n\n \n\n\n\n","sub_path":"OpenClassRoom - Data Scientist/2017-05-25 - Regression Lineaire/1-Representation-Data.py","file_name":"1-Representation-Data.py","file_ext":"py","file_size_in_byte":2201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"63031990","text":"#! /usr/bin/python\n\ndef goodness_of_fit(a1, a2):\n\n import math\n\n diff2 = [(x-y)**2 for x,y in zip(a1,a2)]\n return math.sqrt(sum(diff2)/(max(a1)-min(a1))**2/len(a1))\n\ndef main():\n\n import os\n import imp\n enrs = imp.load_source('enrs',os.environ['RICH_ROOT']+'/analytic/enrs.py')\n\n import numpy\n import h5py\n\n left = enrs.Primitive(1,10,0);\n right = enrs.Primitive(1,1,0);\n prof = enrs.RiemannProfile(left,right,5./3.)\n offset = 0.5\n\n h5file = h5py.File('final.h5','r')\n t = h5file['time']\n x = h5file['grid']\n d = h5file['density']\n p = h5file['pressure']\n v = h5file['x_velocity']\n da = [prof.CalcPrim((i-offset)/t).Density for i in x]\n pa = [prof.CalcPrim((i-offset)/t).Pressure for i in x]\n va = [prof.CalcPrim((i-offset)/t).Velocity for i in x]\n\n gof1 = goodness_of_fit(d,da)\n gof2 = goodness_of_fit(p,pa)\n gof3 = goodness_of_fit(v,va)\n\n f = open('gradesheet.txt','w')\n f.write(str(gof1)+'\\n')\n f.write(str(gof2)+'\\n')\n f.write(str(gof3)+'\\n')\n f.close()\n\n return gof1<0.14 and gof2<0.06 and gof3 < 0.13\n\nif __name__=='__main__':\n \n import os\n\n if main():\n os.system('touch test_passed.res')\n else:\n os.system('touch test_failed.res')\n\n","sub_path":"lib/SurfacingAlgorithms/huji-rich-Elad3DFast/tests/newtonian/one_dimensional/riemann_profiles_1/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":1255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"199128378","text":"import MySQLdb\n\n\n# 打开数据库\ndb = MySQLdb.connect(\"localhost\", \"testuser\", \"testuser123\", \"TESTDB\")\n\n\n# 使用cursor()方法获取操作游标\ncursor = db.cursor()\n\n# 使用execute方法执行sql语句\ncursor.execute(\"SELECT VERSION()\")\n\n# 如果数据库表已经存在使用 execute() 方法删除表\ncursor.execute(\"DROP TABLE IF EXISTS EMPLOYEE\")\n\n# 创建数据表SQL 语句\nsql = \"\"\"CREATE TABLE EMPLOYEE(\n FRIST_NAME CHAR(20) NOT NULL,\n LAST_NAME CHAR(20),\n AGE INT,\n SEX CHAR(1),\n INCOME FLOAT)\"\"\"\ncursor.execute(sql) \n\n\n# 使用fetchone()方法获取一条数据\ndata = cursor.fetchone()\n\n\n# SQL 插入语句\nsql = \"\"\"INSERT INTO EMPLOYEE(FIRST_NAME,LAST_NAME,AGE,SEX,INCOME)\n VALUES('MAC','Mohan',20,'M', 2000)\"\"\"\n\ntry:\n # 执行sql语句\n cursor.execute(sql)\n\n # 获取所有记录列表\n results = cursor.fetchall()\n for row in results:\n fname = row[0]\n lname = row[1]\n age = row[2]\n sex = row[3]\n income = row[4]\n print(\"fname=%s, lname=%s, age=%d, sex=%s, income=%d\" % \\\n (fname, lname, age, sex, income))\n # 提交到数据库\n db.commit()\nexcept:\n print(\"Error: unable to fetch data\")\n # rollback in case there is any error\n db.rollback() \n\n\nprint(\"database version : %s\" % data)\n\n# 关闭数据库\n\ndb.close()","sub_path":"practice-mysql.py","file_name":"practice-mysql.py","file_ext":"py","file_size_in_byte":1375,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"193496508","text":"fact = {}\ndef Factorial(a):\n\tif a <= 1:\n\t\treturn 1\n\telse:\n\t\treturn a * Factorial(a - 1)\n\nfor i in range(0, 10):\n\tfact[i] = Factorial(i)\n\ne = 10\nfor i in range(11, 1000000):\n\tsummation = 0\n\tif i % (e * 10) == 0:\n\t\te *= 10\n\tj = 1\n\twhile True:\n\t\tsummation += fact[int(i / j) % 10]\n\t\tif summation > i:\n\t\t\tbreak\n\t\telif j == e:\n\t\t\tbreak\n\t\tj *= 10\n\tif summation == i:\n\t\tprint(i)\n","sub_path":"Dfact.py","file_name":"Dfact.py","file_ext":"py","file_size_in_byte":372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"581332553","text":"class Solution(object):\n def maxProduct(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: int\n \"\"\"\n if not nums:\n return 0\n\n ret = nums[0]\n max_l = [nums[0]]\n min_l = [nums[0]]\n\n for i in range(1, len(nums)):\n max_now = max((max_l[i - 1] * nums[i], min_l[i - 1] * nums[i], nums[i]))\n min_now = min((max_l[i - 1] * nums[i], min_l[i - 1] * nums[i], nums[i]))\n ret = max(ret, max_now)\n max_l.append(max_now)\n min_l.append(min_now)\n\n return ret\n","sub_path":"leetcode/algorithm/maximum-product-subarray.py","file_name":"maximum-product-subarray.py","file_ext":"py","file_size_in_byte":580,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"131437114","text":"#!usr/bin/env python\nfrom calendar import monthrange\nimport h5py\nimport os\nfrom os.path import join, realpath, dirname, exists\nfrom netCDF4 import Dataset\nimport datetime as dt\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfrom mpl_toolkits.basemap import Basemap\nfrom matplotlib.cm import ScalarMappable\nfrom matplotlib.colors import Normalize\n\nfrom ..utils import time_utils\nfrom ..utils import orbit_utils\n\nfrom .earthdata import re_km\n\nimport warnings\n\n# location of MERRA files\nmerra_dir = realpath(join(dirname(__file__), '..', '..', 'MERRA_Data'))\nfilepath = join(\n merra_dir, '{Y}', '{m}', 'MERRA2_400.tavg3_3d_asm_Nv.{strf8}.nc4')\n\n# lat and lon resolutions of merra data\nlonres = 0.625\nlatres = 0.5\n\n\nclass Clouds(object):\n \"\"\"\n The Clouds class transfers data from MERRA-2\n HDF5 files into class attributes to be used \n in other parts of the code. \n \n One of the most important aspects of this class\n is its ability to interpolate cloud fractions \n between the times provided in the files.\n \n Also contains a plotting method to show the \n global cloud cover.\n \"\"\"\n \n def __init__(self, time):\n \"\"\"\n Initialize the Clouds instance.\n \n args:\n \n time - time_utils.Time instance. The\n time associated with the cloud\n data to be read in. As MERRA \n comes in daily files, only the\n year, month and date of the time\n instance are considered\n \"\"\"\n # get the filename for the date in question\n self.fname = filepath.format(Y=time.strftime(\"%Y\"), m=time.strftime(\"%m\"), strf8=time.strftime(\"%Y%m%d\"))\n if not os.path.exists(self.fname):\n raise IOError(\"Specified MERRA data file {0} does not exist\".format(self.fname))\n print(\"Reading cloud data from MERRA file\", self.fname)\n self.merra = Dataset(self.fname, 'r')\n\n # save MERRA datasets as attributes\n # suppress some weird warning: valid_range not used since it cannot be safely cast to variable data type\n with warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\n self.time = self.merra.variables[\"time\"][:]\n self.cloud = self.merra.variables[\"CLOUD\"][:]\n \n \n # Get the times corresponding to each time step in the dataset\n if type(time) == type(dt.datetime(2015,5,5)):\n time = time_utils.Time(time, time)\n d0 = dt.datetime(time.datetime.year, time.datetime.month, time.datetime.day, 1, 30)\n t0 = time_utils.Time(d0, time.ref)\n self.times = [t0 + 60*t for t in self.time]\n \n \n # take the maximum value along the vertical axis\n self.cloud = np.amax(self.cloud, axis=1)\n \n \n # append the first column of lons to the end, \n # so that locations in the range (179.6875 E, 180 E]\n # can get assigned the proper cloud fraction values\n first_cld = np.array([np.vstack(self.cloud[k,:,0]) for k in range(self.cloud.shape[0])])\n self.cloud = np.append(self.cloud, first_cld, axis=2)\n \n def __call__(self, time):\n \"\"\"\n Perform a linear interpolation of \n the cloud data along the time axis.\n \n args:\n \n time - time_utils.Time instance. The\n exact time the data will get\n interpolated to. If the time \n falls outside the first and last\n time steps, then the first or\n last set of values will be returned\n \"\"\"\n # return the first timestep's array for low values\n if type(time) == type(dt.datetime(2015,5,5)):\n time = time_utils.Time(time, time)\n \n if time.datetime <= self.times[0].datetime:\n return self.cloud[0,:,:]\n \n # return the last timestep's array for high values\n elif time.datetime > self.times[-1].datetime:\n return self.cloud[-1,:,:]\n \n else:\n # find the time steps that bound the new time\n for k in range(len(self.times)-1):\n if self.times[k].datetime < time.datetime <= self.times[k+1].datetime:\n lower = k\n upper = k+1\n \n # create the weights for the intepolation\n tlower = 0\n tupper = self.times[upper] - self.times[lower]\n t = time - self.times[lower]\n \n wlower = (t - tupper)/(tlower-tupper)\n wupper = (t - tlower)/(tupper-tlower)\n \n # interpolate cloud values to the new time\n cld = wlower*self.cloud[lower,:,:] + wupper*self.cloud[upper,:,:]\n \n return cld\n \n \n def show_clouds_ortho(self, time, centre, map_bkgd='bluemarble', out='show'):\n \"\"\"\n Show an orthographic projection with \n cloud cover that ranges from transparent\n for small cloud fraction, to white to\n grey.\n \n args:\n \n time - time_utils.Time instance that the\n cloud field will get interpolated to\n and shown for.\n \n centre - tuple of (lat, lon), indicating the\n centre of the orthographic projection\n \n map_bkgd - string to indicate what kind of back-\n ground will be used for the image. \n Can be any of 'bluemarble', 'etopo',\n or 'mask'\n \n out - string to indicate what kind of output\n should be provided\n \"\"\"\n if type(time) == type(dt.datetime(2015, 4, 4)):\n time = time_utils.Time(time, time)\n\n def col(x):\n \"\"\"\n Function that provides RGBA values\n for a given cloud fraction between\n 0 and 1. \n \"\"\"\n \n # get RGB value\n c = 1. - 0.3*x**2.\n \n # transparent if c.f. < 0.1\n if x < 0.1:\n return (c,c,c,0)\n \n # opaque if c.f. > 0.7\n elif x > 0.7:\n return (c,c,c,1.)\n \n # translucent if 0.1 < c.f. < 0.7\n else:\n return (c,c,c,np.clip((x - 0.1)/0.7, 0., 1.)**2.)\n \n # draw map projection\n projection = Basemap(projection='ortho', lat_0=centre[0], lon_0=centre[1], resolution='c')\n \n # determine what kind of background\n # is needed, and draw it\n if map_bkgd == 'bluemarble':\n projection.bluemarble()\n elif map_bkgd == 'etopo':\n projection.etopo()\n elif map_bkgd == 'mask':\n projection.drawlsmask(land_color='limegreen', ocean_color='dodgerblue', resolution='l')\n \n projection.drawcoastlines(linewidth=0.75)\n projection.drawcountries(linewidth=0.5)\n projection.drawmeridians(np.arange(-180, 181, 30), latmax=90)\n projection.drawparallels(np.arange(-90, 91, 15), latmax=90)\n else:\n raise ValueError('invalid map background specification')\n \n # use a call to interpolate the cloud \n # fields to the needed time \n cf = self(time)\n\n # define MERRA lat and lon grid\n lats = np.arange(-90., 90.25, 0.5)\n lons = np.arange(-180., 180.3125, 0.625)\n lons, lats = np.meshgrid(lons, lats)\n \n # define contour levels and colors\n levs = np.linspace(0,1,256, endpoint=True)\n clevs = [col(x) for x in levs]\n \n # get the map projection coordinates\n # and mask the cloud data, so that lcoations\n # on the opposite side of th Earth don't get\n # drawn\n x, y = projection(lons, lats)\n \n cfmask = np.ma.masked_greater(x, 1e15).mask\n cfmasked = np.ma.array(cf, mask=cfmask)\n \n # draw contour\n plt.contourf(x, y, cfmasked, levs, colors=clevs, extend='both')\n \n # supply specified output\n if out == 'show':\n plt.show()\n elif out == '':\n pass\n else:\n plt.savefig(out)\n return cfmask\n\n def get_clouds(self, time, centre, map_bkgd='bluemarble', out='show'):\n \"\"\"show_clouds_ortho, but without the plotting stuff\"\"\"\n # use a call to interpolate the cloud \n # fields to the needed time \n cf = self(time)\n return cf\n\nclass CloudCollection(object):\n \n \"\"\"\n Very Similar class to the above Clouds, but\n the CloudCollection class has the ability to\n concatenate arrays from multiple MERRA files,\n so that more than 1 day's worth of values can\n be collected.\n \n The methods to interpolate to specfied times\n and to show an orthographic projection of the\n cloud field values are repeated as well.\n \"\"\"\n \n def __init__(self, time_middle, obs_period):\n \"\"\"\n Intialize the CloudCollention instance.\n \n args:\n \n time_middle - the midpoint in time of the\n period for which cloud fields\n are needed\n \n obs_period - The total length of time for which\n the cloud fields are needed. Units\n are in seconds\n \"\"\"\n print(\"Opening cloud files for a time interval centered at {} UTC\".format(time_middle.datetime))\n \n self.time_middle = time_middle\n self.obs_period = obs_period\n \n # find the endpooints of the time period\n obs_start = time_middle - (obs_period//2+3600)\n obs_end = time_middle + (obs_period//2+3600)\n self.obs_start = obs_start\n self.obs_end = obs_end\n \n # The first time step in the MERRA daily files\n # is at 01:30 UTC, and the final one is at 22:30\n # UTC. Therefore, if the endpoints sit outside this\n # range, an extra day will have to be loaded so \n # that all the necessary interpolations can be done\n merra_start = dt.datetime(obs_start.datetime.year, obs_start.datetime.month, obs_start.datetime.day, 1, 30)\n merra_end = dt.datetime(obs_end.datetime.year, obs_end.datetime.month, obs_end.datetime.day, 22, 30)\n \n # check if the starting point occurs\n # before 01:30 UTC\n if obs_start.datetime < merra_start:\n merra_start -= dt.timedelta(seconds=86400)\n \n # check if the ending point occurs\n # after 22:30 UTC \n if obs_end.datetime > merra_end:\n merra_end += dt.timedelta(seconds=86400)\n \n # append all the necessary dates to a list\n dates = []\n cdate = merra_start.date()\n while cdate <= merra_end.date():\n dates.append(dt.datetime(cdate.year, cdate.month, cdate.day, 12))\n cdate += dt.timedelta(days=1)\n \n # create a cloud instance for each day,\n # and concatenate their arrays together\n # to get the full arrays for the CloudCollection \n clouds = []\n for d in dates:\n t = time_utils.Time(d, time_middle.ref)\n clouds.append(Clouds(t))\n \n self.cloud = np.concatenate([c.cloud for c in clouds], axis=0)\n self.times = np.concatenate([c.times for c in clouds])\n \n \n def __call__(self, time):\n \"\"\"\n Equivalent Function to the call method\n for Clouds. Interpolates cloud fields to\n the specified time\n \"\"\"\n if time.datetime < self.obs_start.datetime or time.datetime > self.obs_end.datetime:\n raise ValueError(\"Specified time is outside the observation period\")\n else:\n for k in range(len(self.times)-1):\n if self.times[k].datetime < time.datetime <= self.times[k+1].datetime:\n lower = k\n upper = k+1\n \n tlower = 0\n tupper = self.times[upper] - self.times[lower]\n t = time - self.times[lower]\n \n wlower = (t - tupper)/(tlower-tupper)\n wupper = (t - tlower)/(tupper-tlower)\n \n cld = wlower*self.cloud[lower,:,:] + wupper*self.cloud[upper,:,:]\n \n return cld\n \n def show_clouds_ortho(self, time, center, map_bkgd='bluemarble', out='show'):\n \"\"\"\n Equivalent function to the show function\n in Clouds. provide a map of the cloud fields\n at the specified time\n \"\"\"\n def col(x):\n \n c = 1. - 0.3*x**2.\n \n if x < 0.1:\n return (c,c,c,0)\n elif x > 0.8:\n return (c,c,c,1.)\n else:\n return (c,c,c,np.clip((x - 0.1)/0.7, 0., 1.)**2.)\n \n \n \n projection = Basemap(projection='ortho', lat_0=center[0], lon_0=center[1], resolution='l') \n \n if map_bkgd == 'bluemarble':\n projection.bluemarble(zorder=1)\n elif map_bkgd == 'etopo':\n projection.etopo(zorder=1)\n elif map_bkgd == 'mask':\n projection.drawlsmask(land_color='limegreen', ocean_color='dodgerblue', resolution='l', zorder=1)\n projection.drawcoastlines(linewidth=0.75, zorder=3)\n projection.drawcountries(linewidth=0.5, zorder=3)\n projection.drawmeridians(np.arange(-180, 181, 30), latmax=90, linewidth=0.5, zorder=3)\n projection.drawparallels(np.arange(-90, 91, 15), latmax=90, linewidth=0.5, zorder=3)\n else:\n raise ValueError('invalid map background specification')\n \n cf = self(time)\n \n lats = np.arange(-90., 90.25, 0.5)\n lons = np.arange(-180., 180.3125, 0.625)\n lons, lats = np.meshgrid(lons, lats)\n \n\n levs = np.linspace(0,1,256, endpoint=True)\n clevs = [col(x) for x in levs]\n \n \n x,y = projection(lons, lats)\n \n cfmask = np.ma.masked_greater(x, 1e15).mask\n cfmasked = np.ma.array(cf, mask=cfmask)\n \n plt.contourf(x, y, cfmasked, levs, colors=clevs, extend='both', zorder=2)\n \n if out == 'show':\n plt.show()\n elif out == '':\n pass\n else:\n plt.savefig(out)\n \n \ndef monthly_cloudiness(year, month, centre, cld_threshold=0.05, out='show', data_type='avg per day'):\n \"\"\"\n Method to show how cloudy different regions are\n over the course of a month. This is shown by an\n average number of cloud free hours per day. \n \n This code is largely a replica of Dan's code to\n do the same process, but this will likely run \n much faster. \n \n args:\n \n year - int. year which the data will be shown\n for.\n \n month - int. month which the data will be shown\n for\n \n centre - tuple of (lat,lon) indicating the centre\n of the orthographic projection onto which\n values will be plotted\n \n cld_threshold - maxmimum cloud fraction for a grid\n square to be considered cloud-free\n \n out - string to indicate what kind of output\n to provide\n \n data_type - what kind of data to provide. 'daylight fraction'\n will give the fraction of daylight hours that\n are cloud free, and 'avg per day' will give the\n average number of cloud free daylight hours per day\n \"\"\"\n \n def sun_rise_set(date, latitude):\n \"\"\"\n Determine the sunrise and sunset times,\n given a date and latitude\n \n https://en.wikipedia.org/wiki/Sunrise_equation\n \n args:\n \n date - dt.date instance of the day of the year\n \n latitude - array of latitudes to get the sun\n rise and sun set times\n \"\"\"\n j = int(date.strftime('%j'))\n \n \n solar_dec = 23.45*np.sin(2.*np.pi*((284.+j)/365.))\n A = -1.*np.tan(np.radians(latitude))*np.tan(np.radians(solar_dec))\n \n half_day_length = np.zeros(A.size, dtype=np.float64)\n \n interval = np.ma.masked_inside(A, -1., 1.).mask\n \n \n half_day_length[A < -1.] = 12.\n half_day_length[interval] = np.degrees(np.arccos(A[interval]))/15.\n \n rise = 12. - half_day_length\n set = 12. + half_day_length\n \n return (rise, set, A)\n \n \n filepath = \"/wrk8/MERRA/goldsmr5.gesdisc.eosdis.nasa.gov/%Y/%m/MERRA2_400.tavg3_3d_asm_Nv.%Y%m%d.nc4\"\n \n # construct a range of dates for the month\n d0 = dt.datetime(year, month, 1)\n d1 = dt.datetime(year+1, 1, 1) if month == 12 else dt.datetime(year, month+1, 1)\n date_range = time_utils.Time.arange(d0, d1, dt.timedelta(days=1))\n \n total_cloud_free = np.zeros((361, 577))\n total_daylight = np.zeros((361,577))\n \n for d in date_range:\n \n print(d)\n \n \n # load in the MERRA data for the day\n try: \n merra = Dataset(d.strftime(filepath), \"r\")\n \n except IOError:\n raise IOError('specified MERRA file %s does not exist' % d.strftime(filepath))\n \n \n clear_daylight = np.zeros((8,361,577))\n \n # get all the needed data\n lat = merra['lat'][...]\n lon = merra['lon'][...]\n cld = merra['CLOUD'][...]\n time = merra['time'][...]\n \n \n cld_first_column = cld[...,:1]\n cld = np.append(cld, cld_first_column, axis=-1)\n \n lon = np.append(lon, lon[-1] + 0.625)\n \n sunrise, sunset, _ = sun_rise_set(d, lat)\n \n cld_clear = np.sum(cld < cld_threshold, axis=1) == 72\n \n time_hours = (time + 90.) / 60.\n local_offset = (24./360.)*lon\n \n time_start = (time_hours[:,np.newaxis]*np.ones((1,lat.size)))[...,np.newaxis] + local_offset[np.newaxis,np.newaxis,:]\n time_finish = time_start + 3.\n \n time_start[time_start < 0.] += 24.\n time_start[time_start > 24.] -= 24.\n time_finish[time_finish < 0.] += 24.\n time_finish[time_finish > 24.] -= 24.\n \n \n sunrise = sunrise[np.newaxis,:,np.newaxis]*np.ones((time.size,1,lon.size))\n sunset = sunset[np.newaxis,:,np.newaxis]*np.ones((time.size,1,lon.size))\n \n before_sunrise = time_start <= sunrise\n after_sunrise = time_start > sunrise\n \n before_sunset = time_finish <= sunset\n after_sunset = time_finish > sunset\n \n within_interval = cld_clear*before_sunrise*after_sunset\n add_time = sunset[within_interval] - sunrise[within_interval]\n add_time[add_time < 0.] = 0.\n clear_daylight[within_interval] = add_time\n \n starts_in_interval = cld_clear*before_sunrise*before_sunset\n add_time = time_finish[starts_in_interval] - sunrise[starts_in_interval]\n add_time[add_time < 0.] = 0.\n clear_daylight[starts_in_interval] = add_time\n \n ends_in_interval = cld_clear*after_sunset*after_sunrise\n add_time = sunset[ends_in_interval] - time_start[ends_in_interval]\n add_time[add_time < 0.] = 0.\n clear_daylight[ends_in_interval] = add_time\n \n full_interval = cld_clear*after_sunrise*before_sunset\n add_time = time_finish[full_interval] - time_start[full_interval]\n add_time[add_time < 0.] = 0.\n clear_daylight[full_interval] = add_time \n \n clear_daylight_daily = np.sum(clear_daylight, axis=0)\n daylight_daily = sunset[0] - sunrise[0]\n \n total_cloud_free += clear_daylight_daily\n total_daylight += daylight_daily\n \n merra.close()\n \n \n lon_grid, lat_grid = np.meshgrid(lon, lat)\n \n projection = Basemap(projection='robin', lat_0=centre[0], lon_0=centre[1], resolution='l')\n \n projection.drawcoastlines()\n projection.drawcountries()\n\n \n x, y = projection(lon_grid, lat_grid)\n \n cmask = np.ma.masked_greater(x, 1e15).mask\n cmasked = np.ma.array(total_cloud_free, mask=cmask)\n \n if data_type == 'daylight fraction':\n \n daylight_filled = np.ma.masked_values(total_daylight, 0.).filled(1.) \n fraction = cmasked / daylight_filled\n \n vmin = 0.\n vmax = 1. \n \n sm = ScalarMappable(Normalize(vmin=vmin, vmax=vmax), cmap='nipy_spectral_r')\n levs = np.linspace(vmin, vmax, 256)\n clevs = [sm.to_rgba(lev) for lev in levs]\n \n elif data_type == 'avg per day':\n total_hours = len(date_range)\n fraction = cmasked / total_hours\n \n vmin = 0.\n vmax = 12. \n \n sm = ScalarMappable(Normalize(vmin=vmin, vmax=vmax), cmap='nipy_spectral_r')\n levs = np.linspace(vmin, vmax, 256)\n clevs = [sm.to_rgba(lev) for lev in levs]\n \n else:\n raise ValueError\n \n \n ctr = plt.contourf(x, y, fraction, levs, colors=clevs, extend='both')\n \n \n cbar = plt.colorbar(ctr, orientation='vertical', aspect=25, fraction=0.1, shrink = 0.625)\n cbar.set_ticks(np.linspace(vmin, vmax, 11))\n \n if data_type == 'daylight fraction':\n cbar.set_label(d0.strftime('Fraction of Cloud-Free Daylight for %B %Y'))\n elif data_type == 'avg per day':\n \n cbar.set_ticks(np.arange(0, 13, 2))\n \n if out == 'show':\n plt.show()\n elif out == '':\n pass\n else:\n plt.savefig(os.path.join('../figures/cloud_free_daylight_maps/', out), dpi=240)\n plt.close()\n \n\n \ndef animate(centre, start_time, end_time, map_bkgd='bluemarble'):\n \"\"\"\n Create an animation of cloud fields maps\n for each of the timesteps within a \n CloudCollection instance.\n \n args:\n \n centre - tuple of (lat,lon) indicating the centre\n of the orthographic projection.\n \n start_time - dt.datetime or time_utils.Time instance\n of the start time of the interval\n \n end_time - dt.datetime or time_utils.Time instance\n of the end time of the interval\n \n map_bkgd - string indicating what kind of background\n to use.\n \"\"\"\n \n if isinstance(start_time, dt.datetime):\n t1 = time_utils.Time(start_time, start_time)\n else:\n t1 = start_time\n \n if isinstance(end_time, dt.datetime):\n t2 = time_utils.Time(end_time, t1.ref)\n else:\n t2 = end_time.change_ref(t1.ref)\n \n # find the midpoint and length of the time interval\n delta = t2 - t1\n midpoint = t1 + 0.5*delta\n \n # initialize cloud fields\n cld = CloudCollection(midpoint, delta)\n \n # create a temporary directory to store the\n # images that will make up the video\n os.system('mkdir ../figures/animation_temp')\n \n # Draw cloud map for each time step in the CloudCollection,\n # saving each one as a separate image\n for k,t in enumerate(cld.times):\n try:\n print(t.datetime)\n cld.show_clouds_ortho(t, centre, out='', map_bkgd=map_bkgd)\n \n plt.title(t.strftime('%d %b %Y %H:%M:%S'))\n \n plt.savefig('../figures/animation_temp/cloud{:03}.png'.format(k))\n plt.close()\n except ValueError:\n pass\n \n # create the video, and then delete the individual images\n # and the temporary folder \n os.system('mencoder -o ../figures/clouds_{0}_{1}.avi mf://../figures/animation_temp/cloud*.png -ovc lavc -lavcopts vcodec=msmpeg4v2 -mf fps=2'.format(t1.strftime('%y%m%d'), t2.strftime('%y%m%d'))) \n os.system('rm ../figures/animation_temp/*.png')\n os.system('rmdir ../figures/animation_temp') \n \n \ndef download_merra_data(year, month):\n '''\n Notes:\n This function was created using these instructions from NASA:\n https://disc.gsfc.nasa.gov/data-access#mac_linux_wget\n Also note that MERRA data is only published for up to about 2 months ago\n (from whever you're reading this,\n e.g. I only have data up to July and it's mid-September).\n If you try to get data from the future none will download\n (nothing will crash though).\n Inputs:\n year (string): the year you want to download\n month (string): the month you want to download\n No outputs, just saves the MERRA data to wherever you tell it to go.\n '''\n print(\"Downloading MERRA Data for\", year, month)\n print(\"this will take approximately forever\")\n user = \"CallumMcCracken\" # input(\"Enter earthdata.nasa.gov username: \")\n password = \"ThisIsMyV3ryStrongPassword\" # input(\"Enter password: \")\n cwd = os.getcwd() # so we can get back here at the end\n os.chdir(os.path.expanduser(\"~\"))\n if not os.path.exists(\".netrc\"):\n os.system(\"touch .netrc\")\n os.system('echo \"machine urs.earthdata.nasa.gov '\n + 'login {} '.format(user)\n + 'password {}\" >> .netrc'.format(password))\n os.system('chmod 777 .netrc') # so future people can edit\n os.system('touch .urs_cookies')\n \n # (this will take a bit of an eternity to run)\n os.chdir(merra_dir)\n wget_format = \"https://goldsmr5.gesdisc.eosdis.nasa.gov/opendap/MERRA2/M2T3NVASM.5.12.4/{year}/{month:02}/MERRA2_400.tavg3_3d_asm_Nv.{year}{month:02}{day:02}.nc4.nc?CLOUD[0:7][0:71][0:360][0:575],lat[0:360],time[0:7],lon[0:575],lev\"\n for day in range(1, monthrange(year, month)[1]+1):\n print(\"working on day\", day)\n os.system(\"wget --load-cookies ~/.urs_cookies \"\n \"--save-cookies ~/.urs_cookies \"\n \"--keep-session-cookies \"\n \"--cut-dirs 3 \"\n \"-r -c -nH -nd -np \"\n \"-A MERRA2_400.tavg3_3d_asm_Nv.{year}{month:02}{day:02}* \".format(year=year, month=month, day=day) +\n \"--content-disposition \" +\n wget_format.format(year=year, month=month, day=day))\n print(\"done day\", day)\n\n for fname in os.listdir():\n # rename all .nc4.nc files as just .nc4, plus move them\n if fname.endswith(\"nc4.nc\"):\n new_filename = fname[:-3]\n new_dir = join(\"{}\".format(year), \"{:02}\".format(month))\n if not exists(new_dir):\n os.mkdir(new_dir)\n os.system(\"mv \"+fname+\" \"+join(new_dir, new_filename))\n # remove wget logs\n if \"wget-log\" in fname:\n os.remove(fname)\n\n # at the end return us to the cwd\n os.chdir(cwd)\n\n \nif __name__ == '__main__':\n dtime = dt.datetime(2015,10,23)\n time = time_utils.Time(dtime,dtime)\n clouds = Clouds(time).show_clouds_ortho(time, (0,0), map_bkgd='mask') \n","sub_path":"modules/earth_data/clouds.py","file_name":"clouds.py","file_ext":"py","file_size_in_byte":27211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"316367499","text":"\nimport os\nfrom math import pi\nimport numpy as np\nfrom matplotlib import pyplot as plt\nimport bayesik as bik\n\n\n\n\nnlinks = 2\nfnameNPZ = os.path.join( bik.dirREPO, 'Data', f'links{nlinks}.npz')\nscale = [1000, 1000] + [180/pi]*nlinks\nwith np.load(fnameNPZ) as Z:\n\tQ = scale * Z['Q']\n\tQB = scale * Z['QB']\n\tQLS = scale * Z['QLS']\nEB0,ELS0 = QB - Q, QLS - Q\n\n\n\n\nplt.close('all')\nfig,AX = plt.subplots( 2, 2, figsize=(6,4) )\nfontname = u'Times New Roman'\n\nc0,a0,ec0 = 'k', 1, '0.7'\ncB,aB,ecB = 'b', 0.7, '0'\n\n### posture model errors:\nrangesLS = [(-4,4), (-4,4), (-2,2), (-2,2)]\nrangesB = rangesLS\nfor i,ax in enumerate(AX.ravel()):\n\tax.hist(ELS0[:,i], bins=21, range=rangesLS[i], color=c0, ec=ec0, alpha=a0, zorder=-1, label='Least-squares')\n\tax.hist(EB0[:,i], bins=21, range=rangesB[i], color=cB, ec=ecB, alpha=aB, label='Bayesian')\n\tif i==0:\n\t\tleg = ax.legend(loc='upper right', bbox_to_anchor=(1.1,1))\n\t\tplt.setp(leg.get_texts(), size=8, name=fontname)\n\n\n### label axes:\n[ax.set_ylabel('Frequency', name=fontname, size=14) for ax in AX[:,0]]\nlabels = [r'$r_x$ $\\textrm{error (mm)}$', r'$r_y$ $\\textrm{error (mm)}$', r'$\\phi_1$ $\\textrm{error (mm)}$', r'$\\phi_2$ $\\textrm{error (mm)}$']\n[ax.set_xlabel(s, size=14, usetex=True, name=fontname) for ax,s in zip(AX.ravel(),labels)]\n[plt.setp(ax.get_xticklabels() + ax.get_yticklabels(), name=fontname, size=9) for ax in AX.ravel()]\n\n\nplt.tight_layout()\nplt.show()\n\n\n# plt.savefig( os.path.join( bik.dirREPO, 'Appendix', 'ipynb', 'figs', f'error-{nlinks}-link.png') )\n\n\n\n","sub_path":"Python/figE1.py","file_name":"figE1.py","file_ext":"py","file_size_in_byte":1545,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"549940846","text":"# Definition for singly-linked list.\nclass ListNode(object):\n def __init__(self, x):\n self.val = x\n self.next = None\n\nclass Solution(object):\n def _connect(self, c, L, carry):\n while L is not None:\n val = ( L.val + carry ) % 10\n carry = ( L.val + carry ) // 10\n L.val = val\n c.next = L\n c = c.next\n L = L.next\n if carry == 1:\n c.next = ListNode(1)\n\n def addTwoNumbers(self, l1, l2):\n \"\"\"\n :type l1: ListNode\n :type l2: ListNode\n :rtype: ListNode\n \"\"\"\n carry = 0\n head = ListNode(0)\n curr = head\n while l1 is not None or l2 is not None:\n if l1 is None:\n self._connect (curr, l2, carry)\n return head.next\n if l2 is None:\n self._connect (curr, l1, carry)\n return head.next\n\n val = (l1.val + l2.val + carry) % 10\n carry = (l1.val + l2.val + carry) // 10\n tmp = ListNode(val)\n curr.next = tmp\n curr = curr.next\n l1 = l1.next\n l2 = l2.next\n if carry == 1:\n curr.next = ListNode(1)\n return head.next\n\n\n\n\n\nsol = Solution()\ndef createList(L):\n head = ListNode(0)\n curr = head\n for i in L:\n tmp = ListNode(i)\n curr.next = tmp\n curr = curr.next\n # printList (head.next)\n return head.next\ndef printList(L):\n while L is not None:\n print (L.val, end=' ')\n L = L.next\n print(\"\\n\")\nprint(\"L1\")\nL1 = createList([9,9])\nprint(\"L2\")\nL2 = createList([1])\nresult = sol.addTwoNumbers(L1, L2)\nprintList(result)\n","sub_path":"2018_google/002_sol.py","file_name":"002_sol.py","file_ext":"py","file_size_in_byte":1704,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"650025799","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.2 (3180)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/__unit_test__/ally/cdm/impl/local_filesystem.py\n# Compiled at: 2013-10-02 09:54:40\n\"\"\"\nCreated on Jan 9, 2012\n\n@package support - cdm\n@copyright 2012 Sourcefabric o.p.s.\n@license http: // www.gnu.org / licenses / gpl - 3.0.txt\n@author: Mugur Rus\n\nProvides unit testing for the local filesystem module.\n\"\"\"\nimport package_extender\npackage_extender.PACKAGE_EXTENDER.setForUnitTest(True)\nfrom ally.cdm.impl.local_filesystem import HTTPDelivery, LocalFileSystemCDM, LocalFileSystemLinkCDM\nfrom ally.cdm.spec import PathNotFound\nfrom ally.zip.util_zip import normOSPath\nfrom datetime import datetime\nfrom io import BytesIO\nfrom os import makedirs, remove, sep, stat\nfrom os.path import join, dirname, isfile, isdir\nfrom shutil import rmtree\nfrom tempfile import NamedTemporaryFile, TemporaryDirectory\nimport json, re, unittest\nnormpath = lambda txt: re.sub('[\\\\W]+', '', txt)\n\nclass TestHTTPDelivery(unittest.TestCase):\n\n def testHTTPDelivery(self):\n d = HTTPDelivery()\n d.serverURI = 'http://localhost:8080/content/'\n self.assertEqual(d.getURI('somedir/somefile.jpg'), 'http://localhost:8080/content/somedir/somefile.jpg', 'Computing the URI')\n\n def testLocalFilesystemCDM(self):\n d = HTTPDelivery()\n rootDir = TemporaryDirectory()\n d.serverURI = 'http://localhost/content/'\n d.repositoryPath = rootDir.name\n cdm = LocalFileSystemCDM()\n cdm.delivery = d\n try:\n srcTmpFile = NamedTemporaryFile(delete=False)\n srcTmpFile.close()\n dstPath = 'testdir1/tempfile.txt'\n cdm.publishFromFile(dstPath, srcTmpFile.name)\n dstFilePath = join(d.getRepositoryPath(), normOSPath(dstPath))\n self.assertTrue(isfile(dstFilePath))\n self.assertEqual(datetime.fromtimestamp(stat(dstFilePath).st_mtime), cdm.getTimestamp(dstPath))\n finally:\n rmtree(dirname(dstFilePath))\n remove(srcTmpFile.name)\n\n try:\n inFileName = join('dir1', 'subdir2', 'file1.txt')\n dstPath = join('testdir2', 'tempfile2.txt')\n cdm.publishFromFile(dstPath, join(dirname(__file__), 'test.zip', inFileName))\n dstFilePath = join(d.getRepositoryPath(), normOSPath(dstPath))\n self.assertTrue(isfile(dstFilePath))\n finally:\n rmtree(dirname(dstFilePath))\n\n srcTmpDir = TemporaryDirectory()\n dirs = ('test1/subdir1', 'test2/subdir1')\n for dir in dirs:\n fullPath = join(srcTmpDir.name, dir)\n makedirs(fullPath)\n with open(join(fullPath, 'text.html'), 'w') as (_f):\n pass\n\n try:\n cdm.publishFromDir('testdir3', srcTmpDir.name)\n dstDirPath = join(d.getRepositoryPath(), 'testdir3')\n for dir in dirs:\n dstFilePath = join(dstDirPath, dir, 'text.html')\n self.assertTrue(isfile(dstFilePath))\n self.assertEqual(datetime.fromtimestamp(stat(dstFilePath).st_mtime), cdm.getTimestamp(join('testdir3', dir, 'text.html')))\n\n filePath = 'testdir3/test1/subdir1/text.html'\n self.assertTrue(isfile(join(d.getRepositoryPath(), filePath)))\n cdm.republish(filePath, filePath + '.new')\n self.assertTrue(isfile(join(d.getRepositoryPath(), filePath + '.new')))\n cdm.republish(filePath + '.new', filePath)\n cdm.remove(filePath)\n self.assertFalse(isfile(join(d.getRepositoryPath(), filePath)))\n dirPath = 'testdir3/test2'\n self.assertTrue(isdir(join(d.getRepositoryPath(), dirPath)))\n cdm.remove(dirPath)\n self.assertFalse(isdir(join(d.getRepositoryPath(), dirPath)))\n finally:\n rmtree(dstDirPath)\n\n try:\n cdm.publishFromDir('testdir4', join(dirname(__file__), 'test.zip', 'dir1'))\n dstDirPath = join(d.getRepositoryPath(), 'testdir4')\n dstFilePath = join(dstDirPath, 'subdir1', 'file1.txt')\n self.assertTrue(isfile(dstFilePath))\n dstFilePath = join(dstDirPath, 'subdir2', 'file2.txt')\n self.assertTrue(isfile(dstFilePath))\n dstFilePath = join(dstDirPath, 'sometestfile.txt')\n with open(dstFilePath, 'w') as (_f):\n pass\n cdm.publishFromDir('testdir4', join(dirname(__file__), 'test.zip', 'dir1'))\n self.assertTrue(isfile(dstFilePath))\n finally:\n rmtree(dstDirPath)\n\n try:\n path = join('testdir5', 'somecontent.txt')\n cdm.publishContent(path, BytesIO(b'test'))\n dstFilePath = join(d.getRepositoryPath(), path)\n self.assertTrue(isfile(dstFilePath))\n finally:\n rmtree(join(d.getRepositoryPath(), dirname(path)))\n\n try:\n path = join('testdir6', 'somecontent2.txt')\n cdm.publishFromFile(path, BytesIO(b'test 2'))\n dstFilePath = join(d.getRepositoryPath(), path)\n self.assertTrue(isfile(dstFilePath))\n finally:\n rmtree(join(d.getRepositoryPath(), dirname(path)))\n\n def testLocalFileSystemLinkCDM(self):\n d = HTTPDelivery()\n rootDir = TemporaryDirectory()\n d.serverURI = 'http://localhost/content/'\n d.repositoryPath = rootDir.name\n cdm = LocalFileSystemLinkCDM()\n cdm.delivery = d\n try:\n exceptionRaised = False\n cdm.publishFromFile('a/../../b', 'somefile.txt')\n except PathNotFound:\n exceptionRaised = True\n\n self.assertTrue(exceptionRaised, 'No exception was raised on out of repository path')\n try:\n srcTmpFile = NamedTemporaryFile()\n dstFile = join('testdir7', 'tempfile.txt')\n cdm.publishFromFile(dstFile, srcTmpFile.name)\n dstLinkPath = join(d.getRepositoryPath(), dstFile + cdm._linkExt)\n self.assertTrue(isfile(dstLinkPath))\n with open(dstLinkPath) as (f):\n links = json.load(f)\n self.assertIsInstance(links, list)\n self.assertEqual(links[0][0], 'FS')\n self.assertEqual(srcTmpFile.name, links[0][1])\n self.assertEqual(datetime.fromtimestamp(stat(srcTmpFile.name).st_mtime), cdm.getTimestamp('testdir7/tempfile.txt'))\n finally:\n rmtree(dirname(dstLinkPath))\n\n try:\n dstFile = join('testdir8', 'tempfile2.txt')\n inFileName = join('dir1', 'subdir2', 'file1.txt')\n srcFilePath = join(dirname(__file__), 'test.zip', inFileName)\n cdm.publishFromFile(dstFile, srcFilePath)\n dstLinkPath = join(d.getRepositoryPath(), dstFile + cdm._linkExt)\n self.assertTrue(isfile(dstLinkPath))\n with open(dstLinkPath) as (f):\n links = json.load(f)\n self.assertEqual(links[0][0], 'ZIP')\n zipPath = links[0][1]\n inPath = normOSPath(links[0][2], True)\n linkPath = join(zipPath, inPath)\n self.assertEqual(normpath(linkPath), normpath(srcFilePath))\n self.assertEqual(datetime.fromtimestamp(stat(join(dirname(__file__), 'test.zip')).st_mtime), cdm.getTimestamp('testdir8/tempfile2.txt'))\n finally:\n rmtree(dirname(dstLinkPath))\n\n srcTmpDir = TemporaryDirectory()\n dirs = (join(srcTmpDir.name, 'test1/subdir1'), join(srcTmpDir.name, 'test2/subdir1'))\n for dir in dirs:\n makedirs(dir)\n with open(join(dir, 'text.html'), 'w+') as (_f):\n pass\n\n try:\n cdm.publishFromDir('testlink1', srcTmpDir.name)\n dstLinkPath = join(d.getRepositoryPath(), 'testlink1' + cdm._linkExt)\n self.assertTrue(isfile(dstLinkPath))\n with open(dstLinkPath) as (f):\n links = json.load(f)\n self.assertEqual(links[0][0], 'FS')\n self.assertEqual(srcTmpDir.name, links[0][1])\n self.assertEqual(datetime.fromtimestamp(stat(join(srcTmpDir.name, 'test1/subdir1/text.html')).st_mtime), cdm.getTimestamp('testlink1/test1/subdir1/text.html'))\n delPath1 = 'testlink1/test1/subdir1/text.html'\n cdm.remove(delPath1)\n self.assertTrue(isfile(join(d.getRepositoryPath(), delPath1 + '.deleted')))\n delPath2 = 'testlink1/test1'\n cdm.remove(delPath2)\n self.assertTrue(isfile(join(d.getRepositoryPath(), delPath2 + '.deleted')))\n finally:\n rmtree(join(d.getRepositoryPath(), 'testlink1'))\n remove(dstLinkPath)\n\n try:\n srcFilePath = join(dirname(__file__), 'test.zip', 'dir1') + sep\n cdm.publishFromFile('testlink2', srcFilePath)\n dstLinkPath = join(d.getRepositoryPath(), 'testlink2' + cdm._linkExt)\n self.assertTrue(isfile(dstLinkPath))\n with open(dstLinkPath) as (f):\n links = json.load(f)\n self.assertEqual(links[0][0], 'ZIP')\n zipPath = links[0][1]\n inPath = normOSPath(links[0][2], True)\n link = join(zipPath, inPath)\n self.assertEqual(link, srcFilePath)\n self.assertEqual(datetime.fromtimestamp(stat(join(dirname(__file__), 'test.zip')).st_mtime), cdm.getTimestamp('testlink2/subdir1/file1.txt'))\n delPath1 = 'testlink2/subdir1/file1.txt'\n cdm.remove(delPath1)\n self.assertTrue(isfile(join(d.getRepositoryPath(), delPath1 + '.deleted')))\n delPath2 = 'testlink2/subdir1/'\n self.assertTrue(isdir(join(d.getRepositoryPath(), delPath2)))\n cdm.remove(delPath2)\n self.assertTrue(isfile(join(d.getRepositoryPath(), delPath2.rstrip('/') + '.deleted')))\n self.assertFalse(isdir(join(d.getRepositoryPath(), delPath2)))\n self.assertFalse(isfile(join(d.getRepositoryPath(), delPath1 + '.deleted')))\n finally:\n rmtree(join(d.getRepositoryPath(), 'testlink2'))\n remove(dstLinkPath)\n\n\nif __name__ == '__main__':\n unittest.main()","sub_path":"pycfiles/ally_py-0.9.0-py3.2/local_filesystem.cpython-32.py","file_name":"local_filesystem.cpython-32.py","file_ext":"py","file_size_in_byte":10318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"601947139","text":"__author__ = 'jason.parent@carneylabs.com (Jason Parent)'\n\n# Django imports...\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.urlresolvers import reverse\nfrom django.shortcuts import redirect\nfrom django.shortcuts import render\n\n# Local imports...\nfrom .forms import ProfileForm\n\n\n@login_required\ndef profile_view(request):\n return render(request, 'accounts/profile.html')\n\n\n@login_required\ndef profile_edit_view(request):\n form = ProfileForm(instance=request.user)\n\n if request.method == 'POST':\n form = ProfileForm(instance=request.user, data=request.POST, files=request.FILES)\n\n if form.is_valid():\n form.save()\n\n return redirect(reverse('profile'))\n\n return render(request, 'accounts/profile_edit.html', {\n 'form': form\n })","sub_path":"accounts/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"489507881","text":"#!/usr/bin/env python\n\n\nimport rospy\nfrom tf_sim.msg import Float32Stamped\n\ndef step_gen():\n\n pub = rospy.Publisher('input_val', Float32Stamped, queue_size=10)\n\n rospy.init_node('step_sim')\n\n mg = rospy.get_param('~magnitude', 60)\n freq = rospy.get_param('~frequency', 10)\n\n\n rospy.loginfo('Simulating step of magnitude %d', mg)\n\n rate = rospy.Rate(freq)\n\n msg = Float32Stamped()\n\n while not rospy.is_shutdown():\n\n msg.data = mg\n msg.header.stamp = rospy.get_rostime()\n\n pub.publish(msg)\n\n rate.sleep()\n\nif __name__ == '__main__':\n\n global mg\n\n try:\n step_gen()\n except rospy.ROSInterruptException:\n pass\n","sub_path":"Izar Castorina - Entrega 2 PCSE/Parte 2/tf_sim/scripts/step_sim_node.py","file_name":"step_sim_node.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"503722136","text":"\"\"\"\nVNode顶点\nArcNode 弧\n邻接表\n\"\"\"\n\n\nclass ArcNode:\n def __init__(self, info, vnode=None, nextArcNode=None):\n self.vnode = vnode\n self.nextArcNode = nextArcNode\n self.info = info\n\n\nclass Vnode:\n def __init__(self, data: str, firstArc: ArcNode = None):\n self.data = data\n self.firstArc = firstArc\n\n\nclass AdjvexStruct:\n def __init__(self, vnodeLs: list, kind: int, vnum: int = None, arcnum: int = None):\n self.vnodeLs = vnodeLs\n self.vnum = vnum\n self.arcnum = arcnum\n self.kind = kind\n\n\n# 构建有向图邻接表\n# 逻辑结构如 directed_1.jpg 所示\ndef buildAdj():\n v1 = Vnode('v1')\n v2 = Vnode('v2')\n v3 = Vnode('v3')\n v4 = Vnode('v4')\n\n arc1 = ArcNode(\"arc1\")\n arc2 = ArcNode(\"arc2\")\n arc3 = ArcNode(\"arc3\")\n arc4 = ArcNode(\"arc4\")\n arc5 = ArcNode(\"arc5\")\n arc6 = ArcNode(\"arc6\")\n\n v1.firstArc = arc1\n arc1.vnode = v2\n arc1.nextArcNode = arc2\n arc2.vnode = v3\n\n v2.firstArc = arc3\n arc3.vnode = v3\n arc3.nextArcNode = arc4\n arc4.vnode = v4\n\n v3.firstArc = arc5\n arc5.vnode = v4\n\n v4.firstArc = arc6\n arc6.vnode = v3\n\n vnodeLs = [v1, v2, v3, v4]\n directGraph = AdjvexStruct(vnodeLs, 1)\n\n for vnode in vnodeLs:\n print(vnode.data + \" - > \", end='')\n arc = vnode.firstArc\n while arc is not None:\n print(arc.info + \" - > \", end='')\n arc = arc.nextArcNode\n print()\n\n\n# 构建有向图「逆邻接表」\n# 逻辑结构如 directed_1.jpg 所示\ndef buildUnAdj():\n v1 = Vnode('v1')\n v2 = Vnode('v2')\n v3 = Vnode('v3')\n v4 = Vnode('v4')\n\n arc1 = ArcNode(\"arc1\")\n arc2 = ArcNode(\"arc2\")\n arc3 = ArcNode(\"arc3\")\n arc4 = ArcNode(\"arc4\")\n arc5 = ArcNode(\"arc5\")\n arc6 = ArcNode(\"arc6\")\n\n v2.firstArc = arc1\n arc1.vnode = v1\n\n v3.firstArc = arc2\n arc2.vnode = v1\n arc2.nextArcNode = arc3\n arc3.vnode = v2\n arc3.nextArcNode = arc6\n arc6.vnode = v4\n\n v4.firstArc = arc4\n arc4.vnode = v2\n arc4.nextArcNode = arc5\n arc5.vnode = v3\n\n vnodeLs = [v1, v2, v3, v4]\n directGraph = AdjvexStruct(vnodeLs, 1)\n\n for vnode in vnodeLs:\n print(vnode.data + \" - > \", end='')\n arc = vnode.firstArc\n while arc is not None:\n print(arc.info + \" - > \", end='')\n arc = arc.nextArcNode\n print()\n\n\n# 构建无向图邻接表\n# 逻辑结构如 undirected_1.jpg 所示\ndef buildUndirect():\n v1 = Vnode('v1')\n v2 = Vnode('v2')\n v3 = Vnode('v3')\n v4 = Vnode('v4')\n v5 = Vnode('v5')\n\n arc12 = ArcNode(\"arc1: v1 -> v2\")\n arc21 = ArcNode(\"arc1: v2 -> v1\")\n arc13 = ArcNode(\"arc2: v1 -> v3\")\n arc31 = ArcNode(\"arc2: v3 -> v1\")\n arc15 = ArcNode(\"arc3: v1 -> v5\")\n arc51 = ArcNode(\"arc3: v5 -> v1\")\n arc25 = ArcNode(\"arc4: v2 -> v5\")\n arc52 = ArcNode(\"arc4: v5 -> v2\")\n arc24 = ArcNode(\"arc5: v2 -> v4\")\n arc42 = ArcNode(\"arc5: v4 -> v2\")\n arc34 = ArcNode(\"arc6: v3 -> v4\")\n arc43 = ArcNode(\"arc6: v4 -> v3\")\n arc35 = ArcNode(\"arc7: v3 -> v5\")\n arc53 = ArcNode(\"arc7: v5 -> v3\")\n\n v1.firstArc = arc12\n arc12.vnode = v2\n arc12.nextArcNode = arc13\n arc13.vnode = v3\n arc13.nextArcNode = arc15\n arc15.vnode = v5\n\n v2.firstArc = arc21\n arc21.vnode = v1\n arc21.nextArcNode = arc24\n arc24.vnode = v4\n arc24.nextArcNode = arc25\n arc25.vnode = v5\n\n v3.firstArc = arc31\n arc31.vnode = v1\n arc31.nextArcNode = arc34\n arc34.vnode = v4\n arc34.nextArcNode = arc35\n arc35.vnode = v5\n\n v4.firstArc = arc42\n arc42.vnode = v2\n arc42.nextArcNode = arc43\n arc43.vnode = v3\n\n v5.firstArc = arc51\n arc51.vnode = v1\n arc51.nextArcNode = arc52\n arc52.vnode = v2\n arc52.nextArcNode = arc53\n arc53.vnode = v3\n\n vls = [v1, v2, v3, v4, v5]\n undirectGraph = AdjvexStruct(vls, 2)\n\n for vn in vls:\n print(vn.data + \" \", end='')\n arc = vn.firstArc\n while arc is not None:\n print(\"(\" + arc.info + \") \", end='')\n arc = arc.nextArcNode\n print()\n\n\nif __name__ == \"__main__\":\n buildUnAdj()\n","sub_path":"graph/adjacency_list/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":4142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"88066323","text":"import numpy as np\n\nfrom .utils import *\nfrom .Functions import *\nfrom .IPMOptimizers import get_optimizer as get_ipm_opt\nfrom .UnconstrainedOptimizers import get_optimizer as get_unc_opt\n\nimport time\n\n\"\"\"Particle output will have the shape [[[]-particles]-timesteps] or paths[time_step][particle_idx]. Vals output has shape [[(obj, barrier, time)]-timestep]\"\"\"\n\n\n\n\ndef optimize(config, verbose=False):\n\n # init num_particles\n all_vals = []\n p_start = get_particles(config)\n p_curr = np.array([np.array(p) for p in p_start])\n p_num_particles = len(p_start)\n\n num_steps = config[\"num_path_steps\"]\n\n # get optimization method\n if config[\"optimization_type\"] == \"Unconstrained\":\n opt = get_unc_opt(config) \n elif config[\"optimization_type\"] == \"IPM\": \n opt = get_ipm_opt(config)\n\n if config[\"seed\"] is not None:\n np.random.seed(config[\"seed\"])\n\n for t in range(num_steps):\n print(t)\n\n p_curr, obj_barrier_vals = opt.update(p_curr, t, full_vals=True)\n p_curr = p_curr.copy()\n all_vals.append(obj_barrier_vals)\n \n return all_vals\n","sub_path":"new_adventure/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1118,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"338595480","text":"\nimport scapy.all as scapy\n\ndef scan(ip):\n arp_request=scapy.ARP(pdst=ip)\n # arp_request.show()\n broadcast = scapy.Ether(dst = \"ff:ff:ff:ff:ff:ff\")\n # broadcast.show()\n # scapy.ls(scapy.ARP())\n concadenate = broadcast/arp_request\n answered =scapy.srp(concadenate ,timeout=1,verbose=False)[0]\n # print(answered.summary())\n print(\"IP\\t\\t\\tMAC address\")\n for i in answered:\n print(i[1].psrc,i[1].hwsrc,sep=\"\\t\\t\")\n\n\n # scapy.arping(ip)\n\nscan(\"192.168.43.47/24\")\n\n\n\n\n\n\n\n\n\n\n","sub_path":"network scanning.py","file_name":"network scanning.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"189833374","text":"from yacs.config import CfgNode as CN\n\nC = CN()\n\nC.epochs = 50\nC.steps_per_epoch=None\nC.batch_size = 128\nC.n_features = 18\nC.n_samples = None\nC.w_optim='SGD'\nC.w_decay_order=2\nC.w_lr = 1e-3\nC.w_momentum=0.0\nC.w_weight_decay=5e-4\nC.a_optim=\"SGD\"\nC.a_decay_order=2\nC.a_lr = 1e-3\nC.a_momentum = 0.0\nC.a_weight_decay = 0.\nC.T = 2\nC.grad_clip = 100.\nC.grad_clip_bilevel=1000.\nC.logging_freq = 200\nC.w_checkpoint_freq = 1\nC.n_informative=7\nC.noise=0.25\nC.featurize_type=\"fourier\"\nC.initial_degree=1\nC.hvp=\"exact\"\nC.ihvp =\"exact\"\nC.inv_hess=\"exact\"\nC.normalize_a_lr=True\nC.w_warm_start=0\nC.log_grad_norm=True\nC.log_alphas=True\nC.alpha_weight_decay=0.\nC.grad_inner_loop_order=-1\nC.grad_outer_loop_order=-1\nC.arch_train_data=\"sotl\"\nC.model_type=\"vgg\"\nC.dataset=\"CIFAR\"\nC.device = 'cuda'\nC.train_arch=True\nC.dry_run=False\nC.mode=\"bilevel\"\nC.hessian_tracking=False\nC.smoke_test=True\nC.rand_seed = None\nC.w_scheduler=None\nC.a_scheduler=None\nC.features=None\nC.loss='ce'\nC.log_suffix = \"\"\nC.optimizer_mode = \"autograd\"\nC.bilevel_w_steps=None\nC.debug=False\nC.recurrent=True\nC.rand_seed=1\nC.adaptive_a_lr = False\nC.softplus_alpha_lr = False\nC.softplus_beta=100\nC.alpha_lr=1e-3\nC.arch_update_frequency=1\nC.loss_threshold=None\nC.progress_bar=True\nC.val_split=0.1\n\ndef get_cfg_defaults():\n \"\"\"Get a yacs CfgNode object with default values for my_project.\"\"\"\n # Return a clone so that the defaults will not be altered\n # This is for the \"local variable\" use pattern\n return C.clone()\n\ncfg = C","sub_path":"sotl_nas/configs/lr/mnist_vgg.py","file_name":"mnist_vgg.py","file_ext":"py","file_size_in_byte":1476,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"7288987","text":"from queue import Queue\nimport os\nimport random\nimport string\nfrom threading import Thread\nimport time\nimport itertools\nimport csv\nimport sys\nimport multiprocessing\n\n''' Shared variable among processes to signal program termination '''\nexit_flag = multiprocessing.Value('i', 0)\n''' Shared variable among processes to sync last row '''\nlast_line = multiprocessing.Value('i', 0)\n\nclass MMChallengeProcessing:\n def __init__(self):\n self.jobqueue = multiprocessing.Queue()\n self.outfile = 'source.csv'\n self.outsize = 10 # MB\n self.file = open(self.outfile, 'w')\n \n self.start = 0\n self.end = 5000\n self.offset = 5000\n self.lock = multiprocessing.Lock()\n\n ''' Function to generate row specified in document, returns string with newline '''\n def generateRow(self, count):\n string1 = ''.join(random.choices(string.ascii_lowercase, k=random.randint(1,32)))\n string2 = ''.join(random.choices(string.ascii_lowercase, k=random.randint(1,32)))\n data = [str(count), str(random.randint(1,10)), string1, string2]\n return ','.join(data) + '\\n'\n\n\n ''' Generates and writes data for range of primary keys given. For example, if 2000, 4000 are passed as arguments, \n function will generate 2000 rows starting with id 2000 and ending at 4000. \n '''\n def calculateAndWrite(self, start, end):\n\n ''' Generate Data '''\n temp = ''\n for index in range(start, end):\n temp += (self.generateRow(index))\n\n ''' Loop until last line meets global start variable. Write and return '''\n while True:\n if exit_flag.value == 1:\n break\n if last_line.value == start:\n #print('writing {} : {}'.format(start, end))\n self.file.write(temp)\n last_line.value = end\n return\n\n ''' Pops jobs off queue to process ''' \n def processJobs(self, jobs):\n while True:\n window = self.jobqueue.get()\n start, end = window\n self.calculateAndWrite(start, end)\n if exit_flag.value == 1:\n break\n\n\n ''' Worker that will continuously add chunks of primary id's to calculate until file size is reached '''\n def loadJobs(self, lock):\n while (os.path.getsize(self.outfile)//1024**2) < self.outsize:\n self.jobqueue.put([self.start, self.end])\n self.start = self.end \n self.end += self.offset\n\n \n print(\"--- %s seconds ---\" % (time.time() - start_time))\n\n print(\"Cleaning up queue to exit\")\n with multiprocessing.Lock():\n exit_flag.value += 1\n while not self.jobqueue.empty():\n self.jobqueue.get()\n\n\n ''' Main function '''\n def run(self):\n\n ''' Write Header Row '''\n self.file.write('id,integer1,string1,string2\\n') \n ''' Flush to avoid rewrite of header '''\n self.file.flush()\n \n ''' Start pushing jobs to queue '''\n t = multiprocessing.Process(target=self.loadJobs, args=(self.lock, ))\n t.start()\n \n ''' Process jobs and write to file '''\n numProcesses = 100\n for work in range(0, numProcesses):\n worker = multiprocessing.Process(target=self.processJobs, args=(self.jobqueue,))\n worker.start()\n \n \nif __name__ == '__main__':\n print(\"Generating File...\")\n start_time = time.time()\n\n mm = MMChallengeProcessing()\n mm.run()\n \n\n \n \n","sub_path":"gen_multiprocessing.py","file_name":"gen_multiprocessing.py","file_ext":"py","file_size_in_byte":3670,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"370576106","text":"import ROOT as r \nimport numpy as np\nimport pickle,math,os\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout\n\nimport tensorflow as tf\nfrom keras.backend import tensorflow_backend as K\n\nfrom keras import optimizers\n\n\ndef getCompiledModelA(nvars,nnodes):\n # optimal so far ( 0.980, 0.966)\n model = Sequential()\n model.add(Dense(30,input_dim=nvars, activation='relu'))\n model.add(Dense(10, activation='relu'))\n model.add(Dense(nnodes, activation='softmax'))\n adam = optimizers.adam(lr=1e-4) \n model.compile(loss='categorical_crossentropy', optimizer=adam, metrics=['accuracy','categorical_crossentropy'])\n return model\n\ndef getCompiledModelB(nvars,nnodes):\n model = Sequential()\n model.add(Dense(30,input_dim=nvars, activation='relu'))\n model.add(Dropout(0.4))\n model.add(Dense(10, activation='relu'))\n model.add(Dropout(0.4))\n model.add(Dense(10, activation='relu'))\n model.add(Dense(nnodes, activation='softmax'))\n adam = optimizers.adam(lr=1e-4) \n model.compile(loss='categorical_crossentropy', optimizer=adam, metrics=['accuracy','categorical_crossentropy'])\n return model\n\ndef freeze_session(session, keep_var_names=None, output_names=None, clear_devices=True):\n \"\"\"\n Freezes the state of a session into a pruned computation graph.\n\n Creates a new computation graph where variable nodes are replaced by\n constants taking their current value in the session. The new graph will be\n pruned so subgraphs that are not necessary to compute the requested\n outputs are removed.\n @param session The TensorFlow session to be frozen.\n @param keep_var_names A list of variable names that should not be frozen,\n or None to freeze all the variables in the graph.\n @param output_names Names of the relevant graph outputs.\n @param clear_devices Remove the device directives from the graph for better portability.\n @return The frozen graph definition.\n \"\"\"\n from tensorflow.python.framework.graph_util import convert_variables_to_constants\n graph = session.graph\n with graph.as_default():\n freeze_var_names = list(set(v.op.name for v in tf.global_variables()).difference(keep_var_names or []))\n output_names = output_names or []\n output_names += [v.op.name for v in tf.global_variables()]\n # Graph -> GraphDef ProtoBuf\n input_graph_def = graph.as_graph_def()\n if clear_devices:\n for node in input_graph_def.node:\n node.device = \"\"\n frozen_graph = convert_variables_to_constants(session, input_graph_def,\n output_names, freeze_var_names)\n return frozen_graph\n\ndef saveTF1p6Model(protobuffer_name):\n frozen_graph = freeze_session(K.get_session(),\n output_names=[out.op.name for out in model.outputs])\n tf.train.write_graph(frozen_graph, os.getcwd(), protobuffer_name, as_text=False)\n print (\"Saved frozen model in \",protobuffer_name)\n \n \nif __name__ == \"__main__\":\n\n from optparse import OptionParser\n parser = OptionParser(usage=\"%prog [options]\")\n parser.add_option(\"-i\", \"--infile\", dest=\"infile\", type=\"string\", default=\"vars.pkl\", help=\"Input pickle file (default: vars.pkl)\");\n parser.add_option(\"-o\", \"--outfile\", dest=\"outfile\", type=\"string\", default=\"trained_model\", help=\"Output pb and h5 fil;e (default: trained_model)\");\n parser.add_option(\"-c\", \"--channel\", dest=\"channel\", type=\"string\", default=\"2lss\", help=\"Final state: 2lss or 3l (default: 2lss)\");\n (options, args) = parser.parse_args()\n\n nvars = 29 if options.channel=='2lss' else 32 if options.channel=='3l' else 35\n\n data = pickle.load( open(options.infile,'rb'))\n sums = np.sum(data['train_y'],axis=0)\n print(sums)\n\n sig = sums[0]\n bkg = sums[1] + sums[2]\n #if options.channel=='2lss': bkg += sums[3]\n\n class_weight = { 0 : float((sig+bkg)/sig),\n 1 : float((sig+bkg)/bkg),\n 2 : float((sig+bkg)/bkg)}\n # if options.channel=='2lss':\n # class_weight.update( {3 : float((sig+bkg)/bkg)} )\n\n print ('weights will be', class_weight)\n\n with tf.Session(config=tf.ConfigProto(\n intra_op_parallelism_threads=50,\n inter_op_parallelism_threads=50)) as sess:\n K.set_session(sess)\n \n #nnodes = 4 if options.channel=='2lss' else 3\n nnodes = 3\n model = getCompiledModelA(nvars,nnodes)\n #model = getCompiledModelB(nvars,nnodes)\n\n history = model.fit( data['train_x'], data['train_y'], epochs=20, batch_size=80, validation_data=(data['test_x'], data['test_y']), class_weight=class_weight)\n\n modelname = os.getcwd()+'/'+os.path.basename(options.outfile).split('.')[0]\n # keras model (H5)\n model.save(modelname+'.h5')\n # tf model (PB)\n saveTF1p6Model(modelname+'.pb')\n \n pickle_out = open(modelname+'.pkl','wb')\n pickle.dump( history.history, pickle_out)\n pickle_out.close()\n \n","sub_path":"TTHAnalysis/python/tools/bdts/trainNet.py","file_name":"trainNet.py","file_ext":"py","file_size_in_byte":5086,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"634591751","text":"# Name: Joshua Fogus\n# Last Modified: February 11, 2021\n# Description: A text box for accepting free-form user input.\n\nimport tkinter as tk\n\n\ndef text_input(parent, label, variable):\n \"\"\" creates a text input attached to the given parent, with\n the given label, and storing the result in the given\n variable. Returns a tuple of the label and input. \"\"\"\n tk_label = tk.Label(parent, text=\"{}: \".format(label))\n tk_label.grid(row=2, column=0, sticky=tk.W, pady=(0, 4))\n\n tk_input = tk.Entry(parent, textvariable=variable)\n tk_input.grid(row=2, column=1, sticky=tk.W, pady=(0, 4))\n\n return tk_label, tk_input\n\n\nif __name__ == \"__main__\":\n print(\"This is not meant to be run as a script. Please import this module.\")","sub_path":"life_generator/gui/text_input.py","file_name":"text_input.py","file_ext":"py","file_size_in_byte":751,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"189073112","text":"#!/usr/bin/env python\r\n#coding:utf-8\r\nimport pandas as pd\r\nfrom requests_html import HTMLSession\r\n\r\nsession = HTMLSession()\r\nurl = \"https://www.jianshu.com/p/85f4624485b9\"\r\nr = session.get(url)\r\n\r\nsel = \"body > div.note > div.post > div.article > div.show-content > div > p > a\"\r\n\r\ndef get_text_link_from_sel(sel):\r\n mylist = []\r\n try:\r\n results = r.html.find(sel)\r\n for result in results:\r\n mytext = result.text\r\n mylink = list(result.absolute_links)[0]\r\n mylist.append((mytext,mylink))\r\n return mylist\r\n except:\r\n return None\r\n\r\ndf = pd.DataFrame(get_text_link_from_sel(sel))\r\ndf.columns = [\"text\",\"link\"]\r\ndf.to_csv(\"output.csv\",encoding=\"gbk\",index=False)\r\n\r\n\r\n#print(r.html.text)\r\n#print(r.html.links)\r\n#print(r.html.absolute_links)\r\n# print(results[0].text)\r\n# print(results[0].absolute_links)\r\n# print(\r\n# list(results[0].absolute_links)\r\n# )\r\n#","sub_path":"Reptile.py","file_name":"Reptile.py","file_ext":"py","file_size_in_byte":926,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"529994012","text":"data = [{\"name\": \"CR90 corvette\", \"class\": \"corvette\", \"cost\": \"3500000\", \"rating\": \n 5.7147, \"drive\": \"2.0\"}, \n{\"name\": \"Star Destroyer\", \"class\": \n \"Star Destroyer\", \"cost\": \"150000000\", \"rating\": 1.3338, \n \"drive\": \"2.0\"}, \n{\"name\": \"Sentinel-class landing craft\", \"class\": \"landing craft\", \"cost\": \"240000\", \n\"rating\": 4.166, \"drive\": \"1.0\"}, \n{\"name\": \"Death Star\", \"class\": \"Deep Space Mobile Battlestation\", \"cost\": \n \"1000000000000\", \"rating\": 4, \"drive\": \"4.0\"}, \n{\"name\": \"Millennium Falcon\", \"class\": \"Light freighter\", \"cost\": \"100000\", \"rating\": 5, \n \"drive\": \"0.5\"}, \n{\"name\": \"Y-wing\", \"class\": \"assault starfighter\", \"cost\": \n \"134999\", \"rating\": 7.407, \"drive\": \"1.0\"}, \n{\"name\": \"X-wing\", \"class\": \"Starfighter\", \"cost\": \"149999\", \"rating\": \n 6.666, \"drive\": \"1.0\"}, \n{\"name\": \"TIE Advanced x1\", \"class\": \"Starfighter\", \"cost\": \"unknown\", \"rating\": 0, \"drive\": \n \"1.0\"}, \n{\"name\": \"Executor\", \"class\": \"Star dreadnought\", \"cost\": \n \"1143350000\", \"rating\": 1.749, \"drive\": \"2.0\"}, \n{\"name\": \"Rebel transport\", \"class\": \"Medium transport\", \"cost\": \"unknown\", \"rating\": \n 0, \"drive\": \"4.0\"}, \n{\"name\": \"Slave 1\", \"class\": \"Patrol craft\", \"cost\": \"unknown\", \"rating\": 0, \"drive\": \"3.0\"}, \n{\"name\": \"Imperial shuttle\", \"class\": \"Armed government transport\", \"cost\": \"240000\", \"rating\": \n 4.1666, \"drive\": \"1.0\"}, \n{\"name\": \"EF76 Nebulon-B escort frigate\", \"class\": \"Escort ship\", \"cost\": \"8500000\", \"rating\": \n 2.352, \"drive\": \"2.0\"}, \n{\"name\": \"Calamari Cruiser\", \"class\": \"Star Cruiser\", \"cost\": \"104000000\", \"rating\": \n 9.615, \"drive\": \"1.0\"}, \n{\"name\": \"A-wing\", \"class\": \"Starfighter\", \"cost\": \"175000\", \"rating\": 5.714, \n \"drive\": \"1.0\"}, \n{\"name\": \"B-wing\", \"class\": \"Assault Starfighter\", \"cost\": \"220000\", \"rating\": \n 9.06, \"drive\": \"2.0\"}, {\"name\": \"Republic Cruiser\", \"class\": \"Space cruiser\", \n \"cost\": \"unknown\", \"rating\": 0, \"drive\": \"2.0\"}, \n{\"name\": \"Droid control ship\", \"class\": \"Droid control ship\", \"cost\": \"unknown\", \n \"rating\": 0, \"drive\": \"2.0\"}, {\"name\": \"Naboo fighter\", \"class\": \"Starfighter\", \n \"cost\": \"200000\", \"rating\": 5, \"drive\": \"1.0\"}, \n{\"name\": \"Naboo Royal Starship\", \"class\": \"yacht\", \"cost\": \"unknown\", \"rating\": 0, \"drive\": \"1.8\"}, {\"name\": \"Scimitar\", \"class\": \n \"Space Transport\", \"cost\": \"55000000\", \"rating\": 2.727, \n \"drive\": \"1.5\"}, \n{\"name\": \"J-type diplomatic barge\", \"class\": \"Diplomatic barge\", \"cost\": \"2000000\", \"rating\": 3.5, \"drive\": \"0.7\"}, \n{\"name\": \"AA-9 Coruscant freighter\", \"class\": \"freighter\", \"cost\": \"unknown\", \n \"rating\": 0, \"drive\": \"unknown\"}, \n{\"name\": \"Jedi starfighter\", \"class\": \"Starfighter\", \"cost\": \"180000\", \n \"rating\": 5.55, \"drive\": \"1.0\"}, \n{\"name\": \"H-type Nubian yacht\", \"class\": \"yacht\", \"cost\": \"unknown\", \n \"rating\": 0, \"drive\": \"0.9\"}, \n{\"name\": \"Republic Assault ship\", \"class\": \"assault ship\", \"cost\": \"unknown\", \"rating\": 0, \"drive\": \n \"0.6\"}, \n{\"name\": \"Solar Sailer\", \"class\": \"yacht\", \"cost\": \"35700\", \"rating\": 4.201, \n \"drive\": \"1.5\"}, {\"name\": \"Trade Federation cruiser\", \"class\": \"capital ship\", \n \"cost\": \"125000000\", \"rating\": 1.2, \"drive\": \"1.5\"}, \n{\"name\": \"Theta-class T-2c shuttle\", \"class\": \"transport\", \n \"cost\": \"1000000\", \"rating\": 1, \"drive\": \"1.0\"}, \n{\"name\": \"Republic attack cruiser\", \"class\": \"star destroyer\", \"cost\": \"59000000\", \"rating\": \n 1.69, \"drive\": \"1.0\"}, \n{\"name\": \"Naboo star skiff\", \"class\": \"yacht\", \"cost\": \"unknown\", \"rating\": 0, \"drive\": \"0.5\"}, \n{\"name\": \"Jedi Interceptor\", \"class\": \"starfighter\", \"cost\": \"320000\", \n \"rating\": 3.125, \"drive\": \"1.0\"}, \n{\"name\": \"arc-170\", \"class\": \"starfighter\", \"cost\": \"155000\", \"rating\": 6.451, \"drive\": \n \"1.0\"}, \n{\"name\": \"Banking clan frigte\", \"class\": \"cruiser\", \"cost\": \"57000000\",\n \"rating\": 1.754, \"drive\": \"1.0\"}, \n{\"name\": \"Belbullab-22 starfighter\", \"class\": \"starfighter\", \"cost\": \"168000\", \n \"rating\": 3.571, \"drive\": \"6\"}, \n{\"name\": \"V-wing\", \"class\": \"starfighter\", \"cost\": \"102500\", \"rating\": 9.756, \n \"drive\": \"1.0\"}]","sub_path":"app/tests/test_data_ships.py","file_name":"test_data_ships.py","file_ext":"py","file_size_in_byte":4062,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"419570712","text":"ok = True\nfor subtask_path, return_value in task.run(\".subtask1/\", output_plugin=\"null\").items():\n if \"01_tasklet_1\" not in subtask_path:\n ok = False\n if return_value != \"OK\":\n ok = False\n\nif ok:\n return \"OK\"\nelse:\n return \"NOT OK\"\n","sub_path":"tests/plugins/content/task/01_run.py","file_name":"01_run.py","file_ext":"py","file_size_in_byte":258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"225883877","text":"# gpl author: Ryan Inch (Imaginer)\n\nimport bpy\nfrom bpy.types import Menu\nfrom . import utils_core\nfrom bl_ui.properties_paint_common import UnifiedPaintPanel\n\n# Particle Tools\n\nparticle_tools = (\n (\"None\", 'NONE'),\n (\"Comb\", 'COMB'),\n (\"Smooth\", 'SMOOTH'),\n (\"Add\", 'ADD'),\n (\"Length\", 'LENGTH'),\n (\"Puff\", 'PUFF'),\n (\"Cut\", 'CUT'),\n (\"Weight\", 'WEIGHT')\n)\n\n# Brush Datapaths\n\nbrush_datapath = {\n 'SCULPT': \"tool_settings.sculpt.brush\",\n 'VERTEX_PAINT': \"tool_settings.vertex_paint.brush\",\n 'WEIGHT_PAINT': \"tool_settings.weight_paint.brush\",\n 'TEXTURE_PAINT': \"tool_settings.image_paint.brush\",\n 'PARTICLE_EDIT': \"tool_settings.particle_edit.tool\"\n}\n\n# Brush Icons\n\nbrush_icon = {\n 'SCULPT': {\n \"BLOB\": 'BRUSH_BLOB',\n \"CLAY\": 'BRUSH_CLAY',\n \"CLAY_STRIPS\": 'BRUSH_CLAY_STRIPS',\n \"CREASE\": 'BRUSH_CREASE',\n \"DRAW\": 'BRUSH_SCULPT_DRAW',\n \"FILL\": 'BRUSH_FILL',\n \"FLATTEN\": 'BRUSH_FLATTEN',\n \"GRAB\": 'BRUSH_GRAB',\n \"INFLATE\": 'BRUSH_INFLATE',\n \"LAYER\": 'BRUSH_LAYER',\n \"MASK\": 'BRUSH_MASK',\n \"NUDGE\": 'BRUSH_NUDGE',\n \"PINCH\": 'BRUSH_PINCH',\n \"ROTATE\": 'BRUSH_ROTATE',\n \"SCRAPE\": 'BRUSH_SCRAPE',\n \"SIMPLIFY\": 'BRUSH_SUBTRACT',\n \"SMOOTH\": 'BRUSH_SMOOTH',\n \"SNAKE_HOOK\": 'BRUSH_SNAKE_HOOK',\n \"THUMB\": 'BRUSH_THUMB'\n },\n\n 'VERTEX_PAINT': {\n \"ADD\": 'BRUSH_ADD',\n \"BLUR\": 'BRUSH_BLUR',\n \"DARKEN\": 'BRUSH_DARKEN',\n \"LIGHTEN\": 'BRUSH_LIGHTEN',\n \"MIX\": 'BRUSH_MIX',\n \"MUL\": 'BRUSH_MULTIPLY',\n \"SUB\": 'BRUSH_SUBTRACT'\n },\n\n 'WEIGHT_PAINT': {\n \"ADD\": 'BRUSH_ADD',\n \"BLUR\": 'BRUSH_BLUR',\n \"DARKEN\": 'BRUSH_DARKEN',\n \"LIGHTEN\": 'BRUSH_LIGHTEN',\n \"MIX\": 'BRUSH_MIX',\n \"MUL\": 'BRUSH_MULTIPLY',\n \"SUB\": 'BRUSH_SUBTRACT'\n },\n\n 'TEXTURE_PAINT': {\n \"CLONE\": 'BRUSH_CLONE',\n \"DRAW\": 'BRUSH_TEXDRAW',\n \"FILL\": 'BRUSH_TEXFILL',\n \"MASK\": 'BRUSH_TEXMASK',\n \"SMEAR\": 'BRUSH_SMEAR',\n \"SOFTEN\": 'BRUSH_SOFTEN'\n }\n}\n\n\nclass BrushesMenu(Menu):\n bl_label = \"Brush\"\n bl_idname = \"VIEW3D_MT_sv3_brushes_menu\"\n\n def draw(self, context):\n mode = utils_core.get_mode()\n layout = self.layout\n settings = UnifiedPaintPanel.paint_settings(context)\n colum_n = utils_core.addon_settings(lists=False)\n\n layout.row().label(text=\"Brush\")\n layout.row().separator()\n\n has_brush = utils_core.get_brush_link(context, types=\"brush\")\n current_brush = eval(\"bpy.context.{}\".format(brush_datapath[mode])) if has_brush else None\n\n # get the current brush's name\n if current_brush and utils_core.get_mode() != 'PARTICLE_EDIT':\n current_brush = current_brush.name\n\n if mode == 'PARTICLE_EDIT':\n # if you are in particle edit mode add the menu items for particle mode\n for tool in particle_tools:\n utils_core.menuprop(\n layout.row(), tool[0], tool[1], brush_datapath[mode],\n icon='RADIOBUT_OFF', disable=True,\n disable_icon='RADIOBUT_ON'\n )\n else:\n column_flow = layout.column_flow(columns=colum_n)\n if utils_core.addon_settings(lists=True) == 'template':\n layout.template_ID_preview(settings, \"brush\",\n new=\"brush.add\", rows=3, cols=colum_n)\n else:\n # iterate over all the brushes\n for item in bpy.data.brushes:\n if mode == 'SCULPT':\n if item.use_paint_sculpt:\n # if you are in sculpt mode and the brush\n # is a sculpt brush add the brush to the menu\n utils_core.menuprop(\n column_flow.row(), item.name,\n 'bpy.data.brushes[\"%s\"]' % item.name,\n brush_datapath[mode], icon=brush_icon[mode][item.sculpt_tool],\n disable=True, custom_disable_exp=(item.name, current_brush),\n path=True\n )\n if mode == 'VERTEX_PAINT':\n if item.use_paint_vertex:\n # if you are in vertex paint mode and the brush\n # is a vertex paint brush add the brush to the menu\n utils_core.menuprop(\n column_flow.row(), item.name,\n 'bpy.data.brushes[\"%s\"]' % item.name,\n brush_datapath[mode], icon=brush_icon[mode][item.vertex_tool],\n disable=True, custom_disable_exp=(item.name, current_brush),\n path=True\n )\n if mode == 'WEIGHT_PAINT':\n if item.use_paint_weight:\n # if you are in weight paint mode and the brush\n # is a weight paint brush add the brush to the menu\n utils_core.menuprop(\n column_flow.row(), item.name,\n 'bpy.data.brushes[\"%s\"]' % item.name,\n brush_datapath[mode], icon=brush_icon[mode][item.vertex_tool],\n disable=True, custom_disable_exp=(item.name, current_brush),\n path=True\n )\n if utils_core.get_mode() == 'TEXTURE_PAINT':\n if item.use_paint_image:\n # if you are in texture paint mode and the brush\n # is a texture paint brush add the brush to the menu\n utils_core.menuprop(\n column_flow.row(), item.name,\n 'bpy.data.brushes[\"%s\"]' % item.name,\n brush_datapath[mode], icon=brush_icon[mode][item.image_tool],\n disable=True, custom_disable_exp=(item.name, current_brush),\n path=True\n )\n","sub_path":"engine/2.80/scripts/addons/space_view3d_brush_menus/brushes.py","file_name":"brushes.py","file_ext":"py","file_size_in_byte":6376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"545860208","text":"import re\ndigit_match = re.compile(r\"\\d\")\nletter_match = re.compile(r\"\\w\")\n\ndef hey(text):\n if text.strip() == \"\":\n return \"Fine. Be that way!\"\n\n no_digits = digit_match.sub(\" \", text)\n has_letters = letter_match.search(no_digits) is not None\n\n if has_letters and text.upper() == text:\n return \"Whoa, chill out!\"\n\n if text.endswith(\"?\"):\n return \"Sure.\"\n\n return \"Whatever.\"\n","sub_path":"all_data/exercism_data/python/bob/710528d7813b41359dcb809944b426ad.py","file_name":"710528d7813b41359dcb809944b426ad.py","file_ext":"py","file_size_in_byte":414,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"638855815","text":"\nfrom gui_design import gui\nimport sys\nimport os\nimport matplotlib.pyplot as plt\nimport numpy as np\n\nfrom wave import plot_tmg\nfrom wave import plot_wave\nfrom wave import plot_bunch_moment\nfrom wave import plot_fft\nfrom wave import plot_all_mom\nfrom process import plot_process\nfrom process import plot_process_mom\nfrom process import process_13_15\nfrom process import betatron_13_15\nfrom process import cal_process_mom\nfrom process import quad_xy\nfrom noise import noise_analysis\nfrom extract import extract_data\nfrom extract import plot_pos\nfrom extract import adiabatic_dump\nfrom extract import betatron_ext\nfrom reflectance import reflectance\nfrom reflectance import process_ref\n##from BBA import BBA_cal\nfrom osc_quad import quad_osc\nfrom osc_quad import quad_osc_1212\nfrom zero_moment import zero\nfrom mabiki import mabiki_em\nfrom kiwamu import kiwamu_em\nfrom cmap import plot_cmap\n\nclass Application(gui.Design,object):\n def __init__(self):\n super(Application,self).__init__()\n self.select.configure(command=self._select)\n\n def _select(self):\n self.mode_num = self.num_mode.get()\n ### Home\n if self.mode_num == 16:\n self.reflesh()\n return\n ### BBA\n elif self.mode_num == 10:\n BBA_cal.extract_cal_vol()\n return\n ### mabiki\n elif self.mode_num == 14:\n mabiki_em.emittance()\n return\n ### mabiki\n elif self.mode_num == 15:\n quad_osc.emittance()\n return\n elif self.mode_num == 12:\n quad_osc_1212.emittance()\n return\n file_index = self.file_list.curselection()\n file_name = self.data_path+self.file_list.get(file_index[0]).replace(' ','')\n\n if not file_index: return\n ### extract\n elif self.mode_num == 5:\n extract_data.extract_vol(file_name)\n ### \n elif os.path.isdir(file_name):\n self.reflesh(path=file_name)\n else : self.act_mode(file_name=file_name)\n \n def act_mode(self,file_name):\n\n self.mode_num = self.num_mode.get()\n ### beam data\n if self.mode_num == 0:\n if 'wave_2' in file_name:\n self.memo(file_name)\n vol,mon = plot_wave.plot_wave(file_name)\n plot_bunch_moment.plot_bunch_moment(file_name,vol,mon)\n plot_fft.plot_fft(file_name,vol)\n elif 'process_2' in file_name:\n self.memo(file_name)\n vol = plot_process.plot_process(file_name)\n plot_process_mom.plot_process_mom(file_name,vol)\n ### adjust timing\n elif self.mode_num == 1:\n print(file_name)\n if 'wave_2' in file_name:\n self.memo(file_name)\n plot_tmg.plot_timing(file_name)\n ### noise\n elif self.mode_num == 2:\n if 'wave_2' in file_name:\n self.memo(file_name)\n noise_analysis.wave_noise(file_name)\n elif 'process_2' in file_name:\n self.memo(file_name)\n noise_analysis.process_noise(file_name)\n ### reflectance\n elif self.mode_num == 9:\n if 'wave_2' in file_name:\n self.memo(file_name)\n reflectance.plot_relation_reflectance(file_name)\n elif 'process_2' in file_name:\n self.memo(file_name)\n process_ref.plot_process_ref(file_name)\n ### #13 and #15\n elif self.mode_num == 3:\n if 'wave_2' in file_name:\n self.memo(file_name)\n #noise_analysis.wave_noise(file_name)\n elif 'process_2' in file_name:\n self.memo(file_name)\n process_13_15.plot_13_15(file_name)\n ### betatron osc. #13 and #15\n elif self.mode_num == 4:\n if 'wave_2' in file_name:\n self.memo(file_name)\n #noise_analysis.wave_noise(file_name)\n elif 'process_2' in file_name:\n self.memo(file_name)\n betatron_13_15.betatron(file_name)\n elif '.txt' in file_name:\n betatron_ext.plot_twiss(file_name)\n ### plot extract data\n elif self.mode_num == 6:\n if 'vol' in file_name:\n plot_pos.plot_row(file_name)\n\n ### plot extract data\n elif self.mode_num == 7:\n if 'vol' in file_name:\n adiabatic_dump.plot_dump(file_name)\n ### quad\n elif self.mode_num == 8:\n if 'process' in file_name:\n quad_xy.quad(file_name)\n ### quad\n\n elif self.mode_num == 11:\n if 'process' in file_name:\n zero.power(file_name)\n elif self.mode_num == 17:\n if 'wave' in file_name:\n plot_cmap.plot(file_name)\n\n elif self.mode_num == 13:\n if 'wave' in file_name:\n kiwamu_em.emittance(file_name)\n return\n\n \n\nif __name__==\"__main__\":\n app = Application()\n app.mainloop()\n \n \n","sub_path":"analysis/quick_analysis.py","file_name":"quick_analysis.py","file_ext":"py","file_size_in_byte":5146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"615128564","text":"# -*- coding: utf-8 -*-\n# Copyright 2018 ICON Foundation\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom unittest import TestCase, main\n\nfrom iconsdk.exception import AddressException, JSONRPCException\nfrom iconsdk.icon_service import IconService\nfrom iconsdk.providers.http_provider import HTTPProvider\nfrom iconsdk.utils.validation import is_score_apis\nfrom tests.example_config import BASE_DOMAIN_URL_V3_FOR_TEST, VERSION_FOR_TEST\n\n\nclass TestGetScoreApi(TestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.icon_service = IconService(HTTPProvider(BASE_DOMAIN_URL_V3_FOR_TEST, VERSION_FOR_TEST))\n # Because governance always has score apis, it is proper for the test.\n cls.governance_address = \"cx0000000000000000000000000000000000000001\"\n\n def test_get_score_api(self):\n # case 0: when getting score apis successfully\n result = self.icon_service.get_score_api(self.governance_address)\n self.assertTrue(result)\n # case 1: when address is wrong - wallet address\n self.assertRaises(AddressException, self.icon_service.get_score_api,\n \"hx882efc17c2f50e0d60142b9c0e746cbafb569d8c\")\n # case 2: when address is wrong - too short\n self.assertRaises(AddressException, self.icon_service.get_score_api,\n \"cx882efc17c2f50e0d60142b9c0e746cbafb\")\n # case 3: when the address is not score id\n self.assertRaises(JSONRPCException, self.icon_service.get_score_api,\n \"cxb0776ee37f5b45bfaea8cff1d8232fbb6122ec32\")\n\n def test_validate_score_apis(self):\n result = self.icon_service.get_score_api(self.governance_address)\n self.assertTrue(result)\n self.assertTrue(is_score_apis(result))\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"tests/api_get/test_get_score_api.py","file_name":"test_get_score_api.py","file_ext":"py","file_size_in_byte":2300,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"362053960","text":"# now package it all into a function\n\nimport numpy as np\nfrom skimage.feature import peak_local_max\n\ndef peak_local_max_3d(image,min_distance,threshold=0):\n\n \"\"\"Find peaks in an 3D image as intensity array.\n Peaks are the local maxima in a region of `2 * min_distance + 1`\n (i.e. peaks are separated by at least `min_distance`).\n If there are multiple local maxima with identical pixel intensities\n inside the region defined with `min_distance`,\n the coordinates of all such pixels are returned.\n\n ----------\n image : ndarray\n Input image.\n min_distance : int\n Minimum number of pixels separating peaks in a region of `2 *\n min_distance + 1` (i.e. peaks are separated by at least\n `min_distance`).\n To find the maximum number of peaks, use `min_distance=1`.\n\n Returns\n -------\n output : ndarray\n * If `indices = True` : [[x1,y1,z1],[x2,y2,z2],...] coordinates of peaks.\n\n Notes\n -----\n The function relies on applying scikit image's 2D peak_local_max function\n to generate a candidate list of 3D maxima which then get elliminated in a\n subsequent step to fulfill the min_distance criterion.\n\n\n \"\"\"\n\n ######### setup\n # make an array of zeros\n accumulator = np.zeros(image.shape)\n\n # accumulator for the coordinates and the coordinate intensities\n coordinateAccumulator = []\n\n ######### 2D\n # find all maxima in every 2D slice of the image\n for iz in range(0,image.shape[0]):\n\n coordinates=peak_local_max(image[iz],min_distance=min_distance)\n\n #write the max values into the accumulator at the right positions\n for coord in coordinates:\n coordValue = image[iz][coord[0],coord[1]]\n accumulator[iz,coord[0],coord[1]] = coordValue\n coordinateAccumulator.append([np.array([iz,coord[0],coord[1]]),coordValue])\n\n\n\n\n ######### 3D\n # Elliminate all that are too close together\n\n\n for maxCandidate in coordinateAccumulator:\n maxCandidate_z = maxCandidate[0][0]\n maxCandidate_x = maxCandidate[0][1]\n maxCandidate_y = maxCandidate[0][2]\n\n maxCandidate_value = maxCandidate[1]\n windowSizeHalf = int(min_distance/2)\n #print(windowSizeHalf)\n\n# print(maxCandidate_x-windowSizeHalf)\n# prnt(windowSizeHalf)\n from_x = max(0,maxCandidate_x-windowSizeHalf)\n to_x = min(image.shape[0],maxCandidate_x+windowSizeHalf)\n\n from_y = max(0,maxCandidate_y-windowSizeHalf)\n to_y = min(image.shape[1],maxCandidate_y+windowSizeHalf)\n\n from_z = max(0,maxCandidate_z-windowSizeHalf)\n to_z = min(image.shape[2],maxCandidate_z+windowSizeHalf)\n\n try:\n if(maxCandidate_value < threshold):\n accumulator[maxCandidate_z,maxCandidate_x,maxCandidate_y] = 0\n\n if(maxCandidate_value < np.amax(accumulator[from_z:to_z,from_x:to_x,from_y:to_y])):\n #print(\"test\")\n accumulator[maxCandidate_z,maxCandidate_x,maxCandidate_y] = 0\n except ValueError: #raised if `y` is empty.\n pass\n ########## output\n #\n result = np.transpose(np.nonzero(accumulator))\n return(result)\n","sub_path":"src/peak_local_max_3d.py","file_name":"peak_local_max_3d.py","file_ext":"py","file_size_in_byte":3202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"622309942","text":"from flask import Flask, jsonify, Response\nfrom flask_cors import CORS\nimport json\n\n# Initialising things needed for the application\napp = Flask(__name__)\nCORS(app)\nwith open('data/movie_metadata.json') as movief:\n jsonMovie = json.load(movief)\nwith open('data/theater_showtimes.json') as theatref:\n jsonTheatre = json.load(theatref)\n\n# API test to make sure server is correctly configured\n@app.route('/test')\ndef test():\n return jsonify({\"Test\": 200})\n\n# Returns all theatres\n# Done to make the solution more dynamic, but for this test's sake with the lower\n@app.route('/get-all-theatres')\ndef getAllTheatres():\n return jsonify(jsonTheatre), 200\n\n# Returns info regarding movie id\n@app.route('/movie-info/', methods=['GET'])\ndef getMovieInfo(id):\n # find info and return it\n for movie in jsonMovie:\n if (movie['id'] == id):\n return jsonify(movie), 200\n \n # if reach end of json and info not found\n return Response(\"{\\\"response_msg\\\": \\\"no such movies with that id\\\"}\", status=204, mimetype='application/json')\n\n# Returns info regarding theatre id\n@app.route('/theatre-info/', methods=['GET'])\ndef getTheatreInfo(id):\n # find info and return it\n for theatre in jsonTheatre:\n if (theatre['id'] == id):\n return jsonify(theatre), 200\n \n # if reach end of json and info not found\n return Response(\"{\\\"response_msg\\\": \\\"no such theatres with that id\\\"}\", status=204, mimetype='application/json')\n\nif __name__ == '__main__':\n app.run(debug=True)\n","sub_path":"backend/venv/backend.py","file_name":"backend.py","file_ext":"py","file_size_in_byte":1529,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"149780140","text":"from keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.optimizers import SGD\nfrom keras.callbacks import EarlyStopping\nfrom keras.layers import Dropout\nfrom tqdm import tqdm\nclass EveningTuna():\n def __init__(self,xtrain,ytrain,xtest,ytest,size=[16,16,16,1],activation=\"relu\",final_activation=\"sigmoid\",learing_rate=[0.4,0.04,0.004,0.0004],beta=1,epochs=1000):\n self.xtrain = xtrain\n self.ytrain = ytrain\n self.xtest = xtest\n self.ytest = ytest\n self.size = size\n self.activation = activation\n self.final_activation = final_activation\n self.learing_rate = learing_rate\n self.beta = beta\n self.epochs = epochs\n self.size2 = len(xtrain[0])\n\n def Eveninglution(self):\n best_stats_jpg = \"\"\n best_val = 0\n for i in tqdm(range(1,self.size[0])):\n for j in range(1,self.size[1]):\n for k in range(1,self.size[2]):\n for l in self.learing_rate:\n classif = Sequential()\n classif.add(Dense(i,activation=self.activation,input_shape=(self.size2,)))\n classif.add(Dropout(0.15))\n classif.add(Dense(j, activation=self.activation))\n classif.add(Dropout(0.15))\n classif.add(Dense(k, activation=self.activation))\n classif.add(Dropout(0.15))\n classif.add(Dense(1, activation=self.final_activation))\n classif.compile(optimizer=SGD(learning_rate=l),loss=\"mean_squared_error\",metrics=[\"accuracy\"])\n classif.fit(self.xtrain,self.ytrain,batch_size=34,epochs=self.epochs,callbacks=[EarlyStopping(monitor=\"loss\")],verbose=False)\n siema = classif.evaluate(self.xtest,self.ytest,verbose=False)\n if siema[1] > best_val:\n best_val = siema[1]\n best_stats_jpg = \"Size = [\"+str(i)+\",\"+str(j)+\",\"+str(k)+\",1] | activation = \"+str(self.activation)+\" | learing rate:\"+str(l)+\" Acc = \"+str(best_val)\n print(best_stats_jpg)\nif __name__ == \"__main__\":\n pass","sub_path":"WlasnaOptuna.py","file_name":"WlasnaOptuna.py","file_ext":"py","file_size_in_byte":2227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"279083681","text":"\"\"\"\nEdge detection for my BSC thesis\n\"\"\"\n# -*- coding: iso-8859-15 -*-\nimport numpy as np\nimport cv2\nimport sys\n\nfrom skimage.exposure import rescale_intensity\nimport json as _json\nfrom pprint import pprint\nfrom multiprocessing import Pool\nfrom matplotlib import pyplot as plt\nimport matplotlib.image as mpimg\nfrom matplotlib.widgets import Slider, Button, RadioButtons\n\n__author__ = \"Fauszt András\"\n__email__ = \"andras.fauszt@gmail.com\"\n\nexec(open(\"C:\\\\ScanControl\\VideoMode\\\\Python\\\\venv2\\\\Scripts\\\\activate_this.py\").read())\n\n\ndef convolve(image, kernel):\n # grab the spatial dimensions of the image, along with\n # the spatial dimensions of the kernel\n (iH, iW) = image.shape[:2]\n (kH, kW) = kernel.shape[:2]\n\n # allocate memory for the output image, taking care to\n # \"pad\" the borders of the input image so the spatial\n # size (i.e., width and height) are not reduced\n pad = (kW - 1) // 2\n image = cv2.copyMakeBorder(image, pad, pad, pad, pad,\n cv2.BORDER_REPLICATE)\n output = np.zeros((iH, iW), dtype=\"float32\")\n\n # loop over the input image, \"sliding\" the kernel across\n # each (x, y)-coordinate from left-to-right and top to\n # bottom\n for y in np.arange(pad, iH + pad):\n for x in np.arange(pad, iW + pad):\n # extract the ROI of the image by extracting the\n # *center* region of the current (x, y)-coordinates\n # dimensions\n roi = image[y - pad:y + pad + 1, x - pad:x + pad + 1]\n\n # perform the actual convolution by taking the\n # element-wise multiplicate between the ROI and\n # the kernel, then summing the matrix\n k = (roi * kernel).sum()\n\n # store the convolved value in the output (x,y)-\n # coordinate of the output image\n output[y - pad, x - pad] = k\n\n # rescale the output image to be in the range [0, 255]\n output = rescale_intensity(output, in_range=(0, 255))\n output = (output * 255).astype(\"uint8\")\n\n # return the output image\n return output\n\n\ndef updateTrackbar1(value):\n global img\n global threshold1\n global threshold2\n threshold1 = value\n cv2.imshow('main-window', cv2.Canny(img, threshold1, threshold2))\n\n\ndef updateTrackbar2(value):\n global img\n global threshold1\n global threshold2\n threshold2 = value\n cv2.imshow('main-window', cv2.Canny(img, threshold1, threshold2))\n\n\nargs = sys.argv\ncv2.namedWindow(\"main-window\", cv2.WINDOW_NORMAL)\nimg = cv2.imread(\"C:\\\\ScanControl\\\\VideoMode\\\\Images\\\\\" + args[1], cv2.IMREAD_GRAYSCALE)\nthreshold1 = 100\nthreshold2 = 200\nprint(sys.path)\nf = open('C:\\\\ScanControl\\\\VideoMode\\\\settings.json')\ndata = _json.load(f)\nf.close()\n\nkernel = None\nmulti = None\n\nfor (key, value) in data.items():\n if key == args[2]:\n kernel = np.empty([len(value), len(value)])\n for i in range(0, len(value)):\n for j in range(0, len(value)):\n if (multi != 0) or (multi != None):\n kernel[i][j] = value[i][j] * multi\n else:\n kernel[i][j] = value[i][j]\n if key == args[3]:\n multi = value\n\nedges = cv2.Canny(img, threshold1, threshold2)\n# convoleOutput = convolve(img, kernel)\nopencvOutput = cv2.filter2D(img, -1, kernel)\ncv2.imshow(\"main-window\", opencvOutput)\n\n# cv2.createTrackbar(\"Threshold1\", \"main-window\", threshold1, 1500, updateTrackbar1)\n# cv2.createTrackbar(\"Threshold2\", \"main-window\", threshold2, 1500, updateTrackbar2)\n# cv2.imshow(\"main-window\", edges)\n\ncv2.waitKey(0)\n","sub_path":"VideoMode/Python/analizeImage.py","file_name":"analizeImage.py","file_ext":"py","file_size_in_byte":3569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"125890186","text":"import os\nimport logging\n\nfrom library.logger import set_logging\nfrom library.config import Config\nfrom library.aws.utility import Sns\n\n\ndef lambda_handler(event, context):\n \"\"\" Lambda handler to initiate to find public AMIs \"\"\"\n set_logging(level=logging.INFO)\n logging.debug(\"Initiating AMIs public access checking\")\n\n try:\n sns_arn = os.environ[\"SNS_ARN\"]\n config = Config()\n\n if not config.publicAMIs.enabled:\n logging.debug(\"AMIs public access checking disabled\")\n return\n\n logging.debug(\"Iterating over each account to initiate AMIs public access check\")\n for account_id, account_name in config.publicAMIs.accounts.items():\n payload = {\"account_id\": account_id,\n \"account_name\": account_name,\n \"regions\": config.aws.regions,\n \"sns_arn\": sns_arn\n }\n logging.debug(f\"Initiating AMIs public access checking for '{account_name}'\")\n Sns.publish(sns_arn, payload)\n except Exception:\n logging.exception(\"Error occurred while initiation of AMIs public access check\")\n return\n\n logging.debug(\"AMIs public access checking initiation done\")\n","sub_path":"hammer/identification/lambdas/ami-public-access-issues-identification/initiate_to_desc_public_ami_issues.py","file_name":"initiate_to_desc_public_ami_issues.py","file_ext":"py","file_size_in_byte":1245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"480453156","text":"import socket\nimport tqdm\nimport os\n\nSERVER_HOST = \"0.0.0.0\"\nSERVER_PORT = 5679\nBUFFER_SIZE = 8192\nSEPARATOR = b\"\"\n\n\n\ndef server():\n os.chdir(r'C:\\Users\\Shockz\\Desktop\\Browser-Data-Trojan-Windows')\n\n #Creo el socket\n sckt = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n #Enlazamos el socket a ese host y puerto\n sckt.bind((SERVER_HOST, SERVER_PORT))\n #5 peticiones max\n sckt.listen(5)\n\n print(f\"[-] Escuchando como {SERVER_HOST}:{SERVER_PORT}\")\n try:\n client_socket, address = sckt.accept()\n print(f\"[-] {address} esta conectado.\")\n\n received = client_socket.recv(BUFFER_SIZE)\n filename, filesize = received.split(SEPARATOR)\n filename = os.path.basename(filename)\n filesize = int(filesize)\n progress = tqdm.tqdm(range(filesize), f\"Receiving {filename}\", unit=\"B\", unit_scale=True, unit_divisor=1024)\n with open(filename, \"wb\") as f:\n while True:\n bytes_read = client_socket.recv(BUFFER_SIZE)\n if not bytes_read:\n break\n f.write(bytes_read)\n # update the progress bar\n progress.update(len(bytes_read))\n close(sckt)\n except (SystemExit,KeyboardInterrupt):\n close(sckt)\n\ndef close(sckt):\n #sckt.shutdown(socket.SHUT_RDWR)\n sckt.close()\n print(\"closed\")\n","sub_path":"Sockets/sckt_server.py","file_name":"sckt_server.py","file_ext":"py","file_size_in_byte":1404,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"315154437","text":"import numpy as np\nimport pandas as pd\n\n\ndef _right_squeeze(arr, stop_dim=0):\n \"\"\"\n Remove trailing singleton dimensions\n\n Parameters\n ----------\n arr : ndarray\n Input array\n stop_dim : int\n Dimension where checking should stop so that shape[i] is not checked\n for i < stop_dim\n\n Returns\n -------\n squeezed : ndarray\n Array with all trailing singleton dimensions (0 or 1) removed.\n Singleton dimensions for dimension < stop_dim are retained.\n \"\"\"\n last = arr.ndim\n for s in reversed(arr.shape):\n if s > 1:\n break\n last -= 1\n last = max(last, stop_dim)\n\n return arr.reshape(arr.shape[:last])\n\n\ndef array_like(obj, name, dtype=np.double, ndim=1, maxdim=None,\n shape=None, order='C', contiguous=False):\n \"\"\"\n Convert array-like to an array and check conditions\n\n Parameters\n ----------\n obj : array_like\n An array, any object exposing the array interface, an object whose\n __array__ method returns an array, or any (nested) sequence.\n name : str\n Name of the variable to use in exceptions\n dtype : {None, numpy.dtype}\n Required dtype. Default is double. If None, does not change the dtype\n of obj (if present) or uses NumPy to automatically detect the dtype\n ndim : {int, None}\n Required number of dimensions of obj. If None, no check is performed.\n If the numebr of dimensions of obj is less than ndim, additional axes\n are inserted on the right. See examples.\n maxdim : {int, None}\n Maximum allowed dimension. Use ``maxdim`` instead of ``ndim`` when\n inputs are allowed to have ndim 1, 2, ..., or maxdim.\n shape : {tuple[int], None}\n Required shape obj. If None, no check is performed. Partially\n restricted shapes can be checked using None. See examples.\n order : {'C', 'F'}\n Order of the array\n contiguous : bool\n Ensure that the array's data is contiguous with order ``order``\n\n Examples\n --------\n Convert a list or pandas series to an array\n >>> import pandas as pd\n >>> x = [0, 1, 2, 3]\n >>> a = array_like(x, 'x', ndim=1)\n >>> a.shape\n (4,)\n\n >>> a = array_like(pd.Series(x), 'x', ndim=1)\n >>> a.shape\n (4,)\n >>> type(a.orig)\n pandas.core.series.Series\n\n Squeezes singleton dimensions when required\n >>> x = np.array(x).reshape((4, 1))\n >>> a = array_like(x, 'x', ndim=1)\n >>> a.shape\n (4,)\n\n Right-appends when required size is larger than actual\n >>> x = [0, 1, 2, 3]\n >>> a = array_like(x, 'x', ndim=2)\n >>> a.shape\n (4, 1)\n\n Check only the first and last dimension of the input\n >>> x = np.arange(4*10*4).reshape((4, 10, 4))\n >>> y = array_like(x, 'x', ndim=3, shape=(4, None, 4))\n\n Check only the first two dimensions\n >>> z = array_like(x, 'x', ndim=3, shape=(4, 10))\n\n Raises ValueError if constraints are not satisfied\n >>> z = array_like(x, 'x', ndim=2)\n Traceback (most recent call last):\n ...\n ValueError: x is required to have ndim 2 but has ndim 3\n\n >>> z = array_like(x, 'x', shape=(10, 4, 4))\n Traceback (most recent call last):\n ...\n ValueError: x is required to have shape (10, 4, 4) but has shape (4, 10, 4)\n\n >>> z = array_like(x, 'x', shape=(None, 4, 4))\n Traceback (most recent call last):\n ...\n ValueError: x is required to have shape (*, 4, 4) but has shape (4, 10, 4)\n \"\"\"\n arr = np.asarray(obj, dtype=dtype, order=order)\n if maxdim is not None:\n if arr.ndim > maxdim:\n msg = '{0} must have ndim <= {1}'.format(name, maxdim)\n raise ValueError(msg)\n elif ndim is not None:\n if arr.ndim > ndim:\n arr = _right_squeeze(arr, stop_dim=ndim)\n elif arr.ndim < ndim:\n arr = np.reshape(arr, arr.shape + (1,) * (ndim - arr.ndim))\n if arr.ndim != ndim:\n msg = '{0} is required to have ndim {1} but has ndim {2}'\n raise ValueError(msg.format(name, ndim, arr.ndim))\n if shape is not None:\n for actual, req in zip(arr.shape, shape):\n if req is not None and actual != req:\n req_shape = str(shape).replace('None, ', '*, ')\n msg = '{0} is required to have shape {1} but has shape {2}'\n raise ValueError(msg.format(name, req_shape, arr.shape))\n if contiguous:\n arr = np.ascontiguousarray(arr, dtype=dtype)\n return arr\n\n\nclass PandasWrapper(object):\n \"\"\"\n Wrap array_like using the index from the original input, if pandas\n\n Parameters\n ----------\n pandas_obj : {Series, DataFrame}\n Object to extract the index from for wrapping\n\n Notes\n -----\n Raises if ``orig`` is a pandas type but obj and and ``orig`` have\n different numbers of elements in axis 0. Also raises if the ndim of obj\n is larger than 2.\n \"\"\"\n\n def __init__(self, pandas_obj):\n self._pandas_obj = pandas_obj\n self._is_pandas = isinstance(pandas_obj, (pd.Series, pd.DataFrame))\n\n def wrap(self, obj, columns=None, append=None, trim_start=0, trim_end=0):\n \"\"\"\n Parameters\n ----------\n :param obj:\n :param columns:\n :param append:\n :param trim_start:\n :param trim_end:\n :return:\n\n Returns\n -------\n wrapper : callable\n Callable that has one required input and one optional:\n\n * `obj`: array_like to wrap\n * `columns`: (optional) Column names or series name, if obj is 1d\n * `trim_start`: (optional, default 0) number of observations to drop\n from the start of the index, so that the index applied is\n index[trim_start:]\n * `trim_start`: (optional, default 0) number of observations to drop\n from the end of the index , so that the index applied is\n index[:nobs - trim_end]\n \"\"\"\n obj = np.asarray(obj)\n if not self._is_pandas:\n return obj\n\n if obj.shape[0] + trim_start + trim_end != self._pandas_obj.shape[0]:\n raise ValueError('obj must have the same number of elements in '\n 'axis 0 as orig')\n index = self._pandas_obj.index\n index = index[trim_start:index.shape[0] - trim_end]\n if obj.ndim == 1:\n if columns is None:\n name = getattr(self._pandas_obj, 'name', None)\n elif isinstance(columns, str):\n name = columns\n else:\n name = columns[0]\n if append is not None:\n name = append if name is None else name + '_' + append\n\n return pd.Series(obj, name=name, index=index)\n elif obj.ndim == 2:\n if columns is None:\n columns = getattr(self._pandas_obj, 'columns', None)\n if append is not None:\n new = []\n for c in columns:\n new.append(append if c is None else str(c) + '_' + append)\n columns = new\n return pd.DataFrame(obj, columns=columns, index=index)\n else:\n raise ValueError('Can only wrap 1 or 2-d array_like')\n","sub_path":"statsmodels/tools/validation/validation.py","file_name":"validation.py","file_ext":"py","file_size_in_byte":7214,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"372312849","text":"#!/usr/bin/python\n\nimport sys, re, glob, os, shutil\n\ndef main():\n \"\"\"\n \"\"\"\n from optparse import OptionParser\n parser = OptionParser()\n parser.add_option('--input', dest='input', type='string', help=\"vcf input file directory path.\")\n parser.add_option('--output', dest='output', type='string', help=\"output file name.\")\n parser.add_option('--output-dir', dest='outputdir', type='string', help=\"output directore.\")\n (opts, args) = parser.parse_args()\n\n if not os.path.isdir(opts.outputdir):\n os.mkdir(opts.outputdir)\n\n if opts.input:\n input_dir=opts.input+'/output/'\n files=glob.glob(input_dir + \"*/table/*.LargeDeletion.txt\")\n else:\n sys.exit()\n \n with open(\"%s/allyears.csv\" % opts.outputdir, \"w\") as output:\n for fname in files:\t\n samplename=os.path.basename(fname).split('.')[0] # sample name (last column)\n with open(fname, 'r') as f:\n next(f)\n for line in f:\n chrname=line.split('\\t')[0]; start=line.split('\\t')[4]; stop=line.split('\\t')[5]\n newline='\\t'.join([chrname, start, stop, 'DEL', samplename])\n if \"NA\" not in newline:\n output.write(newline); output.write(\"\\n\")\n shutil.copy(\"%s/allyears.csv\" % opts.outputdir, opts.output)\n \nif __name__ == '__main__':\n main()\n","sub_path":"tools/contra/contra_large_del.py","file_name":"contra_large_del.py","file_ext":"py","file_size_in_byte":1395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"288660891","text":"from .heuristic_politics import heuristic_politics\nfrom .pairs_generation_politics import pairs_generation_politics\nfrom .crossover_politics import crossover_politics, crossover_pc_politics\nfrom .mutation_politics import mutation_politics\nfrom .selection_politics import selection_politics, selection_politics_dict\n\nfrom copy import deepcopy\n\nbasic_politics = {\n \"normal\": {\n \"heuristic_politics\": heuristic_politics[[\"MonteCarlo\"]],\n \"pairs_generation_politics\": pairs_generation_politics[\"light\"],\n \"crossover_politics\": crossover_pc_politics[\"basic\"],\n \"mutation_politics\": mutation_politics[\"light\"],\n \"selection_politics\": selection_politics_dict[\"light\"],\n },\n \"stuck\": {\n \"pairs_generation_politics\": pairs_generation_politics[\"kicker\"],\n \"crossover_politics\": crossover_pc_politics[\"kicker\"],\n \"mutation_politics\": mutation_politics[\"kicker\"],\n \"selection_politics\": selection_politics_dict[\"kicker\"],\n },\n}\n\nbasic_greedy_politics = deepcopy(basic_politics)\nbasic_greedy_politics[\"normal\"][\"heuristic_politics\"] = heuristic_politics[\n [\"MonteCarlo\", \"NearestCity\"]\n]\n\nbasic_dfs_politics = deepcopy(basic_politics)\nbasic_dfs_politics[\"normal\"][\"heuristic_politics\"] = heuristic_politics[[\"DFS\"]]\n\nbasic_dfs_greedy_politics = deepcopy(basic_politics)\nbasic_dfs_greedy_politics[\"normal\"][\"heuristic_politics\"] = heuristic_politics[\n [\"DFS\", \"NearestCity\"]\n]\n\np1 = deepcopy(basic_politics)\np1[\"normal\"][\"selection_politics\"] = selection_politics[[\"tournament\", \"wheel\"]]\n\np2 = deepcopy(basic_politics)\np2[\"normal\"][\"selection_politics\"] = selection_politics[[\"wheel\"]]\n\np3 = deepcopy(basic_greedy_politics)\np3[\"normal\"][\"selection_politics\"] = selection_politics[[\"tournament\", \"wheel\"]]\n\np4 = deepcopy(basic_greedy_politics)\np4[\"normal\"][\"selection_politics\"] = selection_politics[[\"wheel\"]]\n\nMA = deepcopy(basic_politics)\nMA[\"stuck\"][\"mutation_politics\"] = mutation_politics[\"opt\"]\n\ngenetic_politics = {\n \"Basic\": basic_politics,\n \"Greedy\": basic_greedy_politics,\n \"DFS\": basic_dfs_politics,\n \"DFS, Greedy\": basic_dfs_greedy_politics,\n \"Monte_Wheel\": p2,\n \"Monte_WheelTournament\": p1,\n \"Monte_Tournament\": basic_politics,\n \"Greedy_Wheel\": p4,\n \"Greedy_WheelTournament\": p3,\n \"Greedy_Tournament\": basic_greedy_politics,\n \"MA\": MA,\n}\n","sub_path":"algorithms/genetic_politics/genetic_politics.py","file_name":"genetic_politics.py","file_ext":"py","file_size_in_byte":2352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"148205294","text":"# Definition for a binary tree node.\n# class TreeNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n\nclass Solution(object):\n def bstFromPreorder(self, preorder):\n \"\"\"\n :type preorder: List[int]\n :rtype: TreeNode\n \"\"\"\n if not preorder: return None\n if len(preorder)==1: return TreeNode(preorder[0])\n left, right = 1000, 1000\n for i, num in enumerate(preorder):\n if num < preorder[0]:\n left = i\n break\n for i, num in enumerate(preorder):\n if num > preorder[0]:\n right = i\n break\n root = TreeNode(preorder[0])\n root.left = self.bstFromPreorder(preorder[left:right])\n root.right = self.bstFromPreorder(preorder[right:])\n return root\n","sub_path":"Leetcode/dfs_recursion_backtracking/1008_ConstructBinarySearchTreeFromPreorderTraversal/Solution.py","file_name":"Solution.py","file_ext":"py","file_size_in_byte":876,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"307426966","text":"#!/usr/bin/env python2\n\n# transpiled with BefunCompile v1.3.0 (c) 2017\nimport sys\nimport zlib, base64\n_g = (\"AR+LCAAAAAAABADNWslu40gS/ZVUS7qQ0Cj3ZBIEMUBfZ4A++mDQfeM1Tzz54ye2pKjFLrvs6prsLklcRMXji3ix0It6elLqSe0vS725cb+UrA+d9Ghp7bTVUWujdbaa\"\n + \"l7HBw4FotA0Zti2epbMOWsN+PJtXRxsdHtR4HE/N2tNlkvoTr//nDZj1h6827tf+7sN7Jz1aGY0Gi02wHtGRwb4anhkyWp1W6zv6P8EBOM/jEQsXCHA/HH5RLqLVX7hu\"\n + \"afrnkIElnY6+sgBmwi3v6KPsDCkja7Av4Hm2HmMi8WPAN/psIl8jwi5wxif6+d+GrDPoQprsw8/VtwJ7m09Agu/IWHZV2O+SZwIBiTOJqOvWu9PRMfUfWr8NWdpYIyuC\"\n + \"WZadKhLMSK4ItDGbgY+5SjJ/efVWCy8evqP++i+sj3H2QDi+jAzsTmgyGGMTmRaM4GP1IMiOeU3MrBaGnTVIoI2iHogqkBOjyyqAhdB+hOyx4n0HMjAMPe+ieEyIz6kG\"\n + \"WocHg9CSSWEMnR+YNAq9bCpgwxxrjLOnW9X/55B1miMrM57smZRk+c6jvMgRXV0wEoIODwMGJMngweRd1jVvwImp/vy7yO5x8ZnfwRmGTY0ZpsLYzGEXfAKXy3jYEh+R\"\n + \"wdlNWCLmKEKaavgFIPL7kd2E4g+QWQolNMc75AHNx7yUAhPDvhYk9IDLsEkR3XqSdhsJShSdzUNkl7UM+4e4boze31/mQ3wqMhdDZw0zjpfIBBnyU5YWWzMcAbV4Xk6a\"\n + \"d+MeeEudEAuaM/KPL2/88H5YPoPs/ZMeIqMKxEoYGQw01m9PtUUtprBKobID6Y1Um3S0xYGFVUuWVLhqIy8C6LNtWmNK7ps+N9GU0ZyW/T2yDfDrD4/x/yDOwEAA0OUK\"\n + \"h4TOS4pCY0FFvLmoPac0egMVjPh1RLNKv5WLIKrcPuezaUuvn/t8NGau269s2GNU34ZsXZGLjrrFqYmLi1r1olTa7Xc2QLG0htOZ1w6RTcfcP7dNF49t16m5Nef83GbY\"\n + \"2R/2/fAzyNTHkVGCjuyJlrUOEhgXgdpx6Uss2o6TFRcdWQQF/kPQRhxYTo6Y1xdzbrIDYCkc4UP/rPuTGWYTC8TocM/Zd3sjWQVlMNYh4nFSN3KSirCbCn1Ja/gxayn+\"\n + \"oT8IcjJjDiuNSY/shC43xy437TN8OO9fldn37V6Vvbrl7A1x+GlkiQqrTQGC4ibJSULHYUVyIVVTgsBkp6UQQWJ9FlXsolzlVhQLXOYwjAX6pdJD1C2fymfvw3+0EJdh\"\n + \"8z0SAZ+Bo85ZUT8EZcXF8MUZLC/9NlObzX3B5tXyuSMIIUhgn88OoOSjtcUaOy013IbPZepHTvsBBQFIZGwE6Xa34lBFJawHSGvQ96zmQpJ2miDvyDfGGYjIAQNq3Pze\"\n + \"MsauOfXels9xtgG32fEDZNBqGrTJkomGIOSVEspvJnuHPCScGODJQcAZ21XvtaytQZAaD1efXqZ9f/V7Ysyw7IZfXuvzzY84NiChoAqxSj9KX67htlaLiXOxkxO4Aa3y\"\n + \"WNNDuf4hcsyj9QXdc5le1PKL+zMqnxAHlhGsEVaMB3uptHBShOCB5HTseIuEvxM5wQoYyxkZgqBfjk7PEGqn3X56gfdJAJJFyKdafnHnGeq9Rsq8qbMMMi5RgSG9ADgg\"\n + \"aHziuY+nvBzW2Fu9t7YEVqtFF/jCy7LfnSBZQ6KevTPwYnMZDgTnI0a/e9JbRSmsgUzhHEUaBw0nsMAfOU9zy2KZRxeo9a7aUoMPdngKPETomTi8/Cj/VA/EHYyf98sL\"\n + \"KA066vKtc5AriC+H3rSvGScenRYBkca5y1zkrxms+uKFFtlDZHmRHV0ndhSD050hBTAPG7u+jsyM7fNs3JzVMbf4TkUp/vIyYBqjeUeqsyhSdXS6zq8ZwII0IiyDIx+c\"\n + \"lBqqW7LQZjFIPfcCRm5CLhF/oAePbM3s/ZVJAzI5fR3ZaWqbputso/I5zDoe8wwO30CuPKhQiYg8Ija1m3aJUxPXujUWZXlqpCsSuYKzgblPdK4ffckkid6XkBrrZ9d4\"\n + \"eDm2yweM/hAyNfZRl5ypJOiNK/kc9RyOk/qXogpYrObxdbykaoyqvAVVFZ11RXbX/Bz0Wn5BYZIM/vT42oOEzG17dsBb40hElLjkG1n6RiTfPUlc/uLjz22LejX8W3EN\"\n + \"FfRmfC0x013x5NjJZORqJCtYm4TirIVibYLUJfRLIIJj9LNpo98kuOl1AtTvivZH1n7oU+K88pzNjNVNCG0TAjqFTAmM3YyvQ0xcAfNAJ7h8mZeG+lwjM/4so2UpYdyl\"\n + \"uCYEV4Zc7uxhWG5x7Qf1aI3Q2EMtDf98oJYc7C5SkSpoZiNmfhtLNnFeFuILbiNeS8q/Tm/H1zJBhHdMvV11VRxzE780LHY8LM78WCAIr4l4S7pOC5Cwfgdib05wZw2y\"\n + \"Z3xZ9qMavcW73Fo759apN1LT6m6Lwsw4B5KiA1Rtu36APQNhaedoAc9IjEVQrImqKx5xbx62OM96T20bpWQw3zieY9HDG2kyLc93HLcv1Y3ludQmyRyAI4iwAiA8Vvxm\"\n + \"gGJyycAFGKFdgVc7XPNbLGidhXu/aNu3Rs0KtpY5hsMLlgDQ8Be4oYOazMUlrlmnZJT1Ku9r++mpCEE5MVFIxE6aehqTKQni9B7bg2D0is1XfNcyr0ZoZYInYCBey4kL\"\n + \"SVCzP57+QOG0IJyQkEA4IS/VDbw10NABbhWghll2eKUdfLPtY0S9BRd0TQT/O7njLelWmuJuM75mAmn+qOkpGYt/cNQXpDoaYL0gKbGyXVM5VVfjqEmLU7GpuITRsbwc\"\n + \"VAYTLN7phSXTGZBMC5LZHuuGbXJX4CD4IDCsSzsDMmeGMELvA55Z5fYEqhHhC+iK7W2cchnMj71kfN3xNNxuA66KJA9epcmGLkwEyK1PYyKXjeDQaHv2JcVz7ksKx4xj\"\n + \"kKkf0BMLuupLsHOXN3Q9CLURnBJaVoecLSpE1asT7g6oEyEW/1Z8ArJOCth1fJ3piRmHjdl0zzePknR9HFqrYccPCV3twcmupg0JmugILzO4IVj1WnqsJaf+tS83nN2r\"\n + \"4wTx5kE1irLKDWr2aszTa7v7ew5I1243x82Xrvsm1EXITqTy8tSCqdlUHI4DrUO9N4kUJ5AO0hdkTFefkzp9QUZrYfFKEer8FPKQZCw+HjKGml1DzWGNYilmarDx5kpL\"\n + \"mACOGkb1cE3hapNYcOuQ27ASuDX6pOPCp/E1CvnBdQrOXwQD9/ko7JmN6gOCE0QaMGeTKsCdmwQZ5rOlhov1l2Crm+4SbvVScf5712L2iDOkOQv6EYqPM6Q6cM9wra6s\"\n + \"0ThL5B6MHS4RE8HgfOR6cEpPlnJ9ym6xPTCoIGsTLpwSsgVNg0DDcWORsSMEG7QZahknoKKG2oauBSLnvjqZ3kjk7yxpqbs6K8QMYPD2J/oDCDQ7dPLYzLK/8fMLmZk4\"\n + \"yNqRAhM6BtITPkzF1XgdZCTakLT7aYC4gWpWQu2Ks08jeGt5eRrmSOZtvFS1l+Ulm3OJZeodSFcnJfnLEMnTlM+WFBqox8oaZcAhdmh9r/RiMk4fN+msxphqlXEXFs9t\"\n + \"+07r/PYyUQbythqouTgG6jxNSbPjp7PUoOZLh8lQTQYSKSzBL+nJk5FqBK8+ImFg27z+Hlatu1el5vHUOzddB5qjQFNOrSxSqKmf4tHxg/dYRdzRX0DQKAcYhIwVNuWk\"\n + \"MzejR2LZo1LWQUhg94Y2Bq8+oayjIC7A2VnRZ+wx4I5CWQSJeJvT1kD7nsUPAWmQweOAWqpT6BiedOBso7tAomiK3ti1RiRSU2cq7/L8bJQoK9NI1BWn0XhdJsq3r/2g\"\n + \"yh1n3xVoYCICEC+L3W1fRiGViEvjsZF2VAlTEsie70mucNc/TcL0DxdfCiTUZDDOcGqVbNtY+FLRlhxMMlrvjpDU3NlJZlONQg7PDjWzvc1xsNvV3d9F72fXdKve+2Gp\"\n + \"o/BJYYO2hwL/KtJaoK2BODvfJrcWs5s7b2pKOPJz8ff1dT+7UrWxmq2hgT8A7TeRJiwBsu+OuV++ppVEoA+Q09S/vMPZbyLlS2t6cwoC1cZDuv9/1v8AqAXMMpQqAAA=\")\ng = base64.b64decode(_g)[1:]\nfor i in range(ord(base64.b64decode(_g)[0])):\n g = zlib.decompress(g, 16+zlib.MAX_WBITS)\ng=list(map(ord, g))\ndef gr(x,y):\n if(x>=0 and y>=0 and x<218 and y<50):\n return g[y*218 + x];\n return 0;\ndef gw(x,y,v):\n if(x>=0 and y>=0 and x<218 and y<50):\n g[y*218 + x]=v;\ndef td(a,b):\n return ((0)if(b==0)else(a//b))\ndef tm(a,b):\n return ((0)if(b==0)else(a%b))\ns=[]\ndef sp():\n global s\n if (len(s) == 0):\n return 0\n return s.pop()\ndef sa(v):\n global s\n s.append(v)\ndef sr():\n global s\n if (len(s) == 0):\n return 0\n return s[-1]\ndef _0():\n sa(0)\n sa(49)\n sa(1)\n return 1\ndef _1():\n return (3)if(sp()!=0)else(2)\ndef _2():\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()+((((gr(9,1)-48)*10)+(gr(10,1)-48))*10)+(gr(11,1)-48))\n\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n return 3\ndef _3():\n sa(sr());\n gw(6,0,sp())\n gw(1,3,80)\n gw((gr(1,3)%9)+9,(gr(1,3)/9)+1,gr((gr(1,3)%9)+((gr(6,0)/5)*9)+128,8+((gr(6,0)%5)*9)))\n sa(80)\n sa(80)\n return 4\ndef _4():\n return (94)if(sp()!=0)else(5)\ndef _5():\n sp();\n\n return (7)if(sr()!=-1)else(6)\ndef _6():\n sp();\n sys.stdout.write(str(sp())+\" \")\n sys.stdout.flush()\n return 95\ndef _7():\n gw(1,1,22)\n gw(6,1,729)\n gw(17,9,(gr(17,9)%16)+48)\n gw(8+gr(1,1),9,0)\n sa(sp()-1)\n\n sa(80)\n sa(80)\n return 8\ndef _8():\n return (93)if(sp()!=0)else(9)\ndef _9():\n gw(2,0,1)\n sp();\n sa(gr(6,1)-1)\n sa(0)\n sa(((gr(6,1)-1)%27)+72)\n sa(((gr(6,1)-1)/27)+1)\n gw(((gr(6,1)-1)%27)+35,((gr(6,1)-1)/27)+1,0)\n return 10\ndef _10():\n v0=sp()\n v1=sp()\n gw(v1,v0,sp())\n sa(sr());\n\n return (92)if(sp()!=0)else(11)\ndef _11():\n gw(3,0,0)\n gw(3,2,8)\n gw(2,2,8)\n gw(1,2,2)\n gw(5,2,1)\n sp();\n sa(80)\n sa(gr(17,9)-48)\n gw(4,2,gr(17,9)-48)\n return 12\ndef _12():\n return (28)if(sp()!=0)else(13)\ndef _13():\n sa(sr());\n\n return (91)if(sp()!=0)else(14)\ndef _14():\n gw(1,4,0)\n sp();\n return 15\ndef _15():\n return (16)if(gr(3,0)!=81)else(90)\ndef _16():\n gw(2,4,8)\n gw(3,4,8)\n sa(80)\n return 17\ndef _17():\n return (18)if(gr(9+gr(2,4),1+gr(3,4))!=48)else(80)\ndef _18():\n sa(sr());\n\n return (79)if(sp()!=0)else(19)\ndef _19():\n sp();\n\n return (20)if((gr(1,4))!=0)else(21)\ndef _20():\n gw(1,4,0)\n return 15\ndef _21():\n gw(1,5,0)\n return 22\ndef _22():\n gw(2,5,0)\n gw(3,5,0)\n gw(4,5,0)\n gw(5,5,6561)\n gw(1,6,8)\n gw(2,6,8)\n sa(80)\n return 23\ndef _23():\n sa(gr(9+gr(1,6),1+gr(2,6))-48)\n return 24\ndef _24():\n return (25)if(sp()!=0)else(70)\ndef _25():\n sa(sr());\n\n return (69)if(sp()!=0)else(26)\ndef _26():\n sp();\n\n return (27)if((gr(4,5))!=0)else(55)\ndef _27():\n global t0\n t0=gr(2,0)+1\n gw(2,0,gr(2,0)+1)\n gw(gr(1,1)+gr(2,5),1+gr(3,5),t0)\n gw(1,2,0)\n gw(2,2,gr(2,5))\n gw(3,2,gr(3,5))\n gw(4,2,gr(4,5))\n gw(5,2,gr(2,0))\n return 28\ndef _28():\n gw(gr(2,2)+9,gr(3,2)+1,gr(4,2)+48)\n gw(3,0,gr(3,0)+1)\n gw(1,3,8)\n sa(8)\n return 29\ndef _29():\n gw(35+(gr(2,2)*3)+(gr(1,3)%3),1+(gr(3,2)*3)+(gr(1,3)/3),88)\n\n return (30)if((gr(72+(gr(2,2)*3)+(gr(1,3)%3),1+(gr(3,2)*3)+(gr(1,3)/3)))!=0)else(54)\ndef _30():\n sa(sr());\n return 31\ndef _31():\n return (32)if(sp()!=0)else(33)\ndef _32():\n sa(sp()-1)\n\n sa(sr());\n gw(1,3,sp())\n return 29\ndef _33():\n gw(2,3,8)\n sp();\n sa(8)\n return 34\ndef _34():\n gw(35+(gr(2,3)*3)+((gr(4,2)-1)%3),1+(gr(3,2)*3)+((gr(4,2)-1)/3),88)\n\n return (35)if((gr(72+(gr(2,3)*3)+((gr(4,2)-1)%3),1+(gr(3,2)*3)+((gr(4,2)-1)/3)))!=0)else(53)\ndef _35():\n sa(sr());\n return 36\ndef _36():\n return (37)if(sp()!=0)else(38)\ndef _37():\n sa(sp()-1)\n\n sa(sr());\n gw(2,3,sp())\n return 34\ndef _38():\n gw(3,3,8)\n sp();\n sa(8)\n return 39\ndef _39():\n gw(35+(gr(2,2)*3)+((gr(4,2)-1)%3),1+(gr(3,3)*3)+((gr(4,2)-1)/3),88)\n\n return (40)if((gr(72+(gr(2,2)*3)+((gr(4,2)-1)%3),1+(gr(3,3)*3)+((gr(4,2)-1)/3)))!=0)else(52)\ndef _40():\n sa(sr());\n\n return (51)if(sp()!=0)else(41)\ndef _41():\n gw(2,3,2)\n gw(3,3,2)\n sp();\n sa(8)\n return 42\ndef _42():\n gw(35+((((gr(2,2)/3)*3)+gr(2,3))*3)+((gr(4,2)-1)%3),1+((((gr(3,2)/3)*3)+gr(3,3))*3)+((gr(4,2)-1)/3),88)\n\n return (43)if((gr(72+((((gr(2,2)/3)*3)+gr(2,3))*3)+((gr(4,2)-1)%3),1+((((gr(3,2)/3)*3)+gr(3,3))*3)+((gr(4,2)-1)/3)))!=0)else(50)\ndef _43():\n sa(sr());\n return 44\ndef _44():\n return (49)if(sp()!=0)else(45)\ndef _45():\n sp();\n sa(gr(1,2))\n\n return (46)if((gr(1,2))!=0)else(14)\ndef _46():\n sa(sp()-1)\n\n sa(sr());\n\n return (47)if(sp()!=0)else(48)\ndef _47():\n sp();\n return 13\ndef _48():\n sp();\n return 18\ndef _49():\n sa(sp()-1)\n\n sa(sr());\n gw(2,3,sr()%3)\n sa(sp()/3);\n\n gw(3,3,sp())\n return 42\ndef _50():\n gw(72+((((gr(2,2)/3)*3)+gr(2,3))*3)+((gr(4,2)-1)%3),1+((((gr(3,2)/3)*3)+gr(3,3))*3)+((gr(4,2)-1)/3),gr(5,2))\n sa(sr());\n return 44\ndef _51():\n sa(sp()-1)\n\n sa(sr());\n gw(3,3,sp())\n return 39\ndef _52():\n gw(72+(gr(2,2)*3)+((gr(4,2)-1)%3),1+(gr(3,3)*3)+((gr(4,2)-1)/3),gr(5,2))\n return 40\ndef _53():\n gw(72+(gr(2,3)*3)+((gr(4,2)-1)%3),1+(gr(3,2)*3)+((gr(4,2)-1)/3),gr(5,2))\n sa(sr());\n return 36\ndef _54():\n gw(72+(gr(2,2)*3)+(gr(1,3)%3),1+(gr(3,2)*3)+(gr(1,3)/3),gr(5,2))\n sa(sr());\n return 31\ndef _55():\n gw(1,7,-1)\n gw(2,7,-1)\n gw(3,7,-1)\n gw(5,7,8)\n gw(6,7,8)\n sa(80)\n return 56\ndef _56():\n return (57)if((gr(22+gr(5,7),1+gr(6,7))-gr(2,0))!=0)else(68)\ndef _57():\n sa(sr());\n\n return (67)if(sp()!=0)else(58)\ndef _58():\n gw(5,7,26)\n gw(6,7,26)\n sp();\n sa(728)\n return 59\ndef _59():\n return (66)if((gr(72+gr(5,7),1+gr(6,7))-gr(2,0))!=0)else(60)\ndef _60():\n gw(72+gr(5,7),1+gr(6,7),0)\n gw(35+gr(5,7),1+gr(6,7),0)\n\n return (65)if(gr(9+(gr(5,7)/3),1+(gr(6,7)/3))!=48)else(61)\ndef _61():\n sa(sr());\n return 62\ndef _62():\n return (64)if(sp()!=0)else(63)\ndef _63():\n gw(2,0,gr(2,0)-1)\n gw(22+gr(2,7),1+gr(3,7),0)\n gw(1,5,gr(1,7))\n sp();\n return 22\ndef _64():\n sa(sp()-1)\n\n gw(5,7,sr()%27)\n gw(6,7,sr()/27)\n return 59\ndef _65():\n gw(9+(gr(5,7)/3),1+(gr(6,7)/3),48)\n gw(3,0,gr(3,0)-1)\n sa(sr());\n return 62\ndef _66():\n sa(sr());\n return 62\ndef _67():\n sa(sp()-1)\n\n gw(5,7,sr()%9)\n gw(6,7,sr()/9)\n return 56\ndef _68():\n gw(1,7,gr(9+gr(5,7),1+gr(6,7))-48)\n gw(2,7,gr(5,7))\n gw(3,7,gr(6,7))\n return 57\ndef _69():\n sa(sp()-1)\n\n gw(1,6,sr()%9)\n gw(2,6,sr()/9)\n return 23\ndef _70():\n gw(4,6,0)\n gw(5,6,0)\n gw(6,6,1)\n sa(1)\n return 71\ndef _71():\n return (74)if((gr(35+(gr(1,6)*3)+((gr(6,6)-1)%3),1+(gr(2,6)*3)+((gr(6,6)-1)/3)))!=0)else(72)\ndef _72():\n gw(5,6,gr(5,6)+1)\n\n return (74)if((((1)if((gr(4,6))!=0)else(0))+((1)if(gr(6,6)<=gr(1,5))else(0)))!=0)else(73)\ndef _73():\n gw(4,6,gr(6,6))\n return 74\ndef _74():\n return (78)if(sr()!=9)else(75)\ndef _75():\n sp();\n\n return (76)if((((0)if((gr(5,6))!=0)else(1))+((1)if(gr(5,5)<=gr(5,6))else(0)))!=0)else(77)\ndef _76():\n sa(1)\n return 24\ndef _77():\n gw(2,5,gr(1,6))\n gw(3,5,gr(2,6))\n gw(4,5,gr(4,6))\n gw(5,5,gr(5,6))\n return 76\ndef _78():\n sa(sp()+1)\n\n sa(sr());\n gw(6,6,sp())\n return 71\ndef _79():\n sa(sp()-1)\n\n gw(2,4,sr()%9)\n gw(3,4,sr()/9)\n return 17\ndef _80():\n gw(5,4,0)\n gw(6,4,0)\n gw(4,4,8)\n sa(8)\n return 81\ndef _81():\n return (82)if((gr(35+(gr(2,4)*3)+(gr(4,4)%3),1+(gr(3,4)*3)+(gr(4,4)/3)))!=0)else(89)\ndef _82():\n sa(sr());\n return 83\ndef _83():\n return (88)if(sp()!=0)else(84)\ndef _84():\n global t0\n t0=gr(6,4)\n sp();\n\n return (85)if((gr(6,4))!=0)else(87)\ndef _85():\n global t0\n t0=t0-1\n\n return (18)if((t0)!=0)else(86)\ndef _86():\n gw(1,4,gr(1,4)+1)\n gw(1,2,1)\n gw(2,2,gr(2,4))\n gw(3,2,gr(3,4))\n gw(4,2,gr(5,4))\n gw(5,2,gr(2,0))\n return 28\ndef _87():\n sp();\n return 55\ndef _88():\n sa(sp()-1)\n\n sa(sr());\n gw(4,4,sp())\n return 81\ndef _89():\n gw(6,4,gr(6,4)+1)\n gw(5,4,gr(4,4)+1)\n sa(sr());\n return 83\ndef _90():\n sa((0)if(sr()-49!=0)else(1))\n return 1\ndef _91():\n sa(sp()-1)\n\n gw(3,2,sr()/9)\n gw(2,2,sr()%9)\n gw(1,2,2)\n gw(5,2,1)\n sa(sr());\n sa((sr()%9)+9)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/9);\n\n sa(sp()+1)\n\n v0=sp()\n sa(gr(sp(),v0))\n sa(sp()-48)\n\n sa(sr());\n gw(4,2,sp())\n return 12\ndef _92():\n sa(sp()-1)\n\n sa(sr());\n sa(0)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa((sr()%27)+35)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/27);\n\n sa(sp()+1)\n\n v0=sp()\n v1=sp()\n gw(v1,v0,sp())\n sa(sr());\n sa(0)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa((sr()%27)+72)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/27);\n\n sa(sp()+1)\n return 10\ndef _93():\n sa(sp()-1)\n\n sa(sr());\n sa(sr());\n sa((sr()%9)+9)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/9);\n\n sa(sp()+1)\n\n v0=sp()\n sa(gr(sp(),v0))\n sa(sp()%16);\n\n sa(sp()+48)\n\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa((sr()%9)+9)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/9);\n\n sa(sp()+1)\n\n v0=sp()\n v1=sp()\n gw(v1,v0,sp())\n sa(sr());\n sa(0)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa((sr()%9)+gr(1,1))\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n sa(sp()/9);\n\n sa(sp()+1)\n\n v0=sp()\n v1=sp()\n gw(v1,v0,sp())\n sa(sr());\n return 8\ndef _94():\n global t0\n sa(sp()-1)\n\n sa(sr());\n sa(sr());\n gw(1,3,sp())\n sa(sp()/9);\n\n sa(sp()+((gr(6,0)%5)*9))\n\n sa((gr(1,3)%9)+((gr(6,0)/5)*9)+128)\n v0=sp()\n v1=sp()\n sa(v0)\n sa(v1)\n v0=sp()\n t0=gr(sp(),v0)\n gw((gr(1,3)%9)+9,(gr(1,3)/9)+1,t0)\n sa(sr());\n return 4\nm=[_0,_1,_2,_3,_4,_5,_6,_7,_8,_9,_10,_11,_12,_13,_14,_15,_16,_17,_18,_19,_20,_21,_22,_23,_24,_25,_26,_27,_28,_29,_30,_31,_32,_33,_34,_35,_36,_37,_38,_39,_40,_41,_42,_43,_44,_45,_46,_47,_48,_49,_50,_51,_52,_53,_54,_55,_56,_57,_58,_59,_60,_61,_62,_63,_64,_65,_66,_67,_68,_69,_70,_71,_72,_73,_74,_75,_76,_77,_78,_79,_80,_81,_82,_83,_84,_85,_86,_87,_88,_89,_90,_91,_92,_93,_94]\nc=0\nwhile c<95:\n c=m[c]()\n","sub_path":"compiled/Python2/Euler_Problem-096.py","file_name":"Euler_Problem-096.py","file_ext":"py","file_size_in_byte":14781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"553532188","text":"def main():\n p = argparse.ArgumentParser(usage=__doc__.lstrip())\n p.add_argument('--project', default='scipy/scipy')\n p.add_argument('milestone')\n args = p.parse_args()\n getter = CachedGet('gh_cache.json')\n try:\n milestones = get_milestones(getter, args.project)\n if (args.milestone not in milestones):\n msg = 'Milestone {0} not available. Available milestones: {1}'\n msg = msg.format(args.milestone, ', '.join(sorted(milestones)))\n p.error(msg)\n issues = get_issues(getter, args.project, args.milestone)\n issues.sort()\n finally:\n getter.save()\n prs = [x for x in issues if ('/pull/' in x.url)]\n issues = [x for x in issues if (x not in prs)]\n\n def print_list(title, items):\n print()\n print(title)\n print(('-' * len(title)))\n print()\n for issue in items:\n msg = '- `#{0} <{1}>`__: {2}'\n title = re.sub('\\\\s+', ' ', issue.title.strip())\n if (len(title) > 60):\n remainder = re.sub('\\\\s.*$', '...', title[60:])\n if (len(remainder) > 20):\n remainder = (title[:80] + '...')\n else:\n title = (title[:60] + remainder)\n msg = msg.format(issue.id, issue.url, title)\n print(msg)\n print()\n msg = 'Issues closed for {0}'.format(args.milestone)\n print_list(msg, issues)\n msg = 'Pull requests for {0}'.format(args.milestone)\n print_list(msg, prs)\n return 0","sub_path":"Data Set/bug-fixing-5/398f9a750a163d54267a3574a5a2e99db567b506--bug.py","file_name":"398f9a750a163d54267a3574a5a2e99db567b506--bug.py","file_ext":"py","file_size_in_byte":1534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"419981932","text":"# マルウェアファミリのクラス\r\nclass family:\r\n def __init__(self):\r\n self.name = '' # ファミリ名\r\n self.samples = [] # 検体のリスト\r\n self.num = 0 # 検体数\r\n self.par_vec = [] # Paragraph Vectorのリスト\r\n self.sampling_count = 0 # サンプリングできる回数\r\n self.svm_data = [] # SVMに追加するデータリスト\r\n self.svm_label = [] # SVMに追加するデータリストの教師ラベル\r\n\r\n","sub_path":"MalwareVariantsClassification_YukiNakashima/program/Family.py","file_name":"Family.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"167203974","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Jul 21 12:54:48 2019\r\n积分色素浓度画图 散点和分布图两种\r\n@author: Administrator\r\n\"\"\"\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport sys\r\nsys.path.append('D:/Refresh/py36')\r\nimport support_pie\r\nfrom matplotlib.colors import LogNorm\r\nimport matplotlib as mpl \r\n\r\ndef del_nan(value):\r\n arg_nan = np.where(np.isnan(value))\r\n arg = list(np.arange(len(value)))\r\n for i in range(len(arg_nan[0])):\r\n i = i+1\r\n x = arg_nan[0][-i]\r\n del arg[x]\r\n value = value[arg]\r\n return value,arg\r\n\r\n\r\ndef plot(pigment,df_plt,image):\r\n lon = np.array(df_plt.loc[:,'Longitude [degrees_east]'])\r\n lat = np.array(df_plt.loc[:,'Latitude [degrees_north]'])\r\n value = np.array(df_plt.loc[:,pigment])\r\n lon,lat,value = support_pie.inter_data(lon,lat,value,1000)\r\n# norm = LogNorm(vmin=np.min(value),\r\n# vmax=np.max(value)) \r\n norm = mpl.colors.Normalize(vmin=value.min(),\r\n vmax=value.max()*0.8)\r\n B = plt.contourf(lon,lat,value,25,cmap='rainbow',\r\n norm=norm,alpha=1)\r\n #colorbar \r\n plt.colorbar(B,format='%.1f') \r\n C = plt.contour(lon,lat,value,5,colors='white',linewidths=0.5,\r\n norm=norm)\r\n plt.clabel(C,inline=True,fontsize=12,colors='black',fmt='%.1f')\r\n #岛屿\r\n support_pie.contourf_re(image,lon_left,lon_right,lat_bottom,lat_top)\r\n\r\n \r\ndef plot_scatter(pigment,df_plt,image):\r\n# fig = plt.figure()\r\n support_pie.contourf(image,lon_left,lon_right,lat_bottom,lat_top)\r\n\r\n #station_index = np.array(df_plt['Station'])\r\n station_lon = np.array(df_plt.loc[:,'Longitude [degrees_east]'])\r\n station_lat = np.array(df_plt.loc[:,'Latitude [degrees_north]'])\r\n colors = np.array(df_plt.loc[:,pigment])/5\r\n\r\n norm = LogNorm(vmin=np.min(colors),\r\n vmax=np.max(colors))\r\n plt.scatter(station_lon,station_lat,\r\n c = colors,cmap='rainbow',s = 60,norm=norm) \r\n\r\n cb = plt.colorbar(format='%.1f')\r\n cb_ticks = cb.get_ticks()\r\n cb.set_ticks(cb_ticks[::2])\r\n#读取数据\r\ndf = pd.read_excel('D:/Refresh/data/CHINARE-35/水柱积分藻种数据.xlsx',\r\n sheet_name='Sheet1').set_index('Station')\r\n\r\n\r\nfontsize={'size':14,'weight':'normal'} \r\n\r\nlon_left,lon_right,lat_bottom,lat_top = -108,-88,-73,-65\r\nz = ['Chlorophyll A','Pheophorbide A','Pheophytin A','Fucoxanthin','19-hex',\r\n 'Alloxanthin','Peridinin','Chlorophyll B'] #色素列名\r\ntext = ['Chlorophyll a/mg·$\\mathregular{m^{-2}}$',\r\n 'Pheophorbide a/mg·$\\mathregular{m^{-2}}$',\r\n 'Pheophytin a/mg·$\\mathregular{m^{-2}}$',\r\n 'Fucoxanthin/mg·$\\mathregular{m^{-2}}$',\r\n \"19'-hexanoyloxyfucoxanthin/mg·$\\mathregular{m^{-2}}$\",\r\n 'Alloxanthin/mg·$\\mathregular{m^{-2}}$',\r\n 'Peridinin/mg·$\\mathregular{m^{-2}}$',\r\n 'Chlorophyll b/mg·$\\mathregular{m^{-2}}$'] #题名\r\nindex,column = [4,2] #行和列的数量 【2*2 也可,z和title列数量(4)对应就好】\r\nfig_dir = 'D:/Refresh/data/CHINARE-35/pic/特征色素积分分布-english.png' #图片保存路径\r\n\r\nplt.rcParams['font.sans-serif']=['Times New Roman'] \r\nfig = plt.figure()\r\nplt.rcParams['figure.figsize'] = (10,10)\r\nplt.subplots_adjust(wspace=0.01,hspace=0.1)\r\n\r\nimage = support_pie.loadtif()\r\nnum = 0\r\nfor i in range(index):\r\n for j in range(column):\r\n plt.subplot(index,column,num+1)\r\n plt.gca().xaxis.set_ticks_position('top') \r\n plot(z[num],df,image)\r\n plt.text(-90,-66,support_pie.getChar(num),fontdict=fontsize)\r\n plt.text(-107.7,-66.5,s=text[num],fontdict=fontsize)\r\n\r\n# plt.text(char_x,char_y,support_pie.getChar(num),\r\n# fontdict=fontsize)\r\n if i != 0 :\r\n plt.xticks(())\r\n if j != 0:\r\n plt.yticks(())\r\n num += 1\r\n \r\nplt.savefig(fig_dir,bbox_inches='tight',dpi=1000,pad_inches=0.1)\r\nplt.show()\r\n\r\n###积分chla 分布\r\n#plt.rcParams['figure.figsize'] = (11,6)\r\n#plot('tchla',df,image)\r\n#plt.text(x,y,s='Chlorophyll a/μg·$\\mathregular{m^{-2}}$',\r\n# fontdict=fontsize)\r\n#plt.gca().xaxis.set_ticks_position('top')\r\n#plt.text(-46,-59.5,'f',fontdict=fontsize)\r\n#plt.savefig('D:/Refresh/data/CHINARE-32/pic/CHEMTAX藻种分布/chla积分分布.png',\r\n# bbox_inches='tight',dpi=1000,pad_inches=0.1)","sub_path":"CHINARE-35阿蒙森海东侧海域/图4-3 叶绿素a及其降解产物及特征色素积分浓度分布图.py","file_name":"图4-3 叶绿素a及其降解产物及特征色素积分浓度分布图.py","file_ext":"py","file_size_in_byte":4417,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"5822910","text":"#!/usr/bin/python3\n#-----------------------------------------------------------------------------\n# project: tkinterlite\n# authors: 1966bc\n# mailto: [giuseppe.costanzi@gmail.com]\n# modify: 10/04/2017\n# version: 0.1 \n#-----------------------------------------------------------------------------\n\nfrom tkinter import *\nfrom tkinter import ttk\nfrom tkinter import messagebox\n\nclass Dialog(Toplevel): \n def __init__(self,parent,engine,index=None):\n super().__init__(name='category')\n \n self.resizable(0,0)\n self.parent = parent\n self.engine = engine\n self.index = index\n \n self.enable = BooleanVar()\n\n self.center_me()\n self.init_ui()\n\n def center_me(self):\n\n #center window\n x = (self.master.winfo_screenwidth() - self.master.winfo_reqwidth()) / 2\n y = (self.master.winfo_screenheight() - self.master.winfo_reqheight()) / 2\n self.master.geometry(\"+%d+%d\" % (x, y))\n \n def init_ui(self):\n\n self.panel = self.engine.get_panel_frame(self)\n self.panel.grid(row = 0, column = 0, sticky=N+W+S+E)\n\n Label(self.panel, text=\"Category:\",anchor='w').grid(row=0, sticky='w')\n self.txtCategory = Entry(self.panel,)\n self.txtCategory.grid(row=0, column=1, sticky='w')\n\n Label(self.panel, text=\"Description:\",anchor='w').grid(row=1, sticky='w')\n self.txtDescription = Entry(self.panel,)\n self.txtDescription.grid(row=1, column=1, sticky='w')\n\n Label(self.panel, text=\"Enable:\").grid(row=2, sticky='w')\n self.ckEnable = Checkbutton(self.panel, onvalue=1, offvalue=0, variable = self.enable,)\n self.ckEnable.grid(row=2, column=1, sticky='w')\n\n self.engine.get_save_cancel(self, self)\n \n\n def on_open(self, selected_category = None):\n\n if selected_category is not None:\n self.insert_mode = False\n self.selected_category = selected_category\n msg = \"Update %s\" % (self.selected_category[1],)\n self.set_values()\n else:\n self.insert_mode = True\n msg = \"Insert new category\"\n self.enable.set(1)\n\n self.title(msg)\n self.txtCategory.focus()\n \n def on_save(self, evt):\n\n if self.on_fields_control()==False:\n\n msg = \"Please fill all fields.\"\n messagebox.showwarning(self.engine.title,msg)\n\n else:\n \n if messagebox.askyesno(self.engine.title, \"Do you want to save?\",parent=self) == True:\n\n args = self.get_values()\n\n if self.insert_mode == False:\n\n sql = self.engine.get_update_sql('categories','category_id')\n\n args = self.engine.get_update_sql_args(args, self.selected_category[0])\n \n elif self.insert_mode == True:\n\n sql = self.engine.get_insert_sql('categories',len(args))\n\n self.engine.write(sql,args)\n self.parent.on_open()\n\n if self.index is not None:\n self.parent.lstCategories.see(self.index)\n self.parent.lstCategories.selection_set(self.index)\n \n self.on_cancel()\n\n else:\n msg = \"Operation aborted.\"\n messagebox.showinfo(self.engine.title,msg) \n \n def on_cancel(self, evt=None):\n self.destroy()\n\n def get_values(self,):\n\n return (self.txtCategory.get(),\n self.txtDescription.get(),\n self.enable.get())\n \n def set_values(self,):\n\n self.txtCategory.insert(0, self.selected_category[1])\n self.txtDescription.insert(0, self.selected_category[2])\n self.enable.set(self.selected_category[3])\n\n def on_fields_control(self):\n\n objs = (self.txtCategory,self.txtDescription)\n\n for obj in objs:\n if not obj.get():\n ret = False\n break \n else:\n ret = True\n return ret \n \n","sub_path":"frames/category.py","file_name":"category.py","file_ext":"py","file_size_in_byte":4197,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"449412483","text":"from alpha_vantage.timeseries import TimeSeries\nimport pickle \nimport numpy as np \nimport pandas as pd\nimport time\nimport glob\n\n# ts = TimeSeries(key='2N7U5OBZO5MQT4IL',output_format = 'pandas')\n# # Get json object with the intraday data and another with the call's metadata\n# data, meta_data = ts.get_daily_adjusted('GOOGL', outputsize='compact')\n\n# print(data.columns)\n#print(data['6. volume'])\n\n### Plan ; ^GSPC \n # for every tick in Nasdaq\n # get : [adj_close,volume]\n # [adj_open, volume]\n # save to text file \n\ndef save_NSDAQ():\n # ticks : \n ticks = pd.read_csv('../stock_data/constituents_csv.csv')\n\n donefiles = glob.glob('../stock_data/NASDAQ/*')\n done_ticks = [f.split('/')[-1] for f in donefiles]\n not_working = ['BF.B','CA','CSRA','DPS','EVHC','GGP','LLL','MON','NFX','PX']\n print(len(done_ticks))\n\n for tck in ticks.Symbol.values:\n if tck not in done_ticks and tck not in not_working:\n print(tck)\n ts = TimeSeries(key='2N7U5OBZO5MQT4IL',output_format = 'pandas')\n # Get json object with the intraday data and another with the call's metadata\n data,_ = ts.get_daily_adjusted(tck, outputsize='full')\n tosave = data[['5. adjusted close', '6. volume']]\n np.savetxt('../stock_data/NASDAQ/' + tck,tosave)\n time.sleep(10)\n\nsave_NSDAQ()","sub_path":"Tryout_notebook/get_stoks.py","file_name":"get_stoks.py","file_ext":"py","file_size_in_byte":1374,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"331010880","text":"'''\nCreated on Jan 15, 2013\n\n@author: v-honzha\n'''\n\nclass switch:\n level = 0\n number = 0\n uplink = []\n downlink = []\n def __init__(self, level, number, downlink, uplink):\n self.level = level\n self.number = number\n self.uplink = uplink\n self.downlink = downlink\n\n\n# init switches\nfs01 = [switch(0, 0, range(3), range(12, 15)), \n switch(0, 1, range(3,6), range(15, 18)),\n switch(0, 2, range(6,9), range(18, 21)),\n switch(0, 3, range(9,12), range(21, 24))]\n\nfs02 = [switch(1, 5, range(15, 18), range(21, 24)), \n switch(1, 11, range(12, 15), range(18, 21)),\n switch(1, 16, range(0, 3), range(3, 6)),\n switch(1, 17, range(6, 9), range(9, 12))]\n\nfs03 = [switch(1, 0, range(3), range(12, 15)), \n switch(1, 2, range(6,9), range(18, 21)),\n switch(1, 3, range(9,12), range(21, 24)),]\n\nfs04 = [switch(0, 6, range(3), range(12, 15)), \n switch(0, 7, range(3,6), range(15, 18)),\n switch(0, 8, range(6,9), range(18, 21)),\n switch(0, 9, range(9,12), range(21, 24))]\n\nfs05 = [switch(0, 4, [16, 12, 13], [14, 15, 17]), \n switch(0, 5, range(18,21), range(21, 24)),\n switch(0, 10, range(3), range(6, 9)),\n switch(0, 11, range(3, 6), range(9, 12))]\n\nfs06 = [switch(1, 6, range(3), range(12, 15)), \n switch(1, 7, range(3,6), range(15, 18)),\n switch(1, 8, range(6,9), range(18, 21)),\n switch(1, 9, range(9,12), range(21, 24))]\n\nfs07 = [switch(0, 12, range(3), range(12, 15)), \n switch(0, 13, range(3,6), range(15, 18)),\n switch(0, 14, range(6,9), range(18, 21)),\n switch(0, 15, range(9,12), range(21, 24))]\n\nfs08 = [switch(0, 16, range(3), range(6, 9)), \n switch(0, 17, range(3,6), range(9, 12))]\n\nfs09 = [switch(1, 12, range(3), range(12, 15)), \n switch(1, 13, range(3,6), range(15, 18)),\n switch(1, 14, range(6,9), range(18, 21)),\n switch(1, 15, range(9,12), range(21, 24))]\n\nfs10 = [switch(1, 1, range(3), range(3,6)), \n switch(1, 4, range(6,9), range(9,12)),\n switch(1, 10, range(12,15), range(15,18)),\n switch(2, 3, range(18, 24), [])]\n\ncs02 = [switch(2, 4, range(6), []),\n switch(2, 5, range(6,12), []),\n switch(2, 6, range(12,18), []),\n switch(2, 7, range(18,24), []),\n switch(2, 8, range(24,30), []),\n switch(2, 1, [30,31,32,36,37,38], []),\n switch(2, 2, [33,34,35,39,40,41], []),\n switch(2, 0, range(42,48), []),\n ]\n\nphy_switches = [(fs01, \"fs01\"), (fs02, \"fs02\"), (fs03, \"fs03\"), (fs04, \"fs04\"), (fs05, \"fs05\"), \n (fs06, \"fs06\"), (fs07, \"fs07\"), (fs08, \"fs08\"), (fs09, \"fs09\"), (fs10, \"fs10\")]\n\nserver_mac = \"00:00:01:01:01:01\"\ntor_mac = \"00:00:01:01:01:02\"\nfor phy, phy_name in phy_switches:\n for vrf_id in range(1, len(phy) + 1):\n sw = phy[vrf_id - 1]\n if sw.level == 0: #tor switch\n f = open(\"data/hydrap-tor-{0}-1.soc\".format(sw.number), \"w\")\n \n for l in sw.downlink:\n f.write(\"vlan create {vlan_id} pbm=cpu0,ge{port_id} ubm=ge{port_id}\\n\".\\\n format(vlan_id=l+1024, port_id=l))\n f.write(\"pvlan set ge{port_id} {vlan_id}\\n\".format(port_id=l, vlan_id=l+1024))\n f.write(\"l3 intf add vlan={vlan_id} mac={mac} intf={intf_id} vrf={vrf_id}\\n\".\\\n format(vlan_id=l+1024, mac=tor_mac, intf_id=l+1, vrf_id=vrf_id))\n f.write(\"\\n\")\n \n for l in sw.uplink:\n f.write(\"vlan create {vlan_id} pbm=cpu0,ge{port_id} ubm=ge{port_id}\\n\".\\\n format(vlan_id=l+1024, port_id=l))\n f.write(\"pvlan set ge{port_id} {vlan_id}\\n\".format(port_id=l, vlan_id=l+1024))\n f.write(\"l3 intf add vlan={vlan_id} mac={mac} intf={intf_id} vrf={vrf_id}\\n\".\\\n format(vlan_id=l+1024, mac=tor_mac, intf_id=l+1, vrf_id=vrf_id))\n f.write(\"\\n\")\n \n for dl in sw.downlink:\n f.write(\"l3 egress add mac={mac} intf={intf_id} port=ge{port_id}\\n\".\\\n format(mac=server_mac, intf_id=dl+1, port_id=dl))\n f.write(\"\\n\")\n \n for i in range(len(sw.downlink)):\n dl = sw.downlink[i]\n f.write(\"l3 l3table add vrf={vrf_id} ip={ip} intf={egress_id}\\n\".\\\n format(vrf_id=vrf_id, \\\n ip=\"10.11.{pod_id}.{index}\".format\\\n (pod_id=sw.number, index=i+2), egress_id=(vrf_id-1)*3+100002+i))\n f.close() \n \n","sub_path":"switchConfiguration/Generator.py","file_name":"Generator.py","file_ext":"py","file_size_in_byte":4635,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"43172517","text":"\"\"\"\nCreated on Feb 11, 2015\n\n@author: Jason Bowles\n\"\"\"\nfrom rawdata_emca.runner.config_file_reader import ConfigFileReader\nfrom rawdata_emca import Base_Type\nimport os\nimport sys\nimport traceback\n\nclass RawProcessor(Base_Type):\n \"\"\"\n The raw processor is the entry point into the framework\n \n All directories and parameters that need to be processed should be passed in on instantiation\n \n params['files] should be a list of config files that should be processed, each will be added\n param['file'] gives the option to just add 1 file to be processed\n param['run_date'] optional parameter to define the actual run date of the framework (allow past date processing)\n param['log_file'] lists the file to be used for logging of the framework\n \n this class will then setup logging for the remainder of the run\n \"\"\"\n \n config_files = []\n config_entries = []\n run_date = None\n\n def __init__(self, params):\n \"\"\"\n Constructor\n \"\"\"\n Base_Type.__init__(self)\n self.print_only = False\n if params:\n if 'files' in params:\n self.addall_config_files(params['files'])\n if 'file' in params:\n self.add_config_file(params['file'])\n if 'run_date' in params:\n self.run_date = params['run_date']\n if params.get('print_only',False):\n self.print_only = True\n \n self.force_name = params.get('force',None)\n if 'log_file' in params:\n directory = os.path.dirname(os.path.realpath(params['log_file']))\n if not os.path.exists(directory):\n os.makedirs(directory)\n self.log_file = params['log_file']\n else:\n run_path = os.path.dirname(os.path.realpath(sys.argv[0]))\n directory = os.path.join(run_path,'log')\n if not os.path.exists(directory):\n os.makedirs(directory)\n self.log_file = os.path.join(directory,'rawdata.log')\n \n self.setup_logging()\n self.log_message(\"Kicking Off Raw Data Processor\", log_type='main', status='start', step='root', name='raw_processor')\n \n \n \n def prep_reader(self):\n params = {}\n #params['uuid'] = self.log_uuid\n if self.run_date:\n params['run_date'] = self.run_date\n if self.force_name:\n params['entry_name'] = self.force_name\n reader = ConfigFileReader(params)\n self.log_message(\"Begin Processing Config Files\",status='running')\n for f in self.config_files:\n reader.process_config(f)\n \n return reader\n \n def execute_entries(self):\n final_stat = 0\n try: \n self.log_message(\"About to Prep the Reader\", log_level=self.log_info(),status='start',step='load configs')\n reader = self.prep_reader()\n self.log_message(\"Reader Prepped\", status='complete')\n final_stat = reader.execute_entries(print_order=self.print_only)\n self.log_message('Raw Data Processor has Finished (Number of Failed Entries: {})'.format(final_stat),log_type='main',status='complete',step='root',name='raw_processor',log_level=self.log_debug())\n except Exception as err1:\n step = 'entry loading'\n if self.log_status == 'complete':\n step = 'entry executing'\n self.log_message('Problem while running: ' + str(err1) + ', issue found in: ('+step+')',log_level=self.log_error(),name='raw_processor',status='error')\n traceback.print_exc()\n return final_stat\n \n def execute_single_entry(self, entry_name):\n reader = self.prep_reader()\n return reader.execute_entry(entry_name)\n \n def addall_config_files(self,files):\n if files:\n for f in files:\n self.add_config_file(f)\n \n def add_config_file(self, filename):\n if filename:\n self.config_files.append(filename)","sub_path":"rawdata_emca/runner/raw_processor.py","file_name":"raw_processor.py","file_ext":"py","file_size_in_byte":4090,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"581275615","text":"#Exercise 3\n\n#write a program that prints out all the elements of the list that are less than 5.\n\n#Extras:\n#1. Instead of printing the elements one by one, make a new list that has all the elements less than 5 from this list in it and print out this new list.\n#2. Write this in one line of Python.\n#3. Ask the user for a number and return a list that contains only elements from the original list a that are smaller than that number given by the user.\n\n#create lists\nsourcelist = [1 ,1 ,2 ,3 ,5 ,8 ,13 ,21 ,34 ,55 ,89]\nnewlist = []\n\n#insert element less than 5 from sourcelist to newlist using list comprehension\nnewlist = [element for element in sourcelist if element < 5]\n \nprint(newlist)\n\n#clear newlist\nnewlist.clear()\n\n#insert element less than a number from sourcelist to newlist using list comprehension\nnum = int(input(\"Please insert a number: \"))\n\nnewlist = [element for element in sourcelist if element < num]\n\nprint(\"This is a list of elements that less than \" + str(num) + \" : \" + str(newlist))\n","sub_path":"exercise_3.py","file_name":"exercise_3.py","file_ext":"py","file_size_in_byte":1014,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"512780321","text":"defaultMF_amp = 0.44\ndefaultMF_power = 0.51\ndefaultCSF_amp = 0.50\ndefaultCSF_power = 0.45\ndefaultCSF_max = 3\ndefaultq_power = -0.4\ndefaultq_min = 0.01\ndefault_aMean = 100.0 # log (AU)\ndefault_aSigma = 0.1 # log (AU)\n\n# Eventually we should add in separation properties. (a_mean, a_sigma)\n\nclass MultiplicityUnresolved(object):\n def __init__(self, \n MF_amp=0.44, MF_power=0.51,\n CSF_amp=0.50, CSF_power=0.45, CSF_max=3,\n q_power=-0.4, q_min=0.01):\n \"\"\"\n The properties of stellar companions.\n\n The number of stellar companions, their masses, and separations\n are be described by the following functions:\n\n Multiplicity Fraction -- the number of stellar systems that host \n multiple stars. In other words, the number of primary stars with\n companions. The multiplicity fraction (MF) is typically described\n as: \n B + T + Q + ...\n MF = ---------------------\n S + B + T + Q + ...\n\n where S = single, B = binary, T = triple, Q = quadruple, etc.\n The MF also changes with mass and this dependency can be \n described as a power-law:\n \n `MF(mass) = MF_amp * (mass ** MF_power)`\n\n Companion Star Fraction -- the expected number of companions in\n a multiple system. The companion star fraction (CSF) also \n changes with mass and this dependency can be described as\n a power-law:\n \n `CSF(mass) = CSF_amp * (mass ** CSF_power)`\n\n The companion star fraction is clipped to some maximum\n value, CSF_max. The actual number of companions is drawn \n from a Poisson distribution with an expectation value of CSF.\n\n Mass Ratio (Q) -- The ratio between the companion star \n mass and primary star mass, Q = (m_comp / m_prim ) has\n a probability density function described by a powerlaw:\n\n `P(Q) = Q ** q_power` for q_min <= Q <= 1\n\n Current observations show no significant mass dependence.\n\n Parameters\n ----------\n MF_amp : float\n The amplitude of the power-law describing the Multiplicity \n Fraction as a function of stellar mass. \n\n MF_power : float\n The power of the power-law describing the Multiplicity\n Fraction as a function of stellar mass.\n\n CSF_amp : float\n The amplitude of the power-law describing the companion star \n fraction as a function of stellar mass.\n\n CSF_power : float\n The power of the power-law describing the companion star \n fraction as a function of stellar mass.\n\n CSF_max : float\n The maximum allowed companion star fraction, which is the\n expectation value for the number of companion stars. Given\n a CSF_max = 3, some systems will still have more than 3 \n companions.\n\n q_power : float\n The power of the power-law describing the probability\n density function for the mass ratio.\n\n q_min : float\n The minimum allowed Q value for the probability\n density function of the mass ratio.\n \"\"\"\n \n self.MF_amp = MF_amp\n self.MF_pow = MF_power\n self.CSF_amp = CSF_amp\n self.CSF_pow = CSF_power\n self.CSF_max = CSF_max\n self.q_pow = q_power\n self.q_min = q_min\n\n def multiplicity_fraction(self, mass):\n \"\"\"\n Given a star's mass, determine the probability that the star is in a\n multiple system (multiplicity fraction = MF).\n\n Parameters\n ----------\n mass : float or numpy array\n Mass of primary star.\n\n Return\n ------\n mf : float or numpy array\n Multiplicity Fraction, the fraction of stars at this mass\n that will have one or more companions.\n \"\"\"\n # Multiplicity Fraction\n mf = self.MF_amp * mass ** self.MF_pow\n\n if np.isscalar(mf) and mf > 1:\n mf = 1\n else:\n mf[mf > 1] = 1\n\n return mf\n\n def companion_star_fraction(self, mass):\n \"\"\"\n Given a star's mass, determine the average number of\n companion stars (companion star fraction = CSF).\n\n Parameters\n ----------\n mass : float or numpy array\n Mass of primary star\n\n Return\n ------\n csf : float or numpy array\n Companion Star Fraction, the expected number of companions\n for a star at this mass.\n \"\"\"\n # Companion Star Fraction\n csf = self.CSF_amp * mass ** self.CSF_pow\n \n if np.isscalar(csf) and csf > self.CSF_max:\n csf = self.CSF_max\n else:\n csf[csf > self.CSF_max] = self.CSF_max\n\n return csf\n\n def random_q(self, x):\n \"\"\"\n Generative function for companion mass ratio, equivalent\n to the inverse of the CDF.\n\n `q = m_compnaion / m_primary`\n `P(q) = q ** beta` for q_min <= q <= 1\n\n Parameters\n ----------\n x : float or array_like\n Random number between 0 and 1.\n\n Return\n -------\n q : float or array_like\n companion mass ratio(s)\n \"\"\"\n b = 1.0 + self.q_pow\n q = (x * (1.0 - self.q_min ** b) + self.q_min ** b) ** (1.0 / b)\n\n return q\n\n def random_is_multiple(self, x, MF):\n return x < MF\n\n def random_companion_count(self, x, CSF, MF):\n n_comp = 1 + np.random.poisson((CSF / MF) - 1)\n\n return n_comp\n","sub_path":"popstar/imf/multiplicity.py","file_name":"multiplicity.py","file_ext":"py","file_size_in_byte":5774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"2975570","text":"from mixer.backend.django import mixer\nfrom sports.models import Sport, Sporttype\nfrom django.contrib.auth import get_user_model\nUser = get_user_model()\n\n\ndef test_sporttype_string():\n \"\"\"\n Test Sporttype string\n \"\"\"\n sporttype = mixer.blend(Sporttype)\n assert str(sporttype) == f'{sporttype.title}'\n\n\ndef test_sport_types_added():\n \"\"\"\n Test Sporttype can be added to DB\n \"\"\"\n mixer.blend(Sporttype)\n mixer.blend(Sporttype)\n assert len(Sporttype.objects.all()) == 2\n\n\ndef test_sport_types_added_is_published():\n \"\"\"\n Test Sporttype is not published by default\n \"\"\"\n sporttype = mixer.blend(Sporttype)\n assert sporttype.is_published is False\n\n\ndef test_sport_can_be_added(create_test_user):\n \"\"\"\n Test Sport can be added for Profile\n \"\"\"\n sporttype = mixer.blend(Sporttype)\n user = create_test_user()\n Sport.objects.create(\n profile=user.profile, sporttype=sporttype, level=3)\n\n assert len(Sporttype.objects.all()) == 1\n assert len(User.objects.all()) == 1\n assert len(Sport.objects.all()) == 1\n\n\ndef test_sport_string(create_test_user):\n \"\"\"\n Test Sport string\n \"\"\"\n sporttype = mixer.blend(Sporttype)\n user = create_test_user()\n sport = Sport.objects.create(\n profile=user.profile, sporttype=sporttype, level=3)\n assert str(sport) == f'{sport.profile} - {sport.sporttype} - {sport.level}'\n","sub_path":"backend/sports/tests/test_models.py","file_name":"test_models.py","file_ext":"py","file_size_in_byte":1401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"385606929","text":"\"\"\"\n@@@@@@@@@@@@@@@@@@@@\n@ Peculiar Balance @\n@@@@@@@@@@@@@@@@@@@@\n\nCan we save them? Beta Rabbit is trying to break into a lab that contains the\nonly known zombie cure - but there's an obstacle. The door will only open if a\nchallenge is solved correctly. The future of the zombified rabbit population is\nat stake, so Beta reads the challenge: There is a scale with an object on the\nleft-hand side, whose mass is given in some number of units.\n\nPredictably, the task is to balance the two sides.\nBut there is a catch: You only have this peculiar weight set,\nhaving masses 1, 3, 9, 27, ... units. That is, one for each power of 3.\n\nBeing a brilliant mathematician, Beta Rabbit quickly discovers that any number\nof units of mass can be balanced exactly using this set. To help Beta get into\nthe room, write a method called answer(x), which outputs a list of strings\nrepresenting where the weights should be placed, in order for the two sides to\nbe balanced, assuming that weight on the left has mass x units.\n\nThe first element of the output list should correspond to the 1-unit weight,\nthe second element to the 3-unit weight, and so on. Each string is one of:\n\n\"L\" : put weight on left-hand side\n\"R\" : put weight on right-hand side\n\"-\" : do not use weight\n\nTo ensure that the output is the smallest possible, the last element of the list\nmust not be \"-\". x will always be a positive integer, no larger than 1000000000.\n\"\"\"\n\n# Balanced ternary is a non-standard positional numeral system (a balanced form), useful for comparison logic.\n\ndef inc_one_digit(n):\n if n == 'L':\n return False, '-'\n elif n == '-':\n return False, 'R'\n elif n == 'R':\n return True, 'L'\n\ndef dec_one_digit(n):\n if n == 'L':\n return True, 'R'\n elif n == '-':\n return False, 'L'\n elif n == 'R':\n return False, '-'\n\ndef number(A):\n s = ['-'] # 0\n for _ in xrange(A):\n s = increase(s)\n return s\n\ndef evaluate(A):\n s = 0\n for i, x in enumerate(A):\n s += 3**i * ({'L': -1, '-': 0, 'R': 1}[x])\n return s\n\ndef increase(A):\n B = []\n c = True\n for x in A:\n if c:\n c, x = inc_one_digit(x)\n B.append(x)\n if c:\n B.append('R')\n return B\n\ndef add_one_digit(A, B):\n if A == 'L' and B == 'L':\n return -1, 'R'\n elif A == 'L' and B == '-':\n return 0, 'L'\n elif A == 'L' and B == 'R':\n return 0, '-'\n elif A == '-' and B == 'L':\n return 0, 'L'\n elif A == '-' and B == '-':\n return 0, '-'\n elif A == '-' and B == 'R':\n return 0, 'R'\n elif A == 'R' and B == 'L':\n return 0, '-'\n elif A == 'R' and B == '-':\n return 0, 'R'\n elif A == 'R' and B == 'R':\n return 1, 'L'\n\ndef trim(A):\n for i in range(len(A)):\n if A[len(A)-i-1] != '-':\n return A[:len(A)-i]\n return A\n\ndef add(A, B):\n l = max(len(A), len(B)) + 1\n A = list(A) + (['-' for _ in xrange(l-len(A))])\n B = list(B) + (['-' for _ in xrange(l-len(B))])\n C = ['-' for _ in xrange(l)]\n c = 0\n for x in xrange(l):\n c, v = add_one_digit(A[x], B[x])\n cc, v = add_one_digit(v, C[x])\n c += cc\n C[x] = v\n if c < 0:\n while c < 0:\n x += 1\n cc, v = dec_one_digit(C[x])\n c += 1\n C[x] = v\n if cc: c -= 1\n elif c > 0:\n while c > 0:\n x += 1\n cc, v = inc_one_digit(C[x])\n c -= 1\n C[x] = v\n if cc: c += 1\n return trim(C)\n\ndef ten_to_the_nth(x):\n A = number(1)\n B = ['-']\n for _ in xrange(x):\n for _ in xrange(10):\n B = add(A, B)\n A = B\n B = ['-']\n return A\n\ndef answer(x):\n n = 0\n S = ['-']\n while x:\n A = x % 10\n ten = ten_to_the_nth(n)\n for _ in xrange(A):\n S = add(S, ten)\n x -= A\n x = x / 10\n n += 1\n return S\n","sub_path":"combinatorics/peculiar_balance.py","file_name":"peculiar_balance.py","file_ext":"py","file_size_in_byte":4001,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"139959319","text":"import os\nimport csv\nimport click\nimport random\nfrom enum import Enum\n\nfrom app import app_settings\n\nclass Singleton:\n\tdef __init__(self, cls):\n\t\tself._cls = cls\n\n\tdef get_instance(self):\n\t\ttry:\n\t\t\treturn self._instance\n\t\texcept AttributeError:\n\t\t\tself._instance = self._cls()\n\n\t\treturn self._instance\n\n\tdef __call__(self):\n\t\traise TypeError('Singletons must be accessed through `get_instance()`.')\n\n\tdef __instancecheck__(self, inst):\n\t\treturn isinstance(inst, self._cls)\n\n\ndef get_profiles_paths():\n\treturn [\n os.path.join(app_settings.FIREFOX_PROFILES_PATH, d) for d in os.listdir(app_settings.FIREFOX_PROFILES_PATH)\n if os.path.isdir(os.path.join(app_settings.FIREFOX_PROFILES_PATH, d))\n and 'default' not in d\n ]\n\n\nclass Profile:\n\tdef __init__(self, email, password, proxy=None):\n\t\tself.email = email\n\t\tself.password = password\n\t\tself.proxy = proxy\n\n\tdef __repr__(self):\n\t\treturn f''\n\nclass List:\n\tdef __init__(self):\n\t\tself.profiles = None\n\n\nclass Proxy:\n\tdef __init__(self, ip, port=app_settings.DEFAULT_PORT):\n\t\tself.ip = ip\n\t\tself.port = port\n\n\tdef __repr__(self):\n\t\treturn f''\n\n\nclass Actions(Enum):\n\tSPAM_REPORT_ALL_TO_INBOX\t\t= 0\n\tINBOX_SELECT_ALL_MARK_AS_READ\t= 1\n\tSPAM_SELECT_ALL_MARK_AS_READ\t= 2\n\tINBOX_ARCHIVE_ALL\t\t\t\t= 3\n\tINBOX_OPEN_MESSAGES\t\t\t\t= 4\n\tSPAM_OPEN_MESSAGES\t\t\t\t= 5\n\tINBOX_OPEN_PLUS_CLICK_MESSAGES\t= 6\n\n\ndef load_profiles_from_csv():\n\tprofiles_list = []\n\twith open(app_settings.ACCOUNTS_FILE) as f:\n\t\treader = csv.DictReader(f)\n\t\tfor line in reader:\n\t\t\t# create profile, proxy.\n\t\t\tproxy = None\n\t\t\tif line['proxy']:\n\t\t\t\tparts = line['proxy'].split(':')\n\t\t\t\tif len(parts) == 2:\n\t\t\t\t\tip, port = parts\n\t\t\t\t\tproxy = Proxy(ip, port)\n\t\t\t\telse:\n\t\t\t\t\tproxy = Proxy(parts[0])\n\n\t\t\tif proxy:\n\t\t\t\tprofile = Profile(line['email'], line['password'], proxy)\n\t\t\telse:\n\t\t\t\tprofile = Profile(line['email'], line['password'])\n\t\t\tprofiles_list.append(profile)\n\treturn profiles_list\n\n\ndef show_introduction():\n\tintroduction = f\"\"\"\n+----------------------------------------------------------------------------+\n| |\n| ██╗ ██╗ █████╗ ██╗ ██╗ ██████╗ ██████╗ ██████╗ ██████╗ ████████╗ |\n| ╚██╗ ██╔╝██╔══██╗██║ ██║██╔═══██╗██╔═══██╗ ██╔══██╗██╔═══██╗╚══██╔══╝ |\n| ╚████╔╝ ███████║███████║██║ ██║██║ ██║ ██████╔╝██║ ██║ ██║ |\n| ╚██╔╝ ██╔══██║██╔══██║██║ ██║██║ ██║ ██╔══██╗██║ ██║ ██║ |\n| ██║ ██║ ██║██║ ██║╚██████╔╝╚██████╔╝ ██████╔╝╚██████╔╝ ██║ |\n| ╚═╝ ╚═╝ ╚═╝╚═╝ ╚═╝ ╚═════╝ ╚═════╝ ╚═════╝ ╚═════╝ ╚═╝ |\n| |\n| |\n| {app_settings.APP_NAME} v{app_settings.APP_VERSION} Developed By: {app_settings.DEVELOPED_BY} |\n| |\n| Powred By: Omega Capital. Contact: {app_settings.CONTACT_ME} |\n| |\n+----------------------------------------------------------------------------+\n\n\"\"\"\n\tclick.secho(introduction, fg='bright_black')\n\n\ndef show_actions_list():\n\tactions = get_available_actions()\n\th1 = ' Yahoo actions list:'\n\tclick.secho(h1, fg='cyan')\n\tclick.secho('-' * len(h1), fg='cyan')\n\n\tfor i, action in enumerate(actions, 1):\n\t\taction_string_parts = action.name.split('_')\n\t\taction_pos = action_string_parts[0]\n\t\taction_name = '_'.join(action_string_parts[1:])\n\t\tprint(f' {i} - ({action_pos}) {action_name}')\n\n\ndef get_action():\n\n\tactions = get_available_actions()\n\tshow_actions_list()\n\ttry:\n\t\tclick.secho('\\nPlease choose the action you want ? (Ctrl+C to exit): ', fg='yellow', nl=False)\n\t\tnum = int(input())\n\t\twhile num < 1 or num > len(actions):\n\t\t\tclick.secho('Opps!, Action not found.', fg='red')\n\t\t\tclick.secho('Please choose an action from the list above ? (Ctrl+C to exit): ', fg='yellow', nl=False)\n\t\t\tnum = int(input())\n\texcept KeyboardInterrupt:\n\t\texit()\n\treturn actions[num - 1]\n\n\ndef get_actions():\n\tactions = get_available_actions()\n\tresult = []\n\tshow_actions_list()\n\ttry:\n\t\tclick.secho('\\nPlease choose actions you want ? (Ctrl+C to exit): ', fg='yellow', nl=False)\n\t\tactions_string = [ int(n) for n in str(input()).split(' ')]\n\t\tfor num in actions_string:\n\t\t\twhile num < 1 or num > len(actions):\n\t\t\t\tclick.secho(f'Opps!, Action {num} not found.', fg='red')\n\t\t\t\texit()\n\t\t\t\t# click.secho('Please choose actions from the list above ? (Ctrl+C to exit): ', fg='yellow', nl=False)\n\t\t\t\t# num = int(input())\n\t\t\tresult.append(actions[num - 1])\n\texcept KeyboardInterrupt:\n\t\texit()\n\treturn result\n\n\ndef get_available_actions():\n\tfrom app import Yahoo\n\treturn Yahoo.get_available_actions()\n\n\ndef get_mailbox_messages_range(total_messages):\n\t\"\"\" Get the max and min messages in the inbox \"\"\"\n\tx_min = min(total_messages, app_settings.MESSAGES_MIN_OPEN)\n\tx_max = min(total_messages, app_settings.MESSAGES_MAX_OPEN)\n\treturn x_min, x_max\n\ndef get_amount_of_message(total_messages):\n\t\"\"\"Get the amount of messages to open.\"\"\"\n\treturn random.randint(*get_mailbox_messages_range(total_messages))\n\n\ndef prettify_seconds(sec):\n\tif sec < 60: \t\t# less than 1 minute.\n\t\treturn '%.2f seconds' % sec\n\telif sec / 60 < 60: # less than 1 hour.\n\t\treturn '%.2f minutes' % (sec / 60)\n\telif sec / 3600 < 24: # less than 1 day.\n\t\treturn '%.2f hours' % (sec / 3600)\n\n\t# return in days.\n\treturn '%.2f days' % (sec / 86400)\n\n","sub_path":"app/common.py","file_name":"common.py","file_ext":"py","file_size_in_byte":6217,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"12267034","text":"# coding: utf-8\nimport requests\nimport json\nfrom re import match, IGNORECASE\n\nMESSAGES_API_URL = \"https://api.flowdock.com/messages\"\n\nEMAIL = r\"[a-z0-9!#$%&'*+/=?^_`{|}~-]+(?:\\.[a-z0-9!#$%&'*+/=?^_`{|}~-]+)*@(?:[a-z0-9](?:[a-z0-9-]*[a-z0-9])?\\.)+[a-z0-9](?:[a-z0-9-]*[a-z0-9])?\"\nALPHANUMERIC_UNDERSCORES_WHITESPACE = r'^[a-z0-9_ ]+$'\n\n\n# \"flow_token\": \"cbc4d9ca8000ad74058724084f929fff\",\n# \"event\": \"activity\",\n# \"author\": {\n# \"name\": \"Marty\",\n# \"avatar\": \"https://avatars.githubusercontent.com/u/3017123?v=3\"\n# },\n# \"title\": \"updated ticket\",\n# \"external_thread_id\": \"1234567\",\n# \"thread\": {\n# \"title\": \"Polish the flux capacitor\",\n# \"body\": \"The flux capacitor has been in storage for more than 30 years and it needs to be spick and span for the re-launch.\",\n# \"status\": {\n# \"color\": \"green\",\n# \"value\": \"open\"\n# }\n# }\n\n\nclass MessageAPI(object):\n\tAPI_URL = MESSAGES_API_URL\n\n\tdef __init__(self, flow_token, author, thread_title, thread_body, thread_id, avatar = None):\n\t\tself.msg = {'flow_token':flow_token}\n\t\tself.msg['event'] = 'activity'\n\n\t\tauthor = {'name':author, 'avatar' :avatar}\n\t\tself.msg['author'] = author\n\n\t\tself.msg['external_thread_id'] = thread_id\n\t\tthread = {'title':thread_title, 'body':thread_body}\n\t\tself.msg['thread'] = thread\n\n\tdef __repr__(self):\n\t\treturn \"%s(%s) instance at %s\" % (self.__class__.__name__, self.flow_api_token, hex(id(self)))\n\n\tdef post(self, title):\n\t\tnew_msg = self.msg.copy()\n\t\tnew_msg.update({'title':title})\n\n\t\tprint(json.dumps(new_msg))\n\t\tprint(self.API_URL)\n\t\t\n\t\theaders = {'Content-type': 'application/json', 'Accept': 'text/plain'}\n\n\t\tresponse = requests.post(self.API_URL, data=json.dumps(new_msg), headers = headers,)\n\t\tif not response.ok:\n\t\t\tresponse.raise_for_status()\n\t\treturn True\n\n\nif __name__ == \"__main__\":\n\t# Initialize tasks and results object\n\tinbox = MessageAPI('cbc4d9ca8000ad74058724084f929fff', author='luca', avatar = \"https://avatars.githubusercontent.com/u/3017123?v=3\",\n\t thread_title = 'FTP file download', \n\t\tthread_body='Status of file download', thread_id='jfkdjskafjadskf')\n# With required params only\n\tprint(inbox.post('File updated'))","sub_path":"flowdock/messages.py","file_name":"messages.py","file_ext":"py","file_size_in_byte":2163,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"509188066","text":"\"\"\"\n\nCOLOR PALLETTE - Using k-means clustering to pick out the color palette from a given image\nProcess: \n(1) Read in image\n(2) Resize image\n(3) Pick out RGB values -> convert to LAB values\n(4) PLT - Axes3D Plot LAB values\n(5) Run clusters thru k-means clustering\n(6) Print out gradient of colors\n\n\"\"\"\n\nimport matplotlib.pyplot as plt\n\nfrom mpl_toolkits.mplot3d import Axes3D\nfrom skimage import io, color\n\n\ndef processImage(filename: str) -> list:\n image = io.imread(filename)\n return color.rgb2lab(image)\n\n\ndef plotClusters(filename: str):\n lab = processImage(filename)\n print(lab)\n fig = plt.figure()\n ax = Axes3D(fig)\n ax.scatter(lab[0].flatten(), lab[1].flatten(), lab[2].flatten())\n plt.show()\n\n\ndef findKMeans():\n pass\n\n\ndef createPallete():\n plotClusters(\"test.jpg\")\n findKMeans()\n\n\ndef main():\n try:\n createPallete()\n except Exception as e:\n print(e)\n\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"k-means.py","file_name":"k-means.py","file_ext":"py","file_size_in_byte":954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"55431041","text":"import random\nimport cv2\nimport Augmentor\nimport numpy as np\nimport os, shutil\n\nfrom distutils.dir_util import copy_tree\n\n\ndef unet_augment(sample,vertical_prob,horizontal_prob):\n image, mask= sample['image'], sample['mask']\n\n if (random.random() AUGMENT_DIR 로 복사\n\n path, dirs, _ = list(os.walk(aug_dir))[0]\n for folder in dirs:\n print('folder :', folder)\n augmnt = Augmentor.Pipeline(\n '{}/{}'.format(path, folder), output_directory='', save_format='bmp')\n\n # Add some operations to an existing pipeline.\n augmnt.flip_left_right(probability=0.4)\n # Now we add a vertical flip operation to the pipeline:\n augmnt.flip_top_bottom(probability=0.8)\n # Add a rotate90 operation to the pipeline:\n # augmnt.rotate90(probability=0.5)\n \n # number of images to augment\n augmnt.sample(aug_num)\n\n return 'Done!'\n","sub_path":"U-Net/augment.py","file_name":"augment.py","file_ext":"py","file_size_in_byte":1574,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"394157059","text":"# Constants mapping to the Item table.\nModelNumber = 'ModelNumber'\nbstockCategory = 'bstockCategory'\nbstockSubcategory = 'bstockSubCategory'\nBrand = 'Brand'\nMSRP = 'MSRP'\nMAP = 'MAP'\nWeight = 'Weight'\nShipWeight = 'ShipWeight'\nImageLocation = 'ImageLocation'\nImageMime = 'ImageMime'\n\n# Constants mapping to the Marketplace table.\nidMarketplace = 'idMarketplace'\nmktpName = 'mktpName'\n\n# Constants mapping to the ItemDescription table.\nSKU = 'SKU'\nUPC = 'UPC'\nDescription = 'Description'\nRetailPrice = 'RetailPrice' \nmktCategory = 'mktCategory'\nmktSubcategory = 'mktSubcategory'\n","sub_path":"util/db/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":736,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"318296694","text":"from concurrent.futures import ThreadPoolExecutor\nfrom gettext import gettext as _\nfrom logging import getLogger\nfrom queue import Queue, Empty\nfrom threading import Thread, current_thread\n\nLOG = getLogger(__name__)\n\n\n# The default queuing backlog.\nBACKLOG = 1024\n# The default concurrency.\nCONCURRENT = 10\n\n\nclass Batch:\n \"\"\"\n Provides batching and concurrent execution of download requests.\n\n Attributes:\n requests (generator): A generator of requests.\n concurrent (int): The number of requests to execute in concurrently.\n iterator (RequestIterator): Used to iterate requests as they complete.\n feeder (RequestFeeder): Used to feed submit requests to the executor.\n scratchpad (dict): A scratchpad for sharing information with requests.\n\n Notes:\n * The batch should be used as a context manager.\n * Or, `shutdown()` must be called manually.\n\n Examples:\n\n >>>\n >>> from pulp3.download import HttpRequest\n >>>\n >>> url = 'http://content.org/dog.rpm'\n >>> destination = '/tmp/working/dog.rpm'\n >>> requests = (HttpRequest(url, destination) for _ in range(10))\n >>>\n >>> # Using context manager (highly recommended)\n >>>\n >>> with Batch(requests) as batch:\n >>> for request in batch.download():\n >>> if request.succeeded():\n >>> # Use the downloaded file \\o/\n >>> else:\n >>> # Log something\n >>>\n >>> # Not using context manager\n >>>\n >>> batch = Batch(requests)\n >>> try:\n >>> for request in batch.download():\n >>> if request.succeeded():\n >>> # Use the downloaded file \\o/\n >>> else:\n >>> # Log something\n >>> finally:\n >>> batch.shutdown()\n >>>\n\n \"\"\"\n\n def __init__(self, requests, concurrent=CONCURRENT, backlog=BACKLOG, **scratchpad):\n \"\"\"\n Args:\n requests (generator): A generator of requests.\n concurrent (int): The number of requests to execute in concurrently.\n backlog (int): The number of download requests kept in memory.\n **scratchpad (dict): A scratchpad for sharing information with requests.\n\n \"\"\"\n super(Batch, self).__init__()\n self.requests = requests\n self.concurrent = concurrent\n self.iterator = RequestIterator(backlog)\n self.feeder = RequestFeeder(self)\n self.scratchpad = scratchpad\n if concurrent < 2:\n raise ValueError(_('concurrent may not be < 2'))\n if backlog < concurrent:\n raise ValueError(_('backlog may not be < concurrent'))\n\n def download(self):\n \"\"\"\n Download the batch by executing all of the requests.\n\n Returns:\n RequestIterator: A request iterator.\n The iterator will render the requests in the order completed.\n\n \"\"\"\n LOG.debug(_('%(batch)s - download started'), {'batch': self})\n self.feeder.start()\n return self.iterator\n\n def shutdown(self):\n \"\"\"\n End processing and shutdown the feeder and the thread pool.\n\n Notes:\n This must be called to prevent leaking resources unless the Batch\n is used as a context manager.\n >>>\n >>> with Batch(..) as batch:\n >>> # ...\n\n \"\"\"\n LOG.debug(_('%(batch)s - shutdown'), {'batch': self})\n self.feeder.abort()\n\n def abort(self):\n \"\"\"\n Abort downloading and shutdown.\n \"\"\"\n LOG.debug(_('%(batch)s - download aborted'), {'batch': self})\n self.shutdown()\n\n def __enter__(self):\n return self\n\n def __exit__(self, *unused):\n self.shutdown()\n\n def __str__(self):\n _id = str(id(self))[-4:]\n return _('Batch: id={s} concurrent={c}').format(s=_id, c=self.concurrent)\n\n\nclass BatchExecutor(ThreadPoolExecutor):\n \"\"\"\n Batch thread pool executor.\n \"\"\"\n\n def __init__(self, concurrent=CONCURRENT):\n \"\"\"\n A thread pool executor tailored for the batch.\n The worker queue size is restricted to limit memory usage and to expedite shutdown.\n The shutdown() method is overridden to work properly with a queue\n with limited capacity.\n\n Args:\n concurrent (int): The number of requests to execute in concurrently.\n\n \"\"\"\n super(BatchExecutor, self).__init__(max_workers=concurrent)\n self._work_queue = Queue(maxsize=concurrent)\n\n def drain(self):\n \"\"\"\n Drain the input queue.\n \"\"\"\n try:\n self._work_queue.get(block=False)\n except Empty:\n pass\n\n def shutdown(self, wait=True):\n \"\"\"\n Shutdown the executor.\n Mainly needed to compensate for ThreadPoolExecutor.shutdown() that expects\n an unrestricted queue capacity. Specifically, the terminate signal (None)\n sent to the workers; is read; and blindly put back into the queue for the next\n worker. This results in a (None) message queued in the worker queue after all\n workers have terminate. The queue is drained to prevent the interpreter shutdown\n hook from blocking.\n\n Args:\n wait (bool): Wait for all workers to terminate.\n\n \"\"\"\n if self._shutdown:\n return\n super(BatchExecutor, self).shutdown(wait)\n self.drain()\n\n\nclass RequestFeeder(Thread):\n \"\"\"\n Request feeder.\n A thread used to feed each batched request into the executor.\n May be interrupted and terminated by calling abort().\n\n Attributes:\n batch (Batch): A batch to feed.\n aborted (bool): Feeding has been aborted.\n\n \"\"\"\n\n def __init__(self, batch):\n super(RequestFeeder, self).__init__(name='feeder')\n self.batch = batch\n self.aborted = False\n self.setDaemon(True)\n\n def abort(self, wait=True):\n \"\"\"\n Abort feeding and terminate.\n\n Args:\n wait (bool): Wait for thread to terminate.\n\n \"\"\"\n self.aborted = True\n if wait:\n self.join()\n\n def run(self):\n \"\"\"\n Thread (main) loop.\n Submit each request to the batch executor.\n \"\"\"\n n = 0\n iterator = self.batch.iterator\n requests = self.batch.requests\n scratchpad = self.batch.scratchpad\n with BatchExecutor(self.batch.concurrent) as executor:\n try:\n for request in requests:\n if self.aborted:\n iterator.abort()\n return\n LOG.debug(\n _('%(feeder)s - feed #%(n)d url=%(url)s'),\n {\n 'feeder': self,\n 'n': n,\n 'url': request.url\n })\n scratchpad.update(request.scratchpad)\n request.scratchpad = scratchpad\n future = executor.submit(BatchRequest(request))\n future.add_done_callback(iterator.add)\n n += 1\n except Exception as e:\n iterator.raised(e)\n n += 1\n if n:\n iterator.total = n\n else:\n iterator.empty()\n\n def __str__(self):\n _id = str(id(self))[-4:]\n return _('RequestFeeder: id={s} aborted={a}').format(s=_id, a=self.aborted)\n\n\nclass QueueIterator:\n \"\"\"\n A Queue iterator.\n Each item in the queue is a tuple of: (code, payload).\n\n Attributes:\n queue (Queue): The input queue to be iterated.\n iterated (int): The number of times `__next__()` was called.\n total (int): The total number queued. A value of `-1` indicates\n the total is not yet known.\n\n \"\"\"\n\n NEXT = 'NEXT'\n EXCEPTION = 'EXCEPTION'\n END = 'END'\n\n def __init__(self, backlog=BACKLOG):\n self.queue = Queue(maxsize=backlog)\n self.aborted = False\n self.iterated = 0\n self.total = -1\n\n def abort(self):\n \"\"\"\n Abort iteration.\n Set the abort flag and drain the queue just in case something\n is blocked on put().\n \"\"\"\n if self.aborted:\n return\n self.aborted = True\n self.empty()\n\n def put(self, code, payload=None, block=True):\n \"\"\"\n Enqueue a message.\n\n Args:\n code (str): The message code.\n payload (object): The message payload.\n block (bool): Block when queue is full (default:True).\n\n \"\"\"\n LOG.debug(\n _('%(iterator)s put: code=%(code)s payload=%(payload)s'),\n {\n 'iterator': self,\n 'code': code,\n 'payload': payload\n })\n message = (code, payload)\n self.queue.put(message, block=block)\n\n def add(self, payload):\n \"\"\"\n Add the next object to the input queue to be rendered by `__next__()`.\n\n Args:\n payload: An object to be rendered by `__next__()`.\n\n \"\"\"\n if self.aborted:\n return\n self.put(self.NEXT, payload)\n\n def raised(self, exception):\n \"\"\"\n Add a fatal exception to the input queue. The exception has been raised by\n the object providing the objects to be iterated.\n\n Args:\n exception: An exception to be raised by `__next__()`.\n\n \"\"\"\n if self.aborted:\n return\n self.put(self.EXCEPTION, exception)\n\n def drain(self):\n \"\"\":\n Drain the input queue.\n \"\"\"\n LOG.debug(_('%(iterator)s - input drained'), {'iterator': self})\n while True:\n try:\n self.queue.get(block=False)\n except Empty:\n break\n self.end()\n\n def empty(self):\n \"\"\"\n Add an message to the input queue that signals that the input\n queue will always be empty. The object feeding the queue has nothing\n to be iterated.\n \"\"\"\n self.drain()\n self.end()\n\n def end(self):\n \"\"\"\n Add an message to the input queue that marks the end of input.\n \"\"\"\n self.put(self.END)\n\n def __next__(self):\n \"\"\"\n Get the next enqueued object.\n\n Returns:\n The next enqueued object.\n\n Raises:\n StopIteration: when finished iterating.\n\n \"\"\"\n LOG.debug(_('%(iterator)s - next'), {'iterator': self})\n\n if self.aborted:\n raise StopIteration()\n if self.iterated == self.total:\n raise StopIteration()\n\n code, payload = self.queue.get()\n\n LOG.debug(\n _('%(iterator)s next: code=%(code)s payload=%(payload)s'),\n {\n 'iterator': self,\n 'code': code,\n 'payload': payload\n })\n\n # next\n if code == self.NEXT:\n self.iterated += 1\n return payload\n # fatal\n if code == self.EXCEPTION:\n raise payload\n # empty\n if code == self.END:\n raise StopIteration()\n\n def __iter__(self):\n return self\n\n def __str__(self):\n _id = str(id(self))[-4:]\n description = 'Iterator: id={s} iterated={i}/{t} aborted={a}'\n return _(description.format(\n s=_id,\n i=self.iterated,\n t=self.total,\n a=self.aborted))\n\n\nclass FutureIterator(QueueIterator):\n \"\"\"\n A queue iterator that expects the payload to be a `concurrent.futures.Future`.\n \"\"\"\n\n def __next__(self):\n \"\"\"\n Get the next future and propagate any raised exceptions.\n\n Returns:\n The next `Future.result()`\n\n Raises:\n Anything raised by the object executed.\n\n \"\"\"\n future = super(FutureIterator, self).__next__()\n exception = future.exception()\n if exception:\n LOG.debug(\n _('%(iterator)s - raising: %(exception)s'),\n {\n 'iterator': self,\n 'exception': exception\n })\n raise exception\n else:\n return future.result()\n\n\nclass RequestIterator(FutureIterator):\n \"\"\"\n Provided for semantic clarity.\n \"\"\"\n pass\n\n\nclass BatchRequest:\n \"\"\"\n Batched request wrapper.\n Ensures the request is returned after being executed in the pool.\n\n Attributes:\n request (Request): The wrapped request.\n\n \"\"\"\n\n def __init__(self, request):\n \"\"\"\n\n Args:\n request (Request): The wrapped request.\n\n Returns:\n request (Request): The wrapped request.\n\n Raises:\n Whatever is raised by request()\n\n \"\"\"\n self.request = request\n\n def __call__(self):\n thread = current_thread()\n LOG.debug(_(\n '%(request)s thread=%(thread)s - started'),\n {\n 'thread': thread.getName(),\n 'request': self\n })\n\n try:\n self.request()\n return self.request\n finally:\n LOG.debug(_(\n '%(request)s thread=%(thread)s - end'),\n {\n 'thread': thread.getName(),\n 'request': self\n })\n\n def __str__(self):\n return str(self.request)\n","sub_path":"pulp3/download/batch.py","file_name":"batch.py","file_ext":"py","file_size_in_byte":13618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"405129152","text":"'''\nCreated on 12/06/2014\n\n:author: alfred\n'''\nfrom unittest import TestCase\n\nfrom dirty_models.fields import StringIdField\nfrom dirty_validators.complex import ModelValidate\nfrom werkzeug.exceptions import NotFound\n\nfrom mc_be.commons.fields import ModelIdField\nfrom mc_be.commons.filter_list import FilterList\nfrom mc_be.commons.models.models import PersistentModel\nfrom mc_be.commons.streams.iterator import ListIterator\nfrom mc_be.commons.validators import Unique, RefModelId, EqualField, NotEqualField\nfrom tests import BaseTestCase, mock_session_org_type\n\n\nclass FakeModel(PersistentModel):\n label = StringIdField()\n second_id = StringIdField()\n model_ref = ModelIdField()\n field_1 = StringIdField()\n field_2 = StringIdField()\n\n def get_resource_id(self):\n return 'fake'\n\n\nclass FakeService:\n\n def __init__(self):\n self.results = [FakeModel(id='aaa', second_id='bbb', label='ccc')]\n self.result_load = FakeModel(id='aaa', second_id='bbb', label='ccc')\n\n def list_all(self, filter_list=None, paging=None):\n self.filter_list = filter_list\n\n return ListIterator(self.results)\n\n def load(self, id):\n if self.result_load:\n return self.result_load\n raise NotFound(\"Not found\")\n\n def build_filter_list(self, params):\n self.params = params\n return FilterList()\n\n\nclass UniqueTest(BaseTestCase, TestCase):\n\n def setUp(self):\n super(UniqueTest, self).setUp()\n self.service = FakeService()\n\n def test_success(self):\n class FakeValidator(ModelValidate):\n label = Unique(filter_name='label', service=self.service)\n\n self.service.results = []\n\n model = FakeModel(label='aaa')\n\n validator = FakeValidator()\n\n self.assertTrue(validator.is_valid(model), validator.messages)\n self.assertDictEqual(self.service.params, {'label': 'aaa'})\n\n def test_fail(self):\n class FakeValidator(ModelValidate):\n label = Unique(filter_name='label', service=self.service)\n\n model = FakeModel(label='ccc')\n\n validator = FakeValidator()\n\n self.assertFalse(validator.is_valid(model), validator.messages)\n self.assertDictEqual(validator.messages, {'label': {'alreadyUsed': \"'ccc' is already used by item 'aaa'\"}})\n\n def test_success_same_id(self):\n class FakeValidator(ModelValidate):\n label = Unique(filter_name='label', service=self.service)\n\n model = FakeModel(label='ccc', id='aaa')\n\n validator = FakeValidator()\n\n self.assertTrue(validator.is_valid(model), validator.messages)\n self.assertDictEqual(self.service.params, {'label': 'ccc'})\n\n def test_fail_multi_id(self):\n class FakeValidator(ModelValidate):\n label = Unique(filter_name='label', service=self.service, identifiers=['id', 'second_id'])\n\n model = FakeModel(label='ccc', id='aaa')\n\n validator = FakeValidator()\n\n self.assertFalse(validator.is_valid(model), validator.messages)\n self.assertDictEqual(validator.messages, {'label': {'alreadyUsed': \"'ccc' is already used by item 'aaa:bbb'\"}})\n\n def test_success_multi_id(self):\n class FakeValidator(ModelValidate):\n label = Unique(filter_name='label', service=self.service, identifiers=['id', 'second_id'])\n\n model = FakeModel(label='ccc', second_id='bbb', id='aaa')\n\n validator = FakeValidator()\n\n self.assertTrue(validator.is_valid(model), validator.messages)\n self.assertDictEqual(self.service.params, {'label': 'ccc'})\n\n\nclass RefModelIdTest(BaseTestCase, TestCase):\n\n def setUp(self):\n super(RefModelIdTest, self).setUp()\n self.service = FakeService()\n\n def test_success(self):\n validator = RefModelId(service=self.service)\n\n self.assertTrue(validator.is_valid('aaaa'))\n\n def test_fail_not_found(self):\n validator = RefModelId(service=self.service)\n self.service.result_load = False\n\n self.assertFalse(validator.is_valid('aaaa'))\n self.assertDictEqual(validator.messages, {'notFound': \"Item 'aaaa' not found\"})\n\n def test_success_permission_not_modified(self):\n validator = RefModelId(service=self.service, check_permission='read')\n\n self.assertTrue(validator.is_valid('aaaa', is_modified=False))\n\n @mock_session_org_type('provider')\n def test_success_permission_modified(self):\n validator = RefModelId(service=self.service, check_permission='read')\n\n with self.app.test_request_context():\n self.app.preprocess_request()\n self.assertTrue(validator.is_valid('aaaa', is_modified=True))\n\n @mock_session_org_type('provider')\n def test_fail_permission_modified(self):\n validator = RefModelId(service=self.service, check_permission='shake')\n\n with self.app.test_request_context():\n self.app.preprocess_request()\n self.assertFalse(validator.is_valid('aaaa', is_modified=True))\n self.assertDictEqual(validator.messages,\n {'notAllowed': \"Item 'aaaa' is not allowed by permission 'shake'\"})\n\n\nclass EqualFieldTest(TestCase):\n\n def test_success(self):\n class FakeValidator(ModelValidate):\n field_1 = EqualField(field_name='field_2')\n model = FakeModel(field_1='aaa', field_2='aaa')\n\n validator = FakeValidator()\n\n self.assertTrue(validator.is_valid(model), validator.messages)\n\n def test_fail(self):\n class FakeValidator(ModelValidate):\n field_1 = EqualField(field_name='field_2')\n model = FakeModel(field_1='aaa', field_2='bbb')\n\n validator = FakeValidator()\n\n self.assertFalse(validator.is_valid(model), validator.messages)\n self.assertDictEqual(validator.messages,\n {'field_1': {'notEqualFieldValue': \"Item value: aaa must be equal to field_2 value: bbb\"}})\n\n\nclass NotEqualFieldTest(TestCase):\n\n def test_success(self):\n class FakeValidator(ModelValidate):\n field_1 = NotEqualField(field_name='field_2')\n model = FakeModel(field_1='aaa', field_2='bbb')\n\n validator = FakeValidator()\n\n self.assertTrue(validator.is_valid(model), validator.messages)\n\n def test_fail(self):\n class FakeValidator(ModelValidate):\n field_1 = NotEqualField(field_name='field_2')\n model = FakeModel(field_1='aaa', field_2='aaa')\n\n validator = FakeValidator()\n\n self.assertFalse(validator.is_valid(model), validator.messages)\n self.assertDictEqual(validator.messages,\n {'field_1': {'equalFieldValue': \"Item value: aaa can not be equal to field_2 value: aaa\"}})\n","sub_path":"mc-pybe-release-smip-R4/tests/commons/validators/tests_validators.py","file_name":"tests_validators.py","file_ext":"py","file_size_in_byte":6733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"219467320","text":"import requests, json\nfrom multiprocessing import Process, Manager\nfrom process import extract_repo_fields\n\n\ndef list_repos(token, since):\n res = requests.get(\n f'https://api.github.com/repositories?since={since}',\n headers={'Authorization': f'token {token}'}\n )\n return json.loads(res.content.decode('utf-8'))\n\n\ndef get_repo_details(owner, name, token):\n repo_details = json.loads(requests.get(\n f'https://api.github.com/repos/{owner}/{name}',\n headers={'Authorization': f'token {token}'}\n ).content.decode('utf-8'))\n return repo_details\n\n\ndef get_repo_details_parallel(repos_details, owner, name, token):\n repo_details = json.loads(requests.get(\n f'https://api.github.com/repos/{owner}/{name}',\n headers={'Authorization': f'token {token}'}\n ).content.decode('utf-8'))\n repos_details.append(repo_details)\n\n\ndef get_repos(token, repos):\n repos_details = []\n for repo in repos:\n try:\n owner = repo['owner']['login']\n name = repo['name']\n repo_details = get_repo_details(owner, name, token)\n repo_fields = extract_repo_fields(repo_details)\n repos_details.append(repo_fields)\n except Exception as e:\n print(f'Failed extracting information for {repo[\"id\"]}')\n pass\n return repos_details\n\n\ndef get_repos_parallel(token, repos):\n with Manager() as manager:\n repos_details = manager.list()\n repo_get_processes = []\n for repo in repos:\n owner = repo['owner']['login']\n name = repo['name']\n process = Process(target=get_repo_details_parallel, args=(repos_details, owner, name, token))\n process.start()\n repo_get_processes.append(process)\n for p in repo_get_processes:\n p.join()\n\n repo_field_details = []\n for repo_details in repos_details:\n try:\n repo_fields = extract_repo_fields(repo_details)\n repo_field_details.append(repo_fields)\n except:\n raise\n return repo_field_details\n\n\ndef list_and_get_repo_details(token, since):\n repos = list_repos(token, since)\n repos_details = get_repos(token, repos)\n # repos_details = get_repos_parallel(token, repos)\n return repos_details\n\n\ndef get_rate_remaining(token):\n res = requests.get(\n f'https://api.github.com/rate_limit',\n headers={'Authorization': f'token {token}'}\n )\n rate_details = json.loads(res.content.decode('utf-8'))\n return rate_details['rate']['remaining']","sub_path":"ghrepos-loader/read_ghrepos_serial.py","file_name":"read_ghrepos_serial.py","file_ext":"py","file_size_in_byte":2591,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"580855124","text":"from timeit import default_timer as time\nimport os\nimport sys\nimport logging\nfrom datetime import datetime\nfrom argparse import Namespace\nimport optuna\nimport pandas as pd\nimport numpy as np\nimport torch\nimport chemprop.utils\n\nfrom absl import app\nfrom absl import flags\n\nimport rdkit\nprint(rdkit.__version__)\n\nDELQSAR_ROOT = os.path.abspath(__file__ + '/../../')\nsys.path += [os.path.dirname(DELQSAR_ROOT)]\n\nRESULTS_FILE = os.path.join(DELQSAR_ROOT, 'experiments', 'all_results.csv')\n\nfrom del_qsar import featurizers, splitters, models, losses\nfrom del_qsar.enrichments import R_from_z, R_ranges\n\n\nFLAGS = flags.FLAGS\nflags.DEFINE_string('csv', 'triazine_lib_sEH_SIRT2_QSAR.csv', 'csv filename')\nflags.DEFINE_list('exp', ['sEH_[strep]_tot'], 'Column header(s) for data counts: experiment')\nflags.DEFINE_list('beads', ['beads-linker-only_[strep]_tot'], 'Column header(s) for data counts: beads')\nflags.DEFINE_enum('featurizer', 'graph', ['graph'], 'How molecules are featurized')\nflags.DEFINE_list('splitter', ['random'], 'How molecules are split for training/testing') \nflags.DEFINE_integer('seed', 0, 'Random seed for data splitting and weight initialization')\n\nflags.DEFINE_enum('model_type', 'MoleculeModel', ['MoleculeModel'], 'Model type')\nflags.DEFINE_enum('task_type', 'regression', ['regression', 'classification'], 'Task type')\n\nflags.DEFINE_enum('threshold_type', 'percentile', ['percentile', 'fixed'], \n 'Threshold type (for classification only)')\nflags.DEFINE_float('threshold_val', 99.99, 'Threshold value; exact value or percentile (for classification only)')\n\nflags.DEFINE_enum('loss_fn_train', 'nlogprob', ['nlogprob', 'MSE', 'BCE'], \n 'Loss function during training (note: classifiers automatically use BCE)')\n\nflags.DEFINE_integer('depth', 6, 'Number of message-passing steps')\nflags.DEFINE_integer('hidden_size', 1500, 'Size of hidden layers in MPN')\nflags.DEFINE_integer('ffn_num_layers', 2, 'Number of layers in FFN after MPN encoding')\n\nflags.DEFINE_float('lr', None, 'Initial learning rate')\nflags.DEFINE_integer('max_epochs', 100, 'Maximum number of epochs')\nflags.DEFINE_integer('patience', 5, 'Patience')\nflags.DEFINE_float('max_norm', 5, 'Max norm')\nflags.DEFINE_float('dropout', 0, 'Dropout rate')\nflags.DEFINE_integer('num_workers', 40, 'Number of workers for loading data')\n\nflags.DEFINE_string('out', None, 'Experiment label (subfolder)')\n\nflags.DEFINE_string('device', 'cuda:0', 'Device (set to cuda:0 if using GPU)')\n\nflags.DEFINE_integer('n_trials', 1, 'Number of optuna trials to run')\n\nif not os.path.isdir(os.path.join(DELQSAR_ROOT, 'experiments', 'results')):\n os.mkdir(os.path.join(DELQSAR_ROOT, 'experiments', 'results'))\n \ndt = datetime.today()\nDATE = os.path.join(DELQSAR_ROOT, 'experiments', 'results', \n f'{dt.year}-{str(dt.month).zfill(2)}-{str(dt.day).zfill(2)}')\nif not os.path.isdir(DATE):\n os.mkdir(DATE)\n\n \ndef objective(trial):\n start = time() \n \n INIT_LR = FLAGS.lr\n DROPOUT = FLAGS.dropout\n DEPTH = FLAGS.depth\n HIDDEN_SIZE = FLAGS.hidden_size\n FFN_NUM_LAYERS = FLAGS.ffn_num_layers\n\n SAVE_ROOT = os.path.join(DATE, FLAGS.out)\n if not os.path.isdir(SAVE_ROOT):\n os.mkdir(SAVE_ROOT)\n \n SAVE_SUBFOLDER = os.path.join(SAVE_ROOT, '1_log_files')\n if not os.path.isdir(SAVE_SUBFOLDER):\n os.mkdir(SAVE_SUBFOLDER)\n \n LOG_FILE = os.path.join(SAVE_SUBFOLDER, 'run.log')\n \n logging.basicConfig(\n level=logging.DEBUG,\n format='%(asctime)s %(levelname)s: %(message)s',\n handlers=[\n logging.StreamHandler(sys.stdout),\n ]\n )\n with open(LOG_FILE, 'a') as lf:\n logging.info(f'Trial #{trial.number}')\n lf.write(f'{datetime.now()} INFO: Trial #{trial.number}\\n')\n logging.info(f'Parameters: {trial.params}')\n lf.write(f'{datetime.now()} INFO: Parameters: {trial.params}\\n\\n')\n logging.info('FLAGS:')\n lf.write(f'{datetime.now()} INFO: FLAGS:\\n')\n for f in FLAGS.get_key_flags_for_module(sys.argv[0]):\n logging.info(f.serialize())\n lf.write(f'{datetime.now()} INFO: ' + f.serialize() + '\\n')\n # GPU?\n logging.info(f'CUDA available? {torch.cuda.is_available()}')\n lf.write(f'{datetime.now()} INFO: CUDA available? {torch.cuda.is_available()}\\n')\n\n # Get data\n df_data = pd.read_csv(os.path.join(DELQSAR_ROOT, 'experiments', 'datasets', FLAGS.csv))\n if 'triazine' in FLAGS.csv:\n for col in df_data.columns:\n if ' ' in col:\n df_data = df_data.rename(columns={col: col.replace(' ', '_')})\n\n # Extract counts\n exp_counts = np.array(df_data[FLAGS.exp], dtype='int')\n bead_counts = np.array(df_data[FLAGS.beads], dtype='int')\n exp_tot = np.sum(exp_counts, axis=0) # column sums\n bead_tot = np.sum(bead_counts, axis=0)\n logging.info(f'{len(df_data)} total compounds')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: {len(df_data)} total compounds\\n')\n\n # Featurizer\n smis = df_data['smiles']\n targets = []\n for i in range(exp_counts.shape[0]):\n targets_for_compound = []\n for j in range(exp_counts.shape[1]): \n targets_for_compound.append(R_from_z(bead_counts[i, j], bead_tot[j], \n exp_counts[i, j], exp_tot[j], 0).tolist())\n targets.append(targets_for_compound)\n featurizer = featurizers.GraphFeaturizer(smis, targets)\n x = featurizer.prepare_x()\n\n # Define different splits\n if FLAGS.splitter[0] == 'random':\n splitter = splitters.RandomSplitter()\n train_slice, valid_slice, test_slice = splitter(x, df_data, seed=FLAGS.seed)\n print('Random split:')\n print(f'Train: {train_slice}')\n print(f'Valid: {valid_slice}')\n print(f'Test: {test_slice}')\n with open(LOG_FILE, 'a') as lf:\n lf.write('\\nRandom split:\\n')\n lf.write(f'Train: {train_slice}\\n')\n lf.write(f'Valid: {valid_slice}\\n')\n lf.write(f'Test: {test_slice}\\n\\n')\n else:\n if len(FLAGS.splitter) == 1:\n splitter = splitters.OneCycleSplitter(FLAGS.splitter, LOG_FILE)\n elif len(FLAGS.splitter) == 2:\n splitter = splitters.TwoCycleSplitter(FLAGS.splitter, LOG_FILE)\n elif len(FLAGS.splitter) == 3:\n splitter = splitters.ThreeCycleSplitter(FLAGS.splitter, LOG_FILE)\n else:\n raise ValueError('Unknown splitter')\n train_slice, valid_slice, test_slice = splitter(x, df_data, seed=FLAGS.seed)\n\n logging.info(f'Train length: {len(train_slice):12d}')\n logging.info(f'Valid length: {len(valid_slice):12d}')\n logging.info(f'Test length: {len(test_slice):12d}')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: Train length: {len(train_slice):12d}\\n')\n lf.write(f'{datetime.now()} INFO: Valid length: {len(valid_slice):12d}\\n')\n lf.write(f'{datetime.now()} INFO: Test length: {len(test_slice):12d}\\n')\n\n # For binary classification: get ground truth labels\n if FLAGS.task_type == 'classification':\n R, R_lb, R_ub = R_ranges(bead_counts[:, 0], bead_tot[0], exp_counts[:, 0], exp_tot[0])\n if FLAGS.threshold_type == 'percentile':\n threshold_R = np.percentile(R[train_slice], FLAGS.threshold_val)\n else:\n threshold_R = FLAGS.threshold_val\n logging.info(f'Threshold (R): {threshold_R}')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: Threshold (R): {threshold_R}\\n')\n true_labels_R = torch.zeros((len(R),))\n for idx in range(len(R)):\n if R[idx] > threshold_R:\n true_labels_R[idx] = 1\n else:\n true_labels_R = None \n \n # Build model and train\n BATCH_SIZE = 50\n NUM_TASKS = len(FLAGS.exp)\n \n torch.manual_seed(FLAGS.seed)\n if FLAGS.task_type == 'classification':\n model = models.MoleculeModel(init_lr = INIT_LR, max_lr = 10*INIT_LR,\n dataset_type = 'classification',\n final_lr = INIT_LR, depth = DEPTH,\n hidden_size = HIDDEN_SIZE, \n ffn_num_layers = FFN_NUM_LAYERS,\n dropout = DROPOUT, num_tasks = NUM_TASKS,\n device = FLAGS.device, torch_seed=FLAGS.seed)\n\n model.train_args = Namespace(\n dataset_type = 'classification',\n epochs = 30,\n warmup_epochs = 2.0,\n train_data_size = len(train_slice),\n batch_size = BATCH_SIZE,\n init_lr = INIT_LR,\n max_lr = 10*INIT_LR,\n final_lr = INIT_LR,\n depth = DEPTH,\n hidden_size = HIDDEN_SIZE, \n ffn_num_layers = FFN_NUM_LAYERS,\n dropout = DROPOUT,\n num_lrs = 1\n )\n else:\n model = models.MoleculeModel(init_lr = INIT_LR, max_lr = 10*INIT_LR,\n final_lr = INIT_LR, depth = DEPTH,\n hidden_size = HIDDEN_SIZE, \n ffn_num_layers = FFN_NUM_LAYERS,\n dropout = DROPOUT, num_tasks = NUM_TASKS,\n device = FLAGS.device, torch_seed=FLAGS.seed)\n\n model.train_args = Namespace(\n dataset_type = 'regression',\n epochs = 30,\n warmup_epochs = 2.0,\n train_data_size = len(train_slice),\n batch_size = BATCH_SIZE,\n init_lr = INIT_LR,\n max_lr = 10*INIT_LR,\n final_lr = INIT_LR,\n depth = DEPTH,\n hidden_size = HIDDEN_SIZE, \n ffn_num_layers = FFN_NUM_LAYERS,\n dropout = DROPOUT,\n num_lrs = 1,\n )\n model.optimizer = chemprop.utils.build_optimizer(model, model.train_args)\n model.scheduler = chemprop.utils.build_lr_scheduler(model.optimizer, model.train_args)\n \n # Loss function during training\n if FLAGS.task_type == 'classification':\n model.loss_fn_train = losses.loss_fn_BCE\n elif FLAGS.loss_fn_train == 'nlogprob':\n model.loss_fn_train = losses.loss_fn_nlogprob\n elif FLAGS.loss_fn_train == 'MSE': \n model.loss_fn_train = losses.loss_fn_MSE\n else:\n raise ValueError('Unknown loss function for training') \n\n # Loss function during evaluation (only used for regression models)\n model.loss_fn_eval = losses.loss_fn_nlogprob\n \n logging.info(str(model))\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: ' + str(model) + '\\n')\n # Move to GPU?\n model.to(torch.device(FLAGS.device))\n logging.info(f'Moved model to {FLAGS.device}')\n lf.write(f'{datetime.now()} INFO: Moved model to {FLAGS.device}\\n')\n\n try:\n model.train_on_del(\n x, exp_counts, bead_counts, train_slice, valid_slice, true_labels=true_labels_R,\n batch_size=BATCH_SIZE, max_epochs=FLAGS.max_epochs, \n num_workers=FLAGS.num_workers,\n patience=FLAGS.patience, max_norm=FLAGS.max_norm,\n zscale=lambda epoch: 1 + 9*np.exp(-epoch/2),\n device=FLAGS.device, output_size=NUM_TASKS,\n save_path=os.path.join(SAVE_ROOT, f'best_model_trial_{trial.number}.torch'),\n log_path=LOG_FILE,\n torch_seed=FLAGS.seed,\n )\n except KeyboardInterrupt:\n logging.warning('Training interrupted by KeyboardInterrupt!')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} WARNING: Training interrupted by KeyboardInterrupt!\\n')\n \n # Check for NaN loss\n if model.best_val_loss == float('inf'):\n return model.best_val_loss\n \n # Record loss functions\n with open(os.path.join(SAVE_ROOT, f'losses_trial_{trial.number}.csv'), 'w') as fid:\n fid.write('epoch,train_loss,valid_loss\\n')\n for i in range(len(model.all_valid_losses)):\n fid.write(f'{i+1},{np.squeeze(model.all_train_losses[i])}, {np.squeeze(model.all_valid_losses[i])}\\n')\n train_losses_row_sums = np.sum(model.all_train_losses, axis=1)\n valid_losses_row_sums = np.sum(model.all_valid_losses, axis=1)\n\n # Evaluations\n def evaluate_on_slice(test_slice, slice_label='Test'):\n if model.classification:\n test_roc_auc, test_pr_auc, test_preds = model.evaluate_on_del(\n x, exp_counts, bead_counts, test_slice, batch_size=BATCH_SIZE,\n device=FLAGS.device, true_labels=true_labels_R, num_workers=FLAGS.num_workers,\n )\n \n formatted_roc_auc = ['{0:.5f}'.format(roc_auc) for roc_auc in np.squeeze(test_roc_auc, axis=0)]\n formatted_pr_auc = ['{0:.5f}'.format(pr_auc) for pr_auc in np.squeeze(test_pr_auc, axis=0)]\n logging.info(f'{slice_label} ({len(test_preds)} compounds) ROC AUC = {np.squeeze(formatted_roc_auc)}')\n logging.info(f'{slice_label} ({len(test_preds)} compounds) PR AUC = {np.squeeze(formatted_pr_auc)}')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: {slice_label} ({len(test_preds)} compounds) ROC AUC = {np.squeeze(formatted_roc_auc)}\\n')\n lf.write(f'{datetime.now()} INFO: {slice_label} ({len(test_preds)} compounds) PR AUC = {np.squeeze(formatted_pr_auc)}\\n')\n return test_roc_auc, test_pr_auc, test_preds \n else:\n test_losses, test_enrichments = model.evaluate_on_del(\n x, exp_counts, bead_counts, test_slice, batch_size=BATCH_SIZE,\n num_workers=FLAGS.num_workers, device=FLAGS.device,\n )\n\n avg_test_loss = np.sum(test_losses, axis=0) / test_losses.shape[0]\n formatted_loss = ['{0:.5f}'.format(loss) for loss in avg_test_loss]\n logging.info(f'{slice_label} ({len(test_enrichments)} compounds) average loss = {np.squeeze(formatted_loss)}')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: {slice_label} ({len(test_enrichments)} compounds) average loss = {np.squeeze(formatted_loss)}\\n')\n\n R = np.zeros((NUM_TASKS, len(test_slice)))\n R_lb = np.zeros((NUM_TASKS, len(test_slice)))\n R_ub = np.zeros((NUM_TASKS, len(test_slice)))\n for i in range(NUM_TASKS):\n _R, _R_lb, _R_ub = R_ranges(bead_counts[test_slice, i], bead_tot[i], \n exp_counts[test_slice, i], exp_tot[i])\n R[i] = _R\n R_lb[i] = _R_lb\n R_ub[i] = _R_ub\n\n test_loss_by_POI = []\n test_enrichment_by_POI = []\n for col in test_losses.T:\n test_loss_by_POI.append(col.tolist())\n for col in test_enrichments.T:\n test_enrichment_by_POI.append(col.tolist())\n test_enrichment_by_POI = np.array(test_enrichment_by_POI)\n\n accuracy_vals = np.zeros((NUM_TASKS, 3))\n for i in range(NUM_TASKS):\n low = np.mean(test_enrichment_by_POI[i] < R_lb[i])\n high = np.mean(test_enrichment_by_POI[i] > R_ub[i])\n within = np.mean((test_enrichment_by_POI[i] < R_ub[i]) & (test_enrichment_by_POI[i] > R_lb[i]))\n accuracy_vals[i] = [low, high, within]\n\n # Report fraction of enrichments within (LB, UB)\n for i, POI in enumerate(FLAGS.exp):\n logging.info(f'{POI}')\n logging.info(f'% predicted R < calculated R: {accuracy_vals[i, 0]:.3%}')\n logging.info(f'% predicted R > calculated R: {accuracy_vals[i, 1]:.3%}')\n logging.info(f'% predicted R in (R_lb, R_ub): {accuracy_vals[i, 2]:.3%}')\n with open(LOG_FILE, 'a') as lf:\n for i, POI in enumerate(FLAGS.exp):\n lf.write(f'{datetime.now()} INFO: {POI}\\n')\n lf.write(f'{datetime.now()} INFO: % predicted R < calculated R: {accuracy_vals[i, 0]:.3%}\\n')\n lf.write(f'{datetime.now()} INFO: % predicted R > calculated R: {accuracy_vals[i, 1]:.3%}\\n')\n lf.write(f'{datetime.now()} INFO: % predicted R in (R_lb, R_ub): {accuracy_vals[i, 2]:.3%}\\n\\n')\n\n return test_losses, test_enrichments, avg_test_loss, accuracy_vals\n \n # Evaluate on part of train slice\n try:\n evaluate_on_slice(train_slice[:5000], slice_label='Train subset')\n except ValueError as ve:\n logging.info(str(ve))\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} WARNING: {str(ve)}\\n')\n \n # Evaluate on test set\n if model.classification:\n test_roc_auc, test_pr_auc, test_preds = evaluate_on_slice(test_slice, slice_label='Test set')\n else:\n _losses, _enrichments, avg_test_loss, accuracy_vals = evaluate_on_slice(test_slice, slice_label='Test set')\n\n # Record to results file\n if not os.path.isfile(RESULTS_FILE):\n with open(RESULTS_FILE, 'w') as fid:\n cols = ['Time (of recording)']\n for f in FLAGS.get_key_flags_for_module(sys.argv[0]):\n flag,value = f.serialize().split('=')\n cols.append(flag[2:])\n cols.append('Training losses')\n cols.append('Validation losses')\n cols.append('Test loss')\n cols.append('% predicted R in (R_lb, R_ub)')\n cols.append('Test ROC AUC')\n cols.append('Test PR AUC')\n fid.write('\\t'.join(cols) + '\\n')\n with open(RESULTS_FILE, 'a') as fid:\n cols = [str(datetime.now().now())]\n for f in FLAGS.get_key_flags_for_module(sys.argv[0]):\n flag,value = f.serialize().split('=')\n cols.append(value)\n cols.append(str(list(np.squeeze(model.all_train_losses))))\n cols.append(str(list(np.squeeze(model.all_valid_losses))))\n if model.classification:\n cols.append('')\n cols.append('')\n cols.append(str(np.squeeze(test_roc_auc)))\n cols.append(str(np.squeeze(test_pr_auc)))\n else:\n cols.append(str(np.squeeze(avg_test_loss)))\n acc_vals = accuracy_vals[:, 2]\n formatted_acc_vals = ['{0:.3%}'.format(val) for val in acc_vals]\n print(formatted_acc_vals)\n cols.append(f'{np.squeeze(formatted_acc_vals)}')\n cols.append('')\n cols.append('')\n fid.write('\\t'.join(cols) + '\\n')\n fid.close()\n \n total = time() - start\n m, s = divmod(total, 60)\n h, m = divmod(int(m), 60)\n logging.info(f'Total time for trial #{trial.number}: {h}h {m}m {s:0.2f}s\\n')\n with open(LOG_FILE, 'a') as lf:\n lf.write(f'{datetime.now()} INFO: Total time for trial #{trial.number}: {h}h {m}m {s:0.2f}s\\n\\n')\n lf.close()\n \n return model.best_val_loss\n \ndef main(argv):\n del argv\n study = optuna.create_study()\n study.optimize(objective, n_trials=FLAGS.n_trials)\n\n SAVE_ROOT = os.path.join(DATE, FLAGS.out)\n if not os.path.isdir(SAVE_ROOT):\n os.mkdir(SAVE_ROOT)\n \n SAVE_SUBFOLDER = os.path.join(SAVE_ROOT, '1_log_files')\n if not os.path.isdir(SAVE_SUBFOLDER):\n os.mkdir(SAVE_SUBFOLDER)\n \n LOG_FILE = os.path.join(SAVE_SUBFOLDER, 'run.log')\n TRIALS_LOG = os.path.join(SAVE_SUBFOLDER, 'trials.csv')\n \n trial = study.best_trial\n with open(LOG_FILE, 'a') as lf:\n print('Number of trials: {}'.format(len(study.trials)))\n lf.write(f'\\n{datetime.now()} INFO: ' + 'Number of trials: {}'.format(len(study.trials)) + '\\n')\n \n print(f'Best trial: Trial #{trial.number}')\n lf.write(f'{datetime.now()} INFO: Best trial: Trial #{trial.number}\\n') \n \n print(' Value: {}'.format(trial.value))\n lf.write(f'{datetime.now()} INFO: ' + ' Value: {}'.format(trial.value) + '\\n') \n \n print(' Params: ')\n lf.write(f'{datetime.now()} INFO: ' + ' Params: \\n')\n for key, value in trial.params.items():\n print(' {}: {}'.format(key, value))\n lf.write(f'{datetime.now()} INFO: ' + ' {}: {}'.format(key, value) + '\\n')\n lf.close()\n \n study.trials_dataframe().to_csv(TRIALS_LOG, sep='\\t', index=False)\n \n os.rename(os.path.join(SAVE_ROOT, f'losses_trial_{trial.number}.csv'), \n os.path.join(SAVE_SUBFOLDER, f'losses_trial_{trial.number}.csv'))\n os.rename(os.path.join(SAVE_ROOT, f'best_model_trial_{trial.number}.torch'), \n os.path.join(SAVE_SUBFOLDER, f'best_model_trial_{trial.number}.torch'))\n \nif __name__ == '__main__':\n app.run(main)\n logging.shutdown()\n","sub_path":"experiments/triazine_MPN.py","file_name":"triazine_MPN.py","file_ext":"py","file_size_in_byte":21050,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"161531051","text":"# This is a guess the number game.\r\n\r\n# Import random library\r\nimport random\r\n\r\n# instantiate guessestaken = 0\r\nguessesTaken = 0\r\n\r\n\r\n\r\nprint('Hello! What is your name?\\n')\r\n\r\nmyName = input()\r\n\r\nnumber = random.randint(1, 10)\r\n\r\nprint('Therefore, ' + myName + ', I am thinking of a number between 1 and 10. Guess my number')\r\n\r\n\r\n# checks the number of times of guess taken\r\nwhile guessesTaken < 6:\r\n\r\n print('Take a guess\\n\\t.')\r\n\r\n guess = int(input())\r\n\r\n # add 1 after every guess\r\n guessesTaken = guessesTaken + 1\r\n\r\n #checking the conditions\r\n if guess < number:\r\n\r\n print('Your guess is too low.')\r\n\r\n\r\n\r\n if guess > number:\r\n\r\n print('Your guess is too high.')\r\n\r\n\r\n\r\n if guess == number:\r\n break\r\n\r\n# If user gets answer after first guess\r\nif guess == number:\r\n\r\n guessesTaken = str(guessesTaken)\r\n\r\n print('Great, ' + myName + '! your passed number in ' + guessesTaken + ' guesses!')\r\n\r\n# if user guess wrong answers\r\nif guess != number:\r\n number = str(number)\r\n print('Nope. The number I was thinking of was ' + number)","sub_path":"Day_2/Q3.py","file_name":"Q3.py","file_ext":"py","file_size_in_byte":1088,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"406351819","text":"import cv2\nimport numpy as np\n\nclass WEEDS:\n def __init__(self, imgPath, reSize=(250,250)):\n image = cv2.imread(imgPath)\n self.image = cv2.resize(image, reSize, interpolation = cv2.INTER_AREA)\n self.colorSpace = cv2.COLOR_BGR2LAB\n\n def displayImage(self, title=\"Image Display\", image=None):\n if(image==None):\n cv2.imshow(title, self.image)\n else:\n cv2.imshow(title, image)\n\n def waitClose(self):\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n\n def applyMask(self, r=55):\n maskPlant = np.zeros(self.image.shape[:2], dtype=\"uint8\")\n cv2.circle(maskPlant, (125, 65), r, 255, -1)\n cv2.circle(maskPlant, (225, 75), r, 255, -1)\n cv2.circle(maskPlant, (320, 35), r, 255, -1)\n cv2.circle(maskPlant, (140, 245), r, 255, -1)\n cv2.circle(maskPlant, (255, 245), r, 255, -1)\n cv2.circle(maskPlant, (365, 210), r, 255, -1)\n\n self.image = cv2.bitwise_not(self.image, self.image, mask=maskPlant)\n\n def applyBlur(self, strong=(3,3)):\n self.image = cv2.GaussianBlur(imgB, strong, 0)\n\n def getImage(self):\n return self.image\n\n def extractWeeds_all(self, b_threshold=80, a_threshold=80):\n zeros = np.zeros(self.image.shape[:2], dtype = \"uint8\")\n\n imgLAB = cv2.cvtColor(self.image, self.colorSpace)\n (L, A, B) = cv2.split(imgLAB)\n\n (T_weeds_b, thresh_weeds_b) = cv2.threshold(B, b_threshold, 255, cv2.THRESH_BINARY)\n (T_weeds_a, thresh_weeds_a) = cv2.threshold(A, a_threshold, 255, cv2.THRESH_BINARY)\n imgRGB = cv2.merge([zeros, thresh_weeds_b, thresh_weeds_a])\n self.image = imgRGB\n\n def countPlantArea(self):\n image = self.image\n width = image.shape[1]\n height = image.shape[0]\n greenArea = 0.0\n \n for pixel_w in range(0, width, 1):\n for pixel_h in range(0, height, 1):\n (b, g, r) = image[pixel_h, pixel_w]\n if((b+g+r)>0):\n greenArea += 1 \n\n plantArea = greenArea / (width*height)\n ratio = str(int(plantArea * 100)) + \"%\"\n cv2.putText(image, ratio, (width-150, 50), cv2.FONT_HERSHEY_SIMPLEX, 2, (255, 1, 126), 5)\n self.image = image\n\n def extractWeeds(self, threshold=80, channel=\"B\"):\n zeros = np.zeros(self.image.shape[:2], dtype = \"uint8\")\n\n imgLAB = cv2.cvtColor(self.image, self.colorSpace)\n (L, A, B) = cv2.split(imgLAB)\n\n if(channel==\"B\"):\n selectedChannel = B\n elif(channel==\"L\"):\n selectedChannel = L\n elif(channel==\"A\"):\n selectedChannel = A\n\n (T_weeds, thresh_weeds) = cv2.threshold(selectedChannel, threshold, 255, cv2.THRESH_BINARY)\n imgRGB = cv2.merge([zeros, thresh_weeds, zeros])\n self.image = imgRGB\n #imgGray = cv2.cvtColor(imgRGB, cv2.COLOR_BGR2GRAY)\n","sub_path":"lib/libFarm.py","file_name":"libFarm.py","file_ext":"py","file_size_in_byte":2900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"348090939","text":"from auc.baseusecase import BaseUseCase\nfrom robot.api import logger\nfrom conf.restConstants import *\nimport requests as req\n\n\nclass MiddlewareEnableBackup(BaseUseCase):\n def test_execute_middleware_enable_backup(self):\n \"\"\"\n Function to enable the filesystem backup for a given VM.\n\n Args:\n tenant_id (mandatory): String\ttenant uuid in xstream\n vm_id (mandatory): String vm uuid in xstream\n hostname (mandatory): String hostname of the vm\n retention_days (mandatory): Int no. of retention days\n callback_url (mandatory): String\n Returns:\n Function returns Status Code from the REST API response.\n Following status codes can be returned\n 200: OK\n 404: NOT FOUND\n 401: UNAUTHORIZED\n 500: INTERNAL SERVER ERROR\n \"\"\"\n enable_backup_data = {'TenantID': self.ctx_in['tenant_id'],\n 'VirtualMachineID': self.ctx_in['vm_id'],\n 'VirtualMachineHostName': self.ctx_in['hostname'],\n 'RetentionDays': self.ctx_in['retention_days'],\n 'Callback': self.ctx_in['callback_url']\n }\n try:\n enable_backup_resp = req.post(MIDDLEWARE_ENABLE_BACKUP.format(self.ctx_in['mw_service_host']),\n json=enable_backup_data,\n headers=MIDDLEWARE_SERVICE_HEADER)\n self.ctx_out = True\n except Exception as e:\n logger.info(\"Exception occurred while calling Enable Backup API. \\n \"\n \"Backup enabling did not happen.\\nException: %s\" % e)\n self.ctx_out = False\n return 0\n\n logger.info(\"Status Code = %s\" % enable_backup_resp.status_code)\n logger.info(\"Message = %s\" % enable_backup_resp.text)\n self.ctx_out = enable_backup_resp.status_code\n return enable_backup_resp.status_code\n\n def run_test(self):\n sc = self.test_execute_middleware_enable_backup()\n return sc\n\n def _finalize_context(self):\n assert self.ctx_out == 200, 'Could not enable backup'\n","sub_path":"automation_framework/auc/middleware_enable_backup/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2191,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"468088457","text":"\"\"\"trt_mtcnn.py\n\nThis script demonstrates how to do real-time face detection with\nCython wrapped TensorRT optimized MTCNN engine.\n\"\"\"\n\nimport sys\nimport time\nimport argparse\n\nimport cv2\nfrom utils.camera import add_camera_args, Camera\nfrom utils.display import open_window, set_display, show_fps\nfrom utils.mtcnn import TrtMtcnn\nfrom keras.models import load_model\nimport numpy as np\n\nemotion_dict = {0: 'angry', 1: 'happy', 2: 'neutral', 3: 'sad'}\nsleep_dict = {0: \"awake\", 1: \"drowsy\"}\n\nmodel = load_model(\"./whole_model.hdf5\")\nWINDOW_NAME = 'TrtMtcnnDemo'\nBBOX_COLOR = (0, 255, 0) # green\ncount = 1\nsleep_prev = str(\"awake\")\n\ndef parse_args():\n \"\"\"Parse input arguments.\"\"\"\n desc = ('Capture and display live camera video, while doing '\n 'real-time face detection with TrtMtcnn on Jetson '\n 'Nano')\n parser = argparse.ArgumentParser(description=desc)\n parser = add_camera_args(parser)\n parser.add_argument('--minsize', type=int, default=40,\n help='minsize (in pixels) for detection [40]')\n args = parser.parse_args()\n return args\n\n\ndef show_faces(img, boxes, landmarks):\n \"\"\"Draw bounding boxes and face landmarks on image.\"\"\"\n global sleep_prev\n global count\n img1 = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n for bb, ll in zip(boxes, landmarks):\n x1, y1, x2, y2 = int(bb[0]), int(bb[1]), int(bb[2]), int(bb[3])\n #################################\n roi_color = img1[y1:y2, x1:x2, :]\n cropped_img = np.expand_dims(cv2.resize(roi_color, (224, 224)), 0)\n prediction2 = model.predict(cropped_img)\n sleep = prediction2[0]\n emotion = prediction2[1]\n print(emotion_dict[int(np.argmax(emotion))] , \"\\n\" , sleep_dict[int(np.argmax(sleep))] )\n ################################\n cv2.rectangle(img, (x1, y1), (x2, y2), (206, 174, 17), 1)\n sleep = sleep_dict[int(np.argmax(prediction2[0]))]\n emotion = emotion_dict[int(np.argmax(prediction2[1]))]\n\n if sleep == \"drowsy\" and sleep == sleep_prev:\n count = count+1\n sleep_prev = sleep\n if count > 10:\n cv2.rectangle(img, (x2 + 20, y1+ 40), (x2 + 120, y1 + 55), (28, 13, 191), 13)\n cv2.putText(img, \"ALERT!!\", (x2 + 32, y1 + 54), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (255, 255, 255), 1,\n cv2.LINE_AA)\n else:\n sleep_prev = sleep\n count = 1\n\n cv2.rectangle(img, (x2 + 20, y1 + 5), (x2 + 120, y1 + 20), (206, 174, 17), 13)\n cv2.putText(img, emotion, (x2 + 32, y1 + 18), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (255, 255, 255), 1,\n cv2.LINE_AA)\n\n if sleep == \"awake\":\n cv2.rectangle(img, (x2 + 20, y1 + 40), (x2+ 120, y1 + 55), (206, 174, 17), 13)\n cv2.putText(img, sleep, (x2 + 32, y1 + 53), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (255, 255, 255), 1,\n cv2.LINE_AA)\n elif sleep == \"drowsy\" and count <= 10:\n cv2.rectangle(img, (x2 + 20, y1 + 40), (x2 + 120, y1 + 55), (14, 88, 235), 13)\n cv2.putText(img, sleep, (x2 + 32, y1 + 53), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (255, 255, 255), 1,\n cv2.LINE_AA)\n return img\n\n\ndef loop_and_detect(cam, mtcnn, minsize):\n \"\"\"Continuously capture images from camera and do face detection.\"\"\"\n full_scrn = False\n fps = 0.0\n tic = time.time()\n while True:\n if cv2.getWindowProperty(WINDOW_NAME, 0) < 0:\n break\n img = cam.read()\n if img is not None:\n dets, landmarks = mtcnn.detect(img, minsize=minsize)\n print('{} face(s) found'.format(len(dets)))\n img = show_faces(img, dets, landmarks)\n img = show_fps(img, fps)\n cv2.imshow(WINDOW_NAME, img)\n toc = time.time()\n curr_fps = 1.0 / (toc - tic)\n # calculate an exponentially decaying average of fps number\n fps = curr_fps if fps == 0.0 else (fps*0.95 + curr_fps*0.05)\n tic = toc\n key = cv2.waitKey(1)\n if key == 27: # ESC key: quit program\n break\n elif key == ord('F') or key == ord('f'): # Toggle fullscreen\n full_scrn = not full_scrn\n set_display(WINDOW_NAME, full_scrn)\n\n\ndef main():\n args = parse_args()\n cam = Camera(args)\n cam.open()\n if not cam.is_opened:\n sys.exit('Failed to open camera!')\n\n mtcnn = TrtMtcnn()\n\n cam.start()\n open_window(WINDOW_NAME, args.image_width, args.image_height,\n 'Camera TensorRT MTCNN Demo for Jetson TX2')\n loop_and_detect(cam, mtcnn, args.minsize)\n\n cam.stop()\n cam.release()\n cv2.destroyAllWindows()\n\n del(mtcnn)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Face_Detection_Models/MTCNN/demo_VggFace_mtcnn_jetsontx2.py","file_name":"demo_VggFace_mtcnn_jetsontx2.py","file_ext":"py","file_size_in_byte":4776,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"241517356","text":"from datetime import timedelta\n\nimport airflow\nfrom airflow.models import DAG\nfrom airflow.operators.bash_operator import BashOperator\nfrom airflow.operators.dummy_operator import DummyOperator\nfrom airflow.operators.python_operator import PythonOperator\n\nargs = {\n 'owner': 'Airflow',\n 'start_date': airflow.utils.dates.days_ago(2),\n}\n\ndag = DAG(\n dag_id='dag_2_training_id',\n default_args=args,\n schedule_interval=None,\n dagrun_timeout=timedelta(minutes=60),\n)\n\ndef print_context(execution_date, **context):\n print(execution_date)\n\nPythonOp = PythonOperator(\n task_id='PythonOp',\n provide_context=True,\n python_callable=print_context,\n dag=dag,\n)\n\nBashOp_1 = BashOperator(\n task_id='BashOp_1',\n bash_command='sleep 1',\n dag=dag,\n)\n\nBashOp_2 = BashOperator(\n task_id='BashOp_2',\n bash_command='sleep 5',\n dag=dag,\n)\n\nBashOp_3 = BashOperator(\n task_id='BashOp_3',\n bash_command='sleep 10',\n dag=dag,\n)\n\nDummyOp = DummyOperator(\n task_id='DummyOp',\n dag=dag,\n)\n\nPythonOp >> [BashOp_1,BashOp_2,BashOp_3] >> DummyOp \n","sub_path":"dags/dag_2_training.py","file_name":"dag_2_training.py","file_ext":"py","file_size_in_byte":1080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"353550472","text":"import requests\nimport dryscrape\nfrom bs4 import BeautifulSoup\n\ndef get_drysoup(url):\n\tprint(url)\n\tsession = dryscrape.Session()\n\tsession.set_timeout(60)\n\tsession.visit(url)\n\tpage = session.body()\n\tsoup = BeautifulSoup(page,'html.parser')\n\treturn soup\ndef get_soup(url):\n\tprint(url)\n\t#page = urllib.request.urlopen(url)\n\tpage = requests.get(url)\n\tsoup = BeautifulSoup(page.text,'html.parser')\n\treturn soup\n\t\n\nclass ConsumerAffairs(object):\n\n\tdef __init__(self):\n\t\tself.urls = []\n\n\tdef start_scraping(self, searchterm):\n\t\turl = \"https://www.consumeraffairs.com/search.html?q=\"+searchterm\n\t\tlist = []\n\t\tself.getReviews(url, list)\n\t\tprint(\"****************************************************\",len(list))\n\t\treturn list\t\t\n\n\tdef getReviews(self, url, list):\n\t\tsoup = get_drysoup(url)\n\t\taddr = self.getAddress(soup)\n\t\tres = soup.find(id=\"___gcse_0\")\n\t\t#print(res.findAll(\"a\",{\"class\":\"gs-title\"}))\n\t\tfor item in res.findAll(\"a\",{\"class\":\"gs-title\"}):\n\t\t\td_url = item['href']\n\t\t\tif d_url in self.urls:\n\t\t\t\tcontinue\n\t\t\telse:\t\t\n\t\t\t\tself.urls.append(d_url)\t\n\t\t\t\tdetail = get_soup(d_url)\n\t\t\treviews = detail.findAll(\"div\",{\"class\":\"review--user-post\"})\n\t\t\tprint(len(reviews))\n\t\t\tfor r in reviews:\n\t\t\t\tdict = {}\n\t\t\t\tdict[\"name\"] = r.find(\"span\",{\"class\":\"review__author-name\"}).find(text=True)\n\t\t\t\tdict[\"date\"] = r.find(\"span\",{\"class\":\"review__post-date\"}).find(text=True)\n\t\t\t\tdict[\"title\"] = \"\"\n\t\t\t\tdict[\"content\"] = r.find(\"div\",{\"class\":\"review__body\"}).find('p').find(text=True, recursive=True)\n\t\t\t\tdict[\"address\"] = \"\"\n\t\t\t\tlist.append(dict)\n\t\t\t\t\t\t\n\t\tnextPage = self.getUrl(soup)\n\t\tif nextPage is not None:\n\t\t\tself.getReviews(nextPage, list)\n\t\t\t\t\n\t\tprint(list)\n\t\treturn list\n\n\n\tdef getAddress(self, soup):\n\t\ttry:\n\t\t\treturn \"\\n\".join(soup.find('div',{'itemprop':'address'}).stripped_strings)\n\t\texcept:\n\t\t\treturn \"\"\n\n\tdef getUrl(self, soup):\n\t\tif soup.find(\"div\",{\"class\":\"review--user-post\"}) is not None:\n\t\t\tprint(getUrl)\n\t\t\treturn \"https://www.consumeraffairs.com/\"+soup.find('a',{'class':'next'})['href']\n\t\telse:\n\t\t\treturn None\n\n\n","sub_path":"apps/mentions/scraping/consumeraffairs.py","file_name":"consumeraffairs.py","file_ext":"py","file_size_in_byte":2025,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"34507442","text":"from __future__ import absolute_import\n\nfrom sentry.interfaces.base import Interface\nfrom sentry.interfaces.stacktrace import Stacktrace\nfrom sentry.utils.safe import trim\n\n__all__ = ('Threads',)\n\n\nclass Threads(Interface):\n score = 1900\n\n @classmethod\n def to_python(cls, data):\n threads = []\n\n for thread in data.get('values') or ():\n stacktrace = thread.get('stacktrace')\n if stacktrace is not None:\n stacktrace = Stacktrace.to_python(stacktrace,\n slim_frames=True)\n threads.append({\n 'stacktrace': stacktrace,\n 'id': trim(thread.get('id'), 40),\n 'crashed': bool(thread.get('crashed')),\n 'current': bool(thread.get('current')),\n 'name': trim(thread.get('name'), 200),\n })\n\n return cls(values=threads)\n\n def to_json(self):\n def export_thread(data):\n rv = {\n 'id': data['id'],\n 'current': data['current'],\n 'crashed': data['crashed'],\n 'name': data['name'],\n 'stacktrace': None,\n }\n if data['stacktrace']:\n rv['stacktrace'] = data['stacktrace'].to_json()\n return rv\n\n return {\n 'values': [export_thread(x) for x in self.values],\n }\n\n def get_api_context(self, is_public=False):\n def export_thread(data):\n rv = {\n 'id': data['id'],\n 'current': data['current'],\n 'crashed': data['crashed'],\n 'name': data['name'],\n 'stacktrace': None,\n }\n if data['stacktrace']:\n rv['stacktrace'] = data['stacktrace'].get_api_context(\n is_public=is_public)\n return rv\n\n return {\n 'values': [export_thread(x) for x in self.values],\n }\n\n def get_path(self):\n return 'threads'\n","sub_path":"src/sentry/interfaces/threads.py","file_name":"threads.py","file_ext":"py","file_size_in_byte":2036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"567994059","text":"import networkx as nx\nimport numpy as np\nimport sys\nfrom ba.UtilityClasses import ColType\nimport time\n\nclass GraphClustering:\n\n def __init__(self):\n self.preset_number_of_clusters = None\n self.nearest_neighbour_indices = None\n self.border_points = None\n self.connection_matrix_ = None\n self.sub_connection_matrix_ = None\n self.edges_ = None\n self.graph_ = None\n self.sub_graphs_ = None\n self.graph_clustering_name_ = None\n\n def perform_sub_graph_clustering(self, data_frame, nearest_neighbour_indices, border_points):\n self.nearest_neighbour_indices = nearest_neighbour_indices\n self.border_points = border_points\n self.connection_matrix_ = self._compute_connection_matrix(nearest_neighbour_indices)\n self.sub_connection_matrix_ = self._compute_sub_connection_matrix(border_points, self.connection_matrix_)\n self.edges_ = self._compute_edges(self.sub_connection_matrix_)\n self.graph_ = self._compute_graph(data_frame.df, self.edges_, data_frame.dimensions)\n self.sub_graphs_ = self._compute_sub_graphs(self.graph_)\n\n self.graph_clustering_name_ = self._compute_graph_clustering(self.sub_graphs_, data_frame)\n return self\n\n def perform_sub_graph_clustering_preset_k(self, number_of_clusters, data_frame, nearest_neighbour_indices, border_points_order):\n self.preset_number_of_clusters = number_of_clusters\n self.nearest_neighbour_indices = nearest_neighbour_indices\n self.connection_matrix_ = self._compute_connection_matrix(nearest_neighbour_indices)\n self.edges_ = self._compute_edges(self.connection_matrix_)\n self.graph_ = self._compute_graph(data_frame.df, self.edges_, data_frame.dimensions)\n\n self.graph_ = self._reduce_to_fixed_k(data_frame.df, self.graph_, number_of_clusters, border_points_order)\n self.sub_graphs_ = self._compute_sub_graphs(self.graph_)\n self.graph_clustering_name_ = self._compute_graph_clustering(self.sub_graphs_, data_frame)\n return self\n\n @staticmethod\n def _compute_connection_matrix(nearest_neighbour_indices):\n point_count = nearest_neighbour_indices.shape[0]\n connection_matrix = np.zeros((point_count, point_count))\n for point_array in nearest_neighbour_indices:\n for point in point_array:\n connection_matrix[point_array[0]][point] = 1\n return connection_matrix\n\n @staticmethod\n def _compute_sub_connection_matrix(border_pts, connection_matrix):\n sub_connection_matrix = connection_matrix[:]\n for border_pt in border_pts:\n sub_connection_matrix[border_pt] = np.zeros(connection_matrix[border_pt].shape[0])\n return sub_connection_matrix\n\n @staticmethod\n def _compute_edges(sub_connection_matrix):\n edges = []\n for index, row in enumerate(sub_connection_matrix):\n for idx, point in enumerate(row):\n if point != 0:\n edges.append((index, idx))\n return edges\n\n @staticmethod\n def _compute_graph(df, edges, dimensions):\n g = nx.Graph()\n col_names = list(df.columns.values)[0:dimensions]\n for index, row in df.iterrows():\n col_dict = {}\n for i in range(0, dimensions):\n col_dict[\"a\"+str(i)] = row[col_names[i]]\n g.add_node(index, data=col_dict)\n g.add_edges_from(edges)\n return g\n\n @staticmethod\n def _compute_sub_graphs(graph):\n d = list(nx.connected_component_subgraphs(graph))\n return d\n\n @staticmethod\n def _compute_graph_clustering(sub_graphs, data_frame):\n\n clustering_name = \"graph_clustering \" + str(time.process_time())\n\n data_frame.add_result_name(clustering_name, -2, ColType.CLUSTER_LABEL)\n\n i = 1\n for sub_graph in sub_graphs:\n if len(sub_graph) > 1:\n for node in sub_graph.nodes:\n data_frame.add_result(clustering_name, node, i)\n i += 1\n else:\n for node in sub_graph.nodes:\n data_frame.add_result(clustering_name, node, -1)\n return clustering_name\n\n def _reduce_to_fixed_k(self, df, graph, fixed_number_of_clusters, border_points_order):\n\n sub_clusters = GraphClustering._compute_sub_graphs(graph)\n sub_clusters_without_noise = [x for x in sub_clusters if len(x) > 1]\n current_number_of_clusters = len(sub_clusters_without_noise)\n\n while current_number_of_clusters < fixed_number_of_clusters and len(border_points_order) > 0:\n\n to_remove = border_points_order.pop(0)\n edges_to_remove = [edge for edge in graph.edges if edge[0] == to_remove]\n graph.remove_edges_from(edges_to_remove)\n\n sub_clusters = GraphClustering._compute_sub_graphs(graph)\n sub_clusters_without_noise = [x for x in sub_clusters if len(x) > 1]\n current_number_of_clusters = len(sub_clusters_without_noise)\n\n self.edges_ = graph.edges\n return graph\n","sub_path":"Python Code/ba/GraphSearch.py","file_name":"GraphSearch.py","file_ext":"py","file_size_in_byte":5095,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"467318660","text":"#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Oct 27 14:03:58 2017\n\n@author: sherryluo\n\"\"\"\n\nimport numpy as np\n\nnum_samples = 50\nsample_size = 100\n\ndef create_samples(data_mat):\n\n def draw100(data_mat):\n # randomly draw 100 ratings with replacement\n test_ind = np.random.randint(0, high=data_mat.shape[0], size=sample_size)\n test = data_mat[test_ind]\n \n # the remained ratings serve as prior set\n prior_ind = range(data_mat.shape[0])\n prior_ind = np.setdiff1d(prior_ind, test_ind)\n prior = data_mat[prior_ind]\n return test, prior\n \n tests = []\n for i in range(num_samples):\n test, prior = draw100(data_mat)\n filename = 'prior{}.txt'.format(i)\n with open(filename, 'wb') as f:\n np.savetxt(f, prior.astype(int), fmt='%i', delimiter='\\t')\n tests.append(test)\n \n tests = np.vstack(tests)\n filename2 = 'tests.txt'\n with open(filename2, 'wb') as f2:\n np.savetxt(f2, tests.astype(int), fmt='%i', delimiter='\\t')\n\ndef main():\n datafile = 'ml-100k/u.data'\n u_data = np.loadtxt(datafile, dtype=int)\n u_data = u_data[:, :-1]\n create_samples(u_data)\n \nmain()","sub_path":"collaborative-filtering/sampling.py","file_name":"sampling.py","file_ext":"py","file_size_in_byte":1214,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"357222787","text":"\"\"\"\nA set of request processors that return dictionaries to be merged into a\ntemplate context. Each function takes the request object as its only parameter\nand returns a dictionary to add to the context.\n\nThese are referenced from the setting TEMPLATE_CONTEXT_PROCESSORS and used by\nRequestContext.\n\"\"\"\nfrom settings import SEAFILE_VERSION, SITE_TITLE, SITE_NAME, SITE_BASE, \\\n ENABLE_SIGNUP, MAX_FILE_NAME\ntry:\n from settings import BUSINESS_MODE\nexcept ImportError:\n BUSINESS_MODE = False\n\ntry:\n from settings import ENABLE_FILE_SEARCH\nexcept ImportError:\n ENABLE_FILE_SEARCH = False\n\ndef base(request):\n \"\"\"\n Add seahub base configure to the context.\n \n \"\"\"\n try:\n org = request.user.org\n except AttributeError:\n org = None\n try:\n base_template = request.base_template\n except AttributeError:\n base_template = 'myhome_base.html'\n\n return {\n 'seafile_version': SEAFILE_VERSION,\n 'site_title': SITE_TITLE,\n 'business_mode': BUSINESS_MODE,\n 'cloud_mode': request.cloud_mode,\n 'org': org,\n 'base_template': base_template,\n 'site_name': SITE_NAME,\n 'enable_signup': ENABLE_SIGNUP,\n 'max_file_name': MAX_FILE_NAME,\n 'enable_file_search': ENABLE_FILE_SEARCH,\n }\n\n","sub_path":"base/context_processors.py","file_name":"context_processors.py","file_ext":"py","file_size_in_byte":1306,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"642062645","text":"# def main():\n# numYearBase10 = 2011\n# numYearBase8 = 0o3733\n# numYearBase16 = 0x7DB\n\n# print(\"Year by base 10 : %d, by base 8 : %d, by base 16 : %d\" %(numYearBase10, numYearBase8, numYearBase16))\n\n# numComplex1 = complex(3,4)\n# numComplex2 = 4+3j\n\n# print(\"complex value : \", numComplex1)\n# print(\"Absolute value : \", abs(numComplex2))\n# print(\"Real value : \", numComplex2.real)\n# print(\"Image value : \", numComplex2.imag)\n\n# strDeptName = \"Industrial & Systems Engineering\"\n# strUnivName = \"KAIST\"\n# print(\"Department : \", strDeptName)\n# print(\"Full name or dept : \", (strDeptName+\", \"+strUnivName))\n\n# main()\n\ndef main():\n numTest1 = 10\n numTest2 = 3.0\n numPlus = numTest1 + numTest2\n numMinus = numTest1 - numTest2\n numMultiply = numTest1 * numTest2\n numDivide = numTest1 / numTest2\n numModulo = numTest1 % numTest2\n print(\"{0}, {1}, {2}, {3}, {4}\".format(numPlus, numMinus, numMultiply, numDivide, numModulo))\n \n numDivideInt = numTest1 / int(numTest2)\n print(numDivide, numDivideInt) # in python3 동일한 값이 출력됨\n # integer output을 얻기 위해서는 / 대신에 // 를 사용한다.\n\n numTest2, numTest1 = numTest1, numTest2 # swapping statement\n print(numTest1,numTest2)\n\n print(numTest1 == numTest2)\n print(numTest1 != numTest2)\n print(type(numTest1))\n\n numTest1 = str(numTest1)\n print(type(numTest1), numTest1)\n\n strformula = \"2011 / 7\"\n print(eval(strformula))\n\nmain()","sub_path":"실습/example1-4.py","file_name":"example1-4.py","file_ext":"py","file_size_in_byte":1512,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"251361445","text":"'''\nThis module provides classes to facilitate processing of GriB data using PyGrib\n\nJavier.Delgado@noaa.gov\n'''\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.basemap import Basemap\nimport pygrib\nimport logging as log\n\n\nDEFAULT_CONTOUR_LINE_PLOT_COLOR = 'black'\nDEFAULT_COLOR_MAP = plt.cm.jet\n\nclass GribMessage:\n ''' \n Very light-weight Encapsulation of a GriB message. Currently, it\n just stores it's name, lats, lons, and values. \n The PyGrib grib message is optionally stored as well. This is \n optional since it may take too much memory.\n The class provides methods for plotting the values onto an existing\n Basemap object as a contour or filled contour.\n '''\n def __init__(self, param_name, level, lats, lons, values, plotType, \n grib_message=None, **kwargs):\n self.param_name = param_name # should match parameterName attribute of grib file\n self.level = level\n self.lats = lats\n self.lons = lons\n self.values = values\n self.plot_type = plotType\n if grib_message != None:\n self.grib_message = grib_message\n\n def plot_contour_lines(self, basemap, colors=DEFAULT_CONTOUR_LINE_PLOT_COLOR):\n '''\n Make a contour plot from the GriB message's values onto the given \n Basemap object\n '''\n #small = np.min(self.values)\n #large = np.max(self.values)\n x,y = basemap(self.lons, self.lats)\n # use linspace arg to control how many bins to break the color map into\n #cs = map.contour(x, y, self.values, np.linspace(small,large,11), colors=colors )\n cs = basemap.contour(x, y, self.values, colors=colors ) \n plt.clabel(cs, inline=1, fontsize=10) # put values in contour lines\n\n def plot_contours(self, basemap):\n '''\n Make a filled contour plot from the grib message's values onto the\n given basemap object\n ''' \n x,y = basemap(self.lons, self.lats)\n rangeMin = np.min(self.values)\n rangeMax = np.max(self.values)\n #cs = m.contourf(x,y,self.values, np.linspace(rangeMax,rangeMin,11), cmap=plt.cm.jet )\n cs = basemap.contourf(x,y,self.values, cmap=DEFAULT_COLOR_MAP )\n cb = basemap.colorbar()\n \n","sub_path":"nwpy/dataproc/grib/objects.py","file_name":"objects.py","file_ext":"py","file_size_in_byte":2201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"346733411","text":"# Uses python3\nimport functools\nimport sys\n@functools.lru_cache(None)\ndef fib(n):\n if n<2:\n return n\n return fib(n-1) + fib(n-2)\nif __name__ == '__main__':\n #input = sys.stdin.read()\n n = int(input())\n m= (n+2) %60\n #print((fib(n+2)-1)%10)\n answer = fib(m)%10\n if answer == 0:\n print(answer)\n else:\n print(answer-1)\n #print((fib(m)%10)-1)\n","sub_path":"Algorithmic-ToolBox/week2_algorithmic_warmup/6_last_digit_of_the_sum_of_fibonacci_numbers/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"482352450","text":"import json\nimport re\n\n\ndef read_and_add():\n with open(\"input12.txt\") as f:\n values = f.read()\n r = re.compile(r\"-?\\d+\")\n numbers = r.findall(values)\n sum = 0\n for num in numbers:\n sum += int(num)\n return sum\n\n\ndef read_and_add2():\n with open(\"input12.txt\") as f:\n j = json.load(f)\n\n s = do_sum(j)\n print(s)\n\n\ndef do_sum(arg, ignore_red=False):\n sum_value = 0\n\n if type(arg) is str:\n sum_value += sum_string(arg)\n elif type(arg) is int:\n sum_value += arg\n elif type(arg) is list:\n sum_value += sum([do_sum(x) for x in arg])\n elif type(arg) is dict:\n if ignore_red and \"red\" in arg.values():\n return 0\n else:\n for k, v in arg.items():\n sum_value += do_sum(v)\n\n return sum_value\n\n\ndef sum_string(s):\n c = 0\n try:\n c = int(s)\n except ValueError:\n pass\n\n return c\n","sub_path":"2015/python/day12.py","file_name":"day12.py","file_ext":"py","file_size_in_byte":956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"1525803","text":"# Create a Student class and initialize it with name and roll number. Make methods to :\n# 1. Display - It should display all informations of the student.\n# 2. setAge - It should assign age to student\n# 3. setMarks - It should assign marks to the student.\nfrom datetime import datetime\n\n\nclass Myclass:\n def __init__(self, firstname, lastname, birthday, roll_no):\n self.firstname = firstname\n self.lastname = lastname\n self.birthday = birthday\n self.roll_no = roll_no\n\n def studInfo(self):\n info = {\n \"Fullname\": self.firstname + \" \" + self.lastname,\n \"Birthday\": self.birthday,\n \"Roll Number\": self.roll_no,\n }\n return info\n\n def setAge(self):\n bday = self.birthday.split(\"-\")\n birth = datetime(int(bday[2]), int(bday[0]), int(bday[1]))\n total_days = datetime.today() - birth\n age_obj = str(total_days / 365).split(\" \")\n age = age_obj[0]\n return age + \" \" + \"y/o\"\n\n def setMarks(self, english, math, filipino):\n total = english + math + filipino\n total_mark = total / 3\n return f\"Grade : {total_mark:.1f}\"\n\n\n# x = Myclass('Sheriline', 'Malaca', '03-03-1998', '123421')\n\n# info = x.studInfo()\n# for i in info:\n# print(f'{i} : {info[i]}')\n\n# age = x.setAge()\n# print(f'Age : {age}')\n\n# marks = x.setMarks(80, 85, 89)\n# print(marks)\n\n# Output:\n# Fullname : Sheriline Malaca\n# Birthday : 03-03-1998\n# Roll Number : 123421\n# Age : 21 y/o\n# Grade : 84.7\n\n#######################################################################\n\n# class Derive(Myclass):\n# # print(Myclass('asasasa', 'asasa', '2-2-1997', '2131313'))\n# pass\n\n# print(Myclass('asasasa', 'asasa', '2-2-1997', '2131313').birthday) #2-2-1997\n","sub_path":"D9/class.py","file_name":"class.py","file_ext":"py","file_size_in_byte":1765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"388185650","text":"from twilio.rest import TwilioRestClient\n\naccount_sid = \"Avg900\" # Your Live Credentials Account SID from www.twilio.com/console\nauth_token = \"b567tht\" # Your Live Credentials Auth Token from www.twilio.com/console\n\nclient = TwilioRestClient(account_sid, auth_token)\n\nmessage = client.messages.create(\n body=\"Hello from python.\",\n to=\"verified_number\", # Replace with your phone number\n from_=\"twilio_number\") # Replace with your Twilio number\n\nprint(message.sid)\n","sub_path":"send_text.py","file_name":"send_text.py","file_ext":"py","file_size_in_byte":480,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"79922295","text":"# %load q01_outlier_removal/build.py\n# Default imports\nimport pandas as pd\nimport numpy as np\n\nloan_data = pd.read_csv('data/loan_prediction_uncleaned.csv')\nloan_data = loan_data.drop('Loan_ID', 1)\n\nloan_data\n\ndef outlier_removal(loan_data):\n loan_data1=loan_data.loc[:,['ApplicantIncome','CoapplicantIncome','LoanAmount']]\n loan_data1=loan_data1.dropna()\n loan_data2=loan_data1.sort_values(['ApplicantIncome','CoapplicantIncome','LoanAmount'])\n upper_quartile = np.percentile(loan_data2,95)\n h2=loan_data[loan_data2=': operator.ge,\n '>': operator.gt\n }\n operation = ops.get(op)\n return operation(arg1, arg2)\n\n\ndef create_mask(in_path, band, operator, threshold, write=True, out_dir=None,\n show_output=False,):\n \"\"\"Masking a raster with a given threshold and operatorself.\n\n Creates a a binary output raster in GTiff format. Keeps the input filename\n and adds the suffix '_mask'.\n\n Args:\n in_path (str): full path to input raster for rasterio.open().\n band (int): band number.\n operator (str): select one of '>', '>=', '<', '<=',\n '==', '!='.\n threshold (number): to define the split.\n write (bool): If True, write output to disk.\n out_dir (str): Output directory. If None, outputs will be written along\n input files.\n show_output (bool): If True, plotting the output mask.\n\n Returns:\n mask (ndarray): binary raster [0, 1].\n profile (rasterio profile): to reestablish geographic reference.\n\n Example:\n # files = os.listdir(in_dir)\n # for f in (f for f in files if f.lower().endswith('.tif')\n # create_mask(f, 2, '<', -17, write=True, out_dir='data\\\\output')\n\n TODO:\n * Transform is provided inside profile?\n \"\"\"\n with rio.open(in_path) as src:\n profile = src.profile\n\n try:\n arr = src.read(band)\n except IndexError:\n print(f'WARNING: band index of {in_path} not existing.'\n 'Will be skipped.')\n return None\n\n ras_mask = _cmp(arr, operator, threshold)\n\n if write:\n if out_dir:\n if not os.path.exists(out_dir):\n os.makedirs(out_dir)\n _, fname = os.path.split(in_path)\n out_path = join(out_dir, fname[:-4] + '_mask.tif')\n else:\n out_path = in_path[:-4] + '_mask.tif'\n profile.update(\n dtype=rio.uint8,\n count=1,\n compress='lzw',\n nodata=None)\n with rio.open(out_path, 'w', **profile) as dst:\n dst.write(mask.astype(rio.uint8), 1)\n\n if show_output:\n fig, (ax1, ax2) = plt.subplots(1, 2, figsize=(15, 9))\n show(arr, ax=ax1, cmap='gray', vmin=-25, vmax=0, title='b1')\n show(mask, ax=ax2, cmap='gray', vmin=0, vmax=1, title='mask')\n plt.show()\n\n return ras_mask, profile\n\n\n# if __name__ == '__main__':\n\ngdal.UseExceptions()\n\nhome_dir = 'D:\\\\nw\\\\ws\\\\sarpy'\ndlm_dir = 'D:\\\\nw\\\\04_Geodaten\\\\Basis-DLM\\\\basis-dlm_EPSG25832_Shape'\nshp_path = join(dlm_dir, 'gew01_f.shp')\nshp = gpd.read_file(shp_path)\n\nras_path = join(home_dir, 'Intensity',\n 'Intensity_S1A_IW_SLC__1SDV_20180118T171647_'\n '20180118T171715_020212_0227CA_B637.tif')\nras_path = (\"D:\\\\nw\\\\ws\\\\sarpy\\\\data\\\\\"\n \"Intensity_S1B_IW_SLC__1SDV_20180105T172405_7EE6_clip.tif\")\n\n\n# extract the geometry in GeoJSON format\ngeoms = shp.geometry.values # list of shapely geometries\ngeometry = geoms[0] # shapely geometry\ngeoms = [mapping(geoms[0])]\n# extract the raster values values within the polygon\nwith rasterio.open(ras_path) as src:\n profile = src.profile\n transform = src.transform\n arr = src.read([1, 2, 1])\n\n # rio.plot.show(arr)\n\n # out_image, out_transform = mask(src, geoms, crop=True)\n\nstats = quicklook(ras_path)\n# print(stats)\n# print(arr.shape)\n# rio.plot.show(arr/-25, transform=transform)\n\n\n# shp = shp.to_crs(crs)\n#\n# plt.show()\n","sub_path":"sentools2/dev_sar.py","file_name":"dev_sar.py","file_ext":"py","file_size_in_byte":5688,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"474769152","text":"\"\"\"\nComponents/User Animation Card\n==============================\n\nExample\n-------\n\n.. code-block:: python\n\n from kivymd.app import MDApp\n from kivy.lang import Builder\n from kivy.factory import Factory\n\n from kivymd.toast import toast\n from kivymd.theming import ThemeManager\n from kivymd.uix.useranimationcard import MDUserAnimationCard\n from kivymd.uix.button import MDIconButton\n from kivymd.uix.list import ILeftBodyTouch\n\n # Your content for a contact card.\n Builder.load_string('''\n #:import get_hex_from_color kivy.utils.get_hex_from_color\n\n\n \n orientation: 'vertical'\n padding: dp(10)\n spacing: dp(10)\n adaptive_height: True\n\n MDBoxLayout:\n adaptive_height: True\n\n Widget:\n MDRoundFlatButton:\n text: \"Free call\"\n Widget:\n MDRoundFlatButton:\n text: \"Free message\"\n Widget:\n\n OneLineIconListItem:\n text: \"Video call\"\n IconLeftSampleWidget:\n icon: 'camera-front-variant'\n\n TwoLineIconListItem:\n text: \"Call Viber Out\"\n secondary_text: \"[color=%s]Advantageous rates for calls[/color]\" % get_hex_from_color(app.theme_cls.primary_color)\n IconLeftSampleWidget:\n icon: 'phone'\n\n TwoLineIconListItem:\n text: \"Call over mobile network\"\n secondary_text: \"[color=%s]Operator's tariffs apply[/color]\" % get_hex_from_color(app.theme_cls.primary_color)\n IconLeftSampleWidget:\n icon: 'remote'\n ''')\n\n\n class IconLeftSampleWidget(ILeftBodyTouch, MDIconButton):\n pass\n\n\n class Example(MDApp):\n title = \"Example Animation Card\"\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.user_animation_card = None\n\n def build(self):\n def main_back_callback():\n toast('Close card')\n\n if not self.user_animation_card:\n self.user_animation_card = MDUserAnimationCard(\n user_name=\"Lion Lion\",\n path_to_avatar=\"./assets/african-lion-951778_1280.jpg\",\n callback=main_back_callback)\n self.user_animation_card.box_content.add_widget(\n Factory.TestAnimationCard())\n self.user_animation_card.open()\n\n\n Example().run()\n\"\"\"\n\n\nfrom kivy.animation import Animation\nfrom kivy.clock import Clock\nfrom kivy.core.window import Window\nfrom kivy.lang import Builder\nfrom kivy.metrics import dp, sp\nfrom kivy.properties import ListProperty, ObjectProperty, StringProperty\nfrom kivy.uix.boxlayout import BoxLayout\nfrom kivy.uix.floatlayout import FloatLayout\nfrom kivy.uix.modalview import ModalView\n\nfrom kivymd.theming import ThemableBehavior\nfrom kivymd.uix.behaviors import SpecificBackgroundColorBehavior\nfrom kivymd.uix.button import MDIconButton\n\nBuilder.load_string(\n \"\"\"\n#:import Window kivy.core.window.Window\n#:import StiffScrollEffect kivymd.stiffscroll.StiffScrollEffect\n\n\n\n size_hint_y: None\n height: root.theme_cls.standard_increment\n padding: [root.theme_cls.horizontal_margins - dp(12), 0]\n\n BoxLayout:\n id: left_actions\n orientation: 'horizontal'\n size_hint_x: None\n padding: [0, (self.height - dp(48))/2]\n\n BoxLayout:\n padding: dp(12), 0\n\n MDLabel:\n font_style: 'H6'\n opposite_colors: root.opposite_colors\n theme_text_color: 'Custom'\n text_color: root.specific_text_color\n text: root.title\n shorten: True\n shorten_from: 'right'\n\n BoxLayout:\n id: right_actions\n orientation: 'horizontal'\n size_hint_x: None\n padding: [0, (self.height - dp(48))/2]\n\n\n\n canvas:\n Color:\n rgba:\n root.theme_cls.bg_dark \\\n if root.theme_cls.theme_style == 'Dark' \\\n else root.theme_cls.bg_light\n Rectangle:\n size: self.size\n pos: self.pos\n\n FitImage:\n id: image\n source: root.path_to_avatar\n size_hint: 1, None\n height: Window.height * 40 // 100\n y: Window.height - self.height\n allow_stretch: True\n keep_ratio: False\n\n canvas.after:\n Color:\n rgba: root._primary_color\n Rectangle:\n size: self.size\n pos: self.pos\n\n MDLabel:\n id: user_name\n font_style: 'H4'\n theme_text_color: 'Custom'\n color: 1, 1, 1, 1\n shorten: True\n shorten_from: 'right'\n text: root.user_name\n size_hint_y: None\n height: self.texture_size[1]\n\n ModifiedToolbar:\n id: toolbar\n md_bg_color: 0, 0, 0, 0\n left_action_items: [['arrow-left', lambda x: root._callback_back()]]\n y: Window.height - self.height\n\n ScrollView:\n id: scroll\n y: -image.height\n effect_cls: StiffScrollEffect\n scroll_distance: 100\n\n canvas.before:\n Color:\n rgba:\n root.theme_cls.bg_dark\n Rectangle:\n size: self.size\n pos: self.pos\n\n MDGridLayout:\n id: box_content\n adaptive_height: True\n cols: 1\n\n canvas:\n Color:\n rgba:\n root.theme_cls.bg_dark\n Rectangle:\n size: self.size\n pos: self.pos\n\"\"\"\n)\n\n\nclass MDUserAnimationCard(ThemableBehavior, ModalView):\n user_name = StringProperty()\n path_to_avatar = StringProperty()\n box_content = ObjectProperty()\n callback = ObjectProperty()\n _anim_bottom = True\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self._primary_color = self.theme_cls.primary_color\n self._primary_color[3] = 0\n self.user_animation_card = UserAnimationCard(\n user_name=self.user_name,\n path_to_avatar=self.path_to_avatar,\n _callback_back=self._callback_back,\n _primary_color=self._primary_color,\n )\n self.user_animation_card.ids.user_name.pos = (\n dp(15),\n Window.height - self.user_animation_card.ids.image.height,\n )\n self.box_content = self.user_animation_card.ids.box_content\n self.add_widget(self.user_animation_card)\n\n self._obj_avatar = self.user_animation_card.ids.image\n self._obj_user_name = self.user_animation_card.ids.user_name\n self._obj_toolbar = self.user_animation_card.ids.toolbar\n self._obj_scroll = self.user_animation_card.ids.scroll\n self._set_current_pos_objects()\n\n def _callback_back(self):\n self.dismiss()\n if self.callback:\n self.callback()\n\n def on_open(self):\n self._primary_color = self.theme_cls.primary_color\n self._primary_color[3] = 0\n self.user_animation_card._primary_color = self._primary_color\n\n def _set_current_pos_objects(self):\n self._avatar_y = self._obj_avatar.y\n self._toolbar_y = self._obj_toolbar.y\n self._user_name_y = self._obj_user_name.y\n self._scroll_y = self._obj_scroll.y\n\n def on_touch_move(self, touch):\n if touch.ud[\"swipe_begin\"] < touch.y:\n if self._anim_bottom:\n self._anim_bottom = False\n self.animation_to_top()\n else:\n if not self._anim_bottom:\n self._anim_bottom = True\n self.animation_to_bottom()\n\n def on_touch_down(self, touch):\n touch.ud[\"swipe_begin\"] = touch.y\n return super().on_touch_down(touch)\n\n def on_touch_up(self, touch):\n touch.ud[\"swipe_begin\"] = 0\n\n def animation_to_bottom(self):\n Animation(y=self._scroll_y, d=0.4, t=\"in_out_cubic\").start(\n self._obj_scroll\n )\n Animation(y=self._user_name_y, d=0.5, x=dp(15), t=\"in_out_cubic\").start(\n self._obj_user_name\n )\n Animation(font_size=sp(36), d=0.3, t=\"in_out_cubic\").start(\n self._obj_user_name\n )\n Animation(_primary_color=[0, 0, 0, 0], d=0.3, t=\"in_out_cubic\").start(\n self.user_animation_card\n )\n Animation(y=self._avatar_y, d=0.4, t=\"in_out_cubic\").start(\n self._obj_avatar\n )\n\n def animation_to_top(self):\n user_name_y = (\n Window.height\n - self._obj_toolbar.height\n + (self.theme_cls.standard_increment // 2 - dp(12))\n )\n user_name_x = self.theme_cls.horizontal_margins + dp(12) * 5\n\n Animation(y=-self._obj_toolbar.height, d=0.4, t=\"in_out_cubic\").start(\n self._obj_scroll\n )\n Animation(y=user_name_y, d=0.3, x=user_name_x, t=\"in_out_cubic\").start(\n self._obj_user_name\n )\n Animation(font_size=sp(20), d=0.3, t=\"in_out_cubic\").start(\n self._obj_user_name\n )\n Animation(\n _primary_color=self.theme_cls.primary_color, d=0.3, t=\"in_out_cubic\"\n ).start(self.user_animation_card)\n Animation(y=self._obj_avatar.y + 30, d=0.4, t=\"in_out_cubic\").start(\n self._obj_avatar\n )\n\n\nclass UserAnimationCard(ThemableBehavior, FloatLayout):\n user_name = StringProperty()\n path_to_avatar = StringProperty()\n _callback_back = ObjectProperty()\n _primary_color = ListProperty()\n\n\nclass ModifiedToolbar(\n ThemableBehavior, SpecificBackgroundColorBehavior, BoxLayout\n):\n left_action_items = ListProperty()\n title = StringProperty()\n\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.bind(specific_text_color=self.update_action_bar_text_colors)\n Clock.schedule_once(\n lambda x: self.on_left_action_items(0, self.left_action_items)\n )\n\n def on_left_action_items(self, instance, value):\n self.update_action_bar(self.ids[\"left_actions\"], value)\n\n def update_action_bar(self, action_bar, action_bar_items):\n action_bar.clear_widgets()\n new_width = 0\n for item in action_bar_items:\n new_width += dp(48)\n action_bar.add_widget(\n MDIconButton(\n icon=item[0],\n on_release=item[1],\n opposite_colors=True,\n text_color=self.specific_text_color,\n theme_text_color=\"Custom\",\n )\n )\n action_bar.width = new_width\n\n def update_action_bar_text_colors(self, instance, value):\n for child in self.ids[\"left_actions\"].children:\n child.text_color = self.specific_text_color\n","sub_path":"kivymd/uix/useranimationcard.py","file_name":"useranimationcard.py","file_ext":"py","file_size_in_byte":10876,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"626780877","text":"from django.contrib.auth.models import User, Group\nfrom django.contrib.auth.hashers import make_password\nfrom rest_framework import authentication, permissions\nfrom rest_framework import viewsets\nfrom rest_framework import status\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom rest_framework.decorators import api_view\nfrom rest_framework.views import APIView\nfrom django.shortcuts import redirect\nfrom django.shortcuts import render\nfrom django.http import HttpResponse, JsonResponse, HttpResponseRedirect\nfrom rest_framework.response import Response\nfrom django.contrib.auth import login as do_login\nfrom django.contrib.auth import logout\nfrom django.contrib.auth.forms import AuthenticationForm\nfrom django.contrib.auth import authenticate\nfrom django.contrib.auth.decorators import login_required\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.contrib.auth import logout\nfrom rest_framework.permissions import IsAuthenticated\nfrom django.contrib.auth.views import PasswordResetView, PasswordResetDoneView, PasswordResetConfirmView, PasswordResetCompleteView\nfrom django.contrib import messages #import messages\nfrom .serializers import *\nfrom .models import *\nfrom django.db.models import F\nfrom django.contrib.auth.hashers import make_password\nfrom django.core.mail import send_mail\nfrom django.conf import settings\nfrom django.views.decorators.http import require_http_methods\nfrom django.http import HttpResponse,HttpResponseBadRequest\nfrom django.core import serializers\nimport json\nfrom fcm_django.models import FCMDevice\n\n\n\n#from .serializers import UserSerializer, GroupSerializer,TCategoriaSerializer,TComentarioSerializer,TEscaneosSerializer,TFavoritoSerializer,TGaleriaSerializer,TLocalSerializer,TNotificacionesSerializer,TPermisoSerializer,TRolSerializer,TRolpermisoSerializer,TTelefonoSerializer,TUsuarioSerializer\n#from .models import Categoria,Comentario,Escaneos,Favorito,Galeria,Local,Notificaciones,Permiso,Rol,Rolpermiso,Telefono,User\n\n\nclass UserViewSet(viewsets.ModelViewSet):\n \"\"\"\n API endpoint that allows users to be viewed or edited.\n \"\"\"\n queryset = User.objects.all().order_by('-date_joined')\n serializer_class = UserSerializer\n\n\nclass GroupViewSet(viewsets.ModelViewSet):\n \"\"\"\n API endpoint that allows groups to be viewed or edited.\n \"\"\"\n queryset = Group.objects.all()\n serializer_class = GroupSerializer\n\n\nclass CategoriaViewSet(viewsets.ModelViewSet):\n queryset = Categoria.objects.all()\n serializer_class = TCategoriaSerializer\n\n\nclass ComentarioViewSet(viewsets.ModelViewSet):\n queryset = Comentario.objects.all()\n serializer_class = TComentarioSerializer\n\n\nclass EscaneosViewSet(viewsets.ModelViewSet):\n queryset = Escaneos.objects.all()\n serializer_class = TEscaneosSerializer\n\n\nclass FavoritoViewSet(viewsets.ModelViewSet):\n queryset = Favorito.objects.all()\n serializer_class = TFavoritoSerializer\n\n\nclass GaleriaViewSet(viewsets.ModelViewSet):\n queryset = Galeria.objects.all()\n serializer_class = TGaleriaSerializer\n\n\nclass LocalViewSet(viewsets.ModelViewSet):\n queryset = Local.objects.all()\n serializer_class = TLocalSerializer\n\n\nclass NotificacionesViewSet(viewsets.ModelViewSet):\n queryset = Notificaciones.objects.all()\n serializer_class = TNotificacionesSerializer\n\n\nclass PermisoViewSet(viewsets.ModelViewSet):\n queryset = Permiso.objects.all()\n serializer_class = TPermisoSerializer\n\n\nclass RolViewSet(viewsets.ModelViewSet):\n queryset = Rol.objects.all()\n serializer_class = TRolSerializer\n\n\nclass RolpermisoViewSet(viewsets.ModelViewSet):\n queryset = Rolpermiso.objects.all()\n serializer_class = TRolpermisoSerializer\n\n\nclass TelefonoViewSet(viewsets.ModelViewSet):\n queryset = Telefono.objects.all()\n serializer_class = TTelefonoSerializer\n\n\nclass UsuarioViewSet(viewsets.ModelViewSet):\n queryset = User.objects.all()\n serializer_class = TUsuarioSerializer\n\nclass UsuarioAPPViewSet(viewsets.ModelViewSet):\n queryset = UserAPP.objects.all()\n serializer_class = TUsuarioAPPSerializer\n\n\nclass PublicidadViewSet(viewsets.ModelViewSet):\n queryset = Publicidad.objects.all()\n serializer_class = TPublicidadSerializer\n\n\ndef index(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/index.html',{'permisos': lpermisos})\n\n\ndef tablaUsuario(request):\n #obtener permisos en base al rol anteriro\n lpermisos = obtenerPermisos(request.user)\n user = User.objects.all()\n contexto = {'usuarios': user ,'permisos': lpermisos }\n return render(request, 'productos/tablaUsuario.html', contexto)\n\n\ndef tablaLocal(request):\n #obtener permisos en base al rol anteriro\n lpermisos = obtenerPermisos(request.user)\n local = Local.objects.all() \n contexto = {'locales': local,'permisos': lpermisos }\n return render(request, 'productos/tablaLocal.html', contexto)\n\n\ndef tableCategoria(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n categoria = Categoria.objects.all()\n contexto = {'categorias': categoria,'permisos': lpermisos }\n return render(request, 'productos/tablaCategoria.html', contexto)\n\n\ndef tableFavorito(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n favorito = Favorito.objects.all()\n contexto = {'favoritos': favorito,'permisos': lpermisos}\n return render(request, 'productos/tablaFavorito.html', contexto)\n\n\ndef tableTelefono(request):\n #obtener permisos en base al rol anteriro\n lpermisos = obtenerPermisos(request.user) \n telefono = Telefono.objects.all()\n contexto = {'telefonos': telefono,'permisos': lpermisos }\n return render(request, 'productos/tablaTelefono.html', contexto)\n\n\ndef tableGaleria(request):\n lpermisos = obtenerPermisos(request.user)\n galeria = Galeria.objects.all()\n contexto = {'galerias': galeria ,'permisos': lpermisos}\n return render(request, 'productos/tablaGaleria.html', contexto)\ndef tableGaleria2(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n locales = []\n user = User.objects.filter(email=request.user).first()\n local = Local.objects.filter(adminLocal=user).all()#filtras que locales tiene ese usuario\n for i in local:\n if i.id_local not in locales:\n locales.append(i.id_local)\n galeria = Galeria.objects.filter(id_local__in=locales)\n contexto = {'galerias': galeria,'permisos': lpermisos}\n return render(request, 'productos/tablaGaleria2.html', contexto)\ndef notificaciones(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n notificacion = Notificaciones.objects.all()\n contexto = {'notificaciones': notificacion,'permisos': lpermisos}\n return render(request, 'productos/notificaciones.html', contexto)\ndef localDelete(request):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user)\n if request.method == \"POST\":\n local = Local.objects.get(id_local=request.POST.get('local'))\n local.delete()\n return render(request, 'productos/tablaLocal.html', {\"locales\": Local.objects.all(),'permisos': lpermisos})\n\n\ndef categoriaDelete(request, id_categoria):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user) \n categoria = Categoria.objects.get(id_categoria=id_categoria)\n categoria.delete()\n return render(request, 'productos/tablaCategoria.html', {\"categorias\": Categoria.objects.all(),'permisos': lpermisos})\n\n\ndef favoritoDelete(request, id_favorito):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user) \n favorito = Favorito.objects.get(id_favorito=id_favorito)\n favorito.delete()\n return render(request, 'productos/tablaFavorito.html', {\"favoritos\": Favorito.objects.all(),'permisos': lpermisos})\n\n\ndef telefonoDelete(request, id_telefono):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user) \n telefono = Telefono.objects.get(id_telefono=id_telefono)\n telefono.delete()\n return render(request, 'productos/tablaTelefono.html', {\"telefonos\": Telefono.objects.all(),'permisos': lpermisos})\n\n\ndef galeriaDelete(request, id_contenido):\n #obtener permisos en base al rol anteriro \n lpermisos = obtenerPermisos(request.user) \n galeria = Galeria.objects.get(id_contenido=id_contenido)\n galeria.delete()\n return render(request, 'productos/tablaGaleria.html', {\"galerias\": Galeria.objects.all(),'permisos': lpermisos})\n# ...\n\n\n@csrf_exempt\ndef login(request):\n form = AuthenticationForm()\n if request.method == \"POST\":\n form = AuthenticationForm(data=request.POST)\n if form.is_valid():\n username = form.cleaned_data['username']\n password = form.cleaned_data['password']\n user = authenticate(username=username, password=password)\n print(\"verificando\")\n if user is not None:\n do_login(request, user)\n print(\"si existe\")\n return redirect(\"/tablaLocal\")\n else:\n print(form.is_valid())\n print(\"Password o usuario incorrecto\")\n return render(request, \"productos/login.html\", {'form': form, 'mensaje': \"User o cotrnaseña iconrrecta.\"})\n\n # Si llegamos al final renderizamos el formulario\n return render(request, \"productos/login.html\", {'form': form, 'mensaje': \"\"})\n\ndef obtenerPermisos(user):\n rol = User.objects.all().select_related('id_rol').filter(email=user).values('id_rol')[0]['id_rol']\n dic = Rolpermiso.objects.all().filter(id_rol=rol).values(permiso=F('id_permiso__id_permiso'))\n l = []\n for i in dic:\n if i[\"permiso\"] not in l:\n l.append(i[\"permiso\"])\n return l\n\n@login_required(login_url='/')\ndef logout_view(request):\n logout(request)\n return redirect('/')\n\n\n@login_required(login_url='/')\ndef registrarCategoria(request):\n if request.method == 'POST':\n if(request.POST.get(\"tipo\") != None and request.POST.get(\"descripcion\") != None):\n categoria = Categoria(tipo=request.POST.get(\n \"tipo\"), descripcion=request.POST.get(\"descripcion\"))\n categoria.save()\n # return HttpResponse(status=200)\n return redirect(tablaCategoria)\n return HttpResponse(status=404)\n if request.method == 'GET':\n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/crear/crearCategoria.html',{'permisos': lpermisos})\n\n\n@login_required(login_url='/')\ndef registrarLocal(request):\n if request.method == 'POST':\n imagen = request.FILES.get('imagen')\n if imagen != None:\n if( request.POST.get(\"slogan\") != None and request.POST.get(\"latitud\") != None and request.POST.get(\"longitud\") != None and request.POST.get(\"direccion\") != None and request.POST.get(\"nombrec\") != None and request.POST.get(\"descripcion\") != None):\n categoria = Categoria.objects.filter(id_categoria=request.POST.get(\"categoria\")).first()\n local = Local(latitud=request.POST.get(\"latitud\"), estrellas=0, longitud=request.POST.get(\"longitud\"), slogan=request.POST.get(\"slogan\"), vistas=\n 0, descripcion=request.POST.get(\"descripcion\"), likes=0, direccion=request.POST.get(\"direccion\"), nombre_comercial=request.POST.get(\"nombrec\"), src_logo=request.FILES['imagen'],categoria=categoria)\n local.save()\n # return HttpResponse(status=200)\n #return render(request, 'productos/crear/crearLocal.html',{'permisos': lpermisos })\n else:\n if(request.POST.get(\"estrella\") != None and request.POST.get(\"slogan\") != None and request.POST.get(\"latitud\") != None and request.POST.get(\"longitud\") != None and request.POST.get(\"vista\") != None and request.POST.get(\"direccion\") != None and request.POST.get(\"nombrec\") != None and request.POST.get(\"like\") != None and request.POST.get(\"descripcion\") != None):\n categoria = Categoria.objects.filter(id_categoria=request.POST.get(\"categoria\"))\n local = Local(latitud=request.POST.get(\"latitud\"), estrellas=0, longitud=request.POST.get(\"longitud\"), slogan=request.POST.get(\"slogan\"), vistas=0\n , descripcion=request.POST.get(\"descripcion\"), likes=0, direccion=request.POST.get(\"direccion\"), nombre_comercial=request.POST.get(\"nombrec\"),categoria=categoria)\n local.save()\n # return HttpResponse(status=200)\n #return render(request, 'productos/crear/crearLocal.html',{'permisos': lpermisos })\n return redirect(tablaLocal)\n #return HttpResponse(status=404)\n if request.method == 'GET':\n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/crear/crearLocal.html', {\"categorias\": Categoria.objects.all(),'permisos': lpermisos})\n\n\n@login_required(login_url='/')\ndef editarLocal(request):\n if request.method == 'POST':\n local = Local.objects.get(id_local=request.POST['local'])\n if(request.POST['nombrec'] != None or request.POST['nombrec'] != ''):\n local.nombre_comercial = request.POST['nombrec']\n if(request.POST['descripcion'] != None or request.POST['descripcion'] != ''):\n local.descripcion = request.POST['descripcion']\n if(request.POST['like'] != None or request.POST['like'] != ''):\n local.likes = request.POST['like']\n if(request.POST['estrella'] != None or request.POST['estrella'] != ''):\n local.estrellas = request.POST['estrella']\n if(request.POST['vista'] != None or request.POST['vista'] != ''):\n local.vistas = request.POST['vista']\n if(request.POST['direccion'] != None or request.POST['direccion'] != ''):\n local.direccion = request.POST['direccion']\n if(request.POST['longitud'] != None or request.POST['longitud'] != ''):\n local.direccion = request.POST['longitud']\n if(request.POST['latitud'] != None or request.POST['latitud'] != ''):\n local.direccion = request.POST['latitud']\n if(request.POST['slogan'] != None or request.POST['slogan'] != ''):\n local.direccion = request.POST['slogan']\n imagen = request.FILES.get('imagen')\n if(imagen!=None):\n local.src_logo = request.FILES['imagen']\n local.save()\n # dispositivos=FCMDevice.objects.filter(active=True)\n # dispositivos.send_message(\n # title=\"local se agrego\"+request.POST['nombrec'],\n # body=\"se ha agregado un local\"\n # )\n return redirect(tablaLocal)\n\n\n@login_required(login_url='/')\ndef registrarUsuario(request):\n #lpermisos = obtenerPermisos(request.user)\n print(request.method)\n \n if request.method == 'POST':\n src_imagen = request.FILES.get('imagen')\n if src_imagen!=None:\n if(request.POST.get(\"email\") != None and request.POST.get(\"nombre\") != None and request.POST.get(\"apellido\") != None and request.POST.get(\"contrasena\") != None and request.POST.get(\"telefono\") != None and request.FILES[\"imagen\"] != None):\n rol = Rol.objects.all().filter(id_rol=request.POST.get(\"rol\")).first()\n usuario = User(username=request.POST.get(\"email\").split('@')[0], email=request.POST.get(\"email\"), nombres=request.POST.get(\"nombre\"), first_name=request.POST.get(\"nombre\"), contrasena=request.POST.get(\n \"contrasena\"), password=make_password(request.POST.get(\"contrasena\")), telefono=request.POST.get(\"telefono\"), apellidos=request.POST.get(\"apellido\"), last_name=request.POST.get(\"apellido\"), src_imagen=request.FILES['imagen'],id_rol=rol)\n usuario.save()\n #return redirect(tablaUsuario)\n # return HttpResponse(status=200)\n #return render(request, 'productos/crear/crearUsuario.html',{'permisos': lpermisos })\n else: \n messages.warning(request, 'No se pudo registrar Usuario.')\n else:\n if(request.POST.get(\"email\") != None and request.POST.get(\"nombre\") != None and request.POST.get(\"apellido\") != None and request.POST.get(\"contrasena\") != None and request.POST.get(\"telefono\") != None ):\n rol = Rol.objects.all().filter(id_rol=request.POST.get(\"rol\")).first()\n usuario = User(username=request.POST.get(\"email\").split('@')[0], email=request.POST.get(\"email\"), nombres=request.POST.get(\"nombre\"), first_name=request.POST.get(\"nombre\"), contrasena=request.POST.get(\n \"contrasena\"), password=make_password(request.POST.get(\"contrasena\")), telefono=request.POST.get(\"telefono\"), apellidos=request.POST.get(\"apellido\"), last_name=request.POST.get(\"apellido\"),id_rol=rol)\n usuario.save()\n else: \n messages.warning(request, 'No se pudo registrar Usuario.')\n return redirect(tablaUsuario)\n \n if request.method == 'GET':\n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/crear/crearUsuario.html', {\"roles\": Rol.objects.all(), 'permisos': lpermisos})\n\n\n@login_required(login_url='/')\ndef editarUsuario(request):\n if request.method == 'POST':\n usuario = User.objects.get(email=request.POST['email'])\n if(request.POST['email'] != None or request.POST['email'] != ''):\n usuario.email = request.POST['email']\n if(request.POST['nombres'] != None or request.POST['nombres'] != ''):\n usuario.nombres = request.POST['nombres']\n if(request.POST['apellidos'] != None or request.POST['apellidos'] != ''):\n usuario.apellidos = request.POST['apellidos']\n if(request.POST['contrasena'] != None or request.POST['contrasena'] != ''):\n usuario.contrasena = request.POST['contrasena']\n if(request.POST['telefono'] != None or request.POST['telefono'] != ''):\n usuario.telefono = request.POST['telefono']\n # if(request.POST['id_rol']!=None or request.POST['id_rol']!=''):\n \n # usuario.id_rol=rol\n if(bool(request.FILES.get('imagen', False)) == True):\n usuario.src_imagen = request.FILES['imagen']\n usuario.save()\n return redirect(tablaUsuario)\n #return render(request, 'productos/tablaUsuario.html', {\"usuarios\": User.objects.all()})\n\n\n@login_required(login_url='/')\ndef usuarioDelete(request):\n lpermisos = obtenerPermisos(request.user)\n if request.method == \"POST\":\n usuario = User.objects.get(email=request.POST.get('email'))\n usuario.delete()\n return render(request, 'productos/tablaUsuario.html', {\"usuarios\": User.objects.all() ,'permisos': lpermisos})\n\n@login_required(login_url='/')\ndef registrarPublicidad(request):\n if request.method == 'POST':\n if(request.POST.get(\"descripcion\") != None):\n publicidad = Publicidad(tipo=request.POST.get(\"descripcion\"))\n publicidad.save()\n # return HttpResponse(status=200)\n return render(request, 'productos/crear/crearPublicidad.html')\n return HttpResponse(status=404)\n if request.method == 'GET':\n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/crear/crearPublicidad.html',{'permisos': lpermisos})\n@api_view([\"PUT\"])\n@csrf_exempt\ndef update_favorito(request, favorito_id):\n user = request.user.id\n payload = json.loads(request.body)\n try:\n favorito_item = Favorito.objects.filter(added_by=user, id=favorito_id)\n # returns 1 or 0\n favorito_item.update(**payload)\n favorito = Favorito.objects.get(id=favorito_id)\n serializer = TFavoritoSerializer(favorito)\n return JsonResponse({'favorito': serializer.data}, safe=False, status=status.HTTP_200_OK)\n except ObjectDoesNotExist as e:\n return JsonResponse({'error': str(e)}, safe=False, status=status.HTTP_404_NOT_FOUND)\n except Exception:\n return JsonResponse({'error': 'Something terrible went wrong'}, safe=False, status=status.HTTP_500_INTERNAL_SERVER_ERROR)\n@login_required(login_url='/')\ndef registrarNotificaciones(request):\n if request.method == 'POST':\n if(request.POST.get(\"alcance\") != None and request.POST.get(\"notificacion\") != None):\n notificacion = Notificaciones(alcance=request.POST.get(\n \"alcance\"), notificacion=request.POST.get(\"notificacion\"))\n notificacion.save()\n # return HttpResponse(status=200)\n return redirect(tablaCategoria)\n return HttpResponse(status=404)\n if request.method == 'GET':\n lpermisos = obtenerPermisos(request.user)\n return render(request, 'productos/crear/crearNotificacion.html',{'permisos': lpermisos})\n@csrf_exempt\n@require_http_methods(['POST'])\ndef guardar_token(request):\n body=request.body.decode('utf-8')\n bodyDict=json.loads(body)\n token=bodyDict['token']\n existe=FCMDevice.objects.filter(registration_id=token,active=True)\n if len(existe)>0:\n return HttpResponseBadRequest(json.dumps({'mensaje':'el token ya existe'}))\n dispositivo=FCMDevice()\n dispositivo.registration_id=token\n dispositivo.active=True\n #si el usuario esta logeado procede a enlazar\n if request.user.is_authenticated:\n dispositivo.user=request.user\n try:\n dispositivo.save()\n return HttpResponse(json.dumps({'mensaje':'no se ha podido guardar'}))\n except:\n return HttpResponseBadRequest(json.dumps({'mensaje':'no se ha podido guardar'}))","sub_path":"Admin-Backend/productos/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":21769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"557414885","text":"#!/usr/bin/env python\nimport argparse\n\n# Main code now in lsst.verify; this version for backward-compatibility only.\nfrom lsst.verify.bin.jobReporter import main\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=('Produce a Job object which can either be used '\n 'to build a local report or to ship to SQuaSH.'))\n parser.add_argument('repository', type=str,\n help='Path to a valid gen3 repository')\n parser.add_argument('collection', type=str,\n help='Collection to search for metric measurement values')\n parser.add_argument('--metrics_package', type=str, default=\"validate_drp\",\n help='Name of metrics package to load, defaults to validate_drp.')\n parser.add_argument('--spec', type=str, default=\"design\",\n help='Spec level to apply: minimum, design, or stretch')\n parser.add_argument('--dataset_name', type=str, default=\"validation_data_hsc\",\n help='Name of the dataset for which the report is being generated. '\n 'Defaults to validation_data_hsc.')\n\n args = parser.parse_args()\n main(args.repository, args.collection, args.metrics_package, args.spec, args.dataset_name)\n","sub_path":"bin.src/make_job_document.py","file_name":"make_job_document.py","file_ext":"py","file_size_in_byte":1314,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"191324492","text":"# config:utf-8\n\nimport smtplib\nfrom email.mime.text import MIMEText\nimport datetime\n\njp=\"iso-2022-jp\"\nmessage=MIMEText(\"メール送信テスト\".encode(jp),\"plain\",jp,)\nfrom_address=\"test@aaa.bb\"\nto_address=\"nao.zima@gmail.com\"\n\nmessage[\"Subject\"]=\"件名\"\nmessage[\"From\"]=from_address\nmessage[\"To\"]=to_address\n\n\nserver=smtplib.SMTP('localhost')\nserver.send_message(message)\n","sub_path":"src/send_mail.py","file_name":"send_mail.py","file_ext":"py","file_size_in_byte":376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"561835744","text":"# adapted from pimoroni evdev support for the 7 inch capacitive screen\n# added support for the resistive 3.5 and maybe others that doesn't depend upon SDL 1.2\nimport errno\nimport glob\nimport io\nimport os\nimport queue\nimport struct\nimport time\nfrom collections import namedtuple\nimport pygame\n\nimport select\n\nimport debug\n\nTOUCH_X = 0\nTOUCH_Y = 1\n\nTouchEvent = namedtuple('TouchEvent', ('timestamp', 'type', 'code', 'value'))\n\nEV_SYN = 0\nEV_ABS = 3\n\nABS_X = 0\nABS_Y = 1\n\nEV_KEY = 1\nBTN_TOUCH = 330\nABS_MT_SLOT = 0x2f # 47 MT slot being modified\nABS_MT_POSITION_X = 0x35 # 53 Center X of multi touch position\nABS_MT_POSITION_Y = 0x36 # 54 Center Y of multi touch position\nABS_MT_TRACKING_ID = 0x39 # 57 Unique ID of initiated contact\n\nTS_PRESS = 1\nTS_RELEASE = 0\nTS_MOVE = 2\n\n\nclass Touch(object):\n\tdef __init__(self, slot, x, y):\n\t\tself.slot = slot\n\n\t\tself._x = x\n\t\tself._y = y\n\t\tself.last_x = -1\n\t\tself.last_y = -1\n\n\t\tself._id = -1\n\t\tself.events = []\n\t\tself.on_move = None\n\t\tself.on_press = None\n\t\tself.on_release = None\n\n\t@property\n\tdef position(self):\n\t\treturn self.x, self.y\n\n\t@property\n\tdef last_position(self):\n\t\treturn self.last_x, self.last_y\n\n\t@property\n\tdef valid(self):\n\t\treturn self.id > -1\n\n\t@property\n\tdef id(self):\n\t\treturn self._id\n\n\t@id.setter\n\tdef id(self, value):\n\t\tif value != self._id:\n\t\t\tif value == -1 and not TS_RELEASE in self.events:\n\t\t\t\tself.events.append(TS_RELEASE)\n\t\t\telif not TS_PRESS in self.events:\n\t\t\t\tself.events.append(TS_PRESS)\n\n\t\tself._id = value\n\n\t@property\n\tdef x(self):\n\t\treturn self._x\n\n\t@x.setter\n\tdef x(self, value):\n\t\tif value != self._x and not TS_MOVE in self.events:\n\t\t\tself.events.append(TS_MOVE)\n\t\tself.last_x = self._x\n\t\tself._x = value\n\n\t@property\n\tdef y(self):\n\t\treturn self._y\n\n\t@y.setter\n\tdef y(self, value):\n\t\tif value != self._y and not TS_MOVE in self.events:\n\t\t\tself.events.append(TS_MOVE)\n\t\tself.last_y = self._y\n\t\tself._y = value\n\n\tdef handle_events(self):\n\t\t\"\"\"Run outstanding press/release/move events\"\"\"\n\t\tfor event in self.events:\n\t\t\tif event == TS_MOVE and callable(self.on_move):\n\t\t\t\tself.on_move(event, self)\n\t\t\tif event == TS_PRESS and callable(self.on_press):\n\t\t\t\tself.on_press(event, self)\n\t\t\tif event == TS_RELEASE and callable(self.on_release):\n\t\t\t\tself.on_release(event, self)\n\n\t\tself.events = []\n\n\nclass Touches(list):\n\t@property\n\tdef valid(self):\n\t\treturn [tch for tch in self if tch.valid]\n\n\nclass Touchscreen(object):\n\tTOUCHSCREEN_EVDEV_NAME = ('FT5406 memory based driver', 'Goodix Capactivie Touchscreen')\n\tTOUCHSCREEN_RESISTIVE = 'stmpe-ts'\n\tTOUCHSCREEN28CAP = 'EP0110M09'\n\tEVENT_FORMAT = str('llHHi')\n\tEVENT_SIZE = struct.calcsize(EVENT_FORMAT)\n\n\tdef __init__(self):\n\t\tself.touchdefs = {}\n\t\twith open('../touchdefinitions') as f:\n\t\t\tdefs = f.read().splitlines()\n\t\t\tfor l in defs:\n\t\t\t\ttouchitem = l.split('|')\n\t\t\t\tself.touchdefs[touchitem[0]] = touchitem[1:]\n\t\tprint(self.touchdefs)\n\t\tself._use_multitouch = True\n\t\tself.controller = \"unknown\"\n\t\tself._shiftx = 0\n\t\tself._shifty = 0\n\t\tself._flipx = 0 # 0 for ok else size of x from which to subtract touch value\n\t\tself._flipy = 0 # 0 for ok else size of y from which to subtract touch value\n\t\tself._scalex = 1.0\n\t\tself._scaley = 1.0\n\t\tself._capscreen = True\n\t\tself.a = None\n\t\tself._running = False\n\t\tself._thread = None\n\t\tself._f_poll = select.poll()\n\t\tself._f_device = io.open(self._touch_device(), 'rb', self.EVENT_SIZE)\n\t\tself._f_poll.register(self._f_device, select.POLLIN)\n\t\tself.position = Touch(0, 0, 0)\n\t\tself.touches = Touches([Touch(x, 0, 0) for x in range(10)])\n\t\tself._event_queue = queue.Queue()\n\t\tself._touch_slot = 0\n\n\tdef _run(self):\n\t\tself._running = True\n\t\twhile self._running:\n\t\t\tself.poll()\n\t\t\ttime.sleep(0.00001)\n\n\tdef run(self):\n\t\tself._run()\n\n\tdef stop(self):\n\t\tif self._thread is None:\n\t\t\treturn\n\n\t\tself._running = False\n\t\tself._thread.join()\n\t\tself._thread = None\n\n\t@property\n\tdef _current_touch(self):\n\t\treturn self.touches[self._touch_slot]\n\n\tdef close(self):\n\t\tself._f_device.close()\n\n\tdef __enter__(self):\n\t\treturn self\n\n\tdef __exit__(self, exc_type, exc_value, exc_tb):\n\t\tself.close()\n\n\tdef __iter__(self):\n\t\tpass\n\n\tdef _lazy_read(self):\n\t\twhile self._wait_for_events():\n\t\t\tevent = self._f_device.read(self.EVENT_SIZE)\n\t\t\tif not event:\n\t\t\t\tbreak\n\t\t\tyield event\n\n\tdef _get_pending_events(self):\n\t\tfor event in self._lazy_read():\n\t\t\t(tv_sec, tv_usec, ttype, code, value) = struct.unpack(self.EVENT_FORMAT, event)\n\t\t\tself._event_queue.put(TouchEvent(tv_sec + (tv_usec / 1000000), ttype, code, value))\n\n\tdef _wait_for_events(self, timeout=2):\n\t\treturn self._f_poll.poll(timeout)\n\n\tdef poll(self):\n\t\tself._get_pending_events()\n\n\t\twhile not self._event_queue.empty():\n\t\t\tevent = self._event_queue.get()\n\t\t\tdebug.debugPrint('LLTouch', 'Touch: ' + str(event))\n\t\t\tself._event_queue.task_done()\n\n\t\t\tif event.type == EV_SYN: # Sync\n\t\t\t\tfor tch in self.touches:\n\t\t\t\t\ttch.handle_events()\n\t\t\t\treturn self.touches\n\n\t\t\tif event.type == EV_KEY and not self._capscreen:\n\t\t\t\tif event.code == BTN_TOUCH:\n\t\t\t\t\tself._touch_slot = 0\n\t\t\t\t\t# self._current_touch.id = 1\n\t\t\t\t\tif self.a is None:\n\t\t\t\t\t\tself._current_touch.x = self.position.x\n\t\t\t\t\t\tself._current_touch.y = self.position.y\n\t\t\t\t\telse:\n\t\t\t\t\t\tself._current_touch.x = (self.a[2] + self.a[0] * self.position.x + self.a[\n\t\t\t\t\t\t\t1] * self.position.y) / self.a[6]\n\t\t\t\t\t\tself._current_touch.y = (self.a[5] + self.a[3] * self.position.x + self.a[\n\t\t\t\t\t\t\t4] * self.position.y) / self.a[6]\n\t\t\t\t\tif event.value == 1:\n\t\t\t\t\t\tself._current_touch.events.append(TS_PRESS)\n\t\t\t\t\telse:\n\t\t\t\t\t\tself._current_touch.events.append(TS_RELEASE)\n\n\t\t\tif event.type == EV_ABS: # Absolute cursor position\n\t\t\t\tif event.code == ABS_MT_SLOT:\n\t\t\t\t\tself._touch_slot = event.value\n\n\t\t\t\tif event.code == ABS_MT_TRACKING_ID:\n\t\t\t\t\tself._current_touch.id = event.value\n\n\t\t\t\tif event.code == ABS_MT_POSITION_X:\n\t\t\t\t\ttmp = event.value + self._shiftx\n\t\t\t\t\tif self._flipx != 0:\n\t\t\t\t\t\ttmp = self._flipx - event.value\n\t\t\t\t\tif tmp < 0:\n\t\t\t\t\t\tprint('Negative touch position(x): {}'.format(tmp))\n\t\t\t\t\t\ttmp = 0\n\t\t\t\t\tself._current_touch.x = round(tmp * self._scalex)\n\n\t\t\t\tif event.code == ABS_MT_POSITION_Y:\n\t\t\t\t\ttmp = event.value + self._shifty\n\t\t\t\t\tif self._flipy != 0:\n\t\t\t\t\t\ttmp = self._flipy - event.value\n\t\t\t\t\tif tmp < 0:\n\t\t\t\t\t\tprint('Negative touch position(y): {}'.format(tmp))\n\t\t\t\t\t\ttmp = 0\n\t\t\t\t\tself._current_touch.y = round(tmp * self._scaley)\n\n\t\t\t\tif event.code == ABS_X:\n\t\t\t\t\tself.position.x = event.value\n\n\t\t\t\tif event.code == ABS_Y:\n\t\t\t\t\tself.position.y = event.value\n\n\t\treturn []\n\n\tdef _touch_device(self):\n\t\t# return '/dev/input/touchscreen'\n\t\tfor evdev in glob.glob(\"/sys/class/input/event*\"):\n\t\t\ttry:\n\t\t\t\twith io.open(os.path.join(evdev, 'device', 'name'), 'r') as f:\n\t\t\t\t\tdev = f.read().strip()\n\t\t\t\t\tif dev in self.touchdefs:\n\t\t\t\t\t\tself.controller = dev\n\t\t\t\t\t\tvals = self.touchdefs[dev]\n\t\t\t\t\t\tself._shiftx = int(vals[1])\n\t\t\t\t\t\tself._shifty = int(vals[2])\n\t\t\t\t\t\tself._flipx = int(vals[3])\n\t\t\t\t\t\tself._flipy = int(vals[4])\n\t\t\t\t\t\tself._scalex = float(vals[5])\n\t\t\t\t\t\tself._scaley = float(vals[6])\n\t\t\t\t\t\tself._capscreen = bool(vals[0])\n\t\t\t\t\t\treturn os.path.join('/dev', 'input', os.path.basename(evdev))\n\t\t\t\t\t'''\n\t\t\t\t\tif dev in self.TOUCHSCREEN_EVDEV_NAME:\n\t\t\t\t\t\treturn os.path.join('/dev', 'input', os.path.basename(evdev))\n\t\t\t\t\telif dev == self.TOUCHSCREEN_RESISTIVE:\n\t\t\t\t\t\tself._capscreen = False\n\t\t\t\t\t\twith open('/etc/pointercal', 'r') as pc:\n\t\t\t\t\t\t\tself.a = list(int(x) for x in next(pc).split())\n\t\t\t\t\t\t# set to do corrections? TODO read pointercal and set a flag to correct\n\t\t\t\t\t\treturn os.path.join('/dev', 'input', os.path.basename(evdev))\n\t\t\t\t\telif dev == self.TOUCHSCREEN28CAP:\n\t\t\t\t\t\tself._flipx = 0 # 240 todo auto fix orientation?\n\t\t\t\t\t\tself._flipy = 0 # 320\n\t\t\t\t\t\treturn os.path.join('/dev', 'input', os.path.basename(evdev))\n\t\t\t\t\t'''\n\t\t\texcept IOError as e:\n\t\t\t\tif e.errno != errno.ENOENT:\n\t\t\t\t\traise\n\t\traise RuntimeError('Unable to locate touchscreen device')\n\n\tdef read(self):\n\t\treturn next(iter(self))\n\n\nif __name__ == \"__main__\":\n\timport signal\n\n\tpygame.init()\n\tpygame.fastevent.init()\n\tminx = 1000000\n\tmaxx = -1000000\n\tminy = 1000000\n\tmaxy = -1000000\n\ta = [5724, -6, -1330074, 26, 8427, -1034528, 65536]\n\tb = [34, 952, 38, 943]\n\n\tts = Touchscreen()\n\n\tprint('Test touch points:')\n\n\n\tdef handle_event(event, tch):\n\t\tglobal minx, miny, maxx, maxy\n\t\t# xx = (a[2] + a[0] * touch.x + a[1] * touch.y) / a[6]\n\t\t# yy = (a[5] + a[3] * touch.x + a[4] * touch.y) / a[6]\n\t\t# Xx = (touch.x - b[0]) * 320 / (b[1] - b[0])\n\t\t# Xy = (touch.y - b[2]) * 480 / (b[3] - b[2])\n\t\tprint([\"Release\", \"Press\", \"Move\"][event],\n\t\t\t tch.slot,\n\t\t\t tch.x,\n\t\t\t tch.y)\n\t\tminx = min(minx, tch.x)\n\t\tminy = min(miny, tch.y)\n\t\tmaxx = max(maxx, tch.x)\n\t\tmaxy = max(maxy, tch.y)\n\t\treturn\n\t\t# noinspection PyUnreachableCode\n\t\tif event == 1:\n\t\t\te = pygame.event.Event(pygame.MOUSEBUTTONDOWN, {'pos': (tch.x, tch.y)})\n\t\t\tpygame.fastevent.post(e)\n\t\telif event == 0:\n\t\t\te = pygame.event.Event(pygame.MOUSEBUTTONUP, {'pos': (tch.x, tch.y)})\n\t\t\tpygame.fastevent.post(e)\n\n\n\tfor touch in ts.touches:\n\t\ttouch.on_press = handle_event\n\t\ttouch.on_release = handle_event\n\t\ttouch.on_move = handle_event\n\n\ttry:\n\t\tts.run()\n\texcept KeyboardInterrupt:\n\t\tprint(\"Stopping thread...\")\n\t\tprint('MinX: {} MaxX: {} MinY: {} MaxY: {}'.format(minx, maxx, miny, maxy))\n\t\texit()\n","sub_path":"deprecated/test2touchhandler.py","file_name":"test2touchhandler.py","file_ext":"py","file_size_in_byte":9201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"520234239","text":"import numpy as np\nimport scipy.io\nfrom typing import Iterable, Tuple\n\n\nclass MatlabWeightsParser(object):\n \"\"\"\n Class for handling matconvnet weights.\n \"\"\"\n def __init__(self, weights_path: str, layer_names: Iterable[str]):\n \"\"\"\n :param weights_path: path to matconvnet file.\n :param layer_names: names of convnet for which weight are loading.\n \"\"\"\n\n self._data = scipy.io.loadmat(weights_path)['layers'][0]\n\n self._weights = {\n layer: MatlabWeightsParser.get_conv_layer_weights(self._data[index])\n for index, layer in enumerate(layer_names)\n if layer.startswith('conv')\n }\n\n def get_layer_weights(self, layer_name: str) -> np.ndarray:\n \"\"\"\n Gets weights for layer by name.\n :param layer_name: layer name for which need load weights.\n :return: layer weights.\n \"\"\"\n\n if layer_name not in self._weights:\n raise KeyError('Invalid layer name {}.'.format(\n layer_name\n ))\n\n return self._weights[layer_name]\n\n @staticmethod\n def get_conv_layer_weights(layer_weights: np.ndarray) -> Tuple[np.ndarray]:\n \"\"\"\n Loads convolution weights for matconvnet format into TensorFlow format.\n :param layer_weights: matconvnet weights.\n :return: TensorFlow weights.\n \"\"\"\n\n kernel_weights = layer_weights[0][0][2][0][0]\n\n # matconvnet: layer_weights are\n # [width, height, in_channels, out_channels]\n\n # tensorflow: layer_weights are\n # [height, width, in_channels, out_channels]\n\n kernel_weights = np.transpose(kernel_weights, (1, 0, 2, 3))\n\n bias_weights = np.array(layer_weights[0][0][2][0][1])\n bias_weights = bias_weights.reshape(-1)\n\n return kernel_weights, bias_weights\n","sub_path":"utils/matlab_weight_loader.py","file_name":"matlab_weight_loader.py","file_ext":"py","file_size_in_byte":1844,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"107900199","text":"from django.shortcuts import render,redirect,get_object_or_404\nfrom .models import Movie, Genre, Review\nfrom .forms import ReviewForm\nfrom django.views.decorators.http import require_POST\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponseForbidden,HttpResponse,JsonResponse\nimport requests\nfrom django.core import serializers\nimport pprint\nfrom django.db.models import Avg, Max, Min, Sum\nfrom django.contrib.auth import get_user_model\nimport json\nfrom django.contrib import messages\n\n# Create your views here.\ndef start(request):\n return render(request,'movies/start.html')\n\ndef index(request, page_type):\n movies = Movie.objects.all()\n index_movies = Movie.objects.filter(movie_type=\"now_playing\").order_by('-popularity')[:10] \n # user_rated_movies = Movie.objects.order_by(\"-avgscore\")[:10]\n print(index_movies)\n if page_type == 1:\n top_rated_movies = Movie.objects.filter(movie_type=\"top_rated\").order_by('-popularity')[:10]\n context = {\n 'movies': top_rated_movies,\n }\n elif page_type == 2:\n popul_movies = Movie.objects.filter(movie_type=\"popular\").order_by('-popularity')[:10]\n context = {\n 'movies': popul_movies,\n }\n # elif page_type == 3:\n # if request.user.is_authenticated:\n # for movie in request.user.like_movies.all():\n # url = f'https://api.themoviedb.org/3/movie/{movie.id}/recommendations?api_key=1dfd52c8a24a0f38f40efe41c86be13b&language=ko-KR&page=1'\n # response = requests.get(url).json()\n # pprint.pprint(response[\"results\"])\n # movies = response[\"results\"]\n # pprint.pprint(movies)\n # context = {\n # 'movies': movies\n # }\n # else:\n # context = {\n # 'movies': index_movies\n # }\n else:\n context = {\n 'movies': index_movies\n }\n return render(request,'movies/index.html',context)\n\ndef detail(request,movie_pk):\n movie = get_object_or_404(Movie,pk=movie_pk)\n reviewform = ReviewForm()\n avg_score = 0\n reviews = Review.objects.filter(movie_id=movie_pk)\n\n url = f'https://api.themoviedb.org/3/movie/{movie_pk}/recommendations?api_key=1dfd52c8a24a0f38f40efe41c86be13b&language=ko-KR&page=1'\n response = requests.get(url).json()\n pprint.pprint(response[\"results\"])\n new_movies = response[\"results\"]\n # print(reviews.avgscore)\n # print(reviews.filter(movie_id=movie_pk).aggregate(Avg('score')))\n for review in reviews:\n avg_score += review.score\n if reviews.count() > 0:\n avg_score /= reviews.count()\n avg_score = round(avg_score, 1)\n else:\n avg_score = 0\n peoples = []\n casts = movie.cast.all()\n for cast in casts:\n people = People.objects.filter(id=cast.id).first()\n if people:\n peoples.append(people)\n context = {\n 'movie' : movie,\n 'new_movies': new_movies,\n 'form' : reviewform,\n 'avg_score': avg_score,\n 'peoples': peoples\n }\n return render(request,'movies/detail.html', context)\n\n@login_required\ndef review(request, movie_pk):\n reviews = Review.objects.filter(movie_id=movie_pk)\n movie = get_object_or_404(Movie,pk=movie_pk)\n reviewForm = ReviewForm(request.POST)\n if reviewForm.is_valid():\n review = reviewForm.save(commit=False)\n review.movie_id = movie_pk\n review.user = request.user\n # movie.avgscore = reviews.aggregate(Avg('score'))\n # print(movie.avgscore)\n review.save()\n return redirect('movies:detail', movie_pk)\n context = {\n 'movie' : movie, \n 'form' : reviewForm\n }\n return render(request,'movies/detail.html', context)\n \n@require_POST\ndef review_delete(request, movie_pk, review_pk):\n review = Review.objects.get(pk=review_pk)\n if request.user == review.user:\n review.delete()\n messages.success(request, '댓글이 삭제되었습니다.')\n return redirect('movies:detail', movie_pk)\n else:\n return HttpResponseForbidden\n\ndef review_update(request, movie_pk, review_pk):\n review = get_object_or_404(Review, pk=review_pk)\n review.content = request.POST.get('content') or review.content or ''\n review.score = request.POST.get('score') or review.score\n review.save()\n return redirect('movies:detail',movie_pk)\n\n@login_required\ndef like(request, movie_pk):\n if request.is_ajax():\n movie = get_object_or_404(Movie, pk=movie_pk)\n is_liked = True\n if request.user in movie.like_users.all():\n movie.like_users.remove(request.user)\n is_liked = False\n else:\n movie.like_users.add(request.user)\n is_liked = True\n context = {\n 'is_liked' : is_liked,\n 'like_count' : movie.like_users.count()\n }\n return JsonResponse(context)\n else:\n return HttpResponseForbidden\n\n\ndef recommand(request):\n movie = Movie.objects.all()\n context = {\n 'movies' : movies\n }\n return render(request,'movies/recommand.html',context)\n \ndef search(request):\n query = request.GET.get('search_title')\n if query:\n title_movies = Movie.objects.filter(title__icontains=query)\n description_movies = Movie.objects.filter(description__icontains=query)\n des_movies = description_movies.difference(title_movies)\n context = {\n \"title_movies\" : title_movies,\n \"des_movies\" : des_movies\n }\n return render(request,'movies/search.html',context)\n else:\n return redirect('movies:index', 0)\n","sub_path":"movies/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"570593505","text":"import json\n\nclass config:\n \n def __init__(self):\n with open(\"config.json\") as json_file:\n data = json.load(json_file)\n self.listen_ip = data[\"listen_ip\"]\n self.listen_port = data[\"listen_port\"]\n self.apikey = data[\"apikey\"]\n self.redis_task_ttl = data[\"redis_task_ttl\"]\n self.redis_server = data[\"redis_server\"]\n self.redis_port = data[\"redis_port\"]\n self.redis_core_q = data[\"redis_core_q\"]\n self.redis_task_timeout = data[\"redis_task_timeout\"]\n self.txtfsm_index_file = data[\"txtfsm_index_file\"]\n self.txtfsm_template_server = data[\"txtfsm_template_server\"]\n","sub_path":"backend/core/confload/confload.py","file_name":"confload.py","file_ext":"py","file_size_in_byte":695,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"305805329","text":"def intersection(arrays):\n \"\"\"\n YOUR CODE HERE\n \"\"\"\n # initializing a dict to hold the key value pairs\n dict = {}\n # can be up to 10 lists in all lists\n number_of_lists = len(arrays)\n for each_list in arrays: # for each of the lists from our input (arrays)\n for each_item in each_list: # for each item of each list\n if each_item in dict: # if the item is already in the dictionary\n dict[each_item] += 1 # increase the key count\n else:\n dict[each_item] = 1 # add key count of 1 to dict if it isn't already there\n\n intersections = []\n # we have created a count of how many lists each item exists within\n for key in dict:\n # if it exists in all of them it would be equal to our \"number_of_lists\"\n if dict[key] == number_of_lists:\n # add the values that intersect to our intersection array\n intersections.append(key)\n\n return intersections\n\nif __name__ == \"__main__\":\n arrays = []\n\n arrays.append(list(range(1000000, 2000000)) + [1, 2, 3])\n arrays.append(list(range(2000000, 3000000)) + [1, 2, 3])\n arrays.append(list(range(3000000, 4000000)) + [1, 2, 3])\n\n print(intersection(arrays))\n ","sub_path":"hashtables/ex3/ex3.py","file_name":"ex3.py","file_ext":"py","file_size_in_byte":1236,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"47536688","text":"#!/home/odroid/prefix/default/lib python2\n'''\nCreated on Jan 19, 2018\n\n@author: enzo\n'''\nfrom src import *\nimport matplotlib.pyplot as plt\nimport mpl_toolkits.mplot3d.axes3d as axes3d\n \ndef main():\n # Init all parts of application\n try:\n endEvent = threading.Event()\n processEvent = threading.Event()\n processEvent.clear()\n fig = plt.figure(1) # Plot figure\n ax1 = fig.gca() # Plot data\n #ax1 = fig.add_subplot(111)\n data = Data() # The data class that contains all data\n dspProcces = dsp(data,endEvent,processEvent) # The signalproccesing block\n #time.sleep(5)\n app = Application(dspProcces,data,ax1,fig,endEvent) # Terminal application\n plotevent = app.getPlotEvent()\n \n while endEvent.isSet() == False:\n time.sleep(1)\n if plotevent.isSet():\n plotevent.clear()\n plt.show()\n \n\n except (KeyboardInterrupt,SystemExit):\n app.endprocess()\nif __name__ == '__main__':\n gpsp.start() # Global gps object start\n main()\n","sub_path":"AntChar_backup_180515/Antchar/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1093,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"623728686","text":"# Copyright 2015 gRPC authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"The Python implementation of the GRPC helloworld.Greeter server.\"\"\"\n\nfrom concurrent import futures\n\nimport grpc\n\nimport helloworld_pb2\nimport helloworld_pb2_grpc\n\nfrom datetime import datetime\n\nREPLY=\"done_1\"\nPORT=4444\n\nclass Greeter(helloworld_pb2_grpc.GreeterServicer):\n def SayHello(self, request, context):\n msg=request.name\n print(\"request from the client with the message: %s\" % msg)\n if (msg==\"time\"):\n return helloworld_pb2.HelloReply(message='%s' % time())\n else:\n return helloworld_pb2.HelloReply(message='%s' % refuse())\n def SayHelloAgain(self, request_iterator, context):\n for request in request_iterator:\n print(\"streaming: request from the client with the message %s\" % request.name)\n response=helloworld_pb2.HelloReply(message='streaming: %s! ' % REPLY)\n return response\n\ndef serve():\n server = grpc.server(futures.ThreadPoolExecutor(max_workers=10))\n helloworld_pb2_grpc.add_GreeterServicer_to_server(Greeter(), server)\n server.add_insecure_port('[::]:4444')\n print(\"Server 1 is listening on port 4444 ...\")\n server.start()\n server.wait_for_termination()\n\ndef time():\n now=datetime.now()\n return now.strftime(\"%H:%M:%S\")\n\ndef refuse():\n return \"time was not requested\"\n\nif __name__ == '__main__':\n serve()\n","sub_path":"grpc/server1.py","file_name":"server1.py","file_ext":"py","file_size_in_byte":1922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"493347100","text":"from PyQt4.QtCore import *\nfrom PyQt4.QtGui import *\n\nfrom _actions_ import LActions\nfrom _debug_ import *\nfrom setup import * \n\nimport qrc_resources\n\nif not vars().has_key('ACTIONS'): ACTIONS = 0\n\n#----------------------------------------------------------------------\nclass LMyActions(LActions):\n\n def __init__(self, dataContainer=None, parent=None):\n super(LMyActions, self).__init__(dataContainer, parent)\n self.dataContainer = dataContainer \n self.createActions()\n self.createMenus()\n\n def createActions(self): \n super(LMyActions, self).createActions()\n self.insertARowAction = self.createAction(\"&Insert New Condition\", self.on_insertARowAction_triggered)\n self.removeARowAction = self.createAction(\"&Remove Current Condition\", self.on_removeARowAction_triggered)\n\n def createMenus(self):\n super(LMyActions, self).createMenus()\n self.contextMenuActions = [ self.insertARowAction, self.removeARowAction, ] + self.contextMenuActions\n self.contextMenu = self.getContextMenu(\"Condition Table\")\n\n #----------------------------------------------------------------------\n # Signals/Slots\n\n def on_insertARowAction_triggered(self):\n ''' Insert row '''\n numberOfRows = self.parent().tableModel.rowCount(self)\n self.parent().tableModel.insertRow(numberOfRows)\n\n def on_removeARowAction_triggered(self):\n ''' Remove row '''\n index =self.parent().tableView.currentIndex()\n row = index.row()\n self.parent().tableModel.removeRow(row)\n\n#----------------------------------------------------------------------\n# MAIN\n#----------------------------------------------------------------------\n\nif __name__ == \"__main__\":\n pass\n \n","sub_path":"cls1/eggs/src/lti/widgets.20120401/conditiontable/actions.py","file_name":"actions.py","file_ext":"py","file_size_in_byte":1764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"99580908","text":"\r\nitems = input().split('|')\r\n\r\nbudget = float(input())\r\n\r\nsum_new_prices = 0\r\nnew_budget = 0\r\nsum_old_prices = 0\r\ntotal_profit = 0\r\nbought_products_list = []\r\n\r\nfor i in items:\r\n tokens = i.split('->')\r\n cloth_type = tokens[0]\r\n price = float(tokens[1])\r\n\r\n if cloth_type == 'Clothes' and price <= 50:\r\n if budget >= price:\r\n budget -= price\r\n sum_old_prices += price\r\n price = price * 1.40\r\n sum_new_prices += price\r\n bought_products_list.append(price)\r\n\r\n if cloth_type == 'Shoes' and price <= 35:\r\n if budget >= price:\r\n budget -= price\r\n sum_old_prices += price\r\n price = price * 1.40\r\n sum_new_prices += price\r\n bought_products_list.append(price)\r\n\r\n if cloth_type == 'Accessories' and price <= 20.50:\r\n if budget >= price:\r\n budget -= price\r\n sum_old_prices += price\r\n price = price * 1.40\r\n sum_new_prices += price\r\n bought_products_list.append(price)\r\n\r\ntotal_profit = sum_new_prices - sum_old_prices\r\nnew_budget = budget + sum_new_prices\r\n\r\n\r\nfor item in bought_products_list:\r\n print(f'{item:.2f}', end=' ')\r\nprint()\r\nprint(f'Profit: {total_profit:.2f}')\r\n\r\nif new_budget >= 150:\r\n print('Hello, France!')\r\nelse:\r\n print('Time to go.')\r\n","sub_path":"Python Fundamentals 2020 - 2021/03 - LISTS BASIC/Exercise - 04 - 09.py","file_name":"Exercise - 04 - 09.py","file_ext":"py","file_size_in_byte":1361,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"232053656","text":"# -*- coding: utf8 -*-\n# @author: yinan\n# @time: 18-7-17 下午9:38\n# @filename: util.py\nimport base64\n\nimport os\n\nfrom application import app\nimport time\nimport re\nfrom io import BytesIO\n\nfrom PIL import Image\nfrom qcloud_cos import CosConfig, CosS3Client\n\nfrom application import configs\n\n\nclass CommonUtil(object):\n @staticmethod\n def handle_img(base64_str, filename):\n image_path = configs.SYS_UPLOAD_PATH + filename\n base64_data = re.sub('^data:image/.+;base64,', '', base64_str)\n byte_data = base64.b64decode(base64_data)\n image_data = BytesIO(byte_data)\n img = Image.open(image_data)\n if image_path:\n img.convert(\"RGB\").save(image_path, 'WEBP')\n return img\n\n @staticmethod\n def upload_img(filename):\n tencent_config = configs.TENCENT_OAUTH\n image_path = configs.SYS_UPLOAD_PATH + filename\n config = CosConfig(Region=tencent_config.get('region'), Secret_id=tencent_config.get('secret_id'),\n Secret_key=tencent_config.get('secret_key')) # 获取配置对象\n client = CosS3Client(config)\n remote_name = str(int(time.time())) + '.webp'\n remote_url = 'http://{}.cosgz.myqcloud.com/{}'.format(tencent_config.get('bucket'), remote_name)\n response = None\n try:\n with open(image_path, 'rb') as fp:\n response = client.put_object(\n Bucket=tencent_config.get('bucket'),\n Body=fp.read(),\n Key=remote_name\n )\n os.remove(image_path)\n except Exception as e:\n app.logger.error('upload file to tencent error: {}'.format(e))\n if response is not None and response['ETag'] is not None:\n return remote_url\n else:\n return None\n","sub_path":"application/constant/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":1830,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"399944320","text":"\n\"\"\"\nCreated on Thu Jan 10 3:08:09 2019\n@author: Deloitte Consutling, LLP\n@credits: [Donnally, Ryan rdonnally@deloitte.com,\n Michael Martin michmartin@deloitte.com,\n Whitehead, Noah nowhitehead@deloitte.com,\n Miller, Taylor taymiller@deloitte.com,\n Houdeshell, Benjamin bhoudeshell@deloitte.com,\n Han, Bryan bryhan@deloitte.com,\n Jackson, Amina amijackson@deloitte.com,\n Li, Yingquan yingquli@deloitte.com,\n Barry, Ande elbarry@deloitte.com,\n Grubb, Christopher cgrubb@deloitte.com,\n Stevens, Isaac isstevens@deloitte.com,\n Barsness, Brendan bbarsness@deloitte.com,\n Haddad, Abigail abhaddad@deloitte.com,\n Maroon, Christopher cmaroon@deloitte.com]\n@version: 0.0.1\n@maintainer: Ryan Donnally rdonnally@deloitte.com\n@compatability: Python 3.7, Pandas 0.23.4\n\n\"\"\"\n\nimport pandas as pd\n\n\n###########################################################################\n### Numeric correlations function that returns the top n column correlations\n\ndef get_correlation(df):\n # correlation_table=df.corr(method='kendall')\n z = df.corr().abs() # gets absolute values of a correlation table\n s = z.unstack() # unstacks the data into a single row\n corr_sort = s.sort_values(ascending=False, kind=\"quicksort\") # sorts the values from high to low\n new_corr_sort = (corr_sort[corr_sort != 1]).head(10) # drops the values that are 1 to 1; shows top 10 values\n return new_corr_sort\n\n\n### Show the Sample Data\ndef get_sample_of_cols(self, df):\n if self.get_columns(df) < 6:\n X = df.head()\n else:\n X = df.head().iloc[:, :5]\n return X\n\n\ndef get_fields_and_types(df):\n \"\"\"Returns a dictionary of dataframe columns and dtypes as strings\n\n :param df:\n :return:\n \"\"\"\n pass\n\n\ndef get_variable_type_counts(df):\n \"\"\"returns a dict of type: type_count key: value pairs\"\"\"\n count = df.dtypes.value_counts()\n type_count_dict = {}\n for i in count.index:\n type_count_dict[str(i)] = count[i]\n\n return type_count_dict\n\ndef get_type_counts_with_numeric(df):\n \"\"\"Returns a dict of types and type counts and includes a\n numeric type that is the sum of counts of float columns and int columns\n\n :param df:\n :return:\n \"\"\"\n type_count_dict = get_variable_type_counts(df)\n\n _numeric = []\n for key in _type_count_dict.keys():\n if 'float' in key or 'int' in key:\n _numeric.append(type_count_dict[key])\n\n _total_numeric = sum(_numeric)\n\n type_count_dict['numeric'] = _total_numeric\n\n return type_count_dict\n\n###########################################################################\n###\ndef get_data_summary(df):\n dataset_vars = {\n 'column_names': list(df.columns),\n 'col_type_dict': dict(df.dtypes),\n \"count_rows\": len(df.index),\n \"count_columns\": len(df.columns),\n \"count_missing_values\": df.isnull().sum().sum(), # sum over rows, then over columns,\n \"memoryusage_df\": df.memory_usage(index=False, deep=True).sum(),\n \"memoryusage_avg_col\": df.memory_usage(index=False, deep=True).sum()/len(df.columns),\n # \"Correlation Matrix\": self.get_correlation(df),\n\n # Why do we want to report this\n \"count_unique_values\": df.nunique().sum(),\n \"count_unique_vals_per_col\": df.nunique(axis=0),\n \"percent_unique_values\": df.nunique().sum() / (len(df.index) * len(df.columns)),\n \"percent_unique_per_col\": df.nunique(axis=0) / len(df.index),\n # \"Sample of the Data\": self.get_SampleOfData(df),\n # \"Column Names\": self.get_header_names(df),\n \"count_variable_type\": get_variable_type_counts(df),\n \"percent_missing_values\": df.isnull().sum().sum() / (len(df.index) * len(df.columns))\n }\n return dataset_vars\n\n\n\n\n###########################################################################\n### Function to convert dates into their date decimal values.\n\ndef conversion_to_decimal(date):\n return (float(date.strftime(\"%j\")) - 1) / 366 + float(date.strftime(\"%Y\"))\n\n###########################################################################\n### Function to convert date decimals to their datetime equivalents.\n\ndef conversion_to_date(date):\n year = int(date)\n rem = date - year\n\n base = datetime(year, 1, 1)\n print(base + timedelta(seconds=(base.replace(year=base.year + 1) - base).total_seconds() * rem))\n\n\ndef get_outlier_info(self, pandas_series,\n low_quantile = 0.25,\n high_quantile = 0.75,\n iqr_scale = 1.5):\n quartile_1 = pandas_series.quantile(low_quantile)\n quartile_3 = pandas_series.quantile(high_quantile)\n iqr = quartile_3 - quartile_1\n lower_bound = quartile_1 - (iqr * iqr_scale)\n upper_bound = quartile_3 + (iqr * iqr_scale)\n\n high_count = pandas_series[pandas_series > upper_bound].count()\n low_count = pandas_series[pandas_series < lower_bound].count()\n\n outlier_dict = {'High Cutoff': upper_bound,\n 'Low Cutoff': lower_bound,\n 'High Count': high_count,\n 'Low Count': low_count}\n\n return outlier_dict\n\n\n\n# @TODO: add imputation methods e.g. mean, Gaussian sample, regression, etc.\n# @TODO: sort out why there should separate outlier calculations for different dtypes\ndef outlier_detection(pd_column, impute=True, impute_type = 'mode'):\n \"\"\"\n\n :param pd_column: column from a pandas dataframe\n :param impute: (bool) default = True, whether missing data will be imputed\n :param impute_type:\n :return:\n \"\"\"\n if 'int' in str(pd_column.dtype) or 'float' in (str(pd_column.dtype)):\n if impute:\n if impute_type == 'mode':\n pd_column.fillna(pd.mode()[0], inplace=True)\n outlier_dict = get_outlier_info(pd_column)\n return outlier_dict\n\n else:\n outlier_dict = get_outlier_info(pd_column)\n return outlier_dict\n\n else:\n pass\n\n return outlier_dict\n\n# Support Functions for Datetime\n\n###########################################################################\n### This function is called by datetime_stats - it finds the largest value (year,month,day etc) that changes over the field.\n### Inputs are the min/max from a datetime field, output is a string specifying the largest value that changes\n\ndef largest_value_change(value_min, value_max):\n ValueRange = value_max - value_min\n if ValueRange.days > 365:\n LargeValue = 'Years'\n elif ValueRange.days <= 365 and ValueRange.days > 31:\n LargeValue = 'Months'\n elif ValueRange.days <= 365 and ValueRange.seconds > 86400:\n LargeValue = 'Days'\n elif ValueRange.days == 0 and ValueRange.seconds <= 86400 and ValueRange.seconds > 3600:\n LargeValue = 'Hours'\n elif ValueRange.days == 0 and ValueRange.seconds <= 3600 and ValueRange.seconds > 60:\n LargeValue = 'Minutes'\n elif ValueRange.days == 0 and ValueRange.seconds <= 60 and ValueRange.dt.second > 0:\n LargeValue = 'Seconds'\n else:\n LargeValue = 'Less Than 1 Second'\n return LargeValue\n\n###########################################################################\n### Creates a dict with a frequency distribution of dates by month\n\ndef month_count_dict_create(Field):\n MonthCountDict = {'Jan': 0, 'Feb': 0, 'Mar': 0, 'Apr': 0, 'May': 0, 'Jun': 0, 'Jul': 0, 'Aug': 0, 'Sep': 0,\n 'Oct': 0, 'Nov': 0, 'Dec': 0}\n RefDict = {1: 'Jan', 2: 'Feb', 3: 'Mar', 4: 'Apr', 5: 'May', 6: 'Jun', 7: 'Jul', 8: 'Aug', 9: 'Sep', 10: 'Oct',\n 11: 'Nov', 12: 'Dec'}\n MonthCountDictPrelim = Field.dt.month.value_counts()\n for key in MonthCountDictPrelim.index:\n MonthCountDict[RefDict[key]] = MonthCountDictPrelim[key]\n return MonthCountDict\n\n###########################################################################\n### Creates a dict with a frequency distribution of dates by day of week\ndef weekday_count_dict_create(Field):\n WeekdayCountDict = {'Mon': 0, 'Tue': 0, 'Wed': 0, 'Thu': 0, 'Fri': 0, 'Sat': 0, 'Sun': 0}\n RefDict = {0: 'Mon', 1: 'Tue', 2: 'Wed', 3: 'Thu', 4: 'Fri', 5: 'Sat', 6: 'Sun'}\n WeekdayCountDictPrelim = Field.dt.weekday.value_counts()\n for key in WeekdayCountDictPrelim.index:\n WeekdayCountDict[RefDict[key]] = WeekdayCountDictPrelim[key]\n return WeekdayCountDict\n\n########################################################################\n### Primary function to run object field analysis\n\ndef text_stats(Field):\n # print(Field)\n CleanField = Field[Field.notnull()]\n text_len = CleanField.apply(lambda x: len(str(x)))\n word_count_split = CleanField.apply(lambda x: str(x).split())\n word_count_len = word_count_split.apply(lambda x: len(str(x)))\n word_list = [x for row in word_count_split for x in row]\n word_df = pd.DataFrame(word_list, columns=['Words'])\n word_counts = word_df['Words'].value_counts()\n counts = CleanField.value_counts()\n \"\"\"\n if len(counts) == 1:\n mode_value = counts\n\n else:\n mode_value = 'multiple'\n\n if word_counts[0] != word_counts[1] or len(counts) == 1:\n mode_word = word_counts.keys()[0]\n else:\n mode_word = 'multiple'\n \"\"\"\n special_character = 0\n for x in word_list:\n if x.isalpha() == False and x.isdigit() == False:\n special_character = special_character + 1\n\n TextDict = {}\n TextDict['name_field'] = Field.name\n TextDict['type_category'] = 'object'\n TextDict['dtype'] = Field.dtype\n TextDict['count_records'] = len(CleanField)\n TextDict['memoryusage_col'] = Field.memory_usage(deep=True)\n TextDict['count_missing'] = Field.isna().sum()\n TextDict['count_unique'] = Field.nunique()\n if TextDict['count_records'] > 0:\n TextDict['percent_missing'] = TextDict['count_missing'] / TextDict['count_records']\n TextDict['percent_unique'] = TextDict['count_unique'] / TextDict['count_records']\n else:\n TextDict['percent_missing'] = 0\n TextDict['percent_unique'] = 0\n TextDict['max'] = [text_len.max(), word_count_len.max()]\n TextDict['min'] = [text_len.min(), word_count_len.min()]\n TextDict['mean'] = [text_len.mean(), word_count_len.mean()]\n TextDict['mode'] = [text_len.mode()[0], word_count_len.mode()[0]]\n TextDict['q_25'] = [text_len.quantile(q=.25), word_count_len.quantile(q=.25)]\n TextDict['q_50'] = [text_len.quantile(q=.50), word_count_len.quantile(q=.50)]\n TextDict['q_75'] = [text_len.quantile(q=.75), word_count_len.quantile(q=.75)]\n TextDict['median_length'] = [text_len.median(), word_count_len.median()]\n TextDict['std_length'] = [text_len.std(), word_count_len.std()]\n TextDict['mad'] = [text_len.mad(), word_count_len.mad()]\n TextDict['sem'] = [text_len.sem(), word_count_len.sem()]\n TextDict['var'] = [text_len.var(), word_count_len.var()]\n TextDict['kurt'] = [text_len.kurtosis(), word_count_len.kurtosis()]\n TextDict['skew'] = [text_len.skew(), word_count_len.skew()]\n TextDict['autocorr'] = [text_len.autocorr(), word_count_len.autocorr()]\n\n TextDict['Total_Word_Count'] = sum(x.isalpha() for x in word_list)\n TextDict['Total_Number_Count'] = sum(x.isdigit() for x in word_list)\n TextDict['Total_Special_Count'] = special_character\n TextDict['word_list'] = word_list\n\n return (TextDict)\n\n###########################################################################\n### Primary function to run datetime field analysis\n\ndef datetime_stats(Field):\n CleanField = Field[Field.notnull()]\n # FieldNum = len(Field)\n DatetimeDict = {}\n\n DatetimeDict['memoryusage_col'] = Field.memory_usage(deep=True)\n DatetimeDict['name_field'] = Field.name\n DatetimeDict['type_category'] = 'datetime'\n DatetimeDict['dtype'] = Field.dtype\n DatetimeDict['count_records'] = len(CleanField)\n DatetimeDict['count_missing'] = Field.isna().sum()\n DatetimeDict['count_unique'] = Field.nunique()\n if DatetimeDict['count_records'] > 0:\n DatetimeDict['percent_missing'] = DatetimeDict['count_missing'] / DatetimeDict['count_records']\n DatetimeDict['percent_unique'] = DatetimeDict['count_unique'] / DatetimeDict['count_records']\n else:\n DatetimeDict['percent_missing'] = 0\n DatetimeDict['percent_unique'] = 0\n\n\n DatetimeDict['min'] = CleanField.min()\n DatetimeDict['max'] = CleanField.max()\n DatetimeDict['range'] = (CleanField.max() - CleanField.min())\n NumDates = CleanField.astype('int64')\n DatetimeDict['std'] = pd.to_timedelta(NumDates.std())\n DatetimeDict['kurt'] = NumDates.kurtosis()\n DatetimeDict['skew'] = NumDates.skew()\n DatetimeDict['mean'] = pd.to_datetime(NumDates.mean())\n DatetimeDict['median'] = pd.to_datetime(NumDates.median())\n DatetimeDict['mode'] = str(CleanField.mode()[0])\n DatetimeDict['q_25'] = pd.to_datetime(NumDates.quantile(q=.25))\n DatetimeDict['q_50'] = pd.to_datetime(NumDates.quantile(q=.50))\n DatetimeDict['q_75'] = pd.to_datetime(NumDates.quantile(q=.75))\n DatetimeDict['iqr'] = DatetimeDict['q_75'] - DatetimeDict['q_25']\n # DatetimeDict['YearCountDict'] = CleanField.dt.year.value_counts().to_dict()\n # DatetimeDict['MonthCountDict'] = self.month_count_dict_create(CleanField)\n # DatetimeDict['DayCountDict'] = CleanField.dt.day.value_counts().to_dict()\n # DatetimeDict['WeekdayCountDict'] = self.weekday_count_dict_create(CleanField)\n # This is the largest value that changes over the series, from years to days\n DatetimeDict['laregest_measureable_value'] = largest_value_change(DatetimeDict['min'], DatetimeDict['max'])\n # DatetimeDict['outlier'] = self.outlier_detection(Field,'Date')\n return DatetimeDict\n\n###########################################################################\n### Primary function to run numeric field analysis\n\ndef numeric_stats(Field):\n CleanField = Field[Field.notnull()]\n NumericDict = {}\n\n NumericDict['memeryusage_col'] = Field.memory_usage(deep=True)\n NumericDict['name_field'] = Field.name\n NumericDict['type_category'] = 'numeric'\n NumericDict['dtype'] = Field.dtype\n NumericDict['count_records'] = len(CleanField)\n NumericDict['count_missing'] = Field.isna().sum()\n NumericDict['count_unique'] = Field.nunique()\n if NumericDict['count_records'] > 0:\n NumericDict['percent_missing'] = NumericDict['count_missing'] / NumericDict['count_records']\n NumericDict['percent_unique'] = NumericDict['count_unique'] / NumericDict['count_records']\n else:\n NumericDict['percent_missing'] = 0\n NumericDict['percent_unique'] = 0\n\n\n NumericDict['min'] = CleanField.min()\n NumericDict['max'] = CleanField.max()\n NumericDict['median'] = CleanField.median()\n NumericDict['mean'] = CleanField.mean()\n NumericDict['std'] = CleanField.std()\n NumericDict['mad'] = CleanField.mad()\n NumericDict['sem'] = CleanField.sem()\n NumericDict['var'] = CleanField.var()\n NumericDict['q_25'] = CleanField.quantile(q=.25)\n NumericDict['q_50'] = CleanField.quantile(q=.5)\n NumericDict['q_75'] = CleanField.quantile(q=.75)\n NumericDict['kurt'] = CleanField.kurtosis()\n NumericDict['skew'] = CleanField.skew()\n # Need to come up with how many Modes we will be accepting. Something \"ID\" is unique therefore shows VERY long list.\n NumericDict['mode'] = CleanField.mode()[0]\n # NumericDict['outlier'] = self.outlier_detection(Field,'Numeric')\n return NumericDict\n\n###########################################################################\n### The primary Variable function - Dataframe in, Dictionary of fields and stats out. the output for each type is a series\n###vbut can be easily made into a dictionary by changeing the variable_stats_dict to call the dictionary form of each type.\n\ndef variable_stats(dataframe):\n\n df_numeric = dataframe.select_dtypes(include=['int64', 'float64'])\n numeric_stat = df_numeric.apply(lambda x: numeric_stats(x))\n numeric_stat_dict = numeric_stat.to_dict()\n\n df_datetime = dataframe.select_dtypes(include=['datetime'])\n datetime_stat = df_datetime.apply(lambda x: datetime_stats(x))\n datetime_stat_dict = datetime_stat.to_dict()\n\n df_text = dataframe.select_dtypes(include=['object'])\n text_stat = df_text.apply(lambda x: text_stats(x))\n text_stat_dict = text_stat.to_dict()\n\n variable_stats_dict = {'Numeric': numeric_stat, 'Datetime': datetime_stat, 'Text': text_stat}\n\n return variable_stats_dict\n\n###############################################################################\n\n\n","sub_path":"analysis.py","file_name":"analysis.py","file_ext":"py","file_size_in_byte":16562,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"542073428","text":"import pandas as pd\r\nimport datetime as dt\r\nimport os\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\ndef get_folder_name(keyword):\r\n data_container = []\r\n for foldername in os.listdir():\r\n if keyword in foldername:\r\n data_container.append(foldername)\r\n else:\r\n pass\r\n return data_container\r\n\r\ndef read_all_csv(folder_list, selected_column):\r\n data_container = []\r\n total_files = len(folder_list)\r\n counter = 1\r\n for i in folder_list:\r\n df = pd.read_csv(i + '/OpenStudio/' + i + '/ModelToIdf/in.csv', usecols=selected_column)\r\n df['Date/Time'] = create_time()\r\n df1 = df.set_index(['Date/Time'])\r\n df1.index = pd.to_datetime(df1.index)\r\n df1.columns = [i]\r\n data_container.append(df1)\r\n\r\n print(f'{counter}/{total_files} DataFrame created, {total_files - counter} more files')\r\n counter +=1\r\n print('Generating DataFrame done\\nConcatenating...')\r\n dataframe = pd.concat(data_container, axis=1)\r\n print(\"DataFrame ready\")\r\n return dataframe\r\n\r\ndef create_time():\r\n date_time_column = []\r\n base_time = dt.datetime(2019, 1, 1)\r\n timedelta = dt.timedelta(hours=1)\r\n for i in range(8760):\r\n date_time_column.append(str(base_time))\r\n base_time +=timedelta\r\n return date_time_column\r\n\r\n#def plot_option(df):\r\n# resample = ''\r\n# while (resample is not 'daily') or (resample is not 'monthly') or (resample is not 'hourly'):\r\n# resample = input(\"Choose your resample mapping (hourly, daily, or monthly): \")\r\n# if resample == 'hourly':\r\n# axes = df.resample(\"H\").mean().plot(figsize=(18,6))\r\n# return axes\r\n# elif resample == 'daily':\r\n# axes = df.resample(\"D\").mean().plot(figsize=(18,6))\r\n# return axes\r\n# elif resample == 'monthly':\r\n# axes = df.resample(\"M\").mean().plot(figsize=(18,6))\r\n# return axes\r\n\r\nif __name__ == '__main__':\r\n temperature_column = ['Date/Time', 'ROOM1:Zone Mean Air Temperature [C](Hourly)']\r\n file = get_folder_name('00')\r\n df = read_all_csv(file, temperature_column)\r\n ax = df.resample(\"D\").mean().plot(figsize=(18,6))\r\n ax.legend(loc=2)\r\n ax.set_title(label='Far East Flora Dome Indoor Air Temperature Comparison', loc='center')\r\n ax.set_xlabel(xlabel='Date/Time')\r\n ax.set_ylabel(ylabel='Celsius Degree')\r\n\r\n plt.show()","sub_path":"TemperaturePlot.py","file_name":"TemperaturePlot.py","file_ext":"py","file_size_in_byte":2411,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"234745708","text":"# main jobOption\n\n\nfrom AthenaCommon.AthenaCommonFlags import athenaCommonFlags\nathenaCommonFlags.FilesInput=[\"root://eosatlas//eos/atlas/atlascerngroupdisk/trig-daq/validation/test_data/valid1.117050.PowhegPythia_P2011C_ttbar.merge.HITS.e2658_s1967_s1964/RDO.0119996._000032.pool.root.1\"]\n\njp.AthenaCommonFlags.PoolESDOutput=\"theoutput_EFonly.ESD.pool.root\"\nif not 'rec.doWriteESD' in dir():\n rec.doWriteESD=True\n\njp.AthenaCommonFlags.EvtMax=-1\n\nrec.doWriteRDO=False\nrec.doWriteAOD=False \nrec.doAOD=False \nrec.doWriteTAG=False \nrec.doTrigger=True\n\n\ndef muonOnly():\n TriggerFlags.Slices_all_setOff()\n# TriggerFlags.MuonSlice.setAll()\n TriggerFlags.MuonSlice.signatures=[['mu0noL1_fsperf', '', [], ['Muon'], ['RATE:SingleMuon', 'BW:Muon'],-1]]\n\nfrom TriggerJobOpts.TriggerFlags import TriggerFlags\nTriggerFlags.doHLT=True\n\nfrom TriggerMenu.menu.GenerateMenu import GenerateMenu\nGenerateMenu.overwriteSignaturesWith(muonOnly)\n\n\n# main jobOption\n\ninclude (\"RecExCommon/RecExCommon_topOptions.py\")\n\nStoreGateSvc = Service( \"StoreGateSvc\" )\nStoreGateSvc.Dump = True\n\nfrom MuonTruthAlgs.MuonTruthAlgsConf import MuonDetailedTrackTruthMaker\nfrom TrkTruthAlgs.TrkTruthAlgsConf import TrackTruthSelector\nfrom TrkTruthAlgs.TrkTruthAlgsConf import TrackParticleTruthAlg\n\ncols = [\"HLT_TrackCollection_forCB\",\"HLT_TrackCollection_MuonEFCombTrkTracks\"]\ncolsTP = [\"HLT_xAOD__TrackParticleContainer_MuonEFInfo_ExtrapTrackParticles\",\"HLT_xAOD__TrackParticleContainer_MuonEFInfo_CombTrackParticles\"]\n\ntopSequence += MuonDetailedTrackTruthMaker(name=\"TrigMuEFonlyDetailedTrackTruthMaker\", TrackCollectionNames = cols,OutputLevel=INFO )\n\nfor i in range(0, len(cols) ):\n topSequence += TrackTruthSelector(name= cols[i] + \"Selector\",\n DetailedTrackTruthName = cols[i] + \"Truth\",\n OutputName = cols[i] + \"Truth\",OutputLevel=INFO )\n topSequence += TrackParticleTruthAlg(name = cols[i]+\"TruthAlg\",\n TrackTruthName=cols[i]+\"Truth\",\n TrackParticleName = colsTP[i],OutputLevel=INFO )\nfrom MuonTruthAlgs.MuonTruthAlgsConf import Muon__MuonTruthDecorationAlg\ntopSequence += Muon__MuonTruthDecorationAlg(\"EFonlyMuonTruthDecorationAlg\",\n MuonTruthParticleContainerName = \"HLT_EFMuonTruthParticle\",\n CreateTruthSegments = False,OutputLevel=INFO)\nfrom MuonTruthAlgs.MuonTruthAlgsConf import MuonTruthAssociationAlg\ntopSequence += MuonTruthAssociationAlg(\"EFonlyMuonTruthAssociation\",\n MuonContainerName = \"HLT_xAOD__MuonContainer_MuonEFInfo\",\n MuonTruthParticleContainerName = \"HLT_EFMuonTruthParticle\",OutputLevel=INFO)\n\ndoMon = True\nif( doMon ):\n from MuonPhysValMonitoring.MuonPhysValMonitoringConf import MuonPhysValMonitoring__MuonPhysValMonitoringTool\n tool1 = MuonPhysValMonitoring__MuonPhysValMonitoringTool(\"EFonlyMuonPhysValMonitoringTool\")\n tool1.EnableLumi = False\n tool1.DetailLevel = 1\n tool1.OutputLevel = INFO\n tool1.MuonContainerName = \"HLT_xAOD__MuonContainer_MuonEFInfo\"\n tool1.MuonTruthParticleContainerName = \"HLT_EFMuonTruthParticle\"\n \n ToolSvc += tool1\n\n from AthenaMonitoring.AthenaMonitoringConf import AthenaMonManager\n topSequence += AthenaMonManager( \"PhysValMonManager\" ,OutputLevel=DEBUG)\n\n from AthenaCommon.AppMgr import ServiceMgr\n from GaudiSvc.GaudiSvcConf import THistSvc\n ServiceMgr += THistSvc()\n svcMgr.THistSvc.Output += [\"EFonlyMuonMonExample DATAFILE='EFonlyMuonMonExample.root' OPT='RECREATE'\"]\n \n monMan = topSequence.PhysValMonManager\n monMan.ManualDataTypeSetup = True\n monMan.DataType = \"monteCarlo\"\n monMan.Environment = \"altprod\"\n monMan.ManualRunLBSetup = True\n monMan.Run = 1\n monMan.LumiBlock = 1\n \n monMan.FileKey = \"EFonlyMuonMonExample\"\n monMan.AthenaMonTools += [ tool1 ]\n\n\n\n","sub_path":"athena/Trigger/TrigValidation/TrigMuonValidation/share/TrigMuonValidation_RTT_options_EFonly.py","file_name":"TrigMuonValidation_RTT_options_EFonly.py","file_ext":"py","file_size_in_byte":4057,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"33805507","text":"from collections import namedtuple\nimport psycopg2\n\nBonus = namedtuple('Bonus',\n 'id,number,round,category_id,subcategory_id,quinterest_id,tournament_id,leadin,created_at,updated_at,errors_count,formatted_leadin')\nBonusPart = namedtuple('BonusPart',\n 'id,bonus_id,text,answer,formatted_text,formatted_answer,created_at,updated_at,number,wikipedia_url')\nTournament = namedtuple('Tournament', 'id,year,name,difficulty,quality,address,type,link,created_at,updated_at')\nTossup = namedtuple('Tossup', 'id,text,answer,number,tournament_id,category_id,subcategory_id,round,created_at,updated_at,'\n 'quinterest_id,formatted_text,errors_count,formatted_answer,wikipedia_url')\nconnection = None\n\ncategory_id_to_alias_map = {\n 'geo': 20,\n 'geography': 20,\n 'hist': 18,\n 'history': 18,\n 'lit': 15,\n 'literature': 15,\n 'm': 14,\n 'myth': 14,\n 'p': 25,\n 'philo': 25,\n 'r': 19,\n 'religion': 19,\n 'sci': 17,\n 'science': 17,\n 'ss': 22,\n 'socialscience': 22,\n 'trash': 16,\n 'ce': 26,\n 'currentevents': 26,\n 'fa': 21,\n 'finearts': 21\n}\n\nsubcategory_id_to_alias_map = {\n # 40: ['cea', 'ceamerican'],\n # 42: ['ceo', 'ceother'],\n # 35: ['faam', 'fineartsamerican'],\n # 27: ['faav', 'fineartsaudiovisual'],\n # 8: ['faa', 'fineartsauditory'],\n # 45: ['fab', 'fineartsbritish'],\n # 50: ['fae', 'fineartseuropean'],\n # 77: ['fao', 'opera', 'fineartsopera'],\n # 25: ['faot', 'fineartsother'],\n # 2: ['fav', 'fineartsvisual'],\n # 43: ['faw', 'fineartsworld'],\n # 38: ['geoa', 'geoamerican'],\n 'bio': 14,\n 'biology': 14,\n 'chem': 5,\n 'chemistry': 5,\n 'cs': 23,\n 'math': 26,\n 'physics': 18\n\n # 14: ['bio', 'biology'],\n # 5: ['chem', 'chemistry'],\n # 23: ['cs', 'compsci', 'computerscience'],\n # 26: ['math'],\n # 10: ['osci', 'otherscience'],\n # 18: ['physics']\n # TODO finish this later i'm lazy\n}\n\nclass GlobalState:\n sessions = []\n skip_message = None\n\nstate = GlobalState()\n\ndef get_global_state():\n global state\n return state\n\ndef get_db_connection():\n global connection\n if connection is None:\n connection = psycopg2.connect('dbname=quizdb user=postgres')\n return connection\n\n\ntournaments = []\ndef get_tournaments():\n global tournaments\n if len(tournaments) == 0:\n conn = get_db_connection()\n tournaments = read_tournaments(conn)\n return tournaments\n\ndef read_tournaments(conn):\n cursor = conn.cursor()\n cursor.execute('SELECT * FROM tournaments')\n return list(map(Tournament._make, cursor.fetchall()))\n\n\ntournament_series = {'scop': 'scop', 'pace': 'pace', 'rmbat': 'rmbat', 'bhsat': 'bhsat', 'acf_regs': 'acf regionals', 'acf_fall': 'acf fall', 'co': 'chicago open'}\n\n\ndef get_bonus_batch(conn, arguments):\n # Who needs ORMs anyway amirite? This is literally the jankiest shit i've ever written but it works so...\n # TODO: use an ORM\n\n difficulties = []\n categories = []\n subcategories = []\n selected_tournaments = []\n for arg in arguments:\n if arg[0].isdigit():\n # Difficulty or year\n # Check if range\n if len(arg.split('-')) > 1:\n values = arg.split('-')\n difficulties = list(range(int(values[0]), int(values[1]) + 1))\n elif 0 < int(arg) < 10:\n difficulties = [int(arg)]\n else:\n # assume year\n if selected_tournaments[-1][1] == -1:\n selected_tournaments[-1] = (selected_tournaments[-1][0], int(arg))\n else:\n # Category, subcategory, or tournament\n for k in subcategory_id_to_alias_map.keys():\n if arg == k:\n subcategories.append(str(subcategory_id_to_alias_map[k]))\n for k in category_id_to_alias_map.keys():\n if arg == k:\n categories.append(str(category_id_to_alias_map[k]))\n if arg in tournament_series:\n selected_tournaments.append((tournament_series[arg], -1))\n\n category_conditional = '('\n category_conditional += ' OR '.join([f'bonuses.category_id={category}' for category in categories])\n category_conditional += ')'\n\n difficulty_conditional = '('\n difficulty_conditional += ' OR '.join([f'tournaments.difficulty={difficulty}' for difficulty in difficulties])\n difficulty_conditional += ')'\n\n subcategory_conditional = '('\n subcategory_conditional += ' OR '.join([f'bonuses.subcategory_id={subcategory}' for subcategory in subcategories])\n subcategory_conditional += ')'\n\n matching_tournament_records = []\n\n for t in selected_tournaments:\n if t[1] == -1:\n matching_tournament_records.extend([i for i in get_tournaments() if t[0] in i.name.lower()])\n else:\n matching_tournament_records.extend([i for i in get_tournaments() if t[0] in i.name.lower() and str(t[1]) in i.name.lower()])\n\n if len(matching_tournament_records) == 0 and len(selected_tournaments) != 0:\n raise Exception('No tournament matches your query')\n\n tournament_conditional = '(' if len(matching_tournament_records) > 0 else ''\n tournament_conditional += ' OR '.join([f'tournaments.id={t.id}' for t in matching_tournament_records]) + ')'\n\n if len(difficulties) > 0:\n sql_command = 'SELECT bonuses.* FROM bonuses,tournaments WHERE bonuses.tournament_id=tournaments.id AND ' + difficulty_conditional\n if len(subcategories) > 0:\n sql_command += ' AND ' + subcategory_conditional\n if len(categories) > 0:\n sql_command += ' AND ' + category_conditional\n if len(selected_tournaments) > 0:\n sql_command += ' AND ' + tournament_conditional\n elif len(subcategories) > 0 or len(categories) > 0 or len(tournament_conditional) > 0:\n sql_command = 'SELECT bonuses.* FROM bonuses,tournaments WHERE bonuses.tournament_id=tournaments.id'\n if len(subcategories) > 0:\n sql_command += ' AND ' + subcategory_conditional\n if len(categories) > 0:\n sql_command += ' AND ' + category_conditional\n if len(selected_tournaments) > 0:\n sql_command += ' AND ' + tournament_conditional\n else:\n sql_command = 'SELECT * FROM bonuses'\n\n sql_command += \" ORDER BY RANDOM() LIMIT 20;\"\n print(f'Executing {sql_command}')\n return get_bonus_batch_raw(conn, sql_command)\n\n\ndef get_bonus_batch_raw(conn, sql_command):\n bonuses = []\n cursor = conn.cursor()\n cursor.execute(sql_command)\n for bonus in map(Bonus._make, cursor.fetchall()):\n cursor.execute(f\"SELECT * FROM bonus_parts WHERE bonus_id={bonus.id} ORDER BY number\")\n bonus_parts = list(map(BonusPart._make, cursor.fetchall()))\n if len(bonus_parts) == 0:\n # Unusable\n continue\n cursor.execute(f'SELECT * FROM tournaments WHERE id={bonus.tournament_id}')\n tournament = list(map(Tournament._make, cursor.fetchall()))\n assert len(tournament) == 1\n bonuses.append((bonus, bonus_parts, tournament[0]))\n\n return bonuses\n\n\n\n\n\n","sub_path":"pkbot-master/src/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":7185,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"239573308","text":"import pandas as pd\nimport numpy as np\n\n#declaration\n#get data from https://www.bka.de/DE/AktuelleInformationen/StatistikenLagebilder/PolizeilicheKriminalstatistik/PKS2018/Standardtabellen/standardtabellenTatverdaechtige.html?nn=108686\n#download excel\nfile = pd.read_excel(\"STD-TV-01-T20-Tatverdaechtige_excel.xlsx\" , usecols =\"B,C,D\", skiprows=8)\nsave = pd.DataFrame(columns=['Verbrechen'])\nlen = file.shape\ncrimes = [[],[],[],[],[]]\ntemp = 0\n\n\nfor i in range(0, len[0], 3):\n\n #evaluate percent\n \n if file.iloc[i+1,2] > 0:\n temp = file.iloc[i+1,2]/file.iloc[i+2,2]*100\n else:\n temp = 0\n\n #check percent change the range to whatever\n if temp>50 and temp<=100:\n crimes[0].append(str(file.iloc[i,0]))\n crimes[1].append(temp.round(2))\n crimes[2].append(file.iloc[i,2])\n crimes[3].append(file.iloc[i+1,2])\n crimes[4].append(file.iloc[i+2,2])\n \n#add to pandas dataframe \nsave['Verbrechen'] = crimes[0]\nsave['Prozentualer Anteil'] = crimes[1]\nsave['Anzahl Männer'] = crimes[2]\nsave['Anzahl Frauen'] = crimes[3]\nsave['Gesamt'] = crimes[4]\n\nprint(save)\nsave.to_csv('result.csv')","sub_path":"women_over_50.py","file_name":"women_over_50.py","file_ext":"py","file_size_in_byte":1150,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"637605191","text":"import sys\nfrom itertools import permutations\n\nN,M = map(int,input().split())\nDNAs = []\nfor _ in range(N):\n DNAs.append(input().rstrip())\n\nS=['A','C','G','T']\nresult_DNA = []\ndiff=0\nfor i in range(M):\n temp=[]\n max_c=0\n ans=''\n for dna in DNAs:\n temp.append(dna[i])\n\n for s in S:\n c = temp.count(s)\n if c > max_c:\n max_c = c\n ans=s\n \n for dna in DNAs:\n if dna[i] != ans:\n diff+=1\n\n result_DNA.append(ans)\n\nprint(''.join(result_DNA))\nprint(diff)\n\n","sub_path":"백준/Python/알고파/완전탐색,백트래킹/1969%(DNA).py","file_name":"1969%(DNA).py","file_ext":"py","file_size_in_byte":534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"407105","text":"import healpy as hp\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib import cm\nfrom matplotlib.colors import LogNorm\n\n# Set the number of sources and the coordinates for the input\n#nsources = int(2664001)\n#nsources = int(3149)\nnsources = int(1268174)\n\nnside = 64\nnpix = hp.nside2npix(nside)\n\n# npix= 12*nside^2\n# Coordinates and the density field f\n#thetas = np.random.random(nsources) * np.pi\n#phis = np.random.random(nsources) * np.pi * 2.\n\nfs = np.random.randn(nsources)\n\nwith open(\"./eventFile_data_noHead_eta08.txt\") as inputFile:\n lines = inputFile.readlines()\n #print (lines[1].split()[1])\nthetas=[]\nphis=[]\nfor i in range(nsources): \n thetas.append(float(lines[i+1].split()[1]))\n phis.append(float(lines[i+1].split()[2]))\n #if(thetas[0]<0 or thetas[0]>3.14): \n # print(\"theta out of range\")\n#print(thetas)\n\n# Go from HEALPix coordinates to indices\nindices = hp.ang2pix(nside, thetas, phis)\n\n# Initate the map and fill it with the values\nhpxmap = np.zeros(npix, dtype=np.float)\nfor i in range(nsources):\n #hpxmap[indices[i]] += fs[i]\n #hpxmap[indices[i]] += 1.0\n hpxmap[indices[i]] += 12.0*nside*nside/nsources\n#for j in range(npix):\n# if hpxmap[j] < 0.000001:\n# hpxmap[j] = 0.000001\n\nDPI = 100\nSIZE = 800\n\n# Inspect the map\n#plt.figure(1)\n\n#hp.mollview(hpxmap, xsize = SIZE)\n#map_ring = hp.pixelfunc.reorder(hpxmap, inp='NEST', out='RING')\nhp_smoothed = hp.sphtfunc.smoothing(hpxmap, fwhm=np.radians(5))\n#hp_smoothed = hp.sphtfunc.smoothing(hpxmap, fwhm=np.radians(5.0*3.14/180))\n#hp_smoothed = hp.smoothing(hpxmap, fwhm=np.radians(5.0*3.14/180))\n#hp_smoothed = hp.sphtfunc.smoothing(hpxmap, sigma=np.radians(5.0*3.14/180))\n#hp_smoothed = hp.smoothing(hpxmap, fwhm=60, arcmin=True)\n#hp.mollview(hp_smoothed, nest=False, xsize = SIZE, cmap=cm.jet, norm = 'hist', title='CMS UCC smoothed')\n#cmap_tmp = cm.jet\n#cmap_tmp.set_bad(\"gray\")\n#cmap_tmp.set_under(\"white\")\nhp.mollview(hp_smoothed, nest=False, xsize = SIZE, cmap=cm.jet, norm=\"hist\", title='CMS UCC smoothed') #jet/rainbow/seismic/bwr; norm=\"hist\"/LogNorm()\nhp.graticule()\n#plt.savefig(\"plot_uccData_3149m_smoothed.png\", dpi = DPI)\n#plt.savefig(\"plot_rapidity08_uccData_allEvents_smoothed.png\", dpi = DPI)\nplt.savefig(\"plot_rapidity08_uccData_allEvents_withnormhist_smoothed.png\", dpi = DPI)\n#plt.savefig(\"plot_uccData_oneEvents_smoothed.png\", dpi = DPI)\n\n\n'''\n#plt.figure(2)\n# Get the power spectrum\nCl = hp.anafast(hpxmap)\n#print(Cl)\nplt.plot(Cl)\nplt.ylabel('C_{l}')\nplt.savefig('plot_uccData_power_spectrum.png')\n'''\n\n","sub_path":"healpy/tutorial/pbpbUCC/rapidity08_uccData_power_spectrum.py","file_name":"rapidity08_uccData_power_spectrum.py","file_ext":"py","file_size_in_byte":2545,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"255409483","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[7]:\n\n\nimport pandas as pd\nimport matplotlib.pyplot as plt \nfrom sklearn import svm\nfrom sklearn import metrics\nimport joblib\nfrom sklearn.decomposition import PCA\nimport numpy as np\nfrom sklearn.model_selection import train_test_split\nimport seaborn as sns\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.metrics import confusion_matrix\n\n\n# In[8]:\n\n\ndataframe = pd.read_csv(\"Downloads\\Mark2.csv\")\n\n\n# In[9]:\n\n\ndataframe.head()\n\n\n# In[10]:\n\n\ndataframe.shape\n\n\n# In[11]:\n\n\ndataframe.describe()\n\n\n# In[12]:\n\n\nsns.pairplot(dataframe[['cough','chest_pain','fatigue','back_pain','acidity','prognosis']],hue='prognosis')\n\n\n# In[13]:\n\n\ntrain=dataframe.drop('prognosis',axis=1)\ntrain.head()\n\n\n# In[14]:\n\n\nprognosis=dataframe.prognosis\nprognosis.head()\n\n\n# In[15]:\n\n\nX_train, X_test, Y_train, Y_test = train_test_split(train,prognosis, test_size=0.2, random_state=0)\nprint(\"X_train size ==>\", X_train.shape)\nprint(\"X_test size ==>\", X_test.shape)\nprint(\"Y_train size ==>\", Y_train.shape)\nprint(\"Y_test size ==>\", Y_test.shape)\n\n\n# In[16]:\n\n\nclf1=svm.SVC(kernel='linear')\nclf1.fit(X_train,Y_train)\nY_pred=clf1.predict(X_test)\n\n\n# In[18]:\n\n\nprint(\"Accuracy:\" ,metrics.accuracy_score(Y_test,Y_pred))\nprint(\"Precision:\",metrics.precision_score(Y_test,Y_pred,average='micro'))\n\n\n# In[19]:\n\n\nfrom sklearn.linear_model import LogisticRegression\n\n\n# In[20]:\n\n\nclf = LogisticRegression()\nclf.fit(X_train, Y_train)\nY_pred=clf.predict(X_test)\n\n\n# In[23]:\n\n\nprint(\"Accuracy:\" ,metrics.accuracy_score(Y_test,Y_pred))\nprint(\"Precision:\",metrics.precision_score(Y_test,Y_pred,average='micro'))\n\n\n# In[24]:\n\n\nfrom sklearn.preprocessing import StandardScaler\nscaler = StandardScaler()\nscaler.fit(X_train)\n\nX_train = scaler.transform(X_train)\nX_test = scaler.transform(X_test)\n\n\n# In[25]:\n\n\nfrom sklearn.neighbors import KNeighborsClassifier\nclassifier = KNeighborsClassifier(n_neighbors=5)\nclassifier.fit(X_train, Y_train)\n\n\n# In[26]:\n\n\nY_pred = classifier.predict(X_test)\n\n\n# In[28]:\n\n\nprint(\"Accuracy:\" ,metrics.accuracy_score(Y_test,Y_pred))\nprint(\"Precision:\",metrics.precision_score(Y_test,Y_pred,average='micro'))\n\n\n# In[29]:\n\n\nimport pickle\n\n\n# In[30]:\n\n\nwith open('mark11_pickle','wb') as f:\n pickle.dump(clf1,f)\n\n\n# In[31]:\n\n\nwith open('mark11_pickle','rb') as f:\n mp = pickle.load(f)\n\n\n# In[32]:\n\n\nmp.predict(X_test)\n\n\n# In[ ]:\n\n\n\n\n","sub_path":"DD_models/Marks11.py.py","file_name":"Marks11.py.py","file_ext":"py","file_size_in_byte":2430,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"87658576","text":"import datetime\nimport sys\nfrom doctest import testmod\n\nfrom upemtk import *\n\nTAILLE_CASE = 75\nTAILLE_PLATEAU = 8\n\nDIRECTIONS = [(0, 1), (1, 1), (1, 0), (1, -1), (0, -1), (-1, -1), (-1, 0), (-1, 1)]\n\nB = 1 # Blancs\nN = 0 # Noirs\nV = \".\" # Vide\nC = \"?\" # Coup possible\n\n\ndef pixel_vers_case(pixel: tuple, taille_case: int):\n \"\"\"\n Convertit les coordonées d'un pixel en coordonées d'une case du plateau.\n :param tuple pixel: Coordonées du pixel.\n :param int taille_case: Taille des cases.\n :return tuple: Coordonées de la case.\n\n >>> pixel_vers_case((100, 200), 50)\n (2, 4)\n \"\"\"\n i, j = pixel\n return i // taille_case, j // taille_case\n\n\ndef case_vers_pixel(case: tuple, taille_case: int):\n \"\"\"\n Donne le pixel au centre d'une case à partir de ses coordonées.\n :param tuple case: Coordonées de la case.\n :param int taille_case: Taille des cases.\n :return tuple: Coordonées du pixel.\n\n >>> case_vers_pixel((0, 0), 50)\n (25.0, 25.0)\n \"\"\"\n i, j = case\n return (i + .5) * taille_case, (j + .5) * taille_case\n\n\ndef init_plateau(taille: int):\n \"\"\"\n Génère une liste de listes correspondant au plateau.\n :param int taille: Longueur et largueur du plateau.\n :return list: Liste de listes d'entiers.\n\n >>> init_plateau(2)\n [['.', '.'], ['.', '.']]\n \"\"\"\n resultat = []\n for i in range(taille):\n resultat.append([V] * taille)\n return resultat\n\n\ndef dessine_plateau(plateau: list, ex_plateau: list, boutons: dict, joueur: int):\n \"\"\"\n Dessine le plateau à partir d'une liste de listes ainsi que la barre inférieure.\n :param list plateau: Liste contenant le plateau.\n :param list ex_plateau: Plateau du tour précédent.\n :param dict boutons: Dictionnaires des boutons du jeu.\n :param int joueur: Joueur actuel.\n \"\"\"\n index = 0\n y = 0\n efface_tout()\n while y < len(plateau):\n x = 0\n while x < len(plateau[y]):\n rectangle(x * TAILLE_CASE, y * TAILLE_CASE, x * TAILLE_CASE + TAILLE_CASE, y * TAILLE_CASE + TAILLE_CASE,\n tag=str(index), remplissage='darkgreen')\n x_pixel, y_pixel = case_vers_pixel((x, y), TAILLE_CASE)\n if plateau[y][x] == B:\n cercle(x_pixel, y_pixel, 9 * TAILLE_CASE / 20, remplissage=\"white\")\n elif plateau[y][x] == N:\n cercle(x_pixel, y_pixel, 9 * TAILLE_CASE / 20, remplissage=\"black\")\n elif plateau[y][x] == C:\n cercle(x_pixel, y_pixel, 9 * TAILLE_CASE / 20, remplissage=\"green\", couleur=\"darkgreen\")\n index += 1\n x += 1\n y += 1\n if joueur == B:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"white\")\n texte(10, 10 + len(plateau) * TAILLE_CASE, \"Au tour des blancs\")\n if joueur == N:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"black\")\n texte(10, 10 + len(plateau) * TAILLE_CASE, \"Au tour des noirs\", couleur=\"white\")\n boutons[\"quitter\"] = dessine_bouton(len(plateau) * TAILLE_CASE - 5,\n len(plateau) * TAILLE_CASE + regles['taille_barre'] - 5,\n \"Quitter\", ancrage='se')\n if ex_plateau is not None:\n boutons[\"annuler\"] = dessine_bouton(\n len(plateau) * TAILLE_CASE - 10 - (boutons[\"quitter\"][2] - boutons[\"quitter\"][0]),\n len(plateau) * TAILLE_CASE + regles['taille_barre'] - 5, \"Annuler\", ancrage='se')\n\n\ndef dessine_bouton(x: int, y: int, contenu: str, ancrage: str = 'nw', couleur: str = 'white', taille: int = 24,\n largeur: str = None):\n \"\"\"\n Dessine un bouton aux coordonnées renseignées.\n :param x: Abscisse du point d'ancrage.\n :param y: Ordonnée du point d'ancrage.\n :param contenu: Texte à insérer dans le bouton.\n :param ancrage: Ancrage du bouton.\n :param couleur: Couleur de remplissage du bouton.\n :param taille: Taille du texte.\n :param largeur: Chaîne de caractère déterminant la largeur du bouton.\n :return tuple: Coordonées du bouton.\n\n >>> cree_fenetre(0, 0)\n >>> dessine_bouton(0, 0, \"test\")\n (0, 0, 49, 35)\n >>> ferme_fenetre()\n \"\"\"\n if largeur is None:\n largeur = contenu\n longueur, hauteur = taille_texte(largeur, taille=taille)\n hauteur -= hauteur % 5\n if ancrage == 'nw':\n rectangle(x, y, x + longueur + 2 * 5, y + hauteur + 2 * 5, remplissage=couleur)\n texte(x + (longueur + 2 * 5) // 2, y + (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x, y, x + longueur + 2 * 5, y + hauteur + 2 * 5\n elif ancrage == 'n':\n rectangle(x - (longueur + 2 * 5) // 2, y, x + (longueur + 2 * 5) // 2, y + hauteur + 2 * 5, remplissage=couleur)\n texte(x, y + (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x - (longueur + 2 * 5) // 2, y, x + (longueur + 2 * 5) // 2, y + hauteur + 2 * 5\n elif ancrage == 'ne':\n rectangle(x - longueur - 2 * 5, y, x, y + hauteur + 2 * 5, remplissage=couleur)\n texte(x + (longueur + 2 * 5) // 2, y - (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x - longueur - 2 * 5, y, x, y + hauteur + 2 * 5\n elif ancrage == 'e':\n rectangle(x - hauteur - 2 * 5, y - (hauteur + 2 * 5) // 2, x, y + (hauteur + 2 * 5) // 2, remplissage=couleur)\n texte(x - (longueur + 2 * 5) // 2, y, contenu, ancrage='center', taille=taille)\n return x - hauteur - 2 * 5, y - (hauteur + 2 * 5) // 2, x, y + (hauteur + 2 * 5) // 2\n elif ancrage == 'se':\n rectangle(x - longueur - 2 * 5, y - hauteur - 2 * 5, x, y, remplissage=couleur)\n texte(x - (longueur + 2 * 5) // 2, y - (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x - longueur - 2 * 5, y - hauteur - 2 * 5, x, y\n elif ancrage == 's':\n rectangle(x - (longueur + 2 * 5) // 2, y - hauteur - 2 * 5, x + (longueur + 2 * 5) // 2, y, remplissage=couleur)\n texte(x, y - (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x - (longueur + 2 * 5) // 2, y - hauteur - 2 * 5, x + (longueur + 2 * 5) // 2, y\n elif ancrage == 'sw':\n rectangle(x, y - hauteur - 2 * 5, x + longueur + 2 * 5, y, remplissage=couleur)\n texte(x + (longueur + 2 * 5) // 2, y - (hauteur + 2 * 5) // 2, contenu, ancrage='center', taille=taille)\n return x, y - hauteur - 2 * 5, x + longueur + 2 * 5, y\n elif ancrage == 'w':\n rectangle(x, y - (hauteur + 2 * 5) // 2, x + hauteur + 2 * 5, y + (hauteur + 2 * 5) // 2, remplissage=couleur)\n texte(x + (longueur + 2 * 5) // 2, y, contenu, ancrage='center', taille=taille)\n return x, y - (hauteur + 2 * 5) // 2, x + hauteur + 2 * 5, y + (hauteur + 2 * 5) // 2\n elif ancrage == 'center':\n rectangle(x - (longueur + 2 * 5) // 2, y - (hauteur + 2 * 5) // 2, x + (longueur + 2 * 5) // 2,\n y + (hauteur + 2 * 5) // 2, remplissage=couleur)\n texte(x, y, contenu, ancrage='center', taille=taille)\n return x - (longueur + 2 * 5) // 2, y - (hauteur + 2 * 5) // 2, x + (longueur + 2 * 5) // 2, \\\n y + (hauteur + 2 * 5) // 2\n\n\ndef chercher_coups(plateau: list, joueur: int):\n \"\"\"\n Parcoure le plateau et recherche les coups disponibles.\n :param list plateau: Liste contenant le plateau.\n :param int joueur: Joueur actuel.\n :return list: Liste des coups possibles\n\n >>> chercher_coups([[B, B, B], \\\n [N, N, N], \\\n [V, V, B]], B)\n [(0, 2), (1, 2)]\n \"\"\"\n resultat = []\n for y in range(len(plateau)):\n for x in range(len(plateau[y])):\n if coup_valide(plateau, (x, y), joueur):\n resultat.append((x, y))\n return resultat\n\n\ndef verifier_victoire(plateau: list, joueur: int):\n \"\"\"\n Vérifie si un des deux joueurs remporte la partie.\n :param list plateau: Liste contenant le plateau.\n :param int joueur: Joueur en cours.\n :return tuple: Numéro du joueur gagnant, 0 si le jeu continu et booléen symbolisant une victoire écrasante.\n\n >>> verifier_victoire([[B, N, B], \\\n [N, N, N], \\\n [B, N, B]], B)\n (0, False)\n \"\"\"\n compteur_b = 0\n compteur_n = 0\n for y, ligne in enumerate(plateau):\n for x, case in enumerate(ligne):\n if coup_valide(plateau, (x, y), joueur) or coup_valide(plateau, (x, y), joueur * -1):\n return 0, False\n if case == N:\n compteur_n += 1\n elif case == B:\n compteur_b += 1\n if compteur_b > compteur_n:\n return B, abs(compteur_n - compteur_b) > 10\n elif compteur_n > compteur_b:\n return N, abs(compteur_n - compteur_b) > 10\n else:\n return None, False\n\n\ndef explore(plateau: list, case: tuple, direction: tuple):\n \"\"\"\n Explore une direction à partir d'une case pour\n :param list plateau: Liste contenant le plateau.\n :param tuple case: Coordonées de la case.\n :param tuple direction: Direction de la vérification.\n :return list: Liste des cases.\n\n >>> explore([[V, B, B, V], \\\n [B, N, N, B], \\\n [B, N, N, B], \\\n [V, B, B, N]], (0, 0), (1, 1))\n [0, 0, 0]\n \"\"\"\n resulat = []\n x_case, y_case = case[0] + direction[0], case[1] + direction[1]\n\n explorer = True\n while explorer:\n if 0 <= x_case < len(plateau) and 0 <= y_case < len(plateau) and (\n plateau[y_case][x_case] == B or plateau[y_case][x_case] == N):\n resulat.append(plateau[y_case][x_case])\n x_case += direction[0]\n y_case += direction[1]\n else:\n explorer = False\n\n return resulat\n\n\ndef coup_valide(plateau: list, case: tuple, joueur: int):\n \"\"\"\n Détermine si un coup est autorisé.\n :param list plateau: Liste contenant le plateau.\n :param tuple case: Coordonées de la case.\n :param int joueur: Numéro du joueur.\n :return bool: Le coup est autorisé.\n\n >>> coup_valide([[V, B, B, V], \\\n [B, N, N, B], \\\n [B, N, N, B], \\\n [V, B, B, V]], (0, 0), N)\n False\n \"\"\"\n if plateau[case[1]][case[0]] != V and plateau[case[1]][case[0]] != C:\n return False\n for direction in DIRECTIONS:\n ligne = explore(plateau, case, direction)\n if len(ligne) == 0:\n continue\n if ligne[0] == abs(joueur - 1) and joueur in ligne:\n return True\n return False\n\n\ndef retournement(plateau: list, case: tuple, pions: list, joueur: int, direction: tuple):\n \"\"\"\n Retourne les pions nécessaires.\n :param list plateau: Liste contenant le plateau.\n :param tuple case: Coordonée de la case.\n :param tuple pions: Liste de coordonées des pions à retourner.\n :param int joueur: Numéro du joueur.\n :param tuple direction: Direction de la ligne.\n \"\"\"\n x, y = case\n for pion in pions:\n x += direction[0]\n y += direction[1]\n if joueur not in pions:\n return\n if pion == joueur or pion == V:\n break\n elif pion == abs(joueur - 1):\n plateau[y][x] = joueur\n\n\ndef sauvegarder(plateau: list, joueur: int, nom_fichier: str):\n \"\"\"\n Sauvegarde la partie actuelle dans un fichier.\n :param list plateau: Liste contenant le plateau.\n :param int joueur: Joueur actuel.\n :param str nom_fichier: Nom de la sauvegarde à écraser.\n \"\"\"\n if nom_fichier is None:\n nom_fichier = datetime.datetime.now().strftime(\"%d-%m-%Y %H-%M\") + \".rsi\"\n fichier = open(nom_fichier, \"w\")\n\n fichier.write(str(joueur))\n for ligne in plateau:\n fichier.write(\"\\n\")\n for case in ligne:\n fichier.write(str(case))\n fichier.close()\n\n\ndef affiche_victoire(gagnant: int, ecrasante: bool):\n \"\"\"\n Affiche le message de fin de partie.\n :param int gagnant: Numéro du joueur gagnant.\n :param bool ecrasante: Victoire avec plus de 10 pions d'avance.\n \"\"\"\n if gagnant is None:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"white\")\n texte(10, 10 + len(plateau) * TAILLE_CASE, \"Égalité !\")\n elif gagnant == N:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"black\")\n texte(10, 10 + len(plateau) * TAILLE_CASE,\n \"Victoire \" + (\"écrasante \" if ecrasante else \"\") + \"des noirs !\",\n couleur=\"white\")\n elif gagnant == B:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"white\")\n texte(10, 10 + len(plateau) * TAILLE_CASE,\n \"Victoire \" + (\"écrasante \" if ecrasante else \"\") + \"des blancs !\")\n\n\ndef dessine_menu(regles: dict, boutons: dict):\n \"\"\"\n Dessine le menu principal et modifie les variables de jeu.\n :param dict regles: Variables de jeu.\n :param dict boutons: Dictionnaire ou ajouter les coordonnées des boutons.\n \"\"\"\n efface_tout()\n image(TAILLE_PLATEAU * TAILLE_CASE // 2, TAILLE_PLATEAU * TAILLE_CASE // 2, \"background.gif\", ancrage='center')\n texte(TAILLE_PLATEAU * TAILLE_CASE // 2, TAILLE_CASE, \"REVERSI\", taille=50, couleur=\"black\", ancrage='n')\n\n boutons[\"nouvelle_partie\"] = \\\n dessine_bouton(TAILLE_PLATEAU * TAILLE_CASE // 2, 3 * TAILLE_CASE, \"Nouvelle partie\", ancrage='center',\n largeur=\"Afficher les coups disponibles\")\n boutons[\"charger_partie\"] = \\\n dessine_bouton(TAILLE_PLATEAU * TAILLE_CASE // 2, 4 * TAILLE_CASE, \"Charger une partie\", ancrage='center',\n largeur=\"Afficher les coups disponibles\")\n boutons[\"afficher_coups\"] = \\\n dessine_bouton(TAILLE_PLATEAU * TAILLE_CASE // 2, 5 * TAILLE_CASE, \"Afficher les coups disponibles\",\n ancrage='center', largeur=\"Afficher les coups disponibles\",\n couleur=\"green\" if regles[\"afficher_coups\"] else \"red\")\n boutons[\"quitter\"] = dessine_bouton(TAILLE_PLATEAU * TAILLE_CASE // 2, 7 * TAILLE_CASE, \"Quitter\", ancrage='center')\n\n\ndef clone(plateau: list):\n \"\"\"\n Retourne un clone du plateau.\n :param plateau: Plateau à cloner.\n :return: Plateau cloné.\n\n >>> clone([[1, 1, 1], \\\n [1, 1, 1], \\\n [1, 1, 1]])\n [[1, 1, 1], [1, 1, 1], [1, 1, 1]]\n \"\"\"\n clone_plateau = []\n for list in plateau:\n clone_plateau.append(list.copy())\n return clone_plateau\n\n\nregles = {'joueur': N,\n 'nouvelle_partie': True,\n 'afficher_coups': False,\n 'sauvegarde': None}\nex_plateau = None\nboutons = {}\n\nif __name__ == \"__main__\":\n\n cree_fenetre(0, 0)\n regles['taille_barre'] = taille_texte('X')[1] - taille_texte('X')[1] % 5 + 20\n ferme_fenetre()\n\n testmod()\n cree_fenetre(TAILLE_PLATEAU * TAILLE_CASE, TAILLE_PLATEAU * TAILLE_CASE + regles['taille_barre'])\n\n menu = True\n while menu:\n dessine_menu(regles, boutons)\n x, y = attend_clic_gauche()\n\n for bouton in boutons:\n if int(boutons[bouton][0]) <= x <= int(boutons[bouton][2]) and \\\n int(boutons[bouton][1]) <= y <= int(boutons[bouton][3]):\n if bouton == \"nouvelle_partie\":\n menu = False\n elif bouton == \"charger_partie\":\n regles['nouvelle_partie'] = False\n menu = False\n elif bouton == \"afficher_coups\":\n regles['afficher_coups'] = not regles['afficher_coups']\n break\n elif bouton == \"quitter\":\n sys.exit()\n\n if regles['nouvelle_partie']:\n\n regles['joueur'] = N\n plateau = init_plateau(TAILLE_PLATEAU)\n\n plateau[TAILLE_PLATEAU // 2 - 1][TAILLE_PLATEAU // 2 - 1] = B\n plateau[TAILLE_PLATEAU // 2][TAILLE_PLATEAU // 2] = B\n plateau[TAILLE_PLATEAU // 2 - 1][TAILLE_PLATEAU // 2] = N\n plateau[TAILLE_PLATEAU // 2][TAILLE_PLATEAU // 2 - 1] = N\n\n else:\n\n nom_fichier = input(\"Entrer le nom du fichier de sauvegarde : \")\n\n fichier = open(nom_fichier, \"r\")\n lignes = fichier.readlines()\n regles['joueur'] = int(lignes[0])\n\n plateau = []\n for i in range(TAILLE_PLATEAU):\n ligne = []\n for case in lignes[i + 1].strip():\n if case == str(B) or case == str(N):\n ligne.append(int(case))\n else:\n ligne.append(case)\n plateau.append(ligne)\n\n regles['afficher_coups'] = False\n for ligne in plateau:\n if C in ligne:\n regles['afficher_coups'] = True\n break\n\n fichier.close()\n regles['sauvegarde'] = nom_fichier\n\n gagnant = None\n ecrasante = False\n\n # BOUCLE PRINCIPALE\n while True:\n\n coups_dispo = chercher_coups(plateau, regles['joueur'])\n\n if regles['afficher_coups']:\n for case in coups_dispo:\n plateau[case[1]][case[0]] = C\n\n dessine_plateau(plateau, ex_plateau, boutons, regles['joueur'])\n\n if verifier_victoire(plateau, regles['joueur'])[0] != 0:\n gagnant, ecrasante = verifier_victoire(plateau, regles['joueur'])\n break\n\n if len(coups_dispo) == 0:\n rectangle(0, len(plateau) * TAILLE_CASE, len(plateau) * TAILLE_CASE,\n len(plateau) * TAILLE_CASE + regles['taille_barre'], remplissage=\"white\")\n texte(10, 10 + len(plateau) * TAILLE_CASE, \"Les \" + (\n \"blancs\" if regles['joueur'] == B else \"noirs\") + \" ne peuvent pas jouer, ils passent leur tour.\")\n mise_a_jour()\n attente(3)\n regles['joueur'] = abs(regles['joueur'] - 1)\n continue\n\n x, y = attend_clic_gauche()\n\n if boutons[\"quitter\"][0] <= x <= boutons[\"quitter\"][2] and boutons[\"quitter\"][1] <= y <= boutons[\"quitter\"][3]:\n sauvegarder(plateau, regles['joueur'], regles['sauvegarde'])\n sys.exit()\n if \"annuler\" in boutons:\n if boutons[\"annuler\"][0] <= x <= boutons[\"annuler\"][2] and \\\n boutons[\"annuler\"][1] <= y <= boutons[\"annuler\"][3]:\n regles['joueur'] = abs(regles['joueur'] - 1)\n plateau = clone(ex_plateau)\n ex_plateau = None\n continue\n\n for case in coups_dispo:\n plateau[case[1]][case[0]] = V\n\n x_case, y_case = pixel_vers_case((x, y), TAILLE_CASE)\n\n if pixel_vers_case((x, y), TAILLE_CASE) not in coups_dispo:\n continue\n\n ex_plateau = clone(plateau)\n plateau[y_case][x_case] = regles['joueur']\n\n for direction in DIRECTIONS:\n ligne = explore(plateau, (x_case, y_case), direction)\n retournement(plateau, (x_case, y_case), ligne, regles['joueur'], direction)\n\n regles['joueur'] = abs(regles['joueur'] - 1)\n\n affiche_victoire(gagnant, ecrasante)\n attend_fermeture()\n","sub_path":"Projet final/reversi.py","file_name":"reversi.py","file_ext":"py","file_size_in_byte":19550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"125935721","text":"'''\n使用urllib.request请求一个网页内容,并把内容打印出来\n'''\nfrom urllib import request\n\nif __name__ == '__main__':\n\n url = 'https://jobs.zhaopin.com/CZ507343230J00057937812.htm'\n # 打开相应的url并把相应页面作为返回\n rsp = request.urlopen(url)\n\n # 把返回结果读取出来\n # 读取出来内容类型为bytes\n html = rsp.read()\n\n #如果想把bytes内容转换成字符串,需要解码\n html = html.decode('utf-8')\n print(html)","sub_path":"Spider/v1.py","file_name":"v1.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"238723556","text":"# -*- coding: utf-8 -*-\n\nfrom photo_downloader2 import PhotoDownloader\nfrom photo_task_gen import PhotoTaskGenerator\nfrom photo_reader import PhotoReader\nimport threading\n\nclass Downloader(threading.Thread):\n _download_size = 10\n\n def __init__(self):\n super(Downloader, self).__init__()\n self.photo_taskGenerator = PhotoTaskGenerator()\n self.photo_reader = PhotoReader(self.photo_taskGenerator)\n\n def run(self):\n self.photo_reader.start()\n for i in range(self._download_size):\n PhotoDownloader(self.photo_taskGenerator).run()\n\nif __name__ == '__main__':\n downloader = Downloader()\n downloader.start()","sub_path":"downloader/downloader.py","file_name":"downloader.py","file_ext":"py","file_size_in_byte":661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"377013155","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n# Copyright: (c) 2019, Hetzner Cloud GmbH \n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import absolute_import, division, print_function\n\n__metaclass__ = type\n\nANSIBLE_METADATA = {\n \"metadata_version\": \"1.1\",\n \"status\": [\"preview\"],\n \"supported_by\": \"community\",\n}\n\nDOCUMENTATION = \"\"\"\n---\nmodule: hcloud_rdns\n\nshort_description: Create and manage reverse DNS entries on the Hetzner Cloud.\n\nversion_added: \"2.9\"\n\ndescription:\n - Create, update and delete reverse DNS entries on the Hetzner Cloud.\n\nauthor:\n - Lukas Kaemmerling (@lkaemmerling)\n\noptions:\n server:\n description:\n - The name of the Hetzner Cloud server you want to add the reverse DNS entry to.\n type: str\n required: true\n ip_address:\n description:\n - The IP address that should point to I(dns_ptr).\n type: str\n required: true\n dns_ptr:\n description:\n - The DNS address the I(ip_address) should resolve to.\n - Omit the param to reset the reverse DNS entry to the default value.\n type: str\n state:\n description:\n - State of the reverse DNS entry.\n default: present\n choices: [ absent, present ]\n type: str\n\nrequirements:\n - hcloud-python >= 1.3.0\n\nextends_documentation_fragment: hcloud\n\"\"\"\n\nEXAMPLES = \"\"\"\n- name: Create a reverse DNS entry for a server\n hcloud_rdns:\n server: my-server\n ip_address: 123.123.123.123\n dns_ptr: example.com\n state: present\n\n- name: Ensure the reverse DNS entry is absent (remove if needed)\n hcloud_rdns:\n server: my-server\n ip_address: 123.123.123.123\n dns_ptr: example.com\n state: absent\n\"\"\"\n\nRETURN = \"\"\"\nhcloud_rdns:\n description: The reverse DNS entry\n returned: always\n type: complex\n contains:\n server:\n description: Name of the server\n type: str\n returned: always\n sample: my-server\n ip_address:\n description: The IP address that point to the DNS ptr\n type: str\n returned: always\n sample: 123.123.123.123\n dns_ptr:\n description: The DNS that resolves to the IP\n type: str\n returned: always\n sample: example.com\n\"\"\"\n\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ansible.module_utils._text import to_native\nfrom ansible.module_utils.hcloud import Hcloud\nfrom ansible.module_utils.network.common import utils\n\ntry:\n from hcloud import APIException\nexcept ImportError:\n APIException = None\n\n\nclass AnsibleHcloudReverseDNS(Hcloud):\n def __init__(self, module):\n super(AnsibleHcloudReverseDNS, self).__init__(module, \"hcloud_rdns\")\n self.hcloud_server = None\n self.hcloud_rdns = None\n\n def _prepare_result(self):\n return {\n \"server\": to_native(self.hcloud_server.name),\n \"ip_address\": to_native(self.hcloud_rdns[\"ip_address\"]),\n \"dns_ptr\": to_native(self.hcloud_rdns[\"dns_ptr\"]),\n }\n\n def _get_server(self):\n try:\n self.hcloud_server = self.client.servers.get_by_name(\n self.module.params.get(\"server\")\n )\n except APIException as e:\n self.module.fail_json(msg=e.message)\n\n def _get_rdns(self):\n ip_address = self.module.params.get(\"ip_address\")\n if utils.validate_ip_address(ip_address):\n if self.hcloud_server.public_net.ipv4.ip == ip_address:\n self.hcloud_rdns = {\n \"ip_address\": self.hcloud_server.public_net.ipv4.ip,\n \"dns_ptr\": self.hcloud_server.public_net.ipv4.dns_ptr,\n }\n else:\n self.module.fail_json(msg=\"The selected server does not have this IP address\")\n elif utils.validate_ip_v6_address(ip_address):\n for ipv6_address_dns_ptr in self.hcloud_server.public_net.ipv6.dns_ptr:\n if ipv6_address_dns_ptr[\"ip\"] == ip_address:\n self.hcloud_rdns = {\n \"ip_address\": ipv6_address_dns_ptr[\"ip\"],\n \"dns_ptr\": ipv6_address_dns_ptr[\"dns_ptr\"],\n }\n else:\n self.module.fail_json(msg=\"The given IP address is not valid\")\n\n def _create_rdns(self):\n self.module.fail_on_missing_params(\n required_params=[\"dns_ptr\"]\n )\n params = {\n \"ip\": self.module.params.get(\"ip_address\"),\n \"dns_ptr\": self.module.params.get(\"dns_ptr\"),\n }\n\n if not self.module.check_mode:\n self.hcloud_server.change_dns_ptr(**params).wait_until_finished()\n\n self._mark_as_changed()\n self._get_server()\n self._get_rdns()\n\n def _update_rdns(self):\n dns_ptr = self.module.params.get(\"dns_ptr\")\n if dns_ptr != self.hcloud_rdns[\"dns_ptr\"]:\n params = {\n \"ip\": self.module.params.get(\"ip_address\"),\n \"dns_ptr\": dns_ptr,\n }\n\n if not self.module.check_mode:\n self.hcloud_server.change_dns_ptr(**params).wait_until_finished()\n\n self._mark_as_changed()\n self._get_server()\n self._get_rdns()\n\n def present_rdns(self):\n self._get_server()\n self._get_rdns()\n if self.hcloud_rdns is None:\n self._create_rdns()\n else:\n self._update_rdns()\n\n def delete_rdns(self):\n self._get_server()\n self._get_rdns()\n if self.hcloud_rdns is not None:\n if not self.module.check_mode:\n self.hcloud_server.change_dns_ptr(ip=self.hcloud_rdns['ip_address'], dns_ptr=None)\n self._mark_as_changed()\n self.hcloud_rdns = None\n\n @staticmethod\n def define_module():\n return AnsibleModule(\n argument_spec=dict(\n server={\"type\": \"str\", \"required\": True},\n ip_address={\"type\": \"str\", \"required\": True},\n dns_ptr={\"type\": \"str\"},\n state={\n \"choices\": [\"absent\", \"present\"],\n \"default\": \"present\",\n },\n **Hcloud.base_module_arguments()\n ),\n supports_check_mode=True,\n )\n\n\ndef main():\n module = AnsibleHcloudReverseDNS.define_module()\n\n hcloud = AnsibleHcloudReverseDNS(module)\n state = module.params[\"state\"]\n if state == \"absent\":\n hcloud.delete_rdns()\n elif state == \"present\":\n hcloud.present_rdns()\n\n module.exit_json(**hcloud.get_result())\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"env/lib/python3.9/site-packages/ansible/modules/cloud/hcloud/hcloud_rdns.py","file_name":"hcloud_rdns.py","file_ext":"py","file_size_in_byte":6766,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"12899423","text":"import numpy as np\nimport random\n\nclass SimpleMahjongBuffer():\n\n def __init__(self, size=10000):\n self.size = size\n self.s = np.zeros([size, 34, 4, 1], dtype=np.float32)\n self.sp = np.zeros([size, 34, 4, 1], dtype=np.float32)\n self.r = np.zeros([size], dtype=np.float32)\n self.d = np.zeros([size], dtype=np.float32)\n\n self.filled = -1 # up to this index, the buffer is filled\n self.tail = 0 # index to which an experience to be stored next\n\n def __len__(self):\n return self.filled\n\n def append(self, s, r, s_next, done):\n self.s[self.tail] = s\n self.r[self.tail] = r\n self.sp[self.tail] = s_next\n # self.mu[self.tail] = mu\n self.d[self.tail] = done\n\n self.filled = max(self.filled, self.tail)\n self.tail += 1\n if self.tail == self.size:\n self.tail = 0\n\n def sample(self, truncated_steps):\n index = np.random.randint(low=truncated_steps, high=self.filled)\n\n s = np.vstack((self.s[index - truncated_steps:index], self.sp[None, index]))\n ret_tuple = (s,\n self.r[index - truncated_steps:index],\n self.d[index - truncated_steps:index])\n\n return ret_tuple\n\n\n\n\nclass SimpleMahjongBufferPER():\n # Record Episodes\n # Prioritized Experience Replay\n class SumTree:\n def __init__(self, size, scale):\n self.size = size\n self.scale = scale\n self.tree_nodes = np.zeros(2 * size, dtype=np.float32)\n\n def sum(self):\n return self.tree_nodes[1]\n\n def add(self, index, weight):\n index += self.size\n self.tree_nodes[index] = weight ** self.scale\n while index > 0:\n index = index // 2\n self.tree_nodes[index] = self.tree_nodes[index * 2] + self.tree_nodes[index * 2 + 1]\n\n def sample_subtree(self, root):\n if root >= self.size:\n return root - self.size, self.tree_nodes[root] / self.tree_nodes[1]\n pl = self.tree_nodes[root * 2]\n pr = self.tree_nodes[root * 2 + 1]\n s = pl + pr\n pl /= s\n pr /= s\n if random.random() < pl:\n return self.sample_subtree(root * 2)\n else:\n return self.sample_subtree(root * 2 + 1)\n\n def sample(self):\n if self.tree_nodes[1] == 0:\n raise Exception(\"Error! Sampling from empty buffer\")\n return self.sample_subtree(1)\n pass\n\n def __init__(self, size=1024, episode_length=256, priority_eps=10000,\n priority_scale=0.1, IS_scale=1.0, saved=None):\n # Mahjong episode length usually <256\n self.size = size\n self.length = np.zeros([size,], dtype=np.int)\n self.s = np.zeros([size, episode_length, 34, 4, 1], dtype=np.float16)\n self.r = np.zeros([size, episode_length], dtype=np.float32)\n self.d = np.zeros([size, episode_length], dtype=np.float16)\n\n self.IS_scale = IS_scale\n self.priority_scale = priority_scale\n self.priority_eps = priority_eps\n\n self.filled_size = 0 # up to this index, the buffer is filled\n self.tail = 0 # index to which an experience to be stored next\n\n tree_size = 2 ** int(np.ceil(np.log2(size)))\n self.sum_tree = SimpleMahjongBufferPER.SumTree(tree_size, priority_scale)\n self.sum_steps = 0\n self.min_length = 0\n self.max_length = 0\n\n # self.infinity = 1e9\n\n # # initialize priority of saved episodes to inf\n # for i in range(self.saved_size):\n # self.sum_tree.add(i, self.infinity)\n # length = self.length[i]\n # self.sum_steps += length\n # self.min_length = min(self.min_length, length)\n # self.max_length = max(self.max_length, length)\n\n def append_episode(self, s, r, d, weight=0):\n length = len(r)\n self.s[self.tail, :length+1] = s\n self.r[self.tail, :length] = r\n self.d[self.tail, :length] = d\n self.length[self.tail] = length\n\n self.sum_tree.add(self.tail, weight + self.priority_eps)\n\n self.tail = (self.tail + 1) % self.size\n self.filled_size = min(self.filled_size + 1, self.size)\n\n\n def sample_episode(self):\n e_index, e_weight = self.sum_tree.sample()\n\n d = self.d[e_index, :self.length[e_index]]\n s = self.s[e_index, :self.length[e_index]+1]\n r = self.r[e_index, :self.length[e_index]]\n # pi = self.pi[e_index]\n\n return s, r, d, self.length[e_index], e_index, e_weight\n\n # def sample_episode_batch(self, batch_size=32):\n # states_b = []\n # actions_b = []\n # pi_b = []\n # r_b = []\n # done_b = []\n # length_b = []\n # mask_b = []\n # indices = []\n # weights = []\n # for i in range(batch_size):\n # states, actions, pi, r, done, length, mask, e_index, e_weight = self.sample_episode()\n # states_b.append(states)\n # actions_b.append(actions)\n # pi_b.append(pi)\n # r_b.append(r)\n # done_b.append(done)\n # length_b.append(length)\n # mask_b.append(mask)\n # indices.append(e_index)\n # weights.append((length/e_weight/self.sum_steps) ** self.IS_scale)\n #\n # max_l = max(length_b)\n # states_b = np.stack(states_b, axis=0)[:, :max_l+1]\n # actions_b = np.stack(actions_b, axis=0)[:, :max_l]\n # pi_b = np.stack(pi_b, axis=0)[:, :max_l]\n # r_b = np.stack(r_b, axis=0)[:, :max_l]\n # done_b = np.stack(done_b, axis=0)\n # mask_b = np.stack(mask_b, axis=0)\n # weights_b = np.stack(weights, axis=0)\n #\n # return states_b, actions_b, pi_b, r_b, done_b, length_b, mask_b, indices, weights_b","sub_path":"AI/buffer.py","file_name":"buffer.py","file_ext":"py","file_size_in_byte":5903,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"531651914","text":"import os\nimport sys\nimport time\nimport argparse\nimport numpy as np\nimport cv2\nimport torch\nfrom torch.autograd import Variable\n\nsys.path.append(os.getcwd())\nfrom .utils import utils\nfrom .Net import crnn_vgg\nfrom .config import config\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--image_path', type=str,\n default='/home/lyb/ocr/text_det_reg/dataset/images_sentences1/images/000000_00_00.jpg',\n help='the path to your image')\nopt = parser.parse_args()\n\n\ndef get_alphabets(alphabet_path):\n # 获取字符表\n alphabets = utils.generate_alphabets(alphabet_path=alphabet_path)\n nclass = len(alphabets) + 1\n return nclass, alphabets\n\n\ndef process_image(image):\n image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n dst = cv2.fastNlMeansDenoising(image, None, h=10, templateWindowSize=7, searchWindowSize=21)\n ret, image = cv2.threshold(dst, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n h, w = image.shape\n image = cv2.resize(image, (0, 0), fx=config.imgH / h, fy=config.imgH / h, interpolation=cv2.INTER_CUBIC)\n # 不足的,补充白色区域\n image = padding_image(image)\n\n image = (np.reshape(image, (32, config.imgW, 1))).transpose(2, 0, 1)\n # 预处理,转换为torchTensor\n image = preprocess(image)\n return image\n\ndef padding_image(image_):\n h, w = image_.shape\n img = 255. * np.ones((config.imgH, config.imgW))\n if w < config.imgW:\n img[:, :w] = image_\n else:\n img = cv2.resize(image_, (config.imgW, config.imgH), interpolation=cv2.INTER_CUBIC)\n img = np.uint8(img)\n return img\n\ndef preprocess(image_):\n image = image_.astype(np.float32) / 255.\n image = torch.from_numpy(image).type(torch.FloatTensor)\n image.sub_(config.mean).div_(config.std)\n return image\n\n\ndef crnn_recognition(image, model, device, alphabets):\n converter = utils.strLabelConverter(alphabets)\n image = Variable(image)\n image = image.to(device)\n\n model.eval()\n preds = model(image) # (141, batch_size, 6773)\n\n _, preds = preds.max(2) # (141, batch_szie)\n preds = preds.transpose(1, 0).contiguous().view(-1)\n\n preds_size = torch.IntTensor([preds.size(0)])\n sim_pred = converter.decode(preds.data, preds_size.data, raw=False)\n print('results: {0}'.format(sim_pred))\n return 0\n\n\ndef infer_recognition(image_batch):\n return 0\n\n\nif __name__ == '__main__':\n nclass, alphabets = get_alphabets(alphabet_path='./dataset/alphabets.txt')\n\n model = crnn_vgg.CRNN(32, 1, nclass, 256)\n device = torch.device('cpu')\n # device = torch.device('cuda:2') if torch.cuda.is_available() else torch.device('cpu')\n model.to(device)\n # crnn_model_path = './models/crnn_best.pth'\n crnn_model_path = './models/crnn_Rec_ocr_50_0.00478125.pth'\n print('===> loading pretrained model from {}'.format(crnn_model_path))\n model.load_state_dict(torch.load(crnn_model_path, map_location='cpu'))\n # model.load_state_dict(torch.load(crnn_model_path))\n\n start_time = time.time()\n # ------在此输入图片路径-------\n image = cv2.imread(opt.image_path)\n image = process_image(image)\n image = image.view(1, *image.size())\n\n crnn_recognition(image, model, device, alphabets)\n\n finish_time = time.time()\n print('elapsed time of recognition: {0}'.format(finish_time - start_time))\n\n\n\n","sub_path":"CRNN/infer.py","file_name":"infer.py","file_ext":"py","file_size_in_byte":3349,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"64734394","text":"import socket\nimport threading\nimport sys\n\nfrom server import Server\nfrom client import Client\n\nif __name__ == \"__main__\":\n\n if len(sys.argv) > 1:\n if len(sys.argv) == 3:\n try:\n client = Client(sys.argv[1], int(sys.argv[2]))\n except:\n print(\"Error! Type a number.\")\n else:\n client = Client(sys.argv[1])\n else:\n server = Server()\n server.run()\n","sub_path":"_python_chat/chat.py","file_name":"chat.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"425714534","text":"# -*- coding: utf-8 -*-\n\"\"\"\n获取英雄列表,以及对应的页面url\n\"\"\"\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium import webdriver\nimport pandas as pd\n\ndriver = webdriver.PhantomJS(executable_path=r'D:/Program Files/phantomjs-2.1.1-windows/bin/phantomjs')\n\ndriver.get(\"http://pvp.qq.com/web201605/herolist.shtml\")\n\nheros_node=driver.find_elements_by_xpath('/html/body/div[3]/div/div/div[2]/div[2]/ul/li/a')\n\nhero_list=[]\nurl_list=[]\nfor node in heros_node:\n hero_list.append(node.text)\n url_list.append(node.get_attribute('href'))\n \ndata=pd.DataFrame({'hero':hero_list,'url1':url_list})\n\ndriver.get('http://news.17173.com/z/pvp/yxtj/index.shtml')\n\nnodes=driver.find_elements_by_xpath('//*[@id=\"jsheroshow\"]/li/a')\n\nhero_list=[]\nurl_list=[]\nfor node in nodes:\n hero_list.append(node.text)\n url_list.append(node.get_attribute('href'))\n \ntmp=pd.DataFrame({'hero':hero_list,'url2':url_list})\n\ndata=pd.merge(data,tmp,on='hero')\n\ndata.to_csv('D:/Python project/王者荣耀/url.csv')\ndriver.close() \n\n##4399\ndriver.get(\"http://news.4399.com/gonglue/wzlm/daoju/\")\nactions = ActionChains(driver)\nclick_node=driver.find_element_by_xpath('//*[@id=\"hero_more\"]/a')\nactions.click(click_node)\nactions.click(click_node)\nactions.perform()\nheros_node=driver.find_elements_by_xpath('//*[@id=\"hreo_list\"]/li/a')\n\nitem_list=[]\nurl_list=[]\nfor node in heros_node:\n item_list.append(node.text)\n url_list.append(node.get_attribute('href'))\n \ndata=pd.DataFrame({'item':item_list,'url1':url_list}) \ndata.to_csv('D:/Python project/王者荣耀/item_url.csv') ","sub_path":"study/王者荣耀/爬虫代码/get_url.py","file_name":"get_url.py","file_ext":"py","file_size_in_byte":1605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"80863873","text":"import time\nimport requests\nfrom typing import List, Dict\nfrom dataclasses import dataclass\nfrom arize import public_pb2 as pb\nfrom arize.utils.types import ModelTypes, Environments\n\n\n@dataclass(frozen=True)\nclass Schema:\n prediction_id_column_name: str\n feature_column_names: List[str]\n timestamp_column_name: str = None\n prediction_label_column_name: str = None\n prediction_score_column_name: str = None\n actual_label_column_name: str = None\n shap_values_column_names: Dict[str, str] = None\n\n\nclass Client:\n def __init__(\n self,\n api_key: str,\n organization_key: str,\n uri=\"https://api.arize.com/v1\"):\n self._api_key = api_key\n self._organization_key = organization_key\n self._files_uri = uri + \"/files\"\n\n def log(\n self,\n dataframe,\n path: str,\n model_id: str,\n model_version: str,\n batch_id: str,\n model_type: ModelTypes,\n environment: Environments,\n schema: Schema\n ):\n col_idx = {col: idx for idx, col in enumerate(dataframe.columns)}\n h = pb.FileHeader()\n\n if model_type is None:\n raise AttributeError(\"model_type is required\")\n\n if environment is None:\n raise AttributeError(\"environment is required\")\n\n if environment is Environments.TRAINING:\n if schema.prediction_label_column_name is None or schema.actual_label_column_name is None:\n raise AttributeError(\"both prediction and actual label must be specified for Training environment\")\n if schema.shap_values_column_names is not None:\n raise AttributeError(\"shap_values are not supported for Training environments\")\n h.environment = pb.FileHeader.Environment.TRAINING\n elif environment is Environments.VALIDATION:\n if schema.prediction_label_column_name is None or schema.actual_label_column_name is None:\n raise AttributeError(\"both prediction and actual label must be specified for Validation environment\")\n if schema.shap_values_column_names is not None:\n raise AttributeError(\"shap_values are not supported for Validation environments\")\n if batch_id is None:\n raise AttributeError(\"batch_id is required for Validation environment\")\n h.environment = pb.FileHeader.Environment.VALIDATION\n elif environment is Environments.PRODUCTION:\n h.environment = pb.FileHeader.Environment.PRODUCTION\n else:\n raise AttributeError(f\"unknown environment {environment}\")\n\n with open(path, \"wb\") as f:\n header = h.SerializeToString()\n f.write(len(header).to_bytes(8, \"big\", signed=False))\n f.write(header)\n\n current_time = int(time.time())\n for row in dataframe.to_numpy():\n if Environments.TRAINING == environment:\n msg = pb.PreProductionRecord()\n r = msg.training_record.record\n elif Environments.VALIDATION == environment:\n msg = pb.PreProductionRecord()\n msg.validation_record.batch_id = batch_id\n r = msg.validation_record.record\n else:\n msg = pb.Record()\n r = msg\n\n r.prediction_id = row[col_idx[schema.prediction_id_column_name]]\n r.model_id = model_id\n\n t = (\n int(row[col_idx[schema.timestamp_column_name]])\n if schema.timestamp_column_name is not None\n else current_time\n )\n\n for feature_cn in schema.feature_column_names:\n row_val = row[col_idx[feature_cn]]\n feature_val = r.prediction.features[feature_cn]\n if isinstance(row_val, (str, bool)):\n feature_val.string = row_val\n elif isinstance(row_val, int):\n feature_val.int = row_val\n elif isinstance(row_val, float):\n feature_val.double = row_val\n\n if schema.prediction_label_column_name is not None:\n r.prediction.timestamp.seconds = t\n r.prediction.model_version = model_version\n if model_type is ModelTypes.SCORE_CATEGORICAL:\n r.prediction.label.score_categorical.categorical = row[\n col_idx[schema.prediction_label_column_name]\n ]\n r.prediction.label.score_categorical.score = row[\n col_idx[schema.prediction_score_column_name]\n ]\n elif model_type is ModelTypes.CATEGORICAL:\n r.prediction.label.categorical = row[\n col_idx[schema.prediction_label_column_name]\n ]\n elif model_type is model_type.NUMERIC:\n r.prediction.label.numeric = row[\n col_idx[schema.prediction_label_column_name]\n ]\n elif model_type is model_type.BINARY:\n r.prediction.label.binary = row[\n col_idx[schema.prediction_label_column_name]\n ]\n\n if schema.actual_label_column_name is not None:\n if model_type is ModelTypes.CATEGORICAL:\n r.actual.label.categorical = row[col_idx[schema.actual_label_column_name]]\n elif model_type is ModelTypes.NUMERIC:\n r.actual.label.numeric = row[col_idx[schema.actual_label_column_name]]\n elif model_type is ModelTypes.BINARY:\n r.actual.label.binary = row[col_idx[schema.actual_label_column_name]]\n elif model_type is ModelTypes.SCORE_CATEGORICAL:\n if isinstance(schema.actual_label_column_name, tuple):\n r.actual.label.score_categorical.categorical = row[col_idx[schema.actual_label_column_name[0]]]\n r.actual.label.score_categorical.score = row[col_idx[schema.actual_label_column_name[1]]]\n else:\n r.actual.label.score_categorical.categorical = row[col_idx[schema.actual_label_column_name]]\n\n if schema.shap_values_column_names is not None:\n for feature_name, shap_values_cn in schema.shap_values_column_names.items():\n row_val = row[col_idx[shap_values_cn]]\n r.feature_importances.feature_importances[feature_name] = row_val\n\n msg_bytes = msg.SerializeToString()\n f.write(len(msg_bytes).to_bytes(8, \"big\", signed=False))\n f.write(msg_bytes)\n return self._post_file(path)\n\n def _post_file(self, path):\n with open(path, \"rb\") as f:\n return requests.post(\n self._files_uri,\n data=f,\n headers={\n \"authorization\": self._api_key,\n \"organization\": self._organization_key,\n },\n )\n","sub_path":"arize/pandas/logger.py","file_name":"logger.py","file_ext":"py","file_size_in_byte":7376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"129588078","text":"# -*- coding: utf-8\nimport requests\nfrom bs4 import BeautifulSoup\n\n\nreq = requests.get(\"http://www.keumyang.com/mall/KYDetail.ky?ps_goid=7942\")\nhtml = req.text\n# print(html)\nsoup = BeautifulSoup(html, \"html.parser\")\ninfo_ul = soup.select(\"div.detail_wine > div.detail_top > dl > dd > ul > li:nth-child(9) > span.txt\")\n\nprint(info_ul)\n\n","sub_path":"crawling.py","file_name":"crawling.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"129604422","text":"from swampy.TurtleWorld import *\nimport math\n\ndef square( turtle, length ):\n for i in range(4):\n fd( turtle, length )\n lt( turtle ) \n\ndef polygon( turtle, n, length ):\n for i in range(n):\n fd( turtle, length )\n lt( turtle, 360.0 / n ) \n\ndef circle( turtle, r ):\n n = 360\n alpha = 360.0 / n\n alpha = alpha * math.pi / 180.0\n length = math.sqrt( 2 * r * r * ( 1.0 - math.cos( alpha ) ) )\n polygon( turtle, n, length )\n\ndef arc( turtle, r, angle ):\n n = 360\n alpha = 360.0 / n\n alpha = alpha * math.pi / 180.0\n length = math.sqrt( 2 * r * r * ( 1.0 - math.cos( alpha ) ) )\n for i in range( angle ):\n fd( turtle, length )\n lt( turtle, 360.0 / n ) \n\nworld = TurtleWorld()\nbob = Turtle()\nbob.delay = 0.01\n\n#square( bob, 100 )\n\n#polygon( bob, 6, 100 )\n\n#circle( bob, 100 )\n\narc( bob, 100, 135 )\n\nwait_for_user()","sub_path":"swampy/mypolygon.py","file_name":"mypolygon.py","file_ext":"py","file_size_in_byte":882,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"410000879","text":"from talkingdata.feat2 import *\nimport gc\nimport time\nimport datetime\nimport numpy as np\nimport pandas as pd\nimport catboost as cb\nimport lightgbm as lgb\nfrom sklearn.metrics import roc_auc_score,log_loss\n\n\ninplace = False\ndata_path = 'C:/Users/cui/Desktop/python/talkingdata/data/'\n\n\ntrain = pd.read_hdf(data_path + 'train.hdf')\ntest = pd.read_hdf(data_path + 'test.hdf')\nsubmission = pd.DataFrame({'click_id':test.click_id})\ntest['is_attributed'] = 0\ndata = train.append(test.drop('click_id',axis=1))\ndata = pre_treatment(data,'sub_data')\ndel train,test\ngc.collect()\n\ntrain_feat = pd.DataFrame()\nfor date in [7,8,9]:\n train_feat_sub = make_feat(data[data['date']==date].copy(),\n data,'{}_sub'.format(date))\n train_feat_sub = train_feat_sub[train_feat_sub['is_attributed'] == 1].append(\n train_feat_sub[train_feat_sub['is_attributed'] == 0].sample(frac=0.1, random_state=66))\n train_feat = train_feat.append(train_feat_sub)\n del train_feat_sub\n gc.collect()\n\ntest_feat = make_feat(data[data['date']==10].copy(),\n data,'{}_sub'.format(10))\n\n\n\npredictors = [c for c in train_feat.columns if c not in ['app_count','channel_count',\n 'click_time','date','ip','is_attributed']]\n# predictors = ['app', 'device', 'os', 'channel', 'hour','ip_count','device_count','os_count',\n# 'ip_group_rank1','ip_group_rank2','ip_group_rank2']\n\nprint('开始训练...')\nparams = {\n 'task': 'train',\n 'boosting_type': 'gbdt',\n 'objective': 'binary',\n 'metric': 'auc',\n 'max_depth': 8,\n 'num_leaves': 32,\n 'learning_rate': 0.01,\n 'subsample': 0.7,\n 'colsample_bytree': 0.7,\n 'feature_fraction': 0.9,\n 'bagging_fraction': 0.95,\n 'bagging_freq': 5,\n 'verbose': 0,\n 'seed': 66,\n}\nlgb_train = lgb.Dataset(train_feat[predictors], train_feat.is_attributed)\nlgb_test = lgb.Dataset(test_feat[predictors], test_feat.is_attributed,reference=lgb_train)\n\ngbm = lgb.train(params,lgb_train,3000)\n\nprint('开始预测...')\npreds = gbm.predict(test_feat[predictors])\nsubmission['is_attributed'] = preds\nsubmission.to_csv('C:/Users/cui/Desktop/python/talkingdata/sub{}.csv'.format(\n datetime.datetime.now().strftime('%Y%m%d_%H%M%S')), index=False,float_format='%.4f')\n\n\n","sub_path":"talkingdata/submission_all.py","file_name":"submission_all.py","file_ext":"py","file_size_in_byte":2261,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"433523758","text":"from urllib.request import urlopen\nimport json\n\nwhile True:\n id_game= str(input(\"Enter game id:\"))\n if id_game.lower() == \"exit\":\n break\n elif id_game.isdigit() == False:\n print(\"Enter game id!\")\n else:\n url = f\"https://store.steampowered.com/api/appdetails?appids={id_game}&cc=vn\"\n\n #1. Open connection\n conn = urlopen(url)\n\n #2. Read data\n raw_data = conn.read()\n\n #3. Decode data\n text = raw_data.decode(\"utf-8\")\n data = json.loads(text)\n\n if data[id_game][\"success\"]:\n sub_info = data[id_game][\"data\"][\"package_groups\"][0][\"subs\"]\n for i in sub_info:\n game_name = i[\"option_text\"]\n price = (int(i[\"price_in_cents_with_discount\"])/100) * 0.78\n print(game_name)\n print(\"Gia ban\", price)\n print(\"**********************\")\n else:\n print(\"Sai ma game\")\n","sub_path":"session5/aqi.py","file_name":"aqi.py","file_ext":"py","file_size_in_byte":955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"650923780","text":"#!/usr/bin/env python\n# -- coding: utf-8 --\n\nfrom telethon import TelegramClient, sync\nfrom telethon.tl.functions.photos import UploadProfilePhotoRequest, DeletePhotosRequest\nimport requests\nimport socks\nfrom PIL import ImageDraw, Image, ImageFont\nimport time\nimport os\n\ncelsius = '°'\nTEMP_PATH = 'temperature_images'\nREADY_PATH = 'ready_images'\nWEATHER_PATH = 'weather_images'\n\nlocation = 498817 # Saint-Petersburg\nopenweather_api_key = 'abcdefghi'\n\ntelegram_api_id = 1234567\ntelegram_api_hash = 'abcdefghi'\n\nFONT_SIZE = 80\nTEXT_Y_POSITION = 80\n\n#proxy options\n#w = input('Which: ')\n\nhost = '127.0.0.1' # a valid host\nport = 9050 # a valid port\nproxy = (socks.SOCKS5, host, port)\n\nicons = [\"01d\", \"02d\", \"03d\", \"04d\", \"09d\", \"10d\", \"11d\", \"13d\", \"50d\", \"01n\", \"02n\", \"03n\", \"04n\", \"09n\", \"10n\", \"11n\", \"13n\", \"50n\"]\n\n# create all avatars\nprint('Start of temperature images generation')\nfor temperature in range(-99, 99, 1):\n raw = Image.new('RGBA', (200, 200), \"gray\")\n parsed = ImageDraw.Draw(raw)\n length = len(str(temperature))\n if length == 1:\n x_start = 70\n if length == 2:\n x_start = 40\n if length == 3:\n x_start = 30\n\n font = ImageFont.truetype(\"arial.ttf\", FONT_SIZE)\n parsed.text((x_start, TEXT_Y_POSITION), f'{temperature}{celsius}', align=\"center\", font=font)\n raw.save(f'{TEMP_PATH}/{temperature}.png', \"PNG\")\nprint('Generating finished')\n\nprint('Beginning icons download')\nfor ic in icons:\n url = f'http://openweathermap.org/img/wn/{ic}@2x.png'\n r = requests.get(url)\n with open(f'{WEATHER_PATH}/{ic}.png', 'wb') as f:\n f.write(r.content)\n r.raise_for_status()\nprint('Download successfull')\n\nprint('Generate_final_images')\nfor temperature_file in os.listdir(f'{TEMP_PATH}'):\n if temperature_file.endswith(\".png\"):\n new_im = Image.open(f'{TEMP_PATH}/{temperature_file}')\n for weather_file in os.listdir(f'{WEATHER_PATH}'):\n if weather_file.endswith(\".png\"):\n im = Image.open(f'{WEATHER_PATH}/{weather_file}')\n test_img = Image.composite(im, new_im, im)\n temperature_file = temperature_file.strip('.png')\n weather_file = weather_file.replace('.png','')\n test_img.save(f'{READY_PATH}/{temperature_file}_{weather_file}.png')\nprint('done')\n\ndef get_temperature(weather_data):\n return round(weather_data['main']['temp'])\ndef get_conditions(weather_data):\n weath_array = weather_data['weather']\n weath_dict = weath_array[0]\n return weath_dict['icon']\n\n\ndef get_weather(location, api_key):\n url = f'https://api.openweathermap.org/data/2.5/weather?id={location}&units=metric&appid={api_key}'\n r = requests.get(url)\n return r.json()\n\n\nclient = TelegramClient('1', telegram_api_id, telegram_api_hash)\n#client = TelegramClient('anon', telegram_api_id, telegram_api_hash, proxy=(socks.SOCKS5, '127.0.0.1', 9050))\nclient.connect()\n\n\nclient.start()\n\nlast_temperature = -274\n\nwhile True:\n weather_data = get_weather(location, openweather_api_key)\n temperature = get_temperature(weather_data)\n conditions = get_conditions(weather_data)\n print(last_temperature, temperature)\n if temperature == last_temperature:\n time.sleep(15 * 60)\n continue\n\n client(DeletePhotosRequest(client.get_profile_photos('me')))\n file = client.upload_file(f'{READY_PATH}/{temperature}_{conditions}.png')\n client(UploadProfilePhotoRequest(file))\n last_temperature = temperature\n time.sleep(15 * 60)","sub_path":"change_avatar.py","file_name":"change_avatar.py","file_ext":"py","file_size_in_byte":3510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"112234328","text":"\nfrom mcts_node import MCTSNode\nfrom random import choice\nimport random\nfrom timeit import default_timer as time\nfrom math import sqrt, log\n#how do we improve our algorithm?\n#why does backpropagate not work?\n#The flip(i.e how do we use the function for the opponent)\n#is rollout correct?\nnum_nodes = 10\nexplore_faction = 2.\n\nROLLOUTS = 1000\nMAX_DEPTH = 5\n\ndef traverse_nodes(node, board, state, identity):\n \"\"\" Traverses the tree until the end criterion are met.\n Args:\n node: A tree node from which the search is traversing.\n board: The game setup.\n state: The state of the game.\n identity: The bot's identity, either 'red' or 'blue'.\n Returns: A node from which the next stage of the search can proceed.\n \"\"\"\n\n # For first node, will pick root because it is also leaf node\n # UCB - X is how many times won over how many times played (visited)\n # x = node.wins/node.visits (if visits != 0)\n # C is hardcoded, we can choose. Lower C value is exploitation, higher is exploration.\n # Try C = 2\n # Use highest UCB to choose each node down a path until leaf_node is reached\n turn = identity\n current = node\n # index = 0\n # while len(current.child_nodes) != 0:\n # for action in current.child_nodes:\n # current = current.child_nodes[action]\n # print(\"index: \", index)\n # print(\"action: \", action)\n # index += 1\n ucb = {}\n\n temp = node\n max_ucb = 0\n\n i = 0\n updatestate = state\n # a = len(current.child_nodes)\n while len(current.child_nodes) != 0:\n for action in current.child_nodes:\n # print(i)\n i = i +1\n if current.child_nodes[action].visits == 0 :\n #print(len(current.child_nodes))\n child = current.child_nodes[action]\n exploitation = child.wins/child.visits\n\n exploration = 2 * (sqrt((2 * log(child.parent.visits))/ child.visits))\n\n ucb[child] = exploitation + exploration\n\n if ucb[child] > max_ucb:\n max_ucb = ucb[child]\n temp = child \n \n if turn == 1:\n turn = 2\n elif turn == 2:\n turn = 1\n current = current.child_nodes[action]\n updatestate = board.next_state(updatestate, action)\n\n\n leaf_node = current\n\n return leaf_node\n # Hint: return leaf_node\n\n\n\ndef expand_leaf(node, board, state):\n \"\"\" Adds a new leaf to the tree by creating a new child node for the given node.\n Args:\n node: The node for which a child will be added.\n board: The game setup.\n state: The state of the game.\n Returns: The added child node.\n \"\"\"\n\n # Create new node from parent node\n # Can choose action randomly\n # Action list comes from node.untried_actions?\n parent_node = node \n actions1 = board.legal_actions(state)\n state2 = board.next_state(state, actions1[0])\n actions2 = board.legal_actions(state2)\n #print(actions[0])\n # print(\"this is the length before \",len(parent_node.child_nodes) )\n child_node = MCTSNode(parent=parent_node, parent_action=actions1[0], action_list=board.legal_actions(state2))\n try:\n parent_node.child_nodes[actions2[0]] = child_node\n except: \n return child_node\n parent_node.child_nodes[actions1[0]] = child_node\n #print(\"it broke\", actions2, \"also :\", actions1)\n # print(\"this is the length after\",len(parent_node.child_nodes) )\n return child_node\n\n \n # Hint: return new_node\n\n\ndef rollout(board, state):\n \"\"\" Given the state of the game, the rollout plays out the remainder randomly.\n Args:\n board: The game setup.\n state: The state of the game.\n return state/ board at the end - see who won\n \"\"\"\n\n # Return who won this game (board or state)\n # board.is_ended(state) == true when over\n # Same as Rollout bot, can basically copy/paste\n # pass\n if not board.is_ended(state):\n prevmov = board.legal_actions(state)[0]\n nextmove = prevmov\n predistance = float('inf')\n while not board.is_ended(state):\n predistance = float('inf')\n\n for move in board.legal_actions(state):\n if move == prevmov:\n\n continue\n x1 = move[2]\n x2 = prevmov[2]\n y1 = move[3]\n y2 = prevmov[3]\n if prevmov[1]!= move[1] or prevmov[0]!= move[0]:\n distance = sqrt((x1 - x2) ** 2 + (y2 - y1) ** 2 + (move[0] - prevmov[0]) ** 2 + (move[1] - prevmov[1]) ** 2) * 0.5\n else:\n distance = sqrt((x1 - x2) ** 2 + (y2 - y1) ** 2) * 0.5\n \n if predistance >distance:\n predistance = distance\n nextmove = move\n prevmov = nextmove\n state = board.next_state(state,nextmove)\n\n return board.current_player(state)\n\n\ndef backpropagate(node, won):\n \"\"\" Navigates the tree from a leaf node to the root, updating the win and visit count of each node along the path.\n \n Args:\n node: A leaf node.\n won: An indicator of whether the bot won or lost the game.\n \"\"\"\n if node.parent == None:\n node.visits += 1\n node.wins += won\n return node\n node.wins += won\n node.visits += 1\n\n return backpropagate(node.parent, won)\n\n\ndef think(board, state):\n \"\"\" Performs MCTS by sampling games and calling the appropriate functions to construct the game tree.\n Args:\n board: The game setup.\n state: The state of the game.\n Returns: The action to be taken.\n \"\"\"\n identity_of_bot = board.current_player(state)\n root_node = MCTSNode(parent=None, parent_action=None, action_list=board.legal_actions(state))\n i = 0\n start = time()\n currenttime = 0\n\n for step in range(num_nodes):\n # Copy the game for sampling a playthrough\n sampled_game = state\n times = time()\n currenttime = times -start\n # Start at root\n #print(\"this is the length : \",len(root_node.child_nodes))\n\n leaf_node = traverse_nodes(root_node, board, sampled_game, identity_of_bot)\n child_node = expand_leaf(leaf_node, board, state)\n if len(child_node.untried_actions) == 0:\n print(\"drawsss\")\n bestAction = child_node.parent_action\n break\n parent_node = child_node.parent\n actions = parent_node.untried_actions\n wins = rollout(board, board.next_state(state, actions[0])) \n\n temp = child_node\n temp.visits += 1\n temp.wins += wins\n while temp.parent !=None:\n temp = temp.parent\n temp.visits += 1\n temp.wins += wins\n root_node = temp\n #print(currenttime)\n high=-1\n #print(\"test2\")\n for node in root_node.child_nodes:\n temp2=root_node.child_nodes[node]\n if ((temp2.wins/temp2.visits)>high) and node!=None:\n score=(temp2.wins/temp2.visits)\n best=node\n\n return best","sub_path":"pa3/mcts_modified.py","file_name":"mcts_modified.py","file_ext":"py","file_size_in_byte":7119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"145910491","text":"import acm\nael_variables = [\n['vertical_shift', 'Vertical Shift (%)', 'double', None, None, 1, 0, '', None, True],\n['tilt', 'Tilt (%)', 'double', None, None, 1, 0, '', None, True]\n]\n\ndef ael_main_ex( parameters, dictExtra ):\n params = acm.FNamedParameters()\n name = str('')\n vertical = parameters['vertical_shift']\n tilt = parameters['tilt']\n if (vertical != 0.0):\n name = name + str(vertical) + '% vertical'\n if (tilt != 0.0):\n name = name + str(' & ') + str(tilt) + '% tilt'\n elif (tilt != 0.0):\n name = name + str(tilt) + '% tilt'\n params.Name(name)\n params.AddParameter( 'vertical_shift', parameters['vertical_shift'] )\n params.AddParameter( 'tilt', parameters['tilt'] )\n vec = acm.FArray()\n vec.Add( params )\n return vec\n","sub_path":"Extensions/Default/FPythonCode/baseCorrelationScenarioGetParams.py","file_name":"baseCorrelationScenarioGetParams.py","file_ext":"py","file_size_in_byte":794,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"157909712","text":"from Bio import SeqIO\nfrom Bio.SeqUtils import GC\n\nfile = str(input(\"File: \"))\nreads = list(SeqIO.parse(file, \"fasta\"))\n\nif len(reads[0].seq)%3 == 1:\n\treads[0].seq += 'NN'\nelif len(reads[0].seq)%3 == 2:\n\treads[0].seq += 'N'\n\nrf1 = reads[0].seq\nrf2 = reads[0].seq[1:len(reads[0].seq)-2]\nrf3 = reads[0].seq[2:len(reads[0].seq)-1]\n#rf4 = reads[0].seq.reverse_complement()\n#rf5 = reads[0].seq[2:len(reads[0].seq)-1].reverse_complement()\n#rf6 = reads[0].seq[1:len(reads[0].seq)-2].reverse_complement()\n\nseqs1 = rf1.translate()\nseqs2 = rf2.translate()\nseqs3 = rf3.translate()\n#seqs4 = rf4.translate()\n#seqs5 = rf5.translate()\n#seqs6 = rf6.translate()\n\nlstart1 = []\nlstart2 = []\nlstart3 = []\n#lstart4 = []\n#lstart5 = []\n#lstart6 = []\nc1 = 0\nc2 = 0\nc3 = 0\n#c4 = 0\n#c5 = 0\n#c6 = 0\nn1 = 0\nn2 = 0\nn3 = 0\n#n4 = 0\n#n5 = 0\n#n6 = 0\n\ndef locs(x, lstart, c):\n\tplength = 0\n\tstartstop = {}\n\tfor i in range(0, len(x)):\n\t\tif plength == 0:\n\t\t\tstart = int(i)\n\t\tif x[i] != '*':\n\t\t\tplength += 1\n\t\telif x[i] == '*':\n\t\t\tif plength > 100:\n\t\t\t\tlstart.append(start)\n\t\t\t\tstartstop[lstart[c]] = int(i)\n\t\t\t\tc += 1\n\t\t\tplength = 0\n\treturn startstop\n\ndef prot(x, lstart, n):\n\tsequences = {}\n\tproteins = x.split('*')\n\tfor i in range(0, len(proteins)):\n\t\tif len(proteins[i]) > 100:\n\t\t\tsequences[lstart[n]] = ''.join(proteins[i])\n\t\t\tn += 1\n\treturn sequences\n\nlocations1 = locs(seqs1, lstart1, c1)\nproteins1 = prot(seqs1, lstart1, n1)\n\nlocations2 = locs(seqs2, lstart2, c2)\nproteins2 = prot(seqs2, lstart2, n2)\n\nlocations3 = locs(seqs3, lstart3, c3)\nproteins3 = prot(seqs3, lstart3, n3)\n\n#locations4 = locs(seqs4, lstart4, c4)\n#proteins4 = prot(seqs4, lstart4, n4)\n\n#locations5 = locs(seqs5, lstart5, c5)\n#proteins5 = prot(seqs5, lstart5, n5)\n\n#locations6 = locs(seqs6, lstart6, c6)\n#proteins6 = prot(seqs6, lstart6, n6)\n\nprint()\nprint('In reading frame 1:')\nprint(f'The proteins longer than 100 amino acids are: {proteins1}')\nprint(f'The endpoints of the proteins are: {locations1}')\nprint()\nprint('Reading frame 2 is adjusted by +1. In reading frame 2:')\nprint(f'The proteins longer than 100 amino acids are: {proteins2}')\nprint(f'The endpoints of the proteins are: {locations2}')\nprint()\nprint('Reading frame 3 is adjusted by +2. In reading frame 3:')\nprint(f'The proteins longer than 100 amino acids are: {proteins3}')\nprint(f'The endpoints of the proteins are: {locations3}')\nprint()\n#print('Reading frame 4 is the reverse complement of reading frame 1. In reading frame 4:')\n#print(f'The proteins longer than 100 amino acids are: {proteins4}')\n#print(f'The endpoints of the proteins are: {locations4}')\n#print()\n#print('Reading frame 5 is the reverse complement of reading frame 1 adjusted by +1. In reading frame 5:')\n#print(f'The proteins longer than 100 amino acids are: {proteins5}')\n#print(f'The endpoints of the proteins are: {locations5}')\n#print()\n#print('Reading frame 6 is the reverse complement of reading frame 1 adjusted by +2. In reading frame 6:')\n#print(f'The proteins longer than 100 amino acids are: {proteins6}')\n#print(f'The endpoints of the proteins are: {locations6}')\n#print()\n","sub_path":"rvokkarne/bio-translation.py","file_name":"bio-translation.py","file_ext":"py","file_size_in_byte":3066,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"529569228","text":"def find_min(number_list):\n lowest_num = number_list[0]\n for num in number_list:\n if num < lowest_num:\n lowest_num = num\n return lowest_num\n\ndef tally_sort(num_list):\n min_num = find_min(num_list)\n ind_list = []\n sorted_list = []\n filler_list = []\n for num in range(len(num_list)):\n num_list[num]-=min_num\n ind_list.append(0)\n print(num_list)\n print(ind_list)\n for ind in range(len(ind_list)):\n for num in range(len(num_list)):\n if ind == num_list[num]:\n ind_list[ind]+=1\n for tallies in range(len(ind_list)):\n if ind_list[tallies] > 0:\n for num in range(ind_list[tallies]):\n sorted_list.append(ind_list.index(ind_list[tallies]))\n\n\n\n \n \n print(num_list)\n print(ind_list)\n print(sorted_list)\n \n \n \ntally_sort([2, 5, 2, 3, 8, 6, 3])\n","sub_path":"tally_sort.py","file_name":"tally_sort.py","file_ext":"py","file_size_in_byte":913,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"109781141","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n__author__ = \"Рузэль Давлетяров (ruzelapp@yandex.ru)\"\n\nimport os\nimport sys\nimport textwrap\nimport configparser\nfrom urllib.request import urlopen\nfrom urllib.parse import urlparse\n\n# Cторонний модуль - pip install beautifulsoup4\nfrom bs4 import BeautifulSoup\n\n\nif __name__ == '__main__':\n\n url = str(sys.argv[1])\n urlpars = urlparse(url)\n domain = urlpars.netloc\n config = configparser.ConfigParser()\n config.read('config.ini','utf8')\n\n # Определяем путь\n directoryname = os.getcwd() #Если мой код, будет запущен из другой директории, то результат будет тут\n headpath = os.path.join(directoryname,str(urlpars.netloc)) \n tailpath = urlpars.path.replace('/',os.sep) \n path = headpath + tailpath\n\n # Создаем каталоги\n try:\n os.makedirs(path) \n except FileExistsError: #Подовлям исключение, т.к некоторые каталоги могли быть созданы пред. работой\n pass\n\n #Возня с кодировками \n try:\n data = urlopen(url).read().decode('utf8') \n except UnicodeDecodeError:\n try:\n data = urlopen(url).read().decode('cp1251')\n except UnicodeDecodeError:\n data = urlopen(url).read().decode('utf-8','ignore')\n \n #html.parser - стандартный парсер из std Python\n soup = BeautifulSoup(data, \"html.parser\")\n \n #Записываем заголовок\n source = \"\\n\".join(textwrap.wrap(soup.h1.string, width=80, replace_whitespace=False)) + \"\\n\\n\" \n \n #Проход по всем параграфам\n for i in soup.find_all('p'):\n if domain in config: #Если шаблон задан в config.ini\n confset = set(config[domain]['class'].split(','))\n for paren in i.parents:\n if paren.attrs.get('class') != None:\n parentset = set(paren.attrs.get('class'))\n if not parentset.issubset(confset): #Если все классы из атрибута имеются в шаблоне\n lcl_data = i.get_text()\n for link in i.children:\n if link.name == \"a\": #Тут обрабатываем ссылки\n lcl_data = lcl_data.replace(link.text, link.text + ' [' + link['href'] + ']')\n source += \"\\n\".join(textwrap.wrap(lcl_data, width=80, replace_whitespace=False)) + '\\n\\n'\n break \n else:\n lcl_data = i.get_text()\n for link in i.children:\n if link.name == \"a\":\n lcl_data = lcl_data.replace(link.text, link.text + ' [' + link['href'] + ']') \n\n source += \"\\n\".join(textwrap.wrap(lcl_data, width=80, replace_whitespace=False)) + '\\n\\n'\n \n\n # Записываем результат в\n f1 = open(os.path.join(path,'index.txt'), 'w+', encoding=\"utf8\")\n f1.write(source)\n f1.close() \n \n","sub_path":"appreadability.py","file_name":"appreadability.py","file_ext":"py","file_size_in_byte":3182,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"90603056","text":"# 剑指 Offer 14- II. 剪绳子 II\r\n# 给你一根长度为 n 的绳子,请把绳子剪成整数长度的 m 段(m、n都是整数,n>1并且m>1),每段绳子的长度记为 k[0],k[1]...k[m - 1] 。\r\n# 请问 k[0]*k[1]*...*k[m - 1] 可能的最大乘积是多少?例如,当绳子的长度是8时,我们把它剪成长度分别为2、3、3的三段,此时得到的最大乘积是18。\r\n# 答案需要取模 1e9+7(1000000007),如计算初始结果为:1000000008,请返回 1。\r\nclass Solution:\r\n # 5.01 21.45\r\n def cuttingRope(self, n: int) -> int:\r\n # dp[i]代表长度为i所得的最大乘积\r\n dp = [0 for _ in range(n + 1)]\r\n for i in range(2, n + 1):\r\n for j in range(1, i):\r\n dp[i] = max(dp[i], max(j * (i - j), j * dp[i - j]))\r\n return dp[n] % (1000000007)\r\n\r\n # 优秀解答\r\n def cuttingRope(self, n: int) -> int:\r\n sm = [0, 1, 1, 2, 4]\r\n if n <= 4:\r\n return sm[n]\r\n a = n // 3\r\n b = n % 3\r\n if b == 0:\r\n return 3 ** a % 1000000007\r\n else:\r\n return 3 ** (a + b - 2) * (6 - 2 * b) % 1000000007\r\n\r\n # 优秀解答\r\n def cuttingRope(self, n: int) -> int:\r\n\r\n if n <= 3: return n - 1\r\n\r\n res = 1\r\n\r\n while n > 4:\r\n n = n - 3\r\n res = res * 3 % 1000000007\r\n\r\n return res * n % 1000000007","sub_path":"pySrc/CutTheRope2.py","file_name":"CutTheRope2.py","file_ext":"py","file_size_in_byte":1413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"62974214","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Dec 23 10:12:49 2017\n\n@author: Harshvardhan Gazula\n@notes : Contains a variant of the single-shot regression proposed by Eswar.\n Drop site specific columns at each site and use global_beta_vector\n to calculate SSE at each local site.\n@modified: 01/14/2018 weighted average single shot regression\n\"\"\"\n\nimport os\nimport shelve\nimport numpy as np\nfrom numba import jit, prange\nimport pandas as pd\nimport scipy as sp\nfrom mcic_load_data import load_data\n\n@jit(nopython=True)\ndef singleshot_regression(X1, site_01_y1, X2, site_02_y1, X3, site_03_y1, X4,\n site_04_y1):\n\n size_y = site_01_y1.shape[1]\n\n params = np.zeros((X1.shape[1], size_y))\n sse = np.zeros(size_y)\n tvalues = np.zeros((X1.shape[1], size_y))\n rsquared = np.zeros(size_y)\n\n for voxel in prange(size_y):\n print(voxel)\n\n y1 = site_01_y1[:, voxel]\n y2 = site_02_y1[:, voxel]\n y3 = site_03_y1[:, voxel]\n y4 = site_04_y1[:, voxel]\n\n # Start of single shot regression\n beta1 = np.linalg.inv(X1.T @ X1) @ (X1.T @ y1)\n beta2 = np.linalg.inv(X2.T @ X2) @ (X2.T @ y2)\n beta3 = np.linalg.inv(X3.T @ X3) @ (X3.T @ y3)\n beta4 = np.linalg.inv(X4.T @ X4) @ (X4.T @ y4)\n\n # PART 02 - Aggregating parameter values at the remote\n count_y_local = np.array([len(y1), len(y2), len(y3), len(y4)])\n\n # Weighted Average\n count_y_local_float = count_y_local.astype(np.float64)\n sum_params = np.column_stack((beta1, beta2, beta3, beta4))\n avg_beta_vector = sum_params @ count_y_local_float / np.sum(count_y_local)\n\n# =============================================================================\n# # Simple Average\n# avg_beta_vector = (beta1 + beta2 + beta3 + beta4) / 4\n# =============================================================================\n\n params[:, voxel]= avg_beta_vector\n\n # PART 03 - SSE at each local site\n y1_estimate = np.dot(avg_beta_vector, X1.T)\n y2_estimate = np.dot(avg_beta_vector, X2.T)\n y3_estimate = np.dot(avg_beta_vector, X3.T)\n y4_estimate = np.dot(avg_beta_vector, X4.T)\n\n sse1 = np.linalg.norm(y1 - y1_estimate)**2\n sse2 = np.linalg.norm(y2 - y2_estimate)**2\n sse3 = np.linalg.norm(y3 - y3_estimate)**2\n sse4 = np.linalg.norm(y4 - y4_estimate)**2\n\n # At Local\n mean_y_local = np.array([np.mean(y1), np.mean(y2), np.mean(y3), np.mean(y4)])\n\n # At Remote\n mean_y_global = np.sum(\n mean_y_local * count_y_local) / np.sum(count_y_local)\n\n # At Local\n sst1 = np.sum(np.square(y1 - mean_y_global))\n sst2 = np.sum(np.square(y2 - mean_y_global))\n sst3 = np.sum(np.square(y3 - mean_y_global))\n sst4 = np.sum(np.square(y4 - mean_y_global))\n\n cov1 = X1.T @ X1\n cov2 = X2.T @ X2\n cov3 = X3.T @ X3\n cov4 = X4.T @ X4\n\n # PART 05 - Finding rsquared (global)\n SSE_global = sse1 + sse2 + sse3 + sse4\n sse[voxel] = SSE_global\n SST_global = sst1 + sst2 + sst3 + sst4\n r_squared_global = 1 - (SSE_global / SST_global)\n rsquared[voxel] = r_squared_global\n\n # PART 04 - Finding p-value at the Remote\n varX_matrix_global = cov1 + cov2 + cov3 + cov4\n\n dof_global = np.sum(count_y_local) - len(avg_beta_vector)\n\n MSE = SSE_global / dof_global\n var_covar_beta_global = MSE * np.linalg.inv(varX_matrix_global)\n se_beta_global = np.sqrt(np.diag(var_covar_beta_global))\n ts_global = avg_beta_vector / se_beta_global\n\n tvalues[:, voxel] = ts_global\n\n return (params, sse, tvalues, rsquared, dof_global)\n\n\nfolder_index = input('Enter the name of the folder to save results: ')\nfolder_name = folder_index.replace(' ', '_')\nif not os.path.exists(folder_name):\n os.makedirs(folder_name)\n\nX1, site_01_y1, X2, site_02_y1, X3, site_03_y1, X4, site_04_y1, column_name_list = load_data(\n)\n\n(params, sse, tvalues, rsquared, dof_global) = singleshot_regression(\n X1, site_01_y1, X2, site_02_y1, X3, site_03_y1, X4, site_04_y1)\n\n\nps_global = 2 * sp.stats.t.sf(np.abs(tvalues), dof_global)\npvalues = pd.DataFrame(ps_global.transpose(), columns=column_name_list)\nsse = pd.DataFrame(sse.transpose(), columns=['sse'])\nparams = pd.DataFrame(params.transpose(), columns=column_name_list)\ntvalues = pd.DataFrame(tvalues.transpose(), columns=column_name_list)\nrsquared = pd.DataFrame(rsquared.transpose(), columns=['rsquared_adj'])\n\n# %% Writing to a file\nprint('Writing data to a shelve file')\nresults = shelve.open(os.path.join(folder_name, 'singleshotWA_results'))\nresults['params'] = params\nresults['sse'] = sse\nresults['pvalues'] = pvalues\nresults['tvalues'] = tvalues\nresults['rsquared'] = rsquared\nresults.close()\n","sub_path":"mcic_regression_singleshot.py","file_name":"mcic_regression_singleshot.py","file_ext":"py","file_size_in_byte":4865,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"335985491","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.8-x86_64/egg/vint/file_util.py\n# Compiled at: 2013-04-19 06:08:22\nfrom __future__ import unicode_literals\nimport json, logging, os, codecs\n__author__ = b'tchen'\nlogger = logging.getLogger(__name__)\nFIRST_EXTENSIONS = [\n b'.h']\nINSTRUCTION_FILE = b'README'\nEXAM_CONFIG_FILE = b'.interview.json'\nCASE_CONFIG_FILE = b'.case.json'\nEXAM_INSTRUCTION_TEMPLATE = b'\\nHello %(applicant)s, Welcome to exam %(name)s\\n\\n%(description)s\\n\\nInstructions:\\n\\n1. Write the code as fast as you can. Optimize when you have further time.\\n2. Verify the correctness and robustness of your code with proper output.\\n3. When you finish the exam, please go back to this directory (where you see this file), and execute \"vint finish\".\\n This is very important to do so since we will time your exam and submit your result back to the hiring manager.\\n\\nStart your journey now, pal!\\n\\n'\nCASE_INSTRUCTION_TEMPLATE = b\"\\n\\nCase%(position)d: %(name)s\\n\\n%(description)s\\n\\nInstructions:\\n\\n1. You need to code in %(lang)s, with acceptable extentisons: %(extentions)s.\\n2. You'd better to write down the code inside one file unless you find it is not readable.\\n\"\n\ndef write_file(filename, content):\n f = codecs.open(filename, b'w+', encoding=b'utf8')\n f.write(content)\n f.close()\n\n\nclass Template(object):\n\n @staticmethod\n def create_exam_config(exam_path, interview):\n filename = os.path.join(os.getcwd(), exam_path, EXAM_CONFIG_FILE)\n content = json.dumps(interview)\n write_file(filename, content)\n\n @staticmethod\n def create_exam_instruction(exam_path, interview, exam):\n filename = os.path.join(os.getcwd(), exam_path, INSTRUCTION_FILE)\n content = EXAM_INSTRUCTION_TEMPLATE % {b'applicant': interview[b'applicant'], \n b'name': exam[b'name'], \n b'description': exam[b'description']}\n write_file(filename, content)\n\n @staticmethod\n def create_case_config(case_path, case):\n filename = os.path.join(case_path, CASE_CONFIG_FILE)\n content = json.dumps(case)\n write_file(filename, content)\n\n @staticmethod\n def create_case_instruction(case_path, case):\n instruction = os.path.join(case_path, INSTRUCTION_FILE)\n content = CASE_INSTRUCTION_TEMPLATE % {b'position': case[b'position'], \n b'name': case[b'name'], \n b'description': case[b'description'], \n b'lang': case[b'lang'], \n b'extentions': case[b'extentions']}\n write_file(instruction, content)\n\n @staticmethod\n def create_case_code(case_path, case):\n ext = case[b'extentions'].split(b',')[0].strip()\n filename = os.path.join(case_path, b'main%s' % ext)\n write_file(filename, case[b'code'])\n\n\nclass FileUtil(object):\n\n @staticmethod\n def read_content(filename):\n return codecs.open(filename, b'r', encoding=b'utf8').read()\n\n @staticmethod\n def get_valid_files(path, extentions):\n first_list = []\n second_list = []\n normal_list = []\n for root, dirs, files in os.walk(path):\n for f in files:\n for ext in extentions:\n if f.endswith(ext):\n normal_list.append(f)\n break\n\n for f in normal_list:\n for ext in FIRST_EXTENSIONS:\n if f.endswith(ext):\n first_list.append(f)\n break\n else:\n second_list.append(f)\n\n return (\n first_list, second_list)\n\n @staticmethod\n def read_case(path):\n return FileUtil.read_json(path, CASE_CONFIG_FILE)\n\n @staticmethod\n def read_json(path, name):\n filename = os.path.join(path, name)\n return json.load(codecs.open(filename, b'r', encoding=b'utf8'))\n\n @staticmethod\n def read_interview(path):\n return FileUtil.read_json(path, EXAM_CONFIG_FILE)\n\n @staticmethod\n def interview_exists():\n return os.path.exists(EXAM_CONFIG_FILE)","sub_path":"pycfiles/vint-0.1.5-py2.7/file_util.py","file_name":"file_util.py","file_ext":"py","file_size_in_byte":4164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"486263412","text":"from dam import Dam\n\nimport math\nimport numpy as np\nfrom numpy.linalg import norm\nfrom gym.envs.registration import register\n\n#Radial Basis Function\ndef rbf(c,w,s):\n return math.exp(-norm(s-c)/w)\n\ndef spaceout(n,lo,hi):\n step = (hi - lo)/(n+1)\n return [lo+k*step for k in range(1,n+1)]\n \nregister(\n id='DamWrap-v0',\n entry_point='dam_wrap:DamWrap',\n)\n\n#Wrapper for Dam environment\nclass DamWrap(Dam):\n def __init__(self,dreward=2,penalize=False):\n self.centers = spaceout(2,-20,190)\n self.dim = len(self.centers)+1\n self.width = 60 \n \n super(DamWrap,self).__init__()\n \n self.obj_weights = [0.5,0.5]\n assert len(self.obj_weights)==dreward\n assert sum(self.obj_weights)==1\n\n def features(self,s):\n features = [1]\n for c in self.centers:\n features.append(rbf(c,self.width,s))\n return features\n\n def step(self,action,noise,render=False):\n s,r,done,info = super(DamWrap,self).step(action,render,noise)\n \n #Model state: RBFs\n new_s = self.features(s) \n\n #Model reward: convex combination\n new_r = np.dot(self.obj_weights,r)\n\n return np.array(new_s),new_r,done,info\n\n def reset(self,state=None,scalar=False):\n super(DamWrap,self).reset(state)\n if not scalar:\n return np.array(self.features(self.get_state()))\n else:\n return self.get_state()\n","sub_path":"dam/dam_wrap.py","file_name":"dam_wrap.py","file_ext":"py","file_size_in_byte":1464,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"249629905","text":"\"\"\"\nSettings for maintaining and running an AWS Lambda function.\n\nauthor: Shang-Lin Chen\n\n\"\"\"\n\nAWS_PROFILE = 'default' # Profile in .aws/credentials\n # that has S3 and Lambda permissions.\n\nAWS_REGION = 'us-west-2' # AWS region where the lambda function\n # is deployed. Should be the same\n # region as all S3 buckets.\n\n# AWS IAM role that has Lambda and S3 permissions.\nIAM_ROLE = 'put_iam_role_here'\n\nLAMBDA_FUNCTION = 'DecimationFunc' # Name of your lambda function.\n\nLAMBDA_BUCKET = 'lambda-venv-bucket' # Name of the S3 bucket that will \n # store a zip file containing \n # the lambda function and environment.\n\n# Description of what your lambda function does.\n# This is one of the settings sent to AWS.\nLAMBDA_DESCRIPTION = 'Testing Lambda with ObsPy!'\n\nINPUT_BUCKET = 'scedc-pds' # S3 bucket used as s3_input_bucket.\n # This does not need to be modified\n # if using the SCEDC Public Data Set.\n\nOUTPUT_BUCKET = 'output-bucket' # Name of the S3 bucket where Lambda\n # output will be written.\n\nNCORES = 4 # Number of cores to use for calling Lambda function.\n\n","sub_path":"pds-lambda-example/settings_example.py","file_name":"settings_example.py","file_ext":"py","file_size_in_byte":1377,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"426761252","text":"# -*- coding: utf-8 -*-\n\nfrom itertools import chain, product\n\nfrom nose.tools import raises\n\nimport friendlysam as fs\nfrom friendlysam import Storage, FlowNetwork\n\nfrom friendlysam.tests.simple_models import Producer, Consumer, RESOURCE\nfrom friendlysam.tests import default_solver, approx\nfrom friendlysam import Problem, Minimize, Sum\n\ndef test_basic_functionality():\n times = range(1,4)\n\n consumption = lambda t: t * 1.5\n V0 = 10\n\n p = Producer(name='Producer')\n c = Consumer(consumption, name='Consumer')\n s = Storage(RESOURCE, capacity=15, name='Storage')\n s.volume(0).value = V0\n rn = FlowNetwork(RESOURCE)\n rn.connect(p, s)\n rn.connect(s, c)\n\n prob = Problem()\n prob += (part.constraints.make(t) for part, t in product(rn.descendants_and_self, times))\n\n prob.objective = Minimize(Sum(p.cost(t) for t in times))\n\n solution = default_solver.solve(prob)\n\n for t in times:\n c.activity(t).take_value(solution)\n p.activity(t).take_value(solution)\n s.volume(t).take_value(solution)\n\n for t in times:\n assert approx(p.activity(t).value, 0)\n assert approx(c.consumption[RESOURCE](t).value, consumption(t))\n assert approx(s.volume(t).value, s.volume(t-1).value + s.accumulation[RESOURCE](t-1).value)\n","sub_path":"friendlysam/tests/test_storage.py","file_name":"test_storage.py","file_ext":"py","file_size_in_byte":1288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"341455015","text":"\"\"\"\nClassify sentences using trained LinearSVC\n\"\"\"\n\nimport tkinter as tk\nimport pkg_resources\nimport joblib\n\nvectorizer = joblib.load(pkg_resources.resource_filename('Project', 'LinearSVC Approach/vectorizer.joblib'))\nmodel = joblib.load(pkg_resources.resource_filename('Project', 'LinearSVC Approach/model.joblib'))\n\ntext_color = 'black'\n\n\ndef classify_text():\n \"\"\"Classify a text as profane or not-profane\n \"\"\"\n word = ent_word.get(\"1.0\", tk.END)\n result = model.predict(vectorizer.transform([word]))\n if result[0] == 0:\n lbl_classification[\"text\"] = 'Result : Not profane.'\n else:\n lbl_classification[\"text\"] = 'Result : Profane.'\n\n\nwindow = tk.Tk()\nwindow.title(\"Word Classifier\")\n\n\ndef handle_keypress(event):\n \"\"\"Reset classification.\"\"\"\n # print(event.char)\n lbl_classification[\"text\"] = ''\n\n\n# Bind keypress event to handle_keypress()\nwindow.bind(\"\", handle_keypress)\n\nfrm_entry = tk.Frame(master=window)\nlbl_instruction = tk.Label(master=frm_entry, text=\"Enter the sentence to classify :\")\nlbl_instruction.config(font=(\"Courier\", 20))\nent_word = tk.Text(master=frm_entry, width=100, height=5)\nlbl_instruction.grid(row=0, column=0, sticky=\"we\")\nent_word.grid(row=1, column=0, sticky=\"we\")\nbtn_convert = tk.Button(\n master=window,\n text=\"Classify \\N{RIGHTWARDS BLACK ARROW}\",\n command=classify_text\n)\nlbl_classification = tk.Label(master=window, text=\"\", fg=text_color)\nlbl_classification.config(font=(\"Courier\", 40))\nfrm_entry.grid(row=0, column=0, padx=10)\nbtn_convert.grid(row=1, column=0, padx=10, pady=10, sticky=\"w\")\nlbl_classification.grid(row=1, column=0, padx=10, sticky=\"e\")\n\nwindow.mainloop()\n","sub_path":"Gui/classify_sentence.py","file_name":"classify_sentence.py","file_ext":"py","file_size_in_byte":1669,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"379920576","text":"from gui.messages import Message\n\n\nclass Inventory:\n def __init__(self, capacity=26, items=None):\n self.owner = None\n if items is None:\n items = []\n self.capacity = capacity\n self.items = items\n\n def add_to_inv(self, item):\n if len(self.items) + 1 <= self.capacity:\n self.items.append(item)\n return True\n else:\n return False\n\n def remove_from_inv(self, item):\n \"\"\" removes an item from the player main inventory or any quick use slots \"\"\"\n if item in (self.items):\n self.items.remove(item)\n else:\n # check all quick use slots if the item wasn't found\n for e in self.owner.paperdoll.equipped_items:\n if getattr(e, 'slots',0):\n if item in e.qu_inventory.items:\n e.qu_inventory.items.remove(item)\n\n def is_full(self):\n return True if len(self.items) == self.capacity else False\n\nclass QuickUseInv(Inventory):\n def __init__(self, capacity =6, attached_to=None):\n super().__init__(capacity=capacity)\n\n self.attached_to = attached_to\n\n def add_to_qu(self, item):\n owner = self.attached_to.owner\n if owner:\n # # get the index of the first slot returning None\n # #try:\n # #idx = next(idx for idx, v in enumerate(self.items) if v is None)\n # # if that fails (all slots taken) return false\n # except:\n # return False\n #if owner.inventory.add_to_inv(item):\n #else:\n # if item is added to the qu inventory, remove it from the owner inventory\n if self.add_to_inv(item):\n #self.items[idx] = item\n owner.inventory.remove_from_inv(item)\n return True\n else:\n return False\n\n\n def remove_from_qu(self, item):\n owner = self.attached_to.owner\n if owner:\n if owner.inventory.add_to_inv(item):\n self.items[self.items.index(item)] = None\n return True\n else:\n return False\n\n def empty_to_inv(self):\n owner = self.attached_to.owner\n for i in self.items:\n #if i is not None:\n if owner.inventory.add_to_inv(i):\n Message('You move the {0} from your {1} to your backpack.'.format(i.name, self.attached_to))\n else:\n Message('As your backpack is full, you had to drop the {0} on the ground.'.format(i.name, self.attached_to))\n i.drop()\n\n # empty the inventory of removed quick-use equipment\n self.items = []\n # for i in range(self.capacity):\n # self.items[i] = None\n\n def swap_qu_inv(self, other_inv):\n \"\"\" swaps the contents of two quick_use inventories \"\"\"\n for i in self.items:\n if len(other_inv.items) + 1 <= other_inv.capacity:\n other_inv.items.append(i)\n Message('You move the {0} to your new {1}.'.format(i.name, other_inv.attached_to))\n\n # any items still stored are dumped to the inventory\n if len(self.items):\n self.empty_to_inv()\n","sub_path":"components/inventory.py","file_name":"inventory.py","file_ext":"py","file_size_in_byte":3242,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"391183474","text":"#!/usr/bin/env python\n\"\"\" Kafka Bootstrap Script (based on Kafka 1.0.0 release) \"\"\"\nimport datetime\nimport logging\nimport os\nimport shutil\nimport signal\nimport socket\nimport sys\nimport time\nimport urllib.error\nimport urllib.request\nimport uuid\nfrom optparse import OptionParser\n\nimport hostlist\nimport pkg_resources\nfrom pykafka import KafkaClient\n\nfrom pilot.util.ssh_utils import execute_ssh_command, execute_ssh_command_as_daemon\n\nlogging.basicConfig(level=logging.DEBUG)\n\n# For automatic Download and Installation\n# VERSION=\"0.10.1.0\"\nVERSION = \"2.8.1\" \n# KAFKA_DOWNLOAD_URL = \"http://www-us.apache.org/dist/kafka/\" + VERSION + \"/kafka_2.11-\" + VERSION + \".tgz\"\n# KAFKA_DOWNLOAD_URL = \"http://apache.mirrors.lucidnetworks.net/kafka/\"+ VERSION + \"/kafka_2.11-\" + VERSION + \".tgz\"\n# KAFKA_DOWNLOAD_URL = \"http://mirrors.gigenet.com/apache/kafka/\"+ VERSION + \"/kafka_2.11-\" + VERSION + \".tgz\"\n# KAFKA_DOWNLOAD_URL = \"http://mirrors.gigenet.com/apache/kafka/\" + VERSION + \"/kafka_2.13-\" + VERSION + \".tgz\"\n# KAFKA_DOWNLOAD_URL = \"https://downloads.apache.org/kafka/\" + VERSION + \"/kafka_2.13-\" + VERSION + \".tgz\"\n\nKAFKA_DOWNLOAD_URL = \"https://archive.apache.org/dist/kafka/\" + VERSION + \"/kafka_2.13-\" + VERSION + \".tgz\"\nWORKING_DIRECTORY = os.path.join(os.getcwd())\n\n# For using an existing installation\nKAFKA_HOME = os.path.join(os.getcwd(), \"kafka-\" + VERSION)\nKAFKA_CONF_DIR = os.path.join(KAFKA_HOME, \"etc\")\n\nSTOP = False\n\n\ndef handler(signum, frame):\n logging.debug(\"Signal catched. Stop Kafka\")\n global STOP\n STOP = True\n time.sleep(10)\n os.system(\"killall java\")\n sys.exit(0)\n\n\n\nclass KafkaBootstrap():\n\n def __init__(self, working_directory, kafka_home, config_name=\"default\", extension_job_id=None):\n self.working_directory = working_directory\n self.kafka_home = kafka_home\n self.config_name = config_name\n self.jobid = \"kafka-\" + str(uuid.uuid1())\n self.job_working_directory = os.path.join(WORKING_DIRECTORY)\n self.job_conf_dir = os.path.join(self.job_working_directory, \"config\")\n self.extension_job_id = extension_job_id\n self.broker_config_files = {}\n try:\n os.makedirs(self.job_conf_dir)\n os.system(\"rm -rf /tmp/zookeeper*\")\n os.system(\"rm -rf /tmp/kafka-logs*\")\n except:\n pass\n\n def get_server_properties(self, master, hostname, broker_id):\n module = \"pilot.plugins.kafka.configs.\" + self.config_name\n print((\"Access config in module: \" + module + \" File: server.properties\"))\n my_data = pkg_resources.resource_string(module, \"server.properties\").decode(\"utf-8\")\n # print(my_data)\n # Find out external IP\n external_ip = hostname\n try:\n # check whether this host has an external ip that should be used\n external_ip = urllib.request.urlopen('https://ident.me').read().decode('utf8')\n logging.debug(\"External IP discovered: {}\".format(external_ip))\n except:\n logging.debug(\"No external IP discovered.\")\n\n # have at least 4 Kafka log directories containing broker id in config template\n my_data = my_data % (broker_id, hostname, external_ip, broker_id, broker_id, broker_id, broker_id, master)\n my_data = os.path.expandvars(my_data)\n return my_data\n\n def get_zookeeper_properties(self, hostname):\n module = \"pilot.plugins.kafka.configs.\" + self.config_name\n logging.debug(\"Access config in module: \" + module + \" File: zookeeper.properties\")\n my_data = pkg_resources.resource_string(module, \"zookeeper.properties\").decode(\"utf-8\")\n return my_data\n\n #######################################################################################\n ## Get Node List from Resource Management System\n @staticmethod\n def get_pbs_allocated_nodes():\n print(\"Init PBS\")\n pbs_node_file = os.environ.get(\"PBS_NODEFILE\")\n if pbs_node_file == None:\n return [\"localhost\"]\n f = open(pbs_node_file)\n nodes = f.readlines()\n for i in nodes:\n i.strip()\n f.close()\n return list(set(nodes))\n\n @staticmethod\n def get_sge_allocated_nodes():\n logging.debug(\"Init SGE or Local\")\n sge_node_file = os.environ.get(\"PE_HOSTFILE\")\n if sge_node_file == None:\n return [\"localhost\"]\n f = open(sge_node_file)\n sgenodes = f.readlines()\n f.close()\n nodes = []\n for i in sgenodes:\n columns = i.split()\n try:\n for j in range(0, int(columns[1])):\n print((\"add host: \" + columns[0].strip()))\n nodes.append(columns[0] + \"\\n\")\n except:\n pass\n nodes.reverse()\n return list(set(nodes))\n\n @staticmethod\n def get_slurm_allocated_nodes():\n print(\"Init nodefile from SLURM_NODELIST\")\n hosts = os.environ.get(\"SLURM_NODELIST\")\n if hosts == None:\n return [\"localhost\"]\n\n print(\"***** Hosts: \" + str(hosts))\n hosts = hostlist.expand_hostlist(hosts)\n number_cpus_per_node = 1\n if os.environ.get(\"SLURM_CPUS_ON_NODE\") != None:\n number_cpus_per_node = int(os.environ.get(\"SLURM_CPUS_ON_NODE\"))\n freenodes = []\n for h in hosts:\n # for i in range(0, number_cpus_per_node):\n freenodes.append((h + \"\\n\"))\n return list(set(freenodes))\n\n @staticmethod\n def get_nodelist_from_resourcemanager():\n if (os.environ.get(\"PBS_NODEFILE\") != None and os.environ.get(\"PBS_NODEFILE\") != \"\"):\n nodes = KafkaBootstrap.get_pbs_allocated_nodes()\n elif (os.environ.get(\"SLURM_NODELIST\") != None):\n nodes = KafkaBootstrap.get_slurm_allocated_nodes()\n else:\n nodes = KafkaBootstrap.get_sge_allocated_nodes()\n return nodes\n\n #######################################################################################\n def configure_kafka(self):\n logging.debug(\"Kafka Instance Configuration Directory: \" + self.job_conf_dir)\n nodes = self.get_nodelist_from_resourcemanager()\n logging.debug(\"Kafka nodes: \" + str(nodes))\n master = socket.gethostname().split(\".\")[0]\n\n for idx, node in enumerate(nodes):\n path = os.path.join(self.job_conf_dir, \"broker-%d\" % idx)\n try:\n os.makedirs(path)\n except:\n pass #do nothing as path exists\n server_properties_filename = os.path.join(path, \"server.properties\")\n server_properties_file = open(server_properties_filename, \"w\")\n server_properties_file.write(\n self.get_server_properties(master=master, hostname=node.strip(), broker_id=idx))\n server_properties_file.close()\n self.broker_config_files[node] = server_properties_filename\n\n zookeeper_properties_file = open(os.path.join(self.job_conf_dir, \"zookeeper.properties\"), \"w\")\n zookeeper_properties_file.write(self.get_zookeeper_properties(master))\n zookeeper_properties_file.close()\n\n def start_kafka(self):\n logging.debug(\"Start Kafka\")\n os.system(\"killall -s 9 java\")\n os.system(\"pkill -9 java\")\n time.sleep(5)\n\n logging.debug(\"Start Zookeeper\")\n start_command = os.path.join(self.kafka_home, \"bin/zookeeper-server-start.sh\") + \" -daemon \" + os.path.join(\n self.job_conf_dir, \"zookeeper.properties\")\n logging.debug(\"Execute: %s\" % start_command)\n os.system(\". ~/.bashrc & \" + start_command)\n\n logging.debug(\"Start Kafka Cluster\")\n # remove dangling keys referencing localhost\n os.system(\"ssh-keygen -f $HOME/.ssh/known_hosts -R localhost\")\n for node in list(self.broker_config_files.keys()):\n config = self.broker_config_files[node]\n start_command = os.path.join(self.kafka_home, \"bin/kafka-server-start.sh\") + \" -daemon \" + config\n result = execute_ssh_command_as_daemon(node.strip(),\n user=None,\n command=start_command,\n keyfile=None)\n print(\"Host: {} Command: {} Result: {}\".format(node.strip(), start_command, result))\n #logging.debug(\"Execute: %s\" % start_command)\n #os.system(\". ~/.bashrc & \" + start_command)\n\n print((\"Kafka started with configuration: %s\" % self.job_conf_dir))\n\n def check_kafka(self):\n brokers = {}\n try:\n master = socket.gethostname().split(\".\")[0]\n client = KafkaClient(zookeeper_hosts=master + \":2181\")\n brokers = client.brokers\n except:\n pass\n print(\"Found %d brokers: %s\" % (len(list(brokers.keys())), str(brokers)))\n return brokers\n\n def start(self):\n self.configure_kafka()\n self.start_kafka()\n\n ##################################################################################################################\n # Extend cluster\n def start_kafka_extension(self):\n logging.debug(\"Start Kafka\")\n os.system(\"killall -s 9 java\")\n os.system(\"pkill -9 java\")\n time.sleep(5)\n\n logging.debug(\"Start Kafka Cluster\")\n # remove dangling keys referencing localhost\n os.system(\"ssh-keygen -f $HOME/.ssh/known_hosts -R localhost\")\n\n for node in list(self.broker_config_files.keys()):\n config = self.broker_config_files[node]\n start_command = os.path.join(self.kafka_home, \"bin/kafka-server-start.sh\") + \" -daemon \" + config\n result = execute_ssh_command(node.strip(),\n user=None,\n command=start_command,\n keyfile=None)\n print(\"Host: {} Command: {} Result: {}\".format(node.strip(), start_command, result))\n\n #logging.debug(\"Execute: %s\" % start_command)\n #os.system(\". ~/.bashrc & \" + start_command)\n\n print((\"Kafka started with configuration: %s\" % self.job_conf_dir))\n\n def configure_kafka_extension(self):\n logging.debug(\"Kafka Instance Configuration Directory: \" + self.job_conf_dir)\n nodes = self.get_nodelist_from_resourcemanager()\n logging.debug(\"Kafka nodes: \" + str(nodes))\n master = self.find_parent_zookeeper()\n max_id = self.find_max_broker_id()\n for node in nodes:\n idx = max_id + 1\n path = os.path.join(self.job_conf_dir, \"broker-%d\" % idx)\n try:\n os.makedirs(path)\n except:\n pass # do nothing as path exist\n server_properties_filename = os.path.join(path, \"server.properties\")\n server_properties_file = open(server_properties_filename, \"w\")\n server_properties_file.write(\n self.get_server_properties(master=master, hostname=node.strip(), broker_id=idx))\n server_properties_file.close()\n self.broker_config_files[node] = server_properties_filename\n\n def find_max_broker_id(self):\n path_to_parent_kafka_configs = os.path.join(os.getcwd(), \"..\", self.extension_job_id, \"config\")\n files = os.listdir(path_to_parent_kafka_configs)\n max_id = 0\n for i in files:\n bid = 0\n try:\n bid = int(i[-1])\n except:\n pass\n if bid > max_id:\n max_id = bid\n return max_id\n\n def find_parent_zookeeper(self):\n path_to_parent_spark_job = os.path.join(os.getcwd(), \"..\", self.extension_job_id,\n \"config/broker-0/server.properties\")\n print(\"Master of Parent Cluster: %s\" % path_to_parent_spark_job)\n zk = None\n with open(path_to_parent_spark_job, \"r\") as config:\n lines = config.readlines()\n for line in lines:\n if line.startswith(\"zookeeper.connect=\"):\n zk = line.strip().split(\"=\")[1]\n zk = zk.split(\":\")[0]\n\n logging.debug(\"Parent Zookeeper: %s\" % zk)\n return zk\n\n def extend(self):\n self.configure_kafka_extension()\n self.start_kafka_extension()\n\n ##################################################################################################################\n # Stop\n def stop(self):\n self.stop_kafka()\n\n def stop_kafka(self):\n logging.debug(\"Stop Kafka\")\n self.set_env()\n stop_command = os.path.join(KAFKA_HOME, \"bin/kafka-server-stop.sh\")\n logging.debug(\"Execute: %s\" % stop_command)\n stop_command = os.path.join(KAFKA_HOME, \"bin/zookeeper-server-stop.sh\")\n logging.debug(\"Execute: %s\" % stop_command)\n os.system(stop_command)\n\n ##################################################################################################################\n # Utils\n\n\n#########################################################\n# main #\n#########################################################\nif __name__ == \"__main__\":\n\n signal.signal(signal.SIGALRM, handler)\n signal.signal(signal.SIGABRT, handler)\n signal.signal(signal.SIGQUIT, handler)\n signal.signal(signal.SIGINT, handler)\n\n parser = OptionParser()\n parser.add_option(\"-s\", \"--start\", action=\"store_true\", dest=\"start\",\n help=\"start Kafka\", default=True)\n parser.add_option(\"-j\", \"--job\", type=\"string\", action=\"store\", dest=\"jobid\",\n help=\"Job ID of Kafka Cluster to Extend\")\n parser.add_option(\"-q\", \"--quit\", action=\"store_false\", dest=\"start\",\n help=\"terminate Hadoop\")\n parser.add_option(\"-c\", \"--clean\", action=\"store_true\", dest=\"clean\",\n help=\"clean Kafka topics in Zookeeper after termination\")\n\n parser.add_option(\"-n\", \"--config_name\", action=\"store\", type=\"string\", dest=\"config_name\", default=\"default\")\n # parser.add_option(\"-m\", \"--machines\", action=\"store\", type=\"string\", dest=\"config_name\")\n\n (options, args) = parser.parse_args()\n config_name = options.config_name\n logging.debug(\"Bootstrap Kafka on \" + socket.gethostname())\n\n node_list = KafkaBootstrap.get_nodelist_from_resourcemanager()\n number_nodes = len(node_list)\n print(\"nodes: %s\" % str(node_list))\n\n\n #################################################################################################################\n # Download Kafka\n run_timestamp = datetime.datetime.now()\n performance_trace_filename = \"kafka_performance_\" + run_timestamp.strftime(\"%Y%m%d-%H%M%S\") + \".csv\"\n kafka_config_filename = \"kafka_config_\" + run_timestamp.strftime(\"%Y%m%d-%H%M%S\")\n try:\n os.makedirs(WORKING_DIRECTORY)\n except:\n pass\n performance_trace_file = open(os.path.join(WORKING_DIRECTORY, performance_trace_filename), \"a\")\n start = time.time()\n # performance_trace_file.write(\"start_time, %.5f\"%(time.time()))\n\n filename = os.path.basename(KAFKA_DOWNLOAD_URL)\n kafka_home = \"\"\n if not os.path.exists(KAFKA_HOME):\n try:\n os.makedirs(WORKING_DIRECTORY)\n except:\n pass\n\n download_destination = os.path.join(WORKING_DIRECTORY, filename)\n if os.path.exists(download_destination) == False:\n logging.debug(\"Download: %s to %s\" % (KAFKA_DOWNLOAD_URL, download_destination))\n opener = urllib.request.FancyURLopener({})\n opener.retrieve(KAFKA_DOWNLOAD_URL, download_destination);\n else:\n logging.debug(\"Found existing Kafka binaries at: \" + download_destination)\n logging.debug(\"Install Kafka\")\n\n os.chdir(WORKING_DIRECTORY)\n os.system(\"tar -xzf %s\" % filename)\n kafka_home = os.path.join(WORKING_DIRECTORY, os.path.splitext(filename)[0])\n os.environ[\"KAFKA_HOME\"] = kafka_home\n\n end_download = time.time()\n performance_trace_file.write(\"download, %d, %.5f\\n\" % (number_nodes, end_download - start))\n performance_trace_file.flush()\n\n # initialize object for managing kafka clusters\n kafka = KafkaBootstrap(WORKING_DIRECTORY, kafka_home, config_name, options.jobid)\n if options.jobid is not None and options.jobid != \"None\":\n logging.debug(\"Extend Kafka Cluster with PS ID: %s\" % options.jobid)\n kafka.extend()\n end_start = time.time()\n performance_trace_file.write(\"startup-extension, %d, %.5f\\n\" % (number_nodes, (end_start - end_download)))\n performance_trace_file.flush()\n with open(\"kafka_started\", \"w\") as f:\n f.write(str(node_list))\n elif options.start:\n kafka.start()\n number_brokers = 0\n while number_brokers != number_nodes:\n brokers = kafka.check_kafka()\n number_brokers = len(list(brokers.values()))\n logging.debug(\"Number brokers: %d, number nodes: %d\" % (number_brokers, number_nodes))\n time.sleep(1)\n end_start = time.time()\n performance_trace_file.write(\"startup, %d, %.5f\\n\" % (number_nodes, (end_start - end_download)))\n performance_trace_file.flush()\n with open(\"kafka_started\", \"w\") as f:\n f.write(str(node_list))\n else:\n kafka.stop()\n if options.clean:\n directory = \"/tmp/zookeeper/\"\n logging.debug(\"delete: \" + directory)\n shutil.rmtree(directory)\n sys.exit(0)\n\n print(\"Finished launching of Kafka Cluster - Sleeping now\")\n\n while STOP == False:\n logging.debug(\"stop: \" + str(STOP))\n time.sleep(10)\n\n kafka.stop()\n os.remove(os.path.join(WORKING_DIRECTORY, \"kafka_started\"))\n performance_trace_file.write(\"total_runtime, %d, %.5f\\n\" % (number_nodes, time.time() - start))\n performance_trace_file.flush()\n performance_trace_file.close()\n","sub_path":"pilot/plugins/kafka/bootstrap_kafka.py","file_name":"bootstrap_kafka.py","file_ext":"py","file_size_in_byte":18008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"184301952","text":"from SpriteReader import SpriteReader\nfrom MapClasses.CollisionMap import CollisionMap\nfrom MapClasses.BackgroundMap import BackgroundMap\nimport pygame as pyg\nimport init\nfrom InputHandler import InputHandler\nfrom Renderer import Renderer\n\n\ndef main():\n maps = {\n \"collision_map\": CollisionMap(\"src\\maps\\starting_point.txt\"),\n \"background_map\": BackgroundMap()\n }\n screen, scene = init.game()\n running = True\n input_handler = InputHandler(scene, maps)\n sprites = SpriteReader()\n\n renderer = Renderer(screen, scene, maps, sprites)\n while running:\n\n main_loop(input_handler, scene, renderer, maps)\n\n\ndef main_loop(input_handler, scene, renderer, maps):\n\n # This checks for quit and changes scene.player_phsyics\n input_handler.get_events()\n scene.player_physics.update_pos(scene)\n renderer.paint()\n\n\nmain()\n","sub_path":"src/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":861,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"129117667","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[85]:\n\n\nimport cufflinks as cf\nfrom plotly.offline import download_plotlyjs,init_notebook_mode,plot,iplot\nimport pandas as pd\nimport numpy as np\nimport plotly\nimport plotly.plotly as py\nimport plotly.graph_objs as go \n\ninit_notebook_mode(connected=True)\ncf.go_offline()\ndata1 = pd.read_csv(\"master.csv\")\n\nworldMeanSuicides = data1.groupby([\"country\"])[\"suicides_no\"].sum()\ndf1 = pd.DataFrame({\"country\" : worldMeanSuicides.index,\n \"year\" : worldMeanSuicides.values\n })\n\ndef show_values_on_bars(axs):\n def _show_on_single_plot(ax): \n for p in ax.patches:\n _x = p.get_x() + p.get_width() / 2\n _y = p.get_y() + p.get_height()\n value = '{:.2f}'.format(p.get_height())\n ax.text(_x, _y, value, ha=\"center\") \n\n if isinstance(axs, np.ndarray):\n for idx, ax in np.ndenumerate(axs):\n _show_on_single_plot(ax)\n else:\n _show_on_single_plot(axs)\n\ndf = pd.DataFrame({\"year\" : worldMeanSuicides.index,\n \"suicides\" : worldMeanSuicides.values\n })\n\n\ndata = dict(type='choropleth',\n locations = df1[\"country\"],\n locationmode = \"country names\",\n z = df1[\"year\"],\n colorscale = [\n [0, \"rgb(5, 10, 172)\"],\n [0.35, \"rgb(40, 60, 190)\"],\n [0.5, \"rgb(70, 100, 245)\"],\n [0.6, \"rgb(90, 120, 245)\"],\n [0.7, \"rgb(106, 137, 247)\"],\n [1, \"rgb(220, 220, 220)\"]\n ],\n autocolorscale = False,\n reversescale = True,\n colorbar = {\"title\" : \"Number Of Suicides\"},\n marker = go.choropleth.Marker(\n line = go.choropleth.marker.Line(\n color = 'rgb(180,180,180)',\n width = 0.5\n )),\n )\n\nlayout = dict(\n title = 'NUMBER OF SUICIDES PER COUNTRY SINCE 1987-2016',\n geo = dict(\n showframe = False,\n projection = {'type' : \"natural earth\"} )\n)\n\nchoromap = go.Figure(data = [data],layout = layout)\n#iplot(choromap)\nplotly.offline.plot(choromap.iplot())\n\n\n\n\n\n\n\n","sub_path":"Suicides.py","file_name":"Suicides.py","file_ext":"py","file_size_in_byte":2114,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"587908071","text":"from math import *\n\n#standard gravity constant\ngravk = 6.67e-11\n\nmphk = 2236.9362\n\n#fev stands for find escape velosity\ndef fev( mass, dist):\n\tev = sqrt((2 * gravk * mass) / dist)\n\treturn ev\n\nif __name__ == \"__main__\":\n\tnumber_obj = int(input('Input the number of objects to use'))\n\twhile i < number_obj:\n\t\tobj1_m = float(input('Input the mass of object 1 in kg: '))\n\t\tobj1_dkm = float(input('Input the distance from object 1 km: '))\n\t\tobj1_d = obj1_dkm * 1000\n\t\tobj1_ev = fev(obj1_m, obj1_d)/1000\n\n#\t\tobj2_m = float(input('Input the mass of object 2 kg: '))\n#\t\tobj2_dkm = float(input('Input the distance from object 2 km: '))\n#\t\tobj2_d = obj2_dkm * 1000\n#\t\tobj2_ev = fev(obj2_m, obj2_d)/1000\n\n\t\tprint('Escape Velocity of object 1 is ' + str(obj1_ev) + 'KM/s or ' + str(obj1_ev * mphk) + ' MPH')\n#\t\typrint('Escape Velocity of object 2 is ' + str(obj2_ev) + 'KM/s or ' + str(obj2_ev * mphk) + 'MPH')\n","sub_path":"old-cs111/labs/3/excape-vrepeat.py","file_name":"excape-vrepeat.py","file_ext":"py","file_size_in_byte":899,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"148324256","text":"from datetime import date, datetime\n\nfrom dateutil.relativedelta import relativedelta\n\nfrom casexml.apps.case.const import CASE_INDEX_EXTENSION\nfrom casexml.apps.case.mock import CaseStructure, CaseIndex\n\nfrom corehq.apps.locations.models import SQLLocation\nfrom custom.enikshay.private_sector_datamigration.models import (\n Adherence,\n Episode,\n EpisodePrescription,\n LabTest,\n)\n\nfrom dimagi.utils.decorators.memoized import memoized\n\nPERSON_CASE_TYPE = 'person'\nOCCURRENCE_CASE_TYPE = 'occurrence'\nEPISODE_CASE_TYPE = 'episode'\nADHERENCE_CASE_TYPE = 'adherence'\nPRESCRIPTION_CASE_TYPE = 'prescription'\nTEST_CASE_TYPE = 'test'\n\n\ndef get_human_friendly_id():\n return datetime.utcnow().strftime('%Y%m%d%H%M%S%f')[:-3]\n\n\nclass BeneficiaryCaseFactory(object):\n\n def __init__(self, domain, beneficiary):\n self.domain = domain\n self.beneficiary = beneficiary\n\n def get_case_structures_to_create(self):\n person_structure = self.get_person_case_structure()\n ocurrence_structure = self.get_occurrence_case_structure(person_structure)\n episode_structure = self.get_episode_case_structure(ocurrence_structure)\n episode_descendants = [\n self.get_adherence_case_structure(adherence, episode_structure)\n for adherence in self._adherences\n ] + [\n self.get_prescription_case_structure(prescription, episode_structure)\n for prescription in self._prescriptions\n ]\n episode_or_descendants = episode_descendants or [episode_structure]\n\n tests = [\n self.get_test_case_structure(labtest, ocurrence_structure)\n for labtest in self._labtests\n ]\n\n return episode_or_descendants + tests\n\n def get_person_case_structure(self):\n kwargs = {\n 'attrs': {\n 'case_type': PERSON_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'update': {\n 'current_address': self.beneficiary.current_address,\n 'current_episode_type': self.beneficiary.current_episode_type,\n 'dataset': 'real',\n 'first_name': self.beneficiary.firstName,\n 'last_name': self.beneficiary.lastName,\n 'name': ' '.join([self.beneficiary.firstName, self.beneficiary.lastName]),\n 'occupation': '',\n 'phone_number': self.beneficiary.phoneNumber,\n 'secondary_contact_phone_number': self.beneficiary.emergencyContactNo,\n\n 'migration_created_case': 'true',\n 'migration_created_from_record': self.beneficiary.caseId,\n }\n }\n }\n\n if self.beneficiary.age_entered is not None:\n kwargs['attrs']['update']['age'] = self.beneficiary.age_entered\n kwargs['attrs']['update']['age_entered'] = self.beneficiary.age_entered\n else:\n if self.beneficiary.dob is not None:\n kwargs['attrs']['update']['age'] = relativedelta(\n self.beneficiary.creationDate, self.beneficiary.dob\n ).years\n else:\n kwargs['attrs']['update']['age'] = ''\n kwargs['attrs']['update']['age_entered'] = ''\n\n if self.beneficiary.dob is not None:\n kwargs['attrs']['update']['dob'] = self.beneficiary.dob.date()\n kwargs['attrs']['update']['dob_known'] = 'yes'\n else:\n if self.beneficiary.age_entered is not None:\n kwargs['attrs']['update']['dob'] = date(date.today().year - self.beneficiary.age_entered, 7, 1),\n else:\n kwargs['attrs']['update']['dob'] = ''\n kwargs['attrs']['update']['dob_known'] = 'no'\n\n if self.beneficiary.sex is not None:\n kwargs['attrs']['update']['sex'] = self.beneficiary.sex\n\n if self.beneficiary.has_aadhaar_number:\n kwargs['attrs']['update']['aadhaar_number'] = self.beneficiary.identificationNumber\n else:\n kwargs['attrs']['update']['other_id_number'] = self.beneficiary.identificationNumber\n kwargs['attrs']['update']['other_id_type'] = self.beneficiary.other_id_type\n\n if self._episode:\n kwargs['attrs']['update']['hiv_status'] = self._episode.hiv_status\n kwargs['attrs']['update']['current_patient_type_choice'] = self._episode.current_patient_type_choice\n\n agency = (\n self._episode.treating_provider or self.beneficiary.referred_provider\n if self._episode else self.beneficiary.referred_provider\n )\n assert agency is not None\n\n kwargs['attrs']['owner_id'] = SQLLocation.active_objects.get(\n domain=self.domain,\n site_code=agency.nikshayId,\n ).location_id\n\n return CaseStructure(**kwargs)\n\n def get_occurrence_case_structure(self, person_structure):\n kwargs = {\n 'attrs': {\n 'case_type': OCCURRENCE_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'owner_id': '-',\n 'update': {\n 'current_episode_type': self.beneficiary.current_episode_type,\n 'name': 'Occurrence #1',\n 'occurrence_id': get_human_friendly_id(),\n\n 'migration_created_case': 'true',\n 'migration_created_from_record': self.beneficiary.caseId,\n }\n },\n 'indices': [CaseIndex(\n person_structure,\n identifier='host',\n relationship=CASE_INDEX_EXTENSION,\n related_type=PERSON_CASE_TYPE,\n )],\n }\n return CaseStructure(**kwargs)\n\n def get_episode_case_structure(self, occurrence_structure):\n kwargs = {\n 'attrs': {\n 'case_type': EPISODE_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'owner_id': '-',\n 'update': {\n 'adherence_schedule_id': 'schedule_mwf',\n 'date_of_mo_signature': self.beneficiary.dateOfRegn.date(),\n 'dots_99_enabled': 'false',\n 'episode_id': get_human_friendly_id(),\n 'episode_type': self.beneficiary.current_episode_type,\n 'name': self.beneficiary.episode_name,\n 'transfer_in': '',\n\n 'migration_created_case': 'true',\n 'migration_created_from_record': self.beneficiary.caseId,\n }\n },\n 'indices': [CaseIndex(\n occurrence_structure,\n identifier='host',\n relationship=CASE_INDEX_EXTENSION,\n related_type=OCCURRENCE_CASE_TYPE,\n )],\n }\n\n if self._episode:\n rx_start_date = self._episode.rxStartDate.date()\n kwargs['attrs']['date_opened'] = rx_start_date\n kwargs['attrs']['update']['adherence_schedule_date_start'] = rx_start_date\n kwargs['attrs']['update']['date_of_diagnosis'] = self._episode.dateOfDiagnosis.date()\n kwargs['attrs']['update']['disease_classification'] = self._episode.disease_classification\n kwargs['attrs']['update']['episode_pending_registration'] = (\n 'yes' if self._episode.nikshayID is None else 'no'\n )\n kwargs['attrs']['update']['treatment_card_completed_date'] = self._episode.creationDate.date()\n kwargs['attrs']['update']['treatment_initiated'] = 'yes_private'\n kwargs['attrs']['update']['treatment_initiation_date'] = rx_start_date\n kwargs['attrs']['update']['weight'] = int(self._episode.patientWeight)\n\n if self._episode.nikshayID:\n kwargs['attrs']['external_id'] = self._episode.nikshayID\n kwargs['attrs']['update']['nikshay_id'] = self._episode.nikshayID\n\n if self._episode.disease_classification == 'extra_pulmonary':\n kwargs['attrs']['update']['site_choice'] = self._episode.site_choice\n else:\n kwargs['attrs']['update']['episode_pending_registration'] = 'yes'\n kwargs['attrs']['update']['treatment_initiated'] = 'no'\n\n return CaseStructure(**kwargs)\n\n def get_adherence_case_structure(self, adherence, episode_structure):\n kwargs = {\n 'attrs': {\n 'case_type': ADHERENCE_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'owner_id': '-',\n 'update': {\n 'adherence_date': adherence.doseDate.date(),\n 'adherence_value': adherence.adherence_value,\n 'name': adherence.doseDate.date(),\n\n 'migration_created_case': 'true',\n 'migration_created_from_record': adherence.adherenceId,\n }\n },\n 'indices': [CaseIndex(\n episode_structure,\n identifier='host',\n relationship=CASE_INDEX_EXTENSION,\n related_type=EPISODE_CASE_TYPE,\n )],\n }\n return CaseStructure(**kwargs)\n\n def get_prescription_case_structure(self, prescription, episode_structure):\n kwargs = {\n 'attrs': {\n 'case_type': PRESCRIPTION_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'owner_id': '-',\n 'update': {\n 'name': prescription.productName,\n\n 'migration_created_case': 'true',\n 'migration_created_from_record': prescription.prescriptionID,\n }\n },\n 'indices': [CaseIndex(\n episode_structure,\n identifier='host',\n relationship=CASE_INDEX_EXTENSION,\n related_type=EPISODE_CASE_TYPE,\n )],\n }\n return CaseStructure(**kwargs)\n\n def get_test_case_structure(self, labtest, occurrence_structure):\n kwargs = {\n 'attrs': {\n 'case_type': TEST_CASE_TYPE,\n 'close': False,\n 'create': True,\n 'owner_id': '-',\n 'update': {\n 'migration_created_case': 'true',\n }\n },\n 'indices': [CaseIndex(\n occurrence_structure,\n identifier='host',\n relationship=CASE_INDEX_EXTENSION,\n related_type=OCCURRENCE_CASE_TYPE,\n )],\n }\n return CaseStructure(**kwargs)\n\n @property\n @memoized\n def _episode(self):\n episodes = Episode.objects.filter(beneficiaryID=self.beneficiary).order_by('-episodeDisplayID')\n if episodes:\n return episodes[0]\n else:\n return None\n\n @property\n @memoized\n def _adherences(self):\n return list(Adherence.objects.filter(episodeId=self._episode)) if self._episode else []\n\n @property\n @memoized\n def _prescriptions(self):\n return list(EpisodePrescription.objects.filter(beneficiaryId=self.beneficiary))\n\n @property\n @memoized\n def _labtests(self):\n if self._episode:\n return list(LabTest.objects.filter(episodeId=self._episode))\n else:\n return []\n","sub_path":"custom/enikshay/private_sector_datamigration/factory.py","file_name":"factory.py","file_ext":"py","file_size_in_byte":11487,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"454211429","text":"from django.db import models\n\n# (후보들의) 정보를 담을수 있는 공간, 장고에서는 모델이라고 한다.\n# Create your models here.\nclass Candidate(models.Model):\n name = models.CharField(max_length=10)\n introduction = models.TextField()\n area = models.CharField(max_length=15)\n party_numbeer = models.IntegerField(default=0) # 기본값은 1부터 시작.\n\n def __str__(self): # 이클래스를 표현할때 어떻게 표현할지에 대한 내용\n return self.name # 이 클래스를 표현할때는 후보자의 이름으로 표현한다. 브라우저를 확인하면 DB에 저장된 이름이 항목에 표시된다.\n\n\nclass Poll2(models.Model):\n start_date = models.DateTimeField() # 조사를 시작하는 시간\n end_date = models.DateTimeField() # 조사를 종료하는 시간\n area = models.CharField(max_length=15) # 지역\n # id는 모델을 만들면 자동으로 생성된다.\n\n\nclass Choice4(models.Model):\n # 다른 클래스를 사용할때 ForeignKey 를 사용한다. 인수로 on_delete 넣어줘야한다.\n poll = models.ForeignKey(Poll2, on_delete=models.CASCADE)\n candidate = models.ForeignKey(Candidate, on_delete=models.CASCADE)\n votes = models.IntegerField(default=0)\n","sub_path":"elections/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1263,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"383061901","text":"#-*- coding: UTF-8 -*-\nfrom five import grok\nfrom z3c.form import group, field\nfrom datetime import datetime\n\nfrom persistent.list import PersistentList\nfrom plone.dexterity.interfaces import IDexterityContent\n\nfrom zope.annotation.interfaces import IAnnotations\nfrom zope.lifecycleevent.interfaces import IObjectAddedEvent\nfrom zope.component import getUtility\nfrom zope.component import getMultiAdapter\nfrom Acquisition import aq_parent, aq_base, Implicit\nfrom plone.portlets.interfaces import IPortletManager\nfrom plone.portlets.interfaces import IPortletAssignmentMapping\nfrom zope.container.interfaces import INameChooser\nfrom Products.CMFCore.utils import getToolByName\nfrom Products.CMFPlone.utils import safe_unicode\n\nfrom zope.i18n import translate\nfrom zope.i18nmessageid import Message\nfrom zope.interface import implements\nfrom zope.lifecycleevent.interfaces import IObjectRemovedEvent\nfrom zope.component import adapter\nfrom zope.interface import implementer\n\nfrom emc.kb.contents.topic import Itopic\nfrom emc.kb.contents.question import Iquestion\nfrom emc.kb.contents.topicfolder import Itopicfolder\nfrom emc.kb.interfaces import IFollowedEvent\nfrom emc.kb.interfaces import IUnFollowedEvent\nfrom emc.kb.interfaces import IFollowing,IFollowable\nfrom emc.kb.interfaces import ICountNumEvent\nfrom emc.kb.interfaces import PROMOTIONS_PORTLET_COLUMN \nfrom emc.kb.portlets import relatedinformation\nfrom plone.uuid.interfaces import IUUID\n\n\n\nfrom emc.kb import _\n\nFOLLOWED_KEY = 'emc.kb.follow'\n\n@implementer(IFollowing)\n@adapter(IDexterityContent)\nclass Follow(object):\n# grok.provides(IFollowing)\n# grok.context(Itopic)\n \n def __init__(self, context):\n self.context = context\n \n annotations = IAnnotations(context)\n if FOLLOWED_KEY not in annotations.keys():\n annotations[FOLLOWED_KEY] = PersistentList() \n self.followed = annotations[FOLLOWED_KEY]\n \n #Statistics concern the number of\n @property\n def followerNum(self):\n total = len(self.followed)\n return total\n \n #Determine whether to be concerned about\n def available(self, userToken):\n return not(userToken in self.followed)\n# return self.followed.has_key(userToken) \n #Editing statistics concern the number of \n def addFollow(self, userToken):\n if self.available(userToken):\n self.followed.append(userToken)\n else:\n raise KeyError(\"The %s is concerned about\" % userToken)\n #Editing statistics concern the number of \n def delFollow(self, userToken):\n if not self.available(userToken):\n self.followed.remove(userToken)\n else:\n raise KeyError(\"The %s is not concerned about\" % userToken)\n\n@grok.subscribe(IFollowable,IFollowedEvent)\ndef SubscriberFollowed(obj,event):\n mp = getToolByName(obj,'portal_membership')\n# import pdb\n# pdb.set_trace()\n userobject = mp.getAuthenticatedMember()\n username = userobject.getId()\n fwqlist = list(userobject.getProperty('myfollowquestion'))\n fwtlist = list(userobject.getProperty('myfollowtopic'))\n fwlist = list(userobject.getProperty('myfollow'))\n\n \n uuid = IUUID(obj,None)\n if uuid == None:return\n if not (uuid in fwlist):\n fwlist.append(uuid) \n userobject.setProperties(myfollow=fwlist)\n \n if Iquestion.providedBy(obj) and not(uuid in fwqlist):\n fwqlist.append(uuid) \n userobject.setProperties(myfollowquestion=fwqlist) \n if Itopic.providedBy(obj) and not(uuid in fwtlist):\n fwtlist.append(uuid) \n userobject.setProperties(myfollowtopic=fwtlist) \n \n evlute = IFollowing(obj) \n if evlute.available(username):\n evlute.addFollow(username)\n obj.followernum = evlute.followerNum\n obj.reindexObject() \n \n@grok.subscribe(IFollowable,IUnFollowedEvent)\ndef SubscriberUnFollowed(obj,event):\n mp = getToolByName(obj,'portal_membership')\n userobject = mp.getAuthenticatedMember()\n username = userobject.getId()\n fwqlist = list(userobject.getProperty('myfollowquestion'))\n fwtlist = list(userobject.getProperty('myfollowtopic')) \n fwlist = list(userobject.getProperty('myfollow'))\n\n uuid = IUUID(obj,None)\n if uuid == None:return\n if (uuid in fwlist):\n fwlist.remove(uuid)\n userobject.setProperties(myfollow=fwlist)\n if Iquestion.providedBy(obj) and (uuid in fwqlist):\n fwqlist.remove(uuid) \n userobject.setProperties(myfollowquestion=fwqlist) \n if Itopic.providedBy(obj) and (uuid in fwtlist):\n fwtlist.remove(uuid) \n userobject.setProperties(myfollowtopic=fwtlist) \n \n evlute = IFollowing(obj) \n if not evlute.available(username):\n evlute.delFollow(username)\n obj.followernum = evlute.followerNum\n obj.reindexObject() \n\n@grok.subscribe(IFollowable, IObjectRemovedEvent)\ndef delFollow(obj,event):\n followevlute = IFollowing(obj)\n fwlist = followevlute.followed\n if len(fwlist) == 0:\n return\n \n pm = getToolByName(obj, 'portal_membership')\n for userid in fwlist:\n userobject=pm.getMemberById(userid)\n fwqlist = list(userobject.getProperty('myfollowquestion'))\n fwtlist = list(userobject.getProperty('myfollowtopic')) \n fwlist = list(userobject.getProperty('myfollow'))\n# userfollow = list(userobject.getProperty('myfollow'))\n uuid = IUUID(obj,None)\n if uuid in fwlist:\n fwlist.remove(uuid)\n userobject.setProperties(myfollow=fwlist)\n if uuid in fwtlist:\n fwtlist.remove(uuid)\n userobject.setProperties(myfollowtopic=fwtlist)\n if uuid in fwqlist:\n fwqlist.remove(uuid) \n userobject.setProperties(myfollowquestion=fwqlist) \n\n@grok.subscribe(Itopic,ICountNumEvent)\ndef CountNum(obj,event):\n num=obj.visitnum\n visitnum=num+1\n obj.visitnum=visitnum\n\n@grok.subscribe(Itopic,IObjectAddedEvent)\ndef AddTopicPortlet(obj, event):\n \"\"\" this will add the relatedinformation portlet to topic automatically.\"\"\"\n \n parent = aq_parent(obj)\n if Itopic.providedBy(parent):\n return\n \n # A portlet manager is akin to a column\n column = getUtility(IPortletManager, name=PROMOTIONS_PORTLET_COLUMN)\n \n # We multi-adapt the object and the column to an assignment mapping,\n # which acts like a dict where we can put portlet assignments\n manager = getMultiAdapter((obj, column,), IPortletAssignmentMapping)\n \n # We then create the assignment and put it in the assignment manager,\n # using the default name-chooser to pick a suitable name for us.\n assignment = relatedinformation.Assignment()\n chooser = INameChooser(manager)\n manager[chooser.chooseName(None, assignment)] = assignment\n","sub_path":"emc/kb/behaviors/follow.py","file_name":"follow.py","file_ext":"py","file_size_in_byte":6958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"296617587","text":"# -*- coding: utf-8 -*-\n\n#%% md #\n# Exploratory Data Analysis\n\n# Prix de Vente des Proprietées\n\n\n#%%\nimport pandas as pd\nimport os\nfrom pathlib import Path\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport quickda\n\npd.options.display.max_rows = 999\npd.options.display.max_columns = 999\n\nimport seaborn as sns\n\n\n#import warnings\n#warnings.filterwarnings('ignore')\n\n#%%\n# Set Fix Parame\n \nSEED = 1234 # Seed for random number geneartors\n\n#%% md \n### Load data\n\n\n#%%\n\n\n#ROOT_PATH = os.path.abspath(os.path.dirname(__file__))\ndata_folder = '../../data/raw/'\n\ndsets = {} \n\nfilenames = [ \n #'dvf_2018.gz',\n 'dvf_2019.gz'\n ]\n\n#%% \n # Util Fn to read write out a data sample from origina raw data files\ndef pd_read_save(filename, sample_size, output_path):\n # n = sum(1 for line in open(filename, errors = \"ignore\")) - 1 #number of records in file (excludes header)\n # print(\"total rows\")\n # s = sample_size #desired sample sizfre\n # skip = sorted(np.random.random_integers(1,n+1,n-s)) #the 0-indexed header will not be included in the skip list\n df = pd.read_csv(filename, error_bad_lines = False) #, skiprows=skip\n df_sample = df.sample(sample_size, random_state = SEED)\n df_sample.to_csv(output_path, sep = \";\", index = False)\n print('Writing file of shape: ', df_sample.shape)\n return df\n\n\n\n#%%\n# Read saved data samples\nuse_samples = False\n\nfor f in filenames:\n base_name = f.split('.')[0]\n if not use_samples:\n data_path = data_folder + f\n print(data_folder +'samples/'+f.split('.')[0] +'_sample.csv.gz')\n dsets[base_name] = pd_read_save( data_path, sample_size=400000\n , output_path = data_folder +'samples/'+f.split('.')[0] +'_sample.csv.gz')\n else:\n dsets[base_name] = pd.read_csv( data_folder + 'samples/' + base_name + '_sample.csv.gz'\n , error_bad_lines = False, sep = ';'\n )\n \n\n\ndsets.keys()\n\n\n#%% md \n#### Merge the Data Sets \n\n\n#%%\n# Merge the data sets\n\ndata_in = pd.concat([dsets[d] for d in dsets.keys()])\nprint(data_in.shape) \nprint(data_in.tail(2).T)\n\n# Initial pre Analysis showed that there are many duplicated rows.\ndata_in.drop_duplicates(inplace=True)\nprint(data_in.shape) \n\n# =============================================================================\n#%% md \n#### Explore Data \n\n#%%\n\ndata_in.info()\n\n#%%\n# Get Proportion of Null values on Columns\n\n#nulls_df = pd.DataFrame( {'Nulls': data_in.isnull().sum().values\n# , 'Ptj':data_in.isnull().sum() / data_in.shape[0]\n# }\n# )\n#print(nulls_df)\n\n\n# =============================================================================\n#%% md \n### Data Summary\n#Summary of the Row Data Set \n\n#%%\nfrom quickda.explore_data import *\nfrom quickda.clean_data import *\nfrom quickda.explore_numeric import *\nfrom quickda.explore_categoric import *\nfrom quickda.explore_numeric_categoric import *\nfrom quickda.explore_time_series import *\n\n\nsummary = explore(data_in, method=\"summarize\")\n#%%\n\nsummary # Display Raw Data Summary\n\n#summary\n\n#%% md \n### Assign Features on Categories\n\n\n#%%\n#%md Useful FtrsDrop Reason: They can not have missing values\n#valeur_fonciere (target cant have nulls)\n\n\n#%md High null pct ftrs\n\n#adresse_suffixe 0.958\n#ancien_code_commune 0.991\n#ancien_id_parcelle 0.999\n#ancien_nom_commune 0.991\n#code_nature_culture_speciale 0.954\n#code_type_local 0.476\n#lot1_numero 0.687\n#lot1_surface_carrez 0.912\n#lot2_numero 0.935\n#lot2_surface_carrez 0.979\n#lot3_numero 0.989\n#lot3_surface_carrez 0.998\n#lot4_numero 0.996\n#lot4_surface_carrez 0.999\n#lot5_numero 0.998\n#lot5_surface_carrez 1.000\n#nature_culture_speciale 0.954\n#numero_volume 0.997\n\n\n#%md Useful FtrsRedundant Features\n# dtypes count null_sum null_pct nunique\n#ancien_code_commune float64 7004 778519 0.991 532 \n#code_commune object 785523 0 0.000 31251 \n#code_departement object 785523 0 0.000 97 \n#code_nature_culture object 535959 249564 0.318 27 \n#code_nature_culture_speciale object 35761 749762 0.954 113 \n#code_type_local float64 411728 373795 0.476 4 \n\n\n#%md Useful Ftrs\n#surface_reelle_bati 118074 0.590370 NaN 0\n#nombre_pieces_principales 93279 0.466395 NaN 0\n#nature_culture 62293 0.311465 NaN \"N/A\"\n#code_type_local NAN \"N/A\"\n#longitude 4048 0.020240 k-neighb estim\n#latitude 4048 0.020240 k-neighb estim\n\n\n\n\n#%% md \n### Setting Unuseful featues for analysis Using above summary results\n\n\n\n#%% # Assigning Features to differet categorie types\n\nunuseful_ftrs = ['id_mutation', 'adresse_nom_voie', 'adresse_numero']\n\nuseful_high_ptc_ftrs = [ 'adresse_numero'\n , 'nombre_pieces_principales'\n , 'type_local'\n , 'surface_reelle_bati' \n , 'type_local'\n ]\n\n#Un inputable high nulls ftrs\nftrs_high_null_ptc = [f for f in summary.loc[summary.null_pct > 0.40,:].index if f not in useful_high_ptc_ftrs]\n#summary.loc[ftrs_high_null_ptc, 'null_pct']\n\nuseful_code_ftrs = [\n 'code_commune'\n , 'code_postal'\n ]\n\nredundant_ftrs = [c for c in summary.index if 'code' in c and c not in useful_code_ftrs]\nredundant_ftrs\n\n\nunuseful_ftrs += sorted(list(set(ftrs_high_null_ptc).union(set(redundant_ftrs))))\nunuseful_ftrs\n\n\nftrs = list(sorted([f for f in data_in.columns if f not in unuseful_ftrs])) \n\n\n\n#%%\n# Inspect left useful ftrs\nsmmry4 = ['dtypes', 'count', 'null_pct', 'nunique']\nsummary.loc[ftrs, smmry4]\n\n\n\n\n\n#%% md \n### Set useful Feature Groups\n\n\n#%%\n# Set useful Feature Groups\n\nftrs = list(sorted([f for f in data_in.columns if f not in unuseful_ftrs]))\ntarget = 'valeur_fonciere'\nftrs.remove(target)\n\nnum_ftrs = list(summary.loc[summary.index.isin(ftrs) & (summary['dtypes'] != 'object'),:].index)\nnum_ftrs = [f for f in num_ftrs if f not in ['code_postal']] # Remove wrogly assigned num ftrs and target\ndata_in.loc[:,'code_postal'] = data_in.loc[:,'code_postal'].astype('str')\n\nsummary.loc[num_ftrs,smmry4]\n\ndate_ftrs = ['date_mutation']\n\ncat_ftrs = list(summary.loc[summary.index.isin(ftrs)\n & (summary['dtypes'] == 'object')\n & (summary['nunique'] > 2 )\n ,:].index\n )\ncat_ftrs.append('code_postal')\nbool_ftrs = []\n\n# Assert all columns have bee assigned to a ftr class\nall_solumns_assigned = len(set(data_in.columns).symmetric_difference(\n set(unuseful_ftrs + num_ftrs + cat_ftrs + date_ftrs + bool_ftrs + [target])))==0\nassert(all_solumns_assigned)\n\n\ndata_in2 = data_in.loc[:, [target] + ftrs ]\n\n\nfor dt in [date_ftrs, num_ftrs, cat_ftrs]:\n print(summary.loc[dt, smmry4])\n\ndata_in2.tail(1).T\n\n\n#%% md \n##### Check Duplicates\n\n\n\n#%% md #\n#It seems like duplicates are sales with id_parcelle ad num disposition\n#and valeur_fonciere when they are same more than once \n\n\n#%%\n# Explore number of duplicated id parcells, This, inf fact should be a dropped field\nby_id_parcells = data_in.groupby('id_parcelle')\n#by_id_parcells.agg({'id_mutation': 'count'}).id_mutation.value_counts()#.plot.bar() #head()\nby_parcell_mutations = by_id_parcells.agg({'id_mutation': 'count'}).reset_index()\n# print(by_parcell_mutations.head())\n\nnuni_ue_mutations_parcels = by_parcell_mutations.loc[\n by_parcell_mutations.id_mutation>1,'id_parcelle'] #filter parcels wit 2 or more mutations\n\nnuni_ue = (data_in.loc[data_in.id_parcelle.isin(nuni_ue_mutations_parcels.values),:]\n .sort_values(by=['id_parcelle']\n )\n )\nnuni_ue.tail(4).T\n\n\n\n#%% \n\n# Drop duplicate properties\nid_ftrs = ['id_parcelle'\n , 'valeur_fonciere'\n ]\ndata_unique = data_in2.drop_duplicates(subset=id_ftrs, keep='last')\n\n\n#%% \ndata_unique.type_local.value_counts()\n\n#%% \n\n# Select only Sell operations\n\ndata = data_unique.loc[ (data_unique.nature_mutation == 'Vente')\n & (data_unique.type_local.isin(['Appartement', 'Maison']))\n ,:]\ndata.shape\n\n\n\n#%% \n# Update usable ftrs categories Nature mutation will be a not usable ftrs\ncat_ftrs.remove('nature_mutation')\nftrs.remove('nature_mutation')\nunuseful_ftrs.append('nature_mutation')\ncat_ftrs.remove('nom_commune')\nftrs.remove('nom_commune')\nunuseful_ftrs.append('nom_commune')\n\n\n#%%\nftrs\n#%%md \n### Explore Target Feature\n###### First Plotting target_value in thousands to facilitate visualization\n\n\n#%%md \n###### Original target values are very left skewed\n\n#%%\n###### Add sale value in thousands for easiest management\n###### Now plot log of target value\ndf_plot = data.sample(10000, random_state=SEED)\n\nfig, ax = plt.subplots(figsize=(12,8))\n(df_plot.valeur_fonciere/1000).plot.hist(ax = ax, bins = 10)\n#sns.histplot(df_plot.valeur_fonciere/1000, ax=ax, kde=False)\nplt.show()\n\n\n\n#%% md\n#Applying log transformations gives a very centered distribution, There is, however a small separated group of very low valuew which can, maybe be inspected to see if they are outliers.\n\n\n#%%\ndata.numero_disposition\n#%%\n# Now plot log of skewed ftrs\n\nfor f in ['valeur_fonciere', 'surface_reelle_bati', 'surface_terrain']:\n df_plot = data.sample(10000, random_state=SEED) \n fig, ax = plt.subplots(figsize=(12,8))\n sns.histplot(np.log1p(df_plot[f].values), ax=ax, kde=True)\n plt.show()\n\n\n\n#%%\n# We add the scaled target value to use\n\ndata['valeur_fonciere_log'] = np.log1p(data['valeur_fonciere'])\n\n# Now plot log of target value\ndf_plot = data.sample(10000, random_state=SEED)\n\nfig, ax = plt.subplots(figsize=(12,8))\nsns.histplot(np.log1p(df_plot.valeur_fonciere_log.values), ax=ax, kde=True)\nplt.show()\n\n#%% md\n#Potentian outlier vals are sell values under 90 and are approx 1% of data\n\n#%%\n\noutlier_thr = np.expm1(4.5)\nprint('Outlier threshold:', outlier_thr)\nprint( data.loc[data['valeur_fonciere_log']<4.5,:].shape\n , data.loc[data['valeur_fonciere'] 0:\n\t\t\t\t\tbreak\n\t\t\tcleaned_venue = ' '.join(cleaned_venue_tokens)\n\t\t\toutputfile.write('#c' + cleaned_venue + '\\n')\n\n\t\t\tyear = re.search(r'(?<=#t ).+', paper).group()\n\t\t\toutputfile.write('#t' + year + '\\n')\n\n\t\t\tabstract = re.search(r'(?<=#! ).+', paper).group()\n\t\t\tabstract = abstract.lower()\n\t\t\tabstract = abstract.translate(remove_digits)\n\t\t\tabstract_tokens = word_tokenize(abstract)\n\t\t\tcleaned_abs_list = []\n\t\t\tfor each in abstract_tokens:\n\t\t\t\tif each not in pw and each not in string.punctuation:\n\t\t\t\t\tcleaned_abs_list.append(each)\n\t\t\tcleaned_abstract = ' '.join(cleaned_abs_list)\n\t\t\toutputfile.write('#!' + cleaned_abstract + '\\n')\n\n\t\t\toutputfile.write('\\n')\n\n\t\toutputfile.close()","sub_path":"scripts/preprocess_topic.py","file_name":"preprocess_topic.py","file_ext":"py","file_size_in_byte":2746,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"565776019","text":"import sys\n\n#count the arguments\n#starts at 1\narguments = len(sys.argv) - 1\n\n\nclass Predict:\n def __init__(self, temp, rf, snow, time):\n self.temp = temp\n self.rf = rf\n self.snow = snow\n self.time = time\n\n def closed(self):\n if self.temp <= 31 and self.snow > 2 and self.time > 180:\n print(\"Snow day likely\")\n \n else:\n print(\"Not closing, potenital delayed opening\")\n\n\nprediction = Predict(int(sys.argv[1]), int(sys.argv[2]), int(sys.argv[3]), int(sys.argv[4]))\nprediction.closed()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"13414118","text":"\"\"\"\n MergeSort.py 归并排序\n\"\"\"\n\n\ndef merge_sort(_list):\n if len(_list) <= 1:\n return _list\n _mid = len(_list) // 2\n left_list = merge_sort(_list[:_mid])\n right_list = merge_sort(_list[_mid:])\n result = []\n if left_list and right_list:\n while len(left_list) > 0 and len(right_list) > 0:\n result.append(left_list.pop(0)) if left_list[0] <= right_list[0] else result.append(right_list.pop(0))\n result += left_list\n result += right_list\n return result\n\n\nif __name__ == '__main__':\n _l = [1, 2, 3, 4, 222, 3, 4, 6, 99]\n print(merge_sort(_l))\n","sub_path":"MiniModel/Sort/MergeSort.py","file_name":"MergeSort.py","file_ext":"py","file_size_in_byte":602,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"117067855","text":"# On an 8 x 8 chessboard, there is one white rook. There also may be empty squares, white bishops, and black pawns. These are given as characters 'R', '.', 'B', and 'p' respectively. Uppercase characters represent white pieces, and lowercase characters represent black pieces.\n\n# The rook moves as in the rules of Chess: it chooses one of four cardinal directions (north, east, west, and south), then moves in that direction until it chooses to stop, reaches the edge of the board, or captures an opposite colored pawn by moving to the same square it occupies. Also, rooks cannot move into the same square as other friendly bishops.\n\n# Return the number of pawns the rook can capture in one move.\n\n \n\n# Example 1:\n\n\n\n# Input: [[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"p\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"R\",\".\",\".\",\".\",\"p\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"p\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"]]\n# Output: 3\n# Explanation: \n# In this example the rook is able to capture all the pawns.\n# Example 2:\n\n\n\n# Input: [[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\"p\",\"p\",\"p\",\"p\",\"p\",\".\",\".\"],[\".\",\"p\",\"p\",\"B\",\"p\",\"p\",\".\",\".\"],[\".\",\"p\",\"B\",\"R\",\"B\",\"p\",\".\",\".\"],[\".\",\"p\",\"p\",\"B\",\"p\",\"p\",\".\",\".\"],[\".\",\"p\",\"p\",\"p\",\"p\",\"p\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"]]\n# Output: 0\n# Explanation: \n# Bishops are blocking the rook to capture any pawn.\n# Example 3:\n\n\n\n# Input: [[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"p\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"p\",\".\",\".\",\".\",\".\"],[\"p\",\"p\",\".\",\"R\",\".\",\"p\",\"B\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"B\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\"p\",\".\",\".\",\".\",\".\"],[\".\",\".\",\".\",\".\",\".\",\".\",\".\",\".\"]]\n# Output: 3\n# Explanation: \n# The rook can capture the pawns at positions b5, d6 and f5.\n \n\n# Note:\n\n# board.length == board[i].length == 8\n# board[i][j] is either 'R', '.', 'B', or 'p'\n# There is exactly one cell with board[i][j] == 'R'\nclass Solution:\n def numRookCaptures(self, board) -> int:\n for i in range(8):\n flag=False\n for j in range(8):\n if board[i][j] == \"R\":\n flag=True\n break\n if flag:\n break\n ni = i + 1\n ans = 0\n while ni < 8:\n if board[ni][j] == \".\":\n ni += 1\n elif board[ni][j] == \"B\":\n break\n else:\n ans += 1\n break\n ni = i - 1\n while ni >= 0:\n if board[ni][j] == \".\":\n ni -= 1\n elif board[ni][j] == \"B\":\n break\n else:\n ans += 1\n break\n nj = j + 1\n while nj < 8:\n if board[i][nj] == \".\":\n nj += 1\n elif board[i][nj] == \"B\":\n break\n else:\n ans += 1\n break\n nj = j - 1\n while nj >= 0:\n if board[i][nj] == \".\":\n nj -= 1\n elif board[i][nj] == \"B\":\n break\n else:\n ans += 1\n break\n return ans","sub_path":"999_Available_Captures_For_Rook.py","file_name":"999_Available_Captures_For_Rook.py","file_ext":"py","file_size_in_byte":3229,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"608865091","text":"import urllib.request\nimport string\nimport os\nfrom nltk.corpus import stopwords\nfrom nltk.sentiment.vader import SentimentIntensityAnalyzer\nimport math\n\nurl = 'http://www.gutenberg.org/cache/epub/28/pg28.txt'\nresponse = urllib.request.urlopen(url)\ndata = response.read() # a `bytes` object\ntext = data.decode('utf-8')\n# os.chdir(r\"c:\\Users\\gdevina1\\Documents\\GitHub\\text_mining\")\n# cwd = os.getcwd()\n# print(cwd)\n#print(text) # for testing\n\ndef process_online_text(text, skip_header):\n \"\"\"\n Makes a histogram that contains the words from a gutenberg online source file.\n\n filename: string\n skip_header: boolean, whether to skip the Gutenberg header\n \n returns: map from each word to the number of times it appears.\n \"\"\"\n hist = {}\n\n if skip_header:\n text = skip_gutenberg_start_end(text)\n\n text = text.split()\n # text = map(toLowerCase,text)\n\n for i in text:\n i = i.replace('-', ' ')\n strippables = string.punctuation + string.whitespace\n\n for word in i.split():\n # remove punctuation and convert to lowercase\n word = word.strip(strippables)\n word = word.lower()\n\n # update the histogram\n hist[word] = hist.get(word, 0) + 1\n\n return hist\n\ndef process_file(filename, skip_header):\n \"\"\"Makes a histogram that contains the words from a local file.\n\n filename: string\n skip_header: boolean, whether to skip the Gutenberg header\n\n returns: map from each word to the number of times it appears.\n \"\"\"\n hist = {}\n fp = open(filename, encoding='utf8')\n\n if skip_header:\n skip_gutenberg_header(fp)\n\n for line in fp:\n strippables = string.punctuation + string.whitespace\n\n for word in line.split():\n # remove punctuation and convert to lowercase\n word = word.strip(strippables)\n word = word.lower()\n\n # update the histogram\n hist[word] = hist.get(word, 0) + 1\n\n return hist\n\ndef skip_gutenberg_start_end(text):\n \"\"\"\n Reads from text until it finds the line that ends the header.\n\n text: open file object\n\n returns: remaining text starting from 'start' and ends with 'end'\n \"\"\"\n start = \"The Man, the Boy, and the Donkey The Fox and the Goat\"\n i = text.find(start)\n end = \"End of the Project Gutenberg EBook of Aesop's Fables, by Aesop\"\n x = text.find(end)\n\n return text[i+(len(start)):x]\n\ndef extract_fable(text,start,end):\n \"\"\"\n Reads from text until it finds the line that ends the header.\n\n text: open file object\n start: string in text\n end: string in text\n\n returns: remaining text starting from 'start' and ends with 'end'\n \"\"\"\n i = text.find(start)\n x = text.find(end)\n\n return text[i+(len(start)):x+(len(end))]\n\ndef compare(hist,animals):\n \"\"\"\n Finds matches of keys in hist and keys in animals, iterates while counting frequency of word appearance.\n\n hist = map from word to frequency\n animals = text file of list of animals\n\n returns = list of frequency,word pairs\n \"\"\"\n match = {}\n for key in animals:\n if key in hist:\n match[key] = hist.get(key, 0)\n return match\n\ndef no_stop_words(hist, key):\n \"\"\"\n Deletes stop words from dictionary\n\n hist = map from word to frequency\n key = elements to delete\n\n returns = list of frequency,word pairs\n \"\"\"\n hist2 = hist.copy()\n for i in key:\n if i in hist2:\n del hist2[i]\n return hist2\n\ndef sentiment_analysis(text):\n score = SentimentIntensityAnalyzer().polarity_scores(text)\n #return the polarity score of the text\n return score\n\ndef most_common(hist):\n \"\"\"Makes a list of word-freq pairs in descending order of frequency.\n\n hist: map from word to frequency\n\n returns: list of (frequency, word) pairs\n \"\"\"\n t = []\n # remove word from filtered_word_list if it is a stop word\n for key, value in hist.items():\n t.append((value, key))\n t.sort()\n t.reverse()\n return t\n\ndef most_common_all(hist):\n \"\"\"\n Prints a list of word-freq pairs in descending order of frequency.\n\n hist: map from word to frequency\n\n prints: list of (frequency, word) pairs\n \"\"\"\n t = []\n for key, value in hist.items():\n t.append((value, key))\n t.sort()\n t.reverse()\n\n for value, key in t:\n print(\"{:14}{}\".format(key, value))\n\n\ndef most_common_limited(hist, num=10):\n \"\"\"\n Prints the most commons words in a histogram and their frequencies.\n\n hist: histogram (map from word to frequency)\n num: number of words to print\n \"\"\"\n t = most_common(hist)\n for freq, word in t[:num]:\n print(\"{:14}{}\".format(word, freq))\n\ndef total_words(hist):\n \"\"\"\n Returns the total of the frequencies in a histogram.\n \"\"\"\n return sum(hist.values())\n\ndef numerator(fable_one, fable_two):\n \"\"\"\n Calculates the numerator of cosine similarity function using \n\n fable_one: histogram\n fable_two: histogram\n \n returns = number\n \"\"\"\n num = 0\n for word in fable_one:\n if word in fable_two:\n num += fable_one[word]*fable_two[word]\n return num\n\ndef cosine_similarity(fable_one, fable_two):\n \"\"\"\n Calculates cosine similarity of two texts (histograms)\n\n fable_one: histogram\n fable_two: histogram\n\n returns = number \n \"\"\"\n num = numerator(fable_one, fable_two)\n denom = math.sqrt(numerator(fable_one,fable_two)*numerator(fable_two,fable_two))\n return math.acos(num/denom)\n\ndef main():\n hist = process_online_text(text,True)\n animals = list(process_file('animals.txt', False).keys())\n animals.remove(animals[0])\n # print(animals)\n # print(compare(hist, animals))\n hist2 = hist.copy()\n # print(hist2)\n # print(no_stop_words(hist2))\n # most_common(hist)\n # print(no_stop_words(hist2,stop_words))\n stop_words = stopwords.words('english')\n print(\"\")\n print(\"The most common words in the text (without stop words) are:\")\n most_common_limited(no_stop_words(hist2,stop_words))\n print(\"\")\n print(\"------------------------------------------------------------\")\n print(\"\")\n print(\"List of all the animals appearing in Aesop's Fables\")\n most_common_all(compare(hist,animals))\n print(\"\")\n print(\"Total number of animals appearing in text: \", total_words(compare(hist,animals)))\n print(\"------------------------------------------------------------\")\n print(\"\")\n print(\"Total number of words in Aesop's Fables collection: \", total_words(hist))\n print(\"Percentage of total words in text that are animals: {0:.2f}%\" .format(total_words(compare(hist,animals))/total_words(hist) *100))\n print(\"\")\n print(\"------------------------------------------------------------\")\n tabcon= extract_fable(text,start=\"1-21 22-42\",end=\"The Fox and the Goat\")\n tableofcontents = process_online_text(tabcon,False)\n # print(tableofcontents)\n print(\"\")\n print(\"The 5 most common animals in the Aesop's Fables' titles:\")\n most_common_limited(compare(tableofcontents,animals),num=5)\n print(\"\")\n print(\"------------------------------------------------------------\")\n print(\"\")\n print(\"The 5 most common animals in the entire text:\")\n most_common_limited(compare(hist,animals),num=5)\n print(\"\")\n print(\"------------------------------------------------------------\")\n hare_and_tortoise = extract_fable(text, start=\"It is easy to propose impossible remedies.\", end=\"Plodding wins the race.\")\n hist_hare_and_tortoise = process_online_text(hare_and_tortoise,False)\n wolf_sheep_clothing = extract_fable(text, start=\"you cannot reckon.\", end=\"Appearances are deceptive.\")\n hist_wolf_sheep_clothing = process_online_text(wolf_sheep_clothing,False)\n fox_and_grapes = extract_fable(text,start=\"Nothing escapes the master's eye.\", end=\"It is easy to despise what you cannot get.\")\n hist_fox_and_grapes = process_online_text(fox_and_grapes,False)\n print(\"\")\n print(\"Sentiment analysis results of The Hare and The Tortoise:\")\n print(sentiment_analysis(hare_and_tortoise))\n print(\"\")\n print(\"Sentiment analysis results of The Wolf in Sheep's Clothing\")\n print(sentiment_analysis(wolf_sheep_clothing))\n print(\"\")\n print(\"Setiment analysis of The Fox and the Grapes\")\n print(sentiment_analysis(fox_and_grapes))\n print(\"\")\n print(\"------------------------------------------------------------\")\n print(\"\")\n print (\"The Hare and The Tortoise and The Wolf in Sheep's Clothing is {:.2f}% similar.\".format(cosine_similarity(hist_hare_and_tortoise,hist_wolf_sheep_clothing)/(math.pi/2)))\n print(\"\")\n print(\"The Hare and The Tortoise and The Fox and the Grapes is {:.2f}% similar.\".format(cosine_similarity(hist_hare_and_tortoise,hist_fox_and_grapes)/(math.pi/2)))\n print(\"\")\n print(\"The Fox and the Grapes and The Wolf in Sheep's Clothing is {:.2f}% similar.\".format(cosine_similarity(hist_wolf_sheep_clothing,hist_fox_and_grapes)/(math.pi/2)))\n print(\"\")\n print(\"------------------------------------------------------------\")\n print(\"END OF ANALYSIS\")\n print(\"------------------------------------------------------------\")\n \n\nif __name__ == '__main__':\n main()","sub_path":"assignment3.py","file_name":"assignment3.py","file_ext":"py","file_size_in_byte":9305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"221136914","text":"from array import array as arr\n\n# Bubble Sort\n#a= arr('i',[4,3,5,45,23,76,45,98,76,12,24,33])\n#a= arr('i',[1,2,3,4,5,6,7,8,9,10]) #Best case - zero swaps\na= arr('i',[10,9,8,7,6,5,4,3,2,1]) #worst case - n(n-1)/2 swaps\nprint(\"Bubble Sort\")\nprint(\"array before sorting :\")\nprint(a)\n\ndef swap(a,i,j):\n temp = a[i]\n a[i] = a[j]\n a[j] = temp\n\n\n#total_swaps = 0\n#total_comparisons =0\nfor i in range(0,len(a)):\n #swap = 0\n #comparison = 0\n for j in range(0,len(a)-i-1):\n #comparison+=1\n if(a[j]>a[j+1]):\n swap(a,j,j+1)\n #swap+=1\n #total_swaps = total_swaps + swap\n #total_comparisons = total_comparisons + comparison\n #print(f\"Pass {i} : No. of Comparions={comparison}, No of swaps={swap}\")\n\nprint(a)\n\n\n \n","sub_path":"Searching_Sorting/Bubble_Sort.py","file_name":"Bubble_Sort.py","file_ext":"py","file_size_in_byte":768,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"25397433","text":"#!/usr/bin/env python3\n\"\"\"\n1.创建文件名,如果存在,则换一个名字\n2.创建需要录入的内容\n3.写入文件\n\"\"\"\nimport os\n\ndef getfile():\n while True:\n fname=input(\"请输入文件名:\")\n if os.path.exists(fname):\n print(\"%s已经存在,请重试:\" % fname)\n else:\n break\n return fname\n\ndef getcontent():\n print(\"请输入正文:end表示结束\")\n content=[]\n while True:\n line=input(\">\")\n if line == \"end\":\n break\n content.append(line)\n return content\n\ndef wfile(fname,content):\n with open(fname,'w') as fobj:\n fobj.writelines(content) \n\nif __name__==\"__main__\":\n fname=getfile()\n content=getcontent()\n content=['%s\\n' % line for line in content]\n wfile(fname,content)\n\n\n\n\n","sub_path":"STEP05/project/lianxi/mktxt.py","file_name":"mktxt.py","file_ext":"py","file_size_in_byte":747,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"393724870","text":"from .base_pipeline_builder import BasePipelineBuilder\nfrom .pipeline import Pipeline\nfrom .pipeline_step import PipelineStep\n\n\nclass ExtractorPipelineBuilder(BasePipelineBuilder):\n \"\"\"\"\"\"\n\n def __init__(self):\n \"\"\"\"\"\"\n\n self.post_processors = []\n\n def set_mapping(self, v):\n \"\"\"\"\"\"\n\n self.mapping = v\n return self\n\n def set_extractor(self, v):\n \"\"\"\"\"\"\n\n self.extractor = v\n return self\n\n def add_post_processor(self, v, i=None):\n \"\"\"\"\"\"\n\n if i is None:\n self.post_processors.append(v)\n else:\n self.post_processors.insert(i, v)\n return self\n\n def build(self):\n \"\"\"\"\"\"\n\n self.extractor.mapping = self.mapping\n\n extractor_step = PipelineStep(\n self.extractor.extract\n )\n\n post_processor_steps = [\n PipelineStep(post_processor.process)\n for post_processor in self.post_processors\n ]\n\n steps = [\n extractor_step,\n *post_processor_steps\n ]\n\n return Pipeline(\n steps=steps\n )\n","sub_path":"src/magnesium/pipeline/extractor_pipeline_builder.py","file_name":"extractor_pipeline_builder.py","file_ext":"py","file_size_in_byte":1130,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"108316623","text":"import pandas as pd\nimport os\nimport json\n\n \ndef get_details(s):\n a = os.path.basename(s)\n a = a.split('-')\n if a[2].__contains__('.xlsx'):\n a[2] = a[2][:-5]\n return a[0], a[1], a[2] # 0: materia | 1: unidad \\ 2: categoria\n \ndef classify_dataframe(f):\n df = pd.read_excel(f)\n materia, unidad, categ = get_details(f)\n df['Categoria'] = pd.Series([categ]*len(df))\n df['Materia'] = pd.Series([materia]*len(df))\n df['Unidad'] = pd.Series([unidad]*len(df))\n return df\n\ndef get_json_pregs():\n dirname= 'Excels/'\n for root, dirs, files in os.walk(dirname):\n f = files\n \n names = [os.path.join(dirname, filename) for filename in f]\n dfs = [classify_dataframe(name) for name in names]\n \n df = pd.concat(dfs, axis=0)\n \n pr = []\n for i, row in df.iterrows():\n d = {}\n d.update({'Respuestas':row[1:5]})\n d.update({'Preguntas':row['Preguntas']})\n d.update({'Categoria':row['Categoria']})\n d.update({'Materia':row['Materia']})\n d.update({'Unidad':row['Unidad']})\n \n pr = pr + [d]\n \n return json.dumps(pr)\n","sub_path":"noticiados_app/noticiados_app/ex_to_json.py","file_name":"ex_to_json.py","file_ext":"py","file_size_in_byte":1137,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"361332683","text":"#!/bin/env python3\n\nimport arrow\nimport argparse\nimport io\nimport sys\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--start', help = 'Start date of WFH. ISO8601 date format')\nparser.add_argument('--end', help = 'End date of WFH. ISO8601 date format')\nparser.add_argument('--office', help = 'File containing ISO-8601 dates when present at the office')\n\nargs = parser.parse_args()\n\nstart = arrow.get(args.start)\nend = arrow.get(args.end)\nofficeFile = args.office\n\nholidays = []\noffice = []\n\nif officeFile is not None:\n with open(officeFile, 'r') as at_office:\n file_content = at_office.readlines()\n\n for office_date in file_content:\n office.append(office_date.strip())\n\nwith open('holidays.txt', 'r') as publicHolidays:\n file_content = publicHolidays.readlines()\n\n for current_line in file_content:\n holidays.append(current_line.strip())\n\nwith open('wfh.csv', 'w') as schedule:\n schedule.write('Date, Day, Location\\n')\n\n for current_date in arrow.Arrow.span_range('day', start, end):\n date = current_date[0]\n\n if date.isoweekday() in range(1,6):\n str_date = date.format(\"YYYY-MM-DD\")\n\n if str_date in holidays:\n schedule.write(str_date + \",\" + date.format(\"dddd\") + \",Public Holiday\\n\")\n elif str_date in office:\n schedule.write(str_date + \",\" + date.format(\"dddd\") + \",Working at employers office\\n\")\n else:\n schedule.write(str_date + \",\" + date.format(\"dddd\") + \",Working from home\\n\")\n\n","sub_path":"log.py","file_name":"log.py","file_ext":"py","file_size_in_byte":1548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"567737908","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Sep 15 20:17:53 2018\n\n不定數迴圈, break at -9999\n\"\"\"\n\nwhile True:\n y = eval(input(\"Enter a year number: \"))\n if y != -9999:\n if y % 400 == 0 or (y % 4 == 0 and y % 100 != 0):\n print('%d is a leap year.' % (y))\n else:\n print('%d is not a leap year.' % (y))\n else:\n break\n\n\"\"\"\ndef isLeap(y):\n if y % 400 == 0 or (y % 4 == 0 and y % 100 != 0):\n print('%d is a leap year.' % (y))\n else:\n print('%d is not a leap year.' % (y))\n \ndef main():\n while True:\n ye = eval(input(\"Enter a year number: \"))\n if ye != -9999:\n isLeap(ye)\n else:\n break\n \nmain()\n\"\"\" ","sub_path":"textbook/sample-code_s1_s2/s1_loop_3_isLeap.py","file_name":"s1_loop_3_isLeap.py","file_ext":"py","file_size_in_byte":730,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"518828321","text":"\"\"\"\nThis is a simple example of a client program written in Python.\nAgain, this is a very basic example to complement the 'basic_server.c' example.\n\n\nWhen testing, start by initiating a connection with the server by sending the \"init\" message outlined in \nthe specification document. Then, wait for the server to send you a message saying the game has begun. \n\nOnce this message has been read, plan out a couple of turns on paper and hard-code these messages to\nand from the server (i.e. play a few rounds of the 'dice game' where you know what the right and wrong \ndice rolls are). You will be able to edit this trivially later on; it is often easier to debug the code\nif you know exactly what your expected values are. \n\nFrom this, you should be able to bootstrap message-parsing to and from the server whilst making it easy to debug.\nThen, start to add functions in the server code that actually 'run' the game in the background. \n\"\"\"\n\nimport socket\nfrom time import sleep\n# Create a TCP/IP socket\nsock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\n# Connect the socket to the port where the server is listening\nserver_address = ('localhost', 4444)\nprint ('connecting to %s port %s' % server_address)\nsock.connect(server_address)\ncount=0\nsleep(30)\nmessage = 'INIT'.encode()\nsock.sendall(message)\nmoves = [\",MOV,EVEN\",\",MOV,EVEN\",\",MOV,CON,1\",\",MOV,ODD\",\",MOV,CON,1\",\",MOV,ODD\",\",MOV,CON,1\",\",MOV,ODD\"]\ni = 0\ntry:\n while True:\n \n exit = False\n # Look for the response\n amount_received = 0\n amount_expected = len(message)\n \n while amount_received < amount_expected:\n data = sock.recv(1024)\n amount_received += len(data)\n mess = data.decode()\n if \"WELCOME\" in mess:\n id = mess.split(\",\")[1]\n if \"START\" in mess:\n print(\"The games have begun\")\n print ( 'received \"%s\"' % mess)\n sock.sendall((str(id) + moves[i]).encode()) # Client has ID 231\n i = i +1\n elif \"ELIM\" in mess:\n print ( 'received \"%s\"' % mess)\n print(\"We lost, closing connection\")\n exit = True\n break\n elif \"PASS\" in mess or \"FAIL\" in mess:\n print ( 'received \"%s\"' % mess)\n sock.sendall((str(id) + moves[i]).encode())\n i = i+1\n print('Sending con message')\n else:\n print ( 'received \"%s\"' % mess)\n if exit:\n break\nfinally: \n print ('closing socket')\n sock.close()\n","sub_path":"socket_client.py","file_name":"socket_client.py","file_ext":"py","file_size_in_byte":2608,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"584874700","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nProblem 1:Python code for solving initial value problem using backward integration with Euler's method\nCreated on Wed Apr 1 13:28:50 2020\n\n@author: krishnendu\n\"\"\"\n\nimport numpy as np\nfrom scipy.optimize import *\nimport matplotlib.pyplot as plt\nfrom scipy.integrate import *\n#Problem 1: \ndef fun1(y): #defining the function for derivative \n return -9*y\n\ndef func1(x,y): #defining the function for doing the implicit Euler's method\n return x-y-h1*fun1(x)\n\na1=0 #starting point \nb1=1 #ending point\nN1=100 #number of mesh points\nx1=np.linspace(a1,b1,N1) #creating mesh point\nh1=(b1-a1)/(N1-1) #step size\nw1=np.zeros(N1)\nf1=np.zeros(N1)\nw1[0]=np.exp(1) #putting the initial condition\nf1[0]=np.exp(1 ) #putting the initial condition\nfor i in range(N1-1): #iteration \n f1[i+1]=f1[i]+h1*fun1(w1[i]) #finding the guess value for implicit Euler's method using explicit Euler's method\n \n w1[i+1]=newton(func1,f1[i+1],args=(w1[i],)) #finding the value of y\n\nplt.plot(x1,w1,label=\"Implicit Euler's method\") \nplt.xlabel(\"x\",size=18)\nplt.ylabel(\"y\",size=18)\nplt.title(\"Problem 1-(a)\",size=18)\nplt.legend()\nplt.grid()\nplt.show()\n\n#Problem 2: #defining the function for derivative \n\na2=0 #starting point \nb2=1 #ending point\nN2=100 #number of mesh points\nx2=np.linspace(a2,b2,N2) #creating mesh point\nh2=(b2-a2)/(N2-1) #step size\n\ndef fun2(y,x): #defining the function for derivative \n return -20*(y-x)**2+2*x\n\ndef func2(x,y,z): #defining the function for doing the implicit Euler's method\n return x-y-h2*fun2(x,z)\n\nw2=np.zeros(N2)\nf2=np.zeros(N2)\nw2[0]=1/3 #putting the initial value\nf2[0]=1/3 #putting the initial value\nfor i in range(N2-1): #iteration\n f2[i+1]=f2[i]+h2*fun2(w2[i],x2[i]) #finding the guess value for implicit Euler's method using explicit Euler's method\n \n w2[i+1]=newton(func2,f2[i+1],args=(w2[i],x2[i+1],)) #finding value of y\n \nplt.plot(x2,w2,label=\"Implicit Euler's method\")\nplt.xlabel(\"x\",size=18)\nplt.ylabel(\"y\",size=18)\nplt.title(\"Problem 1-(b)\",size=18)\nplt.legend()\nplt.grid()\nplt.show()","sub_path":"problem_1_assign_2.py","file_name":"problem_1_assign_2.py","file_ext":"py","file_size_in_byte":2490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"213345914","text":"\"\"\"\n# Definition for a Node.\nclass Node(object):\n def __init__(self, val, prev, next, child):\n self.val = val\n self.prev = prev\n self.next = next\n self.child = child\n\"\"\"\n\nclass Solution(object):\n def flatten(self, head):\n \"\"\"\n :type head: Node\n :rtype: Node\n \"\"\"\n if head == None or (head.next == None and head.child == None):\n return head\n stack = []\n node = head\n tail = None\n while node != None:\n if node.child != None:\n tmp = node.child\n node.child = None\n if node.next != None:\n stack.append(node.next)\n tmp.prev = node\n node.next = tmp\n node = tmp\n else:\n tail = node\n node = node.next\n while len(stack) > 0:\n top = stack.pop()\n tail.next = top\n top.prev = tail\n tail = top\n while top.next != None:\n top = top.next\n tail = top\n return head\n","sub_path":"dataStructure/LinkedList/Flatten.py","file_name":"Flatten.py","file_ext":"py","file_size_in_byte":930,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"224574278","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\nimport tools.fields\nimport tools.utils\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('frame', '0006_auto_20180524_0937'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='InterfaceLog',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('operating_time', models.DateTimeField(auto_now_add=True, verbose_name='\\u64cd\\u4f5c\\u65f6\\u95f4')),\n ('operator_address', models.GenericIPAddressField(null=True, verbose_name='\\u64cd\\u4f5c\\u7aef\\u5730\\u5740', blank=True)),\n ('operator_type', models.CharField(max_length=500, null=True, verbose_name='\\u64cd\\u4f5c\\u7c7b\\u578b', blank=True)),\n ('operation_object', models.CharField(max_length=500, null=True, verbose_name='\\u64cd\\u4f5c\\u5bf9\\u8c61', blank=True)),\n ('operating_results', models.CharField(max_length=200, null=True, verbose_name='\\u64cd\\u4f5c\\u7ed3\\u679c', blank=True)),\n ('detail_info', models.TextField(max_length=500, null=True, verbose_name='\\u8be6\\u7ec6\\u4fe1\\u606f', blank=True)),\n ('app', models.CharField(max_length=20, verbose_name='\\u5173\\u8054\\u5e94\\u7528')),\n ],\n options={\n 'verbose_name': '\\u63a5\\u53e3\\u65e5\\u5fd7',\n 'verbose_name_plural': '\\u63a5\\u53e3\\u65e5\\u5fd7',\n },\n ),\n migrations.CreateModel(\n name='SessionTicket',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('session_key', tools.fields.FixEncryptedCharField(max_length=255)),\n ('ticket', tools.fields.FixEncryptedCharField(max_length=255)),\n ],\n ),\n migrations.CreateModel(\n name='UserLog',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('operating_time', models.DateTimeField(auto_now_add=True, verbose_name='\\u64cd\\u4f5c\\u65f6\\u95f4')),\n ('operator_address', models.GenericIPAddressField(null=True, verbose_name='\\u64cd\\u4f5c\\u7aef\\u5730\\u5740', blank=True)),\n ('operator_type', models.CharField(max_length=500, null=True, verbose_name='\\u64cd\\u4f5c\\u7c7b\\u578b', blank=True)),\n ('operation_object', models.CharField(max_length=500, null=True, verbose_name='\\u64cd\\u4f5c\\u5bf9\\u8c61', blank=True)),\n ('operating_results', models.CharField(max_length=200, null=True, verbose_name='\\u64cd\\u4f5c\\u7ed3\\u679c', blank=True)),\n ('detail_info', models.TextField(max_length=500, null=True, verbose_name='\\u8be6\\u7ec6\\u4fe1\\u606f', blank=True)),\n ('user', models.CharField(max_length=20, verbose_name='\\u7528\\u6237')),\n ],\n options={\n 'verbose_name': '\\u7528\\u6237\\u65e5\\u5fd7',\n 'verbose_name_plural': '\\u7528\\u6237\\u65e5\\u5fd7',\n },\n ),\n migrations.RemoveField(\n model_name='passwordcodes',\n name='user',\n ),\n migrations.AddField(\n model_name='userprofile',\n name='hw_email',\n field=tools.fields.EncryptedEmailField(max_length=255, null=True, verbose_name='\\u534e\\u4e3a\\u90ae\\u7bb1', blank=True),\n ),\n migrations.AlterField(\n model_name='app',\n name='secret',\n field=models.CharField(help_text='Required. 16 ~ 128 characters. Letters, digits and special character only.', max_length=200, verbose_name=b'Secret'),\n ),\n migrations.AlterField(\n model_name='apptoken',\n name='key',\n field=tools.fields.EncryptedCharField(max_length=255, verbose_name='Key', blank=True),\n ),\n migrations.AlterField(\n model_name='useraccesscodes',\n name='codes',\n field=models.CharField(max_length=128, verbose_name='\\u9a8c\\u8bc1\\u7801'),\n ),\n migrations.AlterField(\n model_name='usercodes',\n name='codes',\n field=models.CharField(max_length=128, verbose_name='\\u9a8c\\u8bc1\\u7801'),\n ),\n migrations.AlterField(\n model_name='userprofile',\n name='is_active_type',\n field=models.IntegerField(default=0, null=True, verbose_name='\\u6ce8\\u9500\\u7c7b\\u578b', blank=True, choices=[(0, '\\u8d26\\u53f7\\u6709\\u6548'), (1, '\\u8d26\\u53f7\\u5230\\u671f'), (2, '\\u8bc1\\u4e66\\u65e0\\u6548'), (3, '\\u90e8\\u95e8\\u6539\\u53d8'), (4, '\\u6ca1\\u6709\\u8d26\\u53f7'), (5, '\\u79bb\\u804c'), (6, '\\u957f\\u671f\\u672a\\u767b\\u5f55'), (7, '\\u5de5\\u53f7\\u53d8\\u66f4')]),\n ),\n migrations.AlterField(\n model_name='userprofile',\n name='phone_number',\n field=tools.fields.EncryptedCharField(blank=True, max_length=255, null=True, verbose_name='\\u624b\\u673a\\u53f7\\u7801', validators=[tools.utils.check_phone]),\n ),\n migrations.AlterField(\n model_name='userprofile',\n name='safe_email',\n field=tools.fields.EncryptedEmailField(max_length=255, null=True, verbose_name='\\u5b89\\u5168\\u90ae\\u7bb1', blank=True),\n ),\n ]\n","sub_path":"WiseEyeIAMService/frame/migrations/0007_auto_20180704_1654.py","file_name":"0007_auto_20180704_1654.py","file_ext":"py","file_size_in_byte":5449,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"310197869","text":"'''\nThe loader is created by the main SaltEventsDaemon. It takes care of loading\nthe given config with a default of '/etc/salt/eventsd'. It also takes care of\ninitializing the logger from saltstack.\n'''\n\nimport sys\nimport logging\nimport salt.log\nimport os\nimport yaml\n\nlogger = salt.log.setup.logging.getLogger(__name__)\nlog = logging.getLogger(__name__)\n\n\nclass SaltEventsdLoader(object):\n '''\n The loader takes care of reading the configfile and\n setting up the correct logger.\n '''\n def __init__(self, config=None, log_level=None, log_file=None, daemonize=False):\n self.config_file = config if config else \"/etc/salt/eventsd\"\n # retrieve current settings from the config file\n self.opts = None\n self._read_yaml(self.config_file)\n\n # make sure we have a 'general' section\n if 'general' in self.opts.keys():\n self.gen_opts = self.opts['general']\n\n # Use log level if explicitly set from cli\n if log_level:\n self.gen_opts['loglevel'] = log_level\n\n # Use log file if explicitly set from cli\n if log_file:\n self.gen_opts['logfile'] = log_file\n\n self.gen_opts['daemonize'] = daemonize\n\n self._init_logger()\n log.info(\"loaded config from {0}\".format(config))\n\n def _init_logger(self):\n '''\n sets up the logger used throughout saltt-eventsd\n '''\n # make sure we have the required settings for our logging\n if ('logfile' in self.gen_opts) and \\\n ('loglevel' in self.gen_opts):\n\n salt.log.setup_logfile_logger(\n self.gen_opts['logfile'],\n self.gen_opts['loglevel'],\n )\n\n # Only log to foreground if not running as a daemon\n if not self.gen_opts['daemonize']:\n salt.log.setup_console_logger(\n log_level=self.gen_opts['loglevel'],\n )\n else:\n # if no log settings found, use defaults\n\n # Only log to foreground if not running as a daemon\n if not self.gen_opts['daemonize']:\n salt.log.setup_console_logger(\n log_level=\"warn\"\n )\n\n salt.log.setup_logfile_logger(\n '/var/log/salt/eventsd',\n 'warn',\n )\n\n def getopts(self):\n '''\n returns the parsed options to the SaltEventsDaemon-Class\n '''\n return self.opts\n\n def _read_yaml(self, path):\n '''\n reads a yaml-formatted configuration file at the given path and\n returns a python dictionary with the pared items in it.\n '''\n try:\n yaml_handle = open(path)\n self.opts = yaml.load(yaml_handle.read())\n except yaml.parser.ParserError as yamlerr:\n print(\"Failed to parse configfile: {0}\".format(path))\n print(yamlerr)\n sys.exit(1)\n except yaml.scanner.ScannerError as yamlerr:\n print(\"Failed to parse configfile: {0}\".format(path))\n print(yamlerr)\n sys.exit(1)\n except IOError as ioerr:\n print(\"Failed to read configfile:\")\n print(os.strerror(ioerr.errno))\n sys.exit(1)\n except OSError as oserr:\n print(\"Failed to read configfile:\")\n print(os.strerror(oserr.errno))\n sys.exit(1)\n","sub_path":"salteventsd/loader.py","file_name":"loader.py","file_ext":"py","file_size_in_byte":3410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"25485943","text":"import numpy as np\nimport itertools\nfrom sklearn.model_selection import train_test_split\n\ndef create_inp_opt(data):\n inp_opt = map(lambda x:x.split('='), data)\n inp_opt_chars = list(map(lambda x: [list(x[0]), list(x[1])], inp_opt))\n \n return inp_opt_chars\n\ndef create_vocab(data):\n inp_opt_merged = list(map(lambda x: list(itertools.chain(*x)), data))\n vocab = set(itertools.chain(*inp_opt_merged))\n vocab = vocab.union({'start', 'end'})\n \n return vocab\n\ndef create_mapping(vocab):\n mapping = {}\n for i in range(len(vocab)):\n mapping[vocab[i]] = i+1\n \n mapping['pad'] = 0\n \n return mapping\n\ndef inp_opt(inp_opt_chars, mapping):\n inp_opt_mapped = list(map(lambda x:[[mapping['start']]+[mapping[i] for i in x[0]]+[mapping['end']], \n [mapping['start']]+[mapping[i] for i in x[1]]+[mapping['end']]], inp_opt_chars))\n \n inp = list(map(lambda x:x[0], inp_opt_mapped))\n opt = list(map(lambda x:x[1], inp_opt_mapped))\n \n return inp, opt\n\nfile = open('dataset.txt','r')\ndata = file.read()\n\ndata = data.split('\\n')[:-1]\n\ntrain, test = train_test_split(data, test_size=0.10, random_state=1)\ntrain, val = train_test_split(train, test_size=0.10, random_state=1)\n\nwith open('train.txt', 'w') as filehandle:\n filehandle.writelines(\"%s\\n\" % exp for exp in train)\n \nwith open('valid.txt', 'w') as filehandle:\n filehandle.writelines(\"%s\\n\" % exp for exp in val)\n \nwith open('test.txt', 'w') as filehandle:\n filehandle.writelines(\"%s\\n\" % exp for exp in test)\n","sub_path":"transformer/dataprocessing.py","file_name":"dataprocessing.py","file_ext":"py","file_size_in_byte":1559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"215358186","text":"from django.conf.urls.defaults import *\n\n# JC: Add urls.py branch to deal with JSON / DATA / API requests. Clean things up.\n# Notes:\n# Allow for endings in / or \"\" where possible\n#\nurlpatterns = patterns('v1',\n (r'api/(?Ptopic)/(?P\\d+)/(?Pstories)/', include('json.gettopic.urls')),\n (r'api/(?Ptopic)/(?P\\d+)/(?Psubtopics)/', include('json.gettopic.urls')),\n (r'api/(?Ptopic)/(?P\\d+)', include('json.gettopic.urls')),\n (r'api/(?Pstory)/', include('json.getuser.urls')),\n (r'api/(?Ptopic)/', include('topic.urls')),\n (r'location$', include('json.location.urls')),\n (r'location/$', include('json.location.urls'))\n)\n","sub_path":"fishbone/urls-api.py","file_name":"urls-api.py","file_ext":"py","file_size_in_byte":724,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"170238353","text":"import numpy as np\nimport sqlite3\n\ndef _insert(db, sql):\n conn = sqlite3.connect(db)\n curs = conn.cursor()\n if isinstance(sql, list):\n for s in sql:\n curs.execute(s)\n else:\n curs.execute(sql)\n conn.commit()\n conn.close()\n\ndef insertSQL(db, name, price, quantity, chain=False):\n pid = np.random.randint(111111,1000000)\n sql = f'''INSERT INTO inventory (pid,name,price,quantity)\n VALUES ({pid},'{name}',{price},{quantity})'''\n if chain:\n return sql\n _insert(db, sql)\n \ndef mass_insert_test(db):\n names = ['hammer','nails','screws','glass','wood','metal',\n 'canvas','pins','glue','paint','pens','bags']\n prices = [5.99,0.50,0.75,11.99,3.75,5.25,\n 1.99,0.25,2.25,6.75,1.25,0.75]\n quants = np.random.randint(200, 500, len(names))\n sql = []\n for i in range(len(names)):\n sql.append((insertSQL(db, names[i], prices[i], quants[i], chain=True)))\n _insert(db, sql)\n","sub_path":"insertModule.py","file_name":"insertModule.py","file_ext":"py","file_size_in_byte":983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"412354484","text":"\"\"\"\nWrite a recursive function called palindrome which will receive a word and an index (always 0). Implement the function,\nso it returns \"{word} is a palindrome\" if the word is a palindrome and \"{word} is not a palindrome\" if the word is not\na palindrome using recursion.\n\n\"\"\"\n\n\ndef palindrome(word, index):\n stop = len(word) // 2\n if index == stop:\n return f'{word} is a palindrome'\n if word[index] == word[len(word)-1-index]:\n return palindrome(word, index+1)\n else:\n return f'{word} is not a palindrome'\n\nprint(palindrome(\"abcba\", 0))\nprint(palindrome(\"peter\", 0))","sub_path":"Functions Advanced/recursion_palindrome.py","file_name":"recursion_palindrome.py","file_ext":"py","file_size_in_byte":601,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"295514502","text":"# Image Scraper\r\n\r\nfrom bs4 import BeautifulSoup\r\nimport requests\r\nfrom PIL import Image\r\nfrom io import BytesIO\r\nimport os\r\n\r\n\r\ndef StartSearch():\r\n search = input(\"Search for:\")\r\n params = {\"q\": search}\r\n # Create a directory name by using the search query, and replacing any spaces with underscores\r\n dir_name = search.replace(\" \", \"_\").lower()\r\n\r\n # .path(): This is a function from the OS library that checks the current path of the program\r\n # isdir(): This is a function from the OS library that checks to see if this is a valid directory name\r\n if not os.path.isdir(dir_name):\r\n # .makedirs(): This is a function from the OS library that creates a directory\r\n # Create the directory:\r\n os.makedirs(dir_name)\r\n\r\n r = requests.get(\"http://www.bing.com/images/search\", params=params)\r\n\r\n soup = BeautifulSoup(r.text, \"html.parser\")\r\n\r\n # Create a list of \"thumb\" class links, obviously for thumbnail images\r\n links = soup.findAll(\"a\", {\"class\": \"thumb\"})\r\n\r\n for item in links:\r\n try:\r\n # Use the requests library, and the .get() function to find each of the href locations of all the links\r\n img_obj = requests.get(item.attrs[\"href\"])\r\n print(\"Getting:\", item.attrs[\"href\"])\r\n # Use the .split() function to split the href location\r\n # [-1]: This is a Python way to index into the last element of a list\r\n title = item.attrs[\"href\"].split(\"/\")[-1]\r\n try:\r\n img = Image.open(BytesIO(img_obj.content))\r\n # Create a new directory called \"scraped_images\" in the actual Python project\r\n # .save(): This is a function from the Image library which needs the directory of the image file on your computer\r\n img.save(\"./\" + dir_name + \"/\" + title, img.format)\r\n except:\r\n print(\"Could not save image\")\r\n except:\r\n print(\"Could not request image\")\r\n # This forces the function to call itself again so it loops\r\n StartSearch()\r\n\r\n\r\n# Call the function:\r\nStartSearch()","sub_path":"python/01PythonStackskillsCourse/06WebScraper/Image Scraper/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2120,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"45832812","text":"#\n# 프로그램 이름: association_basics.py\n# 작성자: Bong Ju Kang\n# 설명: 연관 분석 이해하기\n#\n\n# 필요한 패키지\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\nfrom matplotlib import cm\nfrom numpy.random import RandomState\n\nfrom mlxtend.preprocessing import TransactionEncoder\nfrom mlxtend.frequent_patterns import apriori, association_rules\n\nfrom sklearn.cluster import KMeans\nfrom sklearn.datasets.samples_generator import make_blobs\nfrom sklearn.metrics import silhouette_score, silhouette_samples\n\n# 초기 설정\npng_path = \"./data/png\"\nos.makedirs(png_path, exist_ok=True)\n\n# 한글출력\nplt.rcParams['font.family'] = 'Malgun Gothic'\nplt.rcParams['axes.unicode_minus'] = False\n\n#\n# 본문 예시 예제\n#\n\n# 데이터 생성\ntr_data = [['milk', 'bread'],\n ['butter'],\n ['beer', 'diapers'],\n ['milk', 'bread', 'butter'],\n ['bread']]\n\n# 데이터 전 처리\ntr_encoder = TransactionEncoder()\ntr_encoder.fit(tr_data)\ntr_encoder.columns_\n# ['beer', 'bread', 'butter', 'diapers', 'milk']\n\ntr_encoder_ary = tr_encoder.transform(tr_data)\ntr_encoder_ary = np.where(tr_encoder_ary==True, 1, 0)\ndf = pd.DataFrame(tr_encoder_ary, columns=tr_encoder.columns_ )\nprint(df)\n# beer bread butter diapers milk\n# 0 0 1 0 0 1\n# 1 0 0 1 0 0\n# 2 1 0 0 1 0\n# 3 0 1 1 0 1\n# 4 0 1 0 0 0\n\n# 연관 모델 적합\nfreq_items = apriori(df, min_support=0.4, use_colnames=True, n_jobs=-1)\n\nfreq_items['length'] = freq_items['itemsets'].apply(lambda x: len(x))\n\n# itemsets 컬럼은 frozenset 속성을 갖고 있다. 즉, 변경이 불가능한 집합으로 구성되어 있다.\ntype(freq_items['itemsets'][0])\n# frozenset\n\n# 아래의 2개의 결과는 같다.\nfreq_items[freq_items['itemsets'] == {'bread', 'milk'}]\nfreq_items[freq_items['itemsets'] == {'milk', 'bread'}]\n\n#\n# 선험 알고리즘 예시\n#\n# 데이터 구성\ntr_data = [[1, 2, 3, 4],\n [1, 2, 4],\n [1, 2],\n [2, 3, 4],\n [2, 3],\n [3, 4],\n [2, 4]]\n\n# 데이터 전 처리\ntr_encoder = TransactionEncoder()\ntr_encoder.fit(tr_data)\ntr_encoder.columns_\n\ntr_encoder_ary = tr_encoder.transform(tr_data)\ntr_encoder_ary = np.where(tr_encoder_ary==True, 1, 0)\ndf = pd.DataFrame(tr_encoder_ary, columns=tr_encoder.columns_)\nprint(df)\n\n# 연관 모델 적합\n# 길이가 1개짜리 전부\nfreq_items = apriori(df, min_support=0, max_len=1, use_colnames=True)\nprint(freq_items)\n\n# 지지도가 3/7= 0.43 이상인 경우만\nfreq_items = apriori(df, min_support=3/7, max_len=1, use_colnames=True)\nprint(freq_items)\n\n# 길이가 2개짜리 전부\nfreq_items = apriori(df, min_support=0, max_len=2, use_colnames=True)\nfreq_items['length'] = freq_items['itemsets'].apply(lambda x: len(x))\n\n# 길이가 2개이면서, 지지도 조건을 충족하는 경우\nprint(freq_items[(freq_items['length'] == 2) & (freq_items['support'] >= 3.0/7.0)])\n# support itemsets length\n# 4 0.428571 (1, 2) 2\n# 7 0.428571 (2, 3) 2\n# 8 0.571429 (2, 4) 2\n# 9 0.428571 (3, 4) 2\n\n# 길이가 3개짜리 전부\nfreq_items = apriori(df, min_support=0, max_len=3, use_colnames=True)\nfreq_items['length'] = freq_items['itemsets'].apply(lambda x: len(x))\nprint(freq_items[freq_items['length'] == 3])\n# support itemsets length\n# 10 0.142857 (1, 2, 3) 3\n# 11 0.285714 (1, 2, 4) 3\n# 12 0.142857 (1, 3, 4) 3\n# 13 0.285714 (2, 3, 4) 3\n# 길이가 3이면서 지지도 조건을 충족하는 경우\nprint(freq_items[(freq_items['length'] == 3) & (freq_items['support'] >= 3.0/7.0)])\n# Empty DataFrame\n# Columns: [support, itemsets, length]\n# Index: []\n# 길이가 3인 경우는 고빈도 집합이 없으므로 길이가 2인 경우로 마무리 된다.\n\n#\n# 규칙 생성\n#\n\n# 규칙을 생성하기 위한 고빈도 집합 생성\nfreq_items = apriori(df, min_support=3/7, max_len=3, use_colnames=True)\n\n# 고빈도 집합 기반으로 규칙 생성: support, confidence, lift 기준 적용 가능\nasso_rules = association_rules(freq_items, metric='confidence', min_threshold=1)\nprint(asso_rules.columns)\n# ['antecedents', 'consequents', 'antecedent support',\n# 'consequent support', 'support', 'confidence', 'lift', 'leverage',\n# 'conviction']\n\n# 결과 확인\nasso_rules[['antecedents', 'consequents', 'support', 'confidence', 'lift']]\n\n#\n# 예제: [RETAIL] 데이터를 이용한 예제\n#\n\n#\n# 데이터 구성\n#\n# 데이터 불러오기\nretail = pd.read_excel('https://archive.ics.uci.edu/ml/machine-learning-databases/00352/Online%20Retail.xlsx')\n# retail.to_csv(data_path+'/retail.csv')\nretail.columns\n# Index(['InvoiceNo', 'StockCode', 'Description', 'Quantity', 'InvoiceDate',\n# 'UnitPrice', 'CustomerID', 'Country'],\n# dtype='object')\n\n# 소문자로 변수명를 사용하기 위하여\nnew_columns = retail.columns.str.lower()\nretail.columns = new_columns\n\n# 데이터 탐색\nretail['stockcode'].value_counts()\n# Name: stockcode, Length: 4070, dtype: int64\nretail['country'].value_counts()[:5]\n# United Kingdom 495478\n# Germany 9495\n# France 8557\n# EIRE 8196\n# Spain 2533\n# Name: country, dtype: int64\n\n#\n# 일부 데이터만 구성: Germany에서만 팔린 데이터로 구성\n#\ndf = retail[retail.country=='Germany']\ndf.stockcode.value_counts()\n# Name: stockcode, Length: 1671, dtype: int64\n\n# 데이터 전 처리: 거래 데이터 구성\n\n# 결측값 확인\ndf.isna().sum()\n\n# 거래 데이터 구성\ntrxs = df.groupby(['invoiceno', 'description'])['quantity'].sum().unstack(fill_value=0).reset_index().set_index('invoiceno')\ntrxs_df = pd.DataFrame(np.where(trxs > 1, 1, 0), columns=trxs.columns, index=trxs.index)\n\n# 고빈도 집합 찾기\nsupp_cutoff = 0.05 # 5%\npd.set_option('display.max_columns', 15)\npd.set_option('display.width', 500)\n\nfreq_items = apriori(trxs_df, min_support=supp_cutoff, use_colnames=True)\nfreq_items.sort_values(by='support', ascending=False).head()\n# support itemsets\n# 12 0.442786 (POSTAGE)\n# 18 0.185738 (ROUND SNACK BOXES SET OF4 WOODLAND )\n# 34 0.139303 (ROUND SNACK BOXES SET OF4 WOODLAND , POSTAGE)\n# 17 0.119403 (ROUND SNACK BOXES SET OF 4 FRUITS )\n# 11 0.104478 (PLASTERS IN TIN WOODLAND ANIMALS)\n\n# 고빈도 집합 기반의 규칙 찾기\nrules = association_rules(freq_items, metric='lift', min_threshold=1)\nrules.sort_values(by='lift', ascending=False, inplace=True)\n\n# 후행 품목집합이 있는 경우\nrules[rules.consequents != ''][['antecedents', 'consequents', 'support', 'confidence', 'lift']].head()\n# antecedents consequents support confidence lift\n# 4 (PLASTERS IN TIN WOODLAND ANIMALS) (PLASTERS IN TIN CIRCUS PARADE ) 0.051410 0.492063 5.598383\n# 5 (PLASTERS IN TIN CIRCUS PARADE ) (PLASTERS IN TIN WOODLAND ANIMALS) 0.051410 0.584906 5.598383\n# 34 (ROUND SNACK BOXES SET OF 4 FRUITS ) (ROUND SNACK BOXES SET OF4 WOODLAND , POSTAGE) 0.079602 0.666667 4.785714\n# 31 (ROUND SNACK BOXES SET OF4 WOODLAND , POSTAGE) (ROUND SNACK BOXES SET OF 4 FRUITS ) 0.079602 0.571429 4.785714\n# 26 (ROUND SNACK BOXES SET OF4 WOODLAND ) (ROUND SNACK BOXES SET OF 4 FRUITS ) 0.099502 0.535714 4.486607\n\n","sub_path":"reference/understanding-ml-code/ch24-연관 분석/association_basics.py","file_name":"association_basics.py","file_ext":"py","file_size_in_byte":7643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"405532465","text":"import random\n\nChoice = random.choice([\"Is Jacob stable?: \",\"Does Aiden have good fills?: \",\"Does Jacob play good?: \"])\nList = [\"Is Jacob stable?: \",\"Does Aiden have good fills?: \",\"Does Jacob play good?: \"]\nwhile True:\n Input = input(Choice)\n if Choice == List[0]:\n if Input.lower() in [\"yes\",\"y\",\"ye\"]:\n print(\"Incorrect, doctor says not\")\n break\n elif Input.lower() in [\"no\",\"n\"]:\n print(\"You have 200 IQ\")\n break\n else:\n print(\"Incorrect syntax, try again\")\n elif Choice == List[1]:\n print(\"Is this even a question? Of course he does, jk he needs notes from bekkett and luke\")\n break\n elif Choice == List[2]:\n if Input.lower() in [\"yes\",\"ye\",\"y\"]:\n print(\"Incorrect, he needs to take notes\")\n break\n elif Input.lower() in [\"no\",\"n\"]:\n print(\"True, he needs to take notes\")\n break\n else:\n print(\"Incorrect syntax, try again\")","sub_path":"PythonApplication1/src/Choices.py","file_name":"Choices.py","file_ext":"py","file_size_in_byte":1006,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"48722464","text":"if __name__ == '__main__':\n\tf = open('USDJPY.csv', 'r')\n\tarray = []\n\tfor line in f:\n\t\tvalues = line.split(',')\n\t\tv_open = eval(values[2])\n\t\tv_high = eval(values[3])\n\t\tv_low = eval(values[4])\n\t\tv_close = eval(values[5])\n\t\tv_amount = eval(values[6])\n\t\tarray.append((v_open, v_close, v_high, v_low, v_amount))\n\tf.close()\n\t\n\tpos = 0\n\tsummary = []\n\t\n\tP = 5\n\t\n\twhile pos + P <= len(array):\n\t\tS = 0\n\t\tA = 0\n\t\tH = 0\n\t\tL = 999\n\t\t\n\t\tv_open = array[pos][0]\n\t\tv_close = array[pos + P - 1][1]\n\t\t\n\t\tfor j in range(P):\n\t\t\th = array[pos + j][2]\n\t\t\tl = array[pos + j][3]\n\t\t\taverage = (h + l) / 2\n\t\t\t\n\t\t\tS += average * array[pos + j][4]\n\t\t\tA += array[pos + j][4]\n\t\t\tH = max(H, h)\n\t\t\tL = min(L, l)\n\t\t\n\t\tsummary.append((S / A, v_open, v_close, H, L, A))\n\t\tpos += P\n\t\n\tM = 100000\n\t\n\tfor j in range(len(summary) / M):\n\t\tf = open('train' + str(j + 1) + '.txt', 'w')\n\t\tfor k in range(M):\n\t\t\tpos = j * M + k\n\t\t\tf.write(str(summary[pos]) + '\\n')\n\t\tf.close()\n\t\t\n\tf = open('test.txt', 'w')\n\tpos = (len(summary) / M) * M\n\twhile pos < len(summary):\n\t\tf.write(str(summary[pos]) + '\\n')\n\t\tpos += 1\n\tf.close()\n","sub_path":"history/extract.py","file_name":"extract.py","file_ext":"py","file_size_in_byte":1088,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"424060472","text":"import re\n\nf = open('wb_dump.sql','r')\np_name = re.compile(\"'[A-Z][a-z]+[^']*'\")\np_coordinates = re.compile(\"MULTIPOLYGON\\([^']*'\")\nfor line in f:\n m = p_name.search(line)\n country_name = m.group()[1:(len(m.group())-1)]\n m = p_coordinates.search(line)\n coordinates_string = m.group()[13:(len(m.group())-5)]\n coordinates = coordinates_string.split(' )),((')\n coordinates[0] = coordinates[0][2:len(coordinates[0])]\n f_out = open('countries_borders/'+country_name+'.json','w')\n f_out.write('{\\n')\n f_out.write('\t\"country_name\": '+ '\"' + country_name + '\",\\n')\n f_out.write('\t\"coordinates\": [\\n')\n last_elem = coordinates.pop()\n for elem in coordinates:\n f_out.write('\t\t[\\n')\n terr = elem.split(',')\n last_path = terr.pop()\n for path in terr:\n path = re.sub(' +', ',', path)\n f_out.write('\t\t\t\"'+ path[0:len(path)-1] + '\",\\n')\n last_path = re.sub(' +', ',', last_path)\n f_out.write('\t\t\t\"'+ path[0:len(path)-1] + '\"\\n')\n f_out.write('\t\t],\\n')\n f_out.write('\t\t[\\n')\n terr = last_elem.split(',')\n last_path = terr.pop()\n for path in terr:\n path = re.sub(' +', ',', path)\n f_out.write('\t\t\t\"'+ path[0:len(path)-1] + '\",\\n')\n last_path = re.sub(' +', ',', last_path)\n f_out.write('\t\t\t\"'+ path[0:len(path)-1] + '\"\\n')\n f_out.write('\t\t]\\n')\n f_out.write('\t]\\n')\n f_out.write('}')\n f_out.close()\nf.close()\n","sub_path":"public/countries/python-script.py","file_name":"python-script.py","file_ext":"py","file_size_in_byte":1443,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"143474030","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Jun 25 14:18:30 2018\n\n@author: Fenqiang Zhao, https://github.com/zhaofenqiang\n\nContact: zhaofenqiang0221@gmail.com\n\n\"\"\"\n\nimport torch\nimport numpy as np\nimport torch.nn as nn\n\n\nclass repa_conv_layer(nn.Module):\n \"\"\"Define the convolutional layer on icosahedron discretized sphere using \n rectagular filter in tangent plane\n \n Parameters:\n in_feats (int) - - input features/channels\n out_feats (int) - - output features/channels \n \n Input: \n N x in_feats, tensor\n Return:\n N x out_feats, tensor\n \"\"\" \n def __init__(self, in_feats, out_feats, neigh_indices, neigh_weights):\n super(repa_conv_layer, self).__init__()\n\n self.in_feats = in_feats\n self.out_feats = out_feats\n self.neigh_indices = neigh_indices.reshape(-1) - 1\n self.weight = nn.Linear(25 * in_feats, out_feats)\n self.nodes_number = neigh_indices.shape[0]\n \n neigh_weights = np.reshape(np.tile(neigh_weights, self.in_feats), (neigh_weights.shape[0],neigh_weights.shape[1],3,-1)).astype(np.float32)\n self.neigh_weights = torch.from_numpy(neigh_weights).cuda() \n \n def forward(self, x):\n \n mat = x[self.neigh_indices]\n mat = mat.view(self.nodes_number, 25, 3, -1)\n assert(mat.size() == torch.Size([self.nodes_number, 25, 3, self.in_feats]))\n \n assert(mat.size() == self.neigh_weights.size())\n\n x = torch.mul(mat, self.neigh_weights)\n x = torch.sum(x, 2).view(self.nodes_number, -1)\n assert(x.size() == torch.Size([self.nodes_number, 25 * self.in_feats]))\n \n out = self.weight(x)\n return out\n\n\nclass onering_conv_layer(nn.Module):\n \"\"\"The convolutional layer on icosahedron discretized sphere using \n 1-ring filter\n \n Parameters:\n in_feats (int) - - input features/channels\n out_feats (int) - - output features/channels\n \n Input: \n N x in_feats tensor\n Return:\n N x out_feats tensor\n \"\"\" \n def __init__(self, in_feats, out_feats, neigh_orders, neigh_indices=None, neigh_weights=None):\n super(onering_conv_layer, self).__init__()\n\n self.in_feats = in_feats\n self.out_feats = out_feats\n self.neigh_orders = neigh_orders\n \n self.weight = nn.Linear(7 * in_feats, out_feats)\n \n def forward(self, x):\n \n mat = x[self.neigh_orders].view(len(x), 7*self.in_feats)\n \n out_features = self.weight(mat)\n return out_features\n \n \nclass tworing_conv_layer(nn.Module):\n \"\"\"The convolutional layer on icosahedron discretized sphere using \n 2-ring filter\n \n Parameters:\n in_feats (int) - - input features/channels\n out_feats (int) - - output features/channels\n \n Input: \n N x in_feats tensor\n Return:\n N x out_feats tensor\n \"\"\" \n def __init__(self, in_feats, out_feats, neigh_orders):\n super(tworing_conv_layer, self).__init__()\n\n self.in_feats = in_feats\n self.out_feats = out_feats\n self.neigh_orders = neigh_orders\n \n self.weight = nn.Linear(19 * in_feats, out_feats)\n \n def forward(self, x):\n \n mat = x[self.neigh_orders].view(len(x), 19*self.in_feats)\n \n out_features = self.weight(mat)\n return out_features\n \n\n \nclass pool_layer(nn.Module):\n \"\"\"\n The pooling layer on icosahedron discretized sphere using 1-ring filter\n \n Input: \n N x D tensor\n Return:\n ((N+6)/4) x D tensor\n \n \"\"\" \n\n def __init__(self, neigh_orders, pooling_type='mean'):\n super(pool_layer, self).__init__()\n\n self.neigh_orders = neigh_orders\n self.pooling_type = pooling_type\n \n def forward(self, x):\n \n num_nodes = int((x.size()[0]+6)/4)\n feat_num = x.size()[1]\n x = x[self.neigh_orders[0:num_nodes*7]].view(num_nodes, feat_num, 7)\n if self.pooling_type == \"mean\":\n x = torch.mean(x, 2)\n if self.pooling_type == \"max\":\n x = torch.max(x, 2)\n assert(x[0].size() == torch.Size([num_nodes, feat_num]))\n return x[0], x[1]\n \n assert(x.size() == torch.Size([num_nodes, feat_num]))\n \n return x\n \n \nclass upconv_layer(nn.Module):\n \"\"\"\n The transposed convolution layer on icosahedron discretized sphere using 1-ring filter\n \n Input: \n N x in_feats, tensor\n Return:\n ((Nx4)-6) x out_feats, tensor\n \n \"\"\" \n\n def __init__(self, in_feats, out_feats, upconv_top_index, upconv_down_index):\n super(upconv_layer, self).__init__()\n\n self.in_feats = in_feats\n self.out_feats = out_feats\n self.upconv_top_index = upconv_top_index\n self.upconv_down_index = upconv_down_index\n self.weight = nn.Linear(in_feats, 7 * out_feats)\n \n def forward(self, x):\n \n raw_nodes = x.size()[0]\n new_nodes = int(raw_nodes*4 - 6)\n x = self.weight(x)\n x = x.view(len(x) * 7, self.out_feats)\n x1 = x[self.upconv_top_index]\n assert(x1.size() == torch.Size([raw_nodes, self.out_feats]))\n x2 = x[self.upconv_down_index].view(-1, self.out_feats, 2)\n x = torch.cat((x1,torch.mean(x2, 2)), 0)\n assert(x.size() == torch.Size([new_nodes, self.out_feats]))\n return x\n\n\nclass upsample_interpolation(nn.Module):\n \"\"\"\n The upsampling layer on icosahedron discretized sphere using interpolation\n \n Input: \n N x in_feats, tensor\n Return:\n ((Nx4)-6) x in_feats, tensor\n \n \"\"\" \n\n def __init__(self, upsample_neighs_order):\n super(upsample_interpolation, self).__init__()\n\n self.upsample_neighs_order = upsample_neighs_order\n \n def forward(self, x):\n \n num_nodes = x.size()[0] * 4 - 6\n feat_num = x.size()[1]\n x1 = x[self.upsample_neighs_order].view(num_nodes - x.size()[0], feat_num, 2)\n x1 = torch.mean(x1, 2)\n x = torch.cat((x,x1),0)\n \n return x\n\n\nclass upsample_fixindex(nn.Module):\n \"\"\"\n The upsampling layer on icosahedron discretized sphere using fixed indices 0,\n padding new vertices with 0\n \n Input: \n N x in_feats, tensor\n Return:\n ((Nx4)-6) x in_feats, tensor\n \n \"\"\" \n def __init__(self, upsample_neighs_order):\n super(upsample_fixindex, self).__init__()\n\n self.upsample_neighs_order = upsample_neighs_order\n \n def forward(self, x):\n \n num_nodes = x.size()[0] * 4 - 6\n feat_num = x.size()[1]\n x1 = torch.zeros(num_nodes - x.size()[0], feat_num).cuda()\n x = torch.cat((x,x1),0)\n \n return x\n \n \nclass upsample_maxindex(nn.Module):\n \"\"\"\n The upsampling layer on icosahedron discretized sphere using max indices.\n \n Input: \n N x in_feats, tensor\n Return:\n ((Nx4)-6) x in_feats, tensor\n \n \"\"\" \n\n def __init__(self, num_nodes, neigh_orders):\n super(upsample_maxindex, self).__init__()\n\n self.num_nodes = num_nodes\n self.neigh_orders = neigh_orders\n \n def forward(self, x, max_index):\n \n raw_nodes, feat_num = x.size()\n assert(max_index.size() == x.size())\n x = x.view(-1) \n \n y = torch.zeros(self.num_nodes, feat_num).to(torch.device(\"cuda\"))\n column_ref = torch.zeros(raw_nodes, feat_num)\n for i in range(raw_nodes):\n column_ref[i,:] = i * 7 + max_index[i,:] \n column_index = self.neigh_orders[column_ref.view(-1).long()]\n column_index = torch.from_numpy(column_index).long()\n row_index = np.floor(np.linspace(0.0, float(feat_num), num=raw_nodes*feat_num))\n row_index[-1] = row_index[-1] - 1\n row_index = torch.from_numpy(row_index).long()\n y[column_index, row_index] = x\n \n return y\n\n\n \n","sub_path":"sphericalunet/layers.py","file_name":"layers.py","file_ext":"py","file_size_in_byte":8122,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"423645750","text":"\"\"\"\n10.0 快速排序(Quick Sort)\n\n·取一个元素p(第一个元素),使元素p归位\n-->列表被p分成两部分,左边都比p小,右边都比p大\n-->递归完成排序\n·快速排序框架:\n def quick_sort(data, left, right):\n if left < right:\n mid = partition(data, left, right)\n quick_sort(data, left, mid - 1)\n quick_sort(data, mid + 1, right)\n·时间复杂度:O(nlogn)\n·快速排序的问题:最坏情况(e.g.倒序列表的情况O(n^2)),递归最大深度\n\n===============================================================================\n===============================================================================\n\"\"\"\n\n\n# ========== example ==========\ndef partition(li, left, right):\n tmp = li[left]\n while left < right:\n while left < right and li[right] >= tmp: # 从右边找比tmp小的数\n right -= 1 # 往左走一步\n li[left] = li[right] # 把右边的值写到左边的空位上\n print(li)\n while left < right and li[left] <= tmp:\n left += 1\n li[right] = li[left] # 把左边的值写到右边的空位上\n print(li)\n li[left] = tmp # 把tmp归位\n return left\n\n\ndef quick_sort(li, left, right):\n if left < right: # 至少两个元素\n mid = partition(li, left, right)\n quick_sort(li, left, mid - 1)\n quick_sort(li, mid + 1, right)\n\n\nli = [5, 7, 4, 6, 3, 1, 2, 9, 8]\nli2 = [7, 2, 8, 4, 6, 7, 5, 6, 4, 9]\n# print(li2)\npartition(li, 0, len(li) - 1)\n# quick_sort(li2, 0, len(li2) - 1)\nprint(li)\n","sub_path":"10 Quick Sort-快速排序.py","file_name":"10 Quick Sort-快速排序.py","file_ext":"py","file_size_in_byte":1588,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"165955322","text":"#!/usr/bin/python3\r\n\r\n# Demonstrates point sprites\r\n# Ben Smith\r\n# benjamin.coder.smith@gmail.com\r\n#\r\n# based on pointsprites.cpp\r\n# OpenGL SuperBible\r\n# Program by Richard S. Wright Jr.\r\n\r\n\r\nfrom OpenGL.GL import *\r\nfrom OpenGL.GLUT import *\r\nfrom OpenGL.GLU import *\r\n\r\nfrom PIL import Image\r\nimport time \r\n\r\nESCAPE = b'\\033'\r\n\r\n\r\n\r\nfrom math import cos, sin\r\n\r\nfrom random import randint\r\n\r\nimport sys\r\nsys.path.append(\"../shared\")\r\n\r\nfrom math3d import M3DVector2f\r\n\r\n\r\n# Array of small stars\r\nSMALL_STARS = 100\r\nvSmallStars = [M3DVector2f() for i in range (0, SMALL_STARS)]\r\n\r\nMEDIUM_STARS = 40\r\nvMediumStars = [M3DVector2f() for i in range (0, MEDIUM_STARS)]\r\n\r\nLARGE_STARS = 40\r\nvLargeStars = [M3DVector2f() for i in range (0, LARGE_STARS)]\r\n\r\nSCREEN_X = 800\r\nSCREEN_Y = 600\r\n\r\ndrawMode = 3\r\ntextureObjects = (GLuint * 2)()\r\n\r\ndef InitGL(Width, Height):\r\n\r\n # Turn off blending and all smoothing\r\n if drawMode == 1:\r\n glDisable(GL_BLEND)\r\n glDisable(GL_LINE_SMOOTH)\r\n glDisable(GL_POINT_SMOOTH)\r\n glDisable(GL_TEXTURE_2D)\r\n glDisable(GL_POINT_SPRITE)\r\n \r\n # Turn on antialiasing, and give hint to do the best\r\n # job possible.\r\n if drawMode == 2:\r\n \r\n glBlendFunc(GL_SRC_ALPHA, GL_ONE_MINUS_SRC_ALPHA)\r\n glEnable(GL_BLEND)\r\n glEnable(GL_POINT_SMOOTH)\r\n glHint(GL_POINT_SMOOTH_HINT, GL_NICEST)\r\n glEnable(GL_LINE_SMOOTH)\r\n glHint(GL_LINE_SMOOTH_HINT, GL_NICEST)\r\n glDisable(GL_TEXTURE_2D)\r\n glDisable(GL_POINT_SPRITE)\r\n\r\n # Point Sprites\r\n elif drawMode == 3:\r\n glEnable(GL_BLEND)\r\n glBlendFunc(GL_SRC_COLOR, GL_ONE_MINUS_SRC_COLOR)\r\n glDisable(GL_LINE_SMOOTH)\r\n glDisable(GL_POINT_SMOOTH)\r\n glDisable(GL_POLYGON_SMOOTH)\r\n \r\n # Populate star list\r\n for i in range(0, SMALL_STARS):\r\n vSmallStars[i][0] = float(randint(0, SCREEN_X))\r\n vSmallStars[i][1] = float(randint(0, SCREEN_Y - 100))+100.0\r\n\r\n # Populate star list\r\n for i in range(0, MEDIUM_STARS):\r\n vMediumStars[i][0] = float(randint(0, SCREEN_X * 10))/10.0\r\n vMediumStars[i][1] = float(randint(0, SCREEN_Y - 100))+100.0\r\n\r\n # Populate star list\r\n for i in range(0, LARGE_STARS):\r\n vLargeStars[i][0] = float(randint(0, SCREEN_X * 10))/10.0\r\n vLargeStars[i][1] = float(randint(0, SCREEN_Y - 100) * 10.0)/ 10.0 +100.0\r\n \r\n # Black background\r\n glClearColor(0.0, 0.0, 0.0, 1.0 )\r\n\r\n # Set drawing color to white\r\n glColor3f(0.0, 0.0, 0.0)\r\n\r\n # Load our textures\r\n glGenTextures(2, textureObjects)\r\n glBindTexture(GL_TEXTURE_2D, textureObjects[0])\r\n \r\n # Load this texture map\r\n glTexParameteri(GL_TEXTURE_2D, GL_GENERATE_MIPMAP, GL_TRUE)\r\n img = Image.open('star.png').convert(\"RGB\")\r\n raw_image = img.tobytes()\r\n glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB, img.width, img.height, 0, GL_RGB, GL_UNSIGNED_BYTE, raw_image)\r\n \r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR_MIPMAP_LINEAR)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE)\r\n\r\n glBindTexture(GL_TEXTURE_2D, textureObjects[1])\r\n glTexParameteri(GL_TEXTURE_2D, GL_GENERATE_MIPMAP, GL_TRUE)\r\n img = Image.open('moon.png').convert(\"RGB\")\r\n raw_image = img.tobytes()\r\n glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB, img.width, img.height, 0, GL_RGB, GL_UNSIGNED_BYTE, raw_image)\r\n \r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR_MIPMAP_LINEAR)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE)\r\n glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE)\r\n\r\n glTexEnvi(GL_POINT_SPRITE, GL_COORD_REPLACE, GL_TRUE)\r\n glTexEnvi(GL_TEXTURE_ENV, GL_TEXTURE_ENV_MODE, GL_DECAL)\r\n\r\n # Called to draw scene\r\ndef DrawGLScene():\r\n \r\n x = 700.0 # Location and radius of moon\r\n y = 500.0\r\n r = 50.0\r\n angle = 0.0 # Another looping variable\r\n\r\n # Clear the window\r\n glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT)\r\n \r\n # Everything is white\r\n glColor3f(1.0, 1.0, 1.0)\r\n \r\n if (drawMode == 3):\r\n glEnable(GL_POINT_SPRITE)\r\n glEnable(GL_TEXTURE_2D)\r\n glBindTexture(GL_TEXTURE_2D, textureObjects[0])\r\n glEnable(GL_BLEND)\r\n \r\n # Draw small stars\r\n glPointSize(7.0)\r\n glBegin(GL_POINTS)\r\n for i in range(0, SMALL_STARS):\r\n glVertex2fv(vSmallStars[i])\r\n glEnd()\r\n \r\n # Draw medium sized stars\r\n glPointSize(12.0)\r\n glBegin(GL_POINTS)\r\n for i in range(0, MEDIUM_STARS):\r\n glVertex2fv(vMediumStars[i])\r\n glEnd()\r\n \r\n # Draw largest stars\r\n glPointSize(20.0)\r\n glBegin(GL_POINTS)\r\n for i in range(0, LARGE_STARS):\r\n glVertex2fv(vLargeStars[i])\r\n glEnd()\r\n \r\n \r\n glPointSize(120.0)\r\n if (drawMode == 3):\r\n glDisable(GL_BLEND)\r\n glBindTexture(GL_TEXTURE_2D, textureObjects[1])\r\n \r\n glBegin(GL_POINTS)\r\n glVertex2f(x, y)\r\n glEnd()\r\n \r\n glDisable(GL_TEXTURE_2D)\r\n glDisable(GL_POINT_SPRITE)\r\n\r\n # Draw distant horizon\r\n glLineWidth(3.5)\r\n glBegin(GL_LINE_STRIP)\r\n \r\n glVertex2f(0.0, 25.0)\r\n glVertex2f(50.0, 100.0)\r\n glVertex2f(100.0, 25.0)\r\n glVertex2f(225.0, 115.0)\r\n glVertex2f(300.0, 50.0)\r\n glVertex2f(375.0, 100.0)\r\n glVertex2f(460.0, 25.0)\r\n glVertex2f(525.0, 100.0)\r\n glVertex2f(600.0, 20.0)\r\n glVertex2f(675.0, 70.0)\r\n glVertex2f(750.0, 25.0)\r\n glVertex2f(800.0, 90.0)\r\n \r\n glEnd()\r\n\r\n glutSwapBuffers() \r\n \r\n \r\ndef ReSizeGLScene(w, h):\r\n # Prevent a divide by zero\r\n if(h == 0):\r\n h = 1\r\n \r\n # Set Viewport to window dimensions\r\n glViewport(0, 0, w, h)\r\n\r\n # Reset projection matrix stack\r\n glMatrixMode(GL_PROJECTION)\r\n glLoadIdentity()\r\n\r\n # Establish clipping volume (left, right, bottom, top, near, far)\r\n gluOrtho2D(0.0, SCREEN_X, 0.0, SCREEN_Y)\r\n\r\n\r\n # Reset Model view matrix stack\r\n glMatrixMode(GL_MODELVIEW)\r\n glLoadIdentity()\r\n \r\ndef keyPressed(key, x, y):\r\n if key == ESCAPE:\r\n glutDestroyWindow(window)\r\n sys.exit()\r\n\r\n \r\n# Main program entry point\r\nif __name__ == '__main__':\r\n\r\n glutInit()\r\n glutInitDisplayMode(GLUT_RGBA | GLUT_DOUBLE | GLUT_ALPHA | GLUT_DEPTH)\r\n glutInitWindowSize(640, 480)\r\n glutInitWindowPosition(0, 0)\r\n window = glutCreateWindow(\"Smoothing Out The Jaggies\")\r\n \r\n glutDisplayFunc(DrawGLScene)\r\n\r\n # Uncomment this line to get full screen.\r\n #glutFullScreen()\r\n \r\n #glutIdleFunc(DrawGLScene)\r\n #glutTimerFunc( int(1.0/60.0), update, 0)\r\n \r\n glutReshapeFunc(ReSizeGLScene)\r\n glutKeyboardFunc(keyPressed)\r\n # glutSpecialFunc (specialkeyPressed);\r\n\r\n # Initialize our window. \r\n InitGL(640, 480)\r\n \r\n # Start Event Processing Engine\t\r\n glutMainLoop()\r\n\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"chapt09/pointsprites.py","file_name":"pointsprites.py","file_ext":"py","file_size_in_byte":7741,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"478261339","text":"\nfrom edgetpu.detection.engine import DetectionEngine\nfrom PIL import Image\nimport pyrealsense2 as rs\nimport numpy as np\nimport argparse\nimport imutils\nimport time \nimport sys\nimport cv2\n\nap = argparse.ArgumentParser()\nap.add_argument(\"-m\", \"--model\", required=True, help=\"path to Tensorflow Lite object detection model\")\nap.add_argument(\"-l\", \"--labels\", required=True, help=\"path to labels file\")\nap.add_argument(\"-c\", \"--confidence\", type=float, default=0.3, help = \"Confidence threshold expressed as a decimal\")\nap.add_argument(\"-w\", \"--width\", type = int, default = 500, help = \"Size of image as it's inputed into the model\")\nap.add_argument(\"-i\", \"--info\", type = int, default = 0, help = \"Prints relevent runtime information in the console on startup\")\nargs = vars(ap.parse_args())\n\nlabels={}\n\nfor row in open(args[\"labels\"]):\n (classID, label) = row.strip().split(maxsplit=1)\n labels[int(classID)] = label.strip()\n\nprint(\"Tensorflow-Lite model loading...\")\nmodel = DetectionEngine(args[\"model\"])\n\nprint(\"Initializing Realsense...\")\npipeline = rs.pipeline()\nconfig = rs.config()\ndepth_resolution_x =640\ndepth_resolution_y =480\ncolor_resolution_x =640\ncolor_resolution_y =480\ndepth_fps =30\ncolor_fps =60\nconfig.enable_stream(rs.stream.depth, depth_resolution_x, depth_resolution_y, rs.format.z16, depth_fps) \nconfig.enable_stream(rs.stream.color, color_resolution_x, color_resolution_y, rs.format.bgr8, color_fps)\nprofile = pipeline.start(config)\ndepth_sensor = profile.get_device().first_depth_sensor()\ndepth_scale = depth_sensor.get_depth_scale()\n\nif args[\"info\"] > 0:\n print(\"Depth Input: \",depth_resolution_x,\"x\",depth_resolution_y,\"at\",depth_fps,\"fps\")\n print(\"Color Input: \",color_resolution_x,\"x\",color_resolution_y,\"at\",color_fps,\"fps\")\n\n#align_with = rs.stream.color\n#align = rs.align(align_with)\n\nstart_time = time.time()\nx = 1 # displays the frame rate every 1 second\ncounter = 0\n\n\nwhile True:\n ##wait for a new frame and then get the depth and color frame \n frames = pipeline.wait_for_frames() \n depth_frame = frames.get_depth_frame()\n color_frame = frames.get_color_frame()\n if not depth_frame or not color_frame:\n continue\n \n ##create numpy array of depth and color frames\n depth_image = np.asanyarray(depth_frame.get_data())\n color_image = np.asanyarray(color_frame.get_data())\n ##resize image based upon argument and create a copy to annotate and display\n color_image = imutils.resize(color_image, width=args[\"width\"])\n orig = color_image\n #color_image = cv2.cvtColor(color_image, cv2.COLOR_BGR2RGB)\n color_image = Image.fromarray(color_image)\n\n ##start a timer for inferencing time and feed the frame into the model \n start_inference = time.time()\n results = model.DetectWithImage(color_image, threshold=args[\"confidence\"],\n keep_aspect_ratio=True, relative_coord=False)\n end_inference = time.time()\n \n ##put a bounding box on the result in the copy image\n for r in results:\n bounding_box = r.bounding_box.flatten().astype(\"int\") #try changing r\n (startX, startY, endX, endY) = bounding_box #try changing startx etc\n label = labels[r.label_id]\n cv2.rectangle(orig, (startX, startY), (endX, endY),\n (0, 255, 0), 2)\n \n ##add a dot to the center of the bounding box\n centerX = int((startX - ((startX - endX)*0.5)))\n centerY = int((startY - ((startY - endY)*0.5)))\n cv2.circle(orig, (centerX, centerY), 1, (0, 255, 0), 2 )\n\n #-------Single Point-----------\n ##calculates depth of the center point of the bounding box\n depth = depth_image[centerY, centerX].astype(float)\n depth = depth * depth_scale\n #------------------------------\n \n #-----------Average------------\n ##calculates depth of the center point of the bounding box\n #depth_center = depth_image[centerY, centerX].astype(float)\n #depth_collection = depth_center + depth_image[(centerX+1), (centerY+1)].astype(float) + depth_image[(centerX+1), (centerY-1)].astype(float) + depth_image[(centerX-1), (centerY-1)].astype(float) + depth_image[(centerX-1), (centerY+1)].astype(float)\n #depth_average = depth_collection/5\n #depth = depth_average * depth_scale\n #print(depth)\n #------------------------------\n\n y = startY - 15 if startY - 15 > 15 else startY + 15\n text = \"{}: {:.2f}% {:.3f}\".format(label, r.score*100, depth) \n cv2.putText(orig, text, (startX,y),\n cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 255, 0),2) #try changing font\n\n\n ##create the window to display the result\n cv2.namedWindow('Real Sense Object Detection', cv2.WINDOW_AUTOSIZE)\n cv2.imshow('Real Sense Object Detection', orig)\n key = cv2.waitKey(1) & 0xFF\n\n if args[\"info\"] > 0:\n ##create counter for measuring inference time and frames per second\n counter += 1\n if (time.time() - start_time) > x:\n print(\"FPS: \", counter / (time.time() - start_time))\n fps = []\n fps.append(counter / (time.time() - start_time))\n inference_time = []\n inference_time.append(end_inference - start_inference)\n counter = 0\n start_time = time.time()\n\n if key == ord(\"q\") or key == 27:\n break\n\n##calculate the average FPS and inference time \nif args[\"info\"] > 0:\n fps_sum = 0\n for num in fps:\n fps_sum += num\n fps_average = fps_sum / len(fps)\n print(\"Average FPS:\",fps_average)\n\n inference_sum = 0\n for num in inference_time:\n inference_sum += num\n inference_average = inference_sum / len(inference_time)\n print(\"Average Inference Time:\",inference_average) \n\n \npipeline.stop()\n \n \n\n","sub_path":"Realsense_Object_Detection_Public.py","file_name":"Realsense_Object_Detection_Public.py","file_ext":"py","file_size_in_byte":5780,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"142161884","text":"from django.conf.urls import url\n\nfrom . import views\n\nurlpatterns = [\n url(r'^chat/$', views.index, name='index'),\n url(r'^save-theme/$', views.save_theme ,name=\"save_theme\"),\n url(r'^save-thumb/$', views.save_thumb, name=\"save_thumb\"),\n url(r'^(?P[^/]+)/$', views.room, name='room'),\n]","sub_path":"cms_app/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"221214827","text":"from Spiff_report2 import *\n\nclass Jason_spiff(Spiff_report):\n\tdef __init__(self, spiff_name, fname_eclipse, fname_master):\n\t\tsuper().__init__(spiff_name, fname_eclipse, fname_master)\n\tdef transform_eclipse(self, t_dict, key):\n\t\ttemp = t_dict[key].split(' ')\n\t\tif len(temp) == 1: t_dict[\"CatMatch\"] = temp[0]\n\t\tif temp[0] == \"JASON\": t_dict[\"CatMatch\"] = temp[1]\n\t\telse: t_dict[\"CatMatch\"] = temp[0]\n\t\treturn\n\tdef transform_master(self, t_dict, key):\n\t\ttemp = t_dict[key]\n\t\tt_dict['CatMatchM'] = temp\n\t\treturn\n\tdef cust_clean(self):\n\t\tfor i in range(0, len(self.e_data)): self.e_data.transform_index(i, 'DESC', self.transform_eclipse)\n\t\tfor i in range(0, len(self.m_data)): self.m_data.transform_index(i, 'PART_ID', self.transform_master)\n\tdef cust_match(self):\n\t\tfor i in range(0, len(self.e_data)):\n\t\t\ttemp = self.e_data.get_index(i)\n\t\t\tcatnum = temp[\"CatMatch\"]\n\t\t\tloc = self.m_data.search('CatMatchM', catnum)\n\t\t\tif loc < 0:\n\t\t\t\ttemp['Spiff Match'] = \"N/A\"\n\t\t\t\tcontinue\n\t\t\tspiff = self.m_data.get_index(loc)['Spiff']\n\t\t\ttemp['Spiff Match'] = spiff","sub_path":"Jason_spiff.py","file_name":"Jason_spiff.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"125271251","text":"# import python modules\nimport sys\n\n# import external libraries\nimport numpy as np\nnp.set_printoptions(suppress=True)\n\n# import custom libraries\nfrom filters import *\nfrom fft_features import *\nfrom data_receive import *\nfrom models import *\n# from pi_bridge import *\n\n\ndef main():\n\tfilters = Filter()\n\ttransform = Transform()\n\tmodel = KerasModels()\n\t# pi = piConnectionServer()\n\n\twhile True:\n\t\tx = input(\"Enter 1 to start recording and press 'q' or 'Q' to exit: \")\n\t\tif x == '1':\n\t\t\tdata_getter = GanglionDataReceiver()\n\n\t\t\teeg_recording = data_getter.get_signal()\n\t\t\to1 = eeg_recording[0]\n\t\t\to2 = eeg_recording[1]\n\n\t\t\t# NOTCH WAS CALCULATED BUT WAS PASSED IMMEDIATELY TO BANDPASS SO IT WASNT VISIBLE. PLOTS ERROR\n\t\t\to1_notch = filters.notch_filter(o1)\n\t\t\to2_notch = filters.notch_filter(o2)\n\n\t\t\to1_filtered = filters.bp_filter(o1_notch)\n\t\t\to2_filtered = filters.bp_filter(o2_notch)\n\n\n\t\t\t# SAVING ARRAYS FOR PLOTTING\n\t\t\t# np.savez('alpha_front', o1=o1, o2=o2, o1_notch=o1_notch, o2_notch=o2_notch, o1_filtered=o1_filtered, o2_filtered=o2_filtered)\n\t\t\t# np.savez('alpha_back', o1=o1, o2=o2, o1_notch=o1_notch, o2_notch=o2_notch, o1_filtered=o1_filtered, o2_filtered=o2_filtered)\t\t\t\n\t\t\t# np.savez('alpha_left', o1=o1, o2=o2, o1_notch=o1_notch, o2_notch=o2_notch, o1_filtered=o1_filtered, o2_filtered=o2_filtered)\n\t\t\t# np.savez('alpha_right', o1=o1, o2=o2, o1_notch=o1_notch, o2_notch=o2_notch, o1_filtered=o1_filtered, o2_filtered=o2_filtered)\n\n\t\t\t# Inputs are returned normalized in the range of [0, 1]. Check fft.py normalize()\n\t\t\to1_sum = transform.sum_fourier_trans(o1_filtered)\n\t\t\to2_sum = transform.sum_fourier_trans(o2_filtered)\n\n\t\t\t# append lists and then convert them to np array\n\t\t\tnn_input = np.asarray(o1_sum + o2_sum)\n\t\t\tnn_input.shape = (1, 8)\n\n\t\t\tpredictions = model.nn_model(nn_input)\n\t\t\tpredictions_list = [x.item() for x in predictions]\n\t\t\tpredicted_movement = predictions_list.index(max(predictions_list))\n\t\t\tif predicted_movement == 0:\n\t\t\t\tprint(predicted_movement, ' BACK')\n\t\t\t\t# pi.sendData('BACK')\n\t\t\telif predicted_movement == 1:\n\t\t\t\tprint(predicted_movement, ' FRONT')\n\t\t\t\t# pi.sendData('FRONT')\n\t\t\telif predicted_movement == 2:\n\t\t\t\tprint(predicted_movement, ' LEFT')\n\t\t\t\t# pi.sendData('LEFT')\n\t\t\telse:\n\t\t\t\tprint(predicted_movement, ' RIGHT')\n\t\t\t\t# pi.sendData('RIGHT')\n\n\t\telif x == 'q' or x == 'Q':\n\t\t\t# pi.sendData('q Q Entered. Exiting...')\n\t\t\t# pi.closeConnection()\n\t\t\tsys.exit()\n\n\nif __name__ == '__main__':\n\tmain()","sub_path":"computer_main.py","file_name":"computer_main.py","file_ext":"py","file_size_in_byte":2452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"96243055","text":"from django.shortcuts import render, HttpResponseRedirect, HttpResponse\nfrom django.http import JsonResponse\nfrom django.views.generic import ListView, DetailView\nfrom .models import Posts, Comments\nfrom django.forms import ModelForm\nfrom django.views.generic.edit import CreateView, DeleteView, UpdateView\nfrom django.core.urlresolvers import reverse\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.urls import reverse_lazy\nfrom django.http import Http404\n\n# Create your views here.\n\n\nclass PostListView(ListView):\n\n model = Posts\n template_name = 'index.html'\n\n def get_queryset(self):\n\n queryset = super(PostListView, self).get_queryset()\n q = self.request.GET.get('query')\n\n if q:\n return queryset.filter(description__icontains=q.strip())\n elif self.kwargs.get('user_id'):\n return queryset.filter(user=self.kwargs['user_id'])\n else:\n return queryset\n\n def get_context_data(self, **kwargs):\n context = super(PostListView, self).get_context_data(**kwargs)\n context['form'] = CommentForm\n return context\n\n\nclass CommentForm(ModelForm):\n\n class Meta:\n model = Comments\n fields = ['author', 'comment']\n\n\nclass PostDetailView(DetailView):\n\n model = Posts\n template_name = 'post.html'\n\n def get_context_data(self, **kwargs):\n context = super(PostDetailView, self).get_context_data(**kwargs)\n context['form'] = CommentForm(initial={'post': self.object.pk}, hide_condition=True)\n context['comments'] = Comments.objects.filter(post=self.object.pk)\n return context\n\n\nclass CommentCreate(CreateView):\n\n model = Comments\n form_class = CommentForm\n template_name = 'index.html'\n\n def form_valid(self, form):\n obj = form.save(commit=False)\n obj.post = Posts.objects.get(pk=self.kwargs['post_id'])\n obj.save()\n return HttpResponseRedirect('/')\n\n def get_context_data(self, **kwargs):\n context = super(CommentCreate, self).get_context_data(**kwargs)\n context['object_list'] = Posts.objects.all()\n return context\n\n def get_success_url(self):\n return reverse('home')\n\n\nclass PostForm(ModelForm):\n\n class Meta:\n model = Posts\n fields = ['photo', 'description']\n\n\nclass NewPostCreate(LoginRequiredMixin, CreateView):\n login_url = 'login'\n redirect_field_name = 'index.html'\n\n model = Posts\n form_class = PostForm\n template_name = 'newpost.html'\n\n def form_valid(self, form):\n obj = form.save(commit=False)\n obj.user = self.request.user\n obj.save()\n return HttpResponseRedirect('/')\n\n def get_context_data(self, **kwargs):\n context = super(NewPostCreate, self).get_context_data(**kwargs)\n context['object_list'] = Posts.objects.all()\n return context\n\n\ndef like_post(request):\n post_id = request.GET.get('post_id', None)\n like = 0\n if post_id:\n post = Posts.objects.get(id=int(post_id))\n if post is not None:\n like = post.like + 1\n post.like = like\n post.save()\n return HttpResponse(like)\n\n\ndef create_comment(request, post_id):\n\n new_comment = CommentForm(data=request.POST)\n response_data = {}\n\n if new_comment.is_valid():\n comment = new_comment.save(commit=False)\n comment.post = Posts.objects.get(id=post_id)\n comment.save()\n\n response_data['author'] = comment.author\n response_data['comment'] = comment.comment\n response_data['datetime'] = comment.datetime\n\n return JsonResponse(response_data)\n\n\nclass PostDelete(LoginRequiredMixin, DeleteView):\n\n model = Posts\n template_name = 'posts_confirm_delete.html'\n success_url = reverse_lazy('home')\n\n def get_object(self, queryset=None):\n\n obj = super(PostDelete, self).get_object()\n if not obj.user == self.request.user:\n raise Exception('Unauthenticated user!!!')\n return obj\n","sub_path":"fotogram_app/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4007,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"263595468","text":"import re\n\nimport connexion\nfrom flask import g\n\n\nfrom swagger_server.configuration.exceptions import Unauthorized\nfrom swagger_server.configuration.signal import unauthorized_hook\n\nchave_autenticacao = \"x-api-key\"\n\n\n\n\n\ndef before_request():\n g.username = \"\"\n\n allowed_urls = [\n \"/v2/ui\",\n \"/v2/ui/\",\n \"/v2/swagger.json\",\n \"/v2/ping\",\n \"/v2/ping/\",\n ]\n\n if connexion.request.method == 'OPTIONS':\n return\n\n print(f\"Verificando se a url precisa de autorizacao\")\n\n for allowed_url in allowed_urls:\n if re.match(allowed_url, connexion.request.path):\n print(f\"Url nao precisa de autorizacao\")\n return\n\n x_api_key = connexion.request.headers.get(\"x-api-key\")\n if x_api_key is None or x_api_key!=chave_autenticacao:\n unauthorized_hook.send()\n raise Unauthorized\n\n\n@unauthorized_hook.connect\ndef sinal_nao_autorizado(identity=None,**kwargs):\n return print(\"Sinal de LOGIN NAO AUTORIZADO enviado\")\n","sub_path":"swagger_server/configuration/context_handler.py","file_name":"context_handler.py","file_ext":"py","file_size_in_byte":1001,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"25590762","text":"import model\r\nimport repository\r\n\r\n\r\ndef insert_bookmark(session):\r\n session.execute(\r\n \"INSERT INTO bookmarks (id, title, url, notes, date_added)\"\r\n ' VALUES (5,\"google5\",\"https://www.google5.com/\",\"google5 website\", \"2021-03-24 03:56:33.961691\")'\r\n )\r\n [[bookmark_url]] = session.execute(\r\n \"SELECT url FROM bookmarks WHERE url=:url \",\r\n dict(url=\"https://www.google5.com/\"),\r\n )\r\n return bookmark_url\r\n\r\n\r\ndef test_repository_can_save_a_bookmark(session):\r\n bookmark = model.Bookmark(6,\"google6\",\"https://www.google6.com/\",\"google6 website\", \"2021-03-24 03:56:33.961691\")\r\n\r\n repo = repository.SqlAlchemyRepository(session)\r\n repo.add(bookmark)\r\n session.commit()\r\n\r\n rows = session.execute(\r\n 'SELECT id, title, url, notes, date_added FROM \"bookmarks\"'\r\n )\r\n assert list(rows) == [(6,\"google6\",\"https://www.google6.com/\",\"google6 website\", \"2021-03-24 03:56:33.961691\")]\r\n\r\n\r\n","sub_path":"barky_flask_1/test_repository.py","file_name":"test_repository.py","file_ext":"py","file_size_in_byte":951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"26210082","text":"#! /usr/bin/env python\n# -*- coding: utf-8 -*-\n# __author__ = \"Q1mi\"\n# Date: 2017/8/25\n\n\"\"\"\njinja2演示示例\n\"\"\"\n\nimport socket\n\nimport pymysql\n\n\ndef f1():\n conn = pymysql.connect(host=\"127.0.0.1\", port=3306, user=\"root\", passwd=\"root1234\", db=\"TEST1\")\n cursor = conn.cursor(cursor=pymysql.cursors.DictCursor)\n cursor.execute(\"select uid, name, department_id from userinfo\")\n user_list = cursor.fetchall()\n cursor.close()\n conn.close()\n print(user_list)\n\n with open(\"jinja2示例.html\") as f:\n source_data = f.read()\n\n # 使用jinja2模板渲染\n from jinja2 import Template\n template = Template(source_data)\n data = template.render(user_list=user_list)\n\n return data\n\n\ndef f2():\n return \"ooo->f2\"\n\n\nrouters = [\n (\"/jinja2\", f1),\n (\"/ooo\", f2),\n]\n\n\ndef run():\n s = socket.socket()\n s.bind((\"127.0.0.1\", 8080))\n s.listen(5)\n while True:\n conn, addr = s.accept() # 暂时挂起\n data = conn.recv(8096)\n data = str(data, encoding=\"utf-8\")\n head, body = data.split(\"\\r\\n\\r\\n\")\n value_list = head.split(\"\\r\\n\")\n method, url, protocal = value_list[0].split(\" \")\n\n conn.send(b\"HTTP/1.1 200 OK\\r\\n\\r\\n\")\n\n func_name = None\n for i in routers:\n if i[0] == url:\n func_name = i[1]\n break # 找到一个不需要往下找了\n if func_name: # 如果有这个函数\n response = func_name()\n else:\n response = \"404\"\n\n conn.send(bytes(response, encoding=\"UTF-8\"))\n conn.close()\n\n\nif __name__ == '__main__':\n run()","sub_path":"about_django/Web框架/jinja2渲染示例.py","file_name":"jinja2渲染示例.py","file_ext":"py","file_size_in_byte":1625,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"396048137","text":"from Midas.databases import mongo_to_df, load_df_to_postgres, MongoInterface\nfrom Midas.configs import default_db, raw_data_collection\n\nimport numpy as np\nimport pandas as pd\nfrom pandas.api.types import is_numeric_dtype\nfrom sklearn import preprocessing\nfrom sklearn.decomposition import PCA\n\n\ndef remove_row_with_missing(df):\n return df.dropna()\n\n\ndef remove_col_with_no_data(df):\n to_drop = []\n for col in df.columns:\n # determine if there's only 1 unique value\n if df[col].nunique() == 1:\n to_drop.append(col)\n\n return df.drop(to_drop, axis=\"columns\")\n\n\ndef impute_numeric(series, strategy):\n def _impute_to_mean(series):\n return series.fillna(value=series.mean())\n\n def _impute_to_median(series):\n return series.fillna(value=series.median())\n\n if strategy == \"mean\":\n return _impute_to_mean(series)\n elif strategy == \"median\":\n return _impute_to_median(series)\n else:\n return None\n\n\ndef impute_categorical(series, strategy):\n def _impute_to_missing(series):\n return series.fillna(value=\"missing\")\n\n if strategy == \"fill_with_missing\":\n return _impute_to_missing(series)\n else:\n return None\n\n\ndef imputation(df, label_mapping, numeric_strategy, categorical_strategy):\n if label_mapping:\n for col in df.columns:\n if col in label_mapping[\"numeric\"]:\n df[col] = impute_numeric(df[col], numeric_strategy)\n elif col in label_mapping[\"categorical\"]:\n df[col] = impute_categorical(df[col], categorical_strategy)\n return df\n\n\ndef clean_training_data(filepath, standardize, outliers, variance_retained, label_mapping, numeric_strategy, categorical_strategy):\n df = pd.read_csv(filepath)\n\n if not numeric_strategy:\n numeric_strategy = \"mean\"\n if not categorical_strategy:\n categorical_strategy = \"fill_with_missing\"\n\n return clean_data(df, label_mapping, numeric_strategy, categorical_strategy, outliers, standardize, variance_retained)\n\ndef clean_data(\n df,\n label_mapping,\n # numeric_strategy=\"mean\",\n # categorical_strategy=\"fill_with_missing\",\n numeric_strategy,\n categorical_strategy,\n outliers=None,\n standardize=None,\n variance_retained=0,\n):\n\n # cleaning process\n\n # force datatypes for categorical data to str\n for features in label_mapping[\"categorical\"]:\n df[features] = df[features].astype(str)\n\n df = remove_outliers(df, outliers)\n\n df = standardize_numeric_features(df, label_mapping[\"numeric\"], standardize)\n\n df = imputation(df, label_mapping, numeric_strategy, categorical_strategy)\n\n # this step breaks the data cleaning process\n # df = dimensionality_reduction_using_PCA(df, variance_retained)\n\n return df\n\n\n# Removes outliers in numeric features outside of (Q1 - 1.5 * IQR, Q3 + 1.5 * IQR)\n# When argument outliers is set to 'obs', the entire row (observation) is removed\n# When argument 'outliers is set to 'value', the outlier value is recoded to missing\n\n# Use: outliers = None obs value\ndef remove_outliers(in_df, outliers):\n in_data = in_df.copy()\n if outliers:\n features = list(in_data)\n for feature in features:\n if is_numeric_dtype(in_data[feature]) and in_data[feature].nunique() > 2:\n # get bounds\n sorted_feature = sorted(in_data[feature])\n q1, q3 = np.percentile(sorted_feature, [25, 75])\n iqr = q3 - q1\n lower_bound = q1 - (1.5 * iqr)\n upper_bound = q3 + (1.5 * iqr)\n if (outliers) == \"obs\":\n in_data.drop(\n in_data[in_data[feature] < lower_bound].index, inplace=True\n )\n in_data.drop(\n in_data[in_data[feature] > upper_bound].index, inplace=True\n )\n elif (outliers) == \"value\":\n in_data.loc[in_data[feature] < lower_bound, feature] = np.nan\n in_data.loc[in_data[feature] > upper_bound, feature] = np.nan\n return in_data\n\n\n# Standardizes all numeric features such that each feature mean = 0 and variance = 1\ndef standardize_numeric_features(in_df, columns, standardize):\n in_data = in_df.copy()\n if standardize:\n scaler = preprocessing.StandardScaler()\n for feature in columns:\n try:\n # we might have removed a column as useless\n if is_numeric_dtype(in_data[feature]) and in_data[feature].nunique() > 2:\n in_data[feature] = scaler.fit_transform(in_data[feature].to_frame())\n except KeyError:\n pass\n return in_data\n\n\n# Uses PCA to reduce the dimensionality of the numeric features. The original\n# numeric features are dropped and replaced by a set of new principal components\n# The number of components selected are the minimum needed to ensure that\n# at least x 'variance explained' is retained. In other words, 30 components might be\n# required to ensure that .95 of the variance in the original set of 100 features is\n# retained\n# NOTE: PCA requires imputed data (no missing)\n\n# Use: variance_retained = .95. 0 means don't use PCA. use original.\n\n# FIXME-Ellen:\n# getting error -> Midas/data_cleaning.py:162:17: F821 undefined name 'nonpca_features_df'\ndef dimensionality_reduction_using_PCA(in_df, variance_retained):\n in_data = in_df.copy()\n if variance_retained > 0:\n features = list(in_data)\n scaler = preprocessing.StandardScaler()\n pca = PCA(variance_retained)\n pca_df = pd.DataFrame(index=in_data.index)\n for feature in features:\n if is_numeric_dtype(in_data[feature]) and in_data[feature].nunique() > 2:\n pca_df[feature] = scaler.fit_transform(in_data[feature].to_frame())\n else:\n nonpca_features_df[feature] = in_data[feature]\n pca.fit(pca_df)\n pca_df = pca.transform(pca_df)\n print(\"number of components = \", pca.n_components_)\n return pd.concat([nonpca_features_df, pca_df], axis=1)\n else:\n return in_data\n\n\n# def suggest_dtypes(collection, db=\"raw_data\"):\n\n# mongo_conn = MongoClient(**mongo_connection_info)\n# df = mongo_to_df(mongo_conn[db], collection)\n# # Do analysis here to suggest type labels for features\n# return {\n\n# }\n\n\n\"\"\" TESTING\n\ntrain_transaction = pd.read_csv('train_transaction.csv', index_col=0)\ntrain_id = pd.read_csv('train_identity.csv', index_col=0)\nin_df = train_transaction.merge(\n train_id, how='left', left_on='TransactionID', right_on='TransactionID')\n\n\n#a = remove_outliers(in_df,'obs')\n#b = remove_outliers(in_df,'values')\n#c = standardize_numeric_features(in_df)\n#d = dimensionality_reduction_using_PCA(in_df, .95)\n\n# initialize list of lists\ndata = [[3, 10, 5],\n [5, 15, 6],\n [2, 99, 3],\n [3, 11, 4],\n [3, 13, 3],\n [5, 14, 5],\n [3, 10, 5],\n [-10, 15, 6],\n [2, 14, 3],\n [3, 11, 4],\n [3, 13, 3],\n [5, 14, 5]\n ]\n\n# Create the pandas DataFrame\ndf = pd.DataFrame(data, columns = ['a', 'b', 'c'])\n\nremove_outliers(df, 'obs')\n\nstandardize_numeric_features(df)\n\ndimensionality_reduction_using_PCA(df, .95)\n\"\"\"\n","sub_path":"Midas/data_cleaning.py","file_name":"data_cleaning.py","file_ext":"py","file_size_in_byte":7308,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"139853487","text":"from __future__ import absolute_import, unicode_literals\n\nimport json\n\nfrom dateutil.relativedelta import relativedelta\nfrom django.contrib.auth.models import User\nfrom django.core.urlresolvers import reverse\nfrom django.test.utils import override_settings\nfrom django.utils import timezone\nfrom mock import patch\nfrom temba.types import Contact as TembaContact, Group as TembaGroup\nfrom tracpro.contacts.models import Contact\nfrom tracpro.groups.models import Region, Group\nfrom tracpro.polls.models import Issue, Response, RESPONSE_COMPLETE, RESPONSE_PARTIAL, RESPONSE_EMPTY\nfrom tracpro.test import TracProTest\n\n\nclass RegionTest(TracProTest):\n def test_create(self):\n zabul = Region.create(self.unicef, \"Zabul\", 'G-101')\n jan = self.create_contact(self.unicef, \"Jan\", 'tel:1234', zabul, self.group1, 'C-101')\n bob = User.create(self.unicef, \"Bob\", \"bob@unicef.org\", \"pass\", False, [zabul])\n\n self.assertEqual(zabul.org, self.unicef)\n self.assertEqual(zabul.name, \"Zabul\")\n self.assertEqual(zabul.uuid, 'G-101')\n self.assertEqual(list(zabul.get_contacts()), [jan])\n self.assertEqual(list(zabul.get_users()), [bob])\n\n def test_get_all(self):\n self.assertEqual(len(Region.get_all(self.unicef)), 3)\n self.assertEqual(len(Region.get_all(self.nyaruka)), 1)\n\n @override_settings(CELERY_ALWAYS_EAGER=True, CELERY_EAGER_PROPAGATES_EXCEPTIONS=True, BROKER_BACKEND='memory')\n @patch('dash.orgs.models.TembaClient.get_groups')\n @patch('dash.orgs.models.TembaClient.get_contacts')\n def test_sync_with_groups(self, mock_get_contacts, mock_get_groups):\n mock_get_groups.return_value = [TembaGroup.create(uuid='G-101', name=\"New region\", size=2),\n TembaGroup.create(uuid='G-102', name=\"Other region\", size=1)]\n mock_get_contacts.return_value = [\n TembaContact.create(uuid='C-101', name=\"Jan\", urns=['tel:123'], groups=['G-101', 'G-005'],\n fields=dict(chat_name=\"jan\"), language='eng', modified_on=timezone.now()),\n TembaContact.create(uuid='C-102', name=\"Ken\", urns=['tel:234'], groups=['G-101', 'G-006'],\n fields=dict(chat_name=\"ken\"), language='eng', modified_on=timezone.now())\n ]\n\n # select one new group\n Region.sync_with_groups(self.unicef, ['G-101'])\n self.assertEqual(self.unicef.regions.filter(is_active=True).count(), 1)\n self.assertEqual(self.unicef.regions.filter(is_active=False).count(), 3) # existing de-activated\n\n new_region = Region.objects.get(uuid='G-101')\n self.assertEqual(new_region.name, \"New region\")\n self.assertTrue(new_region.is_active)\n\n # check contact changes\n self.assertEqual(self.unicef.contacts.filter(is_active=True).count(), 2)\n self.assertEqual(self.unicef.contacts.filter(is_active=False).count(), 5) # existing de-activated\n\n jan = Contact.objects.get(uuid='C-101')\n self.assertEqual(jan.name, \"Jan\")\n self.assertEqual(jan.urn, 'tel:123')\n self.assertEqual(jan.region, new_region)\n self.assertTrue(jan.is_active)\n\n # change group and contacts on chatpro side\n Region.objects.filter(name=\"New region\").update(name=\"Huh?\", is_active=False)\n jan.name = \"Janet\"\n jan.save()\n Contact.objects.filter(name=\"Ken\").update(is_active=False)\n\n # re-select new group\n Region.sync_with_groups(self.unicef, ['G-101'])\n\n # local changes should be overwritten\n self.assertEqual(self.unicef.regions.get(is_active=True).name, 'New region')\n self.assertEqual(self.unicef.contacts.filter(is_active=True).count(), 2)\n Contact.objects.get(name=\"Jan\", is_active=True)\n\n\nclass GroupTest(TracProTest):\n def test_create(self):\n group = Group.create(self.unicef, \"Male Teachers\", 'G-101')\n self.assertEqual(group.org, self.unicef)\n self.assertEqual(group.name, \"Male Teachers\")\n self.assertEqual(group.uuid, 'G-101')\n\n def test_get_all(self):\n self.assertEqual(len(Group.get_all(self.unicef)), 3)\n self.assertEqual(len(Group.get_all(self.nyaruka)), 1)\n\n\nclass RegionCRUDLTest(TracProTest):\n def test_list(self):\n url = reverse('groups.region_list')\n\n # log in as a non-administrator\n self.login(self.user1)\n\n response = self.url_get('unicef', url)\n self.assertRedirects(response, 'http://unicef.localhost/users/login/?next=/region/')\n\n # log in as an administrator\n self.login(self.admin)\n\n response = self.url_get('unicef', url)\n self.assertEqual(response.status_code, 200)\n self.assertEqual(len(response.context['object_list']), 3)\n\n def test_most_active(self):\n url = reverse('groups.region_most_active')\n\n five_weeks_ago = timezone.now() - relativedelta(weeks=5)\n five_days_ago = timezone.now() - relativedelta(days=5)\n issue = Issue.objects.create(poll=self.poll1, conducted_on=five_weeks_ago)\n\n # empty response in last month for contact in region #1\n Response.objects.create(flow_run_id=123, issue=issue, contact=self.contact1,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_EMPTY)\n\n # partial response not in last month for contact in region #2\n Response.objects.create(flow_run_id=234, issue=issue, contact=self.contact4,\n created_on=five_weeks_ago, updated_on=five_weeks_ago, status=RESPONSE_PARTIAL)\n\n # partial response in last month for contact in region #2\n Response.objects.create(flow_run_id=345, issue=issue, contact=self.contact4,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_PARTIAL)\n\n # 2 complete responses in last month for contact in region #3\n Response.objects.create(flow_run_id=456, issue=issue, contact=self.contact5,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_COMPLETE)\n Response.objects.create(flow_run_id=567, issue=issue, contact=self.contact5,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_COMPLETE)\n\n # log in as a non-administrator\n self.login(self.user1)\n\n response = self.url_get('unicef', url)\n results = json.loads(response.content)['results']\n self.assertEqual(len(results), 2)\n self.assertEqual(results[0]['id'], self.region3.pk)\n self.assertEqual(results[0]['name'], self.region3.name)\n self.assertEqual(results[0]['response_count'], 2)\n self.assertEqual(results[1]['id'], self.region2.pk)\n self.assertEqual(results[1]['name'], self.region2.name)\n self.assertEqual(results[1]['response_count'], 1)\n\n\nclass GroupCRUDLTest(TracProTest):\n def test_list(self):\n url = reverse('groups.group_list')\n\n # log in as a non-administrator\n self.login(self.user1)\n\n response = self.url_get('unicef', url)\n self.assertRedirects(response, 'http://unicef.localhost/users/login/?next=/group/')\n\n # log in as an administrator\n self.login(self.admin)\n\n response = self.url_get('unicef', url)\n self.assertEqual(response.status_code, 200)\n self.assertEqual(len(response.context['object_list']), 3)\n\n def test_most_active(self):\n url = reverse('groups.group_most_active')\n\n five_weeks_ago = timezone.now() - relativedelta(weeks=5)\n five_days_ago = timezone.now() - relativedelta(days=5)\n issue = Issue.objects.create(poll=self.poll1, conducted_on=five_weeks_ago)\n\n # empty response in last month for contact in group #1\n Response.objects.create(flow_run_id=123, issue=issue, contact=self.contact1,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_EMPTY)\n\n # partial response not in last month for contact in group #2\n Response.objects.create(flow_run_id=234, issue=issue, contact=self.contact3,\n created_on=five_weeks_ago, updated_on=five_weeks_ago, status=RESPONSE_PARTIAL)\n\n # partial response in last month for contact in group #2\n Response.objects.create(flow_run_id=345, issue=issue, contact=self.contact3,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_PARTIAL)\n\n # 2 complete responses in last month for contact in group #3\n Response.objects.create(flow_run_id=456, issue=issue, contact=self.contact5,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_COMPLETE)\n Response.objects.create(flow_run_id=567, issue=issue, contact=self.contact5,\n created_on=five_days_ago, updated_on=five_days_ago, status=RESPONSE_COMPLETE)\n\n # log in as a non-administrator\n self.login(self.user1)\n\n response = self.url_get('unicef', url)\n results = json.loads(response.content)['results']\n self.assertEqual(len(results), 2)\n self.assertEqual(results[0]['id'], self.group3.pk)\n self.assertEqual(results[0]['name'], self.group3.name)\n self.assertEqual(results[0]['response_count'], 2)\n self.assertEqual(results[1]['id'], self.group2.pk)\n self.assertEqual(results[1]['name'], self.group2.name)\n self.assertEqual(results[1]['response_count'], 1)\n\n\nclass UserRegionsMiddlewareTest(TracProTest):\n def test_process_request(self):\n # make anonymous request to home page\n response = self.url_get('unicef', reverse('home.home'))\n self.assertEqual(response.status_code, 302)\n\n # admin user with implicit access to all regions\n self.login(self.admin)\n\n # default to \"All Regions\"\n response = self.url_get('unicef', reverse('home.home'))\n self.assertIsNone(self.client.session['region'])\n\n # check region menu...\n self.assertContains(response, \"Kandahar\", status_code=200)\n self.assertContains(response, \"Khost\")\n self.assertContains(response, \"Kunar\")\n self.assertContains(response, \"All Regions\")\n\n # should come from session this time\n self.url_get('unicef', reverse('home.home'))\n self.assertIsNone(self.client.session['region'])\n\n # any page allows region to be set via _region param\n self.url_get('unicef', reverse('home.home'), {'_region': self.region3.pk})\n self.assertEqual(self.client.session['region'], self.region3.pk)\n\n # can set to region to 0 meaning \"All Regions\"\n self.url_get('unicef', reverse('home.home'), {'_region': 0})\n self.assertIsNone(self.client.session['region'])\n\n # user with access to 2 regions (#2 and #3)\n self.login(self.user2)\n\n # default to first region A-Z\n response = self.url_get('unicef', reverse('home.home'))\n self.assertEqual(self.client.session['region'], self.region2.pk)\n\n # check region menu...\n self.assertContains(response, \"Khost\", status_code=200)\n self.assertContains(response, \"Kunar\")\n self.assertNotContains(response, \"All Regions\")\n\n # can't set to region that user doesn't have access, so defaults back to first\n self.url_get('unicef', reverse('home.home'), {'_region': self.region1.pk})\n self.assertEqual(self.client.session['region'], self.region2.pk)\n\n # user with access to only 1 region\n self.login(self.user1)\n\n # user only has access to region #1 so should default to region #1\n response = self.url_get('unicef', reverse('home.home'))\n self.assertEqual(self.client.session['region'], self.region1.pk)\n\n # no region menu, just region name\n self.assertContains(response, \"Kandahar\", status_code=200)\n self.assertNotContains(response, \"All Regions\")\n","sub_path":"tracpro/groups/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":12019,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"454397992","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build\\bdist.win-amd64\\egg\\djutil\\context_processors.py\n# Compiled at: 2013-08-27 09:18:22\nfrom __future__ import unicode_literals\nfrom django.conf import settings\nfrom django.utils import dateformat\n\ndef analytics(request):\n return {b'ga_tracking_id': getattr(settings, b'GA_TRACKING_ID', b''), \n b'ga_tracking_domain': getattr(settings, b'GA_TRACKING_DOMAIN', b'')}\n\n\ndef version_string(request):\n revision_hash = getattr(settings, b'REVISION_HASH', b'')\n revision_date = getattr(settings, b'REVISION_DATE', b'')\n if revision_date:\n try:\n revision_date = dateformat.format(revision_date, b'j M, G:i')\n except:\n revision_date = b''\n\n revision_env = getattr(settings, b'REVISION_ENV', b'')\n if revision_env:\n revision_env = b', ' + revision_env\n if revision_hash and revision_date:\n version = (b'v. {} ({}{})').format(revision_date, revision_hash, revision_env)\n else:\n version = b''\n return {b'VERSION_STRING': version}","sub_path":"pycfiles/djutil-0.2-py2.7/context_processors.py","file_name":"context_processors.py","file_ext":"py","file_size_in_byte":1168,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"650720202","text":"import urllib2\nimport csv\nimport time\nimport random_stock_tickers\nimport controller\n\nclass YahooFinanceAPIHandler():\n\n def __init__(self, stock_tickers_list, stats_wanted_list):\n self.stock_tickers_list = stock_tickers_list\n self.stats_wanted_list = stats_wanted_list\n self.stat_dict = {\"ask\":\"a\",\n \"average_daily_volume\":\"a2\",\n \"bid\":\"b\",\n \"previous_close\":\"p\",\n \"open\":\"o\",\n \"day_low\":\"g\",\n \"day_high\":\"h\",\n \"day_value_change\":\"w4\",\n \"last_trade\":\"l1\",\n \"last_trade_date\":\"d1\",\n \"last_trade_time\":\"t1\",\n \"52_week_high\":\"k\",\n \"52_week_low\":\"j\",\n \"52_week_range\":\"w\",\n \"dividend_yield\":\"y\",\n \"dividend_per_share\":\"d\",\n \"dividend_pay_date\":\"r1\",\n \"ex_dividend_date\":\"q\",\n \"market_cap\":\"j1\",\n \"name\":\"n\",\n \"revenue\":\"s6\",\n \"symbol\":\"s\",\n \"stock_exchange\":\"x\",\n \"shares_outstanding\":\"j2\",\n \"eps\":\"e\",\n \"ebitda\":\"j4\",\n \"pe\":\"r\"\n }\n\n def create_url(self):\n stock_tickers = self.stock_tickers_list\n stats_wanted_list = self.stats_wanted_list\n stat_dict = self.stat_dict\n url = \"http://finance.yahoo.com/d/quotes.csv?s=\"\n for ticker in stock_tickers:\n if ticker != stock_tickers[-1]:\n url += ticker + \"+\"\n else:\n url += ticker + \"&f=\"\n for stat in stats_wanted_list:\n url += stat_dict[stat]\n return url\n\n def get_unvalidated_stock_table(self):\n url = self.create_url()\n table = []\n while controller.Controller().can_connect_to_url(url) == False:\n time.sleep(60)\n response = urllib2.urlopen(url)\n document = csv.reader(response)\n for row in document:\n table.append(row)\n response.close()\n return table\n\n #this function invalidates a row if ALL entries == \"N/A\"\n \"\"\"def create_validated_table(self):\n unvalidated_table = self.get_unvalidated_stock_table()\n invalid_rows = []\n modified_table = []\n for row in unvalidated_table:\n match_count = 0\n for entry in row:\n if entry == \"N/A\":\n match_count += 1\n if match_count == len(row) - 1:\n invalid_rows.append(row)\n for row in unvalidated_table:\n if row not in invalid_rows:\n modified_table.append(row)\n return modified_table\"\"\"\n\n #this functions invalidates a row if ANY of the entries == \"N/A\"\n def create_validated_table(self):\n unvalidated_table = self.get_unvalidated_stock_table()\n invalid_rows = []\n modified_table = []\n for row in unvalidated_table:\n for entry in row:\n if entry == \"N/A\":\n invalid_rows.append(row)\n for row in unvalidated_table:\n if row not in invalid_rows:\n modified_table.append(row)\n return modified_table\n\n #create_formatted_data returns stock data in a readable format\n def create_formatted_data(self):\n validated_table = self.create_validated_table()\n stats_wanted = self.stats_wanted_list\n text = \"\"\n for row in validated_table:\n for i in range(0, len(stats_wanted)):\n text += stats_wanted[i].upper().replace(\"_\",\" \") + \": \" + row[i].upper() + \"\\n\"\n text += \"\\n\"\n return text\n","sub_path":"yahoo_finance_api_handler.py","file_name":"yahoo_finance_api_handler.py","file_ext":"py","file_size_in_byte":3926,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"554921848","text":"import uasyncio as asyncio\nfrom uasyncio import Event\n\nasync def waiter(event):\n print('Waiting for event')\n await event.wait() # Pause here until event is set\n print('Waiter got event.')\n event.clear() # Flag caller and enable re-use of the event\n\nasync def main():\n event = Event()\n asyncio.create_task(waiter(event))\n await asyncio.sleep(2)\n print('Setting event')\n event.set()\n await asyncio.sleep(1)\n # Caller can check if event has been cleared\n print('Event is {}'.format('set' if event.is_set() else 'clear'))\n\nasyncio.run(main())","sub_path":"Experimental_exercise/uasyncio_test/3.2 Event.py","file_name":"3.2 Event.py","file_ext":"py","file_size_in_byte":576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"135211199","text":"import threading\nimport queue\nfrom CreateBluePrint import CreateBluePrint\nfrom RequestCockpitAPI import RequestCockpitAPI\nfrom cockpit_testing.Framework.utils.utils import BaseTest\nimport time, traceback, sys\nimport argparse\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('-b', help='Run list of blueprints', dest='bpName', action='store', default=[], nargs='+')\n parser.add_argument('-d', help='Run blueprints in a specific directory', dest='bpDirectory', default='', action='store')\n parser.add_argument('-a', help='Use a specific account', dest='account', default='', action='store')\n parser.add_argument('--clone', help=\"Clone blueprints templates' repo\", dest='clone', default=False, action='store_true')\n parser.add_argument('--teardown', help='Delete the account', dest='teardown', default=False, action='store_true')\n parser.add_argument('--no-backend', help='No backend environment', dest='no_backend', default=False,\n action='store_true')\n options = parser.parse_args()\n\n print(' [*] Driver is running ..... ')\n base_test = BaseTest()\n base_test.log('Driver.log')\n base_test.check_cockpit_is_exist()\n\n THREADS_NUMBER = int(base_test.values['threads_number'])\n BLUEPRINT_NAME = options.bpName\n\n create_blueprint = CreateBluePrint(clone=options.clone, no_backend=options.no_backend, bp_dir=options.bpDirectory)\n if options.account:\n create_blueprint.values['account'] = options.account\n elif not options.no_backend:\n create_blueprint.create_account()\n\n create_blueprint.create_blueprint()\n role = {} # {'Thread_name : [ [role_name, service], [role_name, service], .. ]}\n\n\n def get_testService_role(blueprint, thread_name):\n global role\n index = []\n blueprint = blueprint.splitlines()\n for line in blueprint:\n if 'test_' == line[:5]:\n index.append(blueprint.index(line))\n\n if len(index) == 0:\n raise NameError(\"The blueprint doesn't have any test roles.\")\n\n role[thread_name] = []\n for i in index:\n role_line = blueprint[i]\n role_name = role_line[:role_line.find('__')]\n role_service = role_line[role_line.find('__') + 2:-1]\n role[thread_name].append([role_name, role_service])\n\n\n queue = queue.Queue()\n jobs = base_test.get_jobs(BLUEPRINT_NAME)\n for job in jobs:\n queue.put(job)\n\n\n def work():\n while not queue.empty():\n testCasesPath = queue.get()\n bpFileName = testCasesPath[testCasesPath.index('/TestCases/') + 11:]\n print(('\\n [*] Test case : %s' % bpFileName))\n base_test.logging.info('\\n')\n base_test.logging.info('[*] Test case : %s' % bpFileName)\n\n try:\n blueprint = create_blueprint.load_blueprint(testCasesPath=testCasesPath)\n get_testService_role(blueprint=blueprint, thread_name=threading.current_thread().name)\n request_cockpit_api = RequestCockpitAPI()\n request_cockpit_api.create_new_repository(repository=request_cockpit_api.repo['name'])\n request_cockpit_api.send_blueprint(repository=request_cockpit_api.repo['name'],\n blueprint=blueprint)\n\n request_cockpit_api.execute_blueprint(repository=request_cockpit_api.repo['name'],\n blueprint=request_cockpit_api.blueprint['name'])\n request_cockpit_api.run_repository(repository=request_cockpit_api.repo['name'])\n\n testCase_time = request_cockpit_api.get_run_status(repository=request_cockpit_api.repo['name'],\n run_key=request_cockpit_api.repo['key'],\n bpFileName=bpFileName)\n if testCase_time:\n base_test.Testcases_results[bpFileName] = []\n base_test.Testcases_results[bpFileName].append(['TestCase Time', testCase_time])\n for role_item in role[threading.current_thread().name]:\n base_test.Testcases_results[\n bpFileName].append(request_cockpit_api.get_service_data(\n repository=request_cockpit_api.repo['name'],\n role=role_item[0],\n service=role_item[1]))\n else:\n request_cockpit_api.testcase_time = '{:0.2f}'.format(time.time() - request_cockpit_api.start_time)\n error_message = 'ERROR : %s %s' % (\n request_cockpit_api.blueprint['name'], request_cockpit_api.blueprint['log'])\n base_test.Testcases_results[bpFileName] = [['TestCase Time', request_cockpit_api.testcase_time],\n [error_message, role[threading.current_thread().name][0]]]\n except:\n base_test.logging.error(traceback.format_exc())\n\n # Add error message to xml result\n error_message = 'ERROR : %s %s' % (traceback.format_exc(), request_cockpit_api.response_error_content)\n base_test.Testcases_results[bpFileName] = [['TestCase Time', 0], [error_message, 'Unknown service']]\n\n request_cockpit_api.clean_cockpit()\n queue.task_done()\n\n\n for _ in range(THREADS_NUMBER):\n threading.Thread(target=work).start()\n\n queue.join()\n base_test.generate_xml_results()\n if options.teardown:\n create_blueprint.teardown()\n","sub_path":"cockpit_testing/Framework/Driver/Driver.py","file_name":"Driver.py","file_ext":"py","file_size_in_byte":5731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"470530314","text":"import gym\nimport os\nfrom time import sleep\nfrom evolutionary_algorithm.ea.eamain import ConvolutionalNeuralNetwork, SimpleNeuralNetwork\nfrom evolutionary_algorithm.ea.gym_wrapper import RamGymWrapper, MainGymWrapper\nENV_NAME = \"SpaceInvaders-ramNoFrameskip-v4\"\nRAM = True\nMODEL_USED = \"SIMPLE\"\nLOAD_WEIGHTS_PATH = str(os.path.dirname(__file__) + \"/models/\" + ENV_NAME + \"/\" + \"2020-02-20_04-35\" + \"-model.h5\")\n\n\nif RAM:\n env = RamGymWrapper.wrap(gym.make(ENV_NAME))\nelse:\n env = MainGymWrapper.wrap(gym.make(ENV_NAME))\n\nif MODEL_USED == \"SIMPLE\":\n model = SimpleNeuralNetwork((4, 128), env.action_space.n, filepath=LOAD_WEIGHTS_PATH)\nelse:\n model = ConvolutionalNeuralNetwork((4, 84, 84), env.action_space.n, filepath=LOAD_WEIGHTS_PATH)\n\nterminated = False\nstate = env.reset()\ntotal_reward = 0\n\nwhile not terminated:\n action = model.predict(state)\n state, reward, terminated,_ = env.step(action)\n total_reward += reward\n env.render()\n sleep(0.01)\n\nprint(\"Final reward: \" + str(total_reward))\n","sub_path":"evolutionary_algorithm/ea/play.py","file_name":"play.py","file_ext":"py","file_size_in_byte":1020,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"274700055","text":"import voltage\r\nimport asyncio\r\nimport json\r\nimport datetime\r\nimport time\r\n\r\nfrom voltage.ext import commands\r\n\r\n\r\ndef setup(client) -> commands.Cog:\r\n\r\n owner = commands.Cog(\"Owner\")\r\n\r\n\r\n @owner.command(description=\"Change the presence or status of Ikusei!\")\r\n async def status(ctx, *, status, presence=None):\r\n if ctx.author.id in [\r\n \"01H2YMK1VSZM0X4GJAAC7X48TN\",\r\n \"01H213M511VS2W8HMAHHVCSBPT\",\r\n ]:\r\n if not presence:\r\n await client.set_status(status, voltage.PresenceType.online)\r\n return await ctx.send(f\"Changed status to `{status}`\")\r\n else:\r\n if presence.lower() == \"online\":\r\n await client.set_status(status, voltage.PresenceType.online)\r\n return await ctx.send(\r\n f\"Changed status to `{status}` and a presence of `Online!`\"\r\n )\r\n elif presence.lower() == \"idle\":\r\n await client.set_status(status, voltage.PresenceType.idle)\r\n return await ctx.send(\r\n f\"Changed status to `{status}` and a presence of `Idle`!\"\r\n )\r\n elif presence.lower() == \"dnd\" or \"busy\":\r\n await client.set_status(status, voltage.PresenceType.busy)\r\n return await ctx.send(\r\n f\"Changed status to `{status}` and a presence of `Do Not Disturb`!\"\r\n )\r\n else:\r\n return await ctx.send(\"You aren't an owner of this bot!\")\r\n\r\n @owner.command()\r\n async def test(ctx):\r\n embed = voltage.ImageEmbed(url=\"https://i.imgur.com/3LljFXC.jpeg\")\r\n await ctx.send(content=\"[]()\", embed=embed)\r\n\r\n @owner.command(description=\"Test our command\")\r\n async def register(ctx):\r\n\r\n with open(\"json/users.json\", \"r\") as f:\r\n data = json.load(f)\r\n if ctx.author.id in data:\r\n return await ctx.send(\"You're already registered!\")\r\n with open(\"json/users.json\", \"w\") as f:\r\n data[ctx.author.id] = {\r\n \"username\": ctx.author.name,\r\n \"id\": ctx.author.id,\r\n \"bio\": \"User has no bio set!\",\r\n \"beta\": \"False\",\r\n \"ff\": \"False\",\r\n \"notifications\": []\r\n }\r\n json.dump(data, f, indent=2)\r\n embed = voltage.SendableEmbed(description=\"## You're registered!\")\r\n await ctx.send(content=\"[]()\", embed=embed)\r\n\r\n\r\n @owner.command(description=\"Use this after registering\")\r\n async def ar(ctx):\r\n with open(\"json/users.json\", \"r\") as f:\r\n data = json.load(f)\r\n embed = voltage.SendableEmbed(\r\n description=f\"{data[ctx.author.id]['username']}'s profile:\\n\\n**Bio:**\\n{data[ctx.author.id]['bio']}\\n\\n**User's settings:**\\n\\nBeta: `{data[ctx.author.id]['beta']}`\\nFamily Friendly Mode: `{data[ctx.author.id]['ff']}`\"\r\n )\r\n await ctx.send(content=\"[]()\", embed=embed)\r\n\r\n return owner","sub_path":"cogs/owner.py","file_name":"owner.py","file_ext":"py","file_size_in_byte":3080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"158128162","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\n@date: 2020/12/4 上午11:39\n@file: mobilenetv2_inverted_residual.py\n@author: zj\n@description: MobileNetV2 反向残差块\n\"\"\"\nfrom abc import ABC\n\nimport torch.nn as nn\n\n\nclass MobileNetV2InvertedResidual(nn.Module, ABC):\n \"\"\"\n MobileNetV2的反向残差块由一个膨胀卷积和一个深度可分离卷积(depth-wise conv + point-wise conv)组成\n 参考torchvision实现:\n 1. 当膨胀率大于1时,执行膨胀卷积操作;\n 2. 当深度卷积步长为1且输入/输出通道数相同时,执行残差连接\n 3. 反向残差块的最后不执行激活操作\n \"\"\"\n\n def __init__(self,\n # 输入通道数\n in_planes,\n # 输出通道数\n out_planes,\n # 膨胀因子\n expansion_rate=1,\n # 卷积层步长\n stride=1,\n # 卷积层零填充\n padding=1,\n # 卷积层类型\n conv_layer=None,\n # 归一化层类型\n norm_layer=None,\n # 激活层类型\n act_layer=None,\n ):\n super(MobileNetV2InvertedResidual, self).__init__()\n\n if conv_layer is None:\n conv_layer = nn.Conv2d\n if norm_layer is None:\n norm_layer = nn.BatchNorm2d\n if act_layer is None:\n act_layer = nn.ReLU6\n\n # 计算隐藏层输入通道数\n hidden_planes = int(expansion_rate * in_planes)\n features = list()\n if expansion_rate != 1:\n features.append(nn.Sequential(\n conv_layer(in_planes, hidden_planes, kernel_size=1, stride=1, bias=False),\n norm_layer(hidden_planes),\n act_layer(inplace=True)\n ))\n\n # 深度卷积\n features.append(nn.Sequential(\n conv_layer(hidden_planes, hidden_planes, kernel_size=3, stride=stride, padding=padding, bias=False,\n groups=hidden_planes),\n norm_layer(hidden_planes),\n act_layer(inplace=True)\n ))\n\n # 逐点卷积\n features.append(nn.Sequential(\n conv_layer(hidden_planes, out_planes, kernel_size=1, stride=1, bias=False),\n norm_layer(out_planes)\n ))\n\n self.conv = nn.Sequential(*features)\n self.use_res_connect = stride == 1 and in_planes == out_planes\n\n def forward(self, x):\n if self.use_res_connect:\n return x + self.conv(x)\n else:\n return self.conv(x)\n","sub_path":"zcls/model/backbones/mobilenetv2_inverted_residual.py","file_name":"mobilenetv2_inverted_residual.py","file_ext":"py","file_size_in_byte":2611,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"466453153","text":"import os\nos.environ[\"CUDA_DEVICE_ORDER\"]=\"PCI_BUS_ID\"\nos.environ[\"CUDA_VISIBLE_DEVICES\"]=\"0\"\n\nimport pyglet\nfrom keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.optimizers import Adam\n\nfrom common.gym_runner import GymRunner\nfrom common.q_learning_agent import QLearningAgent\n\n\nclass CartPoleAgent(QLearningAgent):\n def __init__(self):\n super(CartPoleAgent, self).__init__(2, 3, maxlen = 100000)\n\n def build_model(self):\n model = Sequential()\n model.add(Dense(30, activation='relu', input_dim=2))\n model.add(Dense(30, activation='relu'))\n model.add(Dense(3))\n model.compile(Adam(lr=0.001), 'mse')\n\n # load the weights of the model if reusing previous training session\n # model.load_weights(\"models/cartpole-v0.h5\")\n return model\n\n\nif __name__ == \"__main__\":\n gym = GymRunner('MountainCar-v0', 'gymresults/mountaincar-v0', tile_coding=True)\n agent = CartPoleAgent()\n gym.train(agent, 3000)\n agent.model.save_weights(\"models/mountaincar-v0.h5\", overwrite=True)\n gym.run(agent, 500)\n gym.env.close()\n # gym.close_and_upload(os.environ['API_KEY'])\n","sub_path":"cartpole/mountain_v0.py","file_name":"mountain_v0.py","file_ext":"py","file_size_in_byte":1160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"439049288","text":"# -*- coding: utf-8 -*-\n\nimport repository\nimport sqlite3\nimport unittest\n\ndb_path = 'temp_hist.db'\n\nclass RepositoryTest(unittest.TestCase):\n\n def setUp(self):\n conn = sqlite3.connect(db_path)\n c = conn.cursor()\n c.execute('DELETE FROM czujnik')\n c.execute('DELETE FROM pomiar')\n c.execute('''INSERT INTO czujnik (id, serial_number, lokalizacja) VALUES(1, 1, 'Pok 11')''')\n c.execute('''INSERT INTO pomiar (id, czujnik_id, pomiar, data) VALUES(1, 1, 20, '2015-11-10')''')\n c.execute('''INSERT INTO pomiar (id, czujnik_id, pomiar, data) VALUES(2, 1, 10, '2015-12-30')''')\n conn.commit()\n conn.close()\n\n def tearDown(self):\n conn = sqlite3.connect(db_path)\n c = conn.cursor()\n c.execute('DELETE FROM czujnik')\n c.execute('DELETE FROM pomiar')\n conn.commit()\n conn.close()\n\n def testGetByIdCzujnik(self):\n czujnik = repository.PomiarRepository().getByCzujnik(1)\n self.assertIsInstance(czujnik, repository.Czujnik, \"Objekt nie jest klasy Czujnik\")\n\n def testGetByIdCzujnikNotFound(self):\n self.assertEqual(repository.PomiarRepository().getByCzujnik(22),\n None, \"Powinno wyjść None\")\n\n def testGetBysrednia(self):\n self.assertEqual(repository.PomiarRepository().sredniaCzujnika(1),15, \"Powinno wyjść 15\")\n\n def testDeleteNotFound(self):\n self.assertRaises(repository.RepositoryException,repository.PomiarRepository().delete,22)\n\n\n\nif __name__ == \"__main__\":\n unittest.main()\n","sub_path":"repositorytest.py","file_name":"repositorytest.py","file_ext":"py","file_size_in_byte":1551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"580362469","text":"import argparse\n\nimport torch.backends.cudnn as cudnn\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nimport torch.utils.data\nfrom torch.utils.data import DataLoader, random_split\nfrom torch.utils.tensorboard import SummaryWriter\nfrom tqdm import tqdm\n\nfrom dataset import Remote\nfrom unet.unet_model import UNet\n\n\nclass Solver(object):\n def __init__(self, config):\n self.model = None\n self.lr = config.lr\n self.epochs = config.epoch\n self.train_batch_size = config.train_batch_size\n self.val_batch_size = config.val_batch_size\n self.criterion = None\n self.optimizer = None\n self.scheduler = None\n self.device = None\n self.cuda = config.cuda\n self.val_percent = 0.1\n self.train_loader = None\n self.val_loader = None\n\n def load_data(self):\n dataset = Remote()\n n_val = int(len(dataset) * self.val_percent)\n n_train = len(dataset) - n_val\n train, val = random_split(dataset, [n_train, n_val])\n self.train_loader = DataLoader(train, batch_size=self.train_batch_size, shuffle=True)\n self.val_loader = DataLoader(val, batch_size=self.val_batch_size, shuffle=False)\n\n def load_model(self):\n if self.cuda:\n self.device = torch.device('cuda')\n cudnn.benchmark = True\n else:\n self.device = torch.device('cpu')\n self.model = UNet(n_channels=3, n_classes=11).to(self.device)\n\n self.optimizer = optim.Adam(self.model.parameters(), lr=self.lr, weight_decay=1e-8)\n # self.scheduler = optim.lr_scheduler.MultiStepLR(self.optimizer, milestones=[75, 150], gamma=0.5)\n self.criterion = nn.CrossEntropyLoss().to(self.device)\n\n def train(self):\n self.model.train()\n loss_tol = 0\n n_total = len(self.train_loader) * 512 * 512 * self.train_batch_size\n n_correct = 0\n pbar = tqdm(total=len(self.train_loader) * self.train_batch_size, unit='img')\n for i, (imgs, masks) in enumerate(self.train_loader):\n imgs = imgs.to(device=self.device, dtype=torch.float32)\n masks = masks.to(device=self.device, dtype=torch.long)\n\n output = self.model(imgs)\n loss = self.criterion(output, masks)\n loss_tol += loss.item()\n\n self.optimizer.zero_grad()\n loss.backward()\n self.optimizer.step()\n\n pred = torch.max(output, 1)[1]\n correct = torch.sum(pred == masks).item()\n n_correct += correct\n\n pbar.set_postfix(**{'train_loss': loss.item() / self.train_batch_size,\n 'train_acc': correct / (self.train_batch_size * 512 * 512)})\n pbar.update(imgs.shape[0])\n pbar.close()\n return loss_tol / len(self.train_loader), n_correct / n_total\n\n def val(self):\n self.model.eval()\n loss_tol = 0\n n_correct = 0\n n_total = len(self.val_loader) * 512 * 512 * self.val_batch_size\n pbar = tqdm(total=len(self.val_loader), desc='val', unit='img')\n with torch.no_grad():\n for i, (imgs, masks) in enumerate(self.val_loader):\n imgs = imgs.to(device=self.device, dtype=torch.float32)\n masks = masks.to(device=self.device, dtype=torch.long)\n\n output = self.model(imgs)\n loss = self.criterion(output, masks)\n loss_tol += loss.item()\n\n pred = torch.max(output, 1)[1]\n n_correct += torch.sum(pred == masks).item()\n\n pbar.update(imgs.shape[0])\n pbar.close()\n return loss_tol / len(self.val_loader), n_correct / n_total\n\n def run(self):\n self.load_data()\n self.load_model()\n accuracy = 0\n writer = SummaryWriter()\n try:\n for epoch in range(self.epochs):\n train_result = self.train()\n # self.scheduler.step(epoch)\n val_result = self.val()\n writer.add_scalar('Loss/train', train_result[0], epoch)\n writer.add_scalar('Acc/train', train_result[1], epoch)\n writer.add_scalar('Loss/val', val_result[0], epoch)\n writer.add_scalar('Acc/val', val_result[1], epoch)\n accuracy = max(accuracy, val_result[1])\n print(f'epoch: {epoch + 1} / {self.epochs} train_loss: {train_result[0]} train_acc: {train_result[1]} '\n f'val_loss: {val_result[0]} val_acc: {val_result[1]}')\n writer.close()\n except KeyboardInterrupt:\n torch.save(self.model.state_dict(), 'INTERRUPTED.pth')\n writer.close()\n torch.save(self.model.state_dict(), 'model.pth')\n\n\ndef main():\n parser = argparse.ArgumentParser(description=\"cifar-10 with PyTorch\")\n parser.add_argument('--lr', default=0.1, type=float, help='learning rate')\n parser.add_argument('--epoch', default=20, type=int, help='number of epochs tp train for')\n parser.add_argument('--train_batch_size', default=2, type=int, help='training batch size')\n parser.add_argument('--val_batch_size', default=1, type=int, help='testing batch size')\n parser.add_argument('--cuda', default=torch.cuda.is_available(), type=bool, help='whether cuda is in use')\n args = parser.parse_args()\n\n solver = Solver(args)\n solver.run()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":5452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"182987137","text":"# Copyright 2016 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport contextlib\nimport socket\n\nfrom selenium import webdriver\nfrom selenium.webdriver.common import by\nfrom selenium.webdriver.common import proxy\nimport xvfbwrapper\n\nfrom stacklight_tests.helpers.ui import ui_settings\n\n\n@contextlib.contextmanager\ndef ui_driver(url, title, wait_element='/html'):\n vdisplay = None\n # Start a virtual display server for running the tests headless.\n if ui_settings.headless_mode:\n vdisplay = xvfbwrapper.Xvfb(width=1920, height=1080)\n args = []\n\n # workaround for memory leak in Xvfb taken from:\n # http://blog.jeffterrace.com/2012/07/xvfb-memory-leak-workaround.html\n args.append(\"-noreset\")\n\n # disables X access control\n args.append(\"-ac\")\n\n if hasattr(vdisplay, 'extra_xvfb_args'):\n # xvfbwrapper 0.2.8 or newer\n vdisplay.extra_xvfb_args.extend(args)\n else:\n vdisplay.xvfb_cmd.extend(args)\n vdisplay.start()\n driver = get_driver(url, wait_element, title)\n try:\n yield driver\n finally:\n driver.quit()\n if vdisplay is not None:\n vdisplay.stop()\n\n\ndef get_driver(url, anchor, title, by_selector_type=by.By.XPATH):\n proxy_address = ui_settings.proxy_address\n # Increase the default Python socket timeout from nothing\n # to something that will cope with slow webdriver startup times.\n # This *just* affects the communication between this test process\n # and the webdriver.\n socket.setdefaulttimeout(60)\n # Start the Selenium webdriver and setup configuration.\n proxy_ex = None\n if proxy_address is not None:\n proxy_ex = proxy.Proxy(\n {\n 'proxyType': proxy.ProxyType.MANUAL,\n 'socksProxy': proxy_address,\n }\n )\n driver = webdriver.Firefox(proxy=proxy_ex)\n if ui_settings.maximize_window:\n driver.maximize_window()\n driver.implicitly_wait(ui_settings.implicit_wait)\n driver.set_page_load_timeout(ui_settings.page_timeout)\n driver.get(url)\n driver.find_element(by_selector_type, anchor)\n assert title in driver.title\n return driver\n\n\ndef get_table(driver, xpath, frame=None):\n if frame:\n driver.switch_to.default_content()\n driver.switch_to.frame(driver.find_element_by_name(frame))\n return driver.find_element_by_xpath(xpath)\n\n\ndef get_table_row(table, row_id):\n return table.find_element_by_xpath(\"tr[{0}]\".format(row_id))\n\n\ndef get_table_size(table):\n return len(table.find_elements_by_xpath(\"tr[position() > 0]\"))\n\n\ndef get_table_cell(table, row_id, column_id):\n row = get_table_row(table, row_id)\n return row.find_element_by_xpath(\"td[{0}]\".format(column_id))\n","sub_path":"stacklight_tests/helpers/ui_tester.py","file_name":"ui_tester.py","file_ext":"py","file_size_in_byte":3313,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"459210969","text":"\nimport numpy as np\n\n\nfrom ml.classifiers.dt.node import Node, LeafNode\nfrom ml.classifiers.dt.cost import gini\n\n\nclass DecisionTree:\n\n def __init__(self, max_depth = 10, min_size = 5, cost = gini, f_count = None):\n self.root = None\n self.max_depth = max_depth\n self.min_size = min_size\n self.cost = cost\n self.f_count = f_count\n\n\n def __str__(self):\n return self.root.to_string()\n\n\n def build_tree(self, X, y, features, depth = 1):\n best_cost = 999999\n best_criteria = None\n best_split = None\n best_labels = None\n\n for feature in features:\n\n for value in np.unique(X.loc[:, feature]):\n if isinstance(value, int) or isinstance(value, float):\n left, right = X.loc[:, feature] < value, X.loc[:, feature] >= value\n else:\n left, right = X.loc[:, feature] == value, X.loc[:, feature] != value\n\n y_left, y_right = y[left], y[right]\n\n if y_left.shape[0] * y_right.shape[0] > 0:\n cost = self.cost([y_left, y_right], np.unique(y))\n\n if best_cost > cost:\n best_cost = cost\n best_criteria = (feature, value)\n best_split = (X.loc[left, :], X.loc[right, :])\n best_labels = (y_left, y_right)\n\n if not best_labels or depth > self.max_depth or len(best_labels[0]) < self.min_size or len(best_labels[1]) < self.min_size:\n return LeafNode(y)\n else:\n return Node(\n best_criteria, \n self.build_tree(best_split[0], best_labels[0], features, depth + 1), \n self.build_tree(best_split[1], best_labels[1], features, depth + 1)\n )\n\n\n def fit(self, X, y):\n if self.f_count:\n features = np.random.choice(list(X), self.f_count, replace = False)\n else:\n features = list(X)\n\n self.root = self.build_tree(X, y, features)\n\n\n def predict(self, X):\n return [self.root.predict(row) for _, row in X.iterrows()]\n \n","sub_path":"ml/classifiers/dt/tree.py","file_name":"tree.py","file_ext":"py","file_size_in_byte":2192,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"397712552","text":"#!/usr/bin/env python\r\n# encoding: utf-8\r\n\r\nimport math\r\n\"\"\"\r\n@version: ??\r\n@author: autmanli\r\n@license: Apache Licence \r\n@file: sushu.py\r\n@time: 2018/10/12 13:29\r\n\"\"\"\r\n\r\n# 验证该数是不是素数,只需要验证该数能够整除从2到他的平方根即可\r\ndef check_sushu(num):\r\n # 求该数的平方根\r\n sqrt = math.floor(math.sqrt(num)) + 1\r\n # 循环从2到平方根的数\r\n for i in range(2, sqrt):\r\n # 如果该数能够整除,说明该数不是素数\r\n if num % i == 0:\r\n return False\r\n return True\r\n\r\n\r\n#filter用法,过滤掉数组中返回值为false的项\r\nsushu_arr=filter(check_sushu,range(1,1000000))\r\n\r\nprint(list(sushu_arr))","sub_path":"sushu.py","file_name":"sushu.py","file_ext":"py","file_size_in_byte":694,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"530327664","text":"from conn import Conn\nfrom flask import jsonify, request\n\n\nclass Tasks:\n def __init__(self):\n self.c = Conn()\n \n def insert_new_task(self):\n \"\"\"This method inserts new task into the database\"\"\" \n title = request.json['title']\n description = request.json['description']\n done = request.json['done']\n self.c.cursor.execute(\n \"\"\"\n INSERT INTO tasks(title, description, done) VALUES('{}', '{}', '{}')\n \"\"\"\n .format(title, description, done))\n self.c.conn.commit()\n self.c.conn.close()\n\n\nif __name__ == \"__main__\":\n task = Tasks()\n task.insert_new_task()\n","sub_path":"package/tasks.py","file_name":"tasks.py","file_ext":"py","file_size_in_byte":672,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"446519046","text":"import openface\nimport cv2\nimport numpy as np \nimport os\nfrom sklearn import svm\n\nfileDir = os.path.dirname(os.path.realpath(__file__))\nmodelDir = os.path.join(fileDir, 'models')\ndlibModelDir = os.path.join(modelDir, 'dlib')\nopenfaceModelDir = os.path.join(modelDir, 'openface')\n\nalign = openface.AlignDlib(os.path.join(dlibModelDir, \"shape_predictor_68_face_landmarks.dat\"))\nnet = openface.TorchNeuralNet(os.path.join(openfaceModelDir, 'nn4.small2.v1.t7'))\n\n\ndef alignImg(img):\n rgbimg = img.getRGB()\n if rgbimg is None:\n return None\n bb = align.getLargestFaceBoundingBox(rgbimg)\n alignedFace = align.align(96, rgbimg, bb, landmarkIndices=openface.AlignDlib.OUTER_EYES_AND_NOSE)\n return alignedFace\n\n\ndef disMed(rep, vectors, n):\n aux=0\n for i in range(3):\n distances=[]\n for j in range(n[i]):\n d = rep - vectors[j+aux]\n d = np.dot(d,d)\n distances.append(d)\n aux+=n[i]\n print(str(np.mean(distances)) + \" - \" + str(np.median(distances)) + \" - \" + str(np.std(distances)))\n \n\ndef main():\n '''\n persons = [\"miguel\", \"rudyer\"]\n persons_rep = []\n\n # Reading the persons that already know\n for person in persons:\n imgs = list(openface.data.iterImgs(\"raw/\"+ person))\n rep = []\n for i,img in enumerate(imgs,start=1):\n aligned = alignImg(img)\n if aligned is None:\n line = person + \" \" + str(i) + \" nao reconhecida\"\n print(line)\n else:\n rep.append(net.forward(aligned))\n persons_rep.append(rep)\n '''\n # Reading the unknown person\n imgs = list(openface.data.iterImgs(\"raw/desconhecido\"))\n img = imgs[0]\n print(img.name)\n img_aligned = alignImg(img)\n if img_aligned is None:\n print(\"nao reconhecido\")\n exit()\n else:\n img_rep = net.forward(img_aligned)\n\n\n imgs = list(openface.data.iterImgs(\"raw/igor\"))\n rep1 = []\n for i,img in enumerate(imgs):\n aligned = alignImg(img)\n if aligned is None:\n line = \"Igor \" + str(i) + \" nao reconhecida\"\n print(line)\n else:\n rep1.append(net.forward(aligned))\n \n imgs = list(openface.data.iterImgs(\"raw/rudyer\"))\n rep2 = []\n for i,img in enumerate(imgs):\n aligned = alignImg(img)\n if aligned is None:\n line = \"Rudyer \" + str(i) + \" nao reconhecida\"\n print(line)\n else:\n rep2.append(net.forward(aligned))\n \n imgs = list(openface.data.iterImgs(\"raw/miguel\"))\n rep3 = []\n for i,img in enumerate(imgs):\n aligned = alignImg(img)\n if aligned is None:\n line = \"Miguel \" + str(i) + \" nao reconhecida\"\n print(line)\n else:\n rep3.append(net.forward(aligned))\n\n # Fiting into the SVC classifier\n samples = rep1 + rep2 + rep3\n labels1 = [\"igor\"] * len(rep1)\n labels2 = [\"rudyer\"] * len(rep2)\n labels3 = [\"miguel\"] * len(rep3)\n labels = labels1 + labels2 + labels3\n \n clf = svm.SVC(gamma='scale', decision_function_shape='ovo', probability=True)\n clf.fit(samples, labels)\n\n # predicting the unknown face\n print(clf.predict(img_rep.reshape(1,-1)))\n\n disMed(img_rep, clf.support_vectors_, clf.n_support_)\n\n #disMed(img_rep, clf.support_vectors_, clf.n_support_[1])\n\n #disMed(img_rep, clf.support_vectors_, clf.n_support_[2])\n\n\n '''\n To do:\n classificar todas as imagens dos rostos conhecidos dentro do svm;\n ao ler um rosto desconhecido, achar as distancias do rosto pros SVM's;\n se menor ou maior que tal distancia (a definir), aplicar clf.predict().\n ''' \n \n\nmain()","sub_path":"classificador.py","file_name":"classificador.py","file_ext":"py","file_size_in_byte":3713,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"9009082","text":"from typing import Dict, List\n\nfrom cloudrail.knowledge.context.aws.resources.ec2.security_group_rule import SecurityGroupRule\n\nfrom cloudrail.knowledge.context.aws.cloudformation.cloudformation_constants import CloudformationResourceType\nfrom cloudrail.knowledge.context.aws.resources_builders.cloudformation.ec2.cloudformation_security_group_rule_base_builder import CloudformationSecurityGroupRuleBaseBuilder\n\n\nclass CloudformationSecurityGroupInlineRuleBuilder(CloudformationSecurityGroupRuleBaseBuilder):\n\n def __init__(self, cfn_by_type_map: Dict[CloudformationResourceType, Dict[str, Dict]]):\n super().__init__(CloudformationResourceType.SECURITY_GROUP, cfn_by_type_map)\n\n def parse_resource(self, cfn_res_attr: dict) -> List[SecurityGroupRule]:\n rules = []\n properties: dict = cfn_res_attr['Properties']\n\n rules.extend(self.parse_security_group_rule(security_group_rule_properties=ingress_rule_property,\n egress=False,\n security_group_id=self.get_resource_id(cfn_res_attr),\n account_id=cfn_res_attr['account_id'],\n region=cfn_res_attr['region']) for ingress_rule_property in self.get_property(properties, 'SecurityGroupIngress', []))\n\n rules.extend(self.parse_security_group_rule(security_group_rule_properties=egress_rule_property,\n egress=True,\n security_group_id=self.get_resource_id(cfn_res_attr),\n account_id=cfn_res_attr['account_id'],\n region=cfn_res_attr['region']) for egress_rule_property in self.get_property(properties, 'SecurityGroupEgress', []))\n\n return rules\n","sub_path":"cloudrail/knowledge/context/aws/resources_builders/cloudformation/ec2/cloudformation_security_group_inline_rule_builder.py","file_name":"cloudformation_security_group_inline_rule_builder.py","file_ext":"py","file_size_in_byte":1939,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"33287631","text":"from rdflib import Graph, plugin\nfrom rdflib.serializer import Serializer\n\n\nclass DataRetriever:\n\n def __init__(self):\n self.data = Graph()\n self.data.parse(\"./test.ttl\", format=\"turtle\")\n\n '''\n print(list(data.query(\"\"\"SELECT DISTINCT ?Concept\n WHERE {\n ?x a ?Concept .\n } LIMIT 1\n \"\"\"))[0].Concept)\n '''\n\n def get_pollutants(self):\n res = self.data.query(\"\"\"select distinct ?x \n WHERE \n {\n ?x a .\n }\n \"\"\")\n return res\n\n def get_stations(self):\n res = self.data.query(\"\"\"select distinct ?x \n WHERE \n {\n ?x a .\n }\n \"\"\")\n tmp = list()\n for row in res:\n d = str(row['x'].toPython())\n tmp.append(d)\n return tmp\n\n\n\n","sub_path":"HandsOn/Group07/application/back-end/data_manager.py","file_name":"data_manager.py","file_ext":"py","file_size_in_byte":1225,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"128031606","text":"#!/usr/bin/env python\n\n# Before running this script:\n#\n# bunzip2 enwiki-latest-pages-articles.xml.bz2\n# mkdir all\n# cd all\n# split -a 3 -b 64m ../enwiki-20130102-pages-articles.xml chunk_\n# cd ..\n#\n# Then:\n#\n# .../create_dataset.py all 64 8gb 8192\n# .../create_dataset.py all 64 4gb 4096\n# .../create_dataset.py all 64 2gb 2048\n# .../create_dataset.py all 64 1gb 1024\n\nimport os, random, shutil, sys\n\ndef choose_chunks(all_dir, chunk_size, dest_dir, dest_size):\n all_chunks = os.listdir(all_dir)\n num_chunks = dest_size/chunk_size\n dest_chunks = random.sample(all_chunks, num_chunks)\n os.mkdir(dest_dir)\n for chunk in dest_chunks:\n shutil.copyfile(\n os.path.join(all_dir, chunk),\n os.path.join(dest_dir, chunk))\n\nif __name__ == '__main__':\n assert len(sys.argv) == 5\n choose_chunks(\n sys.argv[1],\n int(sys.argv[2]),\n sys.argv[3],\n int(sys.argv[4]))\n","sub_path":"hadoop/create_dataset.py","file_name":"create_dataset.py","file_ext":"py","file_size_in_byte":927,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"382330723","text":"# Copyright 2010-2018 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\").\n# You may not use this file except in compliance with the License.\n# A copy of the License is located at\n#\n# http://aws.amazon.com/apache2.0\n#\n# or in the \"license\" file accompanying this file. This file is distributed\n# on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either\n# express or implied. See the License for the specific language governing\n# permissions and limitations under the License.\n\nfrom __future__ import print_function\n\nimport argparse\nfrom awscrt import io, mqtt\nfrom awscrt.io import LogLevel\nimport threading\nimport uuid\n\nTIMEOUT = 5 # seconds given to each step of the test before giving up\nUNIQUE_ID = str(uuid.uuid4()) # prevent simultaneously-running tests from interfering with each other\nCLIENT_ID = 'test_pubsub_' + UNIQUE_ID\nTOPIC = 'test/pubsub/' + UNIQUE_ID\nMESSAGE = 'test message ' + UNIQUE_ID\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--endpoint', required=True, help=\"Connect to this endpoint (aka host-name)\")\nparser.add_argument('--port', type=int, help=\"Override default connection port\")\nparser.add_argument('--cert', help=\"File path to your client certificate, in PEM format\")\nparser.add_argument('--key', help=\"File path to your private key, in PEM format\")\nparser.add_argument('--root-ca', help=\"File path to root certificate authority, in PEM format\")\n\nio.init_logging(LogLevel.Trace, 'stderr')\n\ndef on_connection_interrupted(error_code):\n print(\"Connection has been interrupted with error code\", error_code)\n\ndef on_connection_resumed(return_code, session_present):\n print(\"Connection has been resumed with return code\", return_code, \"and session present:\", session_present)\n\nreceive_results = {}\nreceive_event = threading.Event()\ndef on_receive_message(topic, message):\n receive_results.update(locals())\n receive_event.set()\n\n# Run\nargs = parser.parse_args()\nevent_loop_group = io.EventLoopGroup(1)\nclient_bootstrap = io.ClientBootstrap(event_loop_group)\n\ntls_options = None\nif args.cert or args.key or args.root_ca:\n if args.cert:\n assert args.key\n tls_options = io.TlsContextOptions.create_client_with_mtls_from_path(args.cert, args.key)\n else:\n tls_options = io.TlsContextOptions()\n\n if args.root_ca:\n with open(args.root_ca, mode='rb') as ca:\n rootca = ca.read()\n tls_options.override_default_trust_store(rootca)\n\nif args.port:\n port = args.port\nelif io.is_alpn_available():\n port = 443\n if tls_options:\n tls_options.alpn_list='x-amzn-mqtt-ca'\nelse:\n port = 8883\n\ntls_context = io.ClientTlsContext(tls_options) if tls_options else None\nmqtt_client = mqtt.Client(client_bootstrap, tls_context)\n\n# Connect\nprint(\"Connecting to {}:{} with client-id:{}\".format(args.endpoint, port, CLIENT_ID))\nmqtt_connection = mqtt.Connection(\n client=mqtt_client,\n on_connection_interrupted=on_connection_interrupted,\n on_connection_resumed=on_connection_resumed)\n\nconnect_results = mqtt_connection.connect(\n client_id=CLIENT_ID,\n host_name=args.endpoint,\n port=port).result(TIMEOUT)\nassert(connect_results['session_present'] == False)\n\n# Subscribe\nprint(\"Subscribing to:\", TOPIC)\nqos = mqtt.QoS.AT_LEAST_ONCE\nsubscribe_future, subscribe_packet_id = mqtt_connection.subscribe(\n topic=TOPIC,\n qos=qos,\n callback=on_receive_message)\nsubscribe_results = subscribe_future.result(TIMEOUT)\nassert(subscribe_results['packet_id'] == subscribe_packet_id)\nassert(subscribe_results['topic'] == TOPIC)\nprint(subscribe_results)\nassert(subscribe_results['qos'] == qos)\n\n# Publish\nprint(\"Publishing to '{}': {}\".format(TOPIC, MESSAGE))\npublish_future, publish_packet_id = mqtt_connection.publish(\n topic=TOPIC,\n payload=MESSAGE,\n qos=mqtt.QoS.AT_LEAST_ONCE)\npublish_results = publish_future.result(TIMEOUT)\nassert(publish_results['packet_id'] == publish_packet_id)\n\n# Receive Message\nprint(\"Waiting to receive messsage\")\nassert(receive_event.wait(TIMEOUT))\nassert(receive_results['topic'] == TOPIC)\nassert(receive_results['message'].decode() == MESSAGE)\n\n# Unsubscribe\nprint(\"Unsubscribing from topic\")\nunsubscribe_future, unsubscribe_packet_id = mqtt_connection.unsubscribe(TOPIC)\nunsubscribe_results = unsubscribe_future.result(TIMEOUT)\nassert(unsubscribe_results['packet_id'] == unsubscribe_packet_id)\n\n# Disconnect\nprint(\"Disconnecting\")\nmqtt_connection.disconnect().result(TIMEOUT)\n\n# Done\nprint(\"Test Success\")\n","sub_path":"mqtt_test.py","file_name":"mqtt_test.py","file_ext":"py","file_size_in_byte":4538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"455619867","text":"from PyQt5.QtWidgets import QApplication,QMainWindow,QLineEdit,QMessageBox\r\nfrom PyQt5 import uic\r\nfrom DB import DB\r\nfrom kamera import Kamera as cam2\r\nfrom kamera2 import Kamera as cam1\r\n\r\nclass App(QMainWindow):\r\n def __init__(self):\r\n super().__init__()\r\n self.win = uic.loadUi(r\"GUI\\ilk.ui\")\r\n self.ilDoldur()\r\n self.win.cmbIL.currentIndexChanged.connect(self.ilceDoldur)\r\n self.win.btKaydet.clicked.connect(self.kaydet)\r\n self.win.btKamera.clicked.connect(self.kameraAc)\r\n self.win.btKamera_2.clicked.connect(self.kameraAc2)\r\n self.cam = cam1()\r\n self.cam2 = cam2()\r\n self.win.show()\r\n \r\n\r\n def ilceDoldur(self):\r\n db = DB()\r\n self.win.cmbILCE.clear()\r\n liste = db.ilceListele(self.win.cmbIL.currentIndex())\r\n self.win.cmbILCE.addItem(\"Seçiniz\")\r\n for IlceKod,IlceAd in liste:\r\n self.win.cmbILCE.addItem(IlceAd)\r\n \r\n\r\n def kameraAc(self):\r\n self.cam.adiSoyadi = self.adi+\"_\"+self.soyadi\r\n self.cam.olustur()\r\n self.cam.widg.show()\r\n \r\n def kameraAc2(self):\r\n self.cam2.widg.show()\r\n \r\n\r\n\r\n def ilDoldur(self):\r\n db = DB()\r\n liste = db.ilListele()\r\n self.win.cmbIL.addItem(\"Seçiniz\")\r\n for IlKod,IlAd in liste:\r\n self.win.cmbIL.addItem(IlAd)\r\n\r\n def kaydet(self):\r\n self.adi = self.win.txtAdi.text()\r\n self.soyadi = self.win.txtSoyadi.text() \r\n il = self.win.cmbIL.currentIndex()\r\n ilce = self.win.cmbILCE.currentIndex()\r\n db = DB()\r\n if db.PersonelEkleme(self.adi,self.soyadi,il,ilce):\r\n QMessageBox.information(self,\"Bilgi\",\"Bilgileriniz başarıyla kaydedildi\")\r\n self.win.btKamera.setEnabled(True)\r\n\r\n\r\nif __name__ == '__main__':\r\n import sys\r\n app = QApplication(sys.argv)\r\n ex = App()\r\n sys.exit(app.exec_())","sub_path":"ilk.py","file_name":"ilk.py","file_ext":"py","file_size_in_byte":1951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"417506546","text":"\n\n\"\"\"\n\t@Author: Salman Ahmed\n\n\"\"\"\n\n\n\n\n\nimport numpy as np # linear algebra\nimport pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)\nimport os\nfrom keras.layers import Input, Dense, Reshape, Flatten, Dropout, Concatenate\nfrom keras.layers import BatchNormalization, Activation, ZeroPadding2D, Add\nfrom keras.layers.advanced_activations import PReLU, LeakyReLU\nfrom keras.layers.convolutional import UpSampling2D, Conv2D\nfrom keras.models import Model, Sequential\nfrom keras.optimizers import Adam\n\n\ndef build_generator(noise_shape = (100,), img_shape=(100,3,1)):\n model = Sequential()\n model.add(Dense(256, input_shape=noise_shape))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(1024))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n model.add(Dense(np.prod(img_shape), activation='tanh'))\n model.add(Reshape(img_shape))\n noise = Input(shape=noise_shape)\n img = model(noise)\n return Model(noise, img)\n\n\ndef build_discriminator(img_rows=100, img_cols=3, channels=1):\n img_shape = (img_rows, img_cols, channels)\n model = Sequential()\n model.add(Flatten(input_shape=img_shape))\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(Dense(256))\n model.add(LeakyReLU(alpha=0.2))\n model.add(Dense(1, activation='sigmoid'))\n img = Input(shape=img_shape)\n validity = model(img)\n return Model(img, validity)\n\n\ndata = pd.read_csv(\"data/left_lane.csv\")\ndata.pop('action')\ndata = data.sample(frac=1)\nob_netCols = []\nk = 0\nwhile k < 300:\n ob_netCols.append(\"new_col\"+str(k))\n k += 1\nob_net = data[ob_netCols]\nob_net = np.reshape(ob_net.values, (ob_net.shape[0],100,3,1))\nob_net[ob_net>0] = 0\nob_net[ob_net==-1] = -5\nob_net = ob_net + 10\nob_net = ob_net / 10\nob_net = ob_net * 2\nob_net = ob_net - 1\ngenerator = build_generator()\ndiscriminator = build_discriminator()\ndiscriminator.compile(loss='binary_crossentropy', optimizer=Adam(0.0002, 0.5), metrics=['accuracy'])\ngenerator.compile(loss='binary_crossentropy',optimizer=Adam(0.0002, 0.5))\noverallModel = Sequential()\noverallModel.add(generator)\ndiscriminator.trainable = False\noverallModel.add(discriminator)\noverallModel.compile(loss='binary_crossentropy', optimizer=Adam(0.0002, 0.5))\nBATCH_SIZE = int(ob_net.shape[0]/3)\nindc = []\nj = 0\nwhile j < ob_net.shape[0]:\n indc.append(j)\n j+=1\nfor epoch in range(1000):\n noise = np.random.uniform(0, 1.0, size=(BATCH_SIZE,100))\n fakehR = generator.predict(noise)\n valid = np.ones((BATCH_SIZE,1))\n fake = np.zeros((BATCH_SIZE,1))\n# valid[8] = [[[0]]]\n# valid[13] = [[[0]]]\n\n\n valid = valid * (1 - 0.1)\n fake = fake + (0.1)\n\n # highR = data.sample(n=BATCH_SIZE)\n index = np.random.choice(indc, BATCH_SIZE, replace=False) \n highR = ob_net[index]\n\n discriminator.trainable = True\n d_loss_real = discriminator.train_on_batch(highR, valid)\n d_loss_fake = discriminator.train_on_batch(fakehR, fake)\n k = 0\n d_loss = (d_loss_real[0] + d_loss_fake[0]) * 0.5\n\n\n y = np.ones((BATCH_SIZE,1))\n discriminator.trainable = False\n j = 0\n while j < 5:\n noise = np.random.uniform(0, 1.0, size=(BATCH_SIZE,100))\n \n g_loss = overallModel.train_on_batch(noise,y)\n print (\"Epoch : \",epoch, \" Discriminator Loss : \", d_loss, \" Generator Loss : \", g_loss)\n j+=1\ngenerator.save_weights(\"Left_GEN_WEIGHTS.h5\")\n\n","sub_path":"GANS-Lane_Changes_Implementation/training/left_examples_training.py","file_name":"left_examples_training.py","file_ext":"py","file_size_in_byte":3568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"184499008","text":"from flask import Flask, render_template, flash, request, url_for, redirect, session\nimport numpy as np\nimport pandas as pd\nimport os\nIMAGE_FOLDER = os.path.join('static')\napp = Flask(__name__)\n\n#model = load_model('sentiment_analysis_model.h5')\n\n@app.route('/')\ndef home():\n return render_template('home.html')\n\n@app.route('/predict/',methods=['POST'])\ndef predict():\n if request.method == 'POST':\n message = request.form['message']\n data = [message]\n json_table = data.to_json(orient='records')\n return app.response_class(response=json_table,status=200,mimetype='application/json')\n\nif __name__ == \"__main__\":\n #init()\n app.run(host='127.0.0.0',port=4000)\n","sub_path":"Sentiment_Analysis-Tensorflow/text_accept.py","file_name":"text_accept.py","file_ext":"py","file_size_in_byte":694,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"323381987","text":"import pandas as pd\nimport numpy as np\n# from gensim.models import doc2vec\nfrom gensim import models\nfrom gensim.models.deprecated.doc2vec import LabeledSentence\n\n\n# Filepath to main training dataset.\ntrain_file_path = \"/Users/zhewang/Documents/kaggle/train.csv\"\ndtype = {\n 'id': str,\n 'teacher_id': str,\n 'teacher_prefix': str,\n 'school_state': str,\n 'project_submitted_datetime': str,\n 'project_grade_category': str,\n 'project_subject_categories': str,\n 'project_subject_subcategories': str,\n 'project_title': str,\n 'project_essay_1': str,\n 'project_essay_2': str,\n 'project_essay_3': str,\n 'project_essay_4': str,\n 'project_resource_summary': str,\n 'teacher_number_of_previously_posted_projects': int,\n 'project_is_approved': np.uint8,\n}\n# Read data and store in DataFrame.\ntrain_data = pd.read_csv(train_file_path, sep=',', dtype=dtype, low_memory=True).sample(10000)\nessay1 = train_data['project_essay_1']\nids = train_data['id']\n\ness1_list = []\nfor index, row in train_data.iterrows():\n\tess1_list.append(LabeledSentence(row['project_essay_1'].split(\" \"), [row['id']]))\n#size is the vector length, window means how many words are included in one paragraph \nmodel = models.Doc2Vec(size = 100, window = 200, min_count = 3, workers = 1)\nvocab = model.build_vocab(ess1_list)\nmodel.train(ess1_list, epochs=10, total_words=100)\nmodel.save(\"ess1_model.doc2vec\")\n# model_loaded = models.Doc2Vec.load('ess1_model.doc2vec')\n# print \"the first vector is: \"\n# print model.docvecs[0]\n","sub_path":"src/train-doc2vec.py","file_name":"train-doc2vec.py","file_ext":"py","file_size_in_byte":1526,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"508735822","text":"import json\n\nimport pytest\n\nfrom oapispec.core import utils\nfrom oapispec.core.utils import immutable\n\n\ndef test_merge_simple_dicts_without_precedence():\n a = {'a': 'value'}\n b = {'b': 'other value'}\n assert utils.merge(a, b) == {'a': 'value', 'b': 'other value'}\n\ndef test_merge_simple_dicts_with_precedence():\n a = {'a': 'value', 'ab': 'overwritten'}\n b = {'b': 'other value', 'ab': 'keep'}\n assert utils.merge(a, b) == {'a': 'value', 'b': 'other value', 'ab': 'keep'}\n\ndef test_merge_recursions():\n a = {\n 'a': 'value',\n 'ab': 'overwritten',\n 'nested_a': {\n 'a': 'nested'\n },\n 'nested_a_b': {\n 'a': 'a only',\n 'ab': 'overwritten'\n }\n }\n b = {\n 'b': 'other value',\n 'ab': 'keep',\n 'nested_b': {\n 'b': 'nested'\n },\n 'nested_a_b': {\n 'b': 'b only',\n 'ab': 'keep'\n }\n }\n assert utils.merge(a, b) == {\n 'a': 'value',\n 'b': 'other value',\n 'ab': 'keep',\n 'nested_a': {\n 'a': 'nested'\n },\n 'nested_b': {\n 'b': 'nested'\n },\n 'nested_a_b': {\n 'a': 'a only',\n 'b': 'b only',\n 'ab': 'keep'\n }\n }\n\ndef test_merge_recursions_with_empty():\n a = {}\n b = {\n 'b': 'other value',\n 'ab': 'keep',\n 'nested_b': {\n 'b': 'nested'\n },\n 'nested_a_b': {\n 'b': 'b only',\n 'ab': 'keep'\n }\n }\n assert utils.merge(a, b) == b\n\ndef test_immutable_raises_when_set():\n obj = immutable(x=2, y=23)\n\n with pytest.raises(TypeError):\n obj['x'] = 5\n\n with pytest.raises(TypeError):\n obj.x = 5\n\n with pytest.raises(TypeError):\n del obj['x']\n\n with pytest.raises(TypeError):\n del obj.x\n\ndef test_immutable_none_comparison():\n obj = immutable(x=2, y=23)\n isEqual = obj == None\n assert isEqual is not True\n\ndef test_immutable_repr_dumps_json():\n obj = immutable(x=2, y=23)\n r = repr(obj)\n json.loads(r)\n\ndef test_immutable_str_dumps_json():\n obj = immutable(x=2, y=23)\n s = str(obj)\n json.loads(s)\n\ndef test_immutable_equality():\n a = immutable(x=2, y=23)\n b = immutable(x=2, y=23)\n\n assert a == b\n","sub_path":"tests/core/utils_test.py","file_name":"utils_test.py","file_ext":"py","file_size_in_byte":2327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"499338226","text":"from pathlib import Path\n\ndata_folder = Path(\".\").resolve()\n\n\nmove_instr = {\"up\": (0, -1), \"down\": (0, 1), \"forward\": (1, 1)}\n\n\ndef part1(data):\n pos = [0, 0]\n for direction, distance in data:\n index, unit = move_instr[direction]\n pos[index] += unit * distance\n return pos[0] * pos[1]\n\n\ndef part2(data):\n pos = [0, 0, 0]\n for direction, distance in data:\n index, unit = move_instr[direction]\n if index == 0:\n pos[2] += unit * distance\n else:\n pos[1] += unit * distance\n pos[0] += pos[2] * distance\n return pos[0] * pos[1]\n\n\ndef parse_line(line):\n direction, distance = line.split()\n distance = int(distance)\n return direction, distance\n\n\ndef main():\n data = data_folder.joinpath(\"input.txt\").read_text()\n data = [parse_line(d) for d in data.split(\"\\n\")]\n\n sub_prod = part1(data)\n print(\"Part 1\")\n print(f\"If we multiply the final horizontal position by the final depth we get {sub_prod}\")\n print()\n\n sub_prod = part2(data)\n print(\"Part 2\")\n print(f\"If we multiply the final horizontal position by the final depth we get {sub_prod}\")\n print()\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"2021/02/sol.py","file_name":"sol.py","file_ext":"py","file_size_in_byte":1207,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"269895587","text":"#!/usr/bin/python\n\nimport subprocess\nprofile_url = \"https://www.vpngate.net/common/openvpn_download.aspx?sid=1584878359201&tcp=1&host=219.100.37.144&port=443&hid=15134922&/vpngate_219.100.37.144_tcp_443.ovpn\"\np = subprocess.Popen(\"docker run -it --rm -e profile=\\\"{0}\\\" openvpncli\".format(profile_url), stdout=subprocess.PIPE, shell=True)\n \n(output, err) = p.communicate()\n \np_status = p.wait()\nresults = output.decode('utf-8')\nif \"ERROR: Cannot open TUN/TAP dev /dev/net/tun: No such file or directory\" in results:\n print (\"OK\")\nelse:\n print (\"ERROR\")\n","sub_path":"VPNGateValidation/check.py","file_name":"check.py","file_ext":"py","file_size_in_byte":559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"367800356","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Apr 29 15:33:05 2018\n\n@author: xnguyen\n\"\"\"\n\nimport tensorflow as tf\nimport itertools\n\n\ndef gen():\n for i in itertools.count(1):\n yield (i, [1] * i)\n\nds = tf.data.Dataset.from_generator(\n gen, (tf.int64, tf.int64), (tf.TensorShape([]), tf.TensorShape([None])))\nvalue = ds.make_one_shot_iterator().get_next()\n\nwith tf.Session() as sess:\n for i in range(10): \n print(sess.run(value)) # (1, array([1]))\n #print(sess.run(value)) # (2, array([1, 1]))\n\n\n# Using the generator pattern (an iterable)\nclass generator_recurrent_sin(object):\n def __init__(self, n, time_steps, input_dim, attention_column=13):\n self.n = n \n self.time_steps = time_steps\n self.input_dim = input_dim \n self.attention_column = attention_column\n self.num = 0\n \n def __iter__(self):\n return self.next()\n \n # Python 3 compatibility\n def __next__(self):\n return self.next()\n\n def next(self):\n if self.num < self.n:\n self.num += 1\n x = np.random.standard_normal(size=(self.time_steps, self.input_dim)) \n y = np.random.randint(low=0, high=2, size=(1))\n \n freq = 0\n if y ==0:\n freq = 0.1 * PI\n else:\n freq = 0.5 * PI\n for t in range(self.attention_column, int(self.attention_column+self.input_dim/2)):\n for d in range(self.attention_column):\n x[t, d] = np.sin(t*freq) + 0.05*np.random.randn(1)\n yield (x, y)\n else: \n raise StopIteration()\n\nINPUT_DIM = 32\nTIME_STEPS = 64\nHIDDEN_UNITS = 256\nATTENTION_COLUMN =10\nN=1000\n\ndef get_data_generator():\n return generator_recurrent_sin(N, TIME_STEPS, INPUT_DIM,\n ATTENTION_COLUMN)\n \ndataset = tf.data.Dataset.from_generator(get_data_generator,\n (tf.float32, tf.float32),\n (tf.TensorShape([TIME_STEPS, INPUT_DIM]),\n tf.TensorShape([1])))\nbatch_size=10\n \ndataset = dataset.shuffle(buffer_size=1024) # 1024 files in dataset\ndataset = dataset.repeat() \ndataset = dataset.apply(\n tf.contrib.data.batch_and_drop_remainder(batch_size)\n )\n \ndataset = dataset.prefetch(4)\nimages, labels = dataset.make_one_shot_iterator().get_next()\n\nwith tf.Session() as sess:\n print(sess.run([images, labels]))\n","sub_path":"test_dataset.py","file_name":"test_dataset.py","file_ext":"py","file_size_in_byte":2559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"594506529","text":"# -*- coding: utf-8 -*- \nimport datetime\n\nfrom django.shortcuts import render\nfrom django.http import HttpResponse\nfrom statistic.models import Phone\nfrom .models import Category, PhoneList, Version\nimport json\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.template.context_processors import csrf\n\n@csrf_exempt\ndef phoneData(request):\n if request.method == 'POST':\n device_post = request.POST.get('device')\n key_post = request.POST.get('key')\n if key_post == 'KqePnWoGfHhbLCU4yoPEXi5qXWQk69IE':\n # sync++\n p = Phone.objects.get()\n p.sync_amount += 1\n if device_post == 'iOS':\n # iOS\n p.ios_sync_amount += 1\n elif device_post == 'Android':\n # Android\n p.android_sync_amount += 1\n else:\n # unrecognized device\n return None\n p.save()\n # prepare data\n phoneData = []\n for categoryData in Category.objects.all():\n categoryDict = {}\n categoryDict['name'] = categoryData.category_name\n listArray = []\n for listData in PhoneList.objects.filter(phone_category=categoryData):\n listArray.append({\n 'name': listData.phone_name,\n 'num1': listData.phone_num1,\n 'num2': listData.phone_num2,\n })\n categoryDict['list'] = listArray\n phoneData.append(categoryDict)\n # add Version\n v = Version.objects.get()\n jsonData = {\n 'version': v.version_num,\n 'data': phoneData,\n }\n # ensure_ascii=False Chinese\n return HttpResponse(json.dumps(jsonData, ensure_ascii=False), content_type='application/json')\n else:\n return None\n else:\n return None","sub_path":"phone/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1981,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"5797559","text":"''' graph_controllers\nroutes related to the graphs\n'''\nimport datetime\nimport json\n\nfrom flask import (render_template, request, flash, redirect, url_for, session\n , abort)\nfrom mongoengine import *\n\nfrom application import app\n\n# view controls\nfrom decorators import *\n# mongoengine models\nfrom models import *\n# some utilities\nimport utilities\n# some constants\nimport constants\n\n\ngraph_route = '/organizations//projects//graphs'\n@app.route(graph_route, defaults={'graph_label': None})\n@app.route(graph_route + '/', methods=['GET', 'POST'])\n@verification_required\n@csrf_protect\ndef graphs(org_label, project_label, graph_label):\n ''' graphin things\n /organizations/aquaya/projects/water-quality/graphs\n : view a list of all graphs for the project\n /organizations/aquaya/projects/water-quality/graphs?create=true\n : create a new graph config, immediately redirect to editing\n /organizations/aquaya/projects/water-quality/graphs/ph-vs-time\n : view a graph\n /organizations/aquaya/projects/water-quality/graphs/ph-vs-time?edit=true\n : edit a graph; accepts GET or POST\n '''\n user = User.objects(email=session['email'])[0]\n \n orgs = Organization.objects(label=org_label)\n if not orgs:\n flash('Organization \"%s\" not found, sorry!' % org_label, 'warning')\n return redirect(url_for('organizations'))\n org = orgs[0]\n\n # permission-check\n if org not in user.organizations and not user.admin_rights:\n app.logger.error('%s tried to view a project but was \\\n denied for want of admin rights' % session['email'])\n abort(404)\n \n # find the project\n projects = Project.objects(label=project_label, organization=org) \n if not projects:\n flash('Project \"%s\" not found, sorry!' % project_label, 'warning')\n return redirect(url_for('organizations', org_label=org.label))\n project = projects[0]\n\n if request.method == 'POST':\n # we have a graph_label\n graphs = Graph.objects(label=graph_label, project=project)\n if not graphs:\n abort(404)\n graph = graphs[0]\n\n form_type = request.form.get('form_type', '')\n if form_type == 'info':\n if graph.name != request.form.get('name', ''):\n name = request.form.get('name', '')\n graph.update(set__name = name)\n\n graphs = Graph.objects(project=project).only('label')\n labels = [g.label for g in graphs]\n graph.update(set__label = utilities.generate_label(name\n , labels))\n\n # reload to catch the name change\n graph.reload()\n\n graph.update(set__description = \n request.form.get('description', ''))\n graph.update(set__graph_type = request.form.get('graph_type', ''))\n\n # axes specify a header and come of the form 'header_id__4abcd001'\n xaxis = request.form.get('xaxis', '')\n if xaxis:\n xaxis = xaxis.split('header_id__')[1]\n header = Header.objects(id=xaxis)[0]\n graph.update(set__xaxis = header)\n\n yaxis = request.form.get('yaxis', '')\n if yaxis:\n yaxis = yaxis.split('header_id__')[1]\n header = Header.objects(id=yaxis)[0]\n graph.update(set__yaxis = header)\n\n # pie chart headers are similar to axes..\n pie_header = request.form.get('pie_header', '')\n if pie_header:\n pie_header = pie_header.split('header_id__')[1]\n header = Header.objects(id=pie_header)[0]\n graph.update(set__pie_header = header)\n\n elif form_type == 'filters':\n # extract the 'any filters' vs 'all' distinction\n filter_settings = request.form.get('apply_any_filters', '')\n if filter_settings == 'true':\n graph.update(set__apply_any_filters = True)\n else:\n graph.update(set__apply_any_filters = False)\n\n # attach filter to graph\n requested_filter_ids = request.form.getlist('filters')\n attached_filters = []\n for requested_id in requested_filter_ids:\n prefix, filter_id = requested_id.split('__')\n filters = Filter.objects(id=filter_id)\n if not filters:\n abort(404)\n attached_filters.append(filters[0])\n\n graph.update(set__filters = attached_filters)\n \n elif form_type == 'admin':\n # delete the graph\n name = graph.name\n utilities.delete_graph(graph, session['email'])\n flash('graph \"%s\" was deleted successfully' % name, 'success')\n return redirect(url_for('graphs', org_label=org.label\n , project_label=project.label))\n \n else:\n # bad 'form_type'\n abort(404)\n \n flash('changes saved successfully', 'success')\n return redirect(url_for('graphs', org_label=org.label\n , project_label=project.label, graph_label=graph.label))\n \n if request.method == 'GET':\n if graph_label:\n graphs = Graph.objects(label=graph_label, project=project)\n if not graphs:\n app.logger.error('%s tried to access a graph that does not \\\n exist' % session['email'])\n flash('Graph \"%s\" not found, sorry!' % graph_label\n , 'warning')\n return redirect(url_for('projects'), org_label=org.label\n , project_label=project.label)\n graph = graphs[0]\n\n if request.args.get('edit', '') == 'true':\n # valid graph types\n graph_types = ['line', 'scatter', 'bar', 'chart', 'pie']\n \n available_filters = Filter.objects(project=project)\n \n return render_template('graph_edit.html', graph=graph\n , graph_types = graph_types\n , allowed_graph_types = constants.graph_types\n , available_filters = available_filters)\n\n else:\n # render a graph\n data = []\n project_count = None\n filtered_count = None\n if graph.graph_type == 'line':\n if graph.xaxis and graph.yaxis:\n data, project_count = (\n utilities.generate_line_graph_data(graph))\n filtered_count = len(data)\n else:\n flash('define an x-axis and y-axis for plotting'\n , 'warning')\n\n elif graph.graph_type == 'pie':\n if graph.pie_header:\n data, project_count = (\n utilities.generate_pie_chart_data(graph))\n filtered_count = sum([i['data'] for i in json.loads(data)])\n else:\n flash('define a column to create this pie chart'\n , 'warning')\n\n return render_template('graph.html', graph=graph, data=data\n , project_count=project_count\n , filtered_count = filtered_count)\n\n if request.args.get('create', '') == 'true':\n # create a new graph\n\n # CSRF validation\n token = request.args.get('token', '')\n if not verify_token(token):\n abort(403)\n\n try:\n graph_name = 'graph-%s' % utilities.generate_random_string(6)\n new_graph = Graph(\n creation_time = datetime.datetime.utcnow()\n , creator = user\n , label = graph_name.lower()\n , project = project\n , name = graph_name\n )\n new_graph.save() \n app.logger.info('graph created by %s' % session['email'])\n flash('graph created; please change the defaults', 'success')\n except:\n app.logger.error('graph creation failed for %s' % \\\n session['email'])\n flash('There was an error, sorry :/', 'error')\n return redirect(url_for('projects', org_label=org.label\n , project=project.label))\n \n # redirect to the editing screen\n return redirect(url_for('graphs', org_label=org.label\n , project_label=project.label, graph_label=new_graph.label\n , edit='true'))\n \n # no graph in particular was specified\n graphs = Graph.objects(project=project)\n return render_template('project_graphs.html', project=project\n , graphs=graphs)\n\n","sub_path":"application/graph_controllers.py","file_name":"graph_controllers.py","file_ext":"py","file_size_in_byte":8998,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"652182977","text":"import pandas as pd\nimport matplotlib.pyplot as plt\nimport numpy as np\n# Importamos funcion donde limpiamos nuestra data y creamos un nuevo csv\nfrom limpieza import limpiar_crear\n# Importamos nuestras funcionoes\nfrom funcionesDatos import buscar, bajas_pastel, sobrevivientesBarras, costoTickets\nimport os\n\n\nif __name__ == '__main__':\n # Llamamos la funcion para limpiar y crear un nuevo csv\n # df_raw = pd.read_excel(\"lista.xlsx\")\n # limpiar_crear(df_raw) # retornamos y cramos data.csv\n\n while True:\n os.system(\"cls\") # Limpieza de pantalla\n print(\"Menu:\")\n print(\"1. Buscar un pasajero por codigo de ticket\")\n print(\"2. Gráfica de bajas y sobrevivientes(pastel)\")\n print(\"3. Gráfica de sobrevivientes por clase(barras)\")\n print(\"4. Reporte de costos de tickets\")\n print(\"5. Salir\")\n\n opc = input(\"Seleccione su opción: \")\n\n if opc == \"1\":\n ticket = input(\"Digite el codigo del ticket(Ej: PC 17318 o 113781): \") # Capturando el ticket\n buscar(ticket)\n input()\n\n elif opc == \"2\":\n bajas_pastel()\n input()\n\n elif opc == \"3\":\n sobrevivientesBarras()\n input()\n\n elif opc == \"4\":\n costoTickets()\n input()\n\n elif opc == \"5\":\n print(\"Hasta pronto...\")\n input()\n break\n\n else:\n print(\"Opción incorrecta\")\n input()\n continue","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"98994393","text":"#Örnek Problem:\r\n\"\"\"\r\nmetreküpü 0.79tl olan doğalgaz Mart ayında 346metreküp harcandığına göre;\r\nkaç lira fatura geleceğini hesaplayın\r\n\"\"\"\r\n\r\nmetreküp=0.79\r\nharcanan=346\r\n\r\nfatura = metreküp*harcanan\r\nprint (fatura)\r\n\r\n\r\n# kaynak: https://python-istihza.yazbel.com/etkilesimli_python.html#karakter-dizilerine-giris","sub_path":"7-07.01.2021/ÖrnekProgramlar/ornekproblem5.py","file_name":"ornekproblem5.py","file_ext":"py","file_size_in_byte":330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"147760700","text":"# Importing relevant packages\nfrom flask import Blueprint, render_template, request, session, redirect, url_for\nfrom .models import Item, Watchlist, Bid\nfrom . import db\nfrom sqlalchemy.sql.expression import func\nfrom sqlalchemy import desc\n\n# blueprint\nbp = Blueprint('main', __name__)\n\n# Function for index page\n@bp.route('/')\ndef index():\n func.count()\n items = Item.query.all() # grabs all items in db\n # Does a bunch of filtering based on items\n featured_items = Item.query.order_by(func.random()).limit(4).all()\n CPUCategory = Item.query.order_by(func.random()).filter_by(category='CPU').limit(4)\n GPUCategory = Item.query.order_by(func.random()).filter_by(category='GPU').limit(4)\n MoboCategory = Item.query.order_by(func.random()).filter_by(category='Motherboard').limit(4)\n RAMCategory = Item.query.order_by(func.random()).filter_by(category='RAM').limit(4)\n PSUCategory = Item.query.order_by(func.random()).filter_by(category='Power Supply Unit').limit(4)\n FanCategory = Item.query.order_by(func.random()).filter_by(category='Cooling Fan').limit(4)\n # renders the page with variables\n return render_template('index.html', items=items, featured_items=featured_items, CPUCategory=CPUCategory, GPUCategory=GPUCategory, MoboCategory=MoboCategory, RAMCategory=RAMCategory, PSUCategory=PSUCategory, FanCategory=FanCategory)\n\n# Function that allows the user to search for items by name\n@bp.route('/search')\ndef search():\n # grabs the text inputted, then filters DB based on that\n if request.args['search']:\n ite = \"%\" + request.args['search'] + \"%\"\n items = Item.query.filter(Item.name.like(ite)).all()\n return render_template('index.html', search_items=items)\n else:\n return redirect(url_for('main.index'))\n\n\n\n\n","sub_path":"auction/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"456324839","text":"# 6549 / calculate max rectangle area in the histogram\nfrom collections import deque\nimport sys\n\nwhile True:\n n, *num = list(map(int, sys.stdin.readline().strip().split()))\n if n == 0: # end\n break\n\n num.append(0) # for terminate histogram, append 0 to end\n area, hist = 0, deque()\n for i, h in enumerate(num):\n # if histogram is not empty and top heigh is higher than h\n while hist and num[hist[-1]] > h:\n ih = num[hist.pop()] # height\n # width is from now index to next top height in histogram and -1 or index\n w = i - hist[-1] - 1 if hist else i\n area = max(area, w * ih) # caluclate area\n hist.append(i) # append index\n print(area)\n","sub_path":"stack/6549.py","file_name":"6549.py","file_ext":"py","file_size_in_byte":731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"339211422","text":"#!/usr/bin/env python \n# -*- coding: utf-8 -*- \n# @Time : 3/29/2018 10:53 AM \n# @Author : sunyonghai \n# @File : create_Main.py \n# @Software: ZJ_AI\n# =========================================================\nimport argparse\nimport random\nimport os\nfrom data_processing.utils.io_utils import mkdir\nfrom config import ROOT_HOME\n\ndef _create_Main(path):\n '''\n create the trainval.txt and test.txt for train.\n trainval data : test data = 5:1\n :param path:\n :return:\n '''\n scale = 3\n image_dir = os.path.join(path, 'JPEGImages')\n anno_dir = os.path.join(path, 'Annotations')\n ImageSets_path = os.path.join(path, 'ImageSets')\n main_dir = os.path.join(ImageSets_path, 'Main')\n\n mkdir(main_dir)\n\n imgs = os.listdir(image_dir)\n random.shuffle(imgs)\n\n trainval_test_images = []\n trainval_images = []\n test_images = []\n\n for i in range(len(imgs)):\n s = imgs[i]\n trainval_test_images.append(s.split('.')[0] + '\\n')\n\n for i in range(len(imgs)//scale, len(imgs)):\n s = imgs[i]\n trainval_images.append(s.split('.')[0] + '\\n')\n\n for i in range(len(imgs)//scale):\n s = imgs[i]\n test_images.append(s.split('.')[0] + '\\n')\n\n with open(main_dir+'/trainval_test.txt','w+') as f:\n f.writelines(trainval_test_images)\n print(\"{}, numbers:{}\".format(main_dir + '/trainval_test.txt', len(trainval_test_images)))\n\n with open(main_dir+'/trainval.txt','w+') as f:\n f.writelines(trainval_images)\n print(\"{}, numbers:{}\".format(main_dir + '/trainval.txt', len(trainval_images)))\n with open(main_dir+'/test.txt','w+') as f:\n f.writelines(test_images)\n print(\"{}, numbers:{}\".format(main_dir + '/test.txt', len(test_images)))\n\n print('total: {}'.format(len(imgs)))\n print('step: {}'.format(len(trainval_images)//2+1))\n\ndef create_sub_Main(dirs):\n data_paths = [os.path.join(ROOT_HOME, s) for s in dirs]\n for data_dir in data_paths:\n _create_Main(data_dir)\n\ndef create_subs_Main(data_paths):\n # dirs = ['data/train_data-2018-3-7', 'data/train_data-2018-3-16']\n # data_paths = [os.path.join(ROOT_HOME, s) for s in dirs]\n for data_dir in data_paths:\n _create_Main(data_dir)\n\ndef create_txt(data_dir):\n _create_Main(data_dir)\n\ndef create_txts(data_dirs):\n for data_dir in data_dirs:\n _create_Main(data_dir)\n\n# if __name__ == \"__main__\":\n# dirs = ['data_52/train_data-2018-04-18']\n# data_paths = [os.path.join(ROOT_HOME, s) for s in dirs]\n# create_subs_Main(data_paths)\n\nparser = argparse.ArgumentParser(description='Get the data info')\n# parser.add_argument('-d', '--datadir', help='path in server', default='/home/syh/train_data/data/all_train_data')\n# parser.add_argument('-d', '--datadir', help='path in server', default='/home/syh/train_data/test')\nparser.add_argument('-d', '--datadir', help='path in server', default='/home/syh/tf-faster-rcnn/data/train_data/train_data-2018-04-02')\nargs = parser.parse_args()\n\nif __name__ == \"__main__\":\n # data_dir = '/home/syh/disk/train/all_train_data'\n\n data_dir = args.datadir\n # data_dir = '/home/syh/all_train_data'\n create_txt(data_dir)\n\n\"\"\"\ncd ~/RetinaNet/data_processing\npython create_Main.py -d /home/syh/train_data/data/sub_train_data/train_data-2018-05-11\n\"\"\"","sub_path":"data_processing/create_Main.py","file_name":"create_Main.py","file_ext":"py","file_size_in_byte":3299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"253021988","text":"#PyBank\nimport os\nimport csv\n\ncsvpath = os.path.join('budget_data.csv')\n\n#create lists\nmonths = []\nprofit_loss = []\nmom_dif = []\n\nwith open(csvpath, newline='') as csvfile:\n\n #specify delimiter and variable that holds\n csvreader = csv.reader(csvfile, delimiter=',')\n\n #skip header row\n csv_header = next(csvreader)\n\n #read rows after header\n for row in csvreader:\n\n #append list with months\n months.append(row[0])\n\n #calculate number of months\n num_months = len(months)\n\n #append list with PnL data\n profit_loss.append(float(row[1]))\n\n #calculate net total of PnL\n net_total = sum(profit_loss)\n\n \n \n #calculate MoM difference in PnL\n for i in range(1, len(profit_loss)):\n\n #append list to hold data\n mom_dif.append((profit_loss[i]) - (profit_loss[i-1]))\n\n #calculate average of list\n avg_mom_dif = round(sum(mom_dif) / len(mom_dif), 2)\n\n #find max MoM change\n max_mom_dif = max(mom_dif)\n\n #find min MoM change\n min_mom_dif = min(mom_dif)\n\n #find max date\n max_mom_dif_date = str(months[mom_dif.index(max(mom_dif)) + 1])\n\n #find min date\n min_mom_dif_date = str(months[mom_dif.index(min(mom_dif)) + 1])\n\n with open(\"PyBank_Analysis.txt\", \"w\") as text_file:\n\n #print analysis to txt file\n print(\"Financial Analysis\", file=text_file)\n print(\"---------------------------------\", file=text_file)\n print(f\"Total Months: {num_months}\", file=text_file)\n print(f\"Total: ${net_total}\", file=text_file)\n print(f\"Average Change: ${avg_mom_dif}\", file=text_file)\n print(f\"Greatest Increase in Profits: {max_mom_dif_date} (${max_mom_dif})\", file=text_file)\n print(f\"Greatest Decrease in Profits: {min_mom_dif_date} (${min_mom_dif})\", file=text_file)\n\nprint(open('PyBank_Analysis.txt').read())","sub_path":"PyBank/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1915,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"520560038","text":"from __future__ import unicode_literals\n\nfrom django.contrib.auth.models import AbstractUser\nfrom django.db import models\n\nfrom application import settings\n\n\nclass User(AbstractUser):\n\n pass\n\n\nclass Authored(models.Model):\n author = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name='Author')\n\n class Meta:\n abstract = True\n\n\nclass Titled(models.Model):\n title = models.CharField(max_length=255, verbose_name='Title')\n\n def get_title(self):\n return self.title\n\n class Meta:\n abstract = True\n\n\nclass Dated(models.Model):\n created = models.DateTimeField(auto_now_add=True, verbose_name='Created')\n updated = models.DateTimeField(auto_now=True, verbose_name='Updated')\n\n class Meta:\n abstract = True\n\n\nclass Deletable(models.Model):\n is_deleted = models.BooleanField(default=False, verbose_name='Is deleted')\n\n class Meta:\n abstract = True\n","sub_path":"core/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":912,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"215958576","text":"##########import module#################\nfrom ecmwfapi import ECMWFDataServer\nfrom calendar import monthrange\nimport numpy as np\nfrom retrieve import retrieve\n########################################\n########################################\ninitial_year=2015 \ninitial_month=9 #Could change these variables to define the period\nend_year=2015 #leap year problem has been taken care of in this script already. \nend_month=9 #Can download only one month data or a long period of data\n########################################\nyear_loop=np.arange(initial_year,end_year+1)#could type help(np.arange) in python prompt to see the syntax of this function for +1\nfor year in year_loop:\n if (year!=year_loop[0]) & (year!=year_loop[-1]):\n month_loop=np.arange(1,12+1)\n elif (year== year_loop[0]) & (year==year_loop[-1]):\n month_loop=np.arange(initial_month,end_month+1)\n elif year==year_loop[-1]:\n month_loop=np.arange(1,end_month+1)\n else:\n month_loop=np.arange(initial_month,12+1)\n\n for month in month_loop:\n max_day=list(monthrange(year,month))[1]\n if month<10:\n date_string=str(year)+\"-0\"+str(month)+\"-01/to/\"+str(year)+\"-0\"+str(month)+\"-\"+str(max_day)\n target=str(year)+'0'+str(month)+\"PV_RTW.nc\" #target is just the name of the file.Change it depending on the parameters you choose in retrieve.py\n retrieve(date_string,target) #PV is potential vorticity, R relative humidity\n else: #T temperature and w vertical velocity \n date_string=str(year)+\"-\"+str(month)+\"-01/to/\"+str(year)+\"-\"+str(month)+\"-\"+str(max_day)\n target=str(year)+str(month)+\"PV_RTW.nc\"\n retrieve(date_string,target)\n","sub_path":"ecwmf/era_interim.py","file_name":"era_interim.py","file_ext":"py","file_size_in_byte":1890,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"593351542","text":"\r\n\r\nfrom Vehicle import Vehicle\r\nfrom Order import Order\r\n\r\nimport numpy as np\r\nimport pickle\r\nclass AbstractWorld:\r\n\t\r\n\tdef __init__(self):\r\n\t\t\r\n\t\t[VF,EF] = pickle.load(open(\"./Classes/data/Lehigh.pickle\",'rb'))\r\n\t\tself.Edges=[] \r\n\t\tfor edge in EF:\r\n\t\t\tself.Edges.append( [ edge[0] , edge[1] ,EF[edge][0 ],EF[edge][1 ],EF[edge][2 ] ])\t\t\t\r\n \r\n\t\t\r\n\t\tself.Verticies=[]\r\n\t\tfor v in VF:\r\n\t\t\tself.Verticies.append( [ v,float(VF[v][0]),float(VF[v][1])] )\t\t\t\r\n\t\t \r\n\t\t\r\n\t\tself.v = [v[0] for v in self.Verticies]\r\n\t\t\r\n\t\tself.vShuffled=[v[0] for v in self.Verticies]\r\n\t\tnp.random.seed(47)\r\n\t\tnp.random.shuffle(self.vShuffled)\r\n\t\t\r\n\t\tself.orderId = 0\r\n\t\tself.map = {\"L1\":[\"A\",\"B\"],\"L2\":[\"C\",\"D\"],\"L3\":[\"E\",\"F\"],\"L4\":[\"G\",\"H\"]}\r\n\t\t\r\n\tdef getLocationOfWarehouses(self):\t\r\n\t\ttypes = [\"A\",\"B\",\"C\",\"D\",\"E\",\"F\",\"G\",\"H\"]\r\n\t\tWarehouses = []\r\n\t\tj = 0\r\n\t\tfor t in types:\r\n\t\t\tfor k in range(2):\r\n\t\t\t\tWarehouses.append({\"type\":t,\"location\":self.vShuffled[j]})\r\n\t\t\t\tj+=1\r\n\t\t\r\n\t\treturn Warehouses\r\n\t\t\r\n\tdef getProductionLines(self):\t\r\n\t\ttypes = [\"L1\",\"L2\",\"L3\",\"L4\"]\r\n\t\tProductionLines = []\r\n\t\tj = 20\r\n\t\tfor t in types:\r\n\t\t\tfor k in range(5):\r\n\t\t\t\tProductionLines.append({\r\n\t\t\t\t\"type\":t,\r\n\t\t\t\t\"location\":self.vShuffled[j],\r\n\t\t\t\t\"capacityOfMaterial[tons]\":100})\r\n\t\t\t\tj+=1\r\n\t\t\r\n\t\treturn ProductionLines\r\n\t\t\r\n\tdef getInitialTruckLocations(self):\r\n\t\tnp.random.seed(47)\r\n\t\tvehicles = []\r\n\t\tnp.random.shuffle(self.v)\r\n\t\r\n\t\tfor i, v in enumerate(self.v):\r\n\t\t\tif i%3 == 0:\r\n\t\t\t\tnewVehicle = Vehicle(i,v)\r\n\t\t\t\tnewVehicle.type=\"Truck\"\r\n\t\t\t\tnewVehicle.capacity = np.random.randint(3,30)\r\n\t\t\t\tvehicles.append(newVehicle)\r\n\t\treturn vehicles\r\n\t\t\r\n\tdef getNewOrdersForGivenTime(self,t):\r\n\t\tnewOrders=[]\r\n\t\tnp.random.seed(t)\r\n\t\tif t > 22*60:\r\n\t\t\treturn newOrders\r\n\t\tif np.random.rand() < 0.3:\r\n\t\t\tn = np.random.randint(1,3)\r\n\t\t\tnp.random.shuffle(self.v)\r\n\t\t\tfor j in xrange(n):\r\n\t\t\t\torder = Order(self.orderId)\r\n\t\t\t\torder.finalLocation=np.random.choice(self.vShuffled[40:])\r\n\t\t\t\t\r\n\t\t\t\tl = [e for e in self.map]\r\n\t\t\t\tnp.random.shuffle(l)\r\n\r\n\t\t\t\tfor j in l[0:np.random.randint(2,4)]:\r\n\t\t\t\t\torder.productionProcess.append({\r\n\t\t\t\t\t\t\"processinLine\":j,\r\n\t\t\t\t\t\t\"resourceNeeded\":np.random.choice(self.map[j]),\r\n\t\t\t\t\t\t\"processingTime\":np.random.randint(5,10),\r\n\t\t\t\t\t\t\"materialNeeded[tons]\":np.random.randint(2,10),\t\t\t\t\t\t\r\n\t\t\t\t\t\t})\r\n\t\t\t\t\r\n\t\t\t\t\r\n\t\t\t\t \r\n\t\t\t\tself.orderId+=1\r\n\t\t\t\tnewOrders.append(order)\r\n\t\treturn newOrders\r\n\t\t \r\n\r\n","sub_path":"AbstractWorld.py","file_name":"AbstractWorld.py","file_ext":"py","file_size_in_byte":2380,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"109768746","text":"import chardet\nimport re\nimport requests\nfrom bs4 import BeautifulSoup\n#import cookielib\n\nurl = 'http://www.cdlr.gov.cn/detailnoright.aspx?id=96813'\nreferer = 'http://www.cdlr.gov.cn/detailnoright.aspx?id=96813'\nuser_agent = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13) AppleWebKit/603.1.13 (KHTML, like Gecko) Version/10.1 Safari/603.1.13'\nheaders = {'User-Agent':user_agent,'Referer':referer}\n#headers = {'User-Agent':user_agent}\nhtml = requests.get(url,headers=headers)\n#print(html.text)\n#print(response.headers)\nhtml.encoding = chardet.detect(html.content)['encoding']\n\nsoup = BeautifulSoup(html.text,'lxml' )\n#print(soup.title.string)\n#print(soup.title.string)\n#for a in soup.find_all(class_ = \"\n# print(repr(a))\n#print('nextsibing:%s'%soup.title.next_sibling)\n#for tag in soup.find_all(True):\n# print(tag.name)\n#print(soup.span.string)\na = soup.find_all(\"span\")\nfor i in a:\n print(i)\n","sub_path":"urllib/local_request.py","file_name":"local_request.py","file_ext":"py","file_size_in_byte":901,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"84576587","text":"\"\"\"\r\nVytvari Gabor priznak.\r\n\r\n@author: Katerina Kratochvilova\r\n\"\"\"\r\n\r\nimport numpy as np\r\nfrom skimage.filter import gabor_kernel #skimage.__version__ if it's lower than 0.11 then use filter instead of filters\r\nfrom skimage.filter import gabor_filter # gabor in newer versions\r\nfrom skimage import io\r\nimport skimage\r\nimport math\r\nimport cv2\r\nfrom scipy import ndimage as ndi\r\n\r\n#parametry Gabor funkce\r\norientations = [0, math.pi/4, math.pi/2, math.pi / 4 * 3]\r\nwavelengths = [0.25, 0.5, 1.0] #[2.0, 2.0 * math.sqrt(2), 4.0] #[4, 4.0 * math.sqrt(2), 8, 8 * math.sqrt(2) ]\r\nsigma = 1\r\n\r\ndef create_kernels():\r\n \"\"\"\r\n Vytvari kernely (filtry) podle zadanych parametru.\r\n \r\n Return: \r\n kernels -- vytvorene kernely (filtry).\r\n \"\"\"\r\n #print skimage.__version__\r\n kernels = []\r\n \r\n for theta in orientations:\r\n #print \"Theta: \", theta\r\n for lambd in wavelengths:\r\n kernel = gabor_kernel(1.0/lambd, theta=theta, sigma_x=sigma, sigma_y=sigma) # frequency is 1/lambda\r\n kernels.append(kernel)\r\n \r\n #print kernel.shape\r\n #print kernel\r\n \r\n #io.imshow(np.real(kernel))\r\n #io.imshow(np.imag(kernel))\r\n #io.imshow(io.concatenate_images([np.real(kernel), np.imag(kernel)])) \r\n #io.show() \r\n \r\n \r\n return kernels \r\n \r\n\r\ndef show_kernels(kernels, real=True): #if False - show imaginary part\r\n \"\"\"\r\n Zobrazi kernely. \r\n \r\n Keyword arguments:\r\n kernels -- kernely.\r\n real -- jaka cast se ma zobrazit, pouze realne kernely nabo imaginarni kernely.\r\n \"\"\"\r\n k_size = kernels[0].shape[0]\r\n img = np.zeros((k_size * len(wavelengths), k_size * len(orientations)))\r\n \r\n #print img.shape\r\n for i in range(len(kernels)):\r\n if real:\r\n k = np.real(kernels[i])\r\n else:\r\n k = np.imag(kernels[i])\r\n n = np.linalg.norm(k) # for normalization\r\n #print \"Norm:\", n\r\n r = i % len(wavelengths)\r\n c = i / len(wavelengths)\r\n img[r * k_size : r*k_size + k_size, c*k_size : c*k_size + k_size] = k / n\r\n \r\n io.imshow(img)\r\n io.show()\r\n \r\ndef process_image(img):\r\n \"\"\"\r\n Pouzije filtry na obrazek.\r\n \r\n Keyword arguments:\r\n img -- vstupni obrazek.\r\n Return:\r\n responses -- vysledna matice po pouzitych filtrech. \r\n \"\"\"\r\n responses = []\r\n \r\n for theta in orientations:\r\n for lambd in wavelengths:\r\n real, imag = gabor_filter(img, 1.0/lambd, theta=theta, sigma_x=sigma, sigma_y=sigma, mode='reflect') # frequency is 1/lambda\r\n responses.append((real, imag))\r\n \r\n #or use ndi.convolve(img, kernel, mode='wrap')\r\n \r\n #print len(responses)\r\n \r\n return responses\r\n\r\n\r\ndef count_magnitude(filtered_img_real, filtered_img_imag):\r\n \"\"\"\r\n Spocita magnitudu.\r\n filtered_img_real -- preklopi na osu x, filtered_img_imag preklopi na osu y tim padem mame vektory a spocitame jejich velikost.\r\n \r\n Keyword arguments:\r\n filtered_img_real -- filtrovane obrazky realna cast. \r\n filtered_img_imag -- filtrovane obrazky imaginarni cast.\r\n Return:\r\n magnitude -- vypoctena magnituda.\r\n \"\"\"\r\n magnitude = []\r\n \r\n for i in range(len(filtered_img_real)):\r\n magnitude.append(cv2.magnitude(filtered_img_real[i], filtered_img_imag[i]))\r\n \r\n return magnitude\r\n\r\ndef count_vector(magnitudes, deep): \r\n \"\"\"\r\n Nasklada vektory za sebe.\r\n \r\n Keyword arguments:\r\n phase -- faze. \r\n deep -- velikost vektrou.\r\n Return:\r\n array_histograms_magnitude -- vypocteny vektor.\r\n \"\"\"\r\n\r\n array_histograms_magnitude = np.zeros(shape=(deep*len(magnitudes)), dtype=np.int)\r\n\r\n for i in range(len(magnitudes)): \r\n for x in range(len(magnitudes[i])):\r\n for y in range(len(magnitudes[i][x])):\r\n value = magnitudes[i].item(x,y) \r\n index = ((i*deep))+int((value/deep)) # (i*deep) potrebuju se posunout na i-ty filtrovany obrazek\r\n array_histograms_magnitude[index]=array_histograms_magnitude[index]+1\r\n return array_histograms_magnitude\r\n\r\ndef count_gabor(img): \r\n \"\"\"\r\n Hlavni metoda pro vypocet gabor.\r\n \r\n Keyword arguments:\r\n img -- vstupni obrazek.\r\n Return:\r\n gabor_vector -- vytvoreny priznak.\r\n \"\"\"\r\n img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\r\n kernels = create_kernels()\r\n img = img.astype(float)\r\n #show_kernels(kernels, real=True)\r\n\r\n responses = process_image(img)\r\n vector_deep = 16\r\n filtered_img_real = []\r\n filtered_img_imag = []\r\n \r\n for item in responses:\r\n filtered_img_real.append(item[0])\r\n filtered_img_imag.append(item[1])\r\n \r\n magnitude = count_magnitude(filtered_img_real, filtered_img_imag)\r\n gabor_vector = count_vector(magnitude, vector_deep)\r\n return gabor_vector\r\n #io.imshow(responses[11][0])\r\n #io.show()\r\n \r\n\r\n\r\n\r\n \r\n\r\n\r\n","sub_path":"cd/src/deskriptors/gabor.py","file_name":"gabor.py","file_ext":"py","file_size_in_byte":5227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"347285988","text":"##21000348 Andre Seo\n## Procedural Journey of Chicken\n\nfrom cs1graphics import*\n\ncanvas = Canvas(1000, 300)\ncanvas.setBackgroundColor(\"light blue\")\ncanvas.setTitle(\"Journey of Chicken\")\ncanvas.wait()\n\n#Create the ground\ndef create_ground():\n ground = Rectangle(1000,100)\n ground.setFillColor(\"light green\")\n ground.move(500,250)\n canvas.add(ground)\n\n#Create the sun\ndef create_sun():\n sun = Circle(50)\n sun.setFillColor(\"red\")\n sun.setBorderColor(\"red\")\n sun.move(25,25)\n canvas.add(sun)\n\n\n#Draw a chicken\ndef make_chicken(hen = False):\n layer = Layer()\n make_body(layer, hen)\n wing = make_wing(layer, hen)\n make_eye(layer, hen)\n make_beak(layer, hen)\n make_dots(layer, hen)\n return layer, wing\n\n\n######################################################\n\n\n#Make a body\ndef make_body(layer, hen = False):\n if hen:\n body = Ellipse(70, 80)\n body.setFillColor(\"white\")\n else:\n body = Ellipse(40, 50)\n body.setFillColor(\"yellow\")\n body.move(0,10)\n body.setBorderColor(\"yellow\")\n body.setDepth(20)\n layer.add(body)\n\n\n#Draw a wing\ndef make_wing(layer, hen = False):\n if hen:\n wing = Ellipse(60, 40)\n wing.setFillColor(\"white\")\n wing.setBorderColor(\"yellow\")\n wing.move(15,20)\n else:\n wing = Ellipse(30,20)\n wing.setFillColor(\"yellow\")\n wing.setBorderColor(\"orange\")\n wing.move(10,20)\n wing.adjustReference(-5, -5)\n wing.setDepth(19)\n layer.add(wing)\n return wing\n\n\n#Draw and eye\ndef make_eye(layer, hen = False):\n if hen:\n eye = Circle(3)\n eye.move(-15,-15)\n else:\n eye = Circle(2)\n eye.move(-5, 0)\n eye.setFillColor(\"black\")\n eye.setDepth(18)\n layer.add(eye)\n\n#Draw a beak\ndef make_beak(layer, hen = False):\n if hen:\n beak = Square(8)\n beak.move(-36, 0)\n else:\n beak = Square(4)\n beak.move(-22, 10)\n beak.rotate(45)\n beak.setFillColor(\"orange\")\n beak.setBorderColor(\"orange\")\n beak.setDepth(21)\n layer.add(beak)\n\n\n#Draw dots\ndef make_dots(layer, hen = False):\n if hen:\n head1 = Ellipse(5,8)\n head1.setFillColor(\"red\")\n head1.setBorderColor(\"red\")\n head1.move(600,158)\n head1.setDepth(22)\n layer.add(head1)\n\n head2 = Ellipse(5,8)\n head2.setFillColor(\"red\")\n head2.setBorderColor(\"red\")\n head2.move(594,458)\n head2.setDepth(22)\n layer.add(head2)\n\n######################################################\n\n \n#Draw the group\ndef make_group():##mother hen + baby chicken1\n group = Layer()\n mother_hen, wing = make_chicken(True)\n group.add(mother_hen)\n (chicken1, wing1) = make_chicken()\n chicken1.move(120, 0)\n group.add(chicken1)\n group.move(600,200)\n return group\n\n\ndef make_family():\n group = make_group()\n chicken2, wing2 = make_chicken()\n chicken2.move(800,200)\n canvas.add(group)\n canvas.add(chicken2)\n return group, chicken2, wing2\n\n\n#Create the scene\ndef create_scene():\n create_ground()\n create_sun()\n group, chicken2, wing2 = make_family()\n return group, chicken2, wing2\n\n\ndef move_group(group):\n for i in range(80):\n group.move(-5,-2)\n group.move(-5,2)\n if i == 30:\n text1 = Text(\"OH!\", 20)\n text1.move(800, 160)\n canvas.add(text1)\n elif i == 40:\n canvas.remove(text1)\n text2 = Text(\"WHERE IS MY MOMMY GOING?\", 30)\n text2.move(500, 110)\n canvas.add(text2)\n elif i == 55:\n canvas.remove(text2)\n\n\ndef move_chicken2(chicken2, wing2):\n for i in range(10):\n text3 = Text(\"Waif for ME~\", 25)\n text3.move(500, 110)\n canvas.add(text3)\n for k in range(5):\n chicken2.move(-10, -20)\n wing2.rotate(-10)\n for k in range(5):\n chicken2.move(-10, 20)\n wing2.rotate(10)\n\n\ndef animate_chicken_family():\n group, chicken2, wing2 = create_scene()\n move_group(group)\n move_chicken2(chicken2, wing2)\n\n\ndef main():\n animate_chicken_family()\n canvas.wait()\n canvas.close()\n\n\nmain()\n\n\n\n\n\n\n \n","sub_path":"Homework #12/1 PROCEDURAL_CHICKEN.py","file_name":"1 PROCEDURAL_CHICKEN.py","file_ext":"py","file_size_in_byte":4210,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"610586710","text":"import webbrowser\n\nclass Movie():\n \"\"\" This class provides a way to store movie related information\"\"\"\n \n VALID_RATINGS = [\"G\", \"PG\", \"PG-13\", \"R\"]\n \n def __init__(self, movieTitle, movieStoryline, posterImage, trailerYoutube):\n self.title = movieTitle\n self.storyline = movieStoryline\n self.poster_image_url = posterImage\n self.trailer_youtube_url = trailerYoutube\n\n def showTrailer(self):\n webbrowser.open(self.trailerYoutubeUrl)\n\n\n\n\n\n\n\n\n\n \n","sub_path":"Fundamentos de Programação/movies/media.py","file_name":"media.py","file_ext":"py","file_size_in_byte":498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"244015328","text":"import discord\n\nfrom ..objects.bot import Menel\n\n\ndef setup(bot: Menel):\n @bot.event\n async def on_raw_reaction_add(payload: discord.RawReactionActionEvent):\n if payload.user_id == bot.OWNER and payload.emoji.name == '🗑️':\n try:\n message = await (await bot.fetch_channel(payload.channel_id)).fetch_message(payload.message_id)\n if message.author == bot.user:\n await message.delete()\n except discord.HTTPException as e:\n print(e)","sub_path":"Menel/handlers/reaction_add.py","file_name":"reaction_add.py","file_ext":"py","file_size_in_byte":531,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"367476981","text":"import numpy as np\nimport cv2\nimport os\nimport time\nimport numpy as np\nfrom imutils import face_utils\nimport imutils\n\ndef increase_brightness(img, value=30):\n hsv = cv2.cvtColor(img, cv2.COLOR_BGR2HSV)\n h, s, v = cv2.split(hsv)\n\n lim = 255 - value\n v[v > lim] = 255\n v[v <= lim] += value\n\n final_hsv = cv2.merge((h, s, v))\n img = cv2.cvtColor(final_hsv, cv2.COLOR_HSV2BGR)\n return img\n\ndef main():\n net = cv2.dnn.readNet(\"YOLOFI2.weights\", \"YOLOFI.cfg\")\n cap = cv2.VideoCapture(\"test.mp4\")\n classes = []\n l = 1\n with open(\"obj.names\", \"r\")as f:\n classes = [line.strip() for line in f.readlines()]\n layers_names = net.getLayerNames()\n outputlayers = [layers_names[i[0] - 1] for i in net.getUnconnectedOutLayers()]\n colors = np.random.uniform(0, 255, size=(len(classes), 3))\n font = cv2.FONT_HERSHEY_PLAIN\n frame_id = 0\n dd = -1\n time_now = time.time()\n frame_id = 0\n err = 0\n\n pred_corrects = []\n while True:\n _, frame = cap.read()\n frame_id += 1\n beltcornerdetected = False\n beltdetected = False\n if frame is None:\n break\n height, width, channels = frame.shape\n\n # # Type you code here\n frame = frame[:, 90:-50]\n frame = increase_brightness(frame, 10)\n frame = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n # frame = cv2.bilateralFilter(frame, d=-1, sigmaSpace=20, sigmaColor=5)\n clahe = cv2.createCLAHE(clipLimit=10.0, tileGridSize=(16,16))\n frame = clahe.apply(frame)\n frame = cv2.cvtColor(frame,cv2.COLOR_GRAY2RGB)\n\n blob = cv2.dnn.blobFromImage(frame, 0.00392, (480, 480), (0, 0, 0), True, crop=False)\n net.setInput(blob)\n outs = net.forward(outputlayers)\n class_ids = []\n boxes = []\n shape = []\n confidence = 0\n for out in outs:\n for detection in out:\n scores = detection[5:]\n class_id = np.argmax(scores)\n confidence = scores[class_id]\n\n if confidence > 0.2:\n center_x = int(detection[0] * width)\n center_y = int(detection[1] * height)\n w = int(detection[2] * width)\n h = int(detection[3] * height)\n x = int(center_x - w / 2)\n y = int(center_y - h / 2)\n cv2.rectangle(frame, (x, y), (x + w, y + h), (0, 255, 0), 2)\n if class_id == 1:\n beltcornerdetected = True\n elif class_id == 0:\n beltdetected = True\n\n # output whether prediction is correct\n print(frame_id)\n is_correct = 0\n if beltdetected and frame_id <= 125:\n is_correct = 1\n elif not beltdetected and frame_id > 125:\n is_correct = 1\n print(is_correct)\n pred_corrects.append(is_correct)\n # print(beltdetected)\n cv2.imshow(\"Image\", frame)\n key = cv2.waitKey(1)\n if key == 27:\n break\n print(\"Accuracy: {}\".format(sum(pred_corrects) / len(pred_corrects)))\n cap.release()\n cv2.destroyAllWindows()\n\n\nif __name__ == '__main__':\n main()","sub_path":"BeltDetectionLab2/BeltDetectionLab2.py","file_name":"BeltDetectionLab2.py","file_ext":"py","file_size_in_byte":3512,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"9339376","text":"from collections import deque\n\n\ndef conv_list(n):\n adj_matrix = []\n for _ in range(n):\n adj_list = input().split()\n adj_list = list(map(int, adj_list))\n adj_row = [False] * n\n if adj_list[1] > 0:\n for i in adj_list[2:]:\n adj_row[i - 1] = True\n adj_matrix.append(adj_row)\n return adj_matrix\n\n\ndef depth_first_search(adj_matrix, n):\n visited = [None] * n\n count = 1\n searched = [None] * n\n start = 0\n stack = deque()\n stack.append(start)\n visited[start] = 1\n while count <= 2 * n:\n has_edge = adj_matrix[start]\n i = 0\n while i < n:\n if has_edge[i] and visited[i] is None:\n stack.append(i)\n count += 1\n visited[i] = count\n break\n i += 1\n if i < n:\n start = i\n else:\n searched_node = stack.pop()\n count += 1\n searched[searched_node] = count\n if len(stack) > 0:\n start = stack[-1]\n elif count < 2 * n:\n start = visited.index(None)\n stack.append(start)\n count += 1\n visited[start] = count\n else:\n break\n return visited, searched\n\n\ndef main():\n n = int(input())\n adj_matrix = conv_list(n)\n visited, searched = depth_first_search(adj_matrix, n)\n for i in range(n):\n print(\"{} {} {}\".format(i + 1, visited[i], searched[i]))\n\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"Python_codes/p02238/s659849161.py","file_name":"s659849161.py","file_ext":"py","file_size_in_byte":1556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"123825058","text":"# 5494. 统计所有可行路径.py\nfrom functools import lru_cache\nfrom typing import List\n\n\nclass Solution:\n def countRoutes(self, locations: List[int], start: int,\n finish: int, fuel: int) -> int:\n mod = 10 ** 9 + 7\n begin, end = locations[start], locations[finish]\n if abs(begin - end) > fuel:\n return 0\n n = len(locations)\n\n @lru_cache(None)\n def dfs(i, j, pre):\n if pre == 0:\n if i == j:\n return 1\n return 0\n res = 0\n if i == j:\n res += 1\n for k in range(n):\n if k != i:\n need = abs(locations[i] - locations[k])\n if pre >= need:\n res += dfs(k, j, pre - need)\n return res\n\n return dfs(start, finish, fuel) % mod\n\n\ndef main():\n sol = Solution()\n res = sol.countRoutes([4, 3, 1], 1, 0, 6)\n print(res)\n res = sol.countRoutes([1, 2, 3], 0, 2, 40)\n print(res)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Week14/双周赛/5494. 统计所有可行路径.py","file_name":"5494. 统计所有可行路径.py","file_ext":"py","file_size_in_byte":1092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"383202199","text":"# general imports\nimport os\nimport sys\n\n# smartsim and smartredis imports\nfrom smartsim import Experiment\nfrom smartsim.settings import MpiexecSettings\nfrom smartsim.settings import MpirunSettings\n\n# Define function to parse node list\ndef parseNodeList(fname):\n with open(fname) as file:\n nodelist = file.readlines()\n nodelist = [line.rstrip() for line in nodelist]\n nodelist = [line.split('.')[0] for line in nodelist]\n nNodes = len(nodelist)\n return nodelist, nNodes\n\n# Parse command line arguments\nnodes = int(sys.argv[1])\nppn = int(sys.argv[2])\nsimprocs = int(sys.argv[3])\nsimprocs_pn = int(sys.argv[4])\ndbprocs_pn = int(sys.argv[5])\ndevice = sys.argv[6]\nlogging = sys.argv[7]\nhostfile = sys.argv[8]\n\n# Get nodes of this allocation (job) and split them between the tasks\nnodelist, nNodes = parseNodeList(hostfile)\nprint(f\"\\nRunning on {nNodes} total nodes on Polaris\")\nprint(nodelist, \"\\n\")\nhosts = ','.join(nodelist)\n\n# Initialize the SmartSim Experiment\nPORT = 6780\nexp = Experiment(\"inference-example\", launcher=\"pbs\")\n\n# Set the run settings, including the Fortran executable and how to run it\nFtn_exe = './src/inferenceFtn.exe'\nif (simprocs_pn self.time_samples:\n rand_start = np.random.randint(0, max_len - self.time_samples)\n noise_waveform = numpy_waveform[\n rand_start:rand_start + self.time_samples]\n np_noise_wav = np.array(noise_waveform)\n noise_wav = torch.tensor(np_noise_wav, dtype=torch.float32)\n return self.safe_pad(noise_wav)\n\n def __getitem__(self, idx):\n if self.augment:\n the_time = int(np.modf(time())[0] * 100000000)\n np.random.seed(the_time)\n\n example_sources_paths = self.sources_paths[idx]\n\n _, noise_waveform = wavfile.read(example_sources_paths['noise_path'])\n _, speech_waveform = wavfile.read(example_sources_paths['speech_path'])\n noise_wav = self.get_padded_tensor(noise_waveform)\n speech_wav = self.get_padded_tensor(speech_waveform)\n\n # Also draw a random noise waveform if available.\n if self.extra_noise_paths:\n file_idx = np.random.randint(0, len(self.extra_noise_paths))\n _, extra_noise_waveform = wavfile.read(\n self.extra_noise_paths[file_idx])\n extra_noise_wav = self.get_padded_tensor(extra_noise_waveform)\n else:\n extra_noise_wav = torch.zeros_like(noise_wav)\n\n return speech_wav, noise_wav, extra_noise_wav\n\n def get_generator(self, batch_size=4, shuffle=True, num_workers=4):\n generator_params = {'batch_size': batch_size,\n 'shuffle': shuffle,\n 'num_workers': num_workers,\n 'drop_last': True}\n return torch.utils.data.DataLoader(self, **generator_params,\n pin_memory=True)\n\n\ndef test_generator():\n dataset_root_p = '/mnt/data/ChunkFedEnhance/'\n batch_size = 3\n sample_rate = 16000\n timelength = 4.0\n speaker_ids = [x for x in range(2)]\n time_samples = int(sample_rate * timelength)\n max_abs_snr = 5.\n data_loader = Dataset(\n root_dirpath=dataset_root_p,\n speaker_ids=speaker_ids,\n available_speech_percentage=0.5,\n split='test', sample_rate=sample_rate, timelength=timelength,\n zero_pad=True, augment=True)\n generator = data_loader.get_generator(batch_size=batch_size, num_workers=1)\n\n for speech_wavs, noise_wavs, extra_noise_wavs in generator:\n assert speech_wavs.shape == (batch_size, time_samples)\n assert noise_wavs.shape == (batch_size, time_samples)\n assert extra_noise_wavs.shape == (batch_size, time_samples)\n break\n\nif __name__ == \"__main__\":\n test_generator()\n","sub_path":"fedenhance/dataset_loader/chunked_libri_fsd.py","file_name":"chunked_libri_fsd.py","file_ext":"py","file_size_in_byte":7213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"391339493","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2018/7/31 21:18\n# @Author : yimi\n# @File : send_email.py\n# 构造邮件内容 MIMETEXT\n# 发送邮件:登录SMTP服务器(账号,密码) --> 发送 --> 关闭链接 smtplib\nimport smtplib\nimport os\nfrom email.mime.text import MIMEText\nfrom email.mime.multipart import MIMEMultipart # 声明这个邮件是由多部分组成的\nfrom email.mime.application import MIMEApplication\nfrom task_0810_database.config_file import read_path\n\nclass SendEmail:\n def __init__(self, subject, sender, receiver, mail_text, mail_attach, auth_code):\n self.subject = subject\n self.sender = sender\n self.receiver = receiver\n self.mail_text = mail_text\n self.mail_attach = mail_attach\n self.auth_code = auth_code\n self.msg = MIMEMultipart()\n\n def send_email(self):\n # msg=MIMEMultipart()\n self.msg['Subject'] = self.subject\n self.msg['From'] = self.sender\n self.msg['To'] = self.receiver\n\n # 纯文本邮件体\n msg_text=MIMEText(self.mail_text)\n self.msg.attach(msg_text)\n\n # 附加的附件xls, html, rar\n msg_attachment=MIMEApplication(open(self.mail_attach, 'rb').read())\n msg_attachment.add_header('Content-Disposition','attachment',filename=self.mail_attach)\n self.msg.attach(msg_attachment)\n\n # 登录服务器\n s=smtplib.SMTP_SSL('smtp.qq.com', 465)\n s.login(self.sender, self.auth_code) # 账号, 授权码\n s.sendmail(self.sender,self.receiver, self.msg.as_string())\n s.close()\n\n\nif __name__==\"__main__\":\n subject = '薏米的测试报告'\n sender = '1281018605@qq.com'\n receiver = '1281018605@qq.com'\n # receiver = '281417558@qq.com'\n mail_text = 'api test result'\n html_path = os.path.join(read_path.test_report_path, 'api_test_result.html')\n mail_attach = html_path\n auth_code = 'pxqswpbzogoohahb'\n SendEmail(subject, sender, receiver, mail_text, mail_attach, auth_code).send_email()\n","sub_path":"Interface_AutoTest_07/common/send_email.py","file_name":"send_email.py","file_ext":"py","file_size_in_byte":2044,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"131495016","text":"# Function and PyFunction\nfrom core import aobjects as obj\nimport types as t\nfrom core.error import AILRuntimeError\nimport objects\nimport inspect\nfrom . import types\n\n\ndef pyfunc_func_init(self :obj.AILObject, func :t.FunctionType):\n self['__pyfunction__'] = func\n self['__value__'] = func\n\n\ndef pyfunc_func_call(self :obj.AILObject, *args) -> obj.AILObject:\n fobj = self['__pyfunction__']\n\n try:\n rtn = fobj(*args)\n if rtn is None:\n return obj.null\n return obj.ObjectCreater.new_object(objects.wrapper.WRAPPER_TYPE, rtn)\n except Exception as e:\n return AILRuntimeError(str(e), 'PythonError')\n\n\ndef pyfunc_func_str(self :obj.AILObject):\n return '' % (\n self['__pyfunction__'].__name__, hex(id(self['__pyfunction__'])))\n\n\ndef func_func_init(self, cobj :t.CodeType, globals :dict, name :str):\n self['__code__'] = cobj\n self['__globals__'] = globals\n self['__name__'] = name\n\n\ndef func_func_str(self :obj.AILObject):\n return '' % (self['__name__'], hex(id(self)))\n\n\ndef call(pyfw :obj.AILObject, *args):\n if inspect.isfunction(pyfw):\n try:\n return pyfw(*args)\n except Exception as e:\n return AILRuntimeError(str(e), 'PythonError')\n\n if not isinstance(pyfw, obj.AILObject):\n return AILRuntimeError('Cannot call an object that is not AILObject', 'TypeError')\n\n if pyfw['__class__'] not in (PY_FUNCTION_TYPE, FUNCTION_TYPE):\n return AILRuntimeError('%s is not callable' % pyfw['__class__'], 'TypeError')\n\n cfunc = pyfw['__pyfunction__']\n cfunc(*args)\n\n try:\n return cfunc(*args)\n except Exception as e:\n return AILRuntimeError(str(e), 'PythonError')\n\n\nFUNCTION_TYPE = obj.AILObjectType('', types.I_FUNC_TYPE,\n __init__=func_func_init,\n __str__=func_func_str,\n __repr__=func_func_str)\n\nPY_FUNCTION_TYPE = obj.AILObjectType('', types.I_PYFUNC_TYPE,\n __init__=pyfunc_func_init,\n __call__=pyfunc_func_call,\n __str__=pyfunc_func_str,\n __repr__=pyfunc_func_str)\n\n\ndef convert_to_func_wrapper(pyf):\n import inspect\n\n if obj.compare_type(pyf, PY_FUNCTION_TYPE) or \\\n obj.compare_type(pyf, FUNCTION_TYPE):\n return pyf\n\n if inspect.isfunction(pyf) or inspect.isbuiltin(pyf):\n return obj.ObjectCreater.new_object(\n PY_FUNCTION_TYPE, pyf)\n\n return obj.convert_to_ail_object(pyf)\n","sub_path":"objects/function.py","file_name":"function.py","file_ext":"py","file_size_in_byte":2742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"281992974","text":"# import os\n\nfname = \"lesson_64_sample.txt\"\n\n# try:\n# jabber = open(fname, 'r')\n# # jabber = open(\"C:\\\\Documents and Settings\\\\tim\\\\My Documents\\\\sample.txt\", 'r')\n\n# except IOError:\n# print(\"Could not read file: \", fname)\n\n# for line in jabber:\n# if \"jabberwock\" in line.lower():\n# print(line, end='')\n\n# jabber.close()\n\n# with open(fname, 'r') as jabber:\n# for line in jabber:\n# if \"JAB\" in line.upper():\n# print(line, end='')\n\n\n# with open(fname, 'r') as jabber:\n# line = jabber.readline()\n# while line:\n# print(line, end='') # end = \"\" removes newline \n# line = jabber.readline()\n\n# with open(fname, 'r') as jabber:\n# lines = jabber.readlines()\n# print(lines)\n\n# for line in lines:\n# print(line, end='')\n\n\n# with open(fname, 'r') as jabber:\n# lines = jabber.readlines()\n# print(lines)\n\n# for line in lines[::-1]:\n# print(line, end='')\n\nwith open(fname, 'r') as jabber:\n lines = jabber.read()\n\nfor line in lines[::-1]:\n print(line, end='')\n","sub_path":"Complete Python Masterclass/Source_Codes/lesson_64_fileio.py","file_name":"lesson_64_fileio.py","file_ext":"py","file_size_in_byte":1037,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"528718241","text":"from OpenGL.GL import *\nimport weakref\nimport numpy as np\nimport xml.etree.ElementTree as ET\nfrom PyQt5.QtWidgets import QApplication\nfrom PyQt5.QtWidgets import QOpenGLWidget\nfrom PyQt5.QtCore import QTimer\nfrom PyQt5.QtGui import QSurfaceFormat\nfrom PyQt5.QtGui import QOpenGLVersionProfile\nfrom PyQt5.QtCore import Qt\nimport REESSimulation.api as API\nimport REESSimulation.xml as XML\nimport REESUtility.util as UTIL\nimport REESVisualization.xml as VISXML\nfrom REESVisualization.types import *\nimport REESMesh.factory as MESH_FACTORY\n\n\nOpenGL.ERROR_CHECKING = True\nOpenGL.CHECK_CONTEXT = True\n\n\nclass RenderWidget(QOpenGLWidget):\n\n def __init__(self):\n super(QOpenGLWidget, self).__init__()\n self.resize(400, 400)\n self.m_update_timer = QTimer()\n self.m_update_timer.timeout.connect(self.simulation_step)\n\n self.gl = 0\n\n self.camera = None\n self.light_setup = None\n self.grid = None\n self.clear_color = None\n self.materials = None\n self.scene_graphs = []\n self.movie_recorder = None\n self.contact_scale = 0.5\n\n self.keep_visualization = False # Controls if visualization is loaded whening opening an XML file.\n\n self.engine = None\n\n self.trackball = Trackball() # A track ball used to convert mouse move events into rotations\n self.dolly_mode = False # If on then up-down mouse moves make one moves back and forth towards target\n self.pan_mode = False # If on then one translates in the screen space\n self.trackball_mode = False # If on then one rotates around the camera center\n self.fpv_mode = False # If on then one rotates around the camera position\n self.selection_mode = False # If on then one is making a selection of an object\n self.dolly_sensitivity = 0.025\n self.pan_sensitivity = 0.025\n self.anchor_x = None # Original x-position when doing a mouse operation\n self.anchor_y = None # Original t-position when doing a mouse operation\n self.anchor_eye = None\n self.anchor_center = None\n self.anchor_up = None\n self.height = None # Used to convert from window space into screen space when clicking inside the window\n\n\n def compute_normalized_device_coordinates(self, sx, sy):\n '''\n\n :param sx: X-coordinate of screen position (x-pixel value)\n :param sy: Y-coordinate of screen position (y-pixel value)\n :return:\n '''\n viewport = glGetFloatv(GL_VIEWPORT)\n ratio = self.devicePixelRatio()\n nx = (2.0 * ratio * sx) / viewport[2] - 1.0\n ny = (2.0 * ratio * sy) / viewport[3] - 1.0\n nx = max(-1.0, min(1.0, nx))\n ny = max(-1.0, min(1.0, ny))\n return nx, ny\n\n def initializeGL(self):\n version_profile = QOpenGLVersionProfile()\n version_profile.setVersion(4, 1)\n version_profile.setProfile(QSurfaceFormat.CoreProfile)\n self.gl = self.context().versionFunctions(version_profile)\n if not self.gl:\n raise RuntimeError(\"unable to apply OpenGL version profile\")\n self.gl.initializeOpenGLFunctions()\n\n camera, light_setup, materials, clear_color, grid, mov_rec, contact_scale = VISXML.load('resources/scene.xml')\n\n self.camera = camera\n self.light_setup = light_setup\n self.materials = materials\n self.clear_color = clear_color\n self.grid = grid\n self.movie_recorder = mov_rec\n self.contact_scale = contact_scale\n\n self.scene_graphs.clear()\n self.create_grid_scene_graph()\n\n glClearColor(self.clear_color[0], self.clear_color[1], self.clear_color[2], 0.0)\n glEnable(GL_DEPTH_TEST)\n glDepthFunc(GL_LESS)\n glEnable(GL_CULL_FACE)\n\n def create_rigid_bodies_graph(self):\n\n scene_graph = SceneGraph()\n\n scene_graph.camera = self.camera\n scene_graph.light_setup = self.light_setup\n\n vertex_shader = Shader('resources/shaders/vertex.glsl', GL_VERTEX_SHADER)\n geometry_shader = Shader('resources/shaders/geometry.glsl', GL_GEOMETRY_SHADER)\n fragment_shader = Shader('resources/shaders/fragment.glsl', GL_FRAGMENT_SHADER)\n\n scene_graph.program = ShaderProgram([vertex_shader, geometry_shader, fragment_shader])\n\n for shape in self.engine.shapes.values():\n vao = VAO()\n vertex_data, index_data = create_mesh_array_data(shape.mesh)\n vbo = VBO(vertex_data, index_data)\n shape_node = ShapeNode(scene_graph.program, vao, vbo)\n scene_graph.shape_nodes[shape.name] = shape_node\n\n for body in self.engine.rigid_bodies.values():\n shape_node = scene_graph.shape_nodes[body.shape.name]\n instance = InstanceNode(shape_node, self.materials[body.visual_material], body)\n instance.update_transform(body.r, body.q)\n scene_graph.instance_nodes[body.name] = instance\n\n self.scene_graphs.append(scene_graph)\n\n def create_grid_scene_graph(self):\n scene_graph = SceneGraph()\n\n scene_graph.camera = self.camera\n scene_graph.light_setup = self.light_setup\n\n vertex_shader = Shader('resources/shaders/debug_vertex.glsl', GL_VERTEX_SHADER)\n fragment_shader = Shader('resources/shaders/debug_fragment.glsl', GL_FRAGMENT_SHADER)\n\n scene_graph.program = ShaderProgram([vertex_shader, fragment_shader])\n\n vao = VAO()\n vertex_data, index_data = create_wire_grid_data(self.grid)\n vbo = VBO(vertex_data, index_data, make_triangles=False)\n shape_node = ShapeNode(scene_graph.program, vao, vbo)\n instance = InstanceNode(shape_node, self.materials['grid'])\n scene_graph.instance_nodes['grid'] = instance\n\n self.scene_graphs.append(scene_graph)\n\n def create_contact_points_graph(self):\n scene_graph = SceneGraph()\n\n scene_graph.camera = self.camera\n scene_graph.light_setup = self.light_setup\n\n vertex_shader = Shader('resources/shaders/debug_vertex.glsl', GL_VERTEX_SHADER)\n fragment_shader = Shader('resources/shaders/debug_fragment.glsl', GL_FRAGMENT_SHADER)\n\n scene_graph.program = ShaderProgram([vertex_shader, fragment_shader])\n\n vao = VAO()\n slices = 12\n segments = 12\n\n H = 1.0/(1.0 - 1.0/8.0)\n\n base = MESH_FACTORY.make_sphere(H/8.0,slices, segments)\n shaft = MESH_FACTORY.make_cylinder(H/16.0, 5.0*H/8.0, slices)\n head = MESH_FACTORY.make_cone(H/8.0, 2.0*H/8.0, slices)\n MESH.translate(shaft, V3.make(0.0, 5.0*H/16.0, 0.0))\n MESH.translate(head, V3.make(0.0, 5.0*H/8.0, 0.0))\n\n mesh = MESH.join([base, shaft, head])\n\n vertex_data, index_data = create_mesh_array_data(mesh)\n vbo = VBO(vertex_data, index_data)\n shape_node = ShapeNode(scene_graph.program, vao, vbo)\n instance = ContactPointsInstanceNode(shape_node, self.materials['contact'], self.contact_scale, self.engine)\n scene_graph.instance_nodes['contact'] = instance\n\n self.scene_graphs.append(scene_graph)\n\n def create_joint_connectors_graph(self):\n scene_graph = SceneGraph()\n\n scene_graph.camera = self.camera\n scene_graph.light_setup = self.light_setup\n\n vertex_shader = Shader('resources/shaders/debug_vertex.glsl', GL_VERTEX_SHADER)\n fragment_shader = Shader('resources/shaders/debug_fragment.glsl', GL_FRAGMENT_SHADER)\n\n scene_graph.program = ShaderProgram([vertex_shader, fragment_shader])\n\n vao = VAO()\n slices = 12\n segments = 12\n\n H = 1.0/(1.0 - 1.0/8.0)\n\n base = MESH_FACTORY.make_sphere(H/8.0,slices, segments)\n shaft = MESH_FACTORY.make_cylinder(H/16.0, 5.0*H/8.0, slices)\n head = MESH_FACTORY.make_cone(H/8.0, 2.0*H/8.0, slices)\n MESH.translate(shaft, V3.make(0.0, 5.0*H/16.0, 0.0))\n MESH.translate(head, V3.make(0.0, 5.0*H/8.0, 0.0))\n\n mesh = MESH.join([base, shaft, head])\n\n vertex_data, index_data = create_mesh_array_data(mesh)\n vbo = VBO(vertex_data, index_data)\n shape_node = ShapeNode(scene_graph.program, vao, vbo)\n\n if 'socket' in self.materials:\n socket_material = self.materials['socket']\n else:\n raise RuntimeError('create_joint_connectors_graph(): Missing socket visual material')\n\n if 'plug' in self.materials:\n plug_material = self.materials['plug']\n else:\n raise RuntimeError('create_joint_connectors_graph(): Missing plug visual material')\n\n instance = JointConnectorsInstanceNode(shape_node,\n socket_material,\n plug_material,\n self.contact_scale,\n self.engine\n )\n scene_graph.instance_nodes['connectors'] = instance\n\n self.scene_graphs.append(scene_graph)\n\n def paintGL(self):\n glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT)\n for scene_graph in self.scene_graphs:\n scene_graph.render()\n\n def resizeGL(self, width, height):\n glViewport(0, 0, width, height)\n self.height = height\n\n def mouseMoveEvent(self, e):\n x = e.x()\n y = self.height - e.y()\n\n nx, ny = self.compute_normalized_device_coordinates(x, y)\n\n left = (int(e.buttons()) & Qt.LeftButton) != 0\n middle = (int(e.buttons()) & Qt.MidButton) != 0\n right = (int(e.buttons()) & Qt.RightButton) != 0\n ctrl = (int(QApplication.keyboardModifiers()) & Qt.ControlModifier) != 0\n shift = (int(QApplication.keyboardModifiers()) & Qt.ShiftModifier) != 0\n alt = (int(QApplication.keyboardModifiers()) & Qt.AltModifier) != 0\n\n self.camera.update(self.anchor_eye, self.anchor_center, self.anchor_up)\n\n if self.dolly_mode:\n distance = self.dolly_sensitivity * (y - self.anchor_y)\n self.camera.dolly(-distance)\n\n if self.pan_mode:\n x_distance = self.pan_sensitivity * (x - self.anchor_x)\n y_distance = self.pan_sensitivity * (y - self.anchor_y)\n self.camera.pan(-x_distance, -y_distance)\n\n if self.trackball_mode:\n self.trackball.move_to(nx, ny)\n# self.trackball.move_to(x, y)\n self.camera.orbit(self.trackball.rotation_matrix.transpose())\n\n if self.fpv_mode:\n self.trackball.move_to(nx, ny)\n self.camera.rotate(self.trackball.rotation_matrix)\n\n if self.selection_mode:\n p, r = self.camera.get_ray(nx, ny)\n #self.select_tool.move_selection(p, r, self.camera.dof, self.engine)\n #update_render_manager(self.render_manager, self.engine)\n\n if not self.m_update_timer.isActive():\n self.update()\n\n def mousePressEvent(self, e):\n x = e.x()\n y = self.height - e.y()\n\n nx, ny = self.compute_normalized_device_coordinates(x, y)\n\n left = (int(e.buttons()) & Qt.LeftButton) != 0\n middle = (int(e.buttons()) & Qt.MidButton) != 0\n right = (int(e.buttons()) & Qt.RightButton) != 0\n ctrl = (int(QApplication.keyboardModifiers()) & Qt.ControlModifier) != 0\n shift = (int(QApplication.keyboardModifiers()) & Qt.ShiftModifier) != 0\n alt = (int(QApplication.keyboardModifiers()) & Qt.AltModifier) != 0\n\n if alt and left:\n self.dolly_mode = True\n elif shift and left:\n self.pan_mode = True\n self.camera.center_locked = False\n elif ctrl and left:\n self.selection_mode = True\n elif left:\n self.trackball_mode = True\n elif right:\n self.fpv_mode = True\n self.camera.center_locked = False\n\n self.trackball.reset()\n self.anchor_x = x\n self.anchor_y = y\n self.anchor_eye = np.copy(self.camera.eye)\n self.anchor_center = np.copy(self.camera.center)\n self.anchor_up = np.copy(self.camera.up)\n\n if self.trackball_mode:\n self.trackball.click_at(nx, ny)\n\n if self.fpv_mode:\n self.trackball.click_at(nx, ny)\n\n if self.selection_mode:\n p,r = self.camera.get_ray(nx, ny)\n #self.select_tool.select(p, r, self.engine )\n self.selection_mode = True\n\n def mouseReleaseEvent(self, e):\n #if self.selection_mode:\n # self.select_tool.deselect()\n self.dolly_mode = False\n self.pan_mode = False\n self.trackball_mode = False\n self.fpv_mode = False\n self.camera.center_locked = True\n self.selection_mode = False\n\n def simulation_step(self):\n did_step = API.run(self.engine)\n\n if did_step and self.movie_recorder is not None:\n self.movie_recorder.record(self)\n\n self.update()\n\n def open_file(self, filename):\n xml = ET.parse(filename)\n root = xml.getroot()\n\n if not self.keep_visualization:\n camera, light_setup, materials, clear_color, grid, mov_rec, contact_scale = VISXML.load_from_elementtree(root)\n\n self.camera = camera\n self.light_setup = light_setup\n self.materials = materials\n self.clear_color = clear_color\n self.grid = grid\n self.movie_recorder = mov_rec\n self.contact_scale = contact_scale\n\n self.scene_graphs.clear()\n self.create_grid_scene_graph()\n\n self.engine = XML.load_from_elementtree(root)\n self.create_rigid_bodies_graph()\n self.create_contact_points_graph()\n self.create_joint_connectors_graph()\n\n glClearColor(self.clear_color[0], self.clear_color[1], self.clear_color[2], 0.0)\n\n self.m_update_timer.start(1000/self.engine.solver_params.fps)\n\n self.update()\n\n def save_file(self, filename):\n root = ET.Element('scene')\n\n VISXML.save_to_elementtree(self.camera,\n self.light_setup,\n self.materials,\n self.clear_color,\n self.grid,\n self.movie_recorder,\n self.contact_scale,\n root\n )\n\n XML.save_to_elementtree(self.engine, root)\n\n UTIL.xml_pretty_indent(root)\n tree = ET.ElementTree(root)\n tree.write(filename)\n\n def set_keep_visualization(self, keep_it):\n self.keep_visualization = keep_it\n if keep_it:\n print('RenderWidget: Keeping visualization on file open')\n else:\n print('RenderWidget: Loading visualization on file open')\n\n\ndef initialize_opengl():\n format = QSurfaceFormat()\n format.setProfile(QSurfaceFormat.CoreProfile)\n format.setVersion(4, 5)\n format.setSamples(4)\n QSurfaceFormat.setDefaultFormat(format)\n\n","sub_path":"REESVisualization/widget.py","file_name":"widget.py","file_ext":"py","file_size_in_byte":15149,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"523082594","text":"class Solution:\n def letterCasePermutation(self, S: str):\n if len(S) == 0:\n return ''\n letters = [ele for ele in S if ele.isalpha()]\n pos = [i for i, c in enumerate(S) if c.isalpha()]\n num = [ele for ele in S if ele.isnumeric()]\n\n diff_perm = 2**len(pos)\n\n letters_uplow = [[ele.upper(), ele.lower()] for ele in letters]\n import itertools\n comb = [[*S] for i in range(diff_perm)]\n\n for i, p in enumerate(itertools.product(*letters_uplow)):\n for j, ele in enumerate(p):\n comb[i][pos[j]] = ele\n return [''.join(ele) for ele in comb]\n \ns = Solution()\na = 'f12d'\nprint(s.letterCasePermutation(a))\n\n''' \nthis is very easy, have no problem with it\nbasically take all the alpha values out and turn that into product fo iteration\n'''","sub_path":"subsets/784. Letter Case Permutation.py","file_name":"784. Letter Case Permutation.py","file_ext":"py","file_size_in_byte":842,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"259794372","text":"import mock\nimport unittest\n\nfrom gameengine.controller import GameEngine\nfrom gameengine.commands import RunAwayCommand\n\n\nclass GameEngineTestCase(unittest.TestCase):\n def setUp(self):\n self.game = mock.Mock(is_finished=False)\n self.view = mock.Mock(**{\"read.return_value\": None})\n self.engine = GameEngine(self.game, self.view)\n\n def test_get_command_with_valid_command(self):\n self.view.read.return_value = \"OPEN SESAME\"\n with mock.patch(\"gameengine.controller.commands\") as mock_commands:\n command = self.engine.get_command()\n assert command == mock_commands.get.return_value\n self.view.read.assert_called_once_with(prompt=\"> \")\n mock_commands.get.assert_called_once_with(\"open sesame\")\n\n def test_get_command_with_invalid_command(self):\n self.view.read.side_effect = [\"open sesame\", None]\n with mock.patch(\"gameengine.controller.commands\") as mock_commands:\n mock_commands.get.return_value = None\n command = self.engine.get_command()\n self.view.write.assert_called_with(\"I'm sorry, I don't understand.\")\n\n def test_get_command_returns_run_away_command_on_eof(self):\n command = self.engine.get_command()\n assert isinstance(command, RunAwayCommand)\n","sub_path":"test/unit/test_controller.py","file_name":"test_controller.py","file_ext":"py","file_size_in_byte":1304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"313107403","text":"import pandas as pd\nimport matplotlib.pylab as plt\nfrom statsmodels.tsa.seasonal import seasonal_decompose\nimport argparse\nimport matplotlib.pyplot as plt\nfrom statsmodels.tsa.ar_model import AR\nfrom datetime import datetime \nplt.style.use('ggplot')\n\nmonths = {'January':'1', 'February':'2', 'March':'3', 'April':'4', 'May':'5', 'June':'6',\\\n 'July':'7','August':'8', 'September':'9', 'October':'10', 'November':'11', 'December':'12'}\n\ndef format_data (date_str):\n parts = date_str.split(\"-\")\n date = datetime(int(parts[2]), int(months[parts[1]]), int(parts[0]), 0, 0, 0)\n return date \n\n\ndef plot_data (X):\n\n fig = plt.figure(figsize=(12,6))\n ax = fig.add_subplot(111)\n ax.plot(X,'o')\n plt.show()\n \n\nif __name__ == \"__main__\":\n \n parser = argparse.ArgumentParser()\n parser.add_argument('-i','--input-file',help='Input File')\n window = 24\n\n \n args = parser.parse_args()\n\n df = pd.read_csv(args.input_file) \n \n print(df.shape, df.columns)\n\n df['Date'] = [format_data(d) for d in df['Date'].to_list()] \n df.index = df['Date'].to_list()\n df = df.sort_values(by='Date')\n\n #plot_data(df['Open Price'])\n\n X = df['Open Price']\n\n decomposed = seasonal_decompose(X,freq = 12)\n\n x = decomposed.plot() #See note below about this \n plt.rcParams['figure.figsize']=(35,15)\n plt.style.use('ggplot')\n plt.show()\n\n\n df['stationary']=df['Open Price'].diff()\n X = df['stationary'].dropna()\n train_data = X.iloc[1:len(X)-window]\n test_data = X.iloc[X[len(X)-window:]]\n\n #train the autoregression model\n model = AR(train_data)\n model_fitted = model.fit()\n\n print('The lag value chose is: %s' % model_fitted.k_ar)\n print('The coefficients of the model are:\\n %s' % model_fitted.params)\n \n # make predictions \n predictions = model_fitted.predict(start=len(train_data),\\\n end=len(train_data) + len(test_data)-1, dynamic=False)\n predictions.index = X.index[len(X)-window:]\n\n\n compare_df = pd.concat([df['stationary'].tail(window),predictions],\\\n axis=1).rename(columns={'stationary': 'actual', 0:'predicted'})\n\n #plot the two values\n plt.plot(compare_df['actual'],label='Actual')\n plt.plot(compare_df['predicted'],label='Predicted')\n plt.legend()\n plt.show()\n\n\n\n","sub_path":"predictive/predict_share_price.py","file_name":"predict_share_price.py","file_ext":"py","file_size_in_byte":2239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"128419806","text":"import numpy as np\nimport matplotlib.pyplot as plt\nfrom keras.models import Sequential, Model\nfrom keras.layers import Dense, LSTM, Input\nfrom keras.utils import np_utils\nfrom keras.callbacks import EarlyStopping, ModelCheckpoint\nfrom sklearn.preprocessing import RobustScaler\nfrom sklearn.metrics import mean_squared_error, r2_score\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.decomposition import PCA\nfrom sklearn.datasets import load_boston\nmodelpath = './model/{epoch:02d} - {val_loss:.4f}.hdf5'\n\nes = EarlyStopping(monitor = 'loss', mode = 'min', patience = 10)\ncp = ModelCheckpoint(filepath = modelpath, monitor = 'val_loss',\n mode = 'auto', save_best_only = True)\nscaler = RobustScaler()\npca = PCA(n_components = 10)\n\n\n# 1. 데이터\nx, y = load_boston(return_X_y = True)\n# print(x.shape)\n# print(y.shape)\n\n# 1-1. 데이터 분할\n\n\n# 1-2. Scaling\npca.fit(x)\nx = pca.transform(x)\n# x = scaler.fit_transform(x)\n\nx_train, x_test, y_train, y_test = train_test_split(\n x, y, test_size = 0.2, shuffle = True,\n random_state = 77)\n\nprint(x_train.shape)\nprint(x_test.shape)\nprint(y_train.shape)\nprint(y_test.shape)\n\n\n# 2. 모델링\nmodel = Sequential()\nmodel.add(Dense(32, input_shape = (10, ), activation = 'relu'))\nmodel.add(Dense(32, activation = 'relu'))\nmodel.add(Dense(64, activation = 'relu'))\nmodel.add(Dense(64, activation = 'relu'))\nmodel.add(Dense(128, activation = 'relu'))\nmodel.add(Dense(128, activation = 'relu'))\nmodel.add(Dense(1, activation = 'relu'))\n\nmodel.summary()\n\n\n# 3. 컴파일 및 훈련\nmodel.compile(loss = 'mse', metrics = ['mse'], optimizer = 'adam')\nmodel.fit(x_train, y_train, callbacks = [es, cp],\n epochs = 100, batch_size = 32, verbose = 1,\n validation_split = 0.25) # key_error : 'val_loss', cp에 val_loss를 선언했는데 validation셋을 안줬기 때문\n\n# print(hist.history.keys())\n\n\n# 4. 모델 평가\nres = model.evaluate(x_test, y_test)\nprint(res)\n\ny_pred = model.predict(x_test)\nprint(\"y_pred : \", y_pred)\n\n# 5. 성능 지표 평가\nprint(\"R2 : \", r2_score(y_test, y_pred))\n\n\n'''\nResult\nR2 Score = 0.7801983681184715\n'''","sub_path":"keras/keras73_boston_dnn.py","file_name":"keras73_boston_dnn.py","file_ext":"py","file_size_in_byte":2147,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"414454679","text":"#This is a function that objectifies searching for things\nimport sqlite3\nfrom tkinter import *\nconn = sqlite3.connect('names.db')\n#aconn.row_factory = sqlite3.Row\ncursor = conn.cursor()\n\ndef throwErrorBox(title, message):\n window = tk()\n window.title(title)\n l1 = Label(window, message)\n l1.pack()\n b1 = Button(window, text=\"Okay\", command=window.destroy)\n #Start the loop\n window.mainloop()\n\n\ndef searchFor(searchParam):\n results = cursor.execute(\"SELECT * from names WHERE LIKE ?\", (searchParam,))\n for row in results:\n assert row[3] == row['firstname']\n\n conn.commit()\n conn.close()\n","sub_path":"searchDB.py","file_name":"searchDB.py","file_ext":"py","file_size_in_byte":626,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"159238998","text":"s = '\\n' + ' Еще ''\\n' + ' q werty ' + '\\n' + '\\n' +' 123 45 7 8 9 0 ' + '\\n' + '\\n'\nprint(s)\n\ns = s.replace('\\n', ' ')\nwhile s.find(' ') != -1:\n s = s.replace(' ', ' ')\ns=s.split()\nprint(s)\nif s[0] == 'Еще':\n s.pop(0)\nss=''\ns = ' '.join(s)\nprint(s)\nprint(ss)\nprint(type(s))\n\n\n\n#\n# n = 3\n# m = 4\n# a = []\n# b = []\n# a.append(5)\n# a.append(6)\n# b.append(a)\n# a = []\n# a.append(45)\n# b.append(a)\n# print(a)\n# print(b)","sub_path":"basic/123.py","file_name":"123.py","file_ext":"py","file_size_in_byte":451,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"133609854","text":"#!/usr/bin/env python\r\n# -*- encoding:utf-8 -*-\r\n# 参考URL:https://dev.twitter.com/rest/public/search\r\n\r\nfrom requests_oauthlib import OAuth1Session\r\nimport json\r\n\r\nCONSUMER_KEY = \"****\"\r\nCONSUMER_SECRET = \"****\"\r\nACCESS_TOKEN = '****'\r\nACCESS_TOKEN_SECRET = '****'\r\n\r\nqk = \"****\"\r\n\r\n# ファイルに書き出し\r\nf = open(qk + '.txt', 'a')\r\n\r\nurl = \"https://api.twitter.com/1.1/search/tweets.json\"\r\n\r\n# 指定パラメータ\r\nparams = {\"q\": qk,\r\n \"count\":\"100\",\r\n \"result_type\":\"recent\"}\r\n\r\n# OAuth で GET\r\ntwitter = OAuth1Session(CONSUMER_KEY, CONSUMER_SECRET, ACCESS_TOKEN, ACCESS_TOKEN_SECRET)\r\nreq = twitter.get(url, params = params)\r\n\r\nif req.status_code == 200:\r\n # レスポンスはJSON形式なので parse する\r\n timeline = json.loads(req.text)\r\n\r\n i = 1\r\n # 各ツイートの本文を表示\r\n for tweet in timeline[\"statuses\"]:\r\n try:\r\n print(tweet['text'])\r\n print(\"------------------------\")\r\n \r\n f.write(tweet['text'] + \"\\n\")\r\n f.write(\"------------------------\\n\")\r\n i = i + 1\r\n except: \r\n print(\"Error\")\r\n print(\"------------------------\")\r\n\r\nelse:\r\n # エラーの場合\r\n print (\"Error: %d\" % req.status_code)\r\n\r\nf.close()","sub_path":"search-twitter_oauth.py","file_name":"search-twitter_oauth.py","file_ext":"py","file_size_in_byte":1302,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"71081289","text":"#!/usr/bin/env python\n\nPROJECT = 'virtualenvwrapper-initproject'\nVERSION = '0.0.1'\n\n# Bootstrap installation of Distribute\nfrom setuptools import setup, find_packages\n\nsetup(\n name=PROJECT,\n version=VERSION,\n author='Yannick Formaggio',\n author_email='yannick@thelumberjhack.org',\n url='',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'License :: OSI Approved :: WTPFL v2',\n 'Programming Language :: Python',\n 'Intended Audience :: Developers',\n 'Environment :: Console',\n ],\n platforms=['Any'],\n provides=['virtualenvwrapper.init_project'],\n requires=['virtualenv', 'virtualenvwrapper (>=2.0)'],\n description=(\n 'It creates a python project architecture,'\n 'setup.py, and initialize git repository while'\n 'creating new project using mkproject'\n ),\n long_description=open('README.md').read(),\n namespace_packages=['virtualenvwrapper'],\n packages=find_packages(),\n include_package_data=True,\n entry_points={\n 'virtualenvwrapper.project.post_mkproject': [\n 'user_scripts = virtualenvwrapper.init_project:post_mkproject',\n ]\n },\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1170,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"336953085","text":"from tkinter import *\n\ndef left_click(event):\n frame1['bg'] = 'red'\n frame2['bg'] = 'white'\n frame3['bg'] = 'white'\n\n# еще один способ менять параметры виджета через метод configure\ndef middle_click(event):\n frame1.configure(bg='white')\n frame2.configure(bg='yellow')\n frame3.configure(bg='white')\n\ndef right_click(event):\n frame1['bg'] = 'white'\n frame2['bg'] = 'white'\n frame3['bg'] = 'blue'\n\nroot = Tk()\nroot.title('События мыши')\n\nroot.configure(bg='black')# задает параметры основного экрана\n\nframe1 = Frame(root, width=250, heigh=250, bg='white')\nframe2 = Frame(root, width=250, heigh=250, bg='white')\nframe3 = Frame(root, width=250, heigh=250, bg='white')\n\nframe1.grid(row=0, column=0)\nframe2.grid(row=0, column=1, padx=1, pady=2)# padx задает фрейму бардюры по оси x\nframe3.grid(row=0, column=2)\n\n# вызывает метод для главного окна, чтобы отработало нажатие мыши в любом месте виджета\nroot.bind('', left_click)\nroot.bind('', middle_click)\nroot.bind('', right_click)\nroot.mainloop()\n","sub_path":"tkinter_lesson_06_test.pyw","file_name":"tkinter_lesson_06_test.pyw","file_ext":"pyw","file_size_in_byte":1229,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"295254358","text":"'''\n\nGood morning! Here's your coding interview problem for today.\n\nThis problem was asked by Uber.\n\nGiven an array of integers, return a new array such that each element at index i of the new array is the product of all the numbers in the original array except the one at i.\n\nFor example, if our input was [1, 2, 3, 4, 5], the expected output would be [120, 60, 40, 30, 24]. If our input was [3, 2, 1], the expected output would be [2, 3, 6].\n\nFollow-up: what if you can't use division?\n'''\n\ntest_1 = [1, 2, 3, 4, 5]\ntest_2 = [3, 2, 1]\n\n\n# Problem solved with nested for loops.\ndef multiply_list(number_list):\n output_list = []\n for i in range(0, len(number_list)):\n curr = 1\n tmp_list = number_list.copy()\n tmp_list.pop(i)\n for j in range(0, len(tmp_list)):\n curr = curr * tmp_list[j]\n output_list.append(curr)\n return output_list\n\n\n# Attempting to solve this problem using recursion and without division.\ndef multiply_list_recurse(current_index, input_list=[], output_list=[]):\n if not current_index >= len(input_list):\n output_list.insert(current_index, 1)\n for i in range(0, len(input_list)):\n if not current_index == i:\n output_list[current_index] = output_list[current_index] * input_list[i]\n current_index += 1\n multiply_list_recurse(current_index, input_list, output_list)\n return output_list\n\n\n# Same as above, but without needing an index val passed.\ndef multiply_list_recurse_v2(input_list=[], output_list=[]):\n if not len(output_list) >= len(input_list):\n output_list.insert(len(output_list), 1)\n for i in range(0, len(input_list)):\n if not (len(output_list) - 1) == i:\n output_list[(len(output_list) - 1)] = output_list[(len(output_list) - 1)] * input_list[i]\n multiply_list_recurse_v2(input_list, output_list)\n return output_list\n\n\nprint(multiply_list(test_1))\nprint(multiply_list(test_2))\n\nprint(multiply_list_recurse(0, test_1, []))\nprint(multiply_list_recurse(0, test_2, []))\nprint(multiply_list_recurse(0, [], []))\n\nprint(multiply_list_recurse_v2(test_1, []))\nprint(multiply_list_recurse_v2(test_2, []))\nprint(multiply_list_recurse_v2([], []))","sub_path":"11-26.py","file_name":"11-26.py","file_ext":"py","file_size_in_byte":2230,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"217015057","text":"#!/usr/bin/env python\n\nimport roslib; roslib.load_manifest('smacha')\nimport rospy\nimport smach\nimport smach_ros\n\nfrom smacha_ros.srv import GripperSrv\nfrom smacha_ros.srv import GripperSrvRequest\nfrom smacha_ros.srv import GripperSrvResponse\nfrom geometry_msgs.msg import *\n\ndef gripper_srv(req):\n if req.max_effort > 5.0:\n print('gripper_srv() returning True')\n return GripperSrvResponse(True)\n else:\n print('gripper_srv() returning False')\n return GripperSrvResponse(False)\n\ndef main():\n rospy.init_node('smach_example_actionlib_service_state')\n \n # Register a gripper service\n s = rospy.Service('gripper_srv', GripperSrv, gripper_srv)\n\n # Create a SMACH state machine\n sm0 = smach.StateMachine(outcomes=['succeeded','aborted','preempted'])\n \n # Set userdata\n sm0.userdata.max_effort = 9.0\n sm0.userdata.position = Point()\n sm0.userdata.gripper_input = 9.0\n\n # Open the container\n with sm0:\n # Add states to the container\n\n # Empty request message\n smach.StateMachine.add('TRIGGER_GRIPPER_EMPTY_REQUEST',\n smach_ros.ServiceState('gripper_srv', GripperSrv),\n transitions={'succeeded':'TRIGGER_GRIPPER_FIXED_REQUEST'})\n\n # Fixed request message\n smach.StateMachine.add('TRIGGER_GRIPPER_FIXED_REQUEST',\n smach_ros.ServiceState('gripper_srv', GripperSrv,\n request = GripperSrvRequest(4.0, Point())),\n transitions={'succeeded':'TRIGGER_GRIPPER_USER_DATA_REQUEST'})\n\n # Request from user data\n smach.StateMachine.add('TRIGGER_GRIPPER_USER_DATA_REQUEST',\n smach_ros.ServiceState('gripper_srv', GripperSrv,\n request_slots = ['max_effort',\n 'position']),\n transitions={'succeeded':'TRIGGER_GRIPPER_REQUEST_CALLBACK'})\n\n # Request callback\n @smach.cb_interface(input_keys=['gripper_input'])\n def gripper_request_cb(userdata, request):\n gripper_request = GripperSrvRequest()\n gripper_request.position.x = 2.0\n gripper_request.max_effort = userdata.gripper_input\n return gripper_request\n\n smach.StateMachine.add('TRIGGER_GRIPPER_REQUEST_CALLBACK',\n smach_ros.ServiceState('gripper_srv', GripperSrv,\n request_cb = gripper_request_cb,\n input_keys = ['gripper_input']),\n transitions={'succeeded':'succeeded'})\n\n # Execute SMACH plan\n outcome = sm0.execute()\n\n rospy.signal_shutdown('All done.')\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"smacha_ros/test/executive_smach_tutorials/smach_tutorials/examples/service_test.py","file_name":"service_test.py","file_ext":"py","file_size_in_byte":2934,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"284268598","text":"\"\"\"\nThis file serves as a evaluation interface for the network\n\"\"\"\n# Built in\nimport os\n# Torch\n\n# Own\nimport flag_reader\nfrom class_wrapper import Network\nfrom model_maker import Tandem\nimport data_reader\n\n# Libs\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef compare_truth_pred(pred_file, truth_file):\n \"\"\"\n Read truth and pred from csv files, compute their mean-absolute-error and the mean-squared-error\n :param pred_file: full path to pred file\n :param truth_file: full path to truth file\n :return: mae and mse\n \"\"\"\n pred = np.loadtxt(pred_file, delimiter=' ')\n truth = np.loadtxt(truth_file, delimiter=' ')\n\n mae = np.mean(np.abs(pred-truth), axis=1)\n mse = np.mean(np.square(pred-truth), axis=1)\n\n return mae, mse\n\n\ndef plotMSELossDistrib(pred_file, truth_file, flags):\n mae, mse = compare_truth_pred(pred_file, truth_file)\n plt.figure(figsize=(12, 6))\n plt.hist(mse, bins=100)\n plt.xlabel('Mean Squared Error')\n plt.ylabel('cnt')\n plt.suptitle('Tandem (Avg MSE={:.4e})'.format(np.mean(mse)))\n plt.savefig(os.path.join(os.path.abspath(''), 'data',\n 'Tandem_{}.png'.format(flags.eval_model)))\n plt.show()\n print('Tandem (Avg MSE={:.4e})'.format(np.mean(mse)))\n\ndef evaluate_from_model(model_dir):\n \"\"\"\n Evaluating interface. 1. Retreive the flags 2. get data 3. initialize network 4. eval\n :param model_dir: The folder to retrieve the model\n :return: None\n \"\"\"\n # Retrieve the flag object\n print(\"Retrieving flag object for parameters\")\n flags = flag_reader.load_flags(os.path.join(\"models\", model_dir))\n flags.eval_model = model_dir # Reset the eval mode\n flags.batch_size = 1 # For backprop eval mode, batchsize is always 1\n\n # Get the data\n train_loader, test_loader = data_reader.read_data(x_range=flags.x_range,\n y_range=flags.y_range,\n geoboundary=flags.geoboundary,\n batch_size=flags.batch_size,\n normalize_input=flags.normalize_input,\n data_dir=flags.data_dir)\n print(\"Making network now\")\n\n # Make Network\n ntwk = Network(Tandem, flags, train_loader, test_loader, inference_mode=True, saved_model=flags.eval_model)\n\n # Evaluation process\n print(\"Start eval now:\")\n pred_file, truth_file = ntwk.evaluate()\n\n # Plot the MSE distribution\n plotMSELossDistrib(pred_file, truth_file, flags)\n print(\"Evaluation finished\")\n\n\n\nif __name__ == '__main__':\n # Read the flag, however only the flags.eval_model is used and others are not used\n useless_flags = flag_reader.read_flag()\n\n print(useless_flags.eval_model)\n # Call the evaluate function from model\n evaluate_from_model(useless_flags.eval_model)\n\n","sub_path":"cGAN/evaluate.py","file_name":"evaluate.py","file_ext":"py","file_size_in_byte":2994,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"340362088","text":"from gpiozero import Button, Buzzer\nfrom twython import Twython\nfrom time import sleep\nimport glob\nimport os\nfrom auth import (\n consumer_key,\n consumer_secret,\n access_token,\n access_token_secret\n)\nimport random\noneutton = Button(2)\ntwoutton = Button(3)\nthreeutton = Button(4)\nfourutton = Button(17)\nfiveutton = Button(27)\nsixutton = Button(26)\nbuzzer = Buzzer(22)\none = 0\ntwo = 0\nthree = 0\nfour = 0\nfive = 0\nsix = 0\nprint(\"Initiating\")\nbuzzer.off()\nsleep(5)\nfor i in range(10):\n print(\"Ready?\")\n sleep(random.randint(1,5))\n buzzer.on()\n while True:\n if(oneutton.is_pressed):\n print(\"Kian wins\")\n buzzer.off()\n one += 1\n break\n if(twoutton.is_pressed):\n print(\"Player 2 wins\")\n buzzer.off()\n two += 1\n break\n if(threeutton.is_pressed):\n print(\"Player 3 wins\")\n buzzer.off()\n three += 1\n break\n if(fourutton.is_pressed):\n print(\"Player 4 wins\")\n buzzer.off()\n four += 1\n break\n if(fiveutton.is_pressed):\n print(\"Player 5 wins\")\n buzzer.off()\n five +=1\n break\n if(sixutton.is_pressed):\n print(\"Player 6 wins\")\n buzzer.off()\n six +=1\n break\nwinner = max(one, two, three, four, five, six)\nwhile True:\n if(winner == one):\n print(\"Kian wins game\")\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Victory for Kian!\"\n twitter.update_status(status=message)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n if(winner == two):\n print(\"Player 2 wins game\")\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Player 2 Victory!\"\n twitter.update_status(status=message)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n if(winner == three):\n print(\"Player 3 wins game\")\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Player 3 Victory!\"\n twitter.update_status(status=message)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n if(winner == four):\n print(\"Player 4 wins game\")\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Player 4 Victory!\"\n twitter.update_status(status=message)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n if(winner == five):\n print(\"Player 5 wins game\")\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Player 5 Victory!\"\n twitter.update_status(status=message)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n if(winner == six):\n print(\"Molly wins game\")\n imageFilenames = glob.glob('/home/pi/Desktop/*')\n newestImageFilename = max(imageFilenames, key=os.path.getctime)\n twitter = Twython(consumer_key, consumer_secret, access_token, access_token_secret)\n message=\"Victory for MA!\"\n with open(newestImageFilename, 'rb') as photo:\n twitter.update_status_with_media(status=message,media=photo)\n print(\"Victory Tweeted. Check at @2Stanford2019\")\n break\n else:\n pass\n \n","sub_path":"Button Game 6.py","file_name":"Button Game 6.py","file_ext":"py","file_size_in_byte":3631,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"239211559","text":"# Реализуйте базовый класс Car. У данного класса должны быть следующие атрибуты:\n# speed, color, name, is_police (булево). А также методы: go, stop, turn(direction), которые должны\n# сообщать, что машина поехала, остановилась, повернула (куда). Опишите несколько дочерних классов:\n# TownCar, SportCar, WorkCar, PoliceCar. Добавьте в базовый класс метод show_speed, который должен\n# показывать текущую скорость автомобиля. Для классов TownCar и WorkCar переопределите метод show_speed.\n# При значении скорости свыше 60 (TownCar) и 40 (WorkCar) должно выводиться сообщение о превышении\n# скорости.Создайте экземпляры классов, передайте значения атрибутов. Выполните доступ к атрибутам,\n# выведите результат. Выполните вызов методов и также покажите результат.\n\nfrom random import randint\n\n\nclass Car:\n def __init__(self, speed, color, name, is_police):\n self.color = color\n self.name = name\n self.speed = speed\n self.is_police = is_police\n\n def go(self):\n if self.is_police:\n return f\"Полицейская машина {self.name}, цвет {self.color}, начала движение.\"\n else:\n return f\"Машина {self.name}, цвет {self.color}, начала движение.\"\n\n def show_speed(self):\n return f\"Машина едет со скоростью {self.speed}км/ч.\"\n\n def stop(self):\n return f\"Машина {self.name} остановилась.\"\n\n def turn_direction(self):\n i = randint(1, 3)\n if i == 1:\n return f\"Машина повернула налево.\"\n elif i == 2:\n return f\"Машина поехала прямо.\"\n else:\n return f\"Машина повернула направо.\"\n\n\nclass SportCar(Car):\n def __init__(self, speed, color, name, is_police=False):\n super().__init__(speed, color, name, is_police)\n\n\nclass PoliceCar(Car):\n def __init__(self, speed, color, name, is_police=True):\n super().__init__(speed, color, name, is_police)\n\n\nclass TownCar(Car):\n def __init__(self, speed, color, name, is_police=False):\n super().__init__(speed, color, name, is_police)\n\n def show_speed(self):\n if self.speed > 60:\n return f\"Машина едет со скоростью {self.speed}км/ч.\" + \"\\033[1m\\033[31m {}\\033[0m\".format(\n \"Внимание! Вы превысили скорость!\")\n else:\n return f\"Машина едет со скоростью {self.speed}км/ч.\"\n\n\nclass WorkCar(Car):\n def __init__(self, speed, color, name, is_police=False):\n super().__init__(speed, color, name, is_police)\n\n def show_speed(self):\n if self.speed > 40:\n return f\"Машина едет со скоростью {self.speed}км/ч.\" + \"\\033[1m\\033[31m {}\\033[0m\".format(\n \"Внимание! Вы превысили скорость!\")\n else:\n return f\"Машина едет со скоростью {self.speed}км/ч.\"\n\n\ns = SportCar(120, \"красный\", \"'Астон Мартин'\")\nprint(s.go(), s.turn_direction(), '\\n' + s.show_speed(), s.stop(), '\\n')\np = PoliceCar(100, \"белый\", \"'Шевроле Круз'\")\nprint(p.go(), p.turn_direction(), '\\n' + p.show_speed(), p.stop(), '\\n')\nt = TownCar(70, \"чёрный\", \"'Дацун Мидо'\")\nprint(t.go(), t.turn_direction(), '\\n' + t.show_speed(), t.stop(), '\\n')\nw = WorkCar(50, \"оранжевый\", \"- цементовоз\")\nprint(w.go(), w.turn_direction(), '\\n' + w.show_speed(), w.stop(), '\\n')\n","sub_path":"lesson06/task04.py","file_name":"task04.py","file_ext":"py","file_size_in_byte":4082,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"484449455","text":"import tkinter as tk\nfrom tkinter import *\nfrom datetime import datetime\n\nscreen = tk.Tk(screenName=\"GadgetifyWithGSBlr Shopping\")\nscreen.geometry(\"1600x640\")\nscreen.title(\"GadgetifyWithGSBlr Shopping\")\nscreen.configure(bg=\"grey\")\n\nproduct_list = {\"JBL Basshead earphones: Rs. 1200\" : 1200, \"Bluetooth computer mouse: Rs. 600\": 600, \"HP Mouse: Rs. 200\":200, \n \"Realme 5 pro: Rs. 15000\": 15000, \"Sandisk Pendrive: Rs. 850\": 850}\n\n\n#getting the item purchased\n\ndef get_key(val):\n for key, value in product_list.items():\n if val == value:\n return key\n\n#creating the function \"gen_bill\" that generates the bill and creates a textfile of the same\n\ndef gen_bill():\n timestamp= datetime.now().strftime(\" %d %B %Y, %H:%M:%S \") + \"\\n\"\n shop_name= \" Shop name: GadgetifyWithGSBlr \" + \"\\n\"\n shop_address=\" shop address: 311/5 Akshay nagar, Bangalore, Karnataka, India \" + \"\\n\"\n shop_contact= \" Shop contact no: +91 9988776655 \" + \"\\n\\n\\n\"\n \n prod_pur=prod_select.get()\n name_bill=\"Name of the customer :- \" + name_entry.get() + \"\\n\"\n contact=\"Mobile of the customer :- \" + phone_entry.get() + \"\\n\"\n item_purchased = \"Item Purchased by the customer :- \" + get_key(prod_pur).split(\":\")[0] + \"\\n\"\n item_price = \"Price of the Item purchased:- Rs. \" + str(prod_pur) + \"\\n\"\n quan=\"Quantity of the item purchased:- \" + quantity_entry.get() + \"\\n\"\n pay_mode = \"Mode of Payment:- \" + mop_select.get() + \"\\n\"\n\n\n if(name_entry.get() == \"\" or phone_entry.get() == \"\" or quantity_entry.get() == \"\"):\n l2=Label(screen, text=\"Bill couldn't be generated due to one or more fields missing\", height = 2, width= 100, bg=\"#EA2511\" , fg= \"#050506\")\n l2.grid(row = 8, column= 0, columnspan=10)\n l2.after(3000, lambda: l2.destroy())\n\n else:\n #calculation of tax\n taxable_price =prod_pur* int(quantity_entry.get())\n tax = 0.06* taxable_price\n\n tax_charged= \"Tax on the Item purchased:- Rs. \" + str(tax) + \"\\n\"\n total_price= \"Final Price of the product:- Rs. \" + str(taxable_price + tax) + \"\\n\"\n\n file=open(str(name_entry.get()).split(\" \")[0], \"w\")\n file.write(timestamp)\n file.write(shop_name)\n file.write(shop_address)\n file.write(shop_contact)\n file.write(name_bill)\n file.write(contact)\n file.write(item_purchased)\n file.write(item_price)\n file.write(quan)\n file.write(tax_charged)\n file.write(total_price)\n file.write(pay_mode)\n file.close()\n l1=Label(screen, text=\"Bill generated Successfully....!!!\", height = 2, width= 100, bg=\"#2ECA12\" , fg= \"#DEF031\")\n l1.grid(row = 8, column= 0, columnspan=10)\n l1.after(3000, lambda: l1.destroy())\n \n #clearing fields\n name_entry.delete(0,END)\n phone_entry.delete(0,END)\n quantity_entry.delete(0,END)\n r1_mop.select()\n prod_select.set(1200)\n \n \n\n#create banner\nbanner = Label(screen, text= \"GadgetifyWithGSBlr Shopping Mall\", height = 3, width= 100, bg=\"#1282B6\", fg= \"#FFFFFF\")\nbanner.grid(row=0, column = 0, columnspan=10)\nbanner.config(font=(\"Courier\", 20))\n\n#create fields labels\n\nname = Label(screen, text= \"Name of the Customer\", height = 2, width = 30, bg=\"#1282B6\", fg= \"#FFFFFF\")\nphone = Label(screen, text= \"Mobile No.\", height = 2, width = 30, bg=\"#1282B6\", fg= \"#FFFFFF\")\nmode_of_pay = Label(screen, text= \"Payment Method\", height = 2, width = 30, bg=\"#1282B6\", fg= \"#FFFFFF\")\nitem = Label(screen, text= \"Item You want to Buy\", height = 2, width = 30, bg=\"#1282B6\", fg= \"#FFFFFF\")\nquantity = Label(screen, text= \"Quantity\", height = 2, width = 30, bg=\"#1282B6\", fg= \"#FFFFFF\")\n\n#placing labels on screen\n\nname.grid(row = 1, column= 0, pady = 10)\nphone.grid(row = 2, column= 0, pady = 10)\nmode_of_pay.grid(row = 3, column= 0, pady = 10)\nitem.grid(row = 4, column= 0, pady = 10)\nquantity.grid(row = 5, column= 0, pady = 10)\n\n#creating entry box, radio button\n\nname_entry = Entry(width= 40)\nphone_entry = Entry(width= 40)\nquantity_entry = Entry(width= 40)\n\nmop_select = tk.StringVar()\nr1_mop = Radiobutton(screen, text=\"Online Banking\", value=\"Online Banking\", width= 30, variable = mop_select)\nr2_mop = Radiobutton(screen, text=\"Cash\", value=\"Cash\", width= 30, variable = mop_select)\nr3_mop = Radiobutton(screen, text=\"Card\", value=\"Card\", width= 30, variable = mop_select)\nr1_mop.select()\n\n#placing entry box, radio button on screen\n\nname_entry.grid(row = 1, column = 1, pady=10, ipady=6)\nphone_entry.grid(row = 2, column = 1, pady=10, ipady=6)\nquantity_entry.grid(row = 5, column = 1, pady=10, ipady=6)\n\nr1_mop.grid(row = 3, column = 1, ipady=6)\nr2_mop.grid(row = 3, column = 2, ipady=6)\nr3_mop.grid(row = 3, column = 3, ipady=6)\n\n#creating radio buttons for products list\n\nprod_select= tk.IntVar(screen,1200)\ncount=1\nfor (k,v) in product_list.items():\n Radiobutton(screen, text = k, variable = prod_select, value = v, width =28).grid(row = 4, column = count, ipady= 6)\n count+=1\n\n#creating BuyNow Button and placing that to screen\n\nbuy_now = Button(screen, text=\"Buy Now\", command = gen_bill, width= 40, bg=\"#1282B6\", fg= \"#FFFFFF\")\nbuy_now.grid(row =7, column = 2, ipady=6, pady=10)\n\nscreen.mainloop()","sub_path":"Easy/1. Shopping Cart Problem/solutions/bagofcodes_AnshumaliShaw/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":5566,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"357823008","text":"import sys\n\n\ndef solution(data):\n masks = [0x0, 0x80, 0xE0, 0xF0, 0xF8]\n bits = [0x0, 0x0, 0xC0, 0xE0, 0xF0]\n while data:\n for x in range(4, -1, -1):\n if data[0] & masks[x] == bits[x]:\n break\n if x == 0 or len(data) < x:\n return 0\n for y in range(1, x):\n if data[y] & 0xC0 != 0x80:\n return 0\n data = data[x:]\n return 1\n\n\nif __name__ == \"__main__\":\n n = int(sys.stdin.readline())\n s = list(map(int, sys.stdin.readline().split(\" \")))\n # line = [bin(i)[2:] for i in s]\n # print(line)\n res = solution(s)\n # print(1 if res == 1 else 0)\n print(res)\n","sub_path":"头条/problem4.py","file_name":"problem4.py","file_ext":"py","file_size_in_byte":667,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"296292366","text":"num=int(input('Insira o número que você deseja descobrir se é primo: '))\r\n\r\nprimos=[]\r\n\r\nfor x in range(1, 101):\r\n cont=0\r\n\r\n for y in range(1, x+1):\r\n if x%y==0:\r\n cont+=1\r\n if cont<=2:\r\n primos.append(x)\r\n\r\nif num in primos:\r\n print('{} é um número primo'.format(num))\r\nelse:\r\n print('{} não é um número primo'.format(num))\r\n \r\nind=int(input('Se você deseja visualizar uma lista com números primos digite \"1\": '))\r\n\r\nif ind == 1:\r\n print(primos)\r\n print('OK, até mais!!')\r\nelse:\r\n print('OK, até mais!!')","sub_path":"Python/Exercises/ex52.py","file_name":"ex52.py","file_ext":"py","file_size_in_byte":570,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"419827158","text":"\"\"\"Relu\n\"\"\"\n\nimport tensorflow as tf\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nX = tf.placeholder(tf.float32)\n\nwith tf.name_scope('Relu'):\n fx = tf.nn.relu(X)\n tf.summary.scalar(\"f(x)\", tf.squeeze(fx))\n\ninit = tf.global_variables_initializer()\n\nwith tf.Session() as sess:\n\n # Output graph\n merged = tf.summary.merge_all()\n writer = tf.summary.FileWriter(\"log/Relu/\", graph = sess.graph)\n \n # Run the initializer\n sess.run(init)\n \n for step in range(-10,11):\n a = tf.convert_to_tensor(step, dtype=tf.float32)\n a_r = sess.run([a])\n print(sess.run(a), sess.run(fx, feed_dict={X: a_r}))\n\n sess.run(fx, feed_dict={X: a_r})\n summary = sess.run(merged, feed_dict={X: sess.run([a])})\n writer.add_summary(summary, step)\n\n# with tf.Session() as sess:\n \n# for step in range(-10,11):\n# X = tf.convert_to_tensor(step, dtype=tf.float32)\n# # X = tf.random_uniform([1,1], minval=1.0, maxval=3.0, seed=step) //Or use random number\n# print(sess.run(X), sess.run(tf.nn.relu(X)))","sub_path":"src/TensorFlow/venv/Lab/Tutorials/Activation/Relu.py","file_name":"Relu.py","file_ext":"py","file_size_in_byte":1069,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"237450309","text":"# coding=utf-8\r\nimport torch\r\nfrom torch.autograd import Variable\r\nfrom bert_trigger.config import Config\r\nfrom bert_trigger import BertLstmCrf\r\nimport torch.optim as optim\r\nfrom .tri_utils import load_vocab, read_corpus_tr_id, load_model, save_model\r\nfrom torch.utils.data import TensorDataset\r\nfrom torch.utils.data import DataLoader\r\nimport tqdm\r\nimport argparse\r\n\r\n\r\ndef train(config=None):\r\n \"\"\"Train Model\"\"\"\r\n # load config\r\n if not config:\r\n config = Config()\r\n print('settings:\\n', config)\r\n # load corpus\r\n print('loading corpus.')\r\n vocab = load_vocab(config.vocab)\r\n label_dic = load_vocab(config.tri_id_label_file)\r\n tagset_size = len(label_dic)\r\n # load train and dev dataset\r\n train_data = read_corpus_tr_id(config.tri_id_train_file, max_length=config.max_length, label_dic=label_dic, vocab=vocab)\r\n train_ids = torch.LongTensor([temp[0] for temp in train_data])\r\n train_masks = torch.LongTensor([temp[1] for temp in train_data])\r\n train_tags = torch.LongTensor([temp[2] for temp in train_data])\r\n train_dataset = TensorDataset(train_ids, train_masks, train_tags)\r\n train_loader = DataLoader(train_dataset, shuffle=True, batch_size=config.batch_size)\r\n\r\n dev_data = read_corpus_tr_id(config.tri_id_dev_file, max_length=config.max_length, label_dic=label_dic, vocab=vocab)\r\n dev_ids = torch.LongTensor([temp[0] for temp in dev_data])\r\n dev_masks = torch.LongTensor([temp[1] for temp in dev_data])\r\n dev_tags = torch.LongTensor([temp[2] for temp in dev_data])\r\n dev_dataset = TensorDataset(dev_ids, dev_masks, dev_tags)\r\n dev_loader = DataLoader(dev_dataset, shuffle=True, batch_size=config.batch_size)\r\n # init model\r\n model = BertLstmCrf(config.bert_path, tagset_size, config.bert_embedding, config.rnn_hidden, config.rnn_layer, dropout_ratio=config.dropout_ratio, dropout1=config.dropout1, use_cuda=config.use_cuda)\r\n if config.load_model:\r\n assert config.load_path is not None\r\n model = load_model(model, name=config.load_path)\r\n if config.use_cuda:\r\n model.cuda()\r\n # train model\r\n print('begin training.')\r\n model.train()\r\n optimizer = getattr(optim, config.optim)\r\n optimizer = optimizer(model.parameters(), lr=config.lr, weight_decay=config.weight_decay)\r\n eval_loss = 10000\r\n for epoch in tqdm.tqdm(range(config.base_epoch)):\r\n for i, batch in tqdm.tqdm(enumerate(train_loader)):\r\n model.zero_grad()\r\n inputs, masks, tags = batch\r\n inputs, masks, tags = Variable(inputs), Variable(masks), Variable(tags)\r\n masks = masks.bool()\r\n if config.use_cuda:\r\n inputs, masks, tags = inputs.cuda(), masks.cuda(), tags.cuda()\r\n feats = model(inputs, masks)\r\n loss = model.loss(feats, masks, tags)\r\n loss.backward()\r\n optimizer.step()\r\n # save best model\r\n dev_loss_temp = evaluate(model, dev_loader, epoch, config)\r\n if dev_loss_temp < eval_loss:\r\n print('dev loss: ', eval_loss, ' -> ', dev_loss_temp)\r\n eval_loss = dev_loss_temp\r\n save_model(model, epoch)\r\n return model\r\n\r\n\r\ndef test(config=None, model=None):\r\n \"\"\"Test Model in test file\"\"\"\r\n # load config\r\n if not config:\r\n config = Config()\r\n print('settings:\\n', config)\r\n # load corpus\r\n print('loading corpus')\r\n vocab = load_vocab(config.vocab)\r\n label_dic = load_vocab(config.tri_id_label_file)\r\n tagset_size = len(label_dic)\r\n # load test dataset\r\n test_data = read_corpus_tr_id(config.tri_id_test_file, max_length=config.max_length, label_dic=label_dic, vocab=vocab)\r\n test_ids = torch.LongTensor([temp[0] for temp in test_data])\r\n test_masks = torch.LongTensor([temp[1] for temp in test_data])\r\n test_tags = torch.LongTensor([temp[2] for temp in test_data])\r\n test_dataset = TensorDataset(test_ids, test_masks, test_tags)\r\n test_loader = DataLoader(test_dataset, shuffle=False, batch_size=config.batch_size)\r\n # load trained model\r\n if not model:\r\n model = BertLstmCrf(config.bert_path, tagset_size, config.bert_embedding, config.rnn_hidden, config.rnn_layer, dropout_ratio=config.dropout_ratio, dropout1=config.dropout1, use_cuda=config.use_cuda)\r\n model = load_model(model, name=config.load_path)\r\n if config.use_cuda:\r\n model.cuda()\r\n # evaluate model in test file\r\n print('begin predicting')\r\n evaluate(model, test_loader, 0, config, True)\r\n\r\n\r\ndef predict(config=None, model=None, sent=None):\r\n \"\"\"\r\n Input: raw sentences saved in config.input_file or sent\r\n Output: results of trigger identification saved in config.tri_id_result_file\r\n format: sentence ||| tag (BIO)\r\n \"\"\"\r\n # load config\r\n if not config:\r\n config = Config()\r\n\r\n vocab = load_vocab(config.vocab)\r\n label_dic = load_vocab(config.tri_id_label_file)\r\n tagset_size = len(label_dic)\r\n # load trained model\r\n if not model:\r\n model = BertLstmCrf(config.bert_path, tagset_size, config.bert_embedding, config.rnn_hidden, config.rnn_layer,\r\n dropout_ratio=config.dropout_ratio, dropout1=config.dropout1, use_cuda=config.use_cuda)\r\n model = load_model(model, name=config.load_path)\r\n if config.use_cuda:\r\n model.cuda()\r\n # begin predicting\r\n if (not config.input_file) and sent:\r\n # preprocess sent\r\n sent = sent.lower()\r\n tokens = sent.split()\r\n tokens = tokens[0:min(config.max_length-2, len(tokens))]\r\n tokens_f = ['[CLS]'] + tokens + ['[SEP]']\r\n input_ids = torch.LongTensor([[int(vocab[i]) if i in vocab else int(vocab['[UNK]']) for i in tokens_f]])\r\n input_masks = torch.LongTensor([[1] * len(input_ids[0])])\r\n if config.use_cuda and torch.cuda.is_available():\r\n input_ids, input_masks = input_ids.cuda(), input_masks.cuda()\r\n # predict tags\r\n with torch.no_grad():\r\n feats = model(input_ids, input_masks)\r\n path_score, best_path = model.crf(feats, input_masks)\r\n pred_label = best_path[0].cpu().numpy().tolist()\r\n pred_label = [list(label_dic.keys())[int(x)] for x in pred_label[1:-1]]\r\n return pred_label\r\n else:\r\n with open(config.input_file, 'r', encoding='utf-8') as f:\r\n sents = f.readlines()\r\n data = []\r\n for line in sents:\r\n line = line.lower()\r\n tokens = line.split()\r\n tokens = tokens[0:min(config.max_length - 2, len(tokens))]\r\n tokens_f = ['[CLS]'] + tokens + ['[SEP]']\r\n input_ids = [int(vocab[i]) if i in vocab else int(vocab['[UNK]']) for i in tokens_f]\r\n input_masks = [1] * len(input_ids)\r\n while len(input_ids) < config.max_length:\r\n input_ids.append(0)\r\n input_masks.append(0)\r\n data.append((input_ids, input_masks))\r\n ids = torch.LongTensor([temp[0] for temp in data])\r\n masks = torch.LongTensor([temp[1] for temp in data])\r\n dataset = TensorDataset(ids, masks)\r\n loader = DataLoader(dataset, shuffle=False, batch_size=config.batch_size)\r\n sents = []\r\n pred = []\r\n for i, batch in tqdm.tqdm(enumerate(loader)):\r\n inputs, masks = batch\r\n inputs, masks = Variable(inputs), Variable(masks)\r\n masks = masks.bool()\r\n\r\n # save sentences\r\n for idx in range(inputs.shape[0]):\r\n sents.append(inputs[idx][masks[idx]].cpu().numpy().tolist())\r\n\r\n # predict labels\r\n if config.use_cuda:\r\n inputs, masks = inputs.cuda(), masks.cuda()\r\n with torch.no_grad():\r\n feats = model(inputs, masks)\r\n path_score, best_path = model.crf(feats, masks.byte())\r\n\r\n # save labels\r\n for idx in range(inputs.shape[0]):\r\n pred.append(best_path[idx][masks[idx]].cpu().numpy().tolist())\r\n # save result\r\n save_results(sents, pred, config)\r\n return pred\r\n\r\n\r\ndef evaluate(model, data_loader, epoch, config, save_result=False):\r\n model.eval()\r\n eval_loss = 0\r\n true = [] # true labels\r\n pred = [] # predicted labels\r\n sents = [] # sentences\r\n length = 0\r\n for i, batch in tqdm.tqdm(enumerate(data_loader)):\r\n inputs, masks, tags = batch\r\n length += inputs.size(0)\r\n inputs, masks, tags = Variable(inputs), Variable(masks), Variable(tags)\r\n masks = masks.bool()\r\n\r\n # save sentences\r\n for idx in range(inputs.shape[0]):\r\n sents.append(inputs[idx][masks[idx]].cpu().numpy().tolist())\r\n\r\n if config.use_cuda:\r\n inputs, masks, tags = inputs.cuda(), masks.cuda(), tags.cuda()\r\n feats = model(inputs, masks)\r\n path_score, best_path = model.crf(feats, masks.byte())\r\n loss = model.loss(feats, masks, tags)\r\n eval_loss += loss.item()\r\n\r\n # save labels\r\n for idx in range(inputs.shape[0]):\r\n pred.append(best_path[idx][masks[idx]].cpu().numpy().tolist())\r\n true.append(tags[idx][masks[idx]].cpu().numpy().tolist())\r\n # evaluate model's performace in data loader\r\n label_dic = load_vocab(config.tri_id_label_file)\r\n label_dic = dict(zip(label_dic.values(), label_dic.keys()))\r\n gold_num, predict_num, correct_num = 0, 0, 0\r\n for idx, tag in enumerate(pred):\r\n pred_tag = [label_dic[int(x)] for x in pred[idx]]\r\n true_tag = [label_dic[int(x)] for x in true[idx]]\r\n new_gold_num, new_predict_num, new_correct_num = evaluate_sent(true_tag, pred_tag)\r\n gold_num += new_gold_num\r\n predict_num += new_predict_num\r\n correct_num += new_correct_num\r\n precision = correct_num / predict_num if predict_num else 0\r\n recall = correct_num / gold_num if gold_num else 0\r\n f_score = (2 * precision * recall) / (precision + recall) if precision + recall else 0\r\n print('Gold Num: ', gold_num, ' Pred Num: ', predict_num, ' Corr Num: ', correct_num)\r\n print('Precision: ', precision, ' Recall: ', recall, ' F-score: ', f_score)\r\n # save result\r\n if save_result:\r\n save_results(sents, pred, config)\r\n # return loss\r\n print('eval epoch: {}| loss: {}'.format(epoch, eval_loss/length))\r\n model.train()\r\n return eval_loss\r\n\r\n\r\ndef save_results(sents, pred, config, path=None):\r\n vocab = []\r\n label_dic = load_vocab(config.tri_id_label_file)\r\n label_dic = dict(zip(label_dic.values(), label_dic.keys()))\r\n with open(config.vocab, \"r\", encoding=\"utf-8\") as reader:\r\n while True:\r\n token = reader.readline()\r\n if not token:\r\n break\r\n token = token.strip()\r\n vocab.append(token)\r\n if not path:\r\n path = config.tri_id_result_file\r\n with open(path, 'w', encoding='utf-8') as f:\r\n readable_sents = []\r\n readable_preds = []\r\n for idx in range(len(sents)):\r\n readable_sents.append([vocab[w] for w in sents[idx][1:-1]])\r\n readable_preds.append([label_dic[int(x)] for x in pred[idx][1:-1]])\r\n for idx in range(len(readable_sents)):\r\n f.write(' '.join(readable_sents[idx]))\r\n f.write('|||')\r\n f.write(' '.join(readable_preds[idx]))\r\n f.write('\\n')\r\n\r\n\r\ndef evaluate_sent(true_tag, predict_tag):\r\n gold_num = 0\r\n predict_num = 0\r\n correct_num = 0\r\n sent_len = len(true_tag)\r\n start_flag = False\r\n equal_flag = True\r\n\r\n for i in range(sent_len):\r\n gold_num = gold_num + 1 if 'B' in true_tag[i] else gold_num\r\n predict_num = predict_num + 1 if 'B' in predict_tag[i] else predict_num\r\n if 'B' in true_tag[i]:\r\n start_flag = True\r\n if start_flag and true_tag[i] != predict_tag[i]:\r\n equal_flag = False\r\n if start_flag and ((i < sent_len - 1 and 'I' not in true_tag[i+1]) or i == sent_len - 1):\r\n start_flag = False\r\n if equal_flag and ((i < sent_len - 1 and 'I' not in predict_tag[i+1]) or i == sent_len - 1):\r\n correct_num += 1\r\n equal_flag = True\r\n return gold_num, predict_num, correct_num\r\n\r\n\r\nif __name__ == '__main__':\r\n parser = argparse.ArgumentParser()\r\n parser.add_argument('--do-train', type=bool, default=False,\r\n help='Whether to retrain the model.')\r\n parser.add_argument('--do-eval', type=bool, default=False,\r\n help='Whether to perform evaluation.')\r\n args = parser.parse_args()\r\n if args.do_train:\r\n train()\r\n if args.do_eval:\r\n test()\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"web/utils/tri_id.py","file_name":"tri_id.py","file_ext":"py","file_size_in_byte":12737,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"153785691","text":"from question1 import Array\nfrom timeit import Timer\nfrom memory_profiler import profile\n\n#The beginning of Qn2\nclass SubArray(Array):\n @profile\n def add(self, values):\n print(\"***QUESTION 2***\")\n values.append(34)\n print(\"WE HAVE ADDED 34 IN THE ARRAY: \" , values)\n\n @profile\n def delete(self, values):\n values.remove(9)\n print(\"9 WAS REMOVED FROM THE ARRAY. THE NEW ARRAY IS: \" , values)\n print(\"\")\n\n# Third Question\n #Check if a value is in an array\n @profile\n def check(self, values):\n print(\"***QUESTION 3***\")\n if 5 in values:\n contains = True\n print(contains)\n else:\n print(\"none\")\n\n #Reversing an array\n @profile\n def reverse(self,values):\n print(\"THE REVERSE IS: \")\n for i in reversed(values):\n print(i)\n\n # Adding an item to a specific location\n @profile\n def insertion(self, values):\n values.insert(3, 76)\n print(values)\n print(\"\")\n\nif __name__=='__main__':\n t = Timer(\"SubArray\", \"from __main__ import SubArray\")\n print(\"This is the Runtime for this program: \" , t.timeit())\n","sub_path":"question2_3.py","file_name":"question2_3.py","file_ext":"py","file_size_in_byte":1174,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"532292432","text":"import csv\r\nimport pandas as pd\r\npath_no_backedge=pd.DataFrame(pd.read_csv('finished-paths-back.csv'))\r\n\r\ndata_file = open('percentage-paths-back.csv', 'w+',newline='',encoding='utf-8')\r\ncsv_writer = csv.writer(data_file)\r\nheader=['Equal_Length','Larger_by_1','Larger_by_2','Larger_by_3','Larger_by_4','Larger_by_5','Larger_by_6'\r\n ,'Larger_by_7','Larger_by_8','Larger_by_9','Larger_by_10','Larger_by_more_than_10']\r\ncsv_writer.writerow([header[0],header[1],header[2],header[3],header[4],header[5],header[6],header[7],header[8],header[9],header[10],header[11]])\r\n\r\ntotal_count=0\r\ncount_same=0\r\ncount_remaining=[0,0,0,0,0,0,0,0,0,0,0]\r\n#print(path_no_backedge.iloc[0]['human path length'])\r\n\r\nfor i in range(len(path_no_backedge)):\r\n if path_no_backedge.iloc[i]['Shortest_Path_Length']=='NA':\r\n continue\r\n diff=int(path_no_backedge.iloc[i]['Human_Path_Length'])-int(path_no_backedge.iloc[i]['Shortest_Path_Length'])\r\n total_count+=1\r\n if diff==0:\r\n count_same+=1\r\n elif diff >=1 and diff <=10:\r\n count_remaining[diff-1]+=1\r\n else:\r\n count_remaining[10]+=1\r\ncsv_writer.writerow([round((count_same/total_count)*100,2),round((count_remaining[0]/total_count)*100,2),round((count_remaining[1]/total_count)*100,2),\r\n round((count_remaining[2] / total_count) * 100, 2),round((count_remaining[3]/total_count)*100,2),round((count_remaining[4]/total_count)*100,2),\r\n round((count_remaining[5]/total_count)*100,2),round((count_remaining[6]/total_count)*100,2),round((count_remaining[7]/total_count)*100,2),\r\n round((count_remaining[8]/total_count)*100,2),round((count_remaining[9]/total_count)*100,2),round((count_remaining[10]/total_count)*100,2)\r\n ])\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"percentage-paths-back.py","file_name":"percentage-paths-back.py","file_ext":"py","file_size_in_byte":1781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"99254404","text":"# -*- coding: utf-8 -*-\n\nfrom jsresource import JSResource\n\nDEFAULT_LANGUAGE = 'en'\n\nfile_strings = JSResource(\n 'files/',\n fallback_dictionary = DEFAULT_LANGUAGE\n)\n\nwforms = JSResource(\n 'wforms/',\n fallback_dictionary = DEFAULT_LANGUAGE\n)\n\nfallback_strings_1 = JSResource(\n 'fallback_strings_1/',\n fallback_dictionary = DEFAULT_LANGUAGE\n)\n\nfallback_strings_2 = JSResource(\n 'fallback_strings_2/',\n fallback_dictionary = DEFAULT_LANGUAGE\n)\n\nstrings = JSResource(\n 'strings/',\n fallback_dictionary = DEFAULT_LANGUAGE,\n fallback_resources = [wforms, fallback_strings_1, fallback_strings_2]\n)\n\nfor language in 'en', 'ru':\n print('\\nfor', language, 'file:')\n\n print(file_strings[language]('html.index'))\n print(file_strings[language]('html.templates.first_template'))\n print(file_strings[language]('html.templates.second_template'))\n\n # directly defined strings reedifine nested strings\n print(file_strings[language]('image.1'))\n print(file_strings[language]('image.2'))\n print(file_strings[language]('image.3'))\n\n print(strings[language]('messages.hero.i_found_n_items', message_from = 'First hero', n = 12, hero_sex = 'm'))\n print(strings[language]('messages.hero.i_found_n_items', message_from = 'Second hero', n = 21, hero_sex = 'f'))\n print(strings[language]('messages.hero.i_found_n_items', message_from = 'Third hero', n = 1, hero_sex = 'n'))\n\n print(strings[language]('fallback_string_1'))\n print(strings[language]('fallback_string_2'))\n\n # resolve (sub)dictionary to json string\n print(strings[language].resolve_json('interface'))\n print(file_strings[language].resolve_json(''))\n","sub_path":"example/example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":1666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"556203848","text":"from flask import jsonify, request\n\ndef echo_get():\n name = request.args.get('name', '').strip()\n msg = {\"msg\": \"Hello \" + name} if name else {\"msg\": \"Hello\"}\n return jsonify(msg)\n\ndef echo_post():\n if 'name' in request.form:\n name = request.form['name']\n else:\n name = ''\n msg = {\"msg\": \"Hello \" + name} if name else {\"msg\": \"Hello\"}\n return jsonify(msg)\n","sub_path":"REST/swagger_tutorial_2/swagger.py","file_name":"swagger.py","file_ext":"py","file_size_in_byte":391,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"504571515","text":"# Copyright (C) 2017 Boston College\n# http://www.bostoncollege.edu\n#\n# BC Proprietary Information\n#\n# US Government retains Unlimited Rights\n# Non-Government Users – restricted usage as defined through\n# licensing with STR or via arrangement with Government.\n#\n# In no event shall the initial developers or copyright holders be\n# liable for any damages whatsoever, including - but not restricted\n# to - lost revenue or profits or other direct, indirect, special,\n# incidental or consequential damages, even if they have been\n# advised of the possibility of such damages, except to the extent\n# invariable law, if any, provides otherwise.\n#\n# The Software is provided AS IS with NO\n# WARRANTY OF ANY KIND, INCLUDING THE WARRANTY OF DESIGN,\n# MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE.\n\nimport platform\nimport unittest\nimport numpy\nimport collections\nimport os\nimport ctypes as C\nfrom ctypes import cdll\nimport Shared.Utils.HfgeoLogger as Logger\n\nlogger = Logger.getLogger()\n\n## @package IonoModelEngine\n# Python interface to FOUT1\n\n## The interface class to IrtamPyIface\n# Contains an instance of the C interface library\n#\nclass IrtamPyIface:\n\n ## Constructor sets up all the handle\n #\n def __init__(self):\n\n # Figure out which os this is on so that we can load the correct library\n #\n libName = os.path.join(os.path.dirname(os.path.realpath(__file__)),'lib')\n if platform.system() == 'Darwin':\n libName = os.path.join(libName,'libirtampy_mac.dylib')\n elif platform.system() == 'Linux':\n libName = os.path.join(libName,'libirtampy_lin.so')\n \n self.libhandle = cdll.LoadLibrary( libName )\n self.libhandle.processirtam.argtypes = C.POINTER(C.c_double), \\\n C.POINTER(C.c_double), \\\n C.POINTER(C.c_double), \\\n C.POINTER(C.c_double), \\\n C.POINTER(C.c_double), \\\n C.POINTER(C.c_double), \\\n C.POINTER(C.c_double)\n \n self.irtamInput = collections.namedtuple('processirtam_input', \\\n 'alati, along, xmodip, hourut, tov, param, param_local')\n\n def __del__(self):\n import _ctypes\n logger.warning('IrtamPyIface Destructor')\n _ctypes.dlclose(self.libhandle._handle)\n\n ## Return a named tuple object that represent the input to sao reader.\n # For example: \n # param_local, irtam_tov = processirtam(irtamInput)\n # \n # @retval irtamInput A named tuple that the user need to fill in and pass to procesirtam()\n #\n def getIrtamInput(self):\n return self.irtamInput\n\n ## Call processirtam in the lib \n #\n # @param irtamInput - The named tuple input to getprofile\n #\n # @retval param_local, irtam_tov - IRTAM coefficients and Time of Validity\n # \n def processirtam(self, irtamInput):\n\n # Convert from python type to c type\n #\n alati = C.byref(C.c_double(irtamInput.alati))\n along = C.byref(C.c_double(irtamInput.along))\n xmodip = C.byref(C.c_double(irtamInput.xmodip))\n hourut = C.byref(C.c_double(irtamInput.hourut))\n tov = C.byref(C.c_double(irtamInput.tov))\n param = irtamInput.param.ctypes.data_as(C.POINTER(C.c_double))\n param_local = C.c_double(irtamInput.param_local) \n\n # Calling the C interface\n #\n self.libhandle.processirtam(alati, \\\n along, \\\n xmodip, \\\n hourut, \\\n tov, \\\n param,\\\n param_local)\n \n return param_local.value\n\nclass UnitTest_IrtamPyIface(unittest.TestCase):\n\n def test_irtam(self):\n\n iface = IrtamPyIface()\n\n\t # Get the Input object so we can fill it out\n\t #\n irtamInput = iface.getIrtamInput()\n irtamInput.alati = 32.42\n irtamInput.along = 253.71\n irtamInput.xmodip = 25.0\n irtamInput.hourut = 12.\n irtamInput.tov = 12.\n irtamInput.param = numpy.zeros(1064)\n logger.info(irtamInput.param[0:10])\n\n irtamfile = os.path.dirname(os.path.abspath(__file__)) + '/test_data/IRTAM_hmF2_COEFFS_20140119_1545.ASC' \n data = numpy.genfromtxt(irtamfile)\n irtamInput.param = numpy.reshape(data,1064)\n logger.info(irtamInput.param[0:10])\n irtamInput.param_local = numpy.zeros(1)\n (param_local) = iface.processirtam(irtamInput)\n logger.info(param_local)\n\nif __name__ == \"__main__\":\n logger.setLevel('INFO')\n unittest.main()\n \n\n\n\n\n\n\n\n\n","sub_path":"call-04/Send_RIPE/IonoModelEngine/IRTAM/IrtamPyIface.py","file_name":"IrtamPyIface.py","file_ext":"py","file_size_in_byte":4868,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"257905470","text":"\n\n\"\"\"----------------------------------------------------------------------------\nMODULE\n FSuggestPriceTemplate - Module that handles the suggested price for \n the underlying instrument when trading Repo/Reverse and SecurityLoan. \n The suggested price is the market dirty price rounded to 2 decimals. \n\n (c) Copyright 2011 by SunGard FRONT ARENA. All rights reserved.\n\nRENAME this module to FSuggestPrice.\n\n----------------------------------------------------------------------------\"\"\"\nimport ael\nimport acm\n\ndef suggest_undprice(trade):\n\n if not trade.insaddr:\n print ('No instrument found')\n return 0\n \n if not trade.insaddr.und_insaddr:\n print ('No underlying instrument found')\n return 0\n\n context = acm.GetDefaultContext()\n sheet_type = 'FDealSheet'\n calc_space = acm.Calculations().CreateCalculationSpace( context, sheet_type) \n\n marketPrice_ColumnId = 'Instrument Market Price'\n marketPriceDirty_ColumnId = 'Instrument Market Price Dirty'\n fromQ_ColumnId = 'Standard Calculations From Quote'\n theorPrice_ColumnId = 'Standard Calculations Price Theor'\n fromQtn_ColumnId = 'Standard Calculations From Quotation'\n quote_ColumnId = 'Standard Calculations Quote To Quote'\n quoteDay_ColumnId = 'Standard Calculations From Quote Value Date'\n \n trd = acm.FTrade[trade.trdnbr]\n undIns = acm.FInstrument[trade.insaddr.und_insaddr.insid]\n round = acm.GetFunction('round', 3)\n\n if trade.insaddr.instype == 'BuySellback':\n price = calc_space.CreateCalculation( undIns, marketPrice_ColumnId ).Value().Number()\n if not acm.Math.IsFinite( price ):\n price = calc_space.CreateCalculation( undIns, theorPrice_ColumnId ).Value().Number()\n price = round(price, 2, 'Normal')\n else: \n price = calc_space.CreateCalculation( undIns, marketPriceDirty_ColumnId ).Value().Number()\n if not acm.Math.IsFinite( price ):\n price = calc_space.CreateCalculation( undIns, theorPrice_ColumnId ).Value().Number()\n calc_space.SimulateValue( undIns, marketPrice_ColumnId, price )\n price = calc_space.CreateCalculation( undIns, marketPriceDirty_ColumnId ).Value().Number()\n calc_space.RemoveSimulation( undIns, marketPrice_ColumnId )\n\n price = round(price, 2, 'Normal') \n calc_space.SimulateValue( undIns, fromQ_ColumnId, price )\n calc_space.SimulateValue( undIns, fromQtn_ColumnId, 'Pct of Nominal') \n day = trd.ValueDay()\n calc_space.SimulateValue( undIns, quoteDay_ColumnId, day ) \n price = calc_space.CreateCalculation( undIns, quote_ColumnId ).Value().Number() \n calc_space.RemoveSimulation( undIns, fromQ_ColumnId )\n calc_space.RemoveSimulation( undIns, fromQtn_ColumnId )\n calc_space.RemoveSimulation( undIns, quoteDay_ColumnId )\n\n return price\n\n","sub_path":"Extensions/REMod/FPythonCode/FSuggestPriceTemplate.py","file_name":"FSuggestPriceTemplate.py","file_ext":"py","file_size_in_byte":2898,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"89651377","text":"# lib of temporal difference learning for othello\r\nimport ModifiedBench.Function as Function\r\nimport ModifiedBench.Comparator as Comparator\r\nimport ModifiedBench.pso as pso\r\nimport math\r\nimport random\r\nimport numpy as np\r\nimport copy\r\nimport Problem\r\nimport matplotlib.pyplot as plt\r\nimport Othello\r\n\r\nnpop = 10\r\nps = 100\r\npn = \"othello\"\r\n\r\n\r\ndef cprint(cond,text):\r\n if cond:\r\n print(text)\r\n\r\ndef tdl_othello():\r\n max_games = 1e6\r\n games = 0\r\n # strategy representation-WPC\r\n othello = Problem.Problem()\r\n othello.instantiate(pn)\r\n player = Othello.Player()\r\n player.initialize_weight(othello.xlb,othello.xub) # strategy weight matrix\r\n print(player.w)\r\n # while games < max_games:\r\n pass\r\n\r\n\r\n\r\ndef test_selfplay(weight):\r\n # self-play othello\r\n othello = Problem.Problem()\r\n othello.instantiate(pn)\r\n state = Othello.State()\r\n state.is_end()\r\n player = Othello.Player()\r\n player.w = copy.deepcopy(weight) # strategy weight matrix\r\n while not state.end:\r\n # self-play begins\r\n print(\"round\",state.round,\"turn\", state.turn)\r\n skip, state = player.act(state, update_strategy=False) # black player moves, symbol = 1\r\n state.turn = - state.turn # switch to white player strategy\r\n player.symbol = state.turn\r\n print(\"round\", state.round, \"turn\", state.turn)\r\n skip, state = player.act(state, update_strategy=False) # white player moves, symbol = -1\r\n state.turn = - state.turn # switch to black player strategy\r\n player.symbol = state.turn\r\n state.round += 1\r\n if state.end:\r\n if state.winner == 1:\r\n print(\"Black wins.\")\r\n elif state.winner == -1:\r\n print(\"White wins.\")\r\n else:\r\n print(\"End with a draw\")\r\n\r\n\r\ndef test_tdl(verbose=0):\r\n max_games = 3e6\r\n games = 0\r\n best_players = []\r\n # strategy representation-WPC\r\n othello = Problem.Problem()\r\n othello.instantiate(pn)\r\n player = Othello.Player()\r\n player.initialize_weight(othello.xlb, othello.xub) # strategy weight matrix\r\n # training process\r\n while games < max_games:\r\n if games % (max_games / 20) == 0:\r\n best_players.append(copy.deepcopy(player))\r\n print(\"game\",games)\r\n state = Othello.State()\r\n state.is_end()\r\n while not state.end:\r\n # self-play begins\r\n cprint(verbose == 1, \"round \"+str(state.round)+\" turn \"+str(state.turn))\r\n skip, state = player.act(state, display=False) # black player moves, symbol = 1\r\n state.turn = - state.turn # switch to white player strategy\r\n player.symbol = state.turn\r\n cprint(verbose == 1, \"round \"+str(state.round)+\" turn \"+str(state.turn))\r\n skip, state = player.act(state, display=False) # white player moves, symbol = -1\r\n state.turn = - state.turn # switch to black player strategy\r\n player.symbol = state.turn\r\n state.round += 1\r\n # print(np.round(player.w,2))\r\n if state.end and verbose == 1:\r\n if state.winner == 1:\r\n print(\"Black wins.\")\r\n elif state.winner == -1:\r\n print(\"White wins.\")\r\n elif state.winner == 0:\r\n print(\"End with a draw\")\r\n else:\r\n print(\"Invalid winner!\")\r\n games += 1\r\n print(\"Training is over.\")\r\n print(\"Saving the strategy.\")\r\n np.save(\"./Data/weight_mat/tdl_weight.npy\", player.w)\r\n\r\n\r\nif __name__ == '__main__':\r\n # test_tdl()\r\n data = np.load(\"./Data/opt_process/s0.npy\")\r\n player = Othello.Player()\r\n player.load_weight(data)\r\n print(np.round(player.w, 2))\r\n # test_vs_random_opponents([player],oppo_size=1)\r\n Othello.test_vs_fixed_opponent(player, Othello.hp)\r\n","sub_path":"TDL.py","file_name":"TDL.py","file_ext":"py","file_size_in_byte":3868,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"238636773","text":"# Definition for an interval.\nclass Interval:\n def __init__(self, s=0, e=0):\n self.start = s\n self.end = e\n\nclass Solution:\n\n def union(self, interval_a, interval_b):\n return Interval(interval_a.start, max(interval_a.end, interval_b.end))\n\n def is_intersected(self, interval_a, interval_b):\n if interval_a.start <= interval_b.start <= interval_a.end:\n return True\n else:\n return False\n\n def insert(self, intervals, newInterval):\n intervals.append(newInterval)\n intervals = sorted(intervals, key=lambda x: (x.start, -x.end))\n result = []\n tmp_interval = None\n for interval in intervals:\n if tmp_interval is None:\n tmp_interval = Interval(interval.start, interval.end)\n continue\n if self.is_intersected(tmp_interval, interval):\n tmp_interval = self.union(tmp_interval, interval)\n else:\n result.append(Interval(tmp_interval.start, tmp_interval.end))\n tmp_interval = Interval(interval.start, interval.end)\n if tmp_interval:\n result.append(tmp_interval)\n return result\n","sub_path":"51-100/57.py","file_name":"57.py","file_ext":"py","file_size_in_byte":1201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"264359490","text":"#AWS_Restart\n#Ghanaian Names Days Of The Week\n#Daniel Nii Amu Dodoo\n\n#Populating the days of the week in an array\n\ndaysOfWeek = []\n\nprefixes = ['first', 'second', 'third', 'fourth', 'fifth', 'sixth', 'seventh']\n\nwhile len(daysOfWeek) < 7:\n for prefix in prefixes:\n day = input('Enter the ' + prefix + \" day of the week:\")\n daysOfWeek.append(day.capitalize())\n print(daysOfWeek)\n\n if daysOfWeek.count(day) > 1:\n daysOfWeek.pop()\n day = input ('Enter the' + prefix + \" day of the week again:\")\n daysOfWeek.append(day.capitalize())\n print(daysOfWeek)\n\n \n#User inputs sex and day born\nsex_Male_Female = input(\"Enter Your Sex, Male or Female:\")\nday_Born = input(\"Enter the day of the week you were born:\")\n\n\n#Compares users inputs to days \nif sex_Male_Female == \"Male\" and day_Born == \"Monday\":\n print(\"Kwadwo\")\nelif sex_Male_Female == \"Female\" and day_Born == \"Monday\":\n print(\"Ajua\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Tuesday\":\n print(\"Kwabena\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Tuesday\":\n print(\"Abena\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Wednesday\":\n print(\"Kwaku\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Wednesday\":\n print(\"Akua\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Thursday\":\n print(\"Yaw\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Thursday\":\n print(\"Yaa\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Friday\":\n print(\"Kofi\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Friday\":\n print(\"Afia\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Saturday\":\n print(\"Kwame\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Saturday\":\n print(\"Ama\")\n\nif sex_Male_Female == \"Male\" and day_Born == \"Sunday\":\n print(\"Kwesi\")\n\nelif sex_Male_Female == \"Female\" and day_Born == \"Sunday\":\n print(\"Esi\")\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"Loops_Daniel_Updated.py","file_name":"Loops_Daniel_Updated.py","file_ext":"py","file_size_in_byte":1925,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"21764221","text":"from flask import Flask\nfrom flask import json\nfrom flask import jsonify\nfrom flask import request\nimport os\nimport random\nimport hipchat_client\n\nINSULTS = [\n \"Is your ass jealous of the amount of shit that just came out of your mouth?\",\n \"I bet your brain feels as good as new, seeing that you never use it.\",\n \"I'd like to see things from your point of view but I can't seem to get my head that far up my ass.\",\n \"I could eat a bowl of alphabet soup and shit out a smarter statement than that.\",\n \"If I wanted to kill myself I'd climb your ego and jump to your IQ.\",\n \"If you're gonna be a smartass, first you have to be smart. Otherwise you're just an ass.\",\n \"It's better to let someone think you are an Idiot than to open your mouth and prove it.\",\n \"I have neither the time nor the crayons to explain this to you.\",\n \"Why don't you slip into something more comfortable -- like a coma.\",\n \"I may love to shop but I'm not buying your bullshit.\",\n \"Well I could agree with you, but then we'd both be wrong.\",\n \"You're the reason the gene pool needs a lifeguard.\",\n \"You have two brains cells, one is lost and the other is out looking for it.\",\n \"If I gave you a penny for your thoughts, I'd get change.\",\n \"You're as bright as a black hole, and twice as dense.\",\n \"If you spoke your mind, you'd be speechless.\",\n \"Shock me, say something intelligent.\",\n \"So you've changed your mind, does this one work any better?\",\n \"If your brain was made of chocolate, it wouldn't fill an M&M.\",\n \"I can explain it to you, but I can't understand it for you.\",\n \"You are proof that evolution CAN go in reverse.\",\n \"You do realize makeup isn't going to fix your stupidity?\",\n \"Some drink from the fountain of knowledge; you only gargled.\",\n \"It's kinda sad watching you attempt to fit your entire vocabulary into a sentence.\",\n \"You're a person of rare intelligence. It's rare when you show any.\",\n \"I thought you were attractive, but then you opened your mouth.\",\n 'You stare at frozen juice cans because they say, \"concentrate\".',\n \"You're so stupid you tried to wake a sleeping bag.\",\n \"I don't know what makes you so stupid, but it really works!\",\n \"Aww, it's so cute when you try to talk about things you don't understand.\",\n \"Am I getting smart with you? How would you know?\",\n \"I'm not saying I hate you, but I would unplug your life support to charge my phone.\",\n \"You must have been born on a highway because that's where most accidents happen.\",\n \"You bring everyone a lot of joy, when you leave the room.\",\n \"You shouldn't play hide and seek, no one would look for you.\",\n \"I have neither the time nor the crayons to explain this to you.\"\n]\napp = Flask(__name__)\n\n@app.route(\"/\", methods=['GET'])\ndef hello():\n return \"Hello World!\"\n\n@app.route(\"/\", methods=['POST'])\ndef blame():\n request_data = to_json(request)\n \n requesters_room = _extract_requesters_room(request_data)\n arguments = _extract_arguments(request_data)\n requester = _extract_requester(request_data)\n \n guilty = None\n for argument in arguments:\n if '@' in argument:\n guilty = argument\n \n if not guilty:\n members = hipchat_client.get_room_members(requesters_room)\n \n guilty = '@' + random.choice(members) if members else 'all'\n \n if '--with-violence' in arguments:\n message = 'Hey ' + guilty + ' ! ' + random.choice(INSULTS) + ' (megusta)'\n elif '-h' in arguments or '--help' in arguments:\n message = \"@\" + requester + \" I have neither the time nor the crayons to explain this to you.\"\n elif 'why' in arguments:\n message = \"Because !!!!! https://labibliothequedaelinel.files.wordpress.com/2016/10/image.gif?w=656\"\n else:\n message = 'I blame ' + guilty + ' ! >:-('\n \n return jsonify({\n \"color\": \"red\",\n \"message\": message,\n \"notify\": False,\n \"message_format\": \"text\"\n })\n \ndef to_json(request):\n return json.loads(request.data)\n\ndef _extract_arguments(request_data):\n command = request_data['item']['message']['message']\n\n return command.split(' ')[1:]\n\n\ndef _extract_requester(request_data):\n return request_data['item']['message']['from']['mention_name']\n\n\ndef _extract_requesters_room(request_data):\n return request_data['item']['room']['id']\n\nif __name__ == \"__main__\":\n app.debug = True\n app.run(host=os.getenv('IP', '0.0.0.0'), port=int(os.getenv('PORT', '8080')))","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":4521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"427544139","text":"import os\n\nfrom dvc.utils.fs import move, remove\n\n\nclass TestRemote:\n def test(self, tmp_dir, dvc, remote): # pylint: disable=W0613\n (stage,) = tmp_dir.dvc_gen(\"foo\", \"foo\")\n out = stage.outs[0]\n cache = out.cache_path\n foo_hash = out.hash_info\n foo_hashes = out.get_used_objs().get(None, set())\n\n (stage_dir,) = tmp_dir.dvc_gen(\n {\n \"data_dir\": {\n \"data_sub_dir\": {\"data_sub\": \"data_sub\"},\n \"data\": \"data\",\n \"empty\": \"\",\n }\n }\n )\n\n out_dir = stage_dir.outs[0]\n cache_dir = out_dir.cache_path\n dir_hash = out_dir.hash_info\n dir_hashes = {dir_hash} | {oid for _, _, oid in out_dir.obj}\n\n def _check_status(status, **kwargs):\n for key in (\"ok\", \"missing\", \"new\", \"deleted\"):\n expected = kwargs.get(key, set())\n assert expected == set(getattr(status, key))\n\n # Check status\n status = dvc.cloud.status(foo_hashes)\n _check_status(status, new={foo_hash})\n\n status_dir = dvc.cloud.status(dir_hashes)\n _check_status(status_dir, new=dir_hashes)\n\n # Move cache and check status\n # See issue https://github.com/iterative/dvc/issues/4383 for details\n backup_dir = dvc.odb.local.cache_dir + \".backup\"\n move(dvc.odb.local.cache_dir, backup_dir)\n status = dvc.cloud.status(foo_hashes)\n _check_status(status, missing={foo_hash})\n\n status_dir = dvc.cloud.status(dir_hashes)\n _check_status(status_dir, missing=dir_hashes)\n\n # Restore original cache:\n remove(dvc.odb.local.cache_dir)\n move(backup_dir, dvc.odb.local.cache_dir)\n\n # Push and check status\n dvc.cloud.push(foo_hashes)\n assert os.path.exists(cache)\n assert os.path.isfile(cache)\n\n dvc.cloud.push(dir_hashes)\n assert os.path.isfile(cache_dir)\n\n status = dvc.cloud.status(foo_hashes)\n _check_status(status, ok={foo_hash})\n\n status_dir = dvc.cloud.status(dir_hashes)\n _check_status(status_dir, ok=dir_hashes)\n\n # Remove and check status\n remove(dvc.odb.local.cache_dir)\n\n status = dvc.cloud.status(foo_hashes)\n _check_status(status, deleted={foo_hash})\n\n status_dir = dvc.cloud.status(dir_hashes)\n _check_status(status_dir, deleted=dir_hashes)\n\n # Pull and check status\n dvc.cloud.pull(foo_hashes)\n assert os.path.exists(cache)\n assert os.path.isfile(cache)\n with open(cache, encoding=\"utf-8\") as fd:\n assert fd.read() == \"foo\"\n\n dvc.cloud.pull(dir_hashes)\n assert os.path.isfile(cache_dir)\n\n status = dvc.cloud.status(foo_hashes)\n _check_status(status, ok={foo_hash})\n\n status_dir = dvc.cloud.status(dir_hashes)\n _check_status(status_dir, ok=dir_hashes)\n","sub_path":"dvc/testing/test_remote.py","file_name":"test_remote.py","file_ext":"py","file_size_in_byte":2945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"222764041","text":"# Problem 015\n# Starting in the top left corner of a 2×2 grid, and only being able to\n# move to the right and down, there are exactly 6 routes\n# to the bottom right corner.\n# How many such routes are there through a 20×20 grid?\nanswer = 137846528820\ntarget = 20\n\n\ndef solution():\n # compute all combinations for 'to right' move. 'Down' movement\n # are complementary.\n\n result = 1\n\n for i in range(1, target + 1):\n result = result * (target + i) // i\n\n return result\n\n\nif __name__ == '__main__':\n from timeit import Timer\n solution_time = Timer(solution).timeit(number=100)\n time = round(solution_time / 100, 4)\n print(f'Correct answer is: \\t\\t{answer}')\n print(f'Solution answer is: \\t{solution()}')\n print(f'Average solution execution time: {time}')\n","sub_path":"015/problem_015.py","file_name":"problem_015.py","file_ext":"py","file_size_in_byte":793,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"177942156","text":"from datetime import datetime\n\nimport cv2\nfrom constants import Direction, COLUMN, TIMEOUT_FOR_TRACKER\nfrom human_tracker import HumanTracker\nfrom human_validator import HumanValidator\nfrom logger import Logger\n\n\nclass HumanTrackerHandler:\n human_tracking_dict = {}\n\n @classmethod\n def draw_id_centroid_on_output_frame(cls, frame, centroid, objectID):\n # draw both the ID of the object and the centroid of the\n # object on the output frame\n text = \"ID {}\".format(objectID)\n cv2.putText(frame, text, (centroid[0] - 10, centroid[1] - 10)\n , cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 255, 0), 2)\n cv2.circle(frame, (centroid[0], centroid[1]), 4,\n (0, 255, 0), -1)\n\n @classmethod\n def yield_a_human_tracker_object(cls, objects):\n\n for (objectID, centroid) in objects.items():\n # check to see if a trackable object exists for the current\n # object ID\n human_tracker_object = cls.human_tracking_dict.get(objectID, None)\n\n # if there is no existing trackable object, create one\n if not human_tracker_object:\n Logger.logger().debug(\"Creating a new speed tracker object with object id = {}.\".format(objectID))\n human_tracker_object = HumanTracker(objectID, centroid)\n cls.human_tracking_dict[objectID] = human_tracker_object\n\n else:\n human_tracker_object.centroids.append(centroid)\n human_tracker_object.timestamp_list.append(datetime.now())\n yield human_tracker_object, objectID, centroid\n\n @classmethod\n def clear_object_from_speed_tracking_dict(cls, objectID):\n del (cls.human_tracking_dict[objectID])\n\n @classmethod\n def handle_the_case_where_grace_time_for_tracking_is_over(cls, now, human_tracker_object, keep_dict_items):\n \"\"\"\n This method handles the case where the grace time (TIMEOUT_FOR_TRACKER) for the tracker object is over.\n :param now: timestamp\n :param human_tracker_object: Instance of type HumanTracker.\n :param keep_dict_items: Preserve dictionary items. (for debug purpose)\n :return:\n \"\"\"\n if human_tracker_object.estimated and human_tracker_object.logged and not keep_dict_items:\n # Delete this object from speed tracking dict.\n Logger.logger().debug(\"Deleting objectId {} from the human_tracking_dict.\".format(\n human_tracker_object.objectID))\n cls.clear_object_from_speed_tracking_dict(human_tracker_object.objectID)\n else:\n Logger.logger().debug(\"Computing direction for objectId {} because there are no recorded\"\n \" movements for this object in human_tracking_dict.\".format(\n human_tracker_object.objectID))\n cls.compute_direction(human_tracker_object)\n human_tracker_object.estimated = True\n # Finally log it.\n Logger.logger().debug(\"Perform logging for objectId {} found the human_tracking_dict.\".format(\n human_tracker_object.objectID))\n HumanValidator.validate_column_movement(human_tracker_object, now, None,\n human_tracker_object.objectID)\n\n @classmethod\n def compute_direction_for_dangling_object_ids(cls, keep_dict_items=False):\n \"\"\"\n This method computes direction for dangling objects found in speed_tracking_dict.\n This can happen when the person was tracked only at a few sampling points (column traversal).\n :return:\n \"\"\"\n for object_id, human_tracker_object in cls.human_tracking_dict.copy().items():\n now = datetime.now()\n duration = now - human_tracker_object.timestamp_list[-1]\n if duration.total_seconds() > TIMEOUT_FOR_TRACKER:\n cls.handle_the_case_where_grace_time_for_tracking_is_over(now, human_tracker_object, keep_dict_items)\n\n @classmethod\n def compute_direction(cls, trackable_object):\n \"\"\"\n Compute the direction of the person movement.\n :param trackable_object: object\n :return:\n \"\"\"\n direction = trackable_object.centroids[-1][COLUMN] - trackable_object.centroids[0][COLUMN]\n if direction > 0:\n trackable_object.direction = Direction.ENTER\n else:\n trackable_object.direction = Direction.EXIT\n\n @classmethod\n def record_movement(cls, trackable_object):\n if not trackable_object.estimated:\n cls.compute_direction(trackable_object)\n","sub_path":"Occupancy_Tracker/human_tracker_handler.py","file_name":"human_tracker_handler.py","file_ext":"py","file_size_in_byte":4604,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"575473410","text":"def isPalindromes(texts):\n characters = ''''\"!?,.:;'''\n result = []\n for i in range(len(texts)):\n texts[i] = texts[i].lower().replace(' ','')\n for x in range(len(characters)):\n texts[i] = texts[i].replace(characters[x],'')\n if texts[i] == texts[i][::-1]:\n result.append(True)\n else: result.append(False)\n return result\ntexts = [\"Madam, I'm Adam.\",\"rotator\",\"Hello\",\"nurses run\"]\nprint(isPalindromes(texts))\n","sub_path":"구현/이해강_Palindrome.py","file_name":"이해강_Palindrome.py","file_ext":"py","file_size_in_byte":469,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"255733630","text":"import os\nfrom google.cloud import dialogflow\nimport google.api_core.exceptions\nimport json\nimport sys\nimport argparse\nimport uuid\nfrom environs import Env\n\n\nenv = Env()\nenv.read_env()\n\n\ndef parse_arguments():\n parser = argparse.ArgumentParser()\n parser.add_argument(\n '--action',\n dest='action',\n required=True,\n help='action to do'\n )\n\n parser.add_argument(\n '--path',\n dest='path',\n required=False,\n help='Path to json file with intents'\n )\n\n parser.add_argument(\n '--intent_id',\n dest='intent_id',\n required=False,\n help='id intent to delete'\n )\n\n return parser.parse_args()\n\n\ndef list_intents(project_id):\n intents_client = dialogflow.IntentsClient()\n parent = dialogflow.AgentsClient.agent_path(project_id)\n intents = intents_client.list_intents(request={\"parent\": parent})\n\n for intent in intents:\n print(\"=\" * 20)\n print(f\"Intent name: {intent.name}\")\n print(f\"Intent display_name: {intent.display_name}\")\n print(f\"Action: {intent.action}\\n\")\n print(f\"Root followup intent: {intent.root_followup_intent_name}\")\n print(f\"Parent followup intent: \\\n {intent.parent_followup_intent_name}\\n\")\n\n print(\"Input contexts:\")\n for input_context_name in intent.input_context_names:\n print(\"\\tName: {}\".format(input_context_name))\n\n print(\"Output contexts:\")\n for output_context in intent.output_contexts:\n print(\"\\tName: {}\".format(output_context.name))\n\n\ndef create_intent(\n project_id,\n display_name,\n training_phrases_parts,\n message_texts\n):\n\n \"\"\"Create an intent of the given intent type.\"\"\"\n\n intents_client = dialogflow.IntentsClient()\n\n parent = dialogflow.AgentsClient.agent_path(project_id)\n training_phrases = []\n for training_phrases_part in training_phrases_parts:\n part = dialogflow.Intent.TrainingPhrase.Part(\n text=training_phrases_part\n )\n # Here we create a new training phrase for each provided part.\n training_phrase = dialogflow.Intent.TrainingPhrase(parts=[part])\n training_phrases.append(training_phrase)\n\n text = dialogflow.Intent.Message.Text(text=message_texts)\n message = dialogflow.Intent.Message(text=text)\n\n intent = dialogflow.Intent(\n display_name=display_name,\n training_phrases=training_phrases, messages=[message]\n )\n\n response = intents_client.create_intent(\n request={\"parent\": parent, \"intent\": intent}\n )\n\n print(\"Intent created: {}\".format(response))\n\n\ndef delete_intent(project_id, intent_id):\n \"\"\"Delete intent with the given intent type and intent value.\"\"\"\n\n intents_client = dialogflow.IntentsClient()\n intent_path = intents_client.intent_path(project_id, intent_id)\n intents_client.delete_intent(request={\"name\": intent_path})\n\n\ndef read_file(filepath):\n with open(filepath, 'r') as file:\n file_content = json.load(file)\n return file_content\n\n\ndef main():\n project_id = env('PROJECT_ID')\n args = parse_arguments()\n\n if args.action == 'create' and os.path.exists(args.path):\n try:\n intent_content = read_file(args.path)\n except FileNotFoundError:\n sys.exit('File not found. Exit')\n\n for intent in intent_content:\n questions = intent_content[intent]['questions']\n answer = intent_content[intent]['answer']\n try:\n create_intent(project_id, intent, questions, [answer])\n except google.api_core.exceptions.ServiceUnavailable:\n sys.exit(\"Can't reach google service\")\n except google.api_core.exceptions.FailedPrecondition:\n print(f'Failed precondition for intent with name \"{intent}\"')\n continue\n\n elif args.action == 'list':\n list_intents(project_id)\n\n elif args.action == 'delete' and args.intent_id:\n try:\n isinstance(uuid.UUID(args.intent_id), uuid.UUID)\n except ValueError:\n sys.exit('Invalid intent ID')\n delete_intent(project_id, args.intent_id)\n else:\n sys.exit('Invalid arguments')\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"intent.py","file_name":"intent.py","file_ext":"py","file_size_in_byte":4262,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"468532010","text":"#!/usr/bin/env python\nimport numpy\nimport numpy as np\nimport scipy.special\n\n_alpha = 0.32\n\n\ndef interval(best,lo=np.nan,hi=np.nan):\n \"\"\"\n Pythonized interval for easy output to yaml\n \"\"\"\n return [float(best),[float(lo),float(hi)]]\n\ndef mean_interval(data, alpha=_alpha):\n \"\"\"\n Interval assuming gaussian posterior.\n \"\"\"\n mean =np.mean(data)\n sigma = np.std(data)\n scale = scipy.stats.norm.ppf(1-alpha/2.)\n return interval(mean,mean-scale*sigma,mean+scale*sigma)\n\n\ndef median_interval(data, alpha=_alpha):\n \"\"\"\n Median including bayesian credible interval.\n \"\"\"\n q = [100*alpha/2., 50, 100*(1-alpha/2.)]\n lo,med,hi = numpy.percentile(data,q)\n return interval(med,lo,hi)\n \ndef peak(data, bins=100):\n num,edges = np.histogram(data,bins=bins)\n centers = (edges[1:]+edges[:-1])/2.\n return centers[np.argmax(num)]\n\ndef kde_peak(data, samples=1000):\n \"\"\"\n Identify peak using Gaussian kernel density estimator.\n \"\"\"\n return kde(data,samples)[0]\n\ndef kde(data, samples=1000):\n \"\"\"\n Identify peak using Gaussian kernel density estimator.\n \"\"\"\n # Clipping of severe outliers to concentrate more KDE samples in the parameter range of interest\n mad = np.median(np.fabs(np.median(data) - data))\n cut = (data > np.median(data) - 5. * mad) & (data < np.median(data) + 5. * mad)\n x = data[cut]\n kde = scipy.stats.gaussian_kde(x)\n # No penalty for using a finer sampling for KDE evaluation except computation time\n values = np.linspace(np.min(x), np.max(x), samples) \n kde_values = kde.evaluate(values)\n peak = values[np.argmax(kde_values)]\n return values[np.argmax(kde_values)], kde.evaluate(peak)\n\n\ndef peak_interval(data, alpha=_alpha, samples=1000):\n \"\"\"\n Identify interval using Gaussian kernel density estimator.\n \"\"\"\n peak = kde_peak(data,samples)\n x = np.sort(data.flat); n = len(x)\n # The number of entries in the interval\n window = int(np.rint((1.0-alpha)*n))\n # The start, stop, and width of all possible intervals\n starts = x[:n-window]; ends = x[window:]\n widths = ends - starts\n # Just the intervals containing the peak\n select = (peak >= starts) & (peak <= ends)\n widths = widths[select]\n if len(widths) == 0:\n raise ValueError('Too few elements for interval calculation')\n min_idx = np.argmin(widths)\n lo = x[min_idx]\n hi = x[min_idx+window]\n return interval(peak,lo,hi)\n\ndef min_interval(data, alpha=_alpha):\n x = np.sort(data.flat); n = len(x)\n # The number of entries in the interval\n window = int(np.rint((1.0-alpha)*n))\n # The start, stop, and width of all possible intervals\n starts = x[:n-window]; ends = x[window:]\n widths = ends - starts\n if len(widths) == 0:\n raise ValueError('Too few elements for interval calculation')\n min_idx = np.argmin(widths)\n lo = x[min_idx]\n hi = x[min_idx+window]\n mean = (hi+lo)/2.\n return interval(mean,lo,hi)\n\n\ndef norm_cdf(x):\n # Faster than scipy.stats.norm.cdf\n #https://en.wikipedia.org.wiki/Normal_distribution\n return 0.5*(1 + scipy.special.erf(x/np.sqrt(2)))\n\ndef random_pdf(value,pdf,size=None):\n if size is None: size = 1.0\n cdf = np.cumsum(pdf)\n cdf /= cdf[-1]\n fn = scipy.interpolate.interp1d(cdf, range(0, len(cdf)))\n index = np.rint(fn(np.random.uniform(size=size))).astype(int)\n return value[index]\n\ndef sky(lon=None,lat=None,size=1):\n \"\"\"\n Outputs uniform points on sphere from:\n [0 < lon < 360] & [-90 < lat < 90]\n \"\"\"\n if lon is None:\n umin,umax = 0,1\n else:\n lon = np.asarray(lon)\n lon = np.radians(lon + 360.*(lon<0))\n if lon.size==1: umin=umax=lon/(2*np.pi)\n elif lon.size==2: umin,umax=lon/(2*np.pi)\n else: raise Exception('...')\n \n if lat is None:\n vmin,vmax = -1,1\n else:\n lat = np.asarray(lat)\n lat = np.radians(90 - lat)\n if lat.size==1: vmin=vmax=np.cos(lat)\n elif lat.size==2: vmin,vmax=np.cos(lat)\n else: raise Exception('...')\n\n phi = 2*np.pi*np.random.uniform(umin,umax,size=size)\n theta = np.arcsin(np.random.uniform(vmin,vmax,size=size))\n return np.degrees(phi),np.degrees(theta)\n\nif __name__ == \"__main__\":\n import argparse\n description = \"python script\"\n parser = argparse.ArgumentParser(description=description)\n parser.add_argument('args',nargs=argparse.REMAINDER)\n opts = parser.parse_args(); args = opts.args\n\n import pylab as plt\n ax=plt.subplot(221,projection='aitoff')\n ax.grid(True)\n lon,lat = sky(size=1e3)\n lon,lat= np.radians([lon-360.*(lon>180),lat])\n ax.scatter(lon,lat,marker='.',s=2)\n\n ax=plt.subplot(222,projection='aitoff')\n ax.grid(True)\n lon,lat = sky(size=1e3,lat=[30,45])\n lon,lat= np.radians([lon-360.*(lon>180),lat])\n ax.scatter(lon,lat,marker='.',s=2)\n\n ax=plt.subplot(223,projection='aitoff')\n ax.grid(True)\n lon,lat = sky(size=1e3,lon=[30,45])\n lon,lat= np.radians([lon-360.*(lon>180),lat])\n ax.scatter(lon,lat,marker='.',s=2)\n\n ax=plt.subplot(224,projection='aitoff')\n ax.grid(True)\n lon,lat = sky(size=1e3,lon=[0,45],lat=[30,45])\n lon,lat= np.radians([lon-360.*(lon>180),lat])\n ax.scatter(lon,lat,marker='.',s=2)\n","sub_path":"ugali/utils/stats.py","file_name":"stats.py","file_ext":"py","file_size_in_byte":5266,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"468480744","text":"from django.conf.urls import patterns, include\nfrom django.contrib.auth.decorators import login_required as lr\nfrom django.contrib.auth.decorators import permission_required as pr\n\nfrom src import url\nfrom src.vendor.views import MyTemplateView\nfrom .views import IncidenceRestView\nfrom .views import StatisticsRestView\nfrom xilema.views import GenericExcelView\n\n\nanomalies = patterns(\n '',\n url(r'^$',\n lr(MyTemplateView.as_view(\n template_name='incidence/anomalies.html')),\n name='incidencesanomalies', permissions=[\"list_incidences\", \"export_excel\"]))\n\napi = patterns(\n '',\n url(r'^anomalies/$',\n pr('actions.change_actionclient')(IncidenceRestView.as_view()),\n name='anomalies'),\n\n url(r'^statistics/$',\n pr('actions.change_actionclient')(StatisticsRestView.as_view()),\n name='statistics'))\n\nurlpatterns = patterns(\n '',\n url(r'^$',\n lr(MyTemplateView.as_view(\n template_name='incidence/index.html')),\n name='statisticsincidences', permissions=[\"list_incidences\"]),\n\n url(r'^anomalies/',\n include(anomalies, namespace='anomalies')),\n\n url(r'^export/$',\n lr(GenericExcelView.as_view()),\n name='ex'),\n\n url(r'^api/',\n include(api, namespace='api')))\n","sub_path":"src/apps/incidence/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"192536923","text":"import arcpy\nimport os\nimport pandas as pd\nimport re\n\nfrom Dev.create_basemap import get_rasters\n\ndef process_rasters(rasters, summary, field, threshold):\n summary = summary[['scenario_id_raw', field]]\n for i, raster in enumerate(rasters):\n print(\"\\t{}\".format(i))\n raster = arcpy.Raster(raster)\n fields = [f.name for f in arcpy.ListFields(raster)]\n print(fields)\n exit()\n\n if not i:\n running = raster\n else:\n running += raster\n return running > 0\n\ndef get_fields(durations, weights):\n return [\"{}_{}_%ile\".format(d, w) for d in durations for w in weights]\n\n\ndef main():\n combo_dir = os.path.join(\"..\", \"..\", \"aquatic-model-inputs\", \"bin\", \"Intermediate\", \"CombinedRasters\")\n combo_format = re.compile(\"c_(.{2,3})_(\\d{4})$\")\n summary_file = os.path.join(\"Output\", \"test_summary.csv\")\n durations = [\"1-day\", \"90-day\"]\n weights = [\"unweighted\", \"weighted\"]\n regions = ['07']\n threshold = 90\n\n summary = pd.read_csv(summary_file)\n for region in regions:\n years, rasters = get_rasters(combo_dir, combo_format, region)\n for field in get_fields(durations, weights):\n print(region, field)\n combined = process_rasters(rasters, summary, field, threshold)\n combined.save(out_format.format(region, _class, years[0], years[-1]))\n\n\nmain()\n","sub_path":"ScenarioSelection/Dev/map_exceedances.py","file_name":"map_exceedances.py","file_ext":"py","file_size_in_byte":1383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"378803163","text":"\"\"\"-----------------------------------------------------------------------------\nProject: Prime Brokerage Project\nDepartment: Prime Services\nRequester: Francois Henrion\nDeveloper: Paul Jacot-Guillarmod\nCR Number: 666125 (Initial Deployment)\n\nHISTORY\n================================================================================\nDate Change no Developer Description\n--------------------------------------------------------------------------------\n2011-06-17 685737 Paul J.-Guillarmod Added a function call add TPL to the portfolio swap\n2011-12-08 850928 Herman Hoon Added the Provision function call\n2012-05-08 Peter Kutnik Fixed earlier undocumented change for non cash collateral\n2014-11-20 2450799 Peter Fabian Fixing errors and exception handling\n2015-09-11 3090331 Jakub Tomaga Portfolio independent sweeping\n2018-11-12 CHG1001113453 Tibor Reiss Enable fully funded CFD for MMIBETA2\n2019-03-27 FAPE-65 Tibor Reiss Remove fully funded CFD for MMIBETA2 (yes, you are reading it correctly)\n2019-07-15 FAPE-47 Iryna Shcherbina Create a sweeping report\n2019-11-21 FAPE-147 Tibor Reiss Propagate error\n-----------------------------------------------------------------------------\"\"\"\nimport acm\n\nfrom PS_Functions import (CALENDAR, TODAY, START_DATE_KEYS, START_DATE_LIST, END_DATE_KEYS, END_DATE_LIST, DateGenerator)\nimport PS_FundingSweeper\nimport PS_TimeSeriesFunctions\nfrom at_logging import getLogger, bp_start\nfrom sweeping_report import PSwapSweepingReport\n\n\nLOGGER = getLogger()\n\n\ndef enable_custom_start_date(index, field_values):\n ael_variables[2][9] = (field_values[1] == 'Custom Date')\n return field_values\n\n\ndef enable_custom_end_date(index, field_values):\n ael_variables[4][9] = (field_values[3] == 'Custom Date')\n return field_values\n\n\n# Variable Name, Display Name, Type, Candidate Values, Default, Mandatory, Multiple, Description, Input Hook, Enabled\nael_variables = [['portfolioSwaps', 'Portfolio Swaps', 'FInstrument', None, None, 1, 1, 'Portfolio swaps that will have their funding cashflows extended', None, 1],\n ['startDate', 'Start Date', 'string', START_DATE_KEYS, 'Now', 1, 0, 'Date from which the portfolio swaps will be extended.', enable_custom_start_date, 1],\n ['startDateCustom', 'Start Date Custom', 'string', None, TODAY, 0, 0, 'Custom from date', None, 0],\n ['endDate', 'End Date', 'string', END_DATE_KEYS, 'Now', 1, 0, 'Date to which the portfolio swaps will be extended.', enable_custom_end_date, 1],\n ['endDateCustom', 'End Date Custom', 'string', None, TODAY, 0, 0, 'Custom to date', None, 0],\n ['compoundPortfolio', 'Compound Client Portfolio', 'FCompoundPortfolio', None, None, 0, 0, 'Fixed Income ValStart will be calculated for the compound portfolio to select the overnight funding spread.', None, 1],\n ['collateralPortfolios', 'Compound Collateral Portfolio', 'FCompoundPortfolio', None, None, 0, 1, 'Compound portfolio where the Collateral trades are booked.', None, 1],\n ['resweepTPL', 'Re-Sweep TPL', 'string', ['Yes', 'No'], 'No', 1, 0, 'When running backdated sweeping, indicate whether TPL should be reswept', None, 1],\n ['clientName', 'Short name', 'string', None, 'CLIENT', 0, 0],\n ['sweepingReport', 'Sweeping Report', 'string', None, None, 0, 0, 'Report with detailed breakdown of swept amounts', None, 1],\n ]\n\n\ndef ael_main(ael_dict):\n LOGGER.msg_tracker.reset()\n process_name = \"ps.extend_pswap.{0}\".format(ael_dict[\"clientName\"])\n\n with bp_start(process_name): \n if ael_dict['startDate'] == 'Custom Date':\n start_date = ael_dict['startDateCustom']\n else:\n start_date = START_DATE_LIST[ael_dict['startDate']]\n \n if ael_dict['endDate'] == 'Custom Date':\n end_date = ael_dict['endDateCustom']\n else:\n end_date = END_DATE_LIST[ael_dict['endDate']]\n \n resweep_tpl = ael_dict['resweepTPL'] == 'Yes'\n \n portfolio_swaps = ael_dict['portfolioSwaps']\n report_data = {}\n for portfolio_swap in portfolio_swaps:\n report_data[portfolio_swap.Name()] = ExtendPortfolioSwap(\n portfolio_swap, resweep_tpl, start_date, end_date)\n \n collateral_portfolios = ael_dict['collateralPortfolios']\n if portfolio_swaps:\n portfolio_swap = portfolio_swaps[0]\n portfolio = portfolio_swap.FundPortfolio()\n call_account = acm.FDeposit[portfolio.add_info('PSClientCallAcc')]\n if collateral_portfolios:\n collateral_portfolio = collateral_portfolios[0]\n for date in DateGenerator(start_date, end_date):\n _SetCollateralTimeSeries(collateral_portfolio, date, call_account, 'Collateral Value')\n\n report_filename = ael_dict[\"sweepingReport\"]\n if report_filename:\n try:\n file_path = report_filename.format(date=end_date.replace(\"-\", \"\"))\n report = PSwapSweepingReport(file_path, report_data)\n report.create_report()\n LOGGER.info(\"Wrote secondary output to %s\", file_path)\n except Exception:\n LOGGER.exception(\"Sweeping report wasn't generated.\")\n\n if LOGGER.msg_tracker.errors_counter:\n raise RuntimeError(\"ERRORS occurred. Please check the log.\")\n\n LOGGER.info(\"Completed Successfully\")\n\n\ndef _SetCollateralTimeSeries(collateral_portfolio, date, call_account, column_name):\n time_series_dict = {'FI': 'PS_Collateral_FI',\n 'EQ': 'PS_Collateral_EQ',\n 'Cash': 'PS_Collateral_MM'\n }\n collateral_type_grouper = acm.FAttributeGrouper('Instrument.CollateralType')\n query_folder = _GenerateValStartQuery(collateral_portfolio)\n collateral_dict = PS_FundingSweeper.TradingManagerSweeper(query_folder, date, [column_name], False,\n collateral_type_grouper, 'ZAR')\n collateral_dict_keys = iter(collateral_dict.keys())\n\n if collateral_dict_keys:\n for key in collateral_dict_keys:\n time_series_name = time_series_dict[key]\n value = collateral_dict[key]\n if value:\n value = value[0]\n PS_TimeSeriesFunctions.UpdateTimeSeriesValue(time_series_name, call_account, value, date)\n LOGGER.info('Set time series %s, value %s for %s on %s to %s',\n time_series_name, column_name, call_account.Name(), date, value)\n\n\ndef _GenerateValStartQuery(compound_portfolio):\n \"\"\"Generate a query folder that will be used to calculate Overnight Spread ValStart for the compound portfolio.\"\"\"\n query = acm.CreateFASQLQuery('FTrade', 'AND')\n query.AddAttrNode('Status', 'NOT_EQUAL', acm.EnumFromString('TradeStatus', 'Simulated'))\n query.AddAttrNode('Status', 'NOT_EQUAL', acm.EnumFromString('TradeStatus', 'Void'))\n \n # Add the sub portfolios to the query\n or_node = query.AddOpNode('OR')\n if not compound_portfolio.AllPhysicalPortfolios():\n raise ValueError(\"Portfolio %s has no physical subportfolios\" % compound_portfolio.Name())\n for portfolio in compound_portfolio.AllPhysicalPortfolios():\n or_node.AddAttrNode('Portfolio.Name', 'EQUAL', portfolio.Name())\n \n return query\n\n\ndef _CalculateOvernightSpreadValStart(query, date):\n \"\"\"\n Calculate the total value of Overnight Spread ValStart for the previous banking day. This will be\n used to decide on the spread to be used in the overnight funding.\n \"\"\"\n previous_banking_day = CALENDAR.AdjustBankingDays(date, -1)\n total_val_start = 0\n instrument_val_starts = PS_FundingSweeper.TradingManagerSweeper(query, previous_banking_day,\n ['Overnight Spread ValStart'], True)\n for ins, valStartList in instrument_val_starts.items():\n val_start = valStartList[0]\n total_val_start += val_start\n return total_val_start\n\n\ndef ExtendPortfolioSwap(portfolioSwap, resweepTPL, startDate, endDate):\n \"\"\"Run all the modules needed to extend a portfolio swap for each date between startDate and endDate inclusive.\"\"\"\n report_data = {}\n portfolio = portfolioSwap.FundPortfolio()\n for date in DateGenerator(startDate, endDate):\n PS_TimeSeriesFunctions.UpdateTimeSeries('PSExtExecPremRate', 'PSExtExecPremRate', portfolio, date)\n PS_TimeSeriesFunctions.UpdateTimeSeries('PSExtExecPremNonDMA', 'PSExtExecPremNonDMA', portfolio, date)\n PS_TimeSeriesFunctions.UpdateTimeSeries('PSShortPremRate', 'PSShortPremRate', portfolioSwap, date)\n funding = PS_FundingSweeper.GenerateFunding(portfolioSwap, date)\n provision = PS_FundingSweeper.GenerateProvision(portfolioSwap, date)\n\n # In general we don't want to overwrite the historical TPL resets, unless explicitly choosing to do so.\n tpl = {}\n if date == TODAY or resweepTPL:\n tpl = PS_FundingSweeper.GenerateTotalTPL(portfolioSwap, date)\n report_data[date] = (funding, provision, tpl)\n\n return report_data\n","sub_path":"Python modules/PS_ExtendPortfolioSwap.py","file_name":"PS_ExtendPortfolioSwap.py","file_ext":"py","file_size_in_byte":9358,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"240046318","text":"'''\r\nCreated on Jun 4, 2012\r\n\r\n@author: Nich\r\n'''\r\n\r\ndirs = {\r\n \"e\" : (1, 0),\r\n \"w\" : (-1, 0),\r\n \"d\" : (0, -1),\r\n \"u\" : (0, 1),\r\n }\r\n\r\ndef add_dirs(*dirs):\r\n res = [0, 0]\r\n for d in dirs:\r\n for i, val in enumerate(d):\r\n res[i] = res[i]+val\r\n return tuple(res)","sub_path":"src/MessagePassingMud/map/directions.py","file_name":"directions.py","file_ext":"py","file_size_in_byte":321,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"57643056","text":"import numpy as np\r\n\r\nclass Perceptron(object):\r\n\r\n def __init__(self, inputs, bias, maxIter = 1000, rate = 1):\r\n self.maxIter = maxIter\r\n self.rate = rate\r\n self.weights = np.array(inputs)\r\n self.bias = bias\r\n\r\n def predict(self, inputs):\r\n output = np.dot(inputs, self.weights) + self.bias\r\n if output >= 0:\r\n activation = 1\r\n else:\r\n activation = 0\r\n return activation\r\n\r\n def train(self, train_inputs, labels):\r\n for _ in range(self.maxIter):\r\n for inputs, label in zip(train_inputs, labels):\r\n prediction = self.predict(inputs)\r\n # np.add(self.weights, self.rate * (label - prediction) * inputs, out = self.weights, casting = \"unsafe\")\r\n self.weights = self.weights + self.rate * (label - prediction) * inputs\r\n self.bias = self.bias + self.rate * (label - prediction)\r\n # print(self.weights, self.bias)\r\n # print(self.weights, self.bias)","sub_path":"Perceptron Algorithm/New_Perceptron.py","file_name":"New_Perceptron.py","file_ext":"py","file_size_in_byte":1034,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"3052221","text":"#!/usr/bin/env python3\nfrom pwn import *\nimport pwnlib.shellcraft\nimport pwnlib\nimport time\n\npwn_file = \"./helloctf\"\nelf = ELF(pwn_file)\ncontext.os='linux'\ncontext.arch = 'amd64'\ncontext.terminal = ['tmux', 'splitw', '-h']\n\ncontext.log_level = 100\ncontext.log_level = 'debug'\n\nc = process(pwn_file)\ndbg_script = \"\"\"\n\"\"\"\n#gdb.attach(c,dbg_script)\nc = remote('ctf.adl.tw', '11001')\nc.sendlineafter('CTF', b'A' * (16 + 8) + p64(0x4006c7))\nc.interactive()\nc.close()\n","sub_path":"ctf.adl.tw-2020/helloctf.py","file_name":"helloctf.py","file_ext":"py","file_size_in_byte":462,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"434145406","text":"from bs4 import BeautifulSoup\nimport requests\n\n\ndef get_soup(target):\n response = requests.get(url=target)\n return BeautifulSoup(response.content, 'html.parser')\n\n\ndef form_wiki_href_from_slug(slug):\n if slug[0:3] == \"/w/\" or len(slug) == 0:\n return None\n return \"https://en.wikipedia.org\" + slug\n\n\ndef get_clean_ele_text(ele):\n return ele.text.strip()\n\n\ndef get_current_events_from_soup(soup, day, year, destination):\n wiki_current_itn = soup.find(\n \"div\", {\"aria-labelledby\": \"Topics_in_the_news\"})\n\n wiki_current_itn_featured = wiki_current_itn.find('ul').find_all('li')\n\n destination['featured'] = []\n\n for ele in wiki_current_itn_featured:\n news = {}\n link = ele.find('b').find('a')\n news['title'] = link['title']\n news['href'] = form_wiki_href_from_slug(link['href'])\n destination['featured'].append(news)\n\n wiki_current_ongoing = soup.find(\n \"div\", {\"aria-labelledby\": \"Ongoing_events\"})\n\n ongoing_categories_headers = wiki_current_ongoing.find_all(\"h3\")\n ongoing_categories_lists = wiki_current_ongoing.find_all(\"ul\")\n\n for i, category_header in enumerate(ongoing_categories_headers):\n category = get_clean_ele_text(category_header).lower()\n destination[category] = []\n\n for li in ongoing_categories_lists[i].find_all('li'):\n news = {}\n link = li.find('a')\n news['title'] = link['title']\n news['href'] = form_wiki_href_from_slug(link['href'])\n destination[category].append(news)\n\n scrape_target = \"https://en.wikipedia.org/wiki/Deaths_in_%s\" % (year)\n soup = get_soup(scrape_target)\n\n wiki_recent_deaths_day_header = soup.find('h3')\n\n header_int = int(wiki_recent_deaths_day_header.text)\n\n if header_int > int(day):\n wiki_recent_deaths_day_header = soup.find('h3').find_next_sibling(\n 'h3')\n\n wiki_recent_deaths = wiki_recent_deaths_day_header.find_next_sibling(\n 'ul').find_all('li')\n\n destination['deaths'] = []\n\n for li in wiki_recent_deaths:\n news = {}\n link = li.find('a')\n news['title'] = link['title']\n href = form_wiki_href_from_slug(link['href'])\n\n if href:\n news['href'] = href\n\n destination['deaths'].append(news)\n\n return\n","sub_path":"main/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"345812111","text":"\n\nfrom xai.brain.wordbase.nouns._petard import _PETARD\n\n#calss header\nclass _PETARDS(_PETARD, ):\n\tdef __init__(self,): \n\t\t_PETARD.__init__(self)\n\t\tself.name = \"PETARDS\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"petard\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_petards.py","file_name":"_petards.py","file_ext":"py","file_size_in_byte":238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"9541614","text":"import numpy as np\nimport tensorflow as tf\n\n\n#Restore\n\nwinput = tf.Variable(np.arange(10).reshape(1,10),dtype=tf.float32, name = \"Layer1_Weights\")\nbinput = tf.Variable(np.arange(10).reshape(1,10),dtype = tf.float32, name = \"Layer1_Bias\")\nwoutput = tf.Variable(np.arange(10).reshape(10,1),dtype=tf.float32, name = \"OutPut_Weights\")\nboutput = tf.Variable(np.arange(1).reshape(1,1),dtype = tf.float32, name = \"OutPut_Bias\")\nsaver = tf.train.Saver()\nsess = tf.Session()\nsaver.restore(sess,\"my_net/save_net.ckpt\")\nprint(\"winput\", sess.run(winput))\nprint(\"binput\", sess.run(binput))\nprint(\"woutput\", sess.run(woutput))\nprint(\"boutput\", sess.run(boutput))","sub_path":"myprictice/python/tf/RestoreFbp.py","file_name":"RestoreFbp.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"298142044","text":"'''\nCreated on Oct 30, 2018\n\n@author: daniel\n'''\n\nfrom keras.models import Model, Input\nfrom keras.layers import Convolution2D, Activation, BatchNormalization,MaxPooling2D, Convolution2DTranspose, Dropout, concatenate\nfrom Mylayers import MaxPoolingWithArgmax2D, MaxUnpooling2D\n\ndef inceptionModule(inputs, numFilters = 32):\n \n tower_0 = Convolution2D(numFilters, (1,1), padding='same', kernel_initializer = 'he_normal')(inputs)\n tower_0 = BatchNormalization()(tower_0)\n tower_0 = Activation(\"relu\")(tower_0)\n \n tower_1 = Convolution2D(numFilters, (1,1), padding='same',kernel_initializer = 'he_normal')(inputs)\n tower_1 = BatchNormalization()(tower_1)\n tower_1 = Activation(\"relu\")(tower_1)\n tower_1 = Convolution2D(numFilters, (3,3), padding='same',kernel_initializer = 'he_normal')(tower_1)\n tower_1 = BatchNormalization()(tower_1)\n tower_1 = Activation(\"relu\")(tower_1)\n \n tower_2 = Convolution2D(numFilters, (1,1), padding='same',kernel_initializer = 'he_normal')(inputs)\n tower_2 = BatchNormalization()(tower_2)\n tower_2 = Activation(\"relu\")(tower_2)\n tower_2 = Convolution2D(numFilters, (3,3), padding='same',kernel_initializer = 'he_normal')(tower_2)\n tower_2 = BatchNormalization()(tower_2)\n tower_2 = Activation(\"relu\")(tower_2)\n \n tower_3 = MaxPooling2D((3,3), strides=(1,1), padding='same')(inputs)\n tower_3 = Convolution2D(numFilters, (1,1), padding='same',kernel_initializer = 'he_normal')(tower_3)\n tower_3 = BatchNormalization()(tower_3)\n tower_3 = Activation(\"relu\")(tower_3)\n \n inception_module = concatenate([tower_0, tower_1, tower_2, tower_3], axis = 3)\n return inception_module\n \ndef createUNetInceptionIndexPooling(input_shape = (240,240,1), output_mode=\"sigmoid\", pool_size=(2,2)):\n inputs = Input(input_shape)\n \n numFilters = 16;\n \n conv1 = inceptionModule(inputs, numFilters)\n pool1, mask1 = MaxPoolingWithArgmax2D(pool_size)(conv1)\n \n conv2 = inceptionModule(pool1, 2*numFilters)\n pool2, mask2 = MaxPoolingWithArgmax2D(pool_size)(conv2)\n \n conv3 = inceptionModule(pool2, 4*numFilters)\n drop3 = Dropout(0.5)(conv3)\n pool3, mask3 = MaxPoolingWithArgmax2D(pool_size)(drop3)\n\n conv4 = inceptionModule(pool3, 8*numFilters)\n drop4 = Dropout(0.5)(conv4)\n pool4, mask4 = MaxPoolingWithArgmax2D(pool_size)(drop4)\n \n up6 = MaxUnpooling2D(pool_size)([pool4, mask4])\n merge6 = concatenate([conv4,up6],axis=3)\n conv6 = inceptionModule(merge6, 4*numFilters)\n \n up7 = MaxUnpooling2D(pool_size)([conv6, mask3])\n merge7 =concatenate([conv3,up7],axis=3)\n conv7 = inceptionModule(merge7, 2*numFilters)\n \n up8 = MaxUnpooling2D(pool_size)([conv7, mask2])\n merge8 =concatenate([conv2,up8],axis=3)\n conv8 = inceptionModule(merge8,numFilters)\n \n up9 = MaxUnpooling2D(pool_size)([conv8, mask1])\n merge9 =concatenate([conv1,up9],axis=3)\n conv9 = inceptionModule(merge9, numFilters)\n \n conv9 = Convolution2D(numFilters, 3, activation = 'relu', padding = 'same', kernel_initializer = 'he_normal')(conv9)\n conv9 = Convolution2D(1, 3, activation = 'relu', padding = 'same', kernel_initializer = 'he_normal')(conv9)\n conv10 = Convolution2D(1, 1, activation = output_mode)(conv9)\n\n model = Model(input = inputs, output = conv10)\n\n #model.compile(optimizer = Adam(lr = 1e-4), loss = 'binary_crossentropy', metrics = ['accuracy'])\n \n #model.summary()\n\n return model","sub_path":"UNetFactory/createUNetInceptionIndexPooling.py","file_name":"createUNetInceptionIndexPooling.py","file_ext":"py","file_size_in_byte":3462,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"648240499","text":"from setuptools import setup, find_packages\n\n\ndef read_file(name):\n with open(name) as fobj:\n return fobj.read().strip()\n\n\nLONG_DESCRIPTION = read_file(\"README.md\")\nVERSION = read_file(\"Ctl/VERSION\")\n\nsetup(\n name=\"django-ixpmgr\",\n version=VERSION,\n author=\"20C\",\n author_email=\"code@20c.com\",\n description=\"django overlay for IXP-Manager\",\n long_description=LONG_DESCRIPTION,\n long_description_content_type=\"text/markdown\",\n license=\"LICENSE.txt\",\n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Topic :: Internet\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n ],\n packages=find_packages(\"src\"),\n package_dir={\"\": \"src\"},\n include_package_data=True,\n url=\"https://github.com/20c/django-ixpmgr\",\n download_url=f\"https://github.com/20c/django-ixpmgr/archive/{VERSION}.zip\",\n install_requires=[\"semver==2.10.2\"],\n zip_safe=False,\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"2176926","text":"from django.shortcuts import render , HttpResponse , redirect\nfrom django.http import JsonResponse\nfrom datetime import datetime\nfrom .models import Book , Profile , Comment , Rating\nfrom .forms import BookForm, UserForm , Profile , CommentForm, RegisterForm, ProfileForm , SettingForm , PasswordChangeingForm\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth.models import User\nfrom django.core.mail import send_mail, BadHeaderError\nfrom django.contrib.auth.forms import PasswordResetForm , PasswordChangeForm\nfrom django.contrib.auth.views import PasswordChangeView\nfrom django.template.loader import render_to_string\nfrom django.db.models.query_utils import Q\nfrom django.utils.http import urlsafe_base64_encode\nfrom django.contrib.auth.tokens import default_token_generator\nfrom django.utils.encoding import force_bytes\nfrom django.contrib import messages\nfrom django.urls import path ,reverse_lazy, include\n\n\n# Create your views here.\n\n# Get to HOME page\ndef home(request):\n # book= Book.objects.all()\n return render(request, 'home.html')\n\ndef signin(request):\n return render(request,\"user/signin.html\")\n\ndef search_by_type(request,mtype):\n book= Book.objects.filter(book_type=mtype)\n return render(request, 'book/searchbytype.html', {'books': book} ) \n\n\n#profile\ndef profile(request,pk): \n profile = Profile.objects.get(pk=pk)\n\n return render(request , 'profile/profile_show.html',{'profile':profile} )\n\ndef Profile_edit(request,pk):\n profile_edit = Profile.objects.get(pk=pk)\n if request.method == \"POST\":\n form = ProfileForm(request.POST, instance=profile_edit)\n if form.is_valid():\n profile_edit = form.save()\n messages.add_message(request , messages.SUCCESS , \"Update profile successfully\")\n return redirect('profile', pk=profile_edit.pk)\n else:\n form = ProfileForm(instance=profile_edit)\n return render(request, 'profile/profile_edit.html', {'form': form})\n\n\n\n\n# ------------------------------ book route ----------------------------\n# To display all the books\ndef books(request):\n all_books = Book.objects.all()\n horrorBook= Book.objects.filter(book_type=\"horror\")\n actionBook= Book.objects.filter(book_type=\"action\")\n return render(request , 'book/index.html', {\"all_books\" : all_books, 'horrorBook': horrorBook, 'actionBook': actionBook})\n\n# To view a selector book\ndef view_book(request, pk):\n \n total= 0\n count1=0\n count2=0\n count3=0\n count4=0\n count5=0\n id_ = pk\n rate= Rating.objects.filter(book=pk)\n if rate.count() == 0:\n total=1\n else: \n total=rate.count()\n for rete in rate:\n if rete.score == 1:\n count1 += 1\n if rete.score == 2:\n count2 += 1\n if rete.score == 3:\n count3 += 1\n if rete.score == 4:\n count4 += 1\n if rete.score == 5:\n count5 += 1\n # count1=count1/total*100\n # count2=count2/total*100\n # count3=count3/total*100\n # count4=count4/total*100\n # count5=count5/total*100\n try:\n one_book = Book.objects.get(pk=pk)\n except Exception:\n return HttpResponse(\"error\")\n\n return render(request , 'book/view.html', {\"book\" : one_book, \"count1\" : count1, \"count2\" : count2, \"count3\" : count3 , \"count4\" : count4, \"count5\" : count5})\n\n\ndef show_books(request):\n book= Book.objects.all()\n \n return render(request, 'home.html', {'all_book': book} ) \n\n# To add a new book\n@login_required()\ndef add_book(request):\n if request.method == 'POST':\n form = BookForm(request.POST)\n if form.is_valid():\n form.save()\n messages.add_message(request , messages.SUCCESS , \"Update book successfully\")\n return redirect('/book/')\n else:\n if request.user.profile.user_status:\n form = BookForm()\n return render(request , 'book/book_form.html', {'form': form})\n else:\n return redirect('/home/')\n\n# To Update the book\n@login_required()\ndef edit_book(request,pk):\n edit_book= Book.objects.get(pk=pk)\n form = BookForm(instance=edit_book)\n if (request.method == \"POST\"):\n Edit_book =BookForm(request.POST,instance=edit_book) \n if Edit_book.is_valid() :\n Edit_book.save()\n messages.add_message(request , messages.SUCCESS , \"Added book successfully\")\n return redirect (f'/book/{pk}')\n return render(request , 'book/book_form.html' , {\"form\" : form})\n\n# To Delete the book\n@login_required()\ndef delete_book(request , pk):\n book = Book.objects.get(pk=pk)\n book.delete()\n messages.add_message(request , messages.SUCCESS , \"Deleted book successfully\")\n return redirect('/book/')\n\n# Add a book to a favoriate list\n@login_required()\ndef favorite_book(request , pk):\n user = User.objects.get(pk = request.user.id)\n book = Book.objects.get(pk=pk)\n print(user.profile.fav_books.all)\n if book in user.profile.fav_books.all():\n user.profile.fav_books.remove(book)\n else:\n user.profile.fav_books.add(book)\n\n request.user = user\n print( user.profile.fav_books.all())\n return redirect(f'/book/{pk}/')\n\n# Add a book to a wantToRead\n@login_required()\ndef wantToRead(request , pk):\n user = User.objects.get(pk = request.user.id)\n book = Book.objects.get(pk=pk)\n if book in user.profile.wantToRead.all():\n user.profile.wantToRead.remove(book)\n else:\n user.profile.wantToRead.add(book)\n request.user = user\n print( user.profile.wantToRead.all())\n return redirect(f'/book/{pk}/')\n \n#----------------------- Auth ----------------------------------\ndef singin(request):\n messages.add_message(request , messages.SUCCESS , \"Login successfully\")\n return render(request,\"user/singin.html\")\n\ndef register(request):\n form = RegisterForm()\n if( request.method == \"POST\"):\n user = RegisterForm(request.POST)\n if(user.is_valid()):\n user.save()\n messages.add_message(request , messages.SUCCESS , \"Register successfully\")\n return redirect('/home/')\n else:\n form = user\n messages.add_message(request, messages.ERROR, 'Register failure')\n return render(request, \"user/register.html\", { \"form\": form })\n\n\nclass PasswordChangeView(PasswordChangeView):\n from_class= PasswordChangeingForm\n success_url=reverse_lazy('password_change_success')\n\ndef PasswordChangeDone(request):\n return render(request,\"auth/password_change_done.html\")\n\n#----------------------- profile ----------------------------------\n#show profile \ndef profile(request,pk): \n profile = Profile.objects.get(pk=pk)\n\n return render(request , 'profile/profile_show.html',{'profile':profile} )\n\n#Edit profile \n@login_required()\ndef Profile_edit(request,pk):\n profile_edit = Profile.objects.get(pk=pk)\n if request.method == \"POST\":\n form = ProfileForm(request.POST, request.FILES,instance=profile_edit)\n if form.is_valid():\n profile_edit = form.save()\n messages.add_message(request , messages.SUCCESS , \"Update profile successfully\")\n return redirect('profile', pk=profile_edit.pk)\n else:\n form = ProfileForm(instance=profile_edit)\n return render(request, 'profile/profile_edit.html', {'form': form})\n\n\n\n\n@login_required()\ndef Setting(request,pk):\n if request.method == \"POST\":\n form = SettingForm(request.POST, instance=request.user)\n if form.is_valid():\n profile_Setting = form.save()\n return redirect('profile', pk=request.user.id)\n else:\n form = SettingForm(instance=request.user)\n return render(request, 'profile/setting.html', {'form': form})\n\n# To add a new cooment\n@login_required()\ndef add_comment(request,pk):\n book1=Book.objects.get(pk=pk)\n if request.method == 'POST':\n form = CommentForm(request.POST, request.user)\n if form.is_valid():\n comment = form.save(commit=False)\n comment = Comment.objects.create(\n user =request.user,\n name=comment.name,\n comment=comment.comment,\n )\n comment.book.add(book1)\n comment.save()\n return redirect(f'/book/{pk}/')\n else:\n form = CommentForm()\n \"\"\" messages.add_message(request, messages.WARNING, \"Failed to add your comment\") \"\"\"\n\n return render(request , 'comment/add_comment.html', {'form': form})\n\n#To edit the comment\n@login_required()\ndef edit_comment(request,pk,pkb):\n edit_comment= Comment.objects.get(pk=pk)\n if edit_comment.user == request.user:\n form = CommentForm(instance=edit_comment)\n if (request.method == \"POST\"):\n Edit_comment =CommentForm(request.POST,instance=edit_comment) \n if Edit_comment.is_valid() :\n Edit_comment.save()\n return redirect (f'/book/{pkb}')\n else:\n return render(request , 'comment/add_comment.html' , {\"form\" : form})\n else: \n html = \"you can't do it
\" \n return HttpResponse(html)\n\n# To Delete the comment\n@login_required()\ndef delete_comment(request , pk):\n comment = Comment.objects.get(pk=pk)\n if comment.user == request.user:\n comment.delete()\n messages.add_message(request , messages.SUCCESS , \"The comment deleted successfully\")\n return redirect('home')\n else: \n messages.add_message(request, messages.WARNING, \"Failed to add your comment\")\n\n\ndef Search(request):\n book= Book.objects.all()\n query = request.GET.get(\"q\")\n if query:\n book = Book.objects.filter(Q(name__icontains=query) )\n\n return render(request,'book/Search.html',{'all_books': book})\n\n# To add a new reting for a book\n@login_required()\ndef add_rate(request,pk):\n book1=Book.objects.get(pk=pk)\n val = request.POST['rate']\n print (\"rate : \"+request.POST['rate'])\n print (\"----------------------------\")\n print ( request.user)\n\n if request.method == 'POST':\n rate = Rating.objects.create(\n user =request.user,\n book=book1,\n score = request.POST['rate']\n )\n rate.save()\n messages.add_message(request , messages.SUCCESS , \"Added your rating successfully\")\n return redirect(f'/book/{pk}/')\n messages.add_message(request , messages.WARNING,\"There is an error please try again\")\n return redirect(f'/book/{pk}/')\n\n\n","sub_path":"book/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":10523,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"388900114","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('noccapp', '0004_auto_20150512_1242'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='patient',\n name='user',\n field=models.ForeignKey(verbose_name=b'utente', to=settings.AUTH_USER_MODEL, unique=True),\n ),\n ]\n","sub_path":"noccapp/migrations/0005_auto_20150512_1447.py","file_name":"0005_auto_20150512_1447.py","file_ext":"py","file_size_in_byte":486,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"631381342","text":"# -*- encoding: utf-8 -*-\nimport logging\nimport unittest\nfrom mock import patch,MagicMock\nfrom src.controller.endpoint_handlers.base_endpoint_handler import \\\n BaseEndpointHandler\nfrom src.controller.logs.logger_factory import LoggerFactory\n\n\nclass TestBaseEndpointHandler(unittest.TestCase):\n\n def setUp(self ):\n logger = LoggerFactory.get_logger(\n # This path is set to be used with nosetests, so do not change.\n # Rather change the Pycharm launcher instead if needed and set it\n # to xxx\\ms-cloud\\python\\OutputHandlerNode\n \"./src/controller/logs/processing_node.log\", logging.INFO\n )\n #We disable the stream logger so exceptions do not print\n logger.handlers = logger.handlers[:-1]\n self.handler = BaseEndpointHandler(None, logger)\n\n def tearDown(self ):\n pass\n\n def test_log_endpoint_reads_file(self):\n input = \"Test\\nstring\\nfor\\nlog\\nfile\"\n expected = \"Test
string
for
log
file\"\n\n m = self.get_mock_for_file_open(input, 28)\n with patch('__builtin__.open', m) as mock:\n result = self.handler.handle_log_get()\n self.assertEquals(expected, result)\n\n def get_mock_for_file_open(self, file_text, file_bytes):\n mock = MagicMock(spec=file)\n handle = MagicMock(spec=file)\n handle.write.return_value = None\n handle.read.return_value = file_text\n handle.tell.return_value = file_bytes\n handle.__enter__.return_value = handle\n \"\"\"\n This is the tricky part. You have to specified a return_value to the\n mock intself. This way, when the test_subject calls open() the handle\n mock will be returned.\n\n If you do not assign a return_value to mock, then a random new mock\n will be returned to the test subject, which will have no values set\n for read() or tell().\n \"\"\"\n mock.return_value = handle\n return mock\n\nif __name__ == '__main__':\n unittest.main()","sub_path":"test/test_base_endpoint_handler.py","file_name":"test_base_endpoint_handler.py","file_ext":"py","file_size_in_byte":2017,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"83952569","text":"from django.shortcuts import render\nfrom django.views.generic import View\nfrom django.http import HttpResponse\nfrom django.contrib.auth import authenticate,login\n\nfrom .forms import UserAskForm\nfrom .models import UserAsk\n\n# Create your views here.\n\nclass ContentView(View):\n def get(self, request):\n return render(request, \"contact.html\")\n\n\ndef AskForm(request):\n if request.method == \"POST\":\n send_form = UserAskForm(request.POST) # 实例化 \\ 初始化form\n if send_form.is_valid(): # 验证是否合法(即是否满足forms.py的设置)\n msg = request.POST.get(\"msg\", \"\") # 取出 user_name 和 pass_word\n name = request.POST.get(\"name\", \"\")\n if UserAsk.objects.filter(msg=msg,name=name): # 判断 用户是否存在(如果存在)\n return render(request, \"contact.html\", {\"register_form\": send_form,\n \"msg\": \"请勿重复提交!\"}) # 即 这个 email已经被注册过,不存在就进行下面的操作,传回form回填信息,天使吧验证码保存\n name = request.POST.get(\"name\", \"\") # 取出 user_name\n email = request.POST.get(\"email\", \"\")\n mobile = request.POST.get(\"mobile\", \"\")\n zy = request.POST.get(\"zy\", \"\")\n msg = request.POST.get(\"msg\", \"\")\n content = authenticate(name=name,email=email,mobile=mobile,zy=zy,msg=msg)\n\n user_ask = UserAsk()\n user_ask.name = name\n user_ask.mobile = mobile\n user_ask.email = email\n user_ask.zy = zy\n user_ask.msg = msg\n user_ask.save()\n\n if content is not None:\n login(request,content)\n return render(request,\"contact.html\")\n else:\n return render(request,\"contact.html\",{\"msg\":\"添加信息出错\"})\n elif request.method ==\"GET\":\n return render(request,\"contact.html\")\n\n\n\n\n","sub_path":"apps/content/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1996,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"262644444","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Jan 26 16:31:08 2020\n\n@author: charles\n\"\"\"\nimport time\nimport notify2\nfrom top_news import top_story_sender\n\n\ndef notification_sender():\n \n \n news_recevied = top_story_sender()\n \n notify2.init('notifier')#itializer for d bus\n \n n = notify2.Notification(None) #notification object\n \n n.set_urgency(notify2.URGENCY_NORMAL)\n \n n.set_timeout(10000)\n \n for news in news_recevied:\n n.update(news['title'],news['description'])\n \n n.show()\n \n time.sleep(15000)\n \n \n \n\nnotification_sender()\n\n \n \n \n\n \n","sub_path":"notifier.py","file_name":"notifier.py","file_ext":"py","file_size_in_byte":666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"427930848","text":"import pyglet\nfrom pyglet.window import key, mouse\n\n# set pyglet resource dir\nimage_path = \"../../assets/image\"\npyglet.resource.path = [image_path]\n\n# create window, text and image objects\nwindow = pyglet.window.Window()\nlabel = pyglet.text.Label(\n \"Hello world\",\n font_name=\"Times New Roman\",\n font_size=36,\n x=window.width//2, \n y=window.height//2,\n anchor_x=\"center\", \n anchor_y=\"center\"\n )\nimage = pyglet.resource.image(\"mini.jpg\")\n \n@window.event\ndef on_key_press(symbol, modifiers):\n if symbol == key.A:\n print(\"A\")\n\n if symbol == key.LEFT:\n print(\"LEFT\")\n \n if symbol == key.ENTER:\n print(\"ENTER\")\n\n@window.event\ndef on_mouse_press(x, y, button, modifiers):\n if button == mouse.LEFT:\n print(\"mouse left\")\n\n@window.event\ndef on_draw():\n window.clear()\n image.blit(0, 0)\n label.draw()\n\nif __name__ == \"__main__\":\n pyglet.app.run()\n","sub_path":"example/01-PygletApp/hello_world.py","file_name":"hello_world.py","file_ext":"py","file_size_in_byte":1111,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"286636257","text":"from flask import Flask\nfrom flask import redirect, url_for\n\n\napp = Flask(__name__)\n\n\n@app.route('/')\ndef index():\n return 'Hello World!'\n\n@app.route('/user/')\ndef Hello(username):\n if username == 'victor':\n return f'Hello {username}'\n #return redirect(url_for('index'))\n return redirect('/', 303)\n\n@app.route('/user/')\ndef user_index(username):\n return f'Hello {username}'\n\n@app.route('/post/')\ndef show_post(post_id):\n return f'Post {post_id}'\n\n@app.route('/test')\ndef test():\n print(url_for('index'))\n print(url_for('user_index', username='shiyan'))\n print(url_for('show_post', post_id=2, _external=True))\n print(url_for('show_post', post_id=2, q='python 03'))\n print(url_for('show_post', post_id=2, q='python你好'))\n print(url_for('show_post', post_id=2, _anchor='a'))\n return 'test'\n\n@app.route('/test')\ndef test():\n print('xxx.simplelab.cn/courses/java')\n return redirect('/', 301)\n","sub_path":"flask_code_learn/app_test.py","file_name":"app_test.py","file_ext":"py","file_size_in_byte":975,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"576109641","text":"#! /usr/bin/env python3\n\nimport re\nimport subprocess\nimport os\n\n\ndef path(*path_segments):\n return os.path.join(os.getcwd(), *path_segments)\n\n\ndef open_file(*path_segments):\n file_path = path(*path_segments)\n open(file_path, 'w').close()\n return open(file_path, 'a')\n\ndef get_commands():\n setup_py = open('setup.py', 'r')\n lines = setup_py.read()\n return re.findall(r' {8,}\\'([a-z0-9-]*.py) = jfscripts.[a-z0-9_]*:main\\',',\n lines)\n\n\ndef get_help(command):\n return subprocess.Popen('{} --help'.format(command), shell=True,\n stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n\n\ndef heading(text):\n return '\\n{}\\n{}\\n\\n.. code-block:: text\\n\\n'.format(text, '-' * len(text))\n\n\ndef main():\n commands = get_commands()\n\n header = open(path('README_header.rst'), 'r')\n readme = open_file('README.rst')\n\n # footer = open(path('README_footer.rst'), 'r')\n\n for line in header:\n readme.write(line)\n\n readme.write('\\n')\n for command in commands:\n print(command)\n _help = get_help(command)\n _help.wait()\n readme.write(heading(command))\n\n for line in _help.stdout:\n indented_line = ' ' + line.decode('utf-8')\n readme.write(indented_line)\n\n # for line in footer:\n # readme.write(line)\n\n readme.close()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"_generate_doc.py","file_name":"_generate_doc.py","file_ext":"py","file_size_in_byte":1401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"640487918","text":"import numpy\nfrom keras.utils import np_utils\nfrom keras.datasets import mnist\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Flatten, Conv2D, MaxPooling2D\nfrom com.kailin.api_image import api_image\nfrom com.kailin.api_file import api_file\n\nnumpy.random.seed(10)\n\n(train_image, train_label), (test_image, test_label) = mnist.load_data()\n\n# train_image_hot = train_image.reshape(60000, 784).astype('float32') / 255\n# test_image_hot = test_image.reshape(10000, 784).astype('float32') / 255\n\ntrain_image_hot = train_image.reshape(train_image.shape[0], 28, 28, 1).astype('float32') / 255\ntest_image_hot = test_image.reshape(test_image.shape[0], 28, 28, 1).astype('float32') / 255\n\ntrain_label_hot = np_utils.to_categorical(train_label)\ntest_label_hot = np_utils.to_categorical(test_label)\n\nmodel = Sequential()\nmodel.add(Conv2D(filters=16, kernel_size=(5, 5), padding='same', input_shape=(28, 28, 1), activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Conv2D(filters=36, kernel_size=(5, 5), padding='same', activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\nmodel.add(Dropout(0.25))\nmodel.add(Flatten())\nmodel.add(Dense(units=512, activation='relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(units=128, activation='relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(units=10, activation='softmax'))\n# default version\n# model.add(Dense(input_dim=784, units=1024, kernel_initializer='normal', activation='relu'))\n# model.add(Dropout(0.5))\n# model.add(Dense(units=512, kernel_initializer='normal', activation='relu'))\n# model.add(Dropout(0.5))\n# model.add(Dense(units=10, kernel_initializer='normal', activation='softmax'))\n\nprint(model.summary())\n\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\n\n# try:\n# # model.load_weights('E:\\pythonwork\\data\\e06.h5')\n# api_file.loadMode(model,'E:\\pythonwork\\data\\e06.h5')\n# except:\nhistory = model.fit(x=train_image_hot, y=train_label_hot, validation_split=0.2, epochs=10, batch_size=300, verbose=2)\napi_image.showTrainHistory(history, 'acc', 'val_acc')\napi_image.showTrainHistory(history, 'loss', 'val_loss')\nscores = model.evaluate(test_image_hot, test_label_hot)\nprint('\\n', scores, '\\n')\n\nprediction = model.predict_classes(test_image_hot)\napi_image.showImageLabelPrediction(test_image, test_label, prediction, 0, 25)\napi_image.confusionMatrix(test_label, prediction)\n\napi_file.saveMode(model, 'E:\\pythonwork\\data\\e06.h5')\n# model.save_weights('E:\\pythonwork\\data\\e06.h5')\n","sub_path":"tensorflow/com/kailin/e06.py","file_name":"e06.py","file_ext":"py","file_size_in_byte":2515,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"130968287","text":"from flask import Flask, request, jsonify\r\nfrom flask_restful import Resource, Api\r\nfrom json import dumps\r\nimport datetime\r\n\r\nclass UserInsuranceOffers(Resource):\r\n def post(self):\r\n # define result structure with null values\r\n result = {\r\n 'auto': None,\r\n 'disability': None,\r\n 'home': None,\r\n 'life': None\r\n }\r\n\r\n # treating when receiving an empty body request\r\n if request.json != None:\r\n # get object user sent\r\n user = request.json\r\n # defining user`s base risk score\r\n base_score = 0\r\n for i in user['risk_questions']:\r\n base_score += i\r\n # set initial risk score for each user`s insurance line\r\n risk_score = {\r\n 'auto': base_score,\r\n 'disability': base_score,\r\n 'home': base_score,\r\n 'life': base_score\r\n } \r\n # get actual year for future comparison\r\n this_year = datetime.datetime.now().year\r\n\r\n if user['income'] == 0:\r\n result['disability'] = 'ineligible'\r\n if len(user['house']) == 0:\r\n result['home'] = 'ineligible' \r\n if len(user['vehicle']) == 0:\r\n result['auto'] = 'ineligible'\r\n\r\n if user['age'] > 60:\r\n result['disability'] = 'ineligible'\r\n result['life'] = 'ineligible'\r\n elif user['age'] < 30:\r\n risk_score['auto'] -= 2\r\n risk_score['disability'] -= 2\r\n risk_score['home'] -= 2\r\n risk_score['life'] -= 2\r\n elif user['age'] >= 30 and user['age'] <= 40:\r\n risk_score['auto'] -= 1\r\n risk_score['disability'] -= 1\r\n risk_score['home'] -= 1\r\n risk_score['life'] -= 1\r\n\r\n if user['income'] > 200000:\r\n risk_score['auto'] -= 1\r\n risk_score['disability'] -= 1\r\n risk_score['home'] -= 1\r\n risk_score['life'] -= 1\r\n\r\n if len(user['house']) > 0 and user['house']['ownership_status'] == 'mortgaged':\r\n risk_score['home'] += 1\r\n\r\n if user['dependents'] > 0:\r\n risk_score['disability'] += 1\r\n risk_score['life'] += 1\r\n\r\n if user['marital_status'] == 'married':\r\n risk_score['disability'] -= 1\r\n risk_score['life'] += 1\r\n\r\n if len(user['vehicle']) > 0 and this_year - user['vehicle']['year'] <= 5:\r\n risk_score['auto'] += 1\r\n\r\n # defining user`s final score\r\n for key in result:\r\n # only allow final score changing if insurance life wasn`t previously defined as ineligible for user\r\n if result[key] != 'ineligible':\r\n if risk_score[key] <= 0:\r\n result[key] = 'economic'\r\n elif risk_score[key] > 0 and risk_score[key] < 3:\r\n result[key] = 'regular'\r\n elif risk_score[key] >= 3:\r\n result[key] = 'responsible' \r\n \r\n return jsonify(result)\r\n","sub_path":"app/main/classes/user_insurance_offers.py","file_name":"user_insurance_offers.py","file_ext":"py","file_size_in_byte":3364,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"93021957","text":"# -*- coding: utf-8 -*-\n\"\"\"\nWritten by Daniel M. Aukes and CONTRIBUTORS\nEmail: danaukesseas.harvard.edu.\nPlease see LICENSE for full license.\n\"\"\"\n\nfrom multiprocessing import Process\nfrom multiprocessing import Queue\nimport time\nimport random\nimport yaml\nimport os\n \n#def run_wrapper(a,b):\n# def wrap1(f):\n# print(a,b)\n# def wrapped_func(*args,**kwargs):\n# d = f(*args,**kwargs)\n# return d\n# return wrapped_func\n# return wrap1\n#\n#@run_wrapper('hey','babe')\n#def e(a,b,c):\n# print('inner',a,b,c)\n# return c\n \nclass DummyClass(object):\n def run(self): \n time.sleep(random.gauss(3,1))\n self.dummy = 'asdfsadfafds'\n\nclass ProcessDataWrapper(object):\n ii = 0\n def __init__(self,inner_process):\n self.index = ProcessDataWrapper.ii\n self.inner_process = inner_process\n ProcessDataWrapper.ii+=1\n def run_outer(self):\n self.time_started = time.time()\n self.inner_process.run()\n self.time_finished = time.time()\n self.time_elapsed = self.time_finished - self.time_started\n def save(self,folder):\n if not os.path.exists(folder):\n os.mkdir(folder)\n with open(os.path.normpath(os.path.join(folder,str(self.index)+'.dat')),'w') as f:\n yaml.dump(self,f)\n \nclass ProcessData(object):\n ii = 0\n def __init__(self):\n self.index = ProcessData.ii\n ProcessData.ii+=1\n def run_outer(self):\n self.time_started = time.time()\n self.run()\n self.time_finished = time.time()\n self.time_elapsed = self.time_finished - self.time_started\n def save(self,folder):\n with open(os.path.normpath(os.path.join(folder,str(self.index)+'.dat')),'w') as f:\n yaml.dump(self,f)\n \nclass DummyClass2(ProcessData):\n def run(self): \n time.sleep(random.gauss(3,1))\n self.dummy = 'asdfsadfafds'\n \nclass ProcessContainer(Process):\n def __init__(self,inner_process):\n Process.__init__(self)\n# self.process_data = ProcessDataWrapper(inner_process)\n self.process_data = inner_process\n def run(self):\n self.process_data.run_outer()\n self.queue.put(self.process_data)\n def set_queue(self,queue):\n self.queue = queue\n \nclass ProcessManager(object):\n sleep_time = .5\n def __init__(self,processes,max_processes=4,debug=True,block_till_finished = True,collect_all_data = True,save_data = False,save_dest = '.',refresh_display = None):\n self.all_processes = [ProcessContainer(p) for p in processes]\n self.processes = self.all_processes.copy()\n self.max_processes = max_processes\n self.debug = debug\n self.collect_all_data = collect_all_data\n if self.collect_all_data:\n self.data= []\n self.save_data = save_data\n self.save_dest = save_dest\n self.block_till_finished = block_till_finished\n self.running = []\n self.done = []\n self.num_processes = len(self.processes)\n self.queue=Queue()\n self.checkalivedead()\n if refresh_display!=None:\n self.refresh_display = refresh_display\n \n def run(self):\n self.time_launched = time.time()\n self.print_debug('launching processes') \n while not not self.processes:\n if self.num_running < self.max_processes:\n num_can_start = self.max_processes - self.num_running \n num_remaining = len(self.processes)\n num_to_start = min(num_can_start,num_remaining)\n new_processes = []\n for ii in range(num_to_start):\n p = self.processes.pop(0)\n new_processes.append(p)\n p.set_queue(self.queue)\n p.start()\n self.running.append(p)\n# for p in new_processes:\n# p.join()\n string = '{0:0.0f} slots, {1:0.0f} remaining, {2:0.0f} started'.format(num_can_start,num_remaining,num_to_start)\n self.print_debug(string)\n self.checkalivedead()\n self.pull_queue()\n time.sleep(self.sleep_time)\n# for processss in self.all_processes:\n# proce.join()\n\n if self.block_till_finished:\n self.print_debug('waiting to clear')\n while not not self.running:\n self.checkalivedead()\n self.pull_queue()\n time.sleep(self.sleep_time)\n\n def checkalivedead(self):\n running = []\n new_done = False\n for process in self.running:\n if process.is_alive():\n running.append(process)\n else:\n self.done.append(process)\n new_done = True\n \n self.running = running\n self.num_running = len(self.running)\n self.num_done= len(self.done)\n if self.debug:\n if new_done:\n time_current = time.time()\n time_elapsed = time_current - self.time_launched\n time_remaining = (time_current-self.time_launched)/self.num_done*self.num_processes\n self.print_debug('{0:.2f} minutes elapsed/{1:.2f} minutes total, {2:d}/{3} finished'.format(time_elapsed/60,time_remaining/60,self.num_done,self.num_processes))\n\n def pull_queue(self):\n newitems = []\n while not self.queue.empty():\n item = self.queue.get()\n if self.collect_all_data:\n self.data.append(item)\n if self.save_data:\n item.save(self.save_dest)\n newitems.append(item)\n try:\n self.refresh_display(newitems)\n except AttributeError:\n pass\n \n def print_debug(self,*args,**kwargs):\n if self.debug:\n print(args,kwargs)\n\n\nif __name__ == '__main__':\n t0 = time.time()\n processes = [DummyClass2() for index in range(20)]\n pm = ProcessManager(processes,max_processes=10,debug=True,block_till_finished = True,collect_all_data = False,save_data = True,save_dest = './data')\n pm.run()\n \n t1 = time.time()\n print('elapsed time: ',t1-t0)\n","sub_path":"dev_tools/process_manager.py","file_name":"process_manager.py","file_ext":"py","file_size_in_byte":6271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"44208307","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\nCopyright (C) 2011 Dariusz Suchojad \n\nThis program is free software: you can redistribute it and/or modify\nit under the terms of the GNU General Public License as published by\nthe Free Software Foundation, either version 3 of the License, or\n(at your option) any later version.\n\nThis program is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\nGNU General Public License for more details.\n\nYou should have received a copy of the GNU General Public License\nalong with this program. If not, see .\n\"\"\"\n\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\n# SQLAlchemy\nfrom sqlalchemy.sql.expression import case, literal_column\n\n# Zato\nfrom zato.common.odb.model import(ChannelAMQP, ChannelWMQ, ChannelZMQ, Cluster, \n ConnDefAMQP, ConnDefWMQ, CronStyleJob, HTTPBasicAuth, HTTPSOAP, IntervalBasedJob, \n Job, OutgoingAMQP, OutgoingFTP, OutgoingS3, OutgoingWMQ, OutgoingZMQ, \n Service, TechnicalAccount, WSSDefinition)\n\n# ##############################################################################\n\ndef internal_channel_list(session, cluster_id):\n \"\"\" All the HTTP/SOAP channels that point to internal services.\n \"\"\"\n return session.query(HTTPSOAP.soap_action, Service.name).\\\n filter(HTTPSOAP.cluster_id==Cluster.id).\\\n filter(HTTPSOAP.service_id==Service.id).\\\n filter(Service.is_internal==True).\\\n filter(Cluster.id==cluster_id)\n\n# ##############################################################################\n\ndef job_list(session, cluster_id):\n \"\"\" All the scheduler's jobs defined in the ODB.\n \"\"\"\n return session.query(Job.id, Job.name, Job.is_active,\n Job.job_type, Job.start_date, Job.extra,\n Service.name.label('service_name'), Service.id.label('service_id'),\n IntervalBasedJob.weeks, IntervalBasedJob.days,\n IntervalBasedJob.hours, IntervalBasedJob.minutes,\n IntervalBasedJob.seconds, IntervalBasedJob.repeats,\n CronStyleJob.cron_definition).\\\n outerjoin(IntervalBasedJob, Job.id==IntervalBasedJob.job_id).\\\n outerjoin(CronStyleJob, Job.id==CronStyleJob.job_id).\\\n filter(Cluster.id==cluster_id).\\\n filter(Job.service_id==Service.id).\\\n order_by('job.name').\\\n all()\n\n# ##############################################################################\n\ndef basic_auth_list(session, cluster_id):\n \"\"\" All the HTTP Basic Auth definitions.\n \"\"\"\n return session.query(HTTPBasicAuth).\\\n filter(Cluster.id==cluster_id).\\\n order_by('http_basic_auth_def.name').\\\n all()\n\ndef tech_acc_list(session, cluster_id):\n \"\"\" All the technical accounts.\n \"\"\"\n return session.query(TechnicalAccount).\\\n order_by(TechnicalAccount.name).\\\n filter(Cluster.id==cluster_id).\\\n all()\n\ndef wss_list(session, cluster_id):\n \"\"\" All the WS-Security definitions.\n \"\"\"\n return session.query(WSSDefinition).\\\n filter(Cluster.id==cluster_id).\\\n order_by('wss_def.name').\\\n all()\n\n# ##############################################################################\n\ndef _def_amqp(session, cluster_id):\n return session.query(ConnDefAMQP.name, ConnDefAMQP.id, ConnDefAMQP.host,\n ConnDefAMQP.port, ConnDefAMQP.vhost, ConnDefAMQP.username,\n ConnDefAMQP.frame_max, ConnDefAMQP.heartbeat, ConnDefAMQP.password).\\\n filter(Cluster.id==ConnDefAMQP.cluster_id).\\\n filter(ConnDefAMQP.def_type=='amqp').\\\n filter(Cluster.id==cluster_id).\\\n order_by(ConnDefAMQP.name)\n\ndef def_amqp(session, cluster_id, id):\n \"\"\" A particular AMQP definition\n \"\"\"\n return _def_amqp(session, cluster_id).\\\n filter(ConnDefAMQP.id==id).\\\n one()\n\ndef def_amqp_list(session, cluster_id):\n \"\"\" AMQP connection definitions.\n \"\"\"\n return _def_amqp(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _def_jms_wmq(session, cluster_id):\n return session.query(ConnDefWMQ.id, ConnDefWMQ.name, ConnDefWMQ.host,\n ConnDefWMQ.port, ConnDefWMQ.queue_manager, ConnDefWMQ.channel,\n ConnDefWMQ.cache_open_send_queues, ConnDefWMQ.cache_open_receive_queues,\n ConnDefWMQ.use_shared_connections, ConnDefWMQ.ssl, ConnDefWMQ.ssl_cipher_spec,\n ConnDefWMQ.ssl_key_repository, ConnDefWMQ.needs_mcd, ConnDefWMQ.max_chars_printed).\\\n filter(Cluster.id==ConnDefWMQ.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(ConnDefWMQ.name)\n\ndef def_jms_wmq(session, cluster_id, id):\n \"\"\" A particular JMS WebSphere MQ definition\n \"\"\"\n return _def_jms_wmq(session, cluster_id).\\\n filter(ConnDefWMQ.id==id).\\\n one()\n\ndef def_jms_wmq_list(session, cluster_id):\n \"\"\" JMS WebSphere MQ connection definitions.\n \"\"\"\n return _def_jms_wmq(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _out_amqp(session, cluster_id):\n return session.query(OutgoingAMQP.id, OutgoingAMQP.name, OutgoingAMQP.is_active,\n OutgoingAMQP.delivery_mode, OutgoingAMQP.priority, OutgoingAMQP.content_type,\n OutgoingAMQP.content_encoding, OutgoingAMQP.expiration, OutgoingAMQP.user_id,\n OutgoingAMQP.app_id, ConnDefAMQP.name.label('def_name'), OutgoingAMQP.def_id).\\\n filter(OutgoingAMQP.def_id==ConnDefAMQP.id).\\\n filter(ConnDefAMQP.id==OutgoingAMQP.def_id).\\\n filter(Cluster.id==ConnDefAMQP.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(OutgoingAMQP.name)\n\ndef out_amqp(session, cluster_id, id):\n \"\"\" An outgoing AMQP connection.\n \"\"\"\n return _out_amqp(session, cluster_id).\\\n filter(OutgoingAMQP.id==id).\\\n one()\n\ndef out_amqp_list(session, cluster_id):\n \"\"\" Outgoing AMQP connections.\n \"\"\"\n return _out_amqp(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _out_jms_wmq(session, cluster_id):\n return session.query(OutgoingWMQ.id, OutgoingWMQ.name, OutgoingWMQ.is_active,\n OutgoingWMQ.delivery_mode, OutgoingWMQ.priority, OutgoingWMQ.expiration,\n ConnDefWMQ.name.label('def_name'), OutgoingWMQ.def_id).\\\n filter(OutgoingWMQ.def_id==ConnDefWMQ.id).\\\n filter(ConnDefWMQ.id==OutgoingWMQ.def_id).\\\n filter(Cluster.id==ConnDefWMQ.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(OutgoingWMQ.name)\n\ndef out_jms_wmq(session, cluster_id, id):\n \"\"\" An outgoing JMS WebSphere MQ connection.\n \"\"\"\n return _out_jms_wmq(session, cluster_id).\\\n filter(OutgoingWMQ.id==id).\\\n one()\n\ndef out_jms_wmq_list(session, cluster_id):\n \"\"\" Outgoing JMS WebSphere MQ connections.\n \"\"\"\n return _out_jms_wmq(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _channel_amqp(session, cluster_id):\n return session.query(ChannelAMQP.id, ChannelAMQP.name, ChannelAMQP.is_active,\n ChannelAMQP.queue, ChannelAMQP.consumer_tag_prefix,\n ConnDefAMQP.name.label('def_name'), ChannelAMQP.def_id,\n Service.name.label('service_name')).\\\n filter(ChannelAMQP.def_id==ConnDefAMQP.id).\\\n filter(ChannelAMQP.service_id==Service.id).\\\n filter(Cluster.id==ConnDefAMQP.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(ChannelAMQP.name)\n\ndef channel_amqp(session, cluster_id, id):\n \"\"\" A particular AMQP channel.\n \"\"\"\n return _channel_amqp(session, cluster_id).\\\n filter(ChannelAMQP.id==id).\\\n one()\n\ndef channel_amqp_list(session, cluster_id):\n \"\"\" AMQP channels.\n \"\"\"\n return _channel_amqp(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _channel_jms_wmq(session, cluster_id):\n return session.query(ChannelWMQ.id, ChannelWMQ.name, ChannelWMQ.is_active,\n ChannelWMQ.queue, ConnDefWMQ.name.label('def_name'), ChannelWMQ.def_id,\n Service.name.label('service_name')).\\\n filter(ChannelWMQ.def_id==ConnDefWMQ.id).\\\n filter(ChannelWMQ.service_id==Service.id).\\\n filter(Cluster.id==ConnDefWMQ.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(ChannelWMQ.name)\n\ndef channel_jms_wmq(session, cluster_id, id):\n \"\"\" A particular JMS WebSphere MQ channel.\n \"\"\"\n return _channel_jms_wmq(session, cluster_id).\\\n filter(ChannelWMQ.id==id).\\\n one()\n\ndef channel_jms_wmq_list(session, cluster_id):\n \"\"\" JMS WebSphere MQ channels.\n \"\"\"\n return _channel_jms_wmq(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _out_zmq(session, cluster_id):\n return session.query(OutgoingZMQ.id, OutgoingZMQ.name, OutgoingZMQ.is_active,\n OutgoingZMQ.address, OutgoingZMQ.socket_type).\\\n filter(Cluster.id==OutgoingZMQ.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(OutgoingZMQ.name)\n\ndef out_zmq(session, cluster_id, id):\n \"\"\" An outgoing ZeroMQ connection.\n \"\"\"\n return _out_zmq(session, cluster_id).\\\n filter(OutgoingZMQ.id==id).\\\n one()\n\ndef out_zmq_list(session, cluster_id):\n \"\"\" Outgoing ZeroMQ connections.\n \"\"\"\n return _out_zmq(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _channel_zmq(session, cluster_id):\n return session.query(ChannelZMQ.id, ChannelZMQ.name, ChannelZMQ.is_active,\n ChannelZMQ.address, ChannelZMQ.socket_type, ChannelZMQ.sub_key, \n Service.name.label('service_name')).\\\n filter(Cluster.id==ChannelZMQ.cluster_id).\\\n filter(Service.id==ChannelZMQ.service_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(ChannelZMQ.name)\n\ndef channel_zmq(session, cluster_id, id):\n \"\"\" An incoming ZeroMQ connection.\n \"\"\"\n return _channel_zmq(session, cluster_id).\\\n filter(ChannelZMQ.id==id).\\\n one()\n\ndef channel_zmq_list(session, cluster_id):\n \"\"\" Incoming ZeroMQ connections.\n \"\"\"\n return _channel_zmq(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _http_soap(session, cluster_id):\n\n tech_acc_case_id = (SecurityDefinition.security_def_type=='tech_acc', \n literal_column('(select tech_account.id from tech_account where tech_account.security_def_id = security_def.id)'))\n tech_acc_case_name = (SecurityDefinition.security_def_type=='tech_acc', \n literal_column('(select tech_account.name from tech_account where tech_account.security_def_id = security_def.id)'))\n \n wss_case_id = (SecurityDefinition.security_def_type=='wss_username_password', \n literal_column('(select wss_def.id from wss_def where wss_def.security_def_id = security_def.id)'))\n wss_case_name = (SecurityDefinition.security_def_type=='wss_username_password', \n literal_column('(select wss_def.name from wss_def where wss_def.security_def_id = security_def.id)'))\n \n basic_auth_case_id = (SecurityDefinition.security_def_type=='basic_auth', \n literal_column('(select http_basic_auth_def.id from http_basic_auth_def where http_basic_auth_def.security_def_id = security_def.id)'))\n basic_auth_case_name = (SecurityDefinition.security_def_type=='basic_auth', \n literal_column('(select http_basic_auth_def.name from http_basic_auth_def where http_basic_auth_def.security_def_id = security_def.id)'))\n \n return session.query(HTTPSOAP.id, HTTPSOAP.name, HTTPSOAP.is_active, \n HTTPSOAP.is_internal, HTTPSOAP.transport, HTTPSOAP.url_path, \n HTTPSOAP.method, HTTPSOAP.soap_action, HTTPSOAP.soap_version, \n Service.id.label('service_id'),\n Service.name.label('service_name'),\n Service.impl_name,\n SecurityDefinition.id.label('security_def_id'),\n SecurityDefinition.security_def_type,\n case([tech_acc_case_id, wss_case_id, basic_auth_case_id]).label('security_id'),\n case([tech_acc_case_name, wss_case_name, basic_auth_case_name]).label('security_name'),\n ).\\\n outerjoin(HTTPSOAPSecurity, HTTPSOAPSecurity.http_soap_id==HTTPSOAP.id).\\\n outerjoin(SecurityDefinition, HTTPSOAPSecurity.security_def_id==SecurityDefinition.id).\\\n filter(Cluster.id==HTTPSOAP.cluster_id).\\\n filter(Service.id==HTTPSOAP.service_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(HTTPSOAP.name)\n\n# No point in creating a new function if we can aliast an already existing one.\nhttp_soap_security_list = _http_soap\n\ndef http_soap(session, cluster_id, id):\n \"\"\" An HTTP/SOAP connection.\n \"\"\"\n return _http_soap(session, cluster_id).\\\n filter(HTTPSOAP.id==id).\\\n one()\n\ndef http_soap_list(session, cluster_id, connection=None, transport=None):\n \"\"\" HTTP/SOAP connections.\n \"\"\"\n q = _http_soap(session, cluster_id)\n \n if connection:\n q = q.filter(HTTPSOAP.connection==connection)\n \n if transport:\n q = q.filter(HTTPSOAP.transport==transport)\n \n return q.all()\n\n# ##############################################################################\n\ndef _out_s3(session, cluster_id):\n return session.query(OutgoingS3.id, OutgoingS3.name, OutgoingS3.is_active,\n OutgoingS3.prefix, OutgoingS3.separator,\n OutgoingS3.key_sync_timeout).\\\n filter(Cluster.id==OutgoingS3.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(OutgoingS3.name)\n\ndef out_s3(session, cluster_id, id):\n \"\"\" An outgoing S3 connection.\n \"\"\"\n return _out_s3(session, cluster_id).\\\n filter(OutgoingS3.id==id).\\\n one()\n\ndef out_s3_list(session, cluster_id):\n \"\"\" Outgoing S3 connections.\n \"\"\"\n return _out_s3(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _out_ftp(session, cluster_id):\n return session.query(OutgoingFTP.id, OutgoingFTP.name, OutgoingFTP.is_active,\n OutgoingFTP.host, OutgoingFTP.port, OutgoingFTP.user, OutgoingFTP.password, \n OutgoingFTP.acct, OutgoingFTP.timeout, OutgoingFTP.dircache).\\\n filter(Cluster.id==OutgoingFTP.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(OutgoingFTP.name)\n\ndef out_ftp(session, cluster_id, id):\n \"\"\" An outgoing FTP connection.\n \"\"\"\n return _out_ftp(session, cluster_id).\\\n filter(OutgoingFTP.id==id).\\\n one()\n\ndef out_ftp_list(session, cluster_id):\n \"\"\" Outgoing FTP connections.\n \"\"\"\n return _out_ftp(session, cluster_id).all()\n\n# ##############################################################################\n\ndef _service(session, cluster_id):\n return session.query(Service.id, Service.name, Service.is_active,\n Service.impl_name, Service.is_internal).\\\n filter(Cluster.id==Service.cluster_id).\\\n filter(Cluster.id==cluster_id).\\\n order_by(Service.name)\n\ndef service(session, cluster_id, id):\n \"\"\" A service.\n \"\"\"\n return _service(session, cluster_id).\\\n filter(Service.id==id).\\\n one()\n\ndef service_list(session, cluster_id):\n \"\"\" All services.\n \"\"\"\n return _service(session, cluster_id).all()\n\n# ##############################################################################\n","sub_path":"code/zato-common/src/zato/common/odb/query.py","file_name":"query.py","file_ext":"py","file_size_in_byte":15841,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"241840598","text":"# -*- coding: utf-8 -*-\n\nfrom __future__ import absolute_import\n\nimport inspect\nimport warnings\nfrom functools import wraps\n\nfrom flask import abort, jsonify, request\nfrom marshmallow.exceptions import ValidationError\nfrom six import string_types\nfrom werkzeug.wrappers import Response as WerkzeugResponse\n\nfrom eduid_common.api.messages import FluxData, error_response\nfrom eduid_common.api.schemas.models import FluxFailResponse, FluxResponseStatus, FluxSuccessResponse\nfrom eduid_common.api.utils import get_user\nfrom eduid_common.session import session\n\n__author__ = 'lundberg'\n\n\ndef require_eppn(f):\n @wraps(f)\n def require_eppn_decorator(*args, **kwargs):\n eppn = session.get('user_eppn', None)\n # If the user is logged in and has a session\n # pass on the request to the decorated view\n # together with the eppn of the logged in user.\n if eppn:\n kwargs['eppn'] = eppn\n return f(*args, **kwargs)\n abort(401)\n\n return require_eppn_decorator\n\n\ndef require_user(f):\n @wraps(f)\n def require_user_decorator(*args, **kwargs):\n user = get_user()\n kwargs['user'] = user\n return f(*args, **kwargs)\n\n return require_user_decorator\n\n\ndef can_verify_identity(f):\n @wraps(f)\n def verify_identity_decorator(*args, **kwargs):\n user = get_user()\n # For now a user can just have one verified NIN\n if user.nins.primary is not None:\n # TODO: Make this a CommonMsg I guess\n return error_response(message='User is already verified')\n # A user can not verify a nin if another previously was verified\n locked_nin = user.locked_identity.find('nin')\n if locked_nin and locked_nin.number != kwargs['nin']:\n # TODO: Make this a CommonMsg I guess\n return error_response(message='Another nin is already registered for this user')\n\n return f(*args, **kwargs)\n\n return verify_identity_decorator\n\n\nclass MarshalWith(object):\n \"\"\"\n Decorator to format the data returned from a Flask view and ensure it conforms to a marshmallow schema.\n\n A common usage is to use this to format the response as a Flux Standard Action\n (https://github.com/redux-utilities/flux-standard-action) by using a schema that has FluxStandardAction\n as superclass, or as a mixin.\n\n See the documentation of the FluxResponse class, or the link above, for more information about the\n on-the-wire format of these Flux Standard Actions.\n \"\"\"\n\n def __init__(self, schema):\n self.schema = schema\n\n def __call__(self, f):\n @wraps(f)\n def marshal_decorator(*args, **kwargs):\n # Call the Flask view, which is expected to return a FluxData instance,\n # or in special cases an WerkzeugResponse (e.g. when a redirect is performed).\n ret = f(*args, **kwargs)\n\n if isinstance(ret, WerkzeugResponse):\n # No need to Marshal again, someone else already did that\n return ret\n\n if isinstance(ret, dict):\n # TODO: Backwards compatibility mode - work on removing the need for this\n ret = FluxData(FluxResponseStatus.OK, payload=ret)\n\n if not isinstance(ret, FluxData):\n raise TypeError('Data returned from Flask view was not a FluxData (or WerkzeugResponse) instance')\n\n if ret.status != FluxResponseStatus.OK:\n _flux_response = FluxFailResponse(request, payload=ret.payload)\n else:\n _flux_response = FluxSuccessResponse(request, payload=ret.payload)\n return jsonify(self.schema().dump(_flux_response.to_dict()))\n\n return marshal_decorator\n\n\nclass UnmarshalWith(object):\n def __init__(self, schema):\n self.schema = schema\n\n def __call__(self, f):\n @wraps(f)\n def unmarshal_decorator(*args, **kwargs):\n try:\n json_data = request.get_json()\n if json_data is None:\n json_data = {}\n unmarshal_result = self.schema().load(json_data)\n kwargs.update(unmarshal_result)\n return f(*args, **kwargs)\n except ValidationError as e:\n response_data = FluxFailResponse(\n request, payload={'error': e.normalized_messages(), 'csrf_token': session.get_csrf_token()}\n )\n return jsonify(response_data.to_dict())\n\n return unmarshal_decorator\n\n\n# https://stackoverflow.com/questions/2536307/how-do-i-deprecate-python-functions/40301488#40301488\ndef deprecated(reason):\n \"\"\"\n This is a decorator which can be used to mark functions\n as deprecated. It will result in a warning being emitted\n when the function is used.\n \"\"\"\n\n if isinstance(reason, string_types):\n\n # The @deprecated is used with a 'reason'.\n #\n # .. code-block:: python\n #\n # @deprecated(\"please, use another function\")\n # def old_function(x, y):\n # pass\n\n def decorator(func1):\n\n if inspect.isclass(func1):\n fmt1 = \"Call to deprecated class {name} ({reason}).\"\n else:\n fmt1 = \"Call to deprecated function {name} ({reason}).\"\n\n @wraps(func1)\n def new_func1(*args, **kwargs):\n warnings.simplefilter('always', DeprecationWarning)\n warnings.warn(\n fmt1.format(name=func1.__name__, reason=reason), category=DeprecationWarning, stacklevel=2\n )\n warnings.simplefilter('default', DeprecationWarning)\n return func1(*args, **kwargs)\n\n return new_func1\n\n return decorator\n\n elif inspect.isclass(reason) or inspect.isfunction(reason):\n\n # The @deprecated is used without any 'reason'.\n #\n # .. code-block:: python\n #\n # @deprecated\n # def old_function(x, y):\n # pass\n\n func2 = reason\n\n if inspect.isclass(func2):\n fmt2 = \"Call to deprecated class {name}.\"\n else:\n fmt2 = \"Call to deprecated function {name}.\"\n\n @wraps(func2)\n def new_func2(*args, **kwargs):\n warnings.simplefilter('always', DeprecationWarning)\n warnings.warn(fmt2.format(name=func2.__name__), category=DeprecationWarning, stacklevel=2)\n warnings.simplefilter('default', DeprecationWarning)\n return func2(*args, **kwargs)\n\n return new_func2\n\n else:\n raise TypeError(repr(type(reason)))\n\n\n@deprecated('Use eduid_common.api.decorators.deprecated instead')\nclass Deprecated(object):\n \"\"\"\n Mark deprecated functions with this decorator.\n\n Attention! Use it as the closest one to the function you decorate.\n\n :param message: The deprecation message\n :type message: str | unicode\n \"\"\"\n\n def __init__(self, message=None):\n self.message = message\n\n def __call__(self, func):\n if self.message is None:\n self.message = 'Deprecated function {!r} called'.format(func.__name__)\n\n @wraps(func)\n def new_func(*args, **kwargs):\n warnings.warn(self.message, category=DeprecationWarning, stacklevel=2)\n return func(*args, **kwargs)\n\n # work around a bug in functools.wraps thats fixed in python 3.2\n if getattr(new_func, '__wrapped__', None) is None:\n new_func.__wrapped__ = func\n return new_func\n","sub_path":"src/eduid_common/api/decorators.py","file_name":"decorators.py","file_ext":"py","file_size_in_byte":7549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"304752754","text":"from django.urls import include, path\nfrom rest_framework.routers import DefaultRouter\n\nfrom .views import (CategoryViewSet, CommentViewSet, GenreViewSet,\n ReviewViewSet, TitleViewSet)\n\nrouter = DefaultRouter()\nrouter.register(r'genres', GenreViewSet, 'Genre')\nrouter.register(r'categories', CategoryViewSet, 'Category')\nrouter.register(r'titles', TitleViewSet, 'Title')\nrouter.register(r'titles/(?P[0-9]+)/reviews', ReviewViewSet)\nrouter.register(\n r'titles/(?P[0-9]+)/reviews/(?P[0-9]+)/comments',\n CommentViewSet)\n\nurlpatterns = [\n path('v1/', include('users.urls')),\n path('v1/', include(router.urls)),\n]\n","sub_path":"api/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"467319553","text":"import pandas as pd\r\ndata=pd.read_csv(\"data.csv\")\r\n\r\n#####Splitting Dataset Into Train and Test##############\r\nfrom sklearn.cross_validation import StratifiedShuffleSplit\r\nsss = StratifiedShuffleSplit(data['Gender'], 3, test_size=0.5, random_state=0)\r\nfor train_index, test_index in sss:\r\n print(\"TRAIN:\", train_index, \"TEST:\", test_index)\r\n X_train, X_test = data[train_index], data[test_index]\r\n y_train, y_test = data[train_index], data[test_index]\r\n\r\n######Converting Categorical To Numerical##############\r\n#from sklearn.preprocessing import LabelEncoder\r\n#le = LabelEncoder()\r\n#var_mod = ['LastLetter','LastTwoLetter','FirstLetter']\r\n#for i in var_mod:\r\n# train[i] = le.fit_transform(train[i])\r\n# test[i] = le.fit_transform(test[i])\r\n#\r\n##Following Gaussian Distribution\r\n#train.loc['LastLetter']['Gender'==1].plot.density()\r\n#train['LastTwoLetter'].plot.density() \r\n#train['FirstLetter'].plot.density()\r\n#\r\n#\r\n#import numpy as np\r\n#from sklearn.naive_bayes import GaussianNB\r\n#model = GaussianNB()\r\n#\r\n######Training Part##############\r\n#numpyMatrix_train = train.as_matrix()\r\n#x=np.array(numpyMatrix_train[:,[2,3,4]]).astype('int')\r\n#y=np.array(numpyMatrix_train[:,1]).astype('int')\r\n#model.fit(x,y)\r\n#\r\n######Testing Part##############\r\n#numpyMatrix_test = test.as_matrix()\r\n#X_test=np.array(numpyMatrix_test[:,[2,3,4]]).astype('int')\r\n#Y_actual=np.array(numpyMatrix_test[:,1]).astype('int')\r\n#Y_test= model.predict(X_test)\r\n#\r\n######Confusion Matrix##############\r\n#from sklearn.metrics import confusion_matrix\r\n#cnf_matrix = confusion_matrix(Y_actual, Y_test)\r\n#print(cnf_matrix)\r\n#\r\n#from sklearn import tree\r\n#clf = tree.DecisionTreeClassifier()\r\n#clf = clf.fit(x, y)\r\n#Y_test_decisiontree=clf.predict(X_test)\r\n#cnf_matrix_decisontree = confusion_matrix(Y_actual, Y_test_decisiontree)\r\n#print(cnf_matrix_decisontree)\r\n","sub_path":"NaiveBayesianPython.py","file_name":"NaiveBayesianPython.py","file_ext":"py","file_size_in_byte":1847,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"268994660","text":"from Myemail import Myemail\nfrom Nifty50 import Nifty50\nfrom keyvaluedao import KeyvalueDAO\nimport sqlite3\n\nprint(\"Invoked the advanced decline\")\nnifty50 = Nifty50()\nkeyvaluedao = KeyvalueDAO()\nconn = sqlite3.connect(\"stock.db\")\nflag = keyvaluedao.getValue(conn, \"advance_notify\")\n\nif flag == \"true\":\n output = nifty50.getnifty50()\n advance = output[\"advances\"]\n decline = output[\"declines\"]\n\n subject = \"Advance: \" + str(advance) + \" Decline: \" + str(decline)\n\n counter = 0\n message = \"\"\n while counter < 50:\n message = message + output[\"data\"][counter][\"symbol\"] + \" \" + output[\"data\"][counter][\"per\"] + \"\\n\"\n counter = counter + 1\n #print(message)\n\n myemail = Myemail()\n myemail.send_email(\"aruna\", \"aruna\", \"report\", subject, message)\n\n\n\n","sub_path":"hourlyadde.py","file_name":"hourlyadde.py","file_ext":"py","file_size_in_byte":786,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"240970077","text":"import pytest\n\nfrom sdc.crypto.decrypter import decrypt\nfrom sdc.crypto.key_store import KeyStore\nfrom sdc.crypto.exceptions import InvalidTokenException\nfrom tests import TEST_DO_NOT_USE_UPSTREAM_PUBLIC_PEM, TEST_DO_NOT_USE_SR_PRIVATE_PEM\nfrom tests import TOO_FEW_TOKENS_JWE, VALID_JWE\n\n\nKEY_PURPOSE_AUTHENTICATION = \"authentication\"\n\n# jwt.io public key signed\nTEST_DO_NOT_USE_PUBLIC_KEY = \"\"\"-----BEGIN PUBLIC KEY-----\nMIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQDdlatRjRjogo3Wojg\nGHFHYLugdUWAY9iR3fy4arWNA1KoS8kVw33cJibXr8bvwUAUparCwlv\ndbH6dvEOfou0/gCFQsHUfQrSDv+MuSUMAe8jzKE4qW+jK+xQU9a03GU\nnKHkkle+Q0pX/g6jXZ7r1/xAK5Do2kQ+X5xK9cipRgEKwIDAQAB\n-----END PUBLIC KEY-----\"\"\"\n\n\nclass TestDecrypter:\n\n key_store = KeyStore({\n \"keys\": {\n \"e19091072f920cbf3ca9f436ceba309e7d814a62\": {'purpose': KEY_PURPOSE_AUTHENTICATION,\n 'type': 'private',\n 'value': TEST_DO_NOT_USE_SR_PRIVATE_PEM},\n \"EQ_USER_AUTHENTICATION_SR_PRIVATE_KEY\": {'purpose': KEY_PURPOSE_AUTHENTICATION,\n 'type': 'private',\n 'value': TEST_DO_NOT_USE_SR_PRIVATE_PEM},\n \"EDCRRM\": {'purpose': KEY_PURPOSE_AUTHENTICATION,\n 'type': 'public',\n 'value': TEST_DO_NOT_USE_PUBLIC_KEY},\n \"709eb42cfee5570058ce0711f730bfbb7d4c8ade\": {'purpose': KEY_PURPOSE_AUTHENTICATION,\n 'type': 'public',\n 'value': TEST_DO_NOT_USE_UPSTREAM_PUBLIC_PEM},\n }\n })\n\n def test_decrypt(self):\n json = decrypt(VALID_JWE, self.key_store, KEY_PURPOSE_AUTHENTICATION)\n assert json == {'user': 'jimmy', 'iat': 1498137519.135479, 'exp': 1.0000000000014982e+21}\n\n def test_decrypt_too_few_tokens_in_jwe(self):\n \"\"\"Tests an InvalidTokenException when the token isn't comprised of 5 parts, seperated by several '.' characters\"\"\"\n with pytest.raises(InvalidTokenException):\n decrypt(TOO_FEW_TOKENS_JWE, self.key_store, KEY_PURPOSE_AUTHENTICATION)\n","sub_path":"tests/test_decrypter.py","file_name":"test_decrypter.py","file_ext":"py","file_size_in_byte":2244,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"236797007","text":"import cv2\n\ncapt = cv2.VideoCapture(0)\n\nclassifer = cv2.CascadeClassifier(\"haarcascade_frontalface_default.xml\")\n\nwhile True:\n\n ret, img = capt.read()\n\n if ret:\n faces = classifer.detectMultiScale(img)\n \n for face in faces:\n x, y, w, h = face\n img = cv2.rectangle(img, (x,y), (x+w, y+h), (255, 0, 0), 4)\n \n cv2.imshow(\"face detection program\", img)\n\n key = cv2.waitKey(1)\n\n if key == ord(\"q\"):\n break;\n\ncapt.release()\ncv2.destroyAllWindows()","sub_path":"Desktop/Santushti/python_work/ml_face_detect.py","file_name":"ml_face_detect.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"479108559","text":"\"\"\"\n-------------------------------------------------------------------------------\nName: ShooterGame.py\nPurpose:\nA simple, fun 2d shooter game based on python and pygame\nAuthor: P. Andy\nCreated on: 06/16/2016\n------------------------------------------------------------------------------\n\"\"\"\n\n# Import necessary library\nimport pygame\nfrom pygame.locals import *\nimport random\nimport math\n\n### Initialize the game\n\n# Screen\npygame.init()\nwidth, height = 1200, 600\nscreen=pygame.display.set_mode((width, height))\n\n# Player movement\nkeys = [False, False, False, False]\nplayer_position = [600,300]\n\n# Bullet\naccuracy = [0,0]\nbullets = []\n\n# Enemy related\nenemy_timer = 100\nenemy_timer2 = 0\nenemies = [[1200,100]]\nhealth_value = 194\n\n### Load images\n\n# The character that the player will control\nplayer = pygame.image.load(\"image/hero.png\").convert_alpha()\nplayer = pygame.transform.scale(player, (100, 100))\n\n# The wooden tiles on the bakcground\nbackground = pygame.image.load(\"image/background.png\")\nbackground = pygame.transform.scale(background, (100, 100))\n\n# The girls that the player must protect\nobjective = pygame.image.load(\"image/girl.png\")\nobjective = pygame.transform.scale(objective, (90, 90))\n\n# The bullet that the character will shoot\nbullet = pygame.image.load(\"image/bullet.png\")\n\n# Enemies that need to be defeated\nslime = pygame.image.load(\"image/slime.png\")\nslime = pygame.transform.scale(slime, (50,50))\nslime1 = slime\n\n# For the health meter\nhealth_bar = pygame.image.load(\"image/healthbar.png\")\nhealth = pygame.image.load(\"image/health.png\")\n\n# Ending\nlose_message = pygame.image.load(\"image/lose_message.png\")\nlose_message = pygame.transform.scale(lose_message, (1200,600))\nwin_message = pygame.image.load(\"image/win_message.png\")\nwin_message = pygame.transform.scale(win_message, (1200,600))\n\n# All the audio\n\n# 3.1 - Load audio\nobjective_damage = pygame.mixer.Sound(\"audio/objective_damage.wav\")\nenemy_death = pygame.mixer.Sound(\"audio/enemy_death.wav\")\nbullet_fire = pygame.mixer.Sound(\"audio/bullet.wav\")\nobjective_damage.set_volume(0.05)\nenemy_death.set_volume(0.05)\nbullet_fire.set_volume(0.05)\npygame.mixer.music.load('audio/bgm.mp3')\npygame.mixer.music.play(-1, 0.0)\npygame.mixer.music.set_volume(0.35)\n\n# Conditions for lose and win\nrunning = 1\nexitcode = 0\n\n### Loop until game is completed\nwhile running:\n\n # To spawn enemies at constant pattern\n enemy_timer -= 1\n\n # Clear the screen\n screen.fill(0)\n\n # Draw the background and objective\n for x in range(width/background.get_width()+1):\n for y in range(height/background.get_height()+1):\n screen.blit(background,(x*100,y*100))\n\n # Draw the objective (girls)\n screen.blit(objective, (0,10))\n screen.blit(objective, (0,110))\n screen.blit(objective, (0,210))\n screen.blit(objective, (0,310))\n screen.blit(objective, (0,410))\n screen.blit(objective, (0,510))\n\n # Complicated math formula involving trigonometry\n # For rotation of the player based on the mouse cursor\n mouse_position = pygame.mouse.get_pos()\n angle = math.atan2(mouse_position[1] - (player_position[1] + 32), mouse_position[0] - (player_position[0] + 26))\n player_rotation = pygame.transform.rotate(player, 360 - angle * 57.29)\n player_position2 = (player_position[0] - player_rotation.get_rect().width / 2,\n player_position[1] - player_rotation.get_rect().height / 2)\n screen.blit(player_rotation, player_position2)\n\n # Code for player firing bullet\n for i in bullets:\n\n # Initialize and set up basic info\n index = 0\n velx = math.cos(i[0]) * 10\n vely = math.sin(i[0]) * 10\n i[1] += velx\n i[2] += vely\n\n # If out of boundary, delete it.\n if i[1] < -64 or i[1] > 1200 or i[2] <- 64 or i[2] > 600:\n bullets.pop(index)\n\n index+=1\n\n # Calculate bullet's rotation and draw it on the screen\n for projectile in bullets:\n bullet2 = pygame.transform.rotate(bullet, 360 - projectile[0] * 57.29)\n screen.blit(bullet2, (projectile[1], projectile[2]))\n\n # Draw the enemy slime\n\n if enemy_timer == 0:\n enemies.append([1200, random.randint(50, 550)])\n enemy_timer = 100 - (enemy_timer2 * 2)\n\n if enemy_timer2 >= 35:\n enemy_timer2 = 35\n\n else:\n enemy_timer2 += 5\n\n index = 0\n\n for enemy in enemies:\n\n if enemy[0] < -64:\n enemies.pop(index)\n enemy[0] -= 5\n\n # Allow the enemies to attack the objective\n enemy_rect = pygame.Rect(slime1.get_rect())\n enemy_rect.top = enemy[1]\n enemy_rect.left = enemy[0]\n\n if enemy_rect.left < 64:\n objective_damage.play()\n health_value -= random.randint(5, 20)\n enemies.pop(index)\n\n # Check if bullet collides with slime\n index1 = 0\n\n for i in bullets:\n\n # Set up rectangles to check collision\n bullet_rect = pygame.Rect(bullet.get_rect())\n bullet_rect.left = i[1]\n bullet_rect.top = i[2]\n\n # If it does, eliminate both the bullet and the enemy\n if enemy_rect.colliderect(bullet_rect):\n enemy_death.play()\n accuracy[0] += 1\n enemies.pop(index)\n bullets.pop(index1)\n\n index1 += 1\n\n # Next slime\n index += 1\n\n for enemy in enemies:\n screen.blit(slime1, enemy)\n\n # Draw the time display which indicates how long the game will last\n font = pygame.font.Font(None, 46)\n survivedtext = font.render(str((30000 - pygame.time.get_ticks()) / 60000)+ \":\" + str((30000 - pygame.time.get_ticks()) / 1000 % 60).zfill(2), True, (0,0,0))\n textRect = survivedtext.get_rect()\n textRect.topright = [1195, 5]\n screen.blit(survivedtext, textRect)\n\n # Draw the health meter, displaying amount of health left\n screen.blit(health_bar, (100,10))\n for health1 in range(health_value):\n screen.blit(health, (health1+103,13))\n\n # Update the screen\n pygame.display.flip()\n\n # Main loop for the event codes\n for event in pygame.event.get():\n\n # Allow the player to end the game at their convenience\n if event.type==pygame.QUIT:\n pygame.quit()\n exit(0)\n\n # Change the status to true when the keys are pressed\n elif event.type == pygame.KEYDOWN:\n if event.key == K_w:\n keys[0] = True\n elif event.key == K_a:\n keys[1] = True\n elif event.key == K_s:\n keys[2] = True\n elif event.key == K_d:\n keys[3] = True\n\n # Revert when the keys are lifted\n elif event.type == pygame.KEYUP:\n if event.key == pygame.K_w:\n keys[0] = False\n elif event.key == pygame.K_a:\n keys[1] = False\n elif event.key==pygame.K_s:\n keys[2] = False\n elif event.key == pygame.K_d:\n keys[3] = False\n\n # Complicated math formula for bullet and its rotation\n elif event.type == pygame.MOUSEBUTTONDOWN:\n bullet_fire.play()\n position = pygame.mouse.get_pos()\n accuracy[1] += 1\n bullets.append([math.atan2(position[1] - (player_position2[1] + 32), position[0]\n - (player_position2[0] + 26)),player_position2[0] + 32, player_position2[1] + 32])\n\n # While status is true, move the player (direction depends on which key)\n if keys[0]:\n player_position[1] -= 5\n elif keys[2]:\n player_position[1] += 5\n elif keys[1]:\n player_position[0] -= 5\n elif keys[3]:\n player_position[0] += 5\n\n # Check whether player loses or wins\n\n # Win condition\n if pygame.time.get_ticks() >= 30000:\n running=0\n exitcode=1\n\n # Lose condition\n elif health_value <= 0:\n running=0\n exitcode=0\n\n # Calculate accuracy percentage (given that player landed at least one shot)\n elif accuracy[1] != 0:\n accuracy1 = accuracy[0] * 1.0 / accuracy[1]*100\n\n else:\n accuracy1 = 0\n\n# Display ending based one whether player wins or loses\n\n# Player loses\nif exitcode == 0:\n pygame.font.init()\n font = pygame.font.Font(None, 46)\n text = font.render(\"Accuracy: \"+str(accuracy1)+\"%\", True, (255,0,0))\n textRect = text.get_rect()\n textRect.centerx = screen.get_rect().centerx\n textRect.centery = screen.get_rect().centery+24\n screen.blit(lose_message, (0,0))\n screen.blit(text, textRect)\n\n# Player wins\nelse:\n pygame.font.init()\n font = pygame.font.Font(None, 46)\n text = font.render(\"Accuracy: \"+str(accuracy1)+\"%\", True, (0,255,0))\n textRect = text.get_rect()\n textRect.centerx = screen.get_rect().centerx\n textRect.centery = screen.get_rect().centery+24\n screen.blit(win_message, (0,0))\n screen.blit(text, textRect)\n\nwhile 1:\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n pygame.quit()\n exit(0)\n pygame.display.flip()","sub_path":"ShooterGame.py","file_name":"ShooterGame.py","file_ext":"py","file_size_in_byte":9156,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"160649123","text":"class Node(object):\n def __init__(self, value):\n self.value = value\n self.left = None\n self.right = None\n\n\nclass BST(object):\n def __init__(self, root):\n self.root = Node(root)\n\n def insert(self, new_val):\n self.insert_helper(self.root, new_val)\n\n def insert_helper(self, current, new_val):\n if new_val < current.value:\n if current.left:\n self.insert_helper(current.left, new_val)\n else:\n current.left = Node(new_val)\n elif new_val > current.value:\n if current.right:\n self.insert_helper(current.right, new_val)\n else:\n current.right = Node(new_val)\n else:\n print(f\"{new_val} is already in the tree\")\n\n def search(self, find_val):\n return self.search_helper(self.root, find_val)\n\n def search_helper(self, current, find_val):\n if current:\n if current.value == find_val:\n return True\n elif current.value > find_val:\n return self.search_helper(current.left, find_val)\n else:\n return self.search_helper(current.right, find_val)\n\n def inorder_list(self, current, L):\n if current:\n L = self.inorder_list(current.left, L)\n L.append(current.value)\n L = self.inorder_list(current.right, L)\n return L\n\n def is_BST(self):\n L = self.inorder_list(self.root, [])\n while len(L) >1 and L[-1] > L[-2]:\n L.pop()\n return len(L) <2\n \nbst = BST(8)\nbst.insert(3)\nbst.insert(10)\nbst.insert(1)\nbst.insert(6)\nbst.insert(9)\nbst.insert(11)\n\nprint(bst.is_BST())","sub_path":"Binary Trees/binary_search_tree.py","file_name":"binary_search_tree.py","file_ext":"py","file_size_in_byte":1699,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"157633599","text":"#--------------------------------------------------------------------------------\n# Just a mini script to review statistical functions & implementing them on Python\n# \n# Currently supported functions:\n# - mean\n# - median\n# - variance\n# - standard deviation\n# - range\n# - 1st Quartile / 25th Percentile / 0.25 Quantile\n# - 3rd Quartile / 75th Percentile / 0.75 Quantile\n# - IQR\n#--------------------------------------------------------------------------------\nimport math\n\n# Function already exists in Python\n# sum()\n# max()\n# min()\n# len()\n# sorted()\n\n# returns the average value of all elements in arr\ndef mean( arr ):\n\treturn sum(arr) / len(arr)\n\n# if array length == EVEN, return avg of the 2 middle values\n# if array length == ODD, return middle value\ndef median( arr ):\n\tarr = sorted(arr)\n\tlength = len(arr)\n\n\tif not(length > 0):\n\t\tprint( \"Array must at least have a length > 0\")\n\t\treturn\n\n\t# if number is even\n\tif( length % 2 == 0):\n\t\t# return avg of the 2 middle values\n\t\treturn ( arr[ int(length/2) ] + arr[ int((length/2)-1) ] ) / 2\n\t# else number is odd\n\telse:\n\t\t# returns the middle value\n\t\treturn arr[ int(length/2) ]\n\n# return variance of all elements in arr\ndef variance( arr ):\n\tmeanVal = mean(arr)\t\t\t# get mean values of array\n\tvar_sum = 0\t\t\t\t\t# sum of (difference to the mean, squared)\n\n\t# Calculating summation term in variance formula\n\tfor i in range( len(arr) ):\n\t\ttemp = arr[i] - meanVal\t\t\t\n\t\tvar_sum += math.pow( temp, 2 )\n\n\treturn var_sum / ( len( arr ) - 1 )\t# return variance = var_sum / (N-1)\n\n# return standard deviation of all elements in arr\ndef sd( arr ):\n\treturn math.sqrt( variance( arr ) )\n\n# return the (min,max) of arr\ndef Range( arr ):\n\treturn max(arr)-min(arr)\n\n#-------------------------------------------------------\n# No standard method to calc. quar_1(), quar_3() & iqr()\n# so I'll be using the method taught in class\n#-------------------------------------------------------\n# return the 1st quartile / 25th percentile / 0.25 quantile of the arr\n# i.e. returns the median of the lower half array\ndef quar_1(arr):\n\tarr = sorted(arr)\n\tlength = len(arr)\n\n\tif not(length > 0):\n\t\tprint( \"Array must at least have a length > 0\")\n\t\treturn\n\n\t# if number is even\n\tif( length % 2 == 0):\n\t\t# divide arr into 2 and return median of lower half\n\t\treturn median( arr[:int(length/2)] )\n\t# else number is odd\n\telse:\n\t\t# ignore the median, then divide arr into 2 and return median of lower half\n\t\treturn median( arr[:int((length-1)/2)] )\n\n# return the 3rd quartile / 75th percentile / 0.75 quantile of the arr\n# i.e. returns the median of the upper half array\ndef quar_3(arr):\n\tarr = sorted(arr)\n\tlength = len(arr)\n\n\tif not(length > 0):\n\t\tprint( \"Array must at least have a length > 0\")\n\t\treturn\n\n\t# if number is even\n\tif( length % 2 == 0):\n\t\t# divide arr into 2 and return median of upper half\n\t\treturn median( arr[int(length/2):] )\n\t# else number is odd\n\telse:\n\t\t# ignore the median, then divide arr into 2 and return median of upper half\n\t\treturn median( arr[int((length+1)/2):] )\n\n# return the IQR value of arr\ndef iqr( arr ):\n\treturn quar_3(arr) - quar_1(arr)\n\n#----------------------------------------------------------------------------------\ntestarr = [5,9,14,4,6,3,1]\n\nprint( \"INPUT ---> \\t\", testarr )\nprint( \"SORTED --> \\t\", sorted(testarr) )\nprint( \"COUNT(#) = \\t\", len(testarr) )\nprint( \"MAX = \\t\", max(testarr) ) \nprint( \"MIN = \\t\", min(testarr) ) \nprint( \"RANGE = \\t\", Range(testarr) )\nprint()\nprint( \"MEAN = \\t\", mean( testarr ) )\nprint( \"1st QUAR = \\t\", quar_1(testarr) ) \nprint( \"MEDIAN = \\t\", median( testarr ) )\nprint( \"3rd QUAR = \\t\", quar_3(testarr) ) \nprint( \"IQR = \\t\", iqr(testarr) ) \nprint( \"VARIANCE = \\t\", variance(testarr) )\nprint( \"SD = \\t\", sd(testarr) ) \n","sub_path":"STAT251/statsCalculator.py","file_name":"statsCalculator.py","file_ext":"py","file_size_in_byte":3740,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"443976316","text":"from app01 import models\nfrom concurrent.futures import ThreadPoolExecutor\nfrom django.shortcuts import render,HttpResponse\nfrom app01.utils.network_device import Epon\nfrom app01.utils.NDMS_settings import INT3\n\ndef epon_create_update(req):\n # objs_epon = models.Device.objects.filter(tag__name=\"EPON\")\n objs_epon = models.Device.objects.filter(tag__name=\"EPON\",tag__version__name='Version 5.20')\n print(\"开始进行EPON设备信息更新\")\n pool = ThreadPoolExecutor(INT3)\n for obj_epon in objs_epon:\n if obj_epon.tag.name == \"EPON\":\n # print(obj,obj.name,obj.IP)\n obj = Epon(obj_epon.IP)\n pool.submit(Epon.epon_check_update, obj)\n # 将子线程对象添加到列表中\n else:\n print(\"wrong\")\n pool.shutdown(wait=False)\n\n return HttpResponse(\"EPON设备信息更新正在运行中\")\n\n","sub_path":"about_NDMS/NDMS/app01/myviews/epon_info_update.py","file_name":"epon_info_update.py","file_ext":"py","file_size_in_byte":877,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"50642268","text":"from django.conf.urls import url\n\nfrom .views import (\n DashboardView,\n\n ProjectCreateView,\n ProjectDeleteView,\n ProjectUpdateView,\n ProjectView,\n\n TaskCreateView,\n TaskDeleteView,\n TaskDetailView,\n TaskDoneView,\n TaskUpdateView,\n TaskAssignTeamMemberUpdateView,\n\n TeamMemberCreateView,\n TeamMemberDetailView,\n TeamMemberUpdateView,\n TeamMemberDeleteView,\n)\n\nurlpatterns = (\n url(r'^$', DashboardView.as_view(), name='dashboard'),\n url(r'^project/create/$', ProjectCreateView.as_view(),\n name='project_create'),\n url(r'^project/(?P[0-9]+)$', ProjectView.as_view(), name='project'),\n url(r'^project/(?P[0-9]+)/update$',\n ProjectUpdateView.as_view(), name='project_update'),\n url(r'^project/(?P[0-9]+)/delete$',\n ProjectDeleteView.as_view(), name='project_delete'),\n\n url(r'^project/(?P[0-9]+)/task_create',\n TaskCreateView.as_view(), name='task_create'),\n url(r'^task/(?P[0-9]+)$',\n TaskDetailView.as_view(), name='task'),\n url(r'^task/(?P[0-9]+)/update$',\n TaskUpdateView.as_view(), name='task_update'),\n url(r'^task/(?P[0-9]+)/update_team_members$',\n TaskAssignTeamMemberUpdateView.as_view(),\n name='task_update_team_members'),\n url(r'^task/(?P[0-9]+)/delete$',\n TaskDeleteView.as_view(), name='task_delete'),\n url(r'^task/(?P[0-9]+)/done$',\n TaskDoneView.as_view(), name='task_done'),\n\n url(r'^teammember/create$',\n TeamMemberCreateView.as_view(), name='team_member_create'),\n url(r'^teammember/(?P[0-9]+)$',\n TeamMemberDetailView.as_view(), name='team_member'),\n url(r'^teammember/(?P[0-9]+)/update$',\n TeamMemberUpdateView.as_view(), name='team_member_update'),\n url(r'^teammember/(?P[0-9]+)/delete$',\n TeamMemberDeleteView.as_view(), name='team_member_delete'),\n\n)\n","sub_path":"remindmyteam.com/dashboard/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1911,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"356450341","text":"#!/usr/bin/env python\n\nclass FenwickTree():\n \"\"\"\n Data structure for providing O(lg(n)) single element alteration\n and range sum query. All indices are zero based.\n \"\"\"\n def __init__(self, num_elements):\n self.num_elements = 4 * num_elements + 1\n self.array = [0] * self.num_elements\n\n @staticmethod\n def _get_fenwick_step_length(num):\n return num&-num\n\n def alter(self, index, amount):\n \"\"\"\n Alter the element at index and add amount to it\n \"\"\"\n while index < self.num_elements:\n self.array[index] += amount\n index += self._get_fenwick_step_length(index)\n\n def query(self, from_index, to_index):\n \"\"\"\n Sum of the elements between from_index and to_index, both inclusive\n \"\"\"\n return self._query(to_index) - self._query(from_index - 1)\n\n def _query(self, index):\n result = 0\n while index > 0:\n result += self.array[index]\n index -= self._get_fenwick_step_length(index)\n return result\n\n\nif __name__ == \"__main__\":\n fw = FenwickTree(10)\n fw.alter(1, 1)\n fw.alter(2, 2)\n fw.alter(3, 3)\n fw.alter(4, 4)\n \n for i in range(1, 4 + 1):\n for j in range(i, 4 + 1):\n expected = 0\n for k in range(i, j + 1):\n expected += k\n assert expected == fw.query(i, j)\n","sub_path":"Data Structures/Fenwick Tree/Python/FenwickTree.py","file_name":"FenwickTree.py","file_ext":"py","file_size_in_byte":1394,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"27870412","text":"#!/usr/bin/env python\nfrom vtkmodules.vtkCommonCore import vtkMath\nfrom vtkmodules.vtkCommonDataModel import vtkSphere\nfrom vtkmodules.vtkCommonSystem import vtkTimerLog\nfrom vtkmodules.vtkFiltersCore import (\n vtkHedgeHog,\n vtkMaskPoints,\n)\nfrom vtkmodules.vtkFiltersModeling import vtkOutlineFilter\nfrom vtkmodules.vtkFiltersPoints import (\n vtkBoundedPointSource,\n vtkFitImplicitFunction,\n vtkPCANormalEstimation,\n)\nfrom vtkmodules.vtkRenderingCore import (\n vtkActor,\n vtkPointGaussianMapper,\n vtkPolyDataMapper,\n vtkRenderWindow,\n vtkRenderWindowInteractor,\n vtkRenderer,\n)\nimport vtkmodules.vtkInteractionStyle\nimport vtkmodules.vtkRenderingFreeType\nimport vtkmodules.vtkRenderingOpenGL2\nfrom vtkmodules.util.misc import vtkGetDataRoot\nVTK_DATA_ROOT = vtkGetDataRoot()\n\n# Interpolate onto a volume\n\n# Parameters for debugging\nNPts = 1000000\nmath = vtkMath()\nmath.RandomSeed(31415)\n\n# create pipeline\n#\npoints = vtkBoundedPointSource()\npoints.SetNumberOfPoints(NPts)\npoints.ProduceRandomScalarsOn()\npoints.ProduceCellOutputOff()\npoints.Update()\n\n# Create a sphere implicit function\nsphere = vtkSphere()\nsphere.SetCenter(0,0,0)\nsphere.SetRadius(0.75)\n\n# Extract points along sphere surface\nextract = vtkFitImplicitFunction()\nextract.SetInputConnection(points.GetOutputPort())\nextract.SetImplicitFunction(sphere)\nextract.SetThreshold(0.005)\nextract.Update()\n\n# Now generate normals from resulting points\nnorms = vtkPCANormalEstimation()\nnorms.SetInputConnection(extract.GetOutputPort())\nnorms.SetSampleSize(20)\nnorms.FlipNormalsOn()\nnorms.SetNormalOrientationToGraphTraversal()\n\n# Time execution\ntimer = vtkTimerLog()\ntimer.StartTimer()\nnorms.Update()\ntimer.StopTimer()\ntime = timer.GetElapsedTime()\nprint(\"Points processed: {0}\".format(NPts))\nprint(\" Time to generate normals: {0}\".format(time))\n#print(hBin)\n#print(hBin.GetOutput())\n\nsubMapper = vtkPointGaussianMapper()\nsubMapper.SetInputConnection(norms.GetOutputPort())\nsubMapper.EmissiveOff()\nsubMapper.SetScaleFactor(0.0)\n\nsubActor = vtkActor()\nsubActor.SetMapper(subMapper)\n\n# Draw the normals\nmask = vtkMaskPoints()\nmask.SetInputConnection(norms.GetOutputPort())\nmask.SetRandomModeType(1)\nmask.SetMaximumNumberOfPoints(250)\n\nhhog = vtkHedgeHog()\nhhog.SetInputConnection(mask.GetOutputPort())\nhhog.SetVectorModeToUseNormal()\nhhog.SetScaleFactor(0.25)\n\nhogMapper = vtkPolyDataMapper()\nhogMapper.SetInputConnection(hhog.GetOutputPort())\n\nhogActor = vtkActor()\nhogActor.SetMapper(hogMapper)\n\n# Create an outline\noutline = vtkOutlineFilter()\noutline.SetInputConnection(points.GetOutputPort())\n\noutlineMapper = vtkPolyDataMapper()\noutlineMapper.SetInputConnection(outline.GetOutputPort())\n\noutlineActor = vtkActor()\noutlineActor.SetMapper(outlineMapper)\n\n# Create the RenderWindow, Renderer and both Actors\n#\nren0 = vtkRenderer()\nrenWin = vtkRenderWindow()\nrenWin.AddRenderer(ren0)\niren = vtkRenderWindowInteractor()\niren.SetRenderWindow(renWin)\n\n# Add the actors to the renderer, set the background and size\n#\nren0.AddActor(subActor)\nren0.AddActor(hogActor)\nren0.AddActor(outlineActor)\nren0.SetBackground(0.1, 0.2, 0.4)\n\nrenWin.SetSize(250,250)\n\ncam = ren0.GetActiveCamera()\ncam.SetFocalPoint(1,1,1)\ncam.SetPosition(0,0,0)\nren0.ResetCamera()\n\niren.Initialize()\n\n# render the image\n#\nrenWin.Render()\n\niren.Start()\n","sub_path":"Filters/Points/Testing/Python/TestPCANormalEstimation2.py","file_name":"TestPCANormalEstimation2.py","file_ext":"py","file_size_in_byte":3305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"47431900","text":"from random import seed\r\nfrom random import random\r\nimport numpy\r\nimport time\r\n\r\ncolors = ['Red', 'Blue', 'Green', 'Yellow', 'Black']\r\n\r\nstates = ['1', '2', '3', '4']\r\n\r\nneighbors = {}\r\nneighbors['1'] = ['2', '3']\r\nneighbors['2'] = ['1', '3', '4']\r\nneighbors['3'] = ['1', '2', '4']\r\nneighbors['4'] = ['2','3']\r\n\r\ncolors_of_states = {}\r\n\r\n#color_and_fitness=numpy.empty((range(states),2))\r\n\r\nfitness_and_state = {}\r\nfitness = {}\r\n\r\n #code sampled from a java implementation of ABC\r\n #control parameters of ABC algorithm\r\nnp = 20 #len(states) # 20 # the number of the colony size (employed bees + onlooker bees\r\n # might need to have this to be changed based on the nodes that\r\n # are given into the problem\r\nfoodNumber = np/2 # the number of food sources, it equals half of\r\n # the colony size\r\nlimit = 100 # if a food source can't be improved by this limit we abandon\r\nmaxCycle = 2500 # the number of cycles the bee can make\r\n\r\n#problem specific variables\r\nD = 100 # the number of parameters of the problem to be optimized\r\nlb = -5.12 #lower bound of the parameters\r\nup = 5.12 # upper bound of the parameters\r\n\r\nruntime = 30\r\n\r\n\r\n#fitness ={} #numpy.empty(foodNumber,1) #vector that holds values of fitness associated with food source \"states\"\r\nprob = {} #numpy.empty(foodNumber,1) #vector that holds the probabilities of a food sources to be chosen\r\n\r\nclass beeColony():\r\n \r\n\r\n def __init__(self, states):\r\n self.states=states\r\n #for state in self.states:\r\n #self.set(state)\r\n \r\n def __str__(self):\r\n \r\n for i in range(len(self.states)):\r\n \r\n #print(\"I am at \", self.position[i], \" meu pbest is \", self.pbest_position[i]) \r\n colors_of_states[self.states[i]] = self.get_color_for_state(self.states[i])\r\n\r\n print (colors_of_states)\r\n \r\n def set2(self,state):\r\n for states in state:\r\n \r\n d=self.calculateFitness(state) \r\n #fitness[state]= d\r\n self.calculateFitness(state)\r\n #fitness_and_state.append([d, state])\r\n \r\n\r\n def calculateFitness(self,state): # im finding the fitness of a node based on \r\n # the Degree of it so how many neighbours\r\n h = 0\r\n for nieghbor in neighbors: # this calculates the fitness for \r\n h+1 # the problem im tryna solve\r\n # numpy.append(color_and_fitness, [state, i]) \r\n return h\r\n \r\n def sortSecond(self,val): #code from https://www.geeksforgeeks.org/python-list-sort/\r\n return val[1] # used to help sort an arrayed list\r\n \r\n # def memorizeBestSource(self): # this should hold the highest number at all time\r\n # or it should hold from the node the least amount\r\n # colors it hasn't taken to colour from this point\r\n \r\n #fitness.sort(reverse =True) # s\r\n #fitness_and_state.sort(self,key =sortSecond, reverse =True) #orders the list from most nodes to least \r\n \r\n\r\n def sendEmployedBees(self,state): # generate a random number from 0 to the # of states\r\n\r\n colors_of_states[state] = colors_of_states.get(state)\r\n fitness[state] = 0 # making the fitness now 0 so that if it selected in\r\n # the future it will already have been assigned a color\r\n\r\n\r\n def calculateProbilities(self): # find the probility that this state chosen is\r\n # better then the ones before or not\r\n # or can have this assign the colours\r\n maxfit = fitness[0]\r\n for state in states:\r\n if (fitness[state]>maxfit):\r\n maxfit = fitness[i]\r\n\r\n # for state in states:\r\n # prob[state]=(0.9*(fitness[state]/maxfit))+0.1\r\n #equation taken from a java code, using it make the probabilities\r\n # probability is calculated using fitness \r\n\r\n def sendOnlookerBees(self,state):\r\n # def get_color_for_state(state):\r\n for color in colors:\r\n if self.promising(state, color):\r\n return colors\r\n\r\n #def sendScoutBees(self):\r\n # null\r\n \r\n def get_color_for_state(self,state):\r\n for color in colors:\r\n if self.promising(state, color):\r\n return color\r\n \r\n def promising(self,state, color):\r\n print(state)\r\n for states2 in states:\r\n for neighbor in neighbors.get(states2): \r\n color_of_neighbor = colors_of_states.get(neighbor)\r\n if color_of_neighbor == color:\r\n return False\r\n\r\n return True\r\n\r\n\r\n\r\n\r\ndef main():\r\n start=time.time()\r\n # for state in states:\r\n # colors_of_states[state] = get_color_for_state(state)\r\n bees =beeColony(states)\r\n #graph=bees.set2(bees.states)\r\n # bees.sendEmployedBees(states)\r\n #bees.calculateProbilities()\r\n bees.sendOnlookerBees(states)\r\n bees.__str__()\r\n #bees.memorizeBestSource()\r\n # bees.sendScoutBees()\r\n \r\n stop = time.time()\r\n duration = stop-start\r\n print (colors_of_states)\r\n print(\"\\n\",\"The time to solve the solution is: \"+str(duration))\r\n \r\n\r\n\r\n\r\nmain() \r\n\r\n\r\n\r\n\r\nstates = ['1','2','3','4','5','6','7','8']\r\nneighbors = {}\r\nneighbors['1'] = ['2', '3', '5']\r\nneighbors['2'] = ['1', '6', '5']\r\nneighbors['3'] = ['1', '4', '5','6']\r\nneighbors['4'] = ['3', '6','8']\r\nneighbors['5'] = ['2', '3', '1','7']\r\nneighbors['6'] = ['2', '3', '4','8']\r\nneighbors['7'] = ['5', '8',]\r\nneighbors['8'] = ['7','6','4']\r\n\r\n\r\n\r\n\r\ndef main2():\r\n start=time.time()\r\n # for state in states:\r\n # colors_of_states[state] = get_color_for_state(state)\r\n bees =beeColony(states)\r\n #graph=bees.set2(bees.states)\r\n # bees.sendEmployedBees(states)\r\n #bees.calculateProbilities()\r\n bees.sendOnlookerBees(states)\r\n bees.__str__()\r\n #bees.memorizeBestSource()\r\n # bees.sendScoutBees()\r\n \r\n stop = time.time()\r\n duration = stop-start\r\n print (colors_of_states)\r\n print(\"\\n\",\"The time to solve the solution is: \"+str(duration))\r\n \r\n\r\n\r\n\r\nmain2()\r\n\r\n\r\n\r\n\r\nstates = ['1','2','3','4','5','6','7','8','9','10','11','12','13','14','15','16','17','18','19','20']\r\nneighbors = {}\r\nneighbors['1'] = ['6', '3', '2']\r\nneighbors['2'] = ['1', '4', '10']\r\nneighbors['3'] = ['1', '6', '4','5']\r\nneighbors['4'] = ['3', '2','16','8','5']\r\nneighbors['5'] = ['6','3','4','13','15','7']\r\nneighbors['6'] = ['1','3','5']\r\nneighbors['7'] = ['5', '9','19']\r\nneighbors['8'] = ['4','14','11','9','12']\r\nneighbors['9'] = ['8','12','7','18']\r\nneighbors['10'] = ['2','16']\r\nneighbors['11'] = ['14','8']\r\nneighbors['12'] = ['8','17','9']\r\nneighbors['13'] = ['5','15']\r\nneighbors['14'] = ['16','11','8']\r\nneighbors['15'] = ['13','5','20']\r\nneighbors['16'] = ['10','4','14']\r\nneighbors['17'] = ['12']\r\nneighbors['18'] = ['9','19']\r\nneighbors['19'] = ['20','7','18']\r\nneighbors['20'] = ['15','19']\r\n\r\n\r\n\r\n\r\n\r\ndef main3():\r\n start=time.time()\r\n # for state in states:\r\n # colors_of_states[state] = get_color_for_state(state)\r\n bees =beeColony(states)\r\n #graph=bees.set2(bees.states)\r\n # bees.sendEmployedBees(states)\r\n #bees.calculateProbilities()\r\n bees.sendOnlookerBees(states)\r\n bees.__str__()\r\n #bees.memorizeBestSource()\r\n # bees.sendScoutBees()\r\n \r\n stop = time.time()\r\n duration = stop-start\r\n print (colors_of_states)\r\n print(\"\\n\",\"The time to solve the solution is: \"+str(duration))\r\n \r\n\r\n\r\n\r\nmain3()\r\n\r\n","sub_path":"ABC2.py","file_name":"ABC2.py","file_ext":"py","file_size_in_byte":7668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"416308092","text":"# For data loading.\r\nfrom torchtext import data, datasets\r\nfrom data_process import ParallelDataset, Batch, subsequent_mask, batch_size_fn, read_corpus\r\nimport time\r\nimport torch\r\nfrom torch.autograd import Variable\r\nimport numpy as np\r\n\r\nfrom model import make_model\r\nfrom optim import LabelSmoothing, NoamOpt\r\nfrom loss import SimpleLossCompute\r\nimport argparse\r\n\r\n\r\ndef run_epoch(data_iter, model, loss_compute):\r\n \"Standard Training and Logging Function\"\r\n start = time.time()\r\n total_tokens = 0\r\n total_loss = 0\r\n tokens = 0\r\n for i, batch in enumerate(data_iter):\r\n out = model.forward(batch.src, batch.trg,\r\n batch.src_mask, batch.trg_mask)\r\n loss = loss_compute(out, batch.trg_y, batch.ntokens)\r\n total_loss += loss\r\n total_tokens += batch.ntokens\r\n tokens += batch.ntokens\r\n if i % opt.display_freq == 1:\r\n elapsed = time.time() - start\r\n print(\"Epoch Step: %d Loss: %f Tokens per Sec: %f\" %\r\n (i, loss / batch.ntokens, tokens / elapsed))\r\n start = time.time()\r\n tokens = 0\r\n return total_loss / total_tokens\r\n\r\n\r\nclass MyIterator(data.Iterator):\r\n def create_batches(self):\r\n if self.train:\r\n def pool(d, random_shuffler):\r\n for p in data.batch(d, self.batch_size * 100):\r\n p_batch = data.batch(\r\n sorted(p, key=self.sort_key),\r\n self.batch_size, self.batch_size_fn)\r\n for b in random_shuffler(list(p_batch)):\r\n yield b\r\n\r\n self.batches = pool(self.data(), self.random_shuffler)\r\n\r\n else:\r\n self.batches = []\r\n for b in data.batch(self.data(), self.batch_size,\r\n self.batch_size_fn):\r\n self.batches.append(sorted(b, key=self.sort_key))\r\n\r\n\r\ndef rebatch(pad_idx, batch):\r\n \"Fix order in torchtext to match ours\"\r\n # ????\r\n src, trg = batch.src[0], batch.trg[0]\r\n return Batch(src, trg, pad_idx)\r\n\r\n\r\n# load\r\ndef load_data(path, max_src_len, max_trg_len, BATCH_SIZE, device):\r\n\r\n # Extra vocabulary sym bols\r\n pad_token = \"\"\r\n unk_token = \"\"\r\n bos_token = \"\"\r\n eos_token = \"\"\r\n\r\n extra_tokens = [pad_token, unk_token, bos_token, eos_token]\r\n\r\n PAD = extra_tokens.index(pad_token)\r\n UNK = extra_tokens.index(unk_token)\r\n BOS = extra_tokens.index(bos_token)\r\n EOS = extra_tokens.index(eos_token)\r\n\r\n SRC = data.Field(sequential=True, use_vocab=False, include_lengths=True, batch_first=True,\r\n pad_token=PAD, unk_token=UNK, init_token=BOS, eos_token=EOS, )\r\n TGT = data.Field(sequential=True, use_vocab=False, include_lengths=True, batch_first=True,\r\n pad_token=PAD, unk_token=UNK, init_token=BOS, eos_token=EOS, )\r\n fields = (SRC, TGT)\r\n\r\n # path = 'small-train.t7'\r\n dataset = torch.load(path)\r\n train_src, train_tgt = dataset['train_src'], dataset['train_tgt']\r\n dev_src, dev_tgt = dataset['dev_src'], dataset['dev_tgt']\r\n\r\n def filter_pred(example):\r\n if len(example.src) <= max_src_len and len(example.trg) <= max_trg_len:\r\n return True\r\n return False\r\n\r\n train = ParallelDataset(train_src, train_tgt, fields=fields, filter_pred=filter_pred)\r\n val = ParallelDataset(dev_src, dev_tgt, fields=fields, filter_pred=filter_pred)\r\n\r\n MIN_FREQ = 1\r\n SRC.build_vocab(train.src, min_freq=MIN_FREQ)\r\n TGT.build_vocab(train.trg, min_freq=MIN_FREQ)\r\n\r\n #########################################\r\n train_iter = MyIterator(train, batch_size=BATCH_SIZE, device=device,\r\n repeat=False, sort_key=lambda x: (len(x.src), len(x.trg)),\r\n batch_size_fn=batch_size_fn, train=True)\r\n valid_iter = MyIterator(val, batch_size=BATCH_SIZE, device=device,\r\n repeat=False, sort_key=lambda x: (len(x.src), len(x.trg)),\r\n batch_size_fn=batch_size_fn, train=False)\r\n return SRC, TGT, train_iter, valid_iter\r\n\r\n\r\ndef greedy_decode(model, src, src_mask, max_len, start_symbol):\r\n memory = model.encode(src, src_mask)\r\n ys = torch.ones(1, 1).fill_(start_symbol).type_as(src.data)\r\n\r\n for i in range(max_len - 1):\r\n out = model.decode(memory, src_mask, Variable(ys), Variable(subsequent_mask(ys.size(1)).type_as(src.data)))\r\n prob = model.generator(out[:, -1])\r\n _, next_word = torch.max(prob, dim=1)\r\n next_word = next_word.data[0]\r\n ys = torch.cat([ys, torch.ones(1, 1).type_as(src.data).fill_(next_word)], dim=1)\r\n return ys\r\n\r\n\r\ndef main(opt):\r\n use_cuda = torch.cuda.is_available()\r\n if use_cuda:\r\n print(\"Use GPU...\")\r\n device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\r\n # device = torch.device(\"cpu\")\r\n SRC, TGT, train_iter, valid_iter = load_data(opt.data_path, max_src_len=opt.max_src_seq_len, max_trg_len=opt.max_tgt_seq_len, BATCH_SIZE=opt.batch_size, device=device)\r\n pad_idx = TGT.vocab.stoi[1]\r\n model = make_model(len(SRC.vocab), len(TGT.vocab), N=opt.n_layers, d_model=opt.d_model, d_ff=opt.d_ff, h=opt.n_heads, dropout=opt.dropout).to(device)\r\n criterion = LabelSmoothing(size=len(TGT.vocab), padding_idx=pad_idx, smoothing=0.1).to(device)\r\n model_opt = NoamOpt(model.src_embed[0].d_model, 1, opt.n_warp_steps,\r\n torch.optim.Adam(model.parameters(), lr=opt.lr, betas=(0.9, 0.98), eps=1e-9))\r\n for epoch in range(opt.epochs):\r\n print('epoch:', epoch)\r\n model.train()\r\n run_epoch((rebatch(pad_idx, b) for b in train_iter),\r\n model,\r\n SimpleLossCompute(model.generator, criterion, model_opt))\r\n model.eval()\r\n run_epoch((rebatch(pad_idx, b) for b in valid_iter), model,\r\n SimpleLossCompute(model.generator, criterion, opt=None))\r\n\r\n torch.save(model, opt.save_model)\r\n\r\n\r\n# ################################\r\nif __name__ == '__main__':\r\n parser = argparse.ArgumentParser(description='Training Hyperparams')\r\n parser.add_argument('-data_path', default='vatex_len_40-train.t7', help='Path to the preprocessed data')\r\n\r\n # network params\r\n parser.add_argument('-d_model', type=int, default=512)\r\n parser.add_argument('-d_ff', type=int, default=2048)\r\n parser.add_argument('-n_heads', type=int, default=8)\r\n parser.add_argument('-n_layers', type=int, default=6)\r\n parser.add_argument('-dropout', type=float, default=0.1)\r\n\r\n # training params\r\n parser.add_argument('-lr', type=float, default=0.02)\r\n parser.add_argument('-epochs', type=int, default=1)\r\n parser.add_argument('-batch_size', type=int, default=128)\r\n parser.add_argument('-max_src_seq_len', type=int, default=50)\r\n parser.add_argument('-max_tgt_seq_len', type=int, default=50)\r\n # parser.add_argument('-max_grad_norm', type=float, default=None)\r\n parser.add_argument('-n_warp_steps', type=int, default=4000)\r\n parser.add_argument('-display_freq', type=int, default=100)\r\n # parser.add_argument('-log', default=None)\r\n # parser.add_argument('-model_path', type=str, default='None')\r\n parser.add_argument('-save_model', type=str, default='final_model.pkl')\r\n\r\n opt = parser.parse_args()\r\n print(opt)\r\n main(opt)\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"Transformer/project.py","file_name":"project.py","file_ext":"py","file_size_in_byte":7421,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"654447831","text":"import os\nimport unittest\n\nfrom app.init_app import socketIo, app\nfrom flask_migrate import Migrate, MigrateCommand\nfrom flask_script import Manager\nfrom app import blueprint\nfrom app.main import db\n\n\nif len(app.blueprints) < 1:\n app.register_blueprint(blueprint)\n app.app_context().push()\n app.config['ERROR_404_HELP'] = False\n\nmanager = Manager(app)\n\nmigrate = Migrate(app, db)\n\nmanager.add_command('db', MigrateCommand)\n\n\n@manager.command\ndef run():\n socketIo.run(app)\n\n\n@manager.command\ndef test():\n tests = unittest.TestLoader().discover('app/test', pattern='test*.py')\n result = unittest.TextTestRunner(verbosity=2).run(tests)\n if result.wasSuccessful():\n return 0\n return 1\n\n\nif __name__ == '__main__':\n manager.run()\n","sub_path":"manage.py","file_name":"manage.py","file_ext":"py","file_size_in_byte":759,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"181510598","text":"#!/usr/bin/env python2\nimport os\nimport argparse\nimport csv\nimport cv2\nfrom PIL import Image\nfrom tqdm import tqdm\nfrom verses_count import verses_count\nfrom ayat import find_ayat, draw, output_aya_segment\nfrom utils import safe_makedir, load_lines\n\n\ndef parse_arguments():\n parser = argparse.ArgumentParser()\n parser.add_argument('--input_path', type=str, required=True,\n help='''Path to input folder containing PNG images''')\n parser.add_argument('--output_path', type=str, required=True,\n help='''Path to output folder to generate verification images in''')\n parser.add_argument('--separator1_path', type=str, required=True,\n help='''Path to separator image template for pages 1 and 2''')\n parser.add_argument('--separator3_path', type=str, required=True,\n help='''Path to separator image template for pages 3 up to the end''')\n parser.add_argument('--count_method', type=str, choices=list(verses_count.keys()), required=True,\n help='''Counting method to use''')\n parser.add_argument('--pages', type=str, default='1..604',\n help='''Comma seprated page numbers or ranges, default is 1..604''')\n parser.add_argument('--start_sura', type=str, default='1',\n help='''Start sura numbers for each page in the input pages, default = 1''')\n parser.add_argument('--start_aya', type=str, default='1',\n help='''Start aya numbers for each page in the input pages, default = 1''')\n parser.add_argument('--start_sura_aya_tsv', type=str,\n help='''If specified, reads start_sura and start_aya numbers from recitation tsv''')\n parser.add_argument('--matching_threshold', type=float, default=0.42,\n help='''Matching threshold to match aya separators, default = 0.42''')\n return parser.parse_args()\n\n\ndef parse_start_tuples(args):\n if args.start_sura_aya_tsv:\n return read_start_tuples_tsv(args.pages, args.start_sura_aya_tsv)\n else:\n return read_start_tuple_args(args.pages, args.start_sura, args.start_aya)\n\n\ndef read_start_tuples_tsv(pages, tsv):\n # read tsv and return tuples (page, sura, ayah) for lines with lineId == 1\n if not os.path.isfile(tsv):\n raise RuntimeError(\"File: '%s' not found\" % tsv)\n\n # open file in universal new line mode: U\n with open(tsv, 'rU') as csvfile:\n reader = csv.reader(csvfile, delimiter='\\t', quotechar='\"')\n # file format: rewayaId\tpageId\tsuraId\tverseId\tlineId\tdata\n pages_dict = {}\n for row in reader:\n if row[4] == '1' and pages_dict.get(int(row[1])) == None:\n pages_dict[int(row[1])] = (int(row[2]), int(row[3]))\n\n tuples = []\n total_pages = 0\n for page in pages.split(','):\n page_range = page.split('..')\n start_page = int(page_range[0])\n end_page = int(page_range[-1])\n total_pages += end_page - start_page + 1\n tuples.append({\n 'start_page': start_page,\n 'end_page': end_page,\n 'start_sura': pages_dict[start_page][0],\n 'start_aya': pages_dict[start_page][1]\n })\n return total_pages, tuples\n\ndef read_start_tuple_args(pages, start_sura, start_aya):\n tuples = []\n pages = pages.split(',')\n suras = start_sura.split(',')\n ayahs = start_aya.split(',')\n total_pages = 0\n\n if not (len(pages) == len(suras) == len(ayahs)):\n raise ValueError('Lengths of pages, start_sura and start_aya arguments must match')\n\n for page, sura, ayah in zip(pages, suras, ayahs):\n page_range = page.split('..')\n start_page = int(page_range[0])\n end_page = int(page_range[-1])\n total_pages += end_page - start_page + 1\n tuples.append({\n 'start_page': start_page,\n 'end_page': end_page,\n 'start_sura': int(sura),\n 'start_aya': int(ayah)\n })\n return total_pages, tuples\n\ndef main_find_ayat(progress_bar, all_pages_lines, count_method, start_page, end_page,\n start_sura, start_aya, separator1_path, separator3_path,\n matching_threshold, input_path, output_path, segments_path):\n # by default, we don't increase the ayah on the top of this loop\n # to handle ayat that span multiple pages - this flag allows us to\n # override this.\n end_of_ayah = False\n sura = start_sura\n ayah = 1\n default_lines_to_skip = 2\n count_ayat = verses_count[count_method]\n\n if start_page == 1: # default behavior for page 1\n # in kofy/makky basmallah is counted as aya 1\n if count_method == 'kofy' or count_method == 'makky':\n lines_to_skip = 1\n else:\n lines_to_skip = 2\n elif start_page == 2: # default behavior for page 2\n lines_to_skip = 2\n elif start_sura == 9 and start_aya < 2: # skip only 1 line (header) for tawba\n lines_to_skip = 1\n elif start_aya < 1: # skip 1 if ayah = 0 (basmalah), skip 2 if ayah = -1 (header)\n lines_to_skip = 1 - start_aya\n else: # don't skip anything, starting from the middle of a sura\n lines_to_skip = 0\n ayah = start_aya\n\n for i in range(start_page, end_page + 1):\n filename = str(i) + '.png'\n lines = all_pages_lines[i]\n\n img_gray = cv2.imread(input_path + filename, -1)\n if i == 1 or i == 2:\n template = cv2.imread(separator1_path, -1)\n else:\n template = cv2.imread(separator3_path, -1)\n\n ayat = find_ayat(img_gray, template, matching_threshold)\n\n tpl_width = template.shape[1]\n\n current_line = 0\n x_pos_in_line = -1\n num_lines = len(lines)\n\n with open(segments_path + str(i).zfill(3) + \".sql\", \"wb\") as segments_file:\n first = True\n for ayah_item in ayat:\n if ((end_of_ayah or not first) and count_ayat[sura - 1] == ayah):\n sura = sura + 1\n ayah = 1\n lines_to_skip = default_lines_to_skip\n if sura == 9:\n lines_to_skip = lines_to_skip - 1\n end_of_ayah = False\n elif (end_of_ayah or not first):\n ayah = ayah + 1\n end_of_ayah = False\n first = False\n y_pos = ayah_item[1]\n\n pos = 0\n for line in range(current_line, num_lines):\n cur_line = lines[line]\n miny = cur_line[0][1]\n maxy = cur_line[1][1]\n cur_line_minx = cur_line[0][0]\n cur_line_maxx = cur_line[1][0]\n if lines_to_skip > 0:\n # it is header if skipping 2 lines (obviously)\n # or skipping 1 line and we are in sura 9 tawba (no basmala)\n # or skipping 1 line and we are in sura 1 fatiha and first line\n # (some counting methods don't have basmalah, in which case they skip only 1 line)\n if lines_to_skip == 2 or lines_to_skip == 1 and (sura == 9 or sura == 1 and line == 0): # header\n vals = (i, line + 1, sura, -1, 1, cur_line_minx, cur_line_maxx, miny, maxy)\n else: # basmala\n vals = (i, line + 1, sura, 0, 1, cur_line_minx, cur_line_maxx, miny, maxy)\n output_aya_segment(vals, img_gray, segments_file)\n lines_to_skip = lines_to_skip - 1\n current_line = current_line + 1\n continue\n pos = pos + 1\n if y_pos <= maxy:\n # we found the line with the ayah\n maxx = cur_line_maxx\n if x_pos_in_line > 0:\n maxx = x_pos_in_line\n minx = ayah_item[0]\n # small value indicating that the separator is the last thing comes in line\n # or page 1 and second line (always basmalah), this is to cover basmalah when counted as aya 1\n if minx < tpl_width/2 or i == 1 and line == 1:\n minx = 0\n end_of_sura = False\n if count_ayat[sura - 1] == ayah:\n end_of_sura = True\n\n # last aya in sura segment must extend to the leftmost, the empty space is ugly\n # also last ayah in page 2 must extend to the leftmost\n if end_of_sura or i == 2 and ayah_item == ayat[-1]:\n minx = 0\n\n vals = (i, line + 1, sura, ayah, pos, minx, maxx, miny, maxy)\n output_aya_segment(vals, img_gray, segments_file)\n\n # check if this is header/basmalah, it must occupy the whole line\n if (vals[3] == -1 or vals[3] == 0) and (vals[5] != 0 or vals[6] != cur_line_maxx):\n raise RuntimeError(\n 'Something is wrong: Header or Basmalah are not occuping the whole line')\n\n if end_of_sura or abs(minx - cur_line_minx) < tpl_width/2:\n x_pos_in_line = -1\n current_line = current_line + 1\n if current_line == num_lines:\n # last line, and no more ayahs - set it to increase\n end_of_ayah = True\n else:\n x_pos_in_line = minx\n break\n else:\n # we add this line\n maxx = cur_line_maxx\n if x_pos_in_line > 0:\n maxx = x_pos_in_line\n x_pos_in_line = -1\n current_line = current_line + 1\n vals = (i, line + 1, sura, ayah, pos, cur_line_minx, maxx,\n miny, maxy)\n output_aya_segment(vals, img_gray, segments_file)\n\n # draw aya separators\n draw(img_gray, template, ayat)\n\n # handle cases when the sura ends on a page, and there are no more\n # ayat. this could mean that we need to adjust lines_to_skip (as is\n # the case when the next sura header is at the bottom) or also add\n # some ayat that aren't being displayed at the moment.\n if end_of_sura:\n # end of sura always means x_pos_in_line is -1\n sura = sura + 1\n ayah = 1\n lines_to_skip = default_lines_to_skip\n if sura == 9:\n lines_to_skip = lines_to_skip - 1\n end_of_ayah = False\n while line + 1 < num_lines and lines_to_skip > 0:\n line = line + 1\n if lines_to_skip == 2 or lines_to_skip == 1 and sura == 9: # header\n vals = (i, line + 1, sura, -1, 1, lines[line][0][0], lines[line][1][0], lines[line][0][1], lines[line][1][1])\n else: # basmala\n vals = (i, line + 1, sura, 0, 1, lines[line][0][0], lines[line][1][0], lines[line][0][1], lines[line][1][1])\n output_aya_segment(vals, img_gray, segments_file)\n lines_to_skip = lines_to_skip - 1\n if lines_to_skip == 0 and line + 1 != num_lines:\n ayah = 0\n\n # we have some lines unaccounted for or stopped mid-line\n if x_pos_in_line != -1 or line + 1 != num_lines:\n if x_pos_in_line == -1:\n line = line + 1\n pos = 0\n ayah = ayah + 1\n # we ignore pages 1 and 2 because they always have empty spaces at the end\n if i > 2:\n for l in range(line, num_lines):\n cur_line = lines[l]\n pos = pos + 1\n maxx = cur_line[1][0]\n if x_pos_in_line > 0:\n maxx = x_pos_in_line\n x_pos_in_line = -1\n vals = (i, l + 1, sura, ayah, pos, cur_line[0][0], maxx,\n cur_line[0][1], cur_line[1][1])\n output_aya_segment(vals, img_gray, segments_file)\n\n # done with detecting segments, now write using cv2\n image_name = output_path + str(i).zfill(3) + \".png\"\n cv2.imwrite(image_name, img_gray)\n\n # now paste segmented to original\n original = Image.open(input_path + filename).convert('RGBA')\n segmented = Image.open(image_name).convert('RGBA')\n segmented.paste(original, mask=original)\n segmented.save(image_name, \"PNG\")\n\n # finally increment progress bar\n progress_bar.update(1)\n\n\nif __name__ == \"__main__\":\n args = parse_arguments()\n total_pages, start_tuples = parse_start_tuples(args)\n\n input_path = args.input_path + '/'\n output_path = safe_makedir(args.output_path + '/ayat/')\n segments_path = safe_makedir(args.output_path + '/segments/')\n lines = load_lines(args.output_path)\n\n print(\"Finding aya boundaries using separator templates into \" + output_path + \"...\")\n with tqdm(total=total_pages) as pbar:\n for tuple in start_tuples:\n main_find_ayat(progress_bar=pbar,\n all_pages_lines=lines,\n start_page=tuple['start_page'],\n end_page=tuple['end_page'],\n start_sura=tuple['start_sura'],\n start_aya=tuple['start_aya'],\n count_method=args.count_method,\n separator1_path=args.separator1_path,\n separator3_path=args.separator3_path,\n matching_threshold=args.matching_threshold,\n input_path=input_path,\n output_path=output_path,\n segments_path=segments_path)\n","sub_path":"detect_ayat.py","file_name":"detect_ayat.py","file_ext":"py","file_size_in_byte":12694,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"4675312","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on 7/21/17\nAuthor: Jihoon Kim\n\"\"\"\n\n\nfrom konlpy.tag import Twitter\nfrom konlpy.utils import pprint\nimport matplotlib.pyplot as plt\nimport matplotlib\nmatplotlib.rc('font', family=\"NanumBarunGothic\")\nimport seaborn as sns\nimport pandas as pd\nfrom wordcloud import WordCloud, ImageColorGenerator\nfrom collections import Counter\nimport numpy as np\nfrom PIL import Image\ndata = pd.read_csv('./data/c_data/덕수궁.csv')\n\ntext_sample = data.review\n\nlength = data.shape[0]\ntwit = Twitter()\npos_counter = Counter()\nfor i in range(length):\n\n twit_pos = twit.pos(text_sample[i])\n pos_counter += Counter(twit_pos)\n print(\"{}th Document is being processed.\".format(i))\n\ncount_df = pd.DataFrame.from_dict(pos_counter, orient='index').reset_index()\ncount_df.rename(columns={'index': 'pos', 0: 'count'}, inplace=True)\nsorted = count_df.sort_values(by='count', ascending=False)\nsorted['pos_word'] = sorted.pos.apply(lambda x: x[0])\nsorted['pos'] = sorted.pos.apply(lambda x: x[1])\n\nav_pos = ['Noun', 'Verb', 'Adjective', 'Adverb']\ngo_index = sorted.pos.isin(av_pos)\nsorted = sorted[go_index]\n\n\nword = list(sorted.pos_word)\nword = ' '.join(word)\npalace_coloring = np.array(Image.open(\"./cablecar.png\"))\nwc = WordCloud(background_color=\"white\", max_words=2000, mask=palace_coloring,\n max_font_size=40, random_state=42)\n# generate word cloud\nwc.generate(word)\n\n# create coloring from image\nimage_colors = ImageColorGenerator(palace_coloring)\n# recolor wordcloud and show\n# we could also give color_func=image_colors directly in the constructor\nplt.figure()\nplt.imshow(wc.recolor(color_func=image_colors), interpolation=\"bilinear\")\nplt.axis(\"off\")\n\nplt.figure()\nplt.imshow(palace_coloring, cmap=plt.cm.gray, interpolation=\"bilinear\")\nplt.axis(\"off\")\nplt.show()","sub_path":"nlp.py","file_name":"nlp.py","file_ext":"py","file_size_in_byte":1800,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"563705907","text":"import nltk\nimport re\nlemmatizer=nltk.stem.WordNetLemmatizer()\n\n# read in and clean corpus\nwith open('test_corpus.txt','r') as f:\n data=f.read()\n\ndef clean_line(file_line):\n sentences=nltk.sent_tokenize(file_line)\n new_sentences=[]\n for sentence in sentences:\n #sentence=re.sub(r'[^a-zA-z\\s]',r'',sentence)\n #sentence=re.sub(\"(?<=[a-z])'(?=[a-z])\", \"\", sentence)\n words=sentence.split()\n words=[word.lower() for word in words]\n #words=[lemmatizer.lemmatize(word) for word in words]\n sentence=' '.join(words)\n new_sentences.append(sentence)\n return new_sentences\n \n\n# test yield \ndef get_sentences(filename): \n with open(filename,'r') as f:\n for line in f:\n yield line\n\n# sentences=get_sentences('test_corpus.txt')\n# type(next(sentences))\n\n# Gold hypernyms\nimport pandas as pd\nhypers=pd.read_table('/home/srawat/Documents/Hearst Pattern Analysis/wbless/wbless.tsv')\nhypers=hypers[hypers.relation=='hyper']\nhypers=hypers.reset_index(drop=True)\nhypos=hypers.word1\nhypers=hypers.word2\n\n# Extractions from corpus\ndef extract_sents(filename,hypos,hypers):\n extractions=[]\n with open(filename,'r') as f:\n for line in f:\n sents=clean_line(line)\n for sent in sents:\n words=sent.split()\n for w1,w2 in zip(hypos,hypers):\n if w1 in words and w2 in words:\n extractions.append((sent,w1,w2))\n return extractions\n\n# Extractions from corpus without pre-processing\ndef extract_sents(filename,hypos,hypers):\n extractions=[]\n with open(filename,'r') as f:\n for line in f:\n sents=nltk.sent_tokenize(line)\n for sent in sents:\n words=sent.split()\n for w1,w2 in zip(hypos,hypers):\n if w1 in words and w2 in words:\n extractions.append((sent,w1,w2))\n return extractions\ntest_extractions=extract_sents('test_corpus.txt',hypos,hypers)\n \n# !TODO significantly less extractions without pre-processing\n\n# Check dep paths between pairs\n# Format: sent,hypo,hyper\nimport networkx as nx\nimport spacy\nnlp=spacy.load('en_core_web_sm')\n\ndef extract_dep_paths(extractions):\n dep_paths=[]\n for sent,hypo,hyper in extractions:\n doc=nlp(sent)\n edges=[]\n for token in doc:\n for child in token.children:\n edges.append(('{0}'.format(token.lower_),'{0}'.format(child.lower_)))\n graph=nx.Graph(edges)\n try:\n path=nx.shortest_path(graph,source=hypo,target=hyper)\n paths=[]\n for word in path:\n for token in doc:\n if word==token.text.lower():\n paths.append(token.text+'-'+token.pos_+'-'+token.dep_)\n paths=' '.join(paths)\n dep_paths.append((paths,hypo,hyper))\n except nx.NetworkXNoPath:\n pass\n return dep_paths\n\ntest_dep_paths=extract_dep_paths(test_extractions)\n\n# All dependencies\nall_deps=[]\nfor i in range(len(test_dep_paths)):\n sent=test_dep_paths[i][0]\n deps=[x.split('-')[2] for x in sent.split()]\n all_deps.append(deps)\n\n# unique dependencies\nimport itertools\ndeps=list(itertools.chain.from_iterable(all_deps))\ndeps=list(set(deps))\n\n# Dict vocab for deps\ndep_dict={}\nfor i in range(len(deps)):\n if deps[i] in dep_dict:\n pass\n else:\n dep_dict[deps[i]]=i\n\n# Encode dep paths extracted and freq\nenc_all_deps=[]\nfor dep in all_deps:\n _=[]\n for obj in dep:\n _.append(dep_dict[obj])\n enc_all_deps.append(_)\n\n# \n# more_than_once=[]\n# for dep in enc_all_deps:\n# for dep2 in enc_all_deps:\n# if dep==dep2:\n# more_than_once.append(dep)\n\ndep_freq_dict={}\nfor dep in enc_all_deps:\n if tuple(dep) in dep_freq_dict:\n dep_freq_dict[tuple(dep)]+=1\n else:\n dep_freq_dict[tuple(dep)]=1\n\n# stanford NLP check\n# check spacy github repo, for dependency matcher from text\n\n \n\n\n\n\n\n# test\ntest=test_dep_paths[1][0]\n[x.split('-')[2] for x in test.split()]\n\n\n# TODO: Freq of dep extractions\n\n\n# test dep\ndoc=nlp(u'Countries such as Spain, France and Germany.')\nspacy.displacy.render(doc,style='dep')\n\nedges=[] \nfor token in doc:\n for child in token.children:\n edges.append(('{0}'.format(token.lower_),'{0}'.format(child.lower_)))\n\ng=nx.Graph(edges)\npath=nx.shortest_path(g,source='spain',target='countries')\npaths=[]\nfor word in path:\n for token in doc:\n if word==token.text.lower():\n paths.append(token.text+'-'+token.pos_+'-'+token.dep_)\n \n","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":4628,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"212654157","text":"from django.conf.urls import url\n\nfrom . import views\n\nurlpatterns = [\n url( r'^$', views.index, name='index' ),\n url( r'^ver/(.*)?/$', views.ver, name='ver' ),\n url( r'manejar/(.*)?/$', views.manejar, name='manejar' ),\n url( r'eliminar/(.*)?/$', views.eliminar, name='eliminar' )\n]","sub_path":"paises/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":294,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"436121881","text":"import hashlib\nimport json\nimport base64\nimport requests\nfrom flask import Flask\nfrom flask import jsonify, make_response, request\nfrom flask_sqlalchemy import SQLAlchemy\nfrom random import randint\nfrom datetime import datetime\nfrom Crypto.Cipher import AES\nfrom fastecdsa import curve, point\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///data'\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(app)\n\nid = None\nrid = None\npub_key = None\nrid_ta = None\nbase_point = None\nprime = (2 ** 256) - (2 ** 224) + (2 ** 192) + (2 ** 96) - 1\nfield = prime\nmax_transmission_time = 2\n\n\nclass SmartMeter(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n shared_token = db.Column(db.String(250))\n pseudo_identity = db.Column(db.String(250))\n otp = db.Column(db.String(8))\n session_key = db.Column(db.String(250))\n\n\ndef length_equalise(x, y):\n if len(x) > len(y):\n y = y + ' ' * (len(x) - len(y))\n elif len(y) > len(y):\n x = x + ' ' * (len(y) - len(x))\n\n return x, y\n\n\ndef xor_word(str1, str2):\n return ''.join(chr(ord(s) ^ ord(c)) for s, c in zip(*length_equalise(str1, str2)))\n\ndef get_nonce():\n return randint(0, field - 1)\n\n\n@app.route(\"/add_substation\", methods=[\"POST\"])\ndef initialize():\n global id, rid, pub_key, rid_ta, base_point\n data = json.loads(request.get_json())\n\n if data is None:\n return make_response(\"Invalid Request\", 400)\n\n id = data['id_ss']\n rid = data['rid_ss']\n pub_key = data['pub_key']\n rid_ta = data['rid_ta']\n base_point = data['base_point']\n\n return make_response(\"Substation Initialised\", 200)\n\n\n@app.route('/add-smart-meter', methods=['POST'])\ndef add_smart_meter():\n data = request.get_json()\n shared_token = data['shared_token']\n smart_meter_rid = data['pseudo_identity']\n otp = randint(10000, 999999)\n # otp = 278324\n new_smart_meter = SmartMeter(shared_token=shared_token, pseudo_identity=smart_meter_rid, otp=otp)\n db.session.add(new_smart_meter)\n db.session.commit()\n requests.get('http://localhost:8000/first-handshake')\n\n return make_response('Smart Meter added', 200)\n\n\n@app.route('/first-handshake', methods=['POST'])\ndef first_contact():\n data = json.loads(request.get_json())\n request_time_string = data['Time']\n request_time = datetime.strptime(request_time_string, '%Y-%m-%d %H:%M:%S.%f')\n current_time = datetime.now()\n timedelta = current_time - request_time\n\n if timedelta.seconds > max_transmission_time:\n return make_response('Stale Request', 400)\n\n phi = data['phi']\n alpha = data['alpha']\n encrypted_message = data['E']\n smart_meter_public_key = data['pub_sm']\n X_i = data['X_i']\n A_i = data['A_i']\n\n hashed_rid = hashlib.sha256((rid + str(phi)).encode()).hexdigest()\n xored_result = xor_word(alpha, hashed_rid)\n shared_token = xored_result.split('@')[0]\n\n a_smart_meter = SmartMeter.query.filter_by(shared_token=shared_token).scalar()\n\n if not a_smart_meter or xored_result.split('@')[1] != request_time_string:\n return make_response('Malicious Request. Dropping computation', 400)\n\n smart_meter_rid = a_smart_meter.pseudo_identity\n\n encryption_obj = AES.new(smart_meter_rid[:16].encode(), AES.MODE_CBC, shared_token.encode())\n encrypted_message = base64.b64decode(encrypted_message)\n original_message = encryption_obj.decrypt(encrypted_message).decode()\n otp = original_message.strip()\n\n if str(a_smart_meter.otp) != otp:\n return make_response('Malicious Request, OTP does not match. Dropping computation', 400)\n\n ci_hashed = hashlib.sha256((str(smart_meter_public_key['x']) + str(smart_meter_public_key['y']) +\n str(pub_key['x']) + str(pub_key['y'])).encode()).hexdigest()\n ci = int(ci_hashed, 16)\n\n lhs = curve.P256.G * A_i\n sm_pub_key = point.Point(smart_meter_public_key['x'], smart_meter_public_key['y'], curve=curve.P256)\n public_key = point.Point(pub_key['x'], pub_key['y'], curve=curve.P256)\n Xi = point.Point(X_i['x'], X_i['y'], curve=curve.P256)\n rhs = Xi + (sm_pub_key + (public_key * ci))\n\n if lhs == rhs:\n rnd_nonce = get_nonce()\n T2 = str(datetime.now())\n y_hashed = hashlib.sha256((smart_meter_rid + rid_ta + str(rnd_nonce) + str(otp) + T2).encode('utf-8')).hexdigest()\n y = int(y_hashed, 16)\n Yi = curve.P256.G * y\n Zi = y * Xi\n\n session_key = hashlib.sha256(str(smart_meter_rid + rid + str(A_i) + str(Zi) + str(Yi)).encode('utf-8')).hexdigest()\n a_smart_meter.session_key = session_key\n db.session.commit()\n Vi = hashlib.sha256((str(session_key) + str(request_time_string) + T2 + str(otp.strip())).encode('utf-8')).hexdigest()\n\n Y = {\n \"x\": Yi.x,\n \"y\": Yi.y\n }\n response = {\n \"Yi\": Y,\n \"Vi\": Vi,\n \"Time\": T2\n }\n return make_response(jsonify(response), 200)\n else:\n return make_response(\"Malicious Request. Verification Failed.\", 400)\n\n\n@app.route('/second-handshake', methods=['POST'])\ndef second_contact():\n data = json.loads(request.get_json())\n request_time_string = data['Time']\n request_time = datetime.strptime(request_time_string, '%Y-%m-%d %H:%M:%S.%f')\n current_time = datetime.now()\n timedelta = current_time - request_time\n\n if timedelta.seconds > max_transmission_time:\n return make_response('Stale Request', 400)\n\n ack = data['ack']\n phi = data['phi']\n alpha = data['alpha']\n hashed_rid = hashlib.sha256((rid + str(phi)).encode()).hexdigest()\n xored_result = xor_word(alpha, hashed_rid)\n shared_token = xored_result.split('@')[0]\n\n a_smart_meter = SmartMeter.query.filter_by(shared_token=shared_token).scalar()\n\n session_key = a_smart_meter.session_key\n ack_calculated = hashlib.sha256((str(session_key) + request_time_string).encode('utf-8')).hexdigest()\n\n if ack == ack_calculated:\n return make_response('Verified', 200)\n else:\n return make_response(\"Malicious Request. Verification Failed.\", 400)\n\n\nif __name__ == '__main__':\n db.create_all()\n app.run(port=8001, debug=True)\n","sub_path":"Substation/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":6203,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"587435983","text":"# Copyright 2020 Google LLC\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\" fci_graph unit tests\n\"\"\"\n#pylint: disable=protected-access\n\nimport numpy\nfrom numpy import linalg\nimport pytest\n\nfrom openfermion import FermionOperator\nfrom openfermion import bravyi_kitaev_code\n\nfrom fqe import wavefunction\nfrom fqe.hamiltonians import *\nfrom fqe.fqe_ops.fqe_ops import (\n NumberOperator,\n S2Operator,\n SzOperator,\n TimeReversalOp,\n)\nfrom fqe.fqe_decorators import build_hamiltonian\nfrom tests.unittest_data import build_hamiltonian as test_hamiltonian\nfrom fqe.util import tensors_equal\n\nimport fqe\n\n\ndef test_fqe_control_dot_vdot():\n \"\"\"Find the dot product of two wavefunctions.\n \"\"\"\n wfn1 = fqe.get_number_conserving_wavefunction(4, 8)\n wfn1.set_wfn(strategy='ones')\n wfn1.normalize()\n assert round(abs(fqe.vdot(wfn1, wfn1) - 1. + .0j), 7) == 0\n assert round(abs(fqe.dot(wfn1, wfn1) - 1. + .0j), 7) == 0\n wfn1.set_wfn(strategy='random')\n wfn1.normalize()\n assert round(abs(fqe.vdot(wfn1, wfn1) - 1. + .0j), 7) == 0\n\n\ndef test_Wavefunction():\n \"\"\"Test free function that construct a Wavefunction object\n \"\"\"\n wfn1 = fqe.Wavefunction([[2, 0, 2]])\n wfn2 = wavefunction.Wavefunction([[2, 0, 2]])\n for key, sector in wfn1._civec.items():\n assert key in wfn2._civec\n assert sector.coeff.shape == wfn2._civec[key].coeff.shape\n\n\ndef test_initialize_new_wavefunction():\n \"\"\"Test initialization of the new wavefunction\n \"\"\"\n nele = 3\n m_s = -1\n norb = 4\n wfn = fqe.get_wavefunction(nele, m_s, norb)\n assert isinstance(wfn, wavefunction.Wavefunction)\n\n\ndef test_initialize_new_wavefunctions_multi():\n \"\"\"Test initialization of the new wavefunction with multiple parameters\n \"\"\"\n multiple = [[4, 0, 4], [4, 2, 4], [3, -3, 4], [1, 1, 4]]\n wfns = fqe.get_wavefunction_multiple(multiple)\n for wfn in wfns:\n assert isinstance(wfn, wavefunction.Wavefunction)\n\n\ndef test_time_evolve():\n \"\"\"Test time_evolve of a wavefunction\n \"\"\"\n wfn1 = fqe.Wavefunction([[2, 0, 2]])\n wfn1.set_wfn('ones')\n wfn2 = wavefunction.Wavefunction([[2, 0, 2]])\n wfn2.set_wfn('ones')\n op = FermionOperator('1^ 3') + FermionOperator('3^ 1')\n time = 1.2\n wfn1 = fqe.time_evolve(wfn1, time, op, True)\n wfn2 = wfn2.time_evolve(time, op, True)\n for key, sector in wfn1._civec.items():\n assert key in wfn2._civec\n assert sector.coeff.shape == wfn2._civec[key].coeff.shape\n assert numpy.allclose(sector.coeff, wfn2._civec[key].coeff)\n\n\ndef test_apply():\n \"\"\"Test time_evolve of a wavefunction\n \"\"\"\n wfn1 = fqe.Wavefunction([[2, 0, 2]])\n wfn1.set_wfn('ones')\n wfn2 = wavefunction.Wavefunction([[2, 0, 2]])\n wfn2.set_wfn('ones')\n op = FermionOperator('1^ 3') + FermionOperator('3^ 1')\n wfn1 = fqe.apply(op, wfn1)\n wfn2 = wfn2.apply(op)\n for key, sector in wfn1._civec.items():\n assert key in wfn2._civec\n assert sector.coeff.shape == wfn2._civec[key].coeff.shape\n assert numpy.allclose(sector.coeff, wfn2._civec[key].coeff)\n\n\ndef test_expectationValue():\n \"\"\"Test time_evolve of a wavefunction\n \"\"\"\n wfn1 = fqe.Wavefunction([[2, 0, 2]])\n wfn1.set_wfn('ones')\n wfn2 = wavefunction.Wavefunction([[2, 0, 2]])\n wfn2.set_wfn('ones')\n op = sparse_hamiltonian.SparseHamiltonian(FermionOperator('1^ 3'))\n ex1 = fqe.expectationValue(wfn1, op)\n ex2 = wfn2.expectationValue(op)\n assert numpy.isclose(ex1, ex2)\n\n\ndef test_apply_generated_unitary():\n \"\"\"Test applying generated unitary transformation\n \"\"\"\n norb = 4\n nele = 3\n time = 0.001\n ops = FermionOperator('1^ 3^ 5 0', 2.0 - 2.j) + FermionOperator(\n '0^ 5^ 3 1', 2.0 + 2.j)\n\n wfn = fqe.get_number_conserving_wavefunction(nele, norb)\n wfn.set_wfn(strategy='random')\n wfn.normalize()\n\n reference = fqe.apply_generated_unitary(wfn, time, 'taylor', ops)\n\n h1e = numpy.zeros((2 * norb, 2 * norb), dtype=numpy.complex128)\n h2e = hamiltonian_utils.nbody_matrix(ops, norb)\n h2e = hamiltonian_utils.antisymm_two_body(h2e)\n hamil = general_hamiltonian.General(tuple([h1e, h2e]))\n compute = wfn.apply_generated_unitary(time, 'taylor', hamil)\n\n for key in wfn.sectors():\n diff = reference._civec[key].coeff - compute._civec[key].coeff\n err = linalg.norm(diff)\n assert err < 1.e-8\n\n\ndef test_cirq_interop(c_or_python):\n \"\"\"Check the transition from a line qubit and back.\n \"\"\"\n fqe.settings.use_accelerated_code = c_or_python\n work = numpy.random.rand(16).astype(numpy.complex128)\n work[0] = 0.0 + 0.0j\n work[15] = 0.0 + 0.0j\n norm = numpy.sqrt(numpy.vdot(work, work))\n numpy.divide(work, norm, out=work)\n wfn = fqe.from_cirq(work, thresh=1.0e-7)\n sec = [(1, -1), (1, 1), (2, -2), (2, 0), (2, 2), (3, -1), (3, 1)]\n assert set(sec) == set(wfn._civec.keys())\n test = fqe.to_cirq(wfn)\n assert numpy.allclose(test, work)\n\n # check with Bravyi-Kitaev\n bc = bravyi_kitaev_code(4)\n wfn = fqe.from_cirq(work, thresh=1.0e-7, binarycode=bc)\n test = fqe.to_cirq(wfn, binarycode=bc)\n assert numpy.allclose(test, work)\n\n\ndef test_get_spin_conserving_wavefunction():\n \"\"\" Test get_spin_conserving_wavefunction\n \"\"\"\n norb = 4\n s_z = 2\n wfn_spin = fqe.get_spin_conserving_wavefunction(s_z, norb)\n\n assert 's_z' in wfn_spin._conserved.keys()\n assert wfn_spin._conserved['s_z'] == 2\n assert wfn_spin.conserve_spin()\n\n ref_sectors = {(2, 2), (4, 2), (6, 2)}\n assert ref_sectors == set(wfn_spin.sectors())\n\n s_z = -2\n\n wfn_spin = fqe.get_spin_conserving_wavefunction(s_z, norb)\n\n assert 's_z' in wfn_spin._conserved.keys()\n assert wfn_spin._conserved['s_z'] == -2\n assert wfn_spin.conserve_spin()\n\n ref_sectors = {(2, -2), (4, -2), (6, -2)}\n assert ref_sectors == set(wfn_spin.sectors())\n\n\ndef test_get_number_conserving_wavefunction():\n \"\"\" Test get_number_conserving_wavefunction\n \"\"\"\n norb = 4\n nel = 2\n wfn_spin = fqe.get_number_conserving_wavefunction(nel, norb)\n\n assert 'n' in wfn_spin._conserved.keys()\n assert wfn_spin._conserved['n'] == 2\n assert wfn_spin.conserve_number()\n\n ref_sectors = {(2, 2), (2, 0), (2, -2)}\n assert ref_sectors == set(wfn_spin.sectors())\n\n\ndef test_operator_constructors():\n \"\"\" Creation of FQE-operators\n \"\"\"\n assert isinstance(fqe.get_s2_operator(), S2Operator)\n assert isinstance(fqe.get_sz_operator(), SzOperator)\n assert isinstance(fqe.get_time_reversal_operator(), TimeReversalOp)\n assert isinstance(fqe.get_number_operator(), NumberOperator)\n\n\ndef test_get_hamiltonian_from_openfermion_raises():\n \"\"\" Check the type check of get_hamiltonian_from_openfermion()\n \"\"\"\n with pytest.raises(AssertionError):\n fqe.get_hamiltonian_from_openfermion([])\n\n\ndef test_get_hamiltonian_from_openfermion():\n \"\"\" Check get_hamiltonian_from_openfermion()\n \"\"\"\n norb = 4\n ops = test_hamiltonian.number_nonconserving_fop(2, norb=norb)\n test = fqe.get_hamiltonian_from_openfermion(ops, norb=norb, \\\n conserve_number=False)\n test2 = build_hamiltonian(ops, norb=norb, conserve_number=False)\n assert test == test2\n\n\ndef test_get_diagonal_hamiltonian():\n \"\"\" Check whether get_diagonal_hamiltonian returns the same value as its\n underlying function is supposed to return.\n \"\"\"\n diag = numpy.zeros((5,), dtype=numpy.complex128)\n e_0 = -4.2\n test = diagonal_hamiltonian.Diagonal(diag, e_0)\n test2 = fqe.get_diagonal_hamiltonian(diag, e_0)\n\n assert test == test2\n\n\ndef test_get_diagonal_coulomb():\n \"\"\" Check whether get_diagonal_coulomb returns the same value as its\n underlying function is supposed to return.\n \"\"\"\n diag = numpy.zeros((5, 5), dtype=numpy.complex128)\n e_0 = -4.2\n test = diagonal_coulomb.DiagonalCoulomb(diag, e_0)\n test2 = fqe.get_diagonalcoulomb_hamiltonian(diag, e_0)\n\n assert test == test2\n\n\n@pytest.mark.parametrize(\"hamiltonian, get_function\", \\\n [ (sso_hamiltonian.SSOHamiltonian, fqe.get_sso_hamiltonian),\n (gso_hamiltonian.GSOHamiltonian, fqe.get_gso_hamiltonian),\n (general_hamiltonian.General, fqe.get_general_hamiltonian),\n (restricted_hamiltonian.RestrictedHamiltonian, \\\n fqe.get_restricted_hamiltonian)\n ])\ndef test_get_hamiltonians(hamiltonian, get_function):\n \"\"\" Check whether other Hamiltonian getters return the same value\n as their underlying functions are supposed to return.\n \"\"\"\n h1e = numpy.random.rand(5, 5).astype(numpy.complex128)\n e_0 = -4.2\n test = hamiltonian((h1e,))\n test2 = get_function((h1e,))\n\n assert test == test2\n\n\ndef test_get_sparse_hamiltonian():\n oper = FermionOperator('0 0^')\n test = sparse_hamiltonian.SparseHamiltonian(oper)\n test2 = fqe.get_sparse_hamiltonian(oper)\n\n assert test == test2\n","sub_path":"tests/_fqe_control_test.py","file_name":"_fqe_control_test.py","file_ext":"py","file_size_in_byte":9485,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"194479654","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nfrom flask import Flask, send_from_directory, request, jsonify\nfrom flask_cors import CORS, cross_origin\napp = Flask(__name__)\nCORS(app)\n\nfrom UploadSectionsHandler import UploadSectionsHandler\nfrom GetSectionsHandler import GetSectionsHandler\nfrom UploadPathHandler import UploadPathHandler\nfrom BoundingBoxHandler import BoundingBoxHandler\nfrom UploadMapHandler import UploadMapHandler\nfrom GetSVGPathHandler import GetSVGPathHandler\nfrom UploadImageHandler import UploadImageHandler\n\nfrom DatabaseManager import DatabaseManager\n\nfrom InvalidUsage import InvalidUsage\nimport sys\n\n@app.errorhandler(InvalidUsage)\ndef handle_invalid_usage(error):\n response = jsonify(error.to_dict())\n response.status_code = error.status_code\n return response\n\n@app.route('/')\ndef hello_world():\n return \"hello world\"\n\n# used to upload the SVG map\n@app.route('/upload-map', methods=['POST'])\ndef uploadMap():\n try:\n return UploadMapHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# used to upload an image for section content\n@app.route('/upload-image', methods=['POST'])\ndef uploadImage():\n try:\n return UploadImageHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# used after path has been adjusted, uploads path, zoom level, osm data and points\n@app.route('/upload-path', methods=['POST'])\ndef path():\n try:\n return UploadPathHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# requests boundingbox of a given mapId\n@app.route('/boundingbox', methods=['GET'])\ndef boundingbox():\n try:\n return BoundingBoxHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# requests the flattened svg path in coordinates\n@app.route('/svg-path', methods=['GET'])\ndef svgPath():\n try:\n return GetSVGPathHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# requests sections for a given mapId\n@app.route('/storytelling-sections', methods=['GET'])\ndef storytellingSections():\n try:\n return GetSectionsHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n# uploads changes to storytelling sections\n@app.route('/upload-storytelling-sections', methods=['POST'])\ndef uploadStorytellingSections():\n try:\n return UploadSectionsHandler().processRequest(request)\n except:\n raise InvalidUsage(str(sys.exc_info()))\n\n\n@app.route('/backend/output/')\ndef send_output(path):\n return send_from_directory('output', path)\n\nif __name__ == \"__main__\":\n DatabaseManager().initializeDatabase()\n app.run(debug=True, threaded=True)\n","sub_path":"backend/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"402136395","text":"import matplotlib.pylab as plt\nimport numpy as np\nimport math\n\n# nasz program bedzie się rozwijał z czasem o nowe przypadki tak by obiąć jak najszerszy zakres wszelkich rzutów\n# część interakcyjna z użytkownikeim jest zaplanowana na późniejsze terminy więc narazie przyjmiemy sobie pewne sałe\n\n# oznaczenia:\ng = 9.81 # można zmienic jak coś narazie jako stała w [m/s^2]\nh = 1.0 # wysokość z jakie rozpoczynamy rzut [m]\nVo = 20 # prędkość początkowa [m/s]\nalfa = 60 # kąt nachylenia wektora Vo od poziomu w stopniach ustalić przdział\ntc = 0.0 # calkowity czas lotu\nH = 0.0 # maksymalna wysokosc ns jaka wzniesie sie cialo od wysokosci poczatkowej\nHmax = 0.0 # maksymalna wysokosc na jaka wzniesie sie cialo\nz = 0.0 # zasieg rzutu\nt = 0.0\nVgr = 8000 #predkosc gazow wyrzucanych z rakiety\na = 5 #tempo utraty masy\nm = 250\nG = 6.7*(10**(-11))\nMz = 6*(10**24) #[kg]\nRz = 6371000 #[m]\nV1 = np.around(np.sqrt((G*Mz)/Rz), 1) #[m/s]\nV2 = np.around((np.sqrt(2) * V1), 1)\nx = []\ny = []\n\ndef skladowe (kat,pirewotna): #progra liczący skłodwe wektorów\n sinalf = np.around(np.sin(np.deg2rad(kat)),4)\n cosalf = np.around(np.cos(np.deg2rad(kat)),4)\n skłX = np.around(pirewotna * cosalf,4)\n skłY = np.around(pirewotna * sinalf,4)\n print('składowa wartosc poczatkowej wzdluz x:', skłX \\\n , 'składowa wartosc poczatkowej wzdluz y:', skłY)\n return [skłX,skłY]\n\n\ndef CzasLotu(g, Vox, Voy, h, beta):\n # x = Vox * t ==> t = x/Vox\n # y = h + Voy*t - (g*t^2)/2 ==> y = h + (Voy/Vox)*x - (g*x^2)/2*Vox^2\n # y = tg(beta)*t ==> y = tg(beta)/Vox * x\n # -(g*x^2)/2*Vox^2 + (Voy/Vox - tg(beta)/Vox) * x + h = 0\n a = - g / (2 * (Vox**2))\n b = (Voy/Vox) - (np.tan(np.deg2rad(beta)) / Vox)\n c = h\n delta = (b * b) - 4 * a * c\n x1 = (-b - math.sqrt(delta)) / (2 * a)\n x2 = (-b + math.sqrt(delta)) / (2 * a)\n if x1 >= x2:\n t = np.around(x1/Vox, 4)\n else:\n t = np.around(x2/Vox, 4)\n print('Czas lotu wynosi t =', t)\n return (t)\n\n\ndef MaxWysokosc(Voy, g): # program liczacy maksymalna wysokosc ciała\n H = np.around((Voy ** 2) / (2 * g), 4)\n Hmax = H + h\n print('maksymalna wysokosc na jaka wzleci cialo to H =', Hmax, 'm')\n return (Hmax)\n\n\ndef Zasieg(Vox, tc):\n z = np.around(Vox * tc, 4)\n print('Zasieg rzutu to z =', z, 'm')\n return (z)\n\ndef PredkosciChwilowe(tc, Vox, Voy, g,t,Hmax,h):\n global x\n global y\n i = 0\n while t < tc:\n Vx = Vox\n Vy = Voy - g * t\n if i >= 1:\n x.append(Vx * t)\n y.append(y[i-1] + Vy * dt)\n else:\n x.append(Vx * t)\n y.append(h + Vy * t)\n dt = 0.1\n t = t + dt\n i += 1\n return x, y\n\ndef Rakieta(m, a, t, Vgr):\n print(\"Rakieta\")\n while (a*t) <= 3*m/4:\n Vr = np.around(Vgr * np.log(m/(m - a * t)), 1)\n t += 0.5\n if Vr >= V1:\n print(\"Rakieta osiagnela I predkosc kosmiczna i moze pozostac na orbicie ziemi\")\n return Vr\n print(\"Rakieta nie osiagnela I predkosci kosmicznej\")\n\ndef InnePlanety(G, Mz, Rz):\n planety = {}\n odp = \"TAK\"\n print(\"Program oblicza I predkosc kosmiczna dla roznych planet. Predkosc jest podana w m/s.\")\n while odp == \"TAK\":\n nazwa = input(\"Wprowadz nazwe planety: \")\n x = float(input(\"Masa planety. Wprowadz liczbe przez ktora ma zostac przemnozona masa ziemi: \"))\n y = float(input(\"Promien planety. Wprowadz liczbe przez ktora ma zostac przemnozony promien ziemi: \"))\n mp = x * Mz\n rp = y * Rz\n v = np.around(np.sqrt((G * mp) / rp), 4)\n planety[nazwa] = v\n print(planety)\n odp = str.upper(input(\"Czy chcesz dodac kolejna planete, wpisz TAK lub NIE: \"))\n return planety\n\nprint('Witaj!!!!')\nh = float(input('Z jakiej wysokości rzucic h = '))\nalfa = float(input('Pod jakim kątem do poziomu rzut ma zostać wykonany alfa = '))\nVo = float(input('Jaka ma byc prędkość początkowa obiektu Vo = '))\nprint(\"\"\"\n 0 - podłoże nie idzie w góre ani w dół\n 1 - podłoże idzie w górę pod kątem beta do poziomu\n 2 - podłoże idzie w dół pod kątem beta do poziomu\n \"\"\")\nwybor = int(input('Twój wybór to: '))\nif wybor == 0:\n beta = 0\nelif wybor == 1:\n beta = float(input('Kąt beta ma wynosić beta = '))\nelif wybor == 2:\n beta = float(input('Kąt beta ma wynosić beta = '))\n beta = 180 - beta\nelse:\n print('Błąd nie ma takiego wyboru!!!')\n\nVox, Voy = skladowe(alfa, Vo)\ntc = CzasLotu(g,Vox,Voy,h,beta)\nHmax = MaxWysokosc(Voy, g)\nz = Zasieg(Vox, tc)\nPredkosciChwilowe(tc,Vox,Voy,g,t,Hmax,h)\n\nplt.plot(x, y)\nplt.title(\"Wykres rzutu\")\nplt.show()\n\n","sub_path":"rzut_poprawawzoru.py","file_name":"rzut_poprawawzoru.py","file_ext":"py","file_size_in_byte":4739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"625098194","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('Departamentos', '0008_auto_20150808_0104'),\n ('Historicos', '0004_auto_20150804_0456'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='registro',\n name='fk_departamento',\n field=models.ForeignKey(to='Departamentos.Departamento', null=True),\n ),\n ]\n","sub_path":"apps/Historicos/migrations/0005_registro_fk_departamento.py","file_name":"0005_registro_fk_departamento.py","file_ext":"py","file_size_in_byte":498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"9363693","text":"'''\nLomuto Partion.\n\nCondition for quick sort:\nLet's taeke an element pivot, and all the elements left to pivot is less than pivot and all the elements right to pivot is greater than pivot.\n\nThe main core for the quick sort is the partion.\n\n1. Make the last element as pivot and partion index as start.\n2. Traverse to the array, check if the ith element is less than or equal to pivot element.\n3. if the ith element is smaller, swap it with partion index. Increment the partion index.\n4. After all the swapping is done with ith and partion index, swap the end element and partion index.\n\n'''\n\n# Lomuto partion\n# def partion(array, start, end):\narray = [10,7,8,9,1,5]\nn = len(array)\nprint(\"Length\",n)\nstart = 0\nend = n-1\npivot = array[end]\nprint(\"Pivot\", pivot)\np_index = start\n\nfor i in range(n-1):\n\tprint(\"i:\",i)\n\tif array[i] <= pivot:\n\t\tprint(\"array[i]:\",array[i],\"pivot\", pivot)\n\t\tarray[i], array[p_index] = array[p_index], array[i]\n\t\tp_index += 1\n# swapping pivot and partion index\n\narray[end], array[p_index] = array[p_index], array[end]\nprint(array)\n\n'''\nHoare Partion:\n\nMost cases first element is pivot element.\n\n1. Make 2nd element as start is 'i' and last is 'j' element as end.\n2. Compare the ith element with pivot element, if the ith element is greater than pivot element, pause the i there.\n3. Now compare the pivot with element in j index, if the jth element is less than pivot swap i and j element.\n4. On moving forward we have to increment the j and for j decrement the j.\n5. Once if j crossed i swap the j and pivot element.\n\n'''\n# Hoare Partion\narray = [10, 7, 9, 8, 1, 5]\nn = len(array)\nprint(\"length:\", n)\nstart = 0\nend = n-1\npIndex = start\npivot = array[pIndex]\nwhile start pivot:\n\t\tend -= 1\n\n\tif start\"\n\nimport argparse, os, logging\nimport json\nfrom bs4 import BeautifulSoup\n\nlogging.basicConfig(filename='gent-simplify.log', level=logging.DEBUG)\n\ndef ckanparse(infile, outfile):\n\t\"\"\"Parse HTML file and store it in a new (simplified) JSON file.\"\"\"\n\tlogging.info('Parsing ' + str(infile))\n\n\tdata = {} \n\n\twith open(infile, 'r') as f:\n\t\tsoup = BeautifulSoup(f.read())\n\t\t\n\t\tdata['title'] = soup.find('h1', class_='title').get_text()\n\n\t\tbody = soup.find('div', class_='field-name-body')\n\t\tdata['body'] = body.find('p').get_text()\n\n\t\tfor k in ('contexts', 'urls', 'fmts', 'tags', 'license'):\n\t\t\tdata[k] = set()\t\n\n\t\tname = soup.find('link', rel='canonical')\n\t\tdata['contexts'].add('http://data.gent.be' + name['href'])\n\n\t\tnode = soup.find('div', class_='node-dataset')\n\n\t\tdiv = node.find('div', class_='field-name-field-basisinformatie')\n\t\tif div:\n\t\t\thref = div.find('a')\n\t\t\tif href:\n\t\t\t\tdata['contexts'].add(href['href'])\n\n\t\tdiv = node.find('fieldset', class_='group-download')\n\t\tif div:\n\t\t\threfs = div.find_all('a')\n\t\t\tfor href in hrefs:\n\t\t\t\tdata['urls'].add(href['href'])\n\t\t\t\tdata['fmts'].add(href.get_text())\n\n\t\tfor label in node.find_all('a', typeof='skos:Concept'):\n\t\t\tdata['tags'].add(label.get_text())\n\n\t# Add defaults for required fields\n\tdata['body'] = data['body'].strip()\n\n\tdata['org'] = ['--Gent']\n\tdata['geo'] = ['----Gent']\n\t\n\tdata['media'] = ['Download', 'Webservice']\n\n\tdata['license'].add('Zie website van de data leverancier')\n\n\t# make entries JSON serializable \n\tfor k in ('contexts', 'urls', 'tags', 'fmts', 'license'): \n\t\tdata[k] = list(data[k])\n\t\n\twith open(outfile, 'w') as f:\n\t\tjson.dump(data, f, indent=4)\n\n\ndef main():\t\n\tparser = argparse.ArgumentParser(description='Simplify Gent HTML files')\n\tparser.add_argument('--indir', help='Input directory', required=True)\n\tparser.add_argument('--outdir', help='Output directory', required=True)\n\t\n\targs = parser.parse_args()\n\t\n\tfor f in os.listdir(args.indir):\n\t\tckanparse(os.path.join(args.indir, f), os.path.join(args.outdir, f))\n\t\t\t\nif __name__ == \"__main__\":\n\tmain()\n","sub_path":"html/gent/gent-simplify.py","file_name":"gent-simplify.py","file_ext":"py","file_size_in_byte":2164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"115583481","text":"from src.utils.swap import swap\n\n\ndef quick_sort(arr):\n return qs(arr, 0, len(arr) - 1)\n\n\ndef qs(arr, start, end):\n if (start < end):\n partitionIdx = partition(arr, start, end)\n\n qs(arr, start, partitionIdx - 1)\n qs(arr, partitionIdx + 1, end)\n\n return arr\n\n\ndef partition(arr, start, end):\n pivot = arr[end]\n partitionIdx = start\n\n for i in range(start, end):\n if arr[i] <= pivot:\n swap(arr, i, partitionIdx)\n partitionIdx += 1\n\n swap(arr, partitionIdx, end)\n\n return partitionIdx\n","sub_path":"src/sort/quick.py","file_name":"quick.py","file_ext":"py","file_size_in_byte":556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"313305289","text":"# -*- coding: utf-8 -*-\nfrom .user_agents import agents\n\nBOT_NAME = 'wuba'\n\nSPIDER_MODULES = ['wuba.spiders']\nNEWSPIDER_MODULE = 'wuba.spiders'\n\n\nROBOTSTXT_OBEY = False\n\nCOOKIES_ENABLED = False\n\n# USER_AGENT = agents\n\nDOWNLOAD_DELAY = 1 # 延迟\n\n# LOG_LEVEL = 'INFO' # 日志级别\n\nITEM_PIPELINES = {\n 'wuba.pipelines.WubaPipeline': 300,\n}\n# # 配置redis\n# SCHEDULER = \"scrapy_redis.scheduler.Scheduler\" #调度\n# DUPEFILTER_CLASS = \"scrapy_redis.dupefilter.RFPDupeFilter\" #去重\n# SCHEDULER_PERSIST = True #不清理Redis队列\n# SCHEDULER_QUEUE_CLASS = \"scrapy_redis.queue.SpiderQueue\" #队列\n\n#配置MongoDB\nMONGODB_HOST = '127.0.0.1'\nMONGODB_PORT = 27017\nMONGODB_DBNAME = \"wuba\"\nMONGODB_DOCNAME = \"zufang\"\n\n# # 配置redis\n# REDIS_HOST = '192.168.1.199'\n# REDIS_PORT = 6379\n","sub_path":"wuba/wuba/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":801,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"70318360","text":"from pytest import fixture\nimport json\n\n\n@fixture(scope=\"session\")\ndef feedback():\n with open(\"feedback.jsonc\", \"r\") as handle:\n fixed_json = \"\".join(\n line for line in handle if not line.strip().startswith(\"//\")\n )\n return json.loads(fixed_json)\n\n\ndef test_dar_uma_nota_de_0_a_10_ao_projeto(feedback):\n nota = feedback.get(\"nota\")\n assert type(nota) is int\n assert 0 <= nota <= 10\n\n\ndef test_falar_dos_pontos_positivos(feedback):\n comentario = feedback.get(\"pontos_positivos\")\n assert type(comentario) is not None\n assert comentario.strip() != \"\"\n\n\ndef test_falar_dos_pontos_negativos(feedback):\n comentario = feedback.get(\"pontos_negativos\")\n assert type(comentario) is not None\n assert comentario.strip() != \"\"\n","sub_path":"tests/test_feedback.py","file_name":"test_feedback.py","file_ext":"py","file_size_in_byte":776,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"342909611","text":"import nnpu\nimport tvm\nimport topi\nfrom nnpu.utils import ScheduleProcHelper\nimport numpy as np\n\nimport argparse\n\nparser = argparse.ArgumentParser(description='test of NNPU Op')\nparser.add_argument('--sim', type=str, help='the simulator to use', \n default='S0', choices=['S0', 'S1', 'SC'])\nargs = parser.parse_args()\n\nenv = nnpu.get_env()\nnnpu.set_device(env, type=args.sim)\n\nwith ScheduleProcHelper():\n env = nnpu.get_env()\n shape = (48, 48)\n insn_shape = (16, 16)\n\n dtype_n, dtype_w = env.cfg['dtype_n'], env.cfg['dtype_w']\n a = tvm.placeholder(shape, dtype_n, 'a')\n b = tvm.placeholder(shape, dtype_n, 'b')\n \n sph = ScheduleProcHelper.current\n\n a_buf, a_dram = nnpu.utils.CopyHtoBuf(a, 'a', sph)\n b_buf, b_dram = nnpu.utils.CopyHtoBuf(b, 'b', sph)\n\n k = tvm.reduce_axis((0, shape[1]), 'k')\n dot_shape = (shape[0], )\n dot_buf = tvm.compute(dot_shape, \n lambda i: tvm.sum(a_buf[i, k].astype(dtype_w) * \n b_buf[i, k].astype(dtype_w), k), \n 'dot_buf')\n sph.MarkScope(dot_buf, 'acc')\n \n res_buf = nnpu.utils.CopyAccToBuf(dot_buf, 'res')\n \n res_host, _ = nnpu.utils.CopyBufToH(res_buf, 'res')\n\n # tensorize\n s = nnpu.create_schedule(res_host.op)\n xo, ro, xi, ri = s[dot_buf].tile(dot_buf.op.axis[0], dot_buf.op.reduce_axis[0],\n insn_shape[0], insn_shape[1])\n s[dot_buf].tensorize(xi, env.intrins.get('MRowDot', shape=insn_shape, \n mode='inc', scope_out='acc'))\n\n print(nnpu.lower(s, [a, b, res_host], simple_mode=True))\n \n func = nnpu.build(s, [a, b, res_host], 'nnpu', 'llvm', name='nnpu_func')\n\n print('------------------- device module 1 asm code: ')\n print(func.imported_modules[0].get_source('asm'))\n \n ctx = tvm.nd.TVMContext(13, 0)\n\n a_np = np.random.randint(size=shape, dtype=a.dtype, low = -32, high = 32)\n #a_np = np.random.random(size=shape).astype(a_host.dtype)\n a_nd = tvm.nd.array(a_np, ctx)\n\n b_np = np.random.randint(size=shape, dtype=b.dtype, low = -32, high = 32) \n b_nd = tvm.nd.array(b_np, ctx)\n c_nd = tvm.nd.array(np.zeros((shape[0], )).astype(res_host.dtype), ctx)\n\n func(a_nd, b_nd, c_nd)\n #print('a = ')\n #print(a_np)\n #print('b = ')\n #print(b_np)\n\n print(c_nd.asnumpy())\n print('ground truth is')\n gt = np.multiply(a_np, b_np, dtype=res_host.dtype)\n gt = np.sum(gt, axis=1)\n print(gt)\n np.testing.assert_allclose(c_nd.asnumpy(), gt)\n print('test passed')","sub_path":"nnpu/tests/test_strided/test_strided_mat_row_dot.py","file_name":"test_strided_mat_row_dot.py","file_ext":"py","file_size_in_byte":2586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"83732581","text":"from pyspark import SparkContext\nfrom pyspark.mllib.feature import HashingTF\nfrom pyspark.mllib.feature import IDF\nfrom pyspark.mllib.linalg import Vectors\nfrom pyspark.mllib.clustering import KMeans, KMeansModel\nfrom numpy import array\nfrom math import sqrt\n\ndef createKmeansVector(vector):\n\tparsedData = [0 for i in range(45000)]\n\tvectorSplit = vector.split(';')\n\tfor vs in vectorSplit:\n\t\tvSplit = vs.split(' ')\n\t\tparsedData[int(vSplit[0])] = float(vSplit[1])\n\treturn parsedData\n\nsc = SparkContext()\nparsedData = sc.textFile(\"hdfs://localhost:8020/pyspark/vector\")\\\n\t.map(lambda x: x.split(':'))\\\n\t.map(lambda x:createKmeansVector(x[1]))\n\t\nclusters = KMeans.train(parsedData, 10, maxIterations=10, runs=10, initializationMode=\"random\")\t\n\nclusters.save(sc, \"myModelPath\")\nsameModel = KMeansModel.load(sc, \"myModelPath\")\n\ndef error(point):\n center = sameModel.centers[sameModel.predict(point)]\n return sqrt(sum([x**2 for x in (point - center)]))\n\nWSSSE = parsedData\\\n\t.map(lambda point: error(point))\\\n\t.reduce(lambda x, y: x + y)\nprint(\"Within Set Sum of Squared Error = \" + str(WSSSE))\n\n#print(parsedData.collect()[1])","sub_path":"reishi_batch/pyspark/kmeans.py","file_name":"kmeans.py","file_ext":"py","file_size_in_byte":1126,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"463006172","text":"#!/usr/bin/env python3\n\nimport sys\nimport os.path\n\ndef bankWithMostBlocks(banks):\n return sorted(list(banks.items()), key=lambda e: e[1], reverse=True)[0][0]\n\ndef main(data):\n dataLength = len(data)\n configurations = set()\n configuration = tuple(list(data.values()))\n\n while configuration not in configurations:\n configurations.add(configuration)\n bankIdx = bankWithMostBlocks(data)\n blocksToRedistribute = data[bankIdx]\n data[bankIdx] = 0\n while blocksToRedistribute > 0:\n bankIdx = (bankIdx + 1) % dataLength\n data[bankIdx] += 1\n blocksToRedistribute -= 1\n configuration = tuple(list(data.values()))\n\n counter = 0\n configTarget = configuration\n\n while True:\n if configuration == configTarget and counter > 0:\n break\n counter += 1\n bankIdx = bankWithMostBlocks(data)\n blocksToRedistribute = data[bankIdx]\n data[bankIdx] = 0\n while blocksToRedistribute > 0:\n bankIdx = (bankIdx + 1) % dataLength\n data[bankIdx] += 1\n blocksToRedistribute -= 1\n configuration = tuple(list(data.values()))\n\n return counter\n\nif __name__ == '__main__':\n testVectors = [\n ({0: 0, 1: 2, 2: 7, 3: 0}, 4),\n ]\n\n testResults = [ main(ti[0]) == ti[1] for ti in testVectors ]\n if all(testResults):\n print(\"All tests passed!\")\n inputFile = os.path.join(\n os.path.dirname(\n os.path.dirname(\n os.path.abspath(sys.argv[0])\n )\n ), 'input'\n )\n if os.path.isfile(inputFile):\n with open(inputFile, 'r') as fh:\n inputData = fh.read()\n inputData = inputData.strip()\n inputData = { i: int(e) for i, e in enumerate( inputData.split() ) }\n print(main(inputData))\n else:\n print(\"Input file not found\")\n else:\n print(testResults)\n print(main({0: 0, 1: 2, 2: 7, 3: 0}))\n","sub_path":"06/python/AoC_6b.py","file_name":"AoC_6b.py","file_ext":"py","file_size_in_byte":2040,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"15956630","text":"\"\"\"\nhttps://www.qt.io/qt-for-python\npyside2 is official interface\n\nsudo pacman -S pyside2\n\n\"\"\"\n\nfrom PySide2 import QtCore, QtWidgets, QtGui\nimport random\nimport sys\n\n\nclass MyWidget(QtWidgets.QWidget):\n def __init__(self):\n QtWidgets.QWidget.__init__(self)\n self.hello = ['a', 'b', 'c']\n self.button = QtWidgets.QPushButton('click me')\n self.text = QtWidgets.QLabel('hello world')\n self.text.setAlignment(QtCore.Qt.AlignCenter)\n self.layout = QtWidgets.QVBoxLayout()\n self.layout.addWidget(self.text)\n self.layout.addWidget(self.button)\n self.setLayout(self.layout)\n self.button.clicked.connect(self.magic)\n\n def magic(self):\n self.text.setText(random.choice(self.hello))\n\n\nif __name__ == '__main__':\n app = QtWidgets.QApplication(sys.argv)\n widget = MyWidget()\n widget.show()\n sys.exit(app.exec_())\n","sub_path":"default_qt.py","file_name":"default_qt.py","file_ext":"py","file_size_in_byte":896,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"139175236","text":"import random\nimport time\n\n# check if target in muyList\ndef inList(myList, target): \n for i in myList: \n if (i == target): \n return True\n return False\n\n# print a multiplication between i and j \ndef printQuestion(i, j):\n print(\"What is\", i, \" x \", j, \" ?\") \n\n\n# print possible answers \ndef printOptions(myList): \n for i in range(len(myList)): \n print(chr((ord('a') + i)), myList[i])\n\n\ndef answerList(number, myList): \n result = []\n for i in myList: \n result.append(number * i)\n return result\n\n# return a list of possible answer \ndef createOption(i, j): \n result = []\n result.append(i*j)\n for duration in range(3):\n temp = random.randint(1,9)\n\n while(True):\n temp = random.randint(1,9)\n if inList(result, temp):\n continue\n break\n result.append(i * temp) \n return result \n\ndef checkAnswer(myAnswerIndex, myList, number, target):\n if myList[myAnswerIndex] == number * target:\n return True\n return False\n\n\n\n\nprint(\"Welcome to Multiplication Game\")\nprogress = [ [0 for i in range(9)] for j in range(9)] \nwhile(True):\n number = input(\"Which number do you want to learn? \")\n number = int(number)\n if number <0 or number > 9: \n print(\"Please input number from 1 to 9\")\n continue\n break \ntime.sleep(1)\n\nprint(\"Are you ready?\")\nfromZeroToNine = list(range(0,9))\nrandom.shuffle(fromZeroToNine)\nfor i in fromZeroToNine:\n options = createOption(number, i)\n random.shuffle(options)\n printQuestion(number, i)\n printOptions(options)\n your_answer= input(\"What is the correct answer? a, b, c ,d\")\n your_answer_index = ord(your_answer) - ord('a')\n #print (\"*****\",your_answer_index)\n if (checkAnswer(your_answer_index, options, number, i)):\n print(your_answer ,\"That's the correct answer \")\n else:\n print(your_answer, \"Sorry, This is not correct\")\n print(\"***************************\")\n print()\n\n\n\n\n\n\n\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2003,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"243762456","text":"\"\"\"\nMaximisation step: Fit a component to the data: Find parameters of the\ncomponent at birth time that fit the data today.\n\nThis requires the component propagation in time.\n\n\"\"\"\n\nimport numpy as np\nimport scipy.optimize\n\n#~ from chronostar import likelihood2\nfrom chronostar.component import SphereComponent\n\ntry:\n #TODO age consideration parameter from .pars file\n age_parameter=False\n if age_parameter:\n from chronostar.likelihood import lnprob_func_gradient_descent\n else:\n from chronostar._likelihood import lnprob_func_gradient_descent\n #TODO from NS, what is _likelihood compared to likelihood.py? \n # Will _likelihood go around the likelihood_w_ages, leaving the age \n # consideration out of the mximisation step?\nexcept ImportError:\n print(\"C IMPLEMENTATION OF lnprob_func_gradient_descent NOT IMPORTED\")\n USE_C_IMPLEMENTATION = False\n TODO = True # NOW WHAT?\n\ndef fit_single_comp_gradient_descent_serial(data, memb_probs=None, \n init_pars=None, Component=SphereComponent, \n convergence_tol=1, optimisation_method='Nelder-Mead'):\n \"\"\"\n Write docs...\n \"\"\"\n\n #~ print('init_pars', init_pars)\n\n init_age = init_pars[-1]\n age_offsets = [-9, -4, -0.4, -0.2, -0.5, 0., 0.1, 0.3, 0.5, 5., 10., 20., 40.]\n #~ age_offsets = [-9, -4, -0.4, -0.2, -0.5, 0., 0.1, 0.3, 0.5, 5., 10., 20., 40.]\n #~ age_offsets = [0., 10.] # for testing\n init_ages = np.abs([init_age + age_offset for age_offset in age_offsets])\n init_guess_comp = Component(emcee_pars=init_pars)\n # Age split hardcoded!\n init_guess_comps = init_guess_comp.split_group_ages(init_ages)\n init_pos = [c.get_emcee_pars() for c in init_guess_comps]\n \n\n # Prepare data: exclude non-members\n # This is the required C format\n a = []\n memb_threshold=1e-5\n nearby_star_mask = np.where(memb_probs > memb_threshold)\n for i in nearby_star_mask[0]:\n tmp = np.hstack((data['means'][i], data['covs'][i].flatten(), memb_probs[i]))\n a.append(tmp)\n a=np.array(a)\n\n return_dict={}\n for i in range(len(init_pos)):\n #~ print('init_pos[%d]'%i, init_pos)\n #~ print('optimisation_method', optimisation_method)\n result = scipy.optimize.minimize(lnprob_func_gradient_descent, \n init_pos[i], args=a, \n tol=convergence_tol, method=optimisation_method)\n return_dict[i] = result\n\n\n keys = list(return_dict.keys()) # Keep the keys so you always have the same order\n result_fun = [[k, return_dict[k].fun] for k in keys]\n result_fun_sorted = sorted(result_fun, key=lambda x: x[1])\n best_key = result_fun_sorted[0][0]\n best_result = return_dict[best_key]\n\n # Identify and create the best component (with best lnprob)\n best_component = Component(emcee_pars=best_result.x)\n \n return best_component, best_result.x, -best_result.fun # Check if really minus. Minus is already in the likelihood...\n\n\ndef maximisation_gradient_descent_serial(data, ncomps=None, \n memb_probs=None, all_init_pars=None, all_init_pos=None,\n convergence_tol=1, Component=SphereComponent,\n optimisation_method='Nelder-Mead', \n idir=None):\n \"\"\"\n What is idir?\n \"\"\"\n \n \"\"\"\n I get lots pf pickling errors.\n \n From stackoverflow:\n \n The multiprocessing module has a major limitation when it comes to IPython use:\n\n Functionality within this package requires that the __main__ module be importable by the children. [...] This means that some examples, such as the multiprocessing.pool.Pool examples will not work in the interactive interpreter. [from the documentation]\n\n Fortunately, there is a fork of the multiprocessing module called multiprocess which uses dill instead of pickle to serialization and overcomes this issue conveniently.\n\n Just install multiprocess and replace multiprocessing with multiprocess in your imports\n \"\"\"\n \n return_dict={}\n for i in range(ncomps):\n best_comp, final_pos, lnprob =\\\n fit_single_comp_gradient_descent_serial(data=data, \n memb_probs=memb_probs[:, i],\n convergence_tol=convergence_tol,\n #~ init_pos=all_init_pos[i],\n init_pars=all_init_pars[i], Component=Component,\n #~ trace_orbit_func=trace_orbit_func,\n optimisation_method=optimisation_method, # e.g. Nelder-Mead\n )\n return_dict[i] = [best_comp, lnprob, final_pos]\n\n\n\n new_comps_list = [return_dict[i][0] for i in range(ncomps)]\n all_lnprob = [return_dict[i][1] for i in range(ncomps)]\n all_final_pos = [return_dict[i][2] for i in range(ncomps)]\n\n return new_comps_list, all_lnprob, all_final_pos\n\n\ndef maximisation_gradient_descent_multiprocessing(data, ncomps=None, \n memb_probs=None, all_init_pars=None, all_init_pos=None,\n convergence_tol=1, Component=SphereComponent,\n optimisation_method='Nelder-Mead', \n idir=None):\n \"\"\"\n What is idir?\n \"\"\"\n \n \"\"\"\n I get lots pf pickling errors.\n \n From stackoverflow:\n \n The multiprocessing module has a major limitation when it comes to IPython use:\n\n Functionality within this package requires that the __main__ module be importable by the children. [...] This means that some examples, such as the multiprocessing.pool.Pool examples will not work in the interactive interpreter. [from the documentation]\n\n Fortunately, there is a fork of the multiprocessing module called multiprocess which uses dill instead of pickle to serialization and overcomes this issue conveniently.\n\n Just install multiprocess and replace multiprocessing with multiprocess in your imports\n \"\"\"\n print('IN maximisation_gradient_descent_multiprocessing')\n \n manager = multiprocessing.Manager()\n return_dict = manager.dict()\n \n #~ global worker # should solve pickle error in ipython, but it doesn't\n \n #~ return_dict={}\n #~ for i in range(ncomps):\n #~ best_comp, final_pos, lnprob =\\\n #~ fit_single_comp_gradient_descent_serial(data=data, \n #~ memb_probs=memb_probs[:, i],\n #~ convergence_tol=convergence_tol,\n #init_pos=all_init_pos[i],\n #~ init_pars=all_init_pars[i], Component=Component,\n #trace_orbit_func=trace_orbit_func,\n #~ optimisation_method=optimisation_method, # e.g. Nelder-Mead\n #~ )\n #~ return_dict[i] = [best_comp, lnprob, final_pos]\n\n\n\n\n def worker(i, return_dict):\n best_comp, final_pos, lnprob =\\\n fit_single_comp_gradient_descent_serial(data=data, \n memb_probs=memb_probs[:, i],\n convergence_tol=convergence_tol,\n init_pars=all_init_pars[i], Component=Component,\n optimisation_method=optimisation_method, # e.g. Nelder-Mead\n )\n\n return_dict[i] = [best_comp, lnprob, final_pos]\n\n\n\n\n jobs = []\n for i in range(ncomps):\n process = multiprocessing.Process(target=worker, \n args=(i, return_dict))\n jobs.append(process)\n\n # Start the processes\n for j in jobs:\n j.start()\n\n # Ensure all of the processes have finished\n for j in jobs:\n j.join()\n\n\n print('END maximisation_gradient_descent_multiprocessing')\n\n\n\n new_comps_list = [return_dict[i][0] for i in range(ncomps)]\n all_lnprob = [return_dict[i][1] for i in range(ncomps)]\n all_final_pos = [return_dict[i][2] for i in range(ncomps)]\n\n return new_comps_list, all_lnprob, all_final_pos\n","sub_path":"chronostar/maximisationC.py","file_name":"maximisationC.py","file_ext":"py","file_size_in_byte":7549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"420781356","text":"#!/usr/bin/env python\n\n#CodeWars 2014\n#\n#Goldbach's Conjecture\n\n# Goldbach's conjecture says that every positive even number greater than 2 is \n# the sum of two prime numbers. This conjecture has never been proven in the \n# general case, but it has been confirmed for numbers much larger than \n# your programming environment's native data type supports.\n#\n# Write a program to print the two prime numbers that sum to a given even integer.\n# Find the two primes with the smallest difference\n#\n# Input\n#\n# Each line of input will be a positive, even integer greater than two, \n# except the last line, which will be zero. The maximum input value will be 1000.\n#\n# 28\n# 992\n# 16\n# 0\n# \n\nimport sys\nfrom math import sqrt\n\ndef isprime(value):\n for divisor in range (2, int(sqrt(value))+1):\n if (value/divisor == int(value/divisor)):\n return 0\n return 1\n\nprint (\"Enter even numbers. 0 to end.\")\nfor line in sys.stdin:\n total = int(line)\n lowestDiff=total\n if (total == 0):\n break\n for low in range(3, int((total/2)+1)):\n if (isprime(low)==1):\n high = total-low\n if (isprime(high)==1):\n # Found an answer. Update if lowest answer is found.\n if (high-low < lowestDiff):\n lowestDiff = high-low\n bestLow = low\n bestHigh = high\n print (bestLow, \"+\", bestHigh, \"=\", total)\n\n","sub_path":"codewars/2014/SampleSolutions2014/prob09_Goldbach.py","file_name":"prob09_Goldbach.py","file_ext":"py","file_size_in_byte":1425,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"246839822","text":"import visualization\nimport xyPlot\nimport displayGroupOdbToolset as dgo\nimport os, glob, inspect\nimport sys\nsys.path.append('C:\\\\Users\\\\doktor\\\\PycharmProjects\\\\InputFilesBuild - Copy')\nimport globalPar as gp\nreload(gp)\n\npath = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))) +'\\\\workingDirectory\\\\'\nos.chdir(path)\n\ngp.scandirs(path)\n\n#odbKeysList = ['SDV11', 'SDV_EE11', 'SDV_EE12', 'SDV_EE13', 'SDV_EE22', 'SDV_EE23', 'SDV_EE33', 'SDV_EP11', 'SDV_EP12',\n# 'SDV_EP13', 'SDV_EP22', 'SDV_EP33', 'SDV_PEEQ', 'SDV_TEP']\n\n\npc = gp.ParametersClass()\nsemicolon =';'\ncomma = ','\n\n# Header\ndispFile_to_view = open('db_to_view_FFT.csv', 'a')\nfile = glob.glob(\"*.odb\")[0]\no1 = session.openOdb(name=path + file)\nodb = session.odbs[path + file]\nstep = odb.steps['Step-1']\n\nodbLabels = gp.get_labels(step)\n\n#Header - only on file initialization\ndispFile_to_view.write('id' + semicolon + 'delta' + semicolon + 'alpha' + semicolon + 'velocity' + semicolon + 'time'\n + semicolon + 's' + semicolon + 'Tm' + semicolon + 'm')\n#Header - only on file initialization\n\nfor key in odbLabels.keys():\n if not odbLabels[key]['labels']:\n dispFile_to_view.write(semicolon + str(key))\n #dispFile.write(comma + str(key))\n else:\n for label in odbLabels[key]['labels']:\n dispFile_to_view.write(semicolon + str(label))\n #dispFile.write(comma + str(label))\nodb.close()\n\n\nfor file in glob.glob(\"*.odb\"):\n print(file)\n o1 = session.openOdb(name=path + file)\n odb = session.odbs[path + file]\n step = odb.steps['Step-1']\n\n #odbLabels = gp.get_labels(step)\n\n # Data rows\n alpha, Dxx = gp.get_additional_parameters(file)\n print(alpha)\n time = pc.s / pc.velocity ##velocity in mm/s\n\n for id, frame in enumerate(step.frames):\n dispFile_to_view.write('\\n' + str(id) + semicolon + str(Dxx) + semicolon + str(alpha) + semicolon +\n str(pc.velocity) + semicolon + str(time) + semicolon + str(pc.s) + semicolon + str(pc.Tm) + semicolon + str(pc.m))\n\n for key in odbLabels.keys():\n if not odbLabels[key]['labels']:\n dispFile_to_view.write(semicolon + str(frame.fieldOutputs[key].values[0].data))\n\n else:\n for label in odbLabels[key]['labels']:\n if key == 'S':\n dispFile_to_view.write(semicolon + str(frame.fieldOutputs[key].getScalarField(componentLabel=label).values[0].data))\n else:\n dispFile_to_view.write(semicolon + str(frame.fieldOutputs[key].getScalarField(componentLabel=label).values[7].data))\n odb.close()\n\ndispFile_to_view.close()","sub_path":"OLD_code/OLD_3_1_abaqus_OdbReader.py","file_name":"OLD_3_1_abaqus_OdbReader.py","file_ext":"py","file_size_in_byte":2717,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"616397321","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Aug 2 00:36:13 2015\n\n@author: Евгений\n\"\"\"\n\ndef push_annual(cursor, var_name, year, val):\n cursor.execute(\"INSERT OR REPLACE INTO annual VALUES (?, ?, ?)\", (var_name, year, val))\n\ndef push_quarter(cursor, var_name, year, quarter, val):\n cursor.execute(\"INSERT OR REPLACE INTO quarter VALUES (?, ?, ?, ?)\", (var_name, year, quarter, val))\n\ndef push_monthly(cursor, var_name, year, month, val):\n cursor.execute(\"INSERT OR REPLACE INTO monthly VALUES (?, ?, ?, ?)\", (var_name, year, month, val))\n\n\nimport sqlite3\nconn = sqlite3.connect('kep.sqlite')\nc = conn.cursor()\n\n# Insert a row of data\nc.execute(\"INSERT OR REPLACE INTO annual VALUES (?, ?, ?)\", ('153', 1000, 0))\n\nc.executescript(\"\"\"\nDELETE FROM \"main\".\"quarterly\";\nDELETE FROM \"main\".\"monthly\";\nDELETE FROM \"main\".\"annual\";\n\"\"\")\n\n# Save (commit) the changes\nconn.commit()\n\n# We can also close the connection if we are done with it.\n# Just be sure any changes have been committed or they will be lost.\n\n\n\n\nimport sqlite3\n\ndef wipe_db_tables():\n conn = sqlite3.connect('kep.sqlite')\n c = conn.cursor()\n c.executescript(\"\"\"\n DELETE FROM \"main\".\"quarterly\";\n DELETE FROM \"main\".\"monthly\";\n DELETE FROM \"main\".\"annual\";\n \"\"\")\n conn.commit()\n conn.close()","sub_path":"src/abandoned/db.py","file_name":"db.py","file_ext":"py","file_size_in_byte":1292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"87"}
+{"seq_id":"451759151","text":"#!/bin/env python2.7\n\"\"\"\nrun_w2v_pipeline.py\n\nA script for running all of the analysis in the Word2Vec/Kmeans\npipeline and saving results to a given directory path.\nThis may also be used as an example if you wish to do it by hand.\n\nEstimated running time for the whole dataset is on the order of 8 hours\nas of March 2015\n\"\"\"\nimport os\nthis_dir, _ = os.path.split(__file__)\nimport sys\nsys.path.append('/'.join(['../../', this_dir])) # Add the previous directory to the path (just a clooj for now)\nimport logging\nimport datetime\nimport numpy as np\nlogging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO) #Set up logging\nfrom pymongo import MongoClient\nfrom pprint import pprint, pformat\nfrom sklearn.externals import joblib\n\nfrom jmAlife.txtMine import * # get my utility files. Automatically loads stoplist into the namespace. \n\n## GLOBAL VARIABLES\nc = MongoClient()\n\ndef main(coll, num_pats, model_type, vec_size, num_clusters, phrase_size, dest_dir):\n # Local Vars\n start = datetime.datetime.now()\n USE_GENERATORS = False\n \n\n # Get the appropriate patent cursor\n pats = c.patents.pat_text.find()\n if coll == \"patns\":\n pats = c.patents.patns.find()\n elif coll == \"pat_text\":\n pass\n else: \n sys.exit(\"%s collection not currently supported.\" %coll)\n\n TOTAL_PATS = pats.count()\n N = TOTAL_PATS\n\n # If input number of pats appropriate, take that number from the cursor.\n \"\"\"\n if num_pats < N and num_pats > 0:\n N = num_pats\n if N > 500000:\n USE_GENERATORS = True\n \"\"\"\n\n if model_type != \"d2v\" and model_type != \"w2v\":\n sys.exit(\"This model not currently supported. Enter 'W2V' for word2vec and 'D2V' for doc2vec\")\n\n logging.info(\"Running your %s/cluster job with vec_size %d and %d clusters. Preprocessing %d patents from %s with %d-grams\" \n %(model_type, vec_size, num_clusters, N, coll, phrase_size))\n \n if N != TOTAL_PATS:\n pats = util.take(N, pats)\n\n # Get gensim objects\n logging.info(\"Getting %d sentences...\" %N)\n ls = util.quick_labeled_sentences(list(pats), stoplist, False, phrase_size)\n logging.info(\"Length of labeled sentences is {}\".format(len(ls)))\n \n logging.info(\"Fitting %s model on %d patents...\" %(model_type, N))\n if model_type == \"d2v\":\n # Run d2v analysis and get word vectors\n model = util.get_d2v(ls, vec_size = vec_size, workers = 1)\n word_vecs, doc_vecs, indices = util.partition(model)\n word2index, doc2index = indices\n elif model_type == \"w2v\":\n # run w2v analysis and get word_vectors\n sentences = map(lambda x: x.words, ls)\n logging.info(\"In line before word2vec is fit, sentences has length {}\".format(len(sentences)))\n model = util.get_w2v(sentences, vec_size, workers = 1)\n word_vecs = model.syn0\n word2index = {word: index for (index, word) in enumerate(model.index2word)}\n else:\n raise RuntimeError(\"Internal Error. Invalid model_type %s\" %model_type)\n \n # Get Filenames\n model_ext = '.word2vec'\n if model_type == 'd2v':\n model_ext = '.doc2vec'\n model_fn = '/'.join([dest_dir, ''.join([model_type,str(vec_size),model_ext])])\n kmeans_fn = '/'.join([dest_dir, ''.join(['kmeans', str(num_clusters), '.pkl'])])\n cluster_parse_fn = '/'.join([dest_dir, 'parsed_clusters.npy'])\n\n logging.info(\"Saving Model to disk in %s\" %dest_dir)\n model.save(model_fn)\n\n # run cluster analysis\n logging.info(\"Getting %d clusters for %d words...\" %(num_clusters, len(word_vecs)))\n kmeans = util.get_kmeans(word_vecs, K = num_clusters)\n parsed_clusters = util.parse_clusters(kmeans, model)\n logging.info(\"Parsed Clusters: %s\" %pformat(parsed_clusters))\n\n # TODO - Get document representations and save them to disk too. \n\n # save to disk\n logging.info(\"Saving Kmeans to disk in %s\" %dest_dir)\n joblib.dump(kmeans, kmeans_fn)\n np.save(cluster_parse_fn, parsed_clusters)\n\n end = datetime.datetime.now()\n logging.info(\"Done. Time elapsed: %s\"%util.strfdelta(end - start, \"{days}d,{hours}h,{minutes}m,{seconds}s\"))\n\nif __name__ == \"__main__\":\n # Handle Command Line Args\n if len(sys.argv) != 8:\n sys.exit(\"Usage: python %s <'W2V' or 'D2V'>