diff --git "a/1476.jsonl" "b/1476.jsonl" new file mode 100644--- /dev/null +++ "b/1476.jsonl" @@ -0,0 +1,459 @@ +{"seq_id": "564975218", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('item', '0004_media_vraag'),\n ]\n\n operations = [\n migrations.AlterModelOptions(\n name='afbeelding',\n options={'verbose_name_plural': 'Afbeeldingen'},\n ),\n migrations.AlterModelOptions(\n name='media',\n options={'verbose_name_plural': 'Media'},\n ),\n migrations.AlterModelOptions(\n name='vraag',\n options={'verbose_name_plural': 'Vragen'},\n ),\n migrations.AlterField(\n model_name='media',\n name='beschrijving',\n field=models.TextField(),\n ),\n migrations.AlterField(\n model_name='vraag',\n name='antwoord',\n field=models.TextField(),\n ),\n migrations.AlterField(\n model_name='vraag',\n name='vraag',\n field=models.TextField(),\n ),\n ]\n", "sub_path": "item/migrations/0005_auto_20150620_0719.py", "file_name": "0005_auto_20150620_0719.py", "file_ext": "py", "file_size_in_byte": 1063, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterModelOptions", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterModelOptions", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterModelOptions", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 26, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 26, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 29, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 29, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 36, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 36, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 39, "usage_type": "name"}]} +{"seq_id": "262909989", "text": "import sys\nimport wia\nfrom gpiozero import LightSensor, Buzzer\nimport logging\n\nlogging.getLogger(\"requests\").setLevel(logging.WARNING)\n\n# API endpoints and keys\nwia.secret_key = \"\"\nwia_sensor_name = \"brightness_sensor\"\n\n# Initialize a new light sensor connected to the fourth GPIO pin\nlight_sensor = LightSensor(4)\n\ntry:\n while True:\n # Identify as the light sensor and publish the current value to the IoT platform\n wia.Sensor.publish(\n name = wia_sensor_name,\n data = \"{0}\".format(1 - light_sensor.value)\n )\nexcept KeyboardInterrupt:\n print(\"Interrupted by user, shutting down.\")\n sys.exit(0)\n", "sub_path": "wia.io/raspberry-pi.py", "file_name": "raspberry-pi.py", "file_ext": "py", "file_size_in_byte": 663, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.WARNING", "line_number": 6, "usage_type": "attribute"}, {"api_name": "wia.secret_key", "line_number": 9, "usage_type": "attribute"}, {"api_name": "gpiozero.LightSensor", "line_number": 13, "usage_type": "call"}, {"api_name": "wia.Sensor.publish", "line_number": 18, "usage_type": "call"}, {"api_name": "wia.Sensor", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "159759736", "text": "# Copyright (C) 2009 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may not\n# use this file except in compliance with the License. You may obtain a copy of\n# the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations under\n# the License.\n\n__author__ = [\n 'Damon Kohler ',\n 'Naranjo Manuel Francisco '\n]\n\nimport collections\nimport json\nimport os\nimport socket\nimport sys\n\nPORT = os.environ.get('AP_PORT')\nHOST = os.environ.get('AP_HOST')\nHANDSHAKE = os.environ.get('AP_HANDSHAKE')\nResult = collections.namedtuple('Result', 'id,result,error')\n\nclass _Android(object):\n def __init__(self, addr=None, debug=False):\n if addr is None:\n addr = HOST, PORT\n self.conn = socket.create_connection(addr)\n self.client = self.conn.makefile()\n self.id = 0\n self.debug = debug\n if HANDSHAKE is not None:\n self._authenticate(HANDSHAKE)\n\n def _rpc(self, method, *args):\n data = {'id': self.id,\n 'method': method,\n 'params': args}\n request = json.dumps(data)\n if self.debug and method != \"log\" and not method.startswith(\"_\"):\n self.log(\"call to %s, params: %s\" % (method, args))\n\n self.client.write(request+'\\n')\n self.client.flush()\n response = self.client.readline()\n self.id += 1\n result = json.loads(response)\n\n if self.debug and method != \"log\" and not method.startswith(\"_\"):\n self.log(str(result))\n\n if result['error'] is not None:\n raise Exception(result['error'])\n # we want to expose the result, not our internals\n return result[\"result\"]\n\n def __getattr__(self, name):\n def rpc_call(*args):\n return self._rpc(name, *args)\n return rpc_call\n\n_Android.reference = None\n\ndef API(addr=None, debug=False):\n if _Android.reference == None:\n # make it singleton\n _Android.reference = _Android(addr, debug)\n return _Android.reference\n", "sub_path": "AIRcam/resources/android.py", "file_name": "android.py", "file_ext": "py", "file_size_in_byte": 2222, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.environ.get", "line_number": 26, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 28, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 28, "usage_type": "attribute"}, {"api_name": "collections.namedtuple", "line_number": 29, "usage_type": "call"}, {"api_name": "socket.create_connection", "line_number": 35, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 54, "usage_type": "call"}]} +{"seq_id": "277091576", "text": "import asyncio\nimport time\nfrom concurrent.futures import ThreadPoolExecutor\n\n\ndef download_img(url):\n print(f\"下载图片:{url}\")\n time.sleep(1)\n print(f\"下载完成:{url}\")\n\n\nasync def main():\n executor = ThreadPoolExecutor(2)\n\n loop = asyncio.get_running_loop()\n tasks = []\n for i in range(10):\n t = loop.run_in_executor(executor, download_img, i)\n tasks.append(t)\n\n await asyncio.wait(tasks)\n\n\nasyncio.run(main())\n", "sub_path": "asyncio/youtuoo/async_demos/s1/s19_2.py", "file_name": "s19_2.py", "file_ext": "py", "file_size_in_byte": 462, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "time.sleep", "line_number": 8, "usage_type": "call"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 13, "usage_type": "call"}, {"api_name": "asyncio.get_running_loop", "line_number": 15, "usage_type": "call"}, {"api_name": "asyncio.wait", "line_number": 21, "usage_type": "call"}, {"api_name": "asyncio.run", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "599610583", "text": "\"\"\"add hidden to product\n\nRevision ID: 9d022c4b57e3\nRevises: f92bf82ceb1d\nCreate Date: 2020-05-27 15:07:24.119142\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = '9d022c4b57e3'\ndown_revision = 'f92bf82ceb1d'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n op.add_column('products', sa.Column('hidden', sa.Boolean(), nullable=True, default=False))\n op.execute(\"UPDATE products SET hidden = false\")\n op.alter_column('products', 'hidden', nullable=False)\n\n\ndef downgrade():\n op.drop_column('products', 'hidden')\n", "sub_path": "src/backend/alembic/versions/2020-05-27_9d022c4b57e3_add_hidden_to_product.py", "file_name": "2020-05-27_9d022c4b57e3_add_hidden_to_product.py", "file_ext": "py", "file_size_in_byte": 588, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "alembic.op.add_column", "line_number": 20, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 20, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 20, "usage_type": "call"}, {"api_name": "sqlalchemy.Boolean", "line_number": 20, "usage_type": "call"}, {"api_name": "alembic.op.execute", "line_number": 21, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 21, "usage_type": "name"}, {"api_name": "alembic.op.alter_column", "line_number": 22, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 22, "usage_type": "name"}, {"api_name": "alembic.op.drop_column", "line_number": 26, "usage_type": "call"}, {"api_name": "alembic.op", "line_number": 26, "usage_type": "name"}]} +{"seq_id": "273717880", "text": "from flask import Flask, render_template, jsonify\nfrom flask_sockets import Sockets\nimport docker\nfrom docker.errors import NotFound\nimport time\nimport configure \nfrom thread_send import threadSend\n\n\napp = Flask(__name__)\nsockets = Sockets(app)\ndocker_client = docker.APIClient(base_url=configure.DOCKER_HOST,\n version=configure.DOCKER_API_VERSION,timeout=configure.TIME_OUT)\n\n@app.route('/containers')\ndef containers():\n return jsonify(docker_client.containers())\n\n\n@app.route('/console/')\ndef console(container_id):\n return render_template('index.html', container_id=container_id)\n\ndef create_exec(container_id):\n command = [\"/bin/sh\",\"-c\",'TERM=xterm-256color; export TERM; [ -x /bin/bash ] && ([ -x /usr/bin/script ] && /usr/bin/script -q -c \"/bin/bash\" /dev/null || exec /bin/bash) || exec /bin/sh']\n create_exec_options = {\n \"tty\": True,\n \"stdin\": True,\n }\n exec_id = docker_client.exec_create(container_id, command, **create_exec_options)\n return exec_id\n\n@sockets.route('/echo/')\ndef echo_socket(ws, container_id):\n try:\n exec_id = create_exec(container_id)\n sock = docker_client.exec_start(exec_id, detach=False, tty=True, stream=False,\n socket=True)\n sock.settimeout(600)\n send = threadSend(ws,sock)\n send.start()\n while not ws.closed:\n message = ws.receive()\n if message is not None:\n sock.send(message)\n except NotFound:\n ws.send(\"not fund container[%s].\" % container_id)\n\nif __name__ == '__main__':\n from gevent import pywsgi\n from geventwebsocket.handler import WebSocketHandler\n server = pywsgi.WSGIServer(('', 5000), app, handler_class=WebSocketHandler)\n server.serve_forever()\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1785, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_sockets.Sockets", "line_number": 11, "usage_type": "call"}, {"api_name": "docker.APIClient", "line_number": 12, "usage_type": "call"}, {"api_name": "configure.DOCKER_HOST", "line_number": 12, "usage_type": "attribute"}, {"api_name": "configure.DOCKER_API_VERSION", "line_number": 13, "usage_type": "attribute"}, {"api_name": "configure.TIME_OUT", "line_number": 13, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "thread_send.threadSend", "line_number": 40, "usage_type": "call"}, {"api_name": "docker.errors.NotFound", "line_number": 46, "usage_type": "name"}, {"api_name": "gevent.pywsgi.WSGIServer", "line_number": 52, "usage_type": "call"}, {"api_name": "gevent.pywsgi", "line_number": 52, "usage_type": "name"}, {"api_name": "geventwebsocket.handler.WebSocketHandler", "line_number": 52, "usage_type": "name"}]} +{"seq_id": "634299510", "text": "# -*- coding: utf-8 -*-\n\n# __title__ = 'parse_url.py'\n# __author__ = 'YangYang'\n# __mtime__ = '2018.12.17'\n# 此类还可以继续优化\nfrom django.urls import reverse\nfrom django.http import QueryDict\n\nclass ParseUrl(object):\n \"\"\"\n 保留原搜索条件\n \"\"\"\n def __init__(self,request,namespace,name,*args,**kwargs):\n \"\"\"\n \n :param request: URL 的request\n :param namespace: URL 的namespace\n :param name: URL的别名\n :param args: 默认传参\n :param kwargs: 默认传参\n \"\"\"\n self.request = request\n self.namespace= namespace\n self.name = name\n self.args = args\n self.kwargs = kwargs\n self.nameinfo = \"%s:%s\"%(self.namespace,self.name)\n\n def memory_reverse_url(self):\n \"\"\"\n 保留搜索的url参数到新的页面\n :return:\n \"\"\"\n base_url = reverse(self.nameinfo,args=self.args,kwargs=self.kwargs)\n if not self.request.GET:\n url = base_url\n else:\n param = self.request.GET.urlencode()\n new_query_dict = QueryDict(mutable=True)\n new_query_dict['_filter'] = param\n url = \"%s?%s\"%(base_url,new_query_dict.urlencode())\n\n return url\n\n def memory_url(self):\n \"\"\"\n 页面提交以后返回原始搜索页面\n :return:\n \"\"\"\n base_url = reverse(self.nameinfo, args=self.args, kwargs=self.kwargs)\n params = self.request.GET.get(\"_filter\")\n if not params:\n return base_url\n url = \"%s?%s\"%(base_url,params)\n return url\n\n\n\n\n\n", "sub_path": "seventh_module/CRM/30.rbac以及stark个人总结/6.parse_url.py", "file_name": "6.parse_url.py", "file_ext": "py", "file_size_in_byte": 1630, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.urls.reverse", "line_number": 35, "usage_type": "call"}, {"api_name": "django.http.QueryDict", "line_number": 40, "usage_type": "call"}, {"api_name": "django.urls.reverse", "line_number": 51, "usage_type": "call"}]} +{"seq_id": "127813336", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n#Source: variables.py\n\nfrom pykeyboard import PyKeyboard\nimport serial\nimport os\n\nk = PyKeyboard()\nser = serial.Serial(None)\n\nwindow_height = 300\nwindow_width = 300\n\nopened = True\n\nsymbols = {'1' : u'π', #6 button version\n '2' : u'Σ',\n '3' : u'α',\n '4' : u'β',\n '5' : u'Δ',\n '6' : u'Ω'}\n \n#symbols = {'1' : 'π', #12 button version\n #'2' : 'Σ',\n #'3' : 'α',\n #'4' : 'β',\n #'5' : 'Δ',\n #'6' : 'Ω',\n #'7' : 'test',\n #'8' : 'another',\n #'9' : 'and again',\n #'10' : 'last one',\n #'11' : 'ayy',\n #'12' : 'lmao'}\n\nrows = 2 #6 button version\ncolumns = 3\n\n#rows = 3 #12 button version\n#columns = 4\n\ncurrent_profile = 0\n\npref_file = os.path.abspath('Quick-Keys Preferences')\nicon_file = os.path.abspath('icon.png')\n", "sub_path": "Quick-Keys-Unicode-Revision/variables.py", "file_name": "variables.py", "file_ext": "py", "file_size_in_byte": 971, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pykeyboard.PyKeyboard", "line_number": 10, "usage_type": "call"}, {"api_name": "serial.Serial", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path", "line_number": 46, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}]} +{"seq_id": "288160315", "text": "import logging\n\nimport cwt\n\nfrom celery import chord\nfrom celery import group\nfrom celery import signature\nfrom wps import tasks\nfrom wps.backends import backend\nfrom wps.tasks import base\n\n__ALL__ = ['Local']\n\nlogger = logging.getLogger('wps.backends')\n\nclass Local(backend.Backend):\n def initialize(self):\n pass\n\n def populate_processes(self):\n logger.info('Registering processes for backend \"local\"')\n\n for name, proc in base.REGISTRY.iteritems():\n self.add_process(name, name.title(), proc.ABSTRACT)\n\n def execute(self, identifier, variables, domains, operations, **kwargs):\n if len(operations) == 0:\n raise Exception('Must supply atleast one operation')\n\n operation = operations.values()[0].parameterize()\n\n variable_dict = dict((x, y.parameterize()) for x, y in variables.iteritems())\n\n domain_dict = dict((x, y.parameterize()) for x, y in domains.iteritems())\n\n target_process = base.get_process(identifier)\n\n logger.info('Retrieved process \"{}\"'.format(identifier))\n\n job = kwargs.get('job')\n\n user = kwargs.get('user')\n\n params = {\n 'job_id': job.id,\n 'user_id': user.id,\n }\n\n logger.info('Variables {}'.format(variable_dict))\n\n logger.info('Domains {}'.format(domain_dict))\n\n logger.info('Operation {}'.format(operation))\n\n return target_process.s({}, variable_dict, domain_dict, operation, **params)\n\n def get_task(self, identifier):\n try:\n task = base.get_process(identifier)\n except:\n if 'CDSpark' in identifier:\n task = tasks.edas_submit\n elif 'Oph' in identifier:\n task = tasks.oph_submit\n\n return task\n\n def workflow(self, root_op, variables, domains, operations, **kwargs):\n job = kwargs.get('job')\n\n user = kwargs.get('user')\n\n params = {\n 'job_id': job.id,\n 'user_id': user.id,\n }\n\n global_domains = dict((x, y.parameterize()) for x, y in domains.iteritems())\n\n logger.info('Building workflow')\n\n def _build(node):\n sub_tasks = []\n\n for name in node.inputs:\n if name in operations:\n sub_tasks.append(_build(operations[name]))\n\n task_variables = dict((x, variables[x].parameterize()) \n for x in node.inputs if x in variables)\n\n if len(sub_tasks) == 0:\n task = self.get_task(node.identifier).s(\n {}, task_variables, global_domains, node.parameterize(), **params)\n else:\n task = self.get_task(node.identifier).s(\n task_variables, global_domains, node.parameterize(), **params)\n\n task = group(sub_tasks) | task\n\n return task\n\n return _build(root_op)\n", "sub_path": "compute/wps/backends/local.py", "file_name": "local.py", "file_ext": "py", "file_size_in_byte": 2921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "wps.backends.backend.Backend", "line_number": 16, "usage_type": "attribute"}, {"api_name": "wps.backends.backend", "line_number": 16, "usage_type": "name"}, {"api_name": "wps.tasks.base.REGISTRY.iteritems", "line_number": 23, "usage_type": "call"}, {"api_name": "wps.tasks.base.REGISTRY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "wps.tasks.base", "line_number": 23, "usage_type": "name"}, {"api_name": "wps.tasks.base.get_process", "line_number": 36, "usage_type": "call"}, {"api_name": "wps.tasks.base", "line_number": 36, "usage_type": "name"}, {"api_name": "wps.tasks.base.get_process", "line_number": 59, "usage_type": "call"}, {"api_name": "wps.tasks.base", "line_number": 59, "usage_type": "name"}, {"api_name": "wps.tasks.edas_submit", "line_number": 62, "usage_type": "attribute"}, {"api_name": "wps.tasks", "line_number": 62, "usage_type": "name"}, {"api_name": "wps.tasks.oph_submit", "line_number": 64, "usage_type": "attribute"}, {"api_name": "wps.tasks", "line_number": 64, "usage_type": "name"}, {"api_name": "celery.group", "line_number": 99, "usage_type": "call"}]} +{"seq_id": "159165413", "text": "import datetime\n\nfrom cor.api import CORModule\nfrom sensor_pb2 import SensorReading\nimport threading\nimport time\nimport socket\nimport sys\nimport serial\n\n__author__ = 'denislavrov'\n\nclass SerialReader(CORModule):\n\tdef force_check(self, message):\n\t\tif \"collectd\" in message.payload[\"sensors\"]:\n\t\t\tself.check()\n\n\tdef check(self):\n\t\twith serial.Serial(self.serial_port) as ser:\n\t\t\tcount = 0\n\t\t\tser.readline() # ignore the first line, its a lie\n\t\t\twhile True:\n\t\t\t\tline = ser.readline().decode(\"ascii\").strip()\n\t\t\t\ttry:\n\t\t\t\t\ttemp = float(line)\n\t\t\t\texcept Exception:\n\t\t\t\t\tprint(\"Bad line: \", line)\n\t\t\t\t\tcontinue\n\t\t\t\tif count % 10 == 0:\n\t\t\t\t\treading = SensorReading()\n\t\t\t\t\treading.location = self.location\n\t\t\t\t\treading.timestamp = int(time.time())\n\t\t\t\t\tvals = {\"temperature\":temp}\n\t\t\t\t\treading.values.update(vals)\n\t\t\t\t\tself.messageout(reading)\n\t\t\t\tcount += 1\n\n\tdef __init__(self, location=\"UNKNOWN\", serial_port=\"/dev/cu.usbmodem1421\", *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\t\tself.location = location\n\t\tself.serial_port = serial_port\n\t\tself.t = threading.Thread(target=self.check)\n\t\tself.t.start()\n", "sub_path": "serial_reader.py", "file_name": "serial_reader.py", "file_ext": "py", "file_size_in_byte": 1106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "cor.api.CORModule", "line_number": 13, "usage_type": "name"}, {"api_name": "serial.Serial", "line_number": 19, "usage_type": "call"}, {"api_name": "sensor_pb2.SensorReading", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 42, "usage_type": "call"}]} +{"seq_id": "44381220", "text": "import logging\n\nimport requests\n\nfrom . import app_settings\n\nlog = logging.getLogger(__name__)\nFB_GRAPH_URL = \"https://graph.facebook.com/v10.0\"\n\n\ndef get_facebook_posts():\n fields = \"full_picture,message,created_time,permalink_url,from\"\n url = f\"{FB_GRAPH_URL}/{app_settings.FB_PAGE_ID}/feed?fields={fields}\"\n url += f\"&access_token={app_settings.FB_PAGE_TOKEN}\"\n url += f\"&limit={app_settings.FB_PAGE_FEED_LIMIT}\"\n response = requests.get(url)\n if response.status_code != 200:\n log.debug(response.json())\n return []\n resp_data = response.json()\n posts = []\n for post in resp_data.get(\"data\"):\n from_data = post.get(\"from\")\n posts.append(\n {\n \"from_name\": from_data.get(\"name\"),\n \"message\": post.get(\"message\"),\n \"image\": post.get(\"full_picture\"),\n \"permalink\": post.get(\"permalink_url\"),\n \"created_at\": post.get(\"created_time\"),\n }\n )\n return posts\n", "sub_path": "api/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 1013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 16, "usage_type": "call"}]} +{"seq_id": "35994945", "text": "import io\nimport sys\n\nimport html2text as html2text\nimport krovetzstemmer\nfrom nltk import PorterStemmer\n\n# Instantiate porter stemmer\nporter = PorterStemmer()\nkrovetz = krovetzstemmer.Stemmer()\n\nif len(sys.argv) < 6:\n print('Usage :')\n print('python 4_6.py ... ')\n\n# Assuming all arguments are file\nfiles = []\nfor arg in range(1, len(sys.argv)):\n files.append(sys.argv[arg])\n\n# Get contents of each file\nresults = {}\nfor idx, file in enumerate(files):\n print('{} of {}. Processing {}'.format(idx + 1, len(files), file))\n print('=' * 30)\n\n # get text content\n h = html2text.HTML2Text()\n h.ignore_links = True\n text = h.handle(u' '.join([line.strip() for line in io.open(file, \"r\", encoding=\"utf-8\").readlines()]))\n\n # remove whitespace\n words = []\n for word in text.split():\n if word.isalpha():\n words.append(word.lower())\n text = u' '.join(words)\n\n porter_result = []\n krovetz_result = []\n for c in words:\n porter_result.append(porter.stem(c))\n krovetz_result.append(krovetz.stem(c))\n\n results[file] = {}\n results[file]['original'] = text\n results[file]['porter'] = u' '.join(porter_result)\n results[file]['krovetz'] = u' '.join(krovetz_result)\n\n# print results\ntxt_results = []\nfor file in results:\n txt_results.append(u'Stemmer result of {}'.format(file))\n txt_results.append(u'{}'.format('=' * 60))\n txt_results.append(u'Original text \\t= {}\\n'.format(results[file]['original']))\n txt_results.append(u'Porter result \\t= {}\\n'.format(results[file]['porter']))\n txt_results.append(u'Krovetz result \\t= {}\\n'.format(results[file]['krovetz']))\n\n num_stems_porter = len(set(results[file]['porter'].split()))\n txt_results.append(u'Number of stems produced by Porter \\t= {}\\n'.format(num_stems_porter))\n\n num_stems_krovetz = len(set(results[file]['krovetz'].split()))\n txt_results.append(u'Number of stems produced by Krovetz \\t= {}\\n'.format(num_stems_krovetz))\n txt_results.append(u'\\n')\n\n print(u'\\n'.join(txt_results))\n\n # also write to file\n f = io.open('4_6-result.txt', \"w\", encoding=\"utf-8\")\n for txt_result in txt_results:\n f.write(txt_result + '\\n')\n", "sub_path": "assignments/a2/4_6_rev3.py", "file_name": "4_6_rev3.py", "file_ext": "py", "file_size_in_byte": 2215, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "nltk.PorterStemmer", "line_number": 9, "usage_type": "call"}, {"api_name": "krovetzstemmer.Stemmer", "line_number": 10, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 19, "usage_type": "attribute"}, {"api_name": "html2text.HTML2Text", "line_number": 28, "usage_type": "call"}, {"api_name": "io.open", "line_number": 30, "usage_type": "call"}, {"api_name": "io.open", "line_number": 69, "usage_type": "call"}]} +{"seq_id": "594275871", "text": "#!/usr/bin/env python\n# coding: utf-8\n\nfrom SALib.sample import saltelli\nfrom SALib.analyze import sobol\nimport numpy as np\nimport pandas as pd\n\nn1 = int(input(\"please input n1: \"))\nn2 = int(input(\"please input n2: \"))\nn3 = int(input(\"please input n3: \"))\n\n#====== input parameter ====================#\nnn = np.matrix([[n1],[n2],[n3]]) #cosine vector\nsampling = 5000 # sampling number for saltelli sampling\ntime = 100 # time for ASR measurement\n#================================================#\n\n\ndirc = str(nn[0,0]) + str(nn[1,0])+str(nn[2,0])\nprint(dirc)\n\n\ndef ASR(t, s11, s22, s33, s12, s13, s23, p0, ts, tv):\n #n = np.matrix([[n11,n12,n13,n14,n15,n16, n17, n18, n19],[n21,n22,n23,n24,n25,n26, n27, n28, n29],[n31,n32,n33,n34,n35,n36, n37, n38, n39]])\n K = 50\n G = 30\n \n n = nn / np.linalg.norm(nn)\n\n Jav = (1-np.exp(-t/tv))/(3*K*10**3)\n Jas = (1-np.exp(-t/ts))/(2*G*10**3)\n \n nTn = (s11*n[0]*n[0] + s22*n[1]*n[1] + s33*n[2]*n[2]\n + 2*s12*n[0]*n[1] + 2*s23*n[1]*n[2] + 2*s13*n[0]*n[2]\n )\n \n sm = (1.0/3.0)*(s11 + s22 + s33)\n \n es = (nTn - sm)*Jas\n ev = (sm - p0)*Jav\n \n e = es + ev\n \n return e*10**6\n\n\nproblem = {\n 'num_vars': 9,\n 'names': [ 's11', 's22', 's33', 's12', 's13', 's23', 'p0', 'ts', 'tv'],\n 'bounds': [[0, 100]*9,\n# [0, 100],\n# [35, 45],\n# [0, 100],\n# [0, 100],\n# [0, 100],\n# [2, 12],\n# [0, 100],\n# [0, 100],\n ]\n}\n\n\n\nparam_values = saltelli.sample(problem, sampling)\n\n\ndef S(t):\n Y = np.zeros([param_values.shape[0]])\n \n for i, (s11, s22, s33, s12, s13, s23, p0, ts, tv) in enumerate(param_values):\n Y[i] = ASR(t, s11, s22, s33, s12, s13, s23, p0, ts, tv)\n \n Si = sobol.analyze(problem, Y, print_to_console=False)\n \n out = np.array([[t]])\n outT = np.array([[t]])\n \n for i in range(0, 9):\n inn = np.array([\n [Si[\"S1\"][i], Si[\"S1_conf\"][i]]\n ])\n out = np.hstack([out, inn])\n\n for iT in range(0, 9):\n innT = np.array([\n [Si[\"ST\"][iT], Si[\"ST_conf\"][iT]]\n ])\n outT = np.hstack([outT, innT])\n \n \n return out, outT\n\n\ncolumns = [\"time[h]\",\n \"s11\",\"s11_err\",\n \"s22\",\"s22_err\",\n \"s33\",\"s33_err\",\n \"s12\",\"s12_err\",\n \"s13\",\"s13_err\",\n \"s23\",\"s23_err\",\n \"p0\",\"p0_err\",\n \"ts\",\"ts_err\",\n \"tv\",\"tv_err\"\n ]\n\nSobol_Si = pd.DataFrame(index=[], columns=columns)\nSobol_ST = pd.DataFrame(index=[], columns=columns)\n\nfor t in range(1, time, 5):\n SI = S(t)\n\n Si_p = pd.DataFrame(data = SI[0], columns = columns)\n Sobol_Si = Sobol_Si.append(Si_p)\n\n ST_p = pd.DataFrame(data = SI[1], columns = columns)\n Sobol_ST = Sobol_ST.append(ST_p)\n\nSobol_Si.to_csv(\"output/\" + \"Si_K_G_\" + dirc + \".csv\", index=False)\nSobol_ST.to_csv(\"output/\" + \"ST_K_G_\" + dirc + \".csv\", index=False)\n", "sub_path": "ASR-K-G.py", "file_name": "ASR-K-G.py", "file_ext": "py", "file_size_in_byte": 3006, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.matrix", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 32, "usage_type": "call"}, {"api_name": "SALib.sample.saltelli.sample", "line_number": 65, "usage_type": "call"}, {"api_name": "SALib.sample.saltelli", "line_number": 65, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 69, "usage_type": "call"}, {"api_name": "SALib.analyze.sobol.analyze", "line_number": 74, "usage_type": "call"}, {"api_name": "SALib.analyze.sobol", "line_number": 74, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.hstack", "line_number": 89, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 107, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 108, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 113, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 116, "usage_type": "call"}]} +{"seq_id": "92863926", "text": "import json\n\nfrom flask import Flask, request\n\napp = Flask(__name__)\n\n@app.route('/01-server')\ndef server01():\n # 接收前端传递过来的参数 - callback\n cb = request.args.get('callback')\n return cb+\"('这是服务器端响应的内容');\"\n\n@app.route('/02-server')\ndef server02():\n cb = request.args.get('callback')\n dic = {\n 'flightNO' : 'CA977',\n 'from' : 'Beijing',\n 'to' : 'LA',\n 'time' : '00:30',\n }\n jsonStr = json.dumps(dic)\n return cb+\"(\"+jsonStr+\");\"\n\n@app.route('/03-jq-cross')\ndef jq_cross():\n cb = request.args.get('callback')\n return cb+\"('服务器端响应回去的数据')\"\n\n@app.route('/03-server')\ndef server03():\n cb = request.args.get('huidiao')\n print(cb)\n return cb+\"('这是使用方案2响应的数据');\"\n\nif __name__ == \"__main__\":\n app.run(debug=True,host='0.0.0.0')", "sub_path": "PythonWeb/Ajax/1809/Day04/1808/AjaxDemo04/run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 870, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 5, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 10, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.args.get", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}]} +{"seq_id": "467100766", "text": "import os, sys\nimport numpy as np\nimport pandas as pd\nfrom sklearn.naive_bayes import GaussianNB\n\nDATASET_TEST = '/dsa/data/all_datasets/titanic_ML/titanic.test.csv'\nOUTPUT_PATH = './'\n\ndef check(model):\n if not os.path.exists(DATASET_TEST):\n raise Exception('Test dataset not found. Please ask instructor for help.')\n \n if not isinstance(model, GaussianNB):\n raise TypeError('Expecting a GaussianNB model.')\n \n try:\n prediction = model.predict(pd.read_csv(DATASET_TEST))\n except Exception as e:\n raise RuntimeError('Unable to perform prediction on test dataset.', e)\n \n if prediction.shape != (419,):\n raise Exception('Resulting prediction has wrong dimension. Expecting: (419,) Received:', prediction.shape)\n\ndef snapshot(model):\n os.system('mkdir -p %s' % OUTPUT_PATH)\n os.system('rm -rf %s' % os.path.join(OUTPUT_PATH, '*.npy'))\n prediction = np.array(model.predict(pd.read_csv(DATASET_TEST)))\n FNAME = os.path.join(OUTPUT_PATH, 'submission.npy')\n np.save(FNAME, prediction, allow_pickle=False, fix_imports=True)\n assert os.path.exists(FNAME)\n", "sub_path": "module1/exercises/submission.py", "file_name": "submission.py", "file_ext": "py", "file_size_in_byte": 1132, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.exists", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "attribute"}, {"api_name": "sklearn.naive_bayes.GaussianNB", "line_number": 13, "usage_type": "argument"}, {"api_name": "pandas.read_csv", "line_number": 17, "usage_type": "call"}, {"api_name": "os.system", "line_number": 25, "usage_type": "call"}, {"api_name": "os.system", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}]} +{"seq_id": "86953124", "text": "from selenium import webdriver\nfrom selenium.common.exceptions import NoSuchElementException\n\nimport bs4\n\nimport time\n\n\nbrowser = webdriver.Firefox() # Запускаем локальную сессию firefox\nurl = \"https://www.aviasales.ru/search/MOW0302PAR20021\"\nbrowser.get(url) # Загружаем страницу\ntime.sleep(20) # Пусть страница загрузится. Вдруг у нас медленный интернет...\ntry:\n tickets = browser.find_elements_by_class_name(\"ticket__container\")\n for ticket in tickets:\n price = ticket.find_element_by_class_name(\"buy-button__price-num\")\n text_price = price.text\n href = ticket.find_element_by_class_name(\"buy-button__link\")\n href_text = href.get_attribute(\"href\")\n routes_containers = ticket.find_elements_by_class_name(\"segment-route\")\n print(\"------------------------------------------------\")\n print(\"Цена: {}\".format(text_price))\n print(\"Ссылка: {}\".format(href_text))\n count = 0\n for route_container in routes_containers:\n times = route_container.find_elements_by_class_name(\"segment-route__time\")\n cities = route_container.find_elements_by_class_name(\"segment-route__city\")\n dates = route_container.find_elements_by_class_name(\"segment-route__date\")\n\n origin_time_text = times[0].text\n origin_city_text = cities[0].text\n origin_date_text = dates[0].text\n\n dest_time_text = times[1].text\n dest_city_text = cities[1].text\n dest_date_text = dates[1].text\n print(\"Туда:\") if count == 0 else print(\"Обратно:\")\n print()\n print(\"Откуда:\")\n print(origin_city_text)\n print(origin_time_text)\n print(origin_date_text)\n print()\n print(\"Куда:\")\n print(dest_city_text)\n print(dest_time_text)\n print(dest_date_text)\n count+=1\n print(\"------------------------------------------------\")\n print(\"\\n\")\n\n\n\nexcept NoSuchElementException:\n assert 0, \"can't open url\"\n\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2184, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "selenium.webdriver.Firefox", "line_number": 9, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 9, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 12, "usage_type": "call"}, {"api_name": "selenium.common.exceptions.NoSuchElementException", "line_number": 54, "usage_type": "name"}]} +{"seq_id": "396526758", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# In[1]:\n\n\nimport h5py\nfrom sklearn.model_selection import train_test_split\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom keras import optimizers\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Conv2D, MaxPooling2D, Flatten, Dropout, BatchNormalization, ReLU, Softmax\nfrom keras.regularizers import l2, l1\nfrom keras.initializers import glorot_normal\nfrom sklearn.preprocessing import MultiLabelBinarizer\nfrom keras.models import load_model\nfrom sklearn.metrics import mean_squared_error\nimport pandas as pd\nimport math\nimport tensorflow as tf\nfrom tensorflow import keras\n#from tensorflow.keras.optimizers import Adam\nfrom keras.callbacks import LearningRateScheduler, ReduceLROnPlateau\n\n\n# In[2]:\n\n\n# Fix K detection of sources - Experiment Part A: 1)\nfilename1 = 'C:/Users/geo_p/OneDrive - Heriot-Watt University/DoA DATA/DoA_DATA_JOURNALS/TRAIN_DATA_16ULA_K2_low_SNR_res1_3D_90deg.h5'\nf1 = h5py.File(filename1, 'r')\nangles = np.transpose(np.array(f1['angles']))\nRy_the = np.array(f1['theor'])\nres = 1\nK=2\n\n\n# In[3]:\n\n\nAn_max = np.max(angles)\nAn_min = np.min(angles)\nv = np.arange(An_min, An_max+res,res)\nprint(v)\n\n\n# In[4]:\n\n\nDNN_outp = v.size\nprint(DNN_outp)\n\n\n# In[5]:\n\n\nangles.shape\n\n\n# In[6]:\n\n\nRy_the.shape\n\n\n# In[7]:\n\n\n[SNRs, n, chan, M, N] = Ry_the.shape\n\n\n# In[8]:\n\n\nX_data0=Ry_the.swapaxes(2,4)\nX_data0.shape\n\n\n# In[9]:\n\n\nX_data = X_data0.reshape([SNRs*n,N,M,chan])\nX_data.shape\n\n\n# In[10]:\n\n\nmlb = MultiLabelBinarizer()\nyTrain_encoded = mlb.fit_transform(angles)\n\n\n# In[11]:\n\n\nyTrain_encoded[1]\n\n\n# In[12]:\n\n\nY_Labels = np.tile(yTrain_encoded, reps=(SNRs,1))\n#Y_Labels = yTrain_encoded\nY_Labels.shape\n\n\n# In[13]:\n\n\n# Split the dataset into training and validation sets\nxTrain, xVal, yTrain, yVal = train_test_split(X_data, Y_Labels, test_size=0.1, random_state=42) # checked\n\n\n# In[16]:\n\n\n# Define the model (CNN) for single source localization\ninput_shape = xTrain.shape[1:]\nkern_size1 = 3\nkern_size2 = 2\n\nmodel = Sequential() \nmodel.add(Conv2D(256, kernel_size=(kern_size1,kern_size1), activation=None, input_shape=input_shape, name=\"Conv2D_1\",padding=\"valid\", strides=(2,2)))\nmodel.add(BatchNormalization(trainable=True))\nmodel.add(ReLU())\nmodel.add(Conv2D(256, kernel_size=(kern_size2,kern_size2), activation=None,name=\"Conv2D_2\", padding=\"valid\"))\nmodel.add(BatchNormalization(trainable=True))\nmodel.add(ReLU())\nmodel.add(Conv2D(256, kernel_size=(kern_size2,kern_size2), activation=None,name=\"Conv2D_3\", padding=\"valid\"))\nmodel.add(BatchNormalization(trainable=True))\nmodel.add(ReLU())\nmodel.add(Conv2D(256, kernel_size=(kern_size2,kern_size2), activation=None,name=\"Conv2D_4\", padding=\"valid\"))\nmodel.add(BatchNormalization(trainable=True))\nmodel.add(ReLU())\nmodel.add(Flatten())\nmodel.add(Dense(4096, activation=\"relu\",name=\"Dense_Layer1\"))\nmodel.add(Dropout(0.3,name=\"Dropout1\"))\nmodel.add(Dense(2048, activation=\"relu\",name=\"Dense_Layer2\"))\nmodel.add(Dropout(0.3,name=\"Dropout2\"))\nmodel.add(Dense(1024, activation=\"relu\",name=\"Dense_Layer3\"))\nmodel.add(Dropout(0.3,name=\"Dropout3\"))\nmodel.add(Dense(DNN_outp, activation=\"sigmoid\", kernel_initializer=glorot_normal(seed=None),name=\"Classif_Layer\"))\nmodel.summary()\n\n\n# In[17]:\n\n\n# Train the model with Adam\n# Train the model with decaying learn rate\n# OPTION 1\ndef schedule(epoch,lr): # use this function to gradually reduce the lr\n if epoch<1:\n return lr\n else:\n return float(lr*tf.math.exp(-0.1))\n# OPTION 2\ndef step_decay(epoch, lr): # or use this function to reduce every epochs_drop by a desired factor\n initial_lr = 0.001\n drop = 0.5\n epochs_drop = 20\n lrate = initial_lr* math.pow(drop, math.floor((1+epoch)/epochs_drop))\n return lrate\n#dlr = LearningRateScheduler(step_decay,verbose=1)\nrlr = ReduceLROnPlateau(monitor='val_loss', factor=0.7, patience=10, verbose=1)\ncbks = [rlr]\n#opt = tf.keras.optimizers.SGD(learning_rate=0.1,momentum=0.9,nesterov=True)\nopt = tf.keras.optimizers.Adam(learning_rate=0.001)\nmodel.compile(optimizer=opt , loss='binary_crossentropy', metrics=[tf.keras.metrics.BinaryAccuracy(name=\"acc\")])\ntrain_history = model.fit(xTrain, yTrain, epochs=200, batch_size=32, shuffle=True, validation_data=(xVal, yVal), callbacks=cbks)\n\n\n# In[18]:\n\n\n# summarize history for accuracy\nf1 = plt.figure(1)\nplt.plot(train_history.history['acc'], label='Training accuracy')\nplt.plot(train_history.history['val_acc'], label='Validation accuracy')\n#plt.title('model accuracy')\nplt.ylabel('Accuracy')\nplt.xlabel('Epoch')\nplt.legend(['Train', 'Val.'], loc='lower right')\nplt.grid()\nplt.show()\n\n# summarize history for loss\nf2 = plt.figure(2)\nplt.plot(train_history.history['loss'], label='Training loss')\nplt.plot(train_history.history['val_loss'], label='Validation loss')\n#plt.title('model loss')\nplt.ylabel('Loss')\nplt.xlabel('Epoch')\nplt.legend(['Train', 'Val.'], loc='upper left')\nplt.grid()\nplt.show()\n\n\n# In[19]:\n\n\n# save the training performance for reporting\nf3 = plt.figure(1)\nplt.subplot(2,1,1)\nplt.plot(train_history.history['acc'], label='Training accuracy')\nplt.plot(train_history.history['val_acc'], label='Validation accuracy')\nplt.ylabel('Accuracy')\nplt.legend(['Train', 'Val.'], loc='lower right')\nplt.grid()\nplt.subplot(2,1,2)\nplt.plot(train_history.history['loss'], label='Training loss')\nplt.plot(train_history.history['val_loss'], label='Validation loss')\nplt.ylabel('Loss')\nplt.xlabel('Epoch')\nplt.legend(['Train', 'Val.'], loc='upper right')\nplt.grid()\nplt.show()\n\n\n# In[20]:\n\n\nf1.savefig(\"binary_acc_Adam_LRdef_90deg_v6c.eps\", dpi=1200, bbox_inches='tight')\nf2.savefig(\"loss_Adam_LRdef_90deg_v6c.eps\", dpi=1200, bbox_inches='tight')\n\n\n# In[21]:\n\n\n# Save the figures to include them in the paper\nf3.savefig(\"training_perf_Adam_LRdef_90deg_v6c.eps\", dpi=1200, bbox_inches='tight')\n\n\n# In[22]:\n\n\nmodel.save('Model_CNN_DoA_class_Data_N16_K2_res1_lowSNR_new_training_RQ_90deg_v6c.h5') \n\n\n# In[ ]:\n\n\n\n\n", "sub_path": "Python_CNN/CNN_training_lowSNR_new_training_RQ_test_90deg.py", "file_name": "CNN_training_lowSNR_new_training_RQ_test_90deg.py", "file_ext": "py", "file_size_in_byte": 5925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "h5py.File", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 44, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.MultiLabelBinarizer", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 103, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 123, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 124, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 125, "usage_type": "call"}, {"api_name": "keras.layers.ReLU", "line_number": 126, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 127, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 128, "usage_type": "call"}, {"api_name": "keras.layers.ReLU", "line_number": 129, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 130, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 131, "usage_type": "call"}, {"api_name": "keras.layers.ReLU", "line_number": 132, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 133, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.layers.ReLU", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 136, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 137, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 138, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 139, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 140, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 141, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 142, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 143, "usage_type": "call"}, {"api_name": "keras.initializers.glorot_normal", "line_number": 143, "usage_type": "call"}, {"api_name": "tensorflow.math.exp", "line_number": 157, "usage_type": "call"}, {"api_name": "tensorflow.math", "line_number": 157, "usage_type": "attribute"}, {"api_name": "math.pow", "line_number": 163, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 163, "usage_type": "call"}, {"api_name": "keras.callbacks.ReduceLROnPlateau", "line_number": 166, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 169, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 169, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.metrics.BinaryAccuracy", "line_number": 170, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 170, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 182, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 182, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 183, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 183, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 193, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 193, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 196, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 204, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 204, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 205, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 205, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 206, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 206, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 207, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 207, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 208, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 208, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 209, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 209, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 210, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 210, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 211, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 211, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 212, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 212, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 213, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 213, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 214, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 214, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 215, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 215, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 216, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 216, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 217, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 217, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 218, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 218, "usage_type": "name"}]} +{"seq_id": "96587472", "text": "# Copyright 2015 Infoblox Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport re\n\nfrom infoblox_client import exceptions as ibc_exc\nfrom neutron_lib import constants as n_const\n\nfrom networking_infoblox.neutron.common import constants as const\n\n\nclass PatternBuilder(object):\n\n def __init__(self, ib_context):\n self.ib_cxt = ib_context\n self.grid_config = self.ib_cxt.grid_config\n\n def get_hostname(self, ip_address, instance_name=None, port_id=None,\n device_owner=None, device_id=None, port_name=None,\n external=False, port_tenant_id=None, tenant_name=None):\n \"\"\"Build fqdn based on patterns for network type and device owner.\n\n Two types of host and domain patterns exist:\n - for external network (optional);\n - for private network (default);\n If pattern for external network is not set, default one is used.\n If device owner is a known one (like dhcp_port, routern interface\n etc.), then per owner patterns are used. Floating ip can be exception\n from this rule if VM is associated with it and instance name is\n present in the pattern.\n \"\"\"\n if external:\n host_pattern = (self.grid_config.external_host_name_pattern or\n self.grid_config.default_host_name_pattern)\n domain_pattern = (self.grid_config.external_domain_name_pattern or\n self.grid_config.default_domain_name_pattern)\n else:\n host_pattern = self.grid_config.default_host_name_pattern\n domain_pattern = self.grid_config.default_domain_name_pattern\n\n if device_owner in const.NEUTRON_DEVICE_OWNER_TO_PATTERN_MAP and (\n device_owner != n_const.DEVICE_OWNER_FLOATINGIP or not\n instance_name or \"{instance_name}\" not in host_pattern):\n host_pattern = (\n const.NEUTRON_DEVICE_OWNER_TO_PATTERN_MAP[device_owner])\n\n pattern = [host_pattern, domain_pattern]\n pattern = '.'.join(el.strip('.') for el in pattern if el)\n return self._build(pattern, ip_address, instance_name, port_id,\n device_id, port_name=port_name,\n port_tenant_id=port_tenant_id,\n tenant_name=tenant_name)\n\n def get_zone_name_pattern(self, subnet_name=None, is_external=False):\n pattern = self.grid_config.default_domain_name_pattern\n if is_external and self.grid_config.external_domain_name_pattern:\n pattern = self.grid_config.external_domain_name_pattern\n return pattern\n\n def get_zone_name(self, subnet_name=None, tenant_name=None,\n port_tenant_id=None, is_external=False):\n pattern = self.get_zone_name_pattern(subnet_name, is_external)\n return self._build(pattern, subnet_name=subnet_name,\n tenant_name=tenant_name,\n port_tenant_id=port_tenant_id)\n\n def _build(self, pattern, ip_address=None, instance_name=None,\n port_id=None, device_id=None, subnet_name=None,\n port_name=None, port_tenant_id=None, tenant_name=None):\n self._validate_pattern(pattern)\n\n subnet = self.ib_cxt.subnet\n network = self.ib_cxt.network\n if not subnet_name:\n subnet_name = (subnet['name'] if subnet.get('name')\n else subnet['id'])\n network_name = (network['name'] if network.get('name')\n else network['id'])\n\n pattern_dict = {\n 'network_id': subnet['network_id'],\n 'network_name': network_name,\n 'tenant_id': port_tenant_id or self.ib_cxt.tenant_id,\n 'tenant_name': tenant_name or self.ib_cxt.tenant_name,\n 'subnet_name': subnet_name,\n 'subnet_id': subnet['id']\n }\n\n if port_id:\n pattern_dict['port_id'] = port_id\n\n if device_id:\n pattern_dict['instance_id'] = device_id\n if instance_name:\n pattern_dict['instance_name'] = re.sub(\"[^A-Za-z0-9-]\", \"-\",\n instance_name.strip())\n else:\n # During port_creation for instance_name is not available,\n # so set it to instance_id\n pattern_dict['instance_name'] = pattern_dict['instance_id']\n\n if ip_address:\n octets = ip_address.split('.')\n ip_addr = ip_address.replace('.', '-').replace(':', '-')\n pattern_dict['ip_address'] = ip_addr\n for i in range(len(octets)):\n octet_key = 'ip_address_octet{i}'.format(i=(i + 1))\n pattern_dict[octet_key] = octets[i]\n\n if port_name:\n pattern_dict['port_name'] = port_name\n elif 'ip_address' in pattern_dict:\n pattern_dict['port_name'] = pattern_dict['ip_address']\n else:\n pattern_dict['port_name'] = port_id\n\n try:\n # Validate grid config pattern with pattern_dict to\n # restrict user to use only those pattern variable in\n # grid config pattern which is available in pattern_dict\n self._validate_pattern_struct(pattern, pattern_dict)\n fqdn = pattern.format(**pattern_dict)\n except (KeyError, IndexError) as e:\n raise ibc_exc.InfobloxConfigException(\n msg=\"Invalid pattern %s\" % e)\n # Return fqdn as lowercase string as NIOS creates all resources\n # in lowercases.\n return fqdn.lower()\n\n @staticmethod\n def _validate_pattern_struct(pattern, pattern_dict):\n # This function fetches all variables from grid config\n # pattern and validate the list with supported pattern\n # variables, raises a KeyError if validation fails.\n input_pattern_list = re.findall(\"\\{(.*?)\\}\", pattern)\n invalid_pattern = set(input_pattern_list) - set(pattern_dict)\n if invalid_pattern:\n raise KeyError(list(invalid_pattern))\n\n @staticmethod\n def _validate_pattern(pattern):\n invalid_values = ['..']\n for val in invalid_values:\n if val in pattern:\n error_message = \"Invalid pattern value {0}\".format(val)\n raise ibc_exc.InfobloxConfigException(msg=error_message)\n", "sub_path": "networking_infoblox/neutron/common/pattern.py", "file_name": "pattern.py", "file_ext": "py", "file_size_in_byte": 6958, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "networking_infoblox.neutron.common.constants.NEUTRON_DEVICE_OWNER_TO_PATTERN_MAP", "line_number": 53, "usage_type": "attribute"}, {"api_name": "networking_infoblox.neutron.common.constants", "line_number": 53, "usage_type": "name"}, {"api_name": "neutron_lib.constants.DEVICE_OWNER_FLOATINGIP", "line_number": 54, "usage_type": "attribute"}, {"api_name": "neutron_lib.constants", "line_number": 54, "usage_type": "name"}, {"api_name": "networking_infoblox.neutron.common.constants.NEUTRON_DEVICE_OWNER_TO_PATTERN_MAP", "line_number": 57, "usage_type": "attribute"}, {"api_name": "networking_infoblox.neutron.common.constants", "line_number": 57, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 107, "usage_type": "call"}, {"api_name": "infoblox_client.exceptions.InfobloxConfigException", "line_number": 136, "usage_type": "call"}, {"api_name": "infoblox_client.exceptions", "line_number": 136, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 147, "usage_type": "call"}, {"api_name": "infoblox_client.exceptions.InfobloxConfigException", "line_number": 158, "usage_type": "call"}, {"api_name": "infoblox_client.exceptions", "line_number": 158, "usage_type": "name"}]} +{"seq_id": "443300363", "text": "import boto3\nimport os\n\ndef lambda_handler(event, context):\n\n\tclient = boto3.client('iam', region_name = os.environ[\"REGION\"])\n\tmy_list = list()\n\tiam_all_users = client.list_users()\n\tfor user in iam_all_users['Users']:\n\t\tmy_list.append(user['UserName'])\n\n\treturn {\n\t\t\t'body': {'userList': my_list},\n\t\t\t'headers': {'Content-Type': 'text/html', 'Access-Control-Allow-Origin': '*'}\n\t}", "sub_path": "src/listIAMUsers/lambda_function.py", "file_name": "lambda_function.py", "file_ext": "py", "file_size_in_byte": 382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "boto3.client", "line_number": 6, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}]} +{"seq_id": "334637165", "text": "import csv\r\nimport datetime\r\nfrom dateutil import parser\r\nitems = open(\"ProductDatabase.csv\", \"r\")\r\nusers = open(\"users.csv\", \"r\")\r\nitems_access = csv.reader(items)\r\nitems_access_copy = []\r\nusers_access = csv.reader(users)\r\nuser_id = 1#This will depend on the USER!!! CHANGE ME IN THE GUI\r\nuser_items = []\r\nfor row in items_access:\r\n items_access_copy.append(row)\r\ndef find_item_date_and_name(item_id):\r\n global items_access_copy\r\n i = 0\r\n arr = []\r\n for row in items_access_copy:\r\n #print(row)\r\n if i == 0:\r\n i = i + 1\r\n else:\r\n if int(row[0]) == item_id:#we have found the item, now access the exp date\r\n exp_date = parser.parse(row[3])\r\n name = row[1]\r\n arr.append(exp_date)\r\n arr.append(name)\r\n return arr\r\ni = 0\r\nfor row in users_access:\r\n if i == 0:\r\n i = i + 1#IGNORE THE HEADER\r\n else:\r\n if int(row[0]) == user_id:#we have found the user, access their items\r\n exp_date_and_name = find_item_date_and_name(int(row[1]))\r\n current_date = datetime.datetime.now()\r\n print (str(exp_date_and_name[1]) + \" will expire in \" + str((exp_date_and_name[0] - current_date).days) + \" days\")\r\n", "sub_path": "items.py", "file_name": "items.py", "file_ext": "py", "file_size_in_byte": 1258, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "csv.reader", "line_number": 6, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 8, "usage_type": "call"}, {"api_name": "dateutil.parser.parse", "line_number": 23, "usage_type": "call"}, {"api_name": "dateutil.parser", "line_number": 23, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "attribute"}]} +{"seq_id": "150260776", "text": "# stdlib\nfrom typing import Any\nfrom typing import Dict\nfrom typing import List\nfrom typing import Optional\nfrom typing import Union\n\n# third party\nfrom result import OkErr\nfrom result import Result\n\n# relative\nfrom ...serde.serializable import serializable\nfrom ...store.document_store import DocumentStore\nfrom ...store.linked_obj import LinkedObject\nfrom ...types.twin_object import TwinObject\nfrom ...types.uid import UID\nfrom ...util.telemetry import instrument\nfrom ..action.action_object import ActionObject\nfrom ..context import AuthedServiceContext\nfrom ..policy.policy import OutputHistory\nfrom ..request.request import SubmitRequest\nfrom ..request.request import UserCodeStatusChange\nfrom ..request.request_service import RequestService\nfrom ..response import SyftError\nfrom ..response import SyftNotReady\nfrom ..response import SyftSuccess\nfrom ..service import AbstractService\nfrom ..service import SERVICE_TO_TYPES\nfrom ..service import TYPE_TO_SERVICE\nfrom ..service import service_method\nfrom ..user.user_roles import GUEST_ROLE_LEVEL\nfrom .user_code import SubmitUserCode\nfrom .user_code import UserCode\nfrom .user_code import UserCodeStatus\nfrom .user_code import load_approved_policy_code\nfrom .user_code_stash import UserCodeStash\n\n\n@instrument\n@serializable()\nclass UserCodeService(AbstractService):\n store: DocumentStore\n stash: UserCodeStash\n\n def __init__(self, store: DocumentStore) -> None:\n self.store = store\n self.stash = UserCodeStash(store=store)\n\n @service_method(path=\"code.submit\", name=\"submit\", roles=GUEST_ROLE_LEVEL)\n def submit(\n self, context: AuthedServiceContext, code: SubmitUserCode\n ) -> Union[UserCode, SyftError]:\n \"\"\"Add User Code\"\"\"\n result = self.stash.set(context.credentials, code.to(UserCode, context=context))\n if result.is_err():\n return SyftError(message=str(result.err()))\n return SyftSuccess(message=\"User Code Submitted\")\n\n def _request_code_execution(\n self,\n context: AuthedServiceContext,\n code: SubmitUserCode,\n reason: Optional[str] = \"\",\n ):\n user_code = code.to(UserCode, context=context)\n result = self.stash.set(context.credentials, user_code)\n if result.is_err():\n return SyftError(message=str(result.err()))\n\n linked_obj = LinkedObject.from_obj(user_code, node_uid=context.node.id)\n\n CODE_EXECUTE = UserCodeStatusChange(\n value=UserCodeStatus.EXECUTE, linked_obj=linked_obj\n )\n changes = [CODE_EXECUTE]\n\n kwargs_keys = code.input_kwargs\n kwargs_value = [list(code.kwargs.values())[0][key] for key in kwargs_keys]\n\n request = SubmitRequest(\n changes=changes, action_id=kwargs_value[0], reason=reason\n )\n method = context.node.get_service_method(RequestService.submit)\n result = method(context=context, request=request)\n\n # The Request service already returns either a SyftSuccess or SyftError\n return result\n\n @service_method(\n path=\"code.request_code_execution\",\n name=\"request_code_execution\",\n roles=GUEST_ROLE_LEVEL,\n )\n def request_code_execution(\n self,\n context: AuthedServiceContext,\n code: SubmitUserCode,\n reason: Optional[str] = \"\",\n ) -> Union[SyftSuccess, SyftError]:\n \"\"\"Request Code execution on user code\"\"\"\n return self._request_code_execution(context=context, code=code, reason=reason)\n\n @service_method(path=\"code.get_all\", name=\"get_all\", roles=GUEST_ROLE_LEVEL)\n def get_all(\n self, context: AuthedServiceContext\n ) -> Union[List[UserCode], SyftError]:\n \"\"\"Get a Dataset\"\"\"\n result = self.stash.get_all(context.credentials)\n if result.is_ok():\n return result.ok()\n return SyftError(message=result.err())\n\n @service_method(path=\"code.get_by_id\", name=\"get_by_id\")\n def get_by_uid(\n self, context: AuthedServiceContext, uid: UID\n ) -> Union[SyftSuccess, SyftError]:\n \"\"\"Get a User Code Item\"\"\"\n result = self.stash.get_by_uid(context.credentials, uid=uid)\n if result.is_ok():\n user_code = result.ok()\n if user_code and user_code.input_policy_state:\n # TODO replace with LinkedObject Context\n user_code.node_uid = context.node.id\n return user_code\n return SyftError(message=result.err())\n\n @service_method(path=\"code.get_all_for_user\", name=\"get_all_for_user\")\n def get_all_for_user(\n self, context: AuthedServiceContext\n ) -> Union[SyftSuccess, SyftError]:\n \"\"\"Get All User Code Items for User's VerifyKey\"\"\"\n # TODO: replace with incoming user context and key\n result = self.stash.get_all(context.credentials)\n if result.is_ok():\n return result.ok()\n return SyftError(message=result.err())\n\n def update_code_state(\n self, context: AuthedServiceContext, code_item: UserCode\n ) -> Union[SyftSuccess, SyftError]:\n result = self.stash.update(context.credentials, code_item)\n if result.is_ok():\n return SyftSuccess(message=\"Code State Updated\")\n return SyftError(message=\"Unable to Update Code State\")\n\n def load_user_code(self, context: AuthedServiceContext) -> None:\n result = self.stash.get_all(credentials=context.credentials)\n if result.is_ok():\n user_code_items = result.ok()\n load_approved_policy_code(user_code_items=user_code_items)\n\n @service_method(path=\"code.call\", name=\"call\", roles=GUEST_ROLE_LEVEL)\n def call(\n self, context: AuthedServiceContext, uid: UID, **kwargs: Any\n ) -> Union[SyftSuccess, SyftError]:\n \"\"\"Call a User Code Function\"\"\"\n try:\n filtered_kwargs = filter_kwargs(kwargs)\n result = self.stash.get_by_uid(context.credentials, uid=uid)\n if not result.is_ok():\n return SyftError(message=result.err())\n\n # Unroll variables\n code_item = result.ok()\n status = code_item.status\n\n # Check if the user has permission to execute the code\n # They can execute if they are root user or if they are the user who submitted the code\n if not (\n context.credentials == context.node.verify_key\n or context.credentials == code_item.user_verify_key\n ):\n return SyftError(\n message=f\"Code Execution Permission: {context.credentials} denied\"\n )\n\n # Check if the code is approved\n if status.for_context(context) != UserCodeStatus.EXECUTE:\n if status.for_context(context) == UserCodeStatus.SUBMITTED:\n return SyftNotReady(\n message=f\"{type(code_item)} Your code is waiting for approval: {status}\"\n )\n return SyftError(\n message=f\"{type(code_item)} Your code cannot be run: {status.for_context(context)}\"\n )\n\n output_policy = code_item.output_policy\n if output_policy is None:\n raise Exception(\"Output policy not approved\", code_item)\n\n # Check if the OutputPolicy is valid\n is_valid = output_policy.valid\n\n if not is_valid:\n if len(output_policy.output_history) > 0:\n result = get_outputs(\n context=context,\n output_history=output_policy.output_history[-1],\n )\n return result.as_empty()\n return is_valid\n\n # Execute the code item\n action_service = context.node.get_service(\"actionservice\")\n result: Result = action_service._user_code_execute(\n context, code_item, filtered_kwargs\n )\n if isinstance(result, str):\n return SyftError(message=result)\n\n # Apply Output Policy to the results and update the OutputPolicyState\n result: Union[ActionObject, TwinObject] = result.ok()\n output_policy.apply_output(context=context, outputs=result)\n code_item.output_policy = output_policy\n update_success = self.update_code_state(\n context=context, code_item=code_item\n )\n if not update_success:\n return update_success\n if isinstance(result, TwinObject):\n return result.mock\n else:\n return result.as_empty()\n except Exception as e:\n return SyftError(message=f\"Failed to run. {e}\")\n\n\ndef get_outputs(context: AuthedServiceContext, output_history: OutputHistory) -> Any:\n # relative\n from ...service.action.action_object import TwinMode\n\n if isinstance(output_history.outputs, list):\n if len(output_history.outputs) == 0:\n return None\n outputs = []\n for output_id in output_history.outputs:\n action_service = context.node.get_service(\"actionservice\")\n result = action_service.get(\n context, uid=output_id, twin_mode=TwinMode.PRIVATE\n )\n if isinstance(result, OkErr):\n result = result.value\n outputs.append(result)\n if len(outputs) == 1:\n return outputs[0]\n return outputs\n else:\n raise NotImplementedError\n\n\ndef filter_kwargs(kwargs: Dict[str, Any]) -> Dict[str, Any]:\n # relative\n from ...types.twin_object import TwinObject\n from ..action.action_object import ActionObject\n from ..dataset.dataset import Asset\n\n filtered_kwargs = {}\n for k, v in kwargs.items():\n value = v\n if isinstance(v, ActionObject):\n value = v.id\n if isinstance(v, TwinObject):\n value = v.id\n if isinstance(v, Asset):\n value = v.action_id\n\n if not isinstance(value, UID):\n raise Exception(f\"Input {k} must have a UID not {type(v)}\")\n filtered_kwargs[k] = value\n return filtered_kwargs\n\n\nTYPE_TO_SERVICE[UserCode] = UserCodeService\nSERVICE_TO_TYPES[UserCodeService].update({UserCode})\n", "sub_path": "packages/syft/src/syft/service/code/user_code_service.py", "file_name": "user_code_service.py", "file_ext": "py", "file_size_in_byte": 10282, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "service.AbstractService", "line_number": 42, "usage_type": "name"}, {"api_name": "store.document_store", "line_number": 43, "usage_type": "name"}, {"api_name": "store.document_store.DocumentStore", "line_number": 43, "usage_type": "name"}, {"api_name": "user_code_stash.UserCodeStash", "line_number": 44, "usage_type": "name"}, {"api_name": "store.document_store.DocumentStore", "line_number": 46, "usage_type": "name"}, {"api_name": "store.document_store", "line_number": 47, "usage_type": "name"}, {"api_name": "user_code_stash.UserCodeStash", "line_number": 48, "usage_type": "call"}, {"api_name": "store.document_store", "line_number": 48, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 52, "usage_type": "name"}, {"api_name": "user_code.SubmitUserCode", "line_number": 52, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 55, "usage_type": "attribute"}, {"api_name": "user_code.UserCode", "line_number": 55, "usage_type": "argument"}, {"api_name": "result.is_err", "line_number": 56, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 57, "usage_type": "call"}, {"api_name": "result.err", "line_number": 57, "usage_type": "call"}, {"api_name": "response.SyftSuccess", "line_number": 58, "usage_type": "call"}, {"api_name": "service.service_method", "line_number": 50, "usage_type": "call"}, {"api_name": "user.user_roles.GUEST_ROLE_LEVEL", "line_number": 50, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 53, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 53, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 53, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 62, "usage_type": "name"}, {"api_name": "user_code.SubmitUserCode", "line_number": 63, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 64, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 66, "usage_type": "argument"}, {"api_name": "context.credentials", "line_number": 67, "usage_type": "attribute"}, {"api_name": "result.is_err", "line_number": 68, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 69, "usage_type": "call"}, {"api_name": "result.err", "line_number": 69, "usage_type": "call"}, {"api_name": "store.linked_obj.LinkedObject.from_obj", "line_number": 71, "usage_type": "call"}, {"api_name": "store.linked_obj.LinkedObject", "line_number": 71, "usage_type": "name"}, {"api_name": "context.node", "line_number": 71, "usage_type": "attribute"}, {"api_name": "request.request.UserCodeStatusChange", "line_number": 73, "usage_type": "call"}, {"api_name": "user_code.UserCodeStatus.EXECUTE", "line_number": 74, "usage_type": "attribute"}, {"api_name": "user_code.UserCodeStatus", "line_number": 74, "usage_type": "name"}, {"api_name": "request.request", "line_number": 81, "usage_type": "name"}, {"api_name": "request.request.SubmitRequest", "line_number": 81, "usage_type": "call"}, {"api_name": "context.node.get_service_method", "line_number": 84, "usage_type": "call"}, {"api_name": "context.node", "line_number": 84, "usage_type": "attribute"}, {"api_name": "request.request_service.RequestService.submit", "line_number": 84, "usage_type": "attribute"}, {"api_name": "request.request_service.RequestService", "line_number": 84, "usage_type": "name"}, {"api_name": "request.request", "line_number": 85, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 97, "usage_type": "name"}, {"api_name": "user_code.SubmitUserCode", "line_number": 98, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 99, "usage_type": "name"}, {"api_name": "service.service_method", "line_number": 90, "usage_type": "call"}, {"api_name": "user.user_roles.GUEST_ROLE_LEVEL", "line_number": 93, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 100, "usage_type": "name"}, {"api_name": "response.SyftSuccess", "line_number": 100, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 100, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 106, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 109, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 110, "usage_type": "call"}, {"api_name": "result.ok", "line_number": 111, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 112, "usage_type": "call"}, {"api_name": "result.err", "line_number": 112, "usage_type": "call"}, {"api_name": "service.service_method", "line_number": 104, "usage_type": "call"}, {"api_name": "user.user_roles.GUEST_ROLE_LEVEL", "line_number": 104, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 107, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 107, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 107, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 107, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 116, "usage_type": "name"}, {"api_name": "types.uid.UID", "line_number": 116, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 119, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 120, "usage_type": "call"}, {"api_name": "result.ok", "line_number": 121, "usage_type": "call"}, {"api_name": "user_code.input_policy_state", "line_number": 122, "usage_type": "attribute"}, {"api_name": "user_code.node_uid", "line_number": 124, "usage_type": "attribute"}, {"api_name": "context.node", "line_number": 124, "usage_type": "attribute"}, {"api_name": "response.SyftError", "line_number": 126, "usage_type": "call"}, {"api_name": "result.err", "line_number": 126, "usage_type": "call"}, {"api_name": "service.service_method", "line_number": 114, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 117, "usage_type": "name"}, {"api_name": "response.SyftSuccess", "line_number": 117, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 117, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 130, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 134, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 135, "usage_type": "call"}, {"api_name": "result.ok", "line_number": 136, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 137, "usage_type": "call"}, {"api_name": "result.err", "line_number": 137, "usage_type": "call"}, {"api_name": "service.service_method", "line_number": 128, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 131, "usage_type": "name"}, {"api_name": "response.SyftSuccess", "line_number": 131, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 131, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 140, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 140, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 142, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 143, "usage_type": "call"}, {"api_name": "response.SyftSuccess", "line_number": 144, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 145, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 141, "usage_type": "name"}, {"api_name": "response.SyftSuccess", "line_number": 141, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 141, "usage_type": "name"}, {"api_name": "context.AuthedServiceContext", "line_number": 147, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 148, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 149, "usage_type": "call"}, {"api_name": "result.ok", "line_number": 150, "usage_type": "call"}, {"api_name": "user_code.load_approved_policy_code", "line_number": 151, "usage_type": "call"}, {"api_name": "context.AuthedServiceContext", "line_number": 155, "usage_type": "name"}, {"api_name": "types.uid.UID", "line_number": 155, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 155, "usage_type": "name"}, {"api_name": "context.credentials", "line_number": 160, "usage_type": "attribute"}, {"api_name": "result.is_ok", "line_number": 161, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 162, "usage_type": "call"}, {"api_name": "result.err", "line_number": 162, "usage_type": "call"}, {"api_name": "result.ok", "line_number": 165, "usage_type": "call"}, {"api_name": "context.credentials", "line_number": 171, "usage_type": "attribute"}, {"api_name": "context.node", "line_number": 171, "usage_type": "attribute"}, {"api_name": "context.credentials", "line_number": 172, "usage_type": "attribute"}, {"api_name": "response.SyftError", "line_number": 174, "usage_type": "call"}, {"api_name": "context.credentials", "line_number": 175, "usage_type": "attribute"}, {"api_name": "user_code.UserCodeStatus.EXECUTE", "line_number": 179, "usage_type": "attribute"}, {"api_name": "user_code.UserCodeStatus", "line_number": 179, "usage_type": "name"}, {"api_name": "user_code.UserCodeStatus.SUBMITTED", "line_number": 180, "usage_type": "attribute"}, {"api_name": "user_code.UserCodeStatus", "line_number": 180, "usage_type": "name"}, {"api_name": "response.SyftNotReady", "line_number": 181, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 184, "usage_type": "call"}, {"api_name": "result.as_empty", "line_number": 201, "usage_type": "call"}, {"api_name": "context.node.get_service", "line_number": 205, "usage_type": "call"}, {"api_name": "context.node", "line_number": 205, "usage_type": "attribute"}, {"api_name": "result.Result", "line_number": 206, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 210, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 213, "usage_type": "name"}, {"api_name": "action.action_object.ActionObject", "line_number": 213, "usage_type": "name"}, {"api_name": "types.twin_object.TwinObject", "line_number": 213, "usage_type": "name"}, {"api_name": "result.ok", "line_number": 213, "usage_type": "call"}, {"api_name": "types.twin_object.TwinObject", "line_number": 221, "usage_type": "argument"}, {"api_name": "result.mock", "line_number": 222, "usage_type": "attribute"}, {"api_name": "result.as_empty", "line_number": 224, "usage_type": "call"}, {"api_name": "response.SyftError", "line_number": 226, "usage_type": "call"}, {"api_name": "service.service_method", "line_number": 153, "usage_type": "call"}, {"api_name": "user.user_roles.GUEST_ROLE_LEVEL", "line_number": 153, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 156, "usage_type": "name"}, {"api_name": "response.SyftSuccess", "line_number": 156, "usage_type": "name"}, {"api_name": "response.SyftError", "line_number": 156, "usage_type": "name"}, {"api_name": "util.telemetry.instrument", "line_number": 40, "usage_type": "name"}, {"api_name": "serde.serializable.serializable", "line_number": 41, "usage_type": "call"}, {"api_name": "context.AuthedServiceContext", "line_number": 229, "usage_type": "name"}, {"api_name": "policy.policy.OutputHistory", "line_number": 229, "usage_type": "name"}, {"api_name": "context.node.get_service", "line_number": 238, "usage_type": "call"}, {"api_name": "context.node", "line_number": 238, "usage_type": "attribute"}, {"api_name": "service.action.action_object.TwinMode.PRIVATE", "line_number": 240, "usage_type": "attribute"}, {"api_name": "service.action.action_object.TwinMode", "line_number": 240, "usage_type": "name"}, {"api_name": "result.OkErr", "line_number": 242, "usage_type": "argument"}, {"api_name": "result.value", "line_number": 243, "usage_type": "attribute"}, {"api_name": "typing.Any", "line_number": 229, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 252, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 252, "usage_type": "name"}, {"api_name": "action.action_object.ActionObject", "line_number": 261, "usage_type": "argument"}, {"api_name": "types.twin_object.TwinObject", "line_number": 263, "usage_type": "argument"}, {"api_name": "dataset.dataset.Asset", "line_number": 265, "usage_type": "argument"}, {"api_name": "types.uid.UID", "line_number": 268, "usage_type": "argument"}, {"api_name": "service.TYPE_TO_SERVICE", "line_number": 274, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 274, "usage_type": "name"}, {"api_name": "service.SERVICE_TO_TYPES", "line_number": 275, "usage_type": "name"}, {"api_name": "user_code.UserCode", "line_number": 275, "usage_type": "name"}]} +{"seq_id": "22226563", "text": "from django.contrib.auth.decorators import permission_required\nfrom django.shortcuts import render_to_response\nfrom django.template import RequestContext\nfrom gestiones.Producto.altabebida.forms import altaBebidaForm\nfrom gestiones.Producto.producto.models import Bebida\n\n\n@permission_required('Administrador.is_admin', login_url=\"login\")\ndef altabebida(request):\n if request.method == 'POST':\n\n formulario = altaBebidaForm(request.POST)\n\n if formulario.is_valid():\n\n #capturamos y limpiamos datos\n nombre = formulario.cleaned_data['nombre']\n precio = formulario.cleaned_data['precio']\n stock = formulario.cleaned_data['stock']\n activo = formulario.cleaned_data['activo']\n enPromocion = formulario.cleaned_data['enPromocion']\n marca = formulario.cleaned_data['marca']\n descuento = formulario.cleaned_data['descuento']\n seccion = formulario.cleaned_data['seccion']\n\n bebida = Bebida.objects.create(nombre=nombre, precio=precio, stock=stock, activo=activo,\n enPromocion=enPromocion, marca=marca, descuento=descuento, seccion=seccion)\n seccion.bebidas.add(bebida)\n seccion.save()\n\n #mostramos que la operacion fue exitosa\n return render_to_response('Producto/altabebida/altabebidaexito.html', {},\n context_instance=RequestContext(request))\n\n return render_to_response('Producto/altabebida/altabebida.html', {'formulario': formulario},\n context_instance=RequestContext(request))\n\n else:\n\n formulario = altaBebidaForm()\n return render_to_response('Producto/altabebida/altabebida.html', {'formulario': formulario},\n context_instance=RequestContext(request))", "sub_path": "gestiones/Producto/altabebida/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1891, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "gestiones.Producto.altabebida.forms.altaBebidaForm", "line_number": 12, "usage_type": "call"}, {"api_name": "gestiones.Producto.producto.models.Bebida.objects.create", "line_number": 26, "usage_type": "call"}, {"api_name": "gestiones.Producto.producto.models.Bebida.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "gestiones.Producto.producto.models.Bebida", "line_number": 26, "usage_type": "name"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 32, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 33, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 35, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 36, "usage_type": "call"}, {"api_name": "gestiones.Producto.altabebida.forms.altaBebidaForm", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.render_to_response", "line_number": 41, "usage_type": "call"}, {"api_name": "django.template.RequestContext", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.permission_required", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "601625347", "text": "import numpy as np\nfrom scipy.optimize import curve_fit\nimport matplotlib.pyplot as plt\nfrom decimal import Decimal\nfrom copy import deepcopy\n\ndef fitfunction(x, p0, p1, p2, p3):\n y = np.zeros(len(x))\n y += p0 * (x <= p1)\n y += (p2 * (x - p1) + p0 ) * (x <= p3) * (x > p1)\n y += (p2 * (p3 - p1) + p0 ) * (x > p3)\n return y\ndef fitfunction_real(x, p0, p1, p2, p3):\n if x <= p1:\n return p0\n if x <=p3:\n return p2 * (x - p1) + p0\n else:\n return p2 * (p3 - p1) + p0\ndef fitfunction2(x, p0, p1, p2, p3, p4):\n y = np.zeros(len(x))\n y += p0 * (x <= p1)\n y += (p2 * (x - p1) + p0 ) * (x <= p3) * (x > p1)\n y += (p2 * (p3 - p1) + p0 ) * (x > p3)\n return y\ndef poly(x, *argv):\n s = 0\n for i, each in enumerate(argv):\n s += x**i * each\n return s\n\nlabelshift = 0\nnbtag = 2\nhighlow = \"\"\nif nbtag == 1:\n labelshift = 0.55\n g1 = 1\n g2 = 100\n g3 = 0\n g4 = 450\n # def fitfunction1(x, p0, p1, p2):# p5, p6):\n # return poly(x, p0, p1, p2)# p5, p6)\n\n # def fitfunction2(x, p0, p1):#, p2):# p5, p6):\n # return poly(x, p0, p1)#, p2)# p5, p6)\nif nbtag == 2:\n labelshift = 0.55\n if highlow == \"low\":\n plt.ylim(top=3)\n plt.ylim(bottom=-0.4)\n g1 = 1.1\n g2 = 20\n g3 = 0\n g4 = 600\n # def fitfunction1(x, p0, p1, p2):# p5, p6):\n # return poly(x, p0, p1, p2)# p5, p6)\n\n # def fitfunction2(x, p0, p1):#, p2):# p5, p6):\n # return poly(x, p0, p1)#, p2)# p5, p6)\n# if nbtag == 1:\n# labelshift = 0.55\n# #middle = 12\n# middle = 10\n# def fitfunction1(x, p0, p1, p2):# p5, p6):\n# #return poly(x, p0, p1, p2)# p5, p6)\n# return poly(x, p0, p1)\n\n# def fitfunction2(x, p0, p1):#, p2):# p5, p6):\n# #return poly(x, p0, p1)#, p2)# p5, p6)\n# return poly(x, p0)\n# if nbtag == 2:\n# labelshift = 0.55\n# #middle = 10\n# middle = 6\n# def fitfunction1(x, p0, p1, p2):# p5, p6):\n# return poly(x, p0, p1)# p5, p6)\n\n# def fitfunction2(x, p0, p1):#, p2):# p5, p6):\n# return poly(x, p0, p1)#, p2)# p5, p6)\ndata_point = []\nmc_point = []\nmc_error = []\nbin_edge = []\nmc_point_z = []\nmc_error_z = []\n\nfilename = \"pTH-\" + highlow + \"mbbcut-\" + str(nbtag) + \"tag\"\nprint(filename)\n# load data\nwith open(\"output/t_make_plot_rescale/\" + filename + \".csv\") as f:\n for each_line in f:\n data_tem = each_line.split(\",\")\n if len(data_tem) > 3:\n bin_edge.append(float(data_tem[0]))\n data_point.append(float(data_tem[1]))\n mc_point.append(float(data_tem[2]))\n mc_error.append(float(data_tem[3]))\n mc_point_z.append(float(data_tem[4]))\n mc_error_z.append(float(data_tem[5]))\n else:\n bin_edge.append(float(data_tem[0]))\n\n#convert to numpy\ndata_point = np.array(data_point)\nmc_point = np.array(mc_point)\nmc_error = np.array(mc_error)\nbin_edge = np.array(bin_edge)\nmc_error_z = np.array(mc_error_z)\nmc_point_z = np.array(mc_point_z)\nmc_error_other = np.sqrt(mc_error**2 - mc_error_z**2)\n\n# calculate bin centre\nbin_centre = (bin_edge[0:-1] + bin_edge[1:])/2\n\n# pop zero\nmask = data_point != 0\ndata_point = data_point[mask]\nmc_point = mc_point[mask]\nmc_error = mc_error[mask]\nmc_point_z = mc_point_z[mask]\nmc_error_z = mc_error_z[mask]\nbin_centre = bin_centre[mask]\nmc_error_other = mc_error_other[mask]\n\n# calculate diff\nmc_o = mc_point - mc_point_z\ndiff = (data_point - mc_o) /mc_point_z\ndiff_error = np.sqrt(data_point/ mc_point_z**2 + mc_error_other**2/ mc_point_z**2 + (data_point - mc_o)**2/ mc_point_z**4 * mc_error_z**2)\n\n# xnew = np.arange(0, 1300, 1)\n# tck = interpolate.splrep(bin_centre, diff, k=5, w =1/diff_error, s = 40)\n# ynew = interpolate.splev(xnew, tck, der=0)\n\n# cs = interpolate.CubicSpline(bin_centre, diff, bc_type =((1, 0.0), (1, 0.0)))\n\n# plt.figure()\n# # plt.plot(xnew, ynew, 'k')\n# plt.plot(xnew, cs(xnew), 'k')\n# plt.errorbar(bin_centre, diff, yerr=diff_error, fmt='o')\n# #plt.yscale(\"log\")\n# plt.title('Cubic-spline interpolation')\n# plt.show()\n\n#fit\nchi2nod = []\nupper = len(diff_error)\nfor i, each in enumerate(diff_error):\n if i == 0:\n continue\n if each > 0.08:\n upper = i\n break\nprint(bin_centre[upper])\npopt1, pcov1 = curve_fit(fitfunction, bin_centre, diff, sigma=diff_error, p0=[1,0,0,bin_centre[upper]-50], bounds=((-np.inf, 0, -np.inf, 0), (np.inf, np.inf, np.inf, bin_centre[upper])) )\nprint(popt1)\nprint(pcov1)\nprint(np.sqrt(np.diag(pcov1)))\nr = diff - fitfunction(bin_centre, *popt1)\nchisq = sum((r / diff_error) ** 2)\nchi2nod.append(chisq/(-len(popt1) + len(bin_centre)))\nprint(chisq/(-len(popt1) + len(bin_centre)))\n\n\n# makeplot\nplt.errorbar(bin_centre, diff, yerr=diff_error, fmt='k.')\n\nxs = np.linspace(0, bin_centre[-1],10000)\nys1 = []\nys2 = []\nys3 = []\nxs1 = []\nxs2 = []\nxs3 = []\nfor each in xs:\n if each <= popt1[1]:\n xs1.append(each)\n ys1.append(fitfunction_real(each, popt1[0], popt1[1], popt1[2], popt1[3]))\n elif each <= popt1[3]:\n xs2.append(each)\n ys2.append(fitfunction_real(each, popt1[0], popt1[1], popt1[2], popt1[3]))\n else:\n xs3.append(each)\n ys3.append(fitfunction_real(each, popt1[0], popt1[1], popt1[2], popt1[3]))\nplt.plot(xs1, ys1, 'g-')\nplt.plot(xs2, ys2, 'r-')\nplt.plot(xs3, ys3, 'b-')\nplt.xlabel(r\"$p_{TH}$ [GeV]\", fontsize=17)\nplt.ylabel(\"reweight factor\", fontsize=17)\n#plt.ylim([0.5,1.5])\n#plt.yscale(\"log\")\nax = plt.gca()\nplt.text(0.05, 0.1 + labelshift, \"$\\chi^2$/ndf: \" + \"{:.5f}\".format(chi2nod[0]), fontsize=15, transform=ax.transAxes)\n# plt.text(0.05, 0.03 + labelshift, \"green chi2/ndf: \" + \"{:.5f}\".format(chi2nod[1]), fontsize=15, transform=ax.transAxes)\ntitle1 = r\"ATLAS\"\ntitle1_1 = r\"Internal\"\ntitle3 = \"2 lep., \" + str(nbtag) + \" b-tag\"\nplt.text(0.05, 0.3 + labelshift, title1, fontsize=25, transform=ax.transAxes, style='italic', fontweight='bold')\nplt.text(0.327, 0.3 + labelshift, title1_1, fontsize=25, transform=ax.transAxes)\nplt.text(0.05, 0.2 + labelshift, title3, fontsize=18, weight='bold', style='italic', transform=ax.transAxes)\nplt.savefig(\"output/slopefit_zlf/\" + filename + \"polyfitresult.pdf\" ,bbox_inches='tight', pad_inches = 0)\nplt.show()\n\npopt1 = popt1.tolist()\npcov1 = np.sqrt(np.diag(pcov1)).tolist()\n\n# print data\nwith open(\"output/slopefit_zlf/\" + filename + \"polyfitresult.csv\", \"w\") as f:\n for each in popt1:\n f.write(str(Decimal(repr(each))) + ',')\n f.write('\\n')\n for each in pcov1:\n f.write(str(Decimal(repr(each))) + ',')", "sub_path": "run/slopefit_ptH_zlf.py", "file_name": "slopefit_ptH_zlf.py", "file_ext": "py", "file_size_in_byte": 6514, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.zeros", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 128, "usage_type": "call"}, {"api_name": "scipy.optimize.curve_fit", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 154, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.errorbar", "line_number": 165, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 165, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 185, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 185, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 186, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 186, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 187, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 192, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 192, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 197, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 198, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.text", "line_number": 199, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 199, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 200, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 200, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 201, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 201, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 204, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 209, "usage_type": "call"}, {"api_name": "decimal.Decimal", "line_number": 212, "usage_type": "call"}]} +{"seq_id": "411236518", "text": "#!/usr/bin/env python3\nimport io\nimport logging\nimport os\nimport subprocess\nimport unittest\n\nfrom artifacts_dir import get_per_repo_artifacts_dir\nfrom volume_for_repo import get_volume_for_current_repo\n\nfrom ..parse_dump import (\n DumpItems, get_frequency_of_selinux_xattrs, ItemFilters, NAME_TO_ITEM_TYPE,\n parse_btrfs_dump, unquote_btrfs_progs_path,\n)\nfrom ..subvol_path import SubvolPath\n\n# `unittest`'s output shortening makes tests much harder to debug.\nunittest.util._MAX_LENGTH = 10e4\n\n\ndef _sibling_path(rel_path):\n return os.path.join(os.path.dirname(os.path.abspath(__file__)), rel_path)\n\n\ndef _parse_bytes_to_list(s):\n return list(parse_btrfs_dump(io.BytesIO(s)))\n\n\nclass ParseBtrfsDumpTestCase(unittest.TestCase):\n def setUp(self):\n self.maxDiff = 10e4\n\n def test_unquote(self):\n self.assertEqual(\n (b'\\a\\b\\x1b\\f\\n\\r\\t\\v ' br'\\XYZ\\F\\0\\O\\P'),\n unquote_btrfs_progs_path(\n # Special escapes\n br'\\a\\b\\e\\f\\n\\r\\t\\v\\ \\\\' +\n # Octal escapes\n ''.join(f'\\\\{ord(c):o}' for c in 'XYZ').encode('ascii') +\n # Unrecognized escapes will be left alone\n br'\\F\\0\\O\\P'\n )\n )\n\n def test_ensure_print_demo_dump_covers_all_operations(self):\n print_demo_dump_sh = _sibling_path('print_demo_dump.sh')\n out_bytes = subprocess.check_output(\n ['sudo', print_demo_dump_sh],\n cwd=get_volume_for_current_repo(1e8, get_per_repo_artifacts_dir()),\n )\n out_lines = out_bytes.rstrip(b'\\n').split(b'\\n')\n # Ensure we have exercised all the implemented operations:\n # https://github.com/kdave/btrfs-progs/blob/master/send-dump.c#L319\n expected_ops = {\n 'chmod',\n 'chown',\n 'clone',\n 'link',\n 'mkdir',\n 'mkfifo',\n 'mkfile',\n 'mknod',\n 'mksock',\n 'remove_xattr',\n 'rename',\n 'rmdir',\n 'set_xattr',\n 'snapshot',\n 'subvol',\n 'symlink',\n 'truncate',\n 'unlink',\n 'update_extent',\n 'utimes',\n 'write',\n }\n self.assertEqual(\n {n.decode() for n in NAME_TO_ITEM_TYPE.keys()},\n expected_ops,\n )\n self.assertEqual(\n expected_ops,\n {l.split(b' ', 1)[0].decode() for l in out_lines if l},\n )\n items = _parse_bytes_to_list(out_bytes)\n # We an item per line, and the items cover the expected operations.\n self.assertEqual(len(items), len(out_lines))\n self.assertEqual(\n {getattr(DumpItems, op_name) for op_name in expected_ops},\n {i.__class__ for i in items},\n )\n\n # The reason we want to parse a gold file instead of, as above, running\n # `print_demo_dump.sh` is explained in `update_gold_print_demo_dump.sh`.\n def test_verify_gold_parse(self):\n with open(_sibling_path('gold_print_demo_dump.out'), 'rb') as infile:\n lines = infile.readlines()\n build_start_time, build_end_time = (\n float(l) for l in [lines[0], lines[-1]]\n )\n orig_items = _parse_bytes_to_list(b''.join(lines[1:-1]))\n items = orig_items\n\n # Our test program does not touch the SELinux context, so if it's\n # set, it will be set to the default, and we can just filter out the\n # most frequent value. We don't want to drop selinux attributes\n # blindly because having varying contexts suggests something broken\n # about the test or our environment.\n selinux_freqs = get_frequency_of_selinux_xattrs(orig_items)\n self.assertGreater(len(selinux_freqs), 0) # `gold` has SELinux attrs\n max_name, _count = max(selinux_freqs.items(), key=lambda p: p[1])\n logging.info(f'This test ignores SELinux xattrs set to {max_name}')\n items = ItemFilters.selinux_xattr(\n items,\n discard_fn=lambda _path, ctx: ctx == max_name,\n )\n items = ItemFilters.normalize_utimes(\n items, start_time=build_start_time, end_time=build_end_time,\n )\n items = list(items)\n\n di = DumpItems\n\n def p(path):\n if isinstance(path, str): # forgive missing `b`s, it's a test\n path = path.encode()\n return SubvolPath._new(path)\n\n def chown(path):\n return di.chown(path=p(path), gid=0, uid=0)\n\n def chmod(path, mode=0o644):\n return di.chmod(path=p(path), mode=mode)\n\n def utimes(path):\n return di.utimes(\n path=p(path),\n atime=build_start_time,\n mtime=build_start_time,\n ctime=build_start_time,\n )\n\n def base_metadata(path, mode=0o644):\n return [chown(path), chmod(path, mode), utimes(path)]\n\n # Future: if we end up doing a lot of mid-list insertions, we can\n # autogenerate the temporary names to match what btrfs does.\n def and_rename(item, real_name, utimes_parent=True):\n yield item\n renamed_item = di.rename(\n path=item.path,\n dest=p(\n os.path.join(os.path.dirname(bytes(item.path)), real_name)\n ),\n )\n yield renamed_item\n if utimes_parent: # Rarely, `btrfs send` breaks the pattern.\n yield utimes(os.path.dirname(bytes(renamed_item.dest)))\n\n # These make it quite easy to update the test after you run\n # `update_gold_print_demo_dump.sh`.\n uuid_create = b'e34c8a50-ffc1-2d41-ab67-9219669ea9f3'\n transid_create = 1993\n uuid_mutate = b'ed28f410-3173-b64f-8769-0ba7c3b6ac6d'\n transid_mutate = 1996\n temp_path_middles = {'create_ops': 1991, 'mutate_ops': 1995}\n temp_path_counter = 256 # I have never seen this initial value change.\n\n def temp_path(prefix):\n nonlocal temp_path_counter\n temp_path_counter += 1\n mid = temp_path_middles[prefix]\n return p(f'{prefix}/o{temp_path_counter}-{mid}-0')\n\n self.assertEqual([\n di.subvol(\n path=p('create_ops'), uuid=uuid_create, transid=transid_create,\n ),\n *base_metadata('create_ops', mode=0o755),\n\n *and_rename(di.mkdir(path=temp_path('create_ops')), b'hello'),\n di.set_xattr(\n path=p('create_ops/hello'),\n name=b'user.test_attr',\n data=b'chickens',\n len=8,\n ),\n *base_metadata('create_ops/hello', mode=0o755),\n\n *and_rename(\n di.mkdir(path=temp_path('create_ops')), b'dir_to_remove'\n ),\n *base_metadata('create_ops/dir_to_remove', mode=0o755),\n\n *and_rename(\n di.mkfile(path=temp_path('create_ops')), b'goodbye',\n utimes_parent=False,\n ),\n di.link(\n path=p('create_ops/hello/world'), dest=p('create_ops/goodbye'),\n ),\n utimes('create_ops'),\n utimes('create_ops/hello'),\n di.truncate(path=p('create_ops/goodbye'), size=0),\n *base_metadata('create_ops/goodbye'),\n\n *and_rename(di.mknod(\n path=temp_path('create_ops'), mode=0o60644, dev=0x7a539b7,\n ), b'buffered'),\n *base_metadata('create_ops/buffered'),\n\n *and_rename(di.mknod(\n path=temp_path('create_ops'), mode=0o20644, dev=0x7a539b7,\n ), b'unbuffered'),\n *base_metadata('create_ops/unbuffered'),\n\n *and_rename(di.mkfifo(path=temp_path('create_ops')), b'fifo'),\n *base_metadata('create_ops/fifo'),\n\n *and_rename(\n di.mksock(path=temp_path('create_ops')), b'unix_sock',\n ),\n *base_metadata('create_ops/unix_sock', mode=0o755),\n\n *and_rename(di.symlink(\n path=temp_path('create_ops'), dest=b'hello/world',\n ), b'goodbye_symbolic'),\n chown('create_ops/goodbye_symbolic'),\n utimes('create_ops/goodbye_symbolic'),\n\n *and_rename(\n di.mkfile(path=temp_path('create_ops')), b'1MB_nuls',\n ),\n di.update_extent(\n path=p('create_ops/1MB_nuls'), offset=0, len=2**20,\n ),\n di.truncate(path=p('create_ops/1MB_nuls'), size=2**20),\n *base_metadata('create_ops/1MB_nuls'),\n\n *and_rename(\n di.mkfile(path=temp_path('create_ops')), b'1MB_nuls_clone',\n ),\n di.clone(\n path=p('create_ops/1MB_nuls_clone'), offset=0, len=2**20,\n from_file=p('create_ops/1MB_nuls'), clone_offset=0,\n ),\n di.truncate(path=p('create_ops/1MB_nuls_clone'), size=2**20),\n *base_metadata('create_ops/1MB_nuls_clone'),\n\n *and_rename(\n di.mkfile(path=temp_path('create_ops')), b'zeros_hole_zeros',\n ),\n di.update_extent(\n path=p('create_ops/zeros_hole_zeros'), offset=0, len=16384,\n ),\n di.update_extent(\n path=p('create_ops/zeros_hole_zeros'), offset=32768, len=16384,\n ),\n di.truncate(path=p('create_ops/zeros_hole_zeros'), size=49152),\n *base_metadata('create_ops/zeros_hole_zeros'),\n\n di.snapshot(\n path=p('mutate_ops'),\n uuid=uuid_mutate,\n transid=transid_mutate,\n parent_uuid=uuid_create,\n parent_transid=transid_create,\n ),\n utimes('mutate_ops'),\n di.rename(\n path=p('mutate_ops/hello'), dest=p('mutate_ops/hello_renamed'),\n ),\n utimes('mutate_ops'),\n utimes('mutate_ops'), # `btrfs send` is not so parsimonious\n\n di.remove_xattr(\n path=p('mutate_ops/hello_renamed'), name=b'user.test_attr',\n ),\n utimes('mutate_ops/hello_renamed'),\n\n di.rmdir(path=p('mutate_ops/dir_to_remove')),\n utimes('mutate_ops'),\n\n di.link(\n path=p('mutate_ops/farewell'), dest=p('mutate_ops/goodbye'),\n ),\n di.unlink(path=p('mutate_ops/goodbye')),\n di.unlink(path=p('mutate_ops/hello_renamed/world')),\n utimes('mutate_ops'),\n utimes('mutate_ops'),\n utimes('mutate_ops/hello_renamed'),\n di.truncate(path=p('mutate_ops/farewell'), size=0),\n utimes('mutate_ops/farewell'),\n\n *and_rename(\n di.mkfile(path=temp_path('mutate_ops')), b'hello_renamed/een',\n ),\n di.write(path=p('mutate_ops/hello_renamed/een'), offset=0, len=5),\n di.truncate(path=p('mutate_ops/hello_renamed/een'), size=5),\n *base_metadata('mutate_ops/hello_renamed/een'),\n ], items)\n\n def test_common_errors(self):\n ok_line = b'mkfile ./cat\\\\ and\\\\ dog\\n' # Drive-by test of unquoting\n self.assertEqual(\n [DumpItems.mkfile(path=SubvolPath._new(b'cat and dog'))],\n _parse_bytes_to_list(ok_line),\n )\n\n with self.assertRaisesRegex(RuntimeError, 'has unexpected format:'):\n _parse_bytes_to_list(b' ' + ok_line)\n\n with self.assertRaisesRegex(RuntimeError, \"unknown item type b'Xmkfi\"):\n _parse_bytes_to_list(b'X' + ok_line)\n\n def test_set_xattr_errors(self):\n\n def make_line(len_k='len', len_v=7, name_k='name', data_k='data'):\n return (\n 'set_xattr ./subvol/file '\n f'{name_k}=MY_ATTR {data_k}=MY_DATA {len_k}={len_v}\\n'\n ).encode('ascii')\n\n # Before breaking it, ensure that `make_line` actually works\n for l in [7, 8]: # \\0-terminated would add 1 char\n self.assertEqual(\n [DumpItems.set_xattr(\n path=SubvolPath._new(b'subvol/file'),\n name=b'MY_ATTR', data=b'MY_DATA', len=l,\n )],\n _parse_bytes_to_list(make_line(len_v=l)),\n )\n\n for bad_line in [\n # Bad field name, non-int value, value inconsistent with data,\n make_line(len_k='Xlen'), make_line(len_v='x7'), make_line(len_v=9),\n # Swap name & data fields, try a bad one\n make_line(data_k='name', name_k='data'), make_line(name_k='nom'),\n ]:\n with self.assertRaisesRegex(RuntimeError, 'in line details:'):\n _parse_bytes_to_list(bad_line)\n\n\nif __name__ == '__main__':\n unittest.main()\n", "sub_path": "infra_macros/macro_lib/convert/container_image/btrfs_diff/tests/test_parse_dump.py", "file_name": "test_parse_dump.py", "file_ext": "py", "file_size_in_byte": 12874, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "unittest.util", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 22, "usage_type": "call"}, {"api_name": "parse_dump.parse_btrfs_dump", "line_number": 26, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 26, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 29, "usage_type": "attribute"}, {"api_name": "parse_dump.unquote_btrfs_progs_path", "line_number": 36, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 48, "usage_type": "call"}, {"api_name": "volume_for_repo.get_volume_for_current_repo", "line_number": 50, "usage_type": "call"}, {"api_name": "artifacts_dir.get_per_repo_artifacts_dir", "line_number": 50, "usage_type": "call"}, {"api_name": "parse_dump.NAME_TO_ITEM_TYPE.keys", "line_number": 79, "usage_type": "call"}, {"api_name": "parse_dump.NAME_TO_ITEM_TYPE", "line_number": 79, "usage_type": "name"}, {"api_name": "parse_dump.DumpItems", "line_number": 90, "usage_type": "argument"}, {"api_name": "parse_dump.get_frequency_of_selinux_xattrs", "line_number": 110, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 113, "usage_type": "call"}, {"api_name": "parse_dump.ItemFilters.selinux_xattr", "line_number": 114, "usage_type": "call"}, {"api_name": "parse_dump.ItemFilters", "line_number": 114, "usage_type": "name"}, {"api_name": "parse_dump.ItemFilters.normalize_utimes", "line_number": 118, "usage_type": "call"}, {"api_name": "parse_dump.ItemFilters", "line_number": 118, "usage_type": "name"}, {"api_name": "parse_dump.DumpItems", "line_number": 123, "usage_type": "name"}, {"api_name": "subvol_path.SubvolPath._new", "line_number": 128, "usage_type": "call"}, {"api_name": "subvol_path.SubvolPath", "line_number": 128, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 159, "usage_type": "call"}, {"api_name": "os.path", "line_number": 159, "usage_type": "attribute"}, {"api_name": "parse_dump.DumpItems.mkfile", "line_number": 307, "usage_type": "call"}, {"api_name": "parse_dump.DumpItems", "line_number": 307, "usage_type": "name"}, {"api_name": "subvol_path.SubvolPath._new", "line_number": 307, "usage_type": "call"}, {"api_name": "subvol_path.SubvolPath", "line_number": 307, "usage_type": "name"}, {"api_name": "parse_dump.DumpItems.set_xattr", "line_number": 328, "usage_type": "call"}, {"api_name": "parse_dump.DumpItems", "line_number": 328, "usage_type": "name"}, {"api_name": "subvol_path.SubvolPath._new", "line_number": 329, "usage_type": "call"}, {"api_name": "subvol_path.SubvolPath", "line_number": 329, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 346, "usage_type": "call"}]} +{"seq_id": "614492034", "text": "import cv2\nimport numpy as np\n\n# capturing the video\ncapturedVideo = cv2.VideoCapture('leapord.mp4')\n\n# reading two frames\ncount, frame1 = capturedVideo.read()\ncount, frame2 = capturedVideo.read()\n\nwhile capturedVideo.isOpened():\n\n # find the difference between first fame and second frame\n differenceOfFrames = cv2.absdiff(frame1, frame2)\n\n # converting frames from BGR to GRAY , Easy to find contours in the gray scale mode\n grayFrame = cv2.cvtColor(differenceOfFrames, cv2.COLOR_BGR2GRAY)\n\n # blurring(frame name, k_size, sigmaX value )\n blur = cv2.GaussianBlur(grayFrame, (5, 5), 0)\n\n # _ we don't need first variable(src, threshold_value, max , type)\n _, thresh = cv2.threshold(blur, 20, 255, cv2.THRESH_BINARY)\n\n dilated = cv2.dilate(thresh, None, iterations=3)\n\n # going to find the contours in the dilated image(img, mode, method )\n contours, _ = cv2.findContours(dilated, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\n\n # iterate through all contours\n for contour in contours:\n\n # save all coordinates of saved contours\n (x, y, w, h) = cv2.boundingRect(contour)\n\n if cv2.contourArea(contour) < 400:\n continue\n elif cv2.contourArea(contour) > 400:\n cv2.rectangle(frame1, (x, y), (x+w, y+h), (0, 255, 0), 2)\n\n cv2.putText(frame1, \"Status: {}\".format('Warning'), (10, 20), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 2)\n cv2.putText(frame1, \"Status: {}\".format('Animal Movement Detected'), (10, 50), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0,255), 2)\n\n # drawing the contours in the original frame (frame, contours, contourID, color, thickness)\n # cv2.drawContours(frame1, contours, -1, (0, 255, 0), 2)\n\n # displaying frame\n cv2.imshow(\"feed\", frame1)\n\n frame1 = frame2\n count, frame2 = capturedVideo.read()\n\n if cv2.waitKey(40) == 27:\n break\n\ncv2.destroyAllWindows()\ncapturedVideo.release()", "sub_path": "CCTV Animal Tracking/CCTV Animal Tracking.py", "file_name": "CCTV Animal Tracking.py", "file_ext": "py", "file_size_in_byte": 1921, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cv2.VideoCapture", "line_number": 5, "usage_type": "call"}, {"api_name": "cv2.absdiff", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.GaussianBlur", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 23, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 28, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 34, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.contourArea", "line_number": 38, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 39, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 41, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 41, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 42, "usage_type": "call"}, {"api_name": "cv2.FONT_HERSHEY_SIMPLEX", "line_number": 42, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 48, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 56, "usage_type": "call"}]} +{"seq_id": "348241885", "text": "import grpc\n\nimport demo_pb2\nimport demo_pb2_grpc\n\ndef run():\n channel = grpc.insecure_channel('localhost:50051')\n stub = demo_pb2_grpc.GreeterStub(channel)\n response = stub.SayHello(demo_pb2.HelloRequest(name=\"goodspeed\"))\n print (\"Greeter client received: \" + response.message)\n\n\nif __name__ == '__main__':\n run()", "sub_path": "python3/grpc/greetter_client.py", "file_name": "greetter_client.py", "file_ext": "py", "file_size_in_byte": 331, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "grpc.insecure_channel", "line_number": 7, "usage_type": "call"}, {"api_name": "demo_pb2_grpc.GreeterStub", "line_number": 8, "usage_type": "call"}, {"api_name": "demo_pb2.HelloRequest", "line_number": 9, "usage_type": "call"}]} +{"seq_id": "278850262", "text": "from DataModelDict import DataModelDict as DM\n\nfrom . import PotentialLAMMPSBuilder\nfrom ..tools import aslist\n\nclass EamBuilder(PotentialLAMMPSBuilder):\n \"\"\" \n PotentialLAMMPS builder class for the classic eam style only, which uses\n pair_coeff lines of the form:\n \n pair_coeff 1 1 paramfile1\n pair_coeff 2 2 paramfile2\n\n Note: other EAM styles like eam/alloy, etc. should use ParamFileBuilder!\n \"\"\"\n\n def __init__(self, paramfiles=None, **kwargs):\n \"\"\"\n Class initializer\n\n Parameters\n ----------\n paramfiles : str or list, optional\n The name(s) of the potential's parameter file(s). There should be\n one parameter file for each element model.\n **kwargs : any, optional\n Any other keyword parameters accepted by PotentialLAMMPSBuilder.\n Default values used by this class: units='metal' and\n atom_style='atomic'.\n \"\"\"\n # Set default values for format\n kwargs['units'] = kwargs.get('units', 'metal')\n kwargs['atom_style'] = kwargs.get('atom_style', 'atomic')\n kwargs['pair_style'] = kwargs.get('pair_style', 'eam')\n\n # Call PotentialLAMMPS's init\n PotentialLAMMPSBuilder.__init__(self, **kwargs)\n \n # Set format-specific parameters\n self.paramfiles = paramfiles\n \n @property\n def paramfiles(self):\n \"The names of the parameter files to use\"\n return self.__paramfiles\n\n @paramfiles.setter\n def paramfiles(self, value):\n if value is not None:\n value = aslist(value)\n self.__paramfiles = value\n\n def buildpaircoeff(self):\n \n if self.symbols is not None:\n symbols = self.symbols\n else:\n symbols = self.elements\n\n if len(symbols) != len(self.paramfiles):\n raise ValueError('a paramfile is needed for each symbol/element')\n\n paircoeffs = []\n for symbol, paramfile in zip(symbols, self.paramfiles):\n paircoeff = DM()\n paircoeff['interaction'] = DM([('symbol', [symbol, symbol])])\n paircoeff['term'] = DM([('file', paramfile)])\n paircoeffs.append(paircoeff)\n \n if len(paircoeffs) == 0:\n paircoeffs = paircoeffs[0]\n \n return paircoeffs\n\n @property\n def supported_pair_styles(self):\n \"\"\"tuple : The list of known pair styles that use this format.\"\"\"\n return ('eam',)", "sub_path": "potentials/build/EamBuilder.py", "file_name": "EamBuilder.py", "file_ext": "py", "file_size_in_byte": 2498, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "tools.aslist", "line_number": 50, "usage_type": "call"}, {"api_name": "DataModelDict.DataModelDict", "line_number": 65, "usage_type": "call"}, {"api_name": "DataModelDict.DataModelDict", "line_number": 66, "usage_type": "call"}, {"api_name": "DataModelDict.DataModelDict", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "66495834", "text": "\n# coding: utf-8\n\n# In[373]:\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math\nimport array\nimport pandas as pd\nget_ipython().magic('matplotlib inline')\n\n\n# In[374]:\n\ndata = np.loadtxt('ellipsoids.txt')\nN=1500\n\n\n# In[375]:\n\n#find covariance matrix\ncov1 = np.cov(data.T)\nX=np.mean(data,axis=0)\nfor i in range(3):\n data[:,i]=data[:,i]-X[i]\nmat = data.T@data\ncov = mat/(N-1)\nprint(cov)\n\n\n# In[376]:\n\n#find the eigenvactor and eigenvalues of cov matrix\nvalue,vector = np.linalg.eig(cov) #vector[:,i] it the eigenvrctor corresponding to the eigenvalue value[i]\nprint(\"eigenvectors are\",vector)\nprint(\"eigenvalues are\",value)\n\n\n# In[377]:\n\n#plot the projecttion of the data into the 2-D principle components\nmat_2d = data@vector[:,0:2]\nprint(mat_2d.shape)\nplt.scatter(mat_2d[:,0],mat_2d[:,1],c='b',s=0.5)\nplt.xlabel('dimension 1')\nplt.ylabel('dimension 2')\nplt.title('Ellipsoids')\nplt.show()\n\n\n# In[378]:\n\n#plot the projection of the data into the 1-D principle components\nmat_1d = data@vector[:,0]\nplt.scatter(mat_1d,np.zeros(1500),color='blue',s=0.5)\nplt.xlabel('dimension 1')\nplt.title('Ellipsoids')\nplt.show()\n\n\n# In[379]:\n\nx=y=np.arange(-2,2,0.1)\nx,y=np.meshgrid(x,y)\nplt.contour(x,y,7/8*x**2-(3**(1/2)/4)*x*y+5/8*y**2,[1])\nplt.axis('scaled')\nplt.xlabel('x1')\nplt.ylabel('x2')\nplt.grid()\nplt.show()\n\n\n# # \n", "sub_path": "Data_PreP/Q4.py", "file_name": "Q4.py", "file_ext": "py", "file_size_in_byte": 1322, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.loadtxt", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.cov", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.linalg.eig", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 35, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.contour", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}]} +{"seq_id": "531494655", "text": "# Example from http://flask.pocoo.org/docs/0.12/patterns/fileuploads/\nimport os\nfrom flask import Flask, request, redirect, url_for, jsonify\nfrom flask import send_from_directory\nfrom werkzeug.utils import secure_filename\n\nUPLOAD_FOLDER = 'uploads'\nALLOWED_EXTENSIONS = set(['png', 'jpg', 'jpeg', 'gif'])\n\napp = Flask(__name__)\napp.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER\n\n\nimport numpy as np\nimport tensorflow as tf\n\nimport keras\nfrom keras.preprocessing import image\nfrom keras.applications.xception import (\n Xception, preprocess_input, decode_predictions)\n# from tensorflow.contrib.keras.python.keras.backend import clear_session\nfrom keras import backend as K\n\nmodel = Xception(\n include_top=True,\n weights='imagenet')\ngraph = K.get_session().graph\n\nimage_size = (299, 299)\n\n\ndef allowed_file(filename):\n return '.' in filename and \\\n filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS\n\n@app.route('/', methods=['GET', 'POST'])\ndef upload_file():\n data = {\"success\": False}\n if request.method == 'POST':\n # check if the post request has the file part\n if 'file' not in request.files:\n return redirect(request.url)\n file = request.files['file']\n # if user does not select file, browser also\n # submit a empty part without filename\n if file.filename == '':\n return redirect(request.url)\n if file and allowed_file(file.filename):\n filename = secure_filename(file.filename)\n filepath = os.path.join(app.config['UPLOAD_FOLDER'])\n file.save(os.path.join(app.config['UPLOAD_FOLDER'], filename))\n # Preprocess image for model prediction\n # This step handles scaling and normalization for Xception\n global graph\n with graph.as_default():\n img = image.load_img(os.path.join(app.config['UPLOAD_FOLDER'], file.filename), target_size=image_size)\n x = image.img_to_array(img)\n x = np.expand_dims(x, axis=0)\n x = preprocess_input(x)\n predictions = model.predict(x)\n results = decode_predictions(predictions, top=3)\n # print(results)\n data[\"predictions\"] = []\n\n # loop over the results and add them to the list of\n # returned predictions\n for (xceptionID, label, prob) in results[0]:\n r = {\"label\": label, \"probability\": float(prob)}\n data[\"predictions\"].append(r)\n\n # indicate that the request was a success\n data[\"success\"] = True\n return jsonify(data)\n # return redirect(url_for('uploaded_file',\n # filename=filename))\n return '''\n \n Upload new File\n

Upload new File

\n
\n

\n \n

\n '''\n@app.route('/uploads/')\ndef uploaded_file(filename):\n return send_from_directory(app.config['UPLOAD_FOLDER'],\n filename)\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n", "sub_path": "Lesson-Plans/Week-21-MachineLearning/4/Extra_Content/Alt_Ins_Flask_CNN/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 3248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 10, "usage_type": "call"}, {"api_name": "keras.applications.xception.Xception", "line_number": 24, "usage_type": "call"}, {"api_name": "keras.backend.get_session", "line_number": 27, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 27, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 41, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 42, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 42, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 42, "usage_type": "name"}, {"api_name": "flask.request.files", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.request.url", "line_number": 47, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 47, "usage_type": "name"}, {"api_name": "werkzeug.utils.secure_filename", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.load_img", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 56, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.image.img_to_array", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.preprocessing.image", "line_number": 57, "usage_type": "name"}, {"api_name": "numpy.expand_dims", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.applications.xception.preprocess_input", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.applications.xception.decode_predictions", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.send_from_directory", "line_number": 87, "usage_type": "call"}]} +{"seq_id": "513887987", "text": "import numpy as np\nimport scipy.linalg as la\nimport random\nfrom functools import reduce\nimport os\n\ndef get_b(n, str):\n if (str == \"r\"):\n arr = [random.uniform(-100, 100) for i in range(n)]\n else:\n arr = [float(i) for i in input().split()]\n return np.array(arr)\n\ndef get_matrix(n, str):\n if (str == \"r\"):\n arr = [[random.uniform(-100, 100) for i in range(n)] for j in range(n)]\n else:\n n = int(input())\n arr = []\n for i in range(n):\n arr.append([float(i) for i in input().split()])\n\n return np.array(arr)\n\ndef correct(bl, str):\n print(str + \" is correct\\n\") if bl else print(str + \" is NOT correct\\n\")\n\ndef main():\n print(\"Rand(r) or stdin(s)?\")\n str = input()\n\n n = random.randint(2, 7)\n a = get_matrix(n, str)\n b = get_b(n, str)\n pluq = get_pivot_q_lu(a)\n\n if (is_zero_det(pluq)):\n print(\"matrix has zero det\")\n x = np.linalg.lstsq(a, b)[0]\n x = np.array(x)\n print(\"one solution x: \")\n print(x)\n #\n print(\"a * x\")\n print(a @ x)\n print(\"b\")\n print(b)\n #\n correct(np.allclose(a @ x, b), \"one solution\")\n os._exit(-1)\n\n print(\"A: \")\n print(a)\n print(\"P: \")\n print(pluq[0])\n print(\"Q: \")\n print(pluq[3])\n print(\"L: \")\n print(pluq[1])\n print(\"U: \")\n print(pluq[2])\n correct(np.allclose(pluq[0][0] @ a @ pluq[3][0], pluq[1] @ pluq[2]), \"PLUQ decomposition\")\n\n print(\"b:\")\n print(b)\n x = solve_sys_q(a, b, pluq)\n print(\"x:\")\n print(x)\n correct(np.allclose(x, la.solve(a, b)), \"sys solution\")\n\n print(\"det:\")\n print(det(a))\n correct(np.allclose(det(a), la.det(a)), \"det\")\n\n a_i = inv(a, pluq)\n print(\"a inversed:\")\n print(a_i)\n correct(np.allclose(a_i, la.inv(a)), \"inversion\")\n\n print(\"cond number\")\n print(np.linalg.cond(a))\n\ndef pivot(a):\n n = (a.shape)[0]\n a1 = a.copy()\n id = [[float(i == j) for i in range(n)] for j in range(n)]\n id_det = 1\n for i in range(n):\n maxc, row = a1[i][i], i\n for j in range(i, n):\n if (a1[j][i] > maxc):\n maxc, row = a1[j][i], j\n if (i != row):\n id_det *= -1\n id[i], id[row] = id[row], id[i]\n tmp = a1[row]\n a1[row] = a1[i]\n a1[i] = tmp\n\n return (np.array(id), id_det)\n\ndef pivot_q(a):\n t = pivot(a.T)\n return (t[0].T, t[1])\n\ndef get_lu(a):\n n = (a.shape)[0]\n l = [[0.0 for x in range(n)] for y in range(n)]\n u = [[0.0 for x in range(n)] for y in range(n)]\n\n for j in range(n):\n\n l[j][j] = 1.0\n for i in range(j + 1):\n s1 = sum(u[k][j] * l[i][k] for k in range(i))\n u[i][j] = a[i][j] - s1\n\n for i in range(j, n):\n s2 = sum(u[k][j] * l[i][k] for k in range(j))\n l[i][j] = (a[i][j] - s2) / u[j][j]\n\n return [np.array(l), np.array(u)]\n\ndef get_pivot_lu(a):\n p = pivot(a)\n return [p] + get_lu(p[0] @ a)\n\ndef get_pivot_q_lu(a):\n q = pivot_q(a)\n return get_pivot_lu(a @ q[0]) + [q]\n\ndef has_nan(a):\n n = a.shape[0]\n for i in range(n):\n for j in range(n):\n if np.isnan(a[i][j]):\n return True\n return False\n\ndef is_zero_det(pluq):\n if (has_nan(pluq[1]) or has_nan(pluq[2])):\n return True\n return False\n\ndef solve_sys_tr_l(a, b):\n x = []\n n = a.shape[0]\n x.append(b[0] / a[0][0])\n for i in range(1, n):\n x.append( (b[i] - sum(a[i][j] * x[j] for j in range(i)))/a[i][i] )\n return np.array(x)\n\ndef solve_sys_tr_r(a, b):\n x = []\n n = a.shape[0]\n x = [0.0 for i in range(n)]\n x[n-1] = b[n-1] / a[n-1][n-1]\n for i in range(n-2, -1, -1):\n x[i] = (b[i] - sum(a[i][j] * x[j] for j in range(i+1, n)))/a[i][i]\n return np.array(x)\n\ndef solve_sys(a, b, plu=None):\n if (plu == None):\n plu = get_pivot_lu(a)\n\n #y = la.solve(plu[1], plu[0][0] @ b)\n y = solve_sys_tr_l(plu[1], plu[0][0] @ b)\n #return la.solve(plu[2], y)\n return solve_sys_tr_r(plu[2], y)\n\ndef solve_sys_q(a, b, pluq=None):\n if (pluq == None):\n pluq = get_pivot_q_lu(a)\n '''\n y = la.solve(pluq[1], pluq[0][0] @ b)\n x = la.solve(pluq[2], y)\n return pluq[3][0] @ x\n '''\n y = solve_sys_tr_l(pluq[1], pluq[0][0] @ b)\n x = solve_sys_tr_r(pluq[2], y)\n return pluq[3][0] @ x\n\ndef det_of_tr(a):\n n = a.shape[0]\n res = 1\n for i in range(n):\n res *= a[i][i]\n return res\n\ndef det(a, pluq=None):\n if (pluq == None):\n pluq = get_pivot_q_lu(a)\n '''\n res = 1\n for i in range(4):\n res *= la.det(pluq[i])\n return res\n '''\n return pluq[0][1] * det_of_tr(pluq[1]) * det_of_tr(pluq[2]) * pluq[3][1]\n\ndef inv(a, pluq=None):\n if (pluq == None):\n pluq = get_pivot_q_lu(a)\n n = a.shape[0]\n res = []\n for i in range(n):\n e = [0.0 for x in range(n)]\n e[i] = 1.0\n #solve_sys_q(a, e, pluq)\n res.append(solve_sys_q(a, e, pluq))\n return np.array(res).T\n\nif __name__ == \"__main__\":\n main()\n\n'''\na = get_matrix()\nprint(a)\nprint(pivot_q(a))\nprint(a @ pivot_q(a))\n'''\n\n'''\nprint(solve_sys(get_matrix(), get_b()))\nprint(\"______________\")\nprint(solve_sys_q(get_matrix(), get_b()))\nprint(\"______________\")\nprint(la.solve(get_matrix(), get_b()))\n'''\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n#\n", "sub_path": "matrix/lu.py", "file_name": "lu.py", "file_ext": "py", "file_size_in_byte": 5346, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "random.uniform", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 12, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 23, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.linalg.lstsq", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 49, "usage_type": "call"}, {"api_name": "os._exit", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 69, "usage_type": "call"}, {"api_name": "scipy.linalg.solve", "line_number": 69, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 69, "usage_type": "name"}, {"api_name": "numpy.allclose", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.linalg.det", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.allclose", "line_number": 78, "usage_type": "call"}, {"api_name": "scipy.linalg.inv", "line_number": 78, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.linalg.cond", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 81, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.isnan", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 211, "usage_type": "call"}]} +{"seq_id": "573450340", "text": "from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom .models import Mom, Child\n\n# Create your views here.\ndef index(request):\n return HttpResponse(\"Test URL\")\n\n\ndef getmom(request):\n print('*** in moms ---')\n mummy = Mom.objects.all()\n for eachmom in mummy:\n print(f'MOM: {eachmom.mom_first_name} {eachmom.mom_last_name}')\n for eachchild in Child.objects.filter(child_mom__mom_first_name=eachmom.mom_first_name):\n print('Kid')\n return HttpResponse(\"mom\")\n\n\ndef getchild(request):\n mummy = Mom.objects.all()\n for eachmom in mummy:\n\n for eachchild in Child.objects.filter(child_mom__mom_first_name=eachmom.mom_first_name):\n print(eachchild.child_first_name)\n\n print(f'MOM: {eachmom.mom_first_name} {eachmom.mom_last_name}')\n\n return HttpResponse(\"child\")\n\n\n\n # return HttpResponse(\"get child\")\n\n", "sub_path": "theProject/theApp/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.http.HttpResponse", "line_number": 7, "usage_type": "call"}, {"api_name": "models.Mom.objects.all", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Mom.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Mom", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Child.objects.filter", "line_number": 15, "usage_type": "call"}, {"api_name": "models.Child.objects", "line_number": 15, "usage_type": "attribute"}, {"api_name": "models.Child", "line_number": 15, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 17, "usage_type": "call"}, {"api_name": "models.Mom.objects.all", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Mom.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Mom", "line_number": 21, "usage_type": "name"}, {"api_name": "models.Child.objects.filter", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Child.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.Child", "line_number": 24, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 29, "usage_type": "call"}]} +{"seq_id": "370752066", "text": "# Copyright (c) 2014-2017, iocage\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted providing that the following conditions\n# are met:\n# 1. Redistributions of source code must retain the above copyright\n# notice, this list of conditions and the following disclaimer.\n# 2. Redistributions in binary form must reproduce the above copyright\n# notice, this list of conditions and the following disclaimer in the\n# documentation and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR\n# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY\n# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL\n# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS\n# OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)\n# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,\n# STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING\n# IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n\"\"\"start module for the cli.\"\"\"\nimport click\n\nimport iocage.lib.errors\nimport iocage.lib.Jails\nimport iocage.lib.Logger\nimport iocage.lib.Config.Jail.File.Fstab\n\n__rootcmd__ = True\n\n\n@click.command(name=\"start\", help=\"Starts the specified jails or ALL.\")\n@click.pass_context\n@click.argument(\"jails\", nargs=-1)\ndef cli(ctx, jails):\n \"\"\"\n Update Jails\n \"\"\"\n logger = ctx.parent.logger\n print_function = ctx.parent.print_events\n\n if len(jails) == 0:\n logger.error(\"No jail selector provided\")\n exit(1)\n\n zfs = iocage.lib.ZFS.ZFS()\n zfs.logger = logger\n host = iocage.lib.Host.Host(logger=logger, zfs=zfs)\n\n filters = jails + (\"template=no\",)\n jails = iocage.lib.Jails.JailsGenerator(\n logger=logger,\n host=host,\n zfs=zfs,\n filters=filters\n )\n\n changed_jails = []\n failed_jails = []\n for jail in jails:\n try:\n changed = print_function(jail.updater.apply())\n if changed is True:\n changed_jails.append(jail)\n except iocage.lib.errors.UpdateFailure:\n failed_jails.append(jail)\n\n if len(failed_jails) > 0:\n return False\n\n if len(changed_jails) == 0:\n jails_input = \" \".join(list(jails))\n logger.error(\n f\"No jail was updated or matched your input: {jails_input}\"\n )\n return False\n\n return True\n", "sub_path": "iocage/cli/update.py", "file_name": "update.py", "file_ext": "py", "file_size_in_byte": 2681, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "iocage.lib.errors.lib.ZFS.ZFS", "line_number": 49, "usage_type": "call"}, {"api_name": "iocage.lib.errors.lib", "line_number": 49, "usage_type": "attribute"}, {"api_name": "iocage.lib.errors", "line_number": 49, "usage_type": "name"}, {"api_name": "iocage.lib.errors.lib.Host.Host", "line_number": 51, "usage_type": "call"}, {"api_name": "iocage.lib.errors.lib", "line_number": 51, "usage_type": "attribute"}, {"api_name": "iocage.lib.errors", "line_number": 51, "usage_type": "name"}, {"api_name": "iocage.lib.errors.lib.Jails.JailsGenerator", "line_number": 54, "usage_type": "call"}, {"api_name": "iocage.lib.errors.lib", "line_number": 54, "usage_type": "attribute"}, {"api_name": "iocage.lib.errors", "line_number": 54, "usage_type": "name"}, {"api_name": "iocage.lib.errors.lib", "line_number": 68, "usage_type": "attribute"}, {"api_name": "iocage.lib.errors", "line_number": 68, "usage_type": "name"}, {"api_name": "click.command", "line_number": 35, "usage_type": "call"}, {"api_name": "click.pass_context", "line_number": 36, "usage_type": "attribute"}, {"api_name": "click.argument", "line_number": 37, "usage_type": "call"}]} +{"seq_id": "551256703", "text": "#Consider only the below columns and prepare a prediction model for predicting Price.\n\n#Corolla<-Corolla[c(\"Price\",\"Age_08_04\",\"KM\",\"HP\",\"cc\",\"Doors\",\"Gears\",\"Quarterly_Tax\",\"Weight\")]\n#3 6 8 12 13 15 16 17\n\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\ndataset = pd.read_csv(\"ToyotaCorolla.csv\", encoding=\"latin1\", engine='python')\nx = dataset.iloc[:, [3, 6, 8, 12, 13, 15, 16, 17]].values\ny = dataset.iloc[:, 2].values\n\n#splitting dataset into train and test\nfrom sklearn.model_selection import train_test_split\nx_train, x_test, y_train, y_test = train_test_split(x, y, test_size = 0.2, random_state=0)\n\n#simple linear regression\nfrom sklearn.linear_model import LinearRegression\nregressor = LinearRegression()\nregressor.fit(x_train, y_train)\n\n#predict the test test\ny_pred = regressor.predict(x_test)\n\n#building an optimal model through backward elimination\nimport statsmodels.formula.api as sm\nx = np.append(arr = np.ones((1436,1)).astype(int), values = x, axis = 1)\n\nx_opt = x[:, [0,1,2,3,4,5,6,7]]\nregressor_ols = sm.OLS(endog = y, exog = x_opt).fit()\nregressor_ols.summary()\n# pvalue = 0.269 \n\nx_opt = x[:, [0,1,2,3,5,6,7]]\nregressor_ols = sm.OLS(endog = y, exog = x_opt).fit()\nregressor_ols.summary()", "sub_path": "Assignments/corolla.py", "file_name": "corolla.py", "file_ext": "py", "file_size_in_byte": 1234, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.linear_model.LinearRegression", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 28, "usage_type": "call"}, {"api_name": "statsmodels.formula.api.OLS", "line_number": 31, "usage_type": "call"}, {"api_name": "statsmodels.formula.api", "line_number": 31, "usage_type": "name"}, {"api_name": "statsmodels.formula.api.OLS", "line_number": 36, "usage_type": "call"}, {"api_name": "statsmodels.formula.api", "line_number": 36, "usage_type": "name"}]} +{"seq_id": "583329044", "text": "# built-in libraries\n# ...\n\n# external libraries\nimport scipy.linalg\n\n# internal libraries\nfrom ouroboros import Type, Image, Node\nfrom ouroboros.lib import libkin\n\n# exports\n__all__ = (\"kin\",\n \"point\")\n\n# constants\nGRAVITY_CONST = 6.67808e-11\n\nkin = Type(\".gee#kin\", libkin.kin,\n libkin.kin._asdict,\n lambda x: libkin.kin(**x))\n\n@Image(\".gee@point\",\n one=Node(evs=(), args=(\"m\",),\n ins=(), reqs=(),\n outs=(), pros=()),\n two=Node(evs=(), args=(\"m\",),\n ins=(), reqs=(),\n outs=(), pros=()),\n fun=Node(evs=(\"i\",), args=(),\n ins=(), reqs=(\"t\", \"y\"),\n outs=(\"o\",), pros=(\"y_dot\",)))\ndef point(env, clk, bod, orb):\n \"\"\"Point gravity\"\"\"\n m1, = next(one.data)\n m2, = next(two.data)\n mu = GRAVITY_CONST * m1 * m2 # m3/kg/s2\n\n yield\n while True:\n t, y = next(fun.data)\n \n (r, v) = y\n r_dot = v\n v_dot = - mu * r / scipy.linalg.norm(r) ** 3\n y_dot = libkin.kin(r_dot, v_dot)\n \n fun.data.send((y_dot,))\n yield (fun.ctrl.send((False,)),)\n", "sub_path": "ob-phys/ob-phys/gee.py", "file_name": "gee.py", "file_ext": "py", "file_size_in_byte": 1145, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "ouroboros.Type", "line_number": 18, "usage_type": "call"}, {"api_name": "ouroboros.lib.libkin.kin", "line_number": 18, "usage_type": "attribute"}, {"api_name": "ouroboros.lib.libkin", "line_number": 18, "usage_type": "name"}, {"api_name": "ouroboros.lib.libkin.kin", "line_number": 19, "usage_type": "attribute"}, {"api_name": "ouroboros.lib.libkin", "line_number": 19, "usage_type": "name"}, {"api_name": "ouroboros.lib.libkin.kin", "line_number": 20, "usage_type": "call"}, {"api_name": "ouroboros.lib.libkin", "line_number": 20, "usage_type": "name"}, {"api_name": "scipy.linalg.linalg.norm", "line_number": 44, "usage_type": "call"}, {"api_name": "scipy.linalg.linalg", "line_number": 44, "usage_type": "attribute"}, {"api_name": "scipy.linalg", "line_number": 44, "usage_type": "name"}, {"api_name": "ouroboros.lib.libkin.kin", "line_number": 45, "usage_type": "call"}, {"api_name": "ouroboros.lib.libkin", "line_number": 45, "usage_type": "name"}, {"api_name": "ouroboros.Image", "line_number": 22, "usage_type": "call"}, {"api_name": "ouroboros.Node", "line_number": 23, "usage_type": "call"}, {"api_name": "ouroboros.Node", "line_number": 26, "usage_type": "call"}, {"api_name": "ouroboros.Node", "line_number": 29, "usage_type": "call"}]} +{"seq_id": "440463614", "text": "#!/usr/bin/env python\n# encoding: utf-8\n\"\"\"\nbase.py\n一些首要的定义什么的\nCreated by qign on 2011-11-15.\nCopyright (c) 2011 __MyCompanyName__. All rights reserved.\n\"\"\"\n\nimport os\nimport sys\nimport time\nfrom datetime import datetime\n\njoin = os.path.join\nbase = os.path.dirname(os.path.abspath(os.path.realpath(__file__)))\n#join(base)\n\ndef format_time(seconds,format=\"%Y-%m-%d %H:%M:%S\"):\n dt = datetime.fromtimestamp(seconds)\n return dt.strftime(format)\n\ndef parse_time(format_time):\n ''' 2011-05-03 12:32:44 -> 1304397164\n '''\n if format_time.find('-') == -1:\n return int(format_time)\n if format_time.find(' ') == -1:\n format_time = format_time + ' 0:0:0'\n days,secs = format_time.split(' ')\n _year,_mon,_day = [int(d) for d in days.split('-')]\n _hour,_min,_sec = [int(s) for s in secs.split(':')]\n return time.mktime([_year,_mon,_day,_hour,_min,_sec,0,0,0])\n\n\ndef main():\n pass\n\n\nif __name__ == '__main__':\n main()\n\n", "sub_path": "base.py", "file_name": "base.py", "file_ext": "py", "file_size_in_byte": 977, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.realpath", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 20, "usage_type": "name"}, {"api_name": "time.mktime", "line_number": 33, "usage_type": "call"}]} +{"seq_id": "188274460", "text": "from datetime import datetime\nimport re\n\n\ndef is_between_dates(tar_filename: str, start_date: str, end_date: str,\n filename_pattern: str = \"mongo-dump-(.*).tar.gz\") -> bool:\n \"\"\"\n Check that current dump date is between start_date and end_date\n :param tar_filename: Tar filename which contains daily dump date\n :param start_date: Start date in ISO format\n :param end_date: End date in ISO format\n :param filename_pattern: Filename pattern, which contains date in ISO format\n :return: True if tar file date is between start date and end date, else - False\n \"\"\"\n iso_format = \"%Y-%m-%d\"\n start_date = datetime.strptime(start_date, iso_format)\n end_date = datetime.strptime(end_date, iso_format)\n tar_date = re.search(filename_pattern, tar_filename).group(1)\n tar_date = datetime.strptime(tar_date, iso_format)\n return start_date <= tar_date <= end_date\n", "sub_path": "landscape/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 17, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 17, "usage_type": "name"}, {"api_name": "re.search", "line_number": 18, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 19, "usage_type": "name"}]} +{"seq_id": "493584664", "text": "from abc import ABCMeta, abstractmethod\nfrom typing import Any, Union, Dict, TYPE_CHECKING # noqa: F401\n\nif TYPE_CHECKING:\n from base.grid import Grid\n from base.cell import Cell\nelse:\n Grid = 'Grid'\n Cell = 'Cell'\n\nDIRECTIONS = ['north', 'n', 'south', 's', 'east', 'e', 'west', 'w']\n\n\nclass Algorithm(metaclass=ABCMeta):\n ''' Base algorithm metaclass '''\n\n def __init__(self) -> None:\n self.step_count = 0 # type: int\n\n @abstractmethod\n def on(self, grid: Grid) -> None:\n ''' Run the algorithm '''\n raise NotImplementedError\n\n def step(self, value: int = 1) -> None:\n ''' Step the count of the algorithm '''\n self.step_count += value\n\n @property\n def name(self) -> str:\n ''' Name of the algorithm '''\n return self.__class__.__name__\n\n\nclass AlgorithmWithLogging(Algorithm, metaclass=ABCMeta):\n ''' Base algorithm metaclass with logging '''\n\n def __init__(self, log: bool = True) -> None:\n super().__init__()\n self.log = log\n\n def _prepareLogGrid(self, grid: Grid) -> None:\n ''' Prepare the grid for logging the algorithm '''\n if not self.log: return\n # 'visit' : List of steps on which the cell was visited\n # 'links' : Directions of the links made by the cell (NOT to the cell)\n data = {'visit': [], 'links': []} # type: Dict\n key = self.name # type: str\n for cell in grid.eachCell():\n cell.data[key] = data\n\n def _logVisit(self, cell: Cell) -> None:\n ''' Log the step number on which the cell was visited '''\n if not self.log: return\n cell.data[self.name]['visit'].append(self.step_count)\n\n def _logLink(self, cell: Cell, other: Cell) -> None:\n ''' Log the direction in which the cell made links on the step specified by the number '''\n if not self.log: return\n\n step_count = self.step_count\n\n if other == cell.north:\n cell.data[self.name]['links'].append((step_count, 'north'))\n elif other == cell.south:\n cell.data[self.name]['links'].append((step_count, 'south'))\n elif other == cell.east:\n cell.data[self.name]['links'].append((step_count, 'east'))\n elif other == cell.west:\n cell.data[self.name]['links'].append((step_count, 'west'))\n", "sub_path": "algorithms/algorithm.py", "file_name": "algorithm.py", "file_ext": "py", "file_size_in_byte": 2331, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 4, "usage_type": "name"}, {"api_name": "base.grid.Grid", "line_number": 8, "usage_type": "name"}, {"api_name": "base.cell.Cell", "line_number": 9, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 14, "usage_type": "name"}, {"api_name": "base.grid.Grid", "line_number": 21, "usage_type": "name"}, {"api_name": "abc.abstractmethod", "line_number": 20, "usage_type": "name"}, {"api_name": "abc.ABCMeta", "line_number": 35, "usage_type": "name"}, {"api_name": "base.grid.Grid", "line_number": 42, "usage_type": "name"}, {"api_name": "base.cell.Cell", "line_number": 52, "usage_type": "name"}, {"api_name": "base.cell.Cell", "line_number": 57, "usage_type": "name"}]} +{"seq_id": "650792723", "text": "#!/usr/bin/env python3\n\nimport requests\nfrom bs4 import BeautifulSoup\nfrom hashlib import md5\nfrom time import sleep\nfrom floodfire_crawler.core.base_list_crawler import BaseListCrawler\nfrom floodfire_crawler.storage.rdb_storage import FloodfireStorage\nimport time\n\nclass UdnListCrawler(BaseListCrawler):\n\n @property\n def url(self):\n return self._url\n\n @url.setter\n def url(self, value):\n self._url = value\n\n def __init__(self, config):\n self.floodfire_storage = FloodfireStorage(config)\n\n def fetch_html(self, url):\n headers = {\n 'User-Agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36',\n }\n response = requests.get(url, headers=headers, timeout=15)\n html = response.text\n return html\n\n def get_last(self):\n return None\n\t\n\t\n def fetch_list(self, soup):\n news = []\n news_rows = soup.find_all(\"dt\", {\"class\": \"lazyload\"})\n #md5hash = md5()\n for news_row in news_rows:\n link_a = news_row.find_all(\"a\")[0]\n md5hash = md5(link_a['href'].split(\"?\")[0].encode('utf-8')).hexdigest()\n raw = {\n 'title': news_row.find_all(\"a\")[-1].text.strip().replace(\" \",\" \").replace(\"\\u200b\",\"\"),\n 'url': \"https://udn.com\"+link_a['href'].split(\"?\")[0],\n 'url_md5': md5hash,\n 'source_id': 8,\n 'category': news_row.find_all(\"a\")[-2].text\n }\n news.append(raw)\n return news\n\n def make_a_round(self):\n #first page\n consecutive = 0\n page_url = self.url\n print(page_url)\n sleep(2)\n html = self.fetch_html(page_url)\n #time stamp for next pages\n stamp = round(time.time()*1000)\n \n soup = BeautifulSoup(html, 'html.parser')\n news_list = self.fetch_list(soup)\n #print(news_list)\n for news in news_list:\n if(self.floodfire_storage.check_list(news['url_md5']) == 0):\n self.floodfire_storage.insert_list(news)\n consecutive = 0\n else:\n print(news['title']+' exist! skip insert.')\n consecutive += 1\n\n total_pages = int(soup.find_all(\"div\",{\"class\":\"showmore\"})[0].a['data-totalpages'])\n\n #next page\n for page in range(2, total_pages+1):\n if consecutive > 20:\n print('News consecutive more than 20, stop crawler!!')\n break\n page_url = \"https://udn.com/news/get_breaks_article/\"+str(page)+\"/1/0?_=\"+str(stamp+page)\n print(page_url)\n sleep(2)\n html = self.fetch_html(page_url)\n soup = BeautifulSoup(html, 'html.parser')\n news_list = self.fetch_list(soup)\n #print(news_list)\n for news in news_list:\n if(self.floodfire_storage.check_list(news['url_md5']) == 0):\n self.floodfire_storage.insert_list(news)\n consecutive = 0\n else:\n print(news['title']+' exist! skip insert.')\n consecutive += 1\n page += 1\n\n\n def run(self):\n self.make_a_round()\n \"\"\"\n news_list = self.fetch_list(soup)\n print(news_list)\n for news in news_list:\n if(self.floodfire_storage.check_list(news['url_md5']) == 0):\n self.floodfire_storage.insert_list(news)\n else:\n print(news['title']+' exist! skip insert.')\n \n last_page = self.get_last(soup)\n print(last_page)\n \"\"\"\n \n", "sub_path": "floodfire_crawler/engine/udn_list_crawler.py", "file_name": "udn_list_crawler.py", "file_ext": "py", "file_size_in_byte": 3706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "floodfire_crawler.core.base_list_crawler.BaseListCrawler", "line_number": 11, "usage_type": "name"}, {"api_name": "floodfire_crawler.storage.rdb_storage.FloodfireStorage", "line_number": 22, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "hashlib.md5", "line_number": 42, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 61, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 63, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 83, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 85, "usage_type": "call"}]} +{"seq_id": "340374829", "text": "# Copyright 2020 VinyMeuh. All rights reserved.\n# Use of the source code is governed by a MIT-style license that can be found in the LICENSE file.\n\nfrom PySide2.QtWidgets import (\n QTableView,\n QWidget\n)\n\n\nclass ImagesListView(QTableView):\n\n def __init__(self, parent: QWidget = None):\n super().__init__(parent)\n", "sub_path": "egophoto/images_viewer/images_list_view.py", "file_name": "images_list_view.py", "file_ext": "py", "file_size_in_byte": 328, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "PySide2.QtWidgets.QTableView", "line_number": 10, "usage_type": "name"}, {"api_name": "PySide2.QtWidgets.QWidget", "line_number": 12, "usage_type": "name"}]} +{"seq_id": "446966544", "text": "# code by chenchiwei\n# -*- coding: UTF-8 -*-\nimport os\nimport numpy as np\nfrom sklearn import tree\nfrom sklearn.ensemble import RandomForestRegressor\nfrom sklearn.externals import joblib\nTMODEL_PATH = \"../Data/TModel\"\nN_EST = 1000\n\n# H 試験サンプル分類結果 测试样本分类结果\n# TrainS オリジナル(ソースドメイン)訓練用サンプル np配列 原训练样本 np数组\n# TrainA サポート(ターゲットドメイン)訓練サンプル 辅助训练样本\n# LabelS オリジナル(ソースドメイン)訓練用サンプルラベ��� 原训练样本标签\n# LabelA サポート(ターゲットドメイン)訓練用サンプルラベル 辅助训练样本标签\n# Test テスト用サンプル 测试样本\n# N 反復回数 迭代次数\n# set_name clf_n = {分類対象の属性}_{学習文書数(モデルの名付け用)}\ndef tradaboost(trans_S, trans_A, label_S, label_A, test, N, set_name, save_flag=False, model_name=\"DecisionTreeClassifier\"):\n test = np.array(test)\n print(\"Shape:{0} Dim:{1}\".format(test.shape, test.ndim))\n test = np.vstack(test)\n # print(\"Shape S: {0} Dim S : {1}\".format(trans_S.ndim, trans_S.shape[0]))\n # print(\"Shape T: {0} Dim T : {1}\".format(trans_A.ndim, trans_A.shape[0]))\n # print(\"Test T: {0} Dim T : {1}\".format(test.ndim, test.shape[0]))\n trans_data = np.concatenate((trans_A, trans_S), axis=0)\n trans_label = np.concatenate((label_A, label_S), axis=0)\n row_A = trans_A.shape[0] # [0, ..., 0(80), 1, ..., 1(80個)]なので160\n row_S = trans_S.shape[0]\n row_T = test.shape[0]\n\n test_data = np.concatenate((trans_data, test), axis=0) #消した l39も修正\n\n # 初期化の重み 初始化权重\n weights_A = np.ones([row_A, 1]) / row_A\n weights_S = np.ones([row_S, 1]) / row_S\n weights = np.concatenate((weights_A, weights_S), axis=0)\n\n bata = 1 / (1 + np.sqrt(2 * np.log(row_A / N)))\n\n # 各反復ごとにラベルとベータの値を保存するか? 存储每次迭代的标签和bata值?\n bata_T = np.zeros([1, N])\n result_label = np.ones([row_A + row_S + row_T, N])\n # result_label = np.ones([row_T, N])\n\n predict = np.zeros([row_T])\n\n print('params initial finished.')\n trans_data = np.asarray(trans_data, order='C')\n trans_label = np.asarray(trans_label, order='C')\n test_data = np.asarray(test_data, order='C')\n # test_data = np.asarray(test, order='C')\n\n for i in range(N):\n P = calculate_P(weights, trans_label)\n # print(\"trans_data\", trans_data.shape)\n # print(\"test_data\", test_data.shape)\n # print(\"trans_label\", trans_label.shape)\n # print(\"P\", P)\n \"\"\"\n Traceback (most recent call last):\n File \"transfer_learning.py\", line 61, in \n transfer_learning(n, age_dic, age_clf, age_qids, iterate)\n File \"transfer_learning.py\", line 33, in transfer_learning\n pred = tradaboost(wdata_train, tdata_train, wlabels_train, tlabels_train, data_test, iterate, \"{0}_{1}\".format(clf, n))\n File \"/home/kataoka/Research/Trans/TrAdaBoost.py\", line 53, in tradaboost\n test_data, P, model_name, set_name)\n ValueError: could not broadcast input array from shape (994) into shape (1014)\n \"\"\"\n result_label[:, i] = train_classify(trans_data, trans_label, test_data, P, model_name, set_name)\n # print('result,', result_label[:, i], row_A, row_S, i, result_label.shape)\n\n error_rate = calculate_error_rate(label_S, result_label[row_A:row_A + row_S, i],\n weights[row_A:row_A + row_S, :])\n print('Error rate:', error_rate)\n if error_rate > 0.5:\n error_rate = 0.5\n if error_rate == 0:\n N = i\n break # 防止过拟合\n # error_rate = 0.001\n\n bata_T[0, i] = error_rate / (1 - error_rate)\n\n # ソースドメインのサンプルの重みを調節 调整源域样本权重\n for j in range(row_S):\n weights[row_A + j] = weights[row_A + j] * np.power(bata_T[0, i],\n np.abs(result_label[row_A + j, i] - label_S[j]))\n\n # ターゲットドメインのサンプルの重みを調節 调整辅域样本权重\n for j in range(row_A):\n weights[j] = weights[j] * np.power(bata, (-np.abs(result_label[j, i] - label_A[j])))\n # print bata_T\n for i in range(row_T):\n # トレーニングデータのラベルをスキップ 跳过训练数据的标签\n left = np.sum(\n result_label[row_A + row_S + i, int(np.ceil(N / 2)):N] * np.log(1 / bata_T[0, int(np.ceil(N / 2)):N]))\n right = 0.5 * np.sum(np.log(1 / bata_T[0, int(np.ceil(N / 2)):N]))\n\n if left >= right:\n predict[i] = 1\n else:\n predict[i] = 0\n print(left, right, predict[i])\n return predict\n\n\ndef calculate_P(weights, label):\n total = np.sum(weights)\n return np.asarray(weights / total, order='C')\n\ndef train_classify(trans_data, trans_label, test_data, P, model_name, set_name, weighted=False):\n # clf = tree.DecisionTreeClassifier(criterion=\"gini\", max_features=\"log2\", splitter=\"random\")\n clf = RandomForestRegressor(criterion='mse', n_estimators=N_EST, n_jobs=2)\n if os.path.exists(TMODEL_PATH) == False: os.mkdir(TMODEL_PATH)\n clf.fit(trans_data, trans_label, sample_weight=P[:, 0])\n name = \"{0}/Trans_{1}_{2}.pkl\".format(TMODEL_PATH, set_name, model_name) if weighted else \"{0}/Trans_{1}_{2}_{3}.pkl\".format(TMODEL_PATH, set_name, model_name, \"weighted\")\n joblib.dump(clf, name)\n return clf.predict(test_data)\n\ndef calculate_error_rate(label_R, label_H, weight):\n total = np.sum(weight)\n\n print(weight[:, 0] / total)\n print(np.abs(label_R - label_H))\n return np.sum(weight[:, 0] / total * np.abs(label_R - label_H))\n", "sub_path": "Trans/TrAdaBoost.py", "file_name": "TrAdaBoost.py", "file_ext": "py", "file_size_in_byte": 5872, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.array", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 110, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestRegressor", "line_number": 114, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 115, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.dump", "line_number": 118, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 118, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 126, "usage_type": "call"}]} +{"seq_id": "513428972", "text": "from django.shortcuts import render\n\nimport json\n\n# rest frameworks\nfrom rest_framework import status\nfrom rest_framework.decorators import api_view\nfrom rest_framework.response import Response\n\n# upload file\nfrom django.core.files.storage import FileSystemStorage\n\n\n# Create your views here.\n\n\n@api_view(['GET', 'POST'])\ndef apiOrders(request):\n data = [{'id': 1, 'name': 'James', 'drink': 'Coffee'}, {\n 'id': 2, 'name': 'John', 'drink': 'Tea'}]\n\n if request.method == \"GET\":\n return Response(json.loads(json.dumps(data)))\n\n elif request.method == \"POST\":\n data.append(request.data)\n print(data)\n return Response(json.loads(json.dumps(data)))\n\n\ndef index(request):\n return render(request, 'index.html')\n\n\ndef simple_upload(request):\n if request.method == \"POST\" and request.FILES['myfile']:\n myfile = request.FILES['myfile']\n fs = FileSystemStorage()\n filename = fs.save(myfile.name, myfile)\n uploaded_file_url = fs.url(filename)\n return render(request, 'index.html', {\n 'uploaded_file_url': uploaded_file_url\n })\n return render(request, 'index.html')\n", "sub_path": "DoctorTable/doctortable/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1159, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "rest_framework.response.Response", "line_number": 23, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 23, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 23, "usage_type": "call"}, {"api_name": "rest_framework.response.Response", "line_number": 28, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 28, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 28, "usage_type": "call"}, {"api_name": "rest_framework.decorators.api_view", "line_number": 17, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.core.files.storage.FileSystemStorage", "line_number": 38, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 44, "usage_type": "call"}]} +{"seq_id": "474585701", "text": "# -*- coding: utf-8 -*-\n# @Time    : 2020/9/16 9:33\n# @Author  : xiaolu\n# @FileName: model.py\n# @Software: PyCharm\nfrom torch import nn\nfrom transformers import BertModel, BertConfig\nfrom torch.nn import CrossEntropyLoss\n\n\nclass Model(nn.Module):\n def __init__(self):\n super(Model, self).__init__()\n self.config = BertConfig.from_pretrained('./roberta_pretrain/bert_config.json')\n self.roberta = BertModel.from_pretrained('./roberta_pretrain/pytorch_model.bin', config=self.config)\n self.num_labels = 2\n self.output = nn.Linear(self.config.hidden_size, self.num_labels)\n\n def forward(self, input_ids=None, attention_mask=None, segment_ids=None, labels=None):\n # input_ids, input_mask, segment_ids, labels=labels_ids\n # last_output, all_layers_output = self.roberta(input_ids=input_ids, token_type_ids=segment_ids, attention_mask=attention_mask)\n final_layer, cls_output, layer_13_output = self.roberta(input_ids=input_ids, token_type_ids=segment_ids, attention_mask=attention_mask)\n # print(len(all_layers_output))\n # exit()\n logits = self.output(cls_output)\n if labels is not None:\n loss_fct = CrossEntropyLoss()\n loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1))\n return loss, logits\n return logits, layer_13_output\n\n\n\n\n\n\n\n\n", "sub_path": "Text_Ranking/ReRank/model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 1376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.nn.Module", "line_number": 11, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 11, "usage_type": "name"}, {"api_name": "transformers.BertConfig.from_pretrained", "line_number": 14, "usage_type": "call"}, {"api_name": "transformers.BertConfig", "line_number": 14, "usage_type": "name"}, {"api_name": "transformers.BertModel.from_pretrained", "line_number": 15, "usage_type": "call"}, {"api_name": "transformers.BertModel", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 17, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 17, "usage_type": "name"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "560204497", "text": "\"\"\"Python client for Nvim.\n\nClient library for talking with Nvim processes via it's msgpack-rpc API.\n\"\"\"\nimport logging\nimport os\n\nfrom .api import DecodeHook, Nvim, SessionHook\nfrom .msgpack_rpc import (socket_session, spawn_session, stdio_session,\n tcp_session)\nfrom .plugins import PluginHost, ScriptHost\n\n\n__all__ = ('tcp_session', 'socket_session', 'stdio_session', 'spawn_session',\n 'start_host', 'DecodeHook', 'Nvim', 'SessionHook')\n\n\ndef start_host(session=None):\n \"\"\"Promote the current process into python plugin host for Nvim.\n\n Start msgpack-rpc event loop for `session`, listening for Nvim requests\n and notifications. It registers Nvim commands for loading/unloading\n python plugins.\n\n The sys.stdout and sys.stderr streams are redirected to Nvim through\n `session`. That means print statements probably won't work as expected\n while this function doesn't return.\n\n This function is normally called at program startup and could have been\n defined as a separate executable. It is exposed as a library function for\n testing purposes only.\n \"\"\"\n logger = logging.getLogger(__name__)\n if 'NVIM_PYTHON_LOG_FILE' in os.environ:\n logfile = os.environ['NVIM_PYTHON_LOG_FILE'].strip()\n handler = logging.FileHandler(logfile, 'w')\n handler.formatter = logging.Formatter(\n '%(asctime)s [%(levelname)s @ '\n '%(filename)s:%(funcName)s:%(lineno)s] %(process)s - %(message)s')\n logging.root.addHandler(handler)\n level = logging.INFO\n if 'NVIM_PYTHON_LOG_LEVEL' in os.environ:\n l = getattr(logging,\n os.environ['NVIM_PYTHON_LOG_LEVEL'].strip(),\n level)\n if isinstance(l, int):\n level = l\n logger.setLevel(level)\n if not session:\n session = stdio_session()\n nvim = Nvim.from_session(session)\n with PluginHost(nvim, preloaded=[ScriptHost]) as host:\n host.run()\n\n\n# Required for python 2.6\nclass NullHandler(logging.Handler):\n def emit(self, record):\n pass\n\n\nif not logging.root.handlers:\n logging.root.addHandler(NullHandler())\n", "sub_path": "neovim/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2188, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 37, "usage_type": "call"}, {"api_name": "logging.root.addHandler", "line_number": 40, "usage_type": "call"}, {"api_name": "logging.root", "line_number": 40, "usage_type": "attribute"}, {"api_name": "logging.INFO", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 44, "usage_type": "attribute"}, {"api_name": "msgpack_rpc.stdio_session", "line_number": 50, "usage_type": "call"}, {"api_name": "api.Nvim.from_session", "line_number": 51, "usage_type": "call"}, {"api_name": "api.Nvim", "line_number": 51, "usage_type": "name"}, {"api_name": "plugins.PluginHost", "line_number": 52, "usage_type": "call"}, {"api_name": "plugins.ScriptHost", "line_number": 52, "usage_type": "name"}, {"api_name": "logging.Handler", "line_number": 57, "usage_type": "attribute"}, {"api_name": "logging.root", "line_number": 62, "usage_type": "attribute"}, {"api_name": "logging.root.addHandler", "line_number": 63, "usage_type": "call"}, {"api_name": "logging.root", "line_number": 63, "usage_type": "attribute"}]} +{"seq_id": "177340059", "text": "#!/usr/bin/env python\n\nimport os\nimport random\nimport sys\n\nfrom pyglet.gl import *\nimport pyglet\nfrom pyglet.window import key\n\nwindow = pyglet.window.Window(640, 480)\nBALL_IMAGE = 'star.png'\nballs_batch = pyglet.graphics.Batch()\nballs = []\nlabel = pyglet.text.Label('Press space to add a ball, backspace to remove',\n font_size=14,\n x=window.width // 2, y=10,\n anchor_x='center')\n\nclass Ball(pyglet.sprite.Sprite):\n ball_image = pyglet.resource.image(BALL_IMAGE)\n width = ball_image.width\n height = ball_image.height\n\n def __init__(self):\n x = window.width/2\n y = window.height/2\n\n super(Ball, self).__init__(self.ball_image, x, y, batch=balls_batch)\n\n self.dx = (random.random() - 0.5) * 1000\n self.dy = (random.random() - 0.5) * 1000\n\n def update(self, dt):\n if self.x <= 0 or self.x + self.width >= window.width:\n self.dx *= -1\n if self.y <= 0 or self.y + self.height >= window.height:\n self.dy *= -1\n self.x += self.dx * dt\n self.y += self.dy * dt\n\n self.x = min(max(self.x, 0), window.width - self.width)\n self.y = min(max(self.y, 0), window.height - self.height)\n\n\n@window.event\ndef on_key_press(symbol, modifiers):\n if symbol == key.SPACE:\n balls.append(Ball())\n elif symbol == key.BACKSPACE:\n if balls:\n del balls[-1]\n elif symbol == key.ESCAPE:\n window.has_exit = True\n\n@window.event\ndef on_draw():\n window.clear()\n balls_batch.draw()\n label.draw()\n\ndef update(dt):\n for ball in balls:\n ball.update(dt)\npyglet.clock.schedule_interval(update, 1.)\n\n\n\n\npyglet.app.run()\n", "sub_path": "py help/10. movement/bounce pyglet examples/bounce.py", "file_name": "bounce.py", "file_ext": "py", "file_size_in_byte": 1725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pyglet.window.Window", "line_number": 11, "usage_type": "call"}, {"api_name": "pyglet.window", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.Batch", "line_number": 13, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pyglet.text.Label", "line_number": 15, "usage_type": "call"}, {"api_name": "pyglet.text", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pyglet.sprite", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pyglet.resource.image", "line_number": 21, "usage_type": "call"}, {"api_name": "pyglet.resource", "line_number": 21, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 31, "usage_type": "call"}, {"api_name": "random.random", "line_number": 32, "usage_type": "call"}, {"api_name": "pyglet.window.key.SPACE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 48, "usage_type": "name"}, {"api_name": "pyglet.window.key.BACKSPACE", "line_number": 50, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 50, "usage_type": "name"}, {"api_name": "pyglet.window.key.ESCAPE", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pyglet.window.key", "line_number": 53, "usage_type": "name"}, {"api_name": "pyglet.clock.schedule_interval", "line_number": 65, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pyglet.app.run", "line_number": 70, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 70, "usage_type": "attribute"}]} +{"seq_id": "637377540", "text": "import json\n\nfrom django.core.urlresolvers import reverse\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom horizon import exceptions\nfrom horizon import forms\nfrom horizon import messages\nfrom openstack_dashboard.api import sds_controller as api\nfrom openstack_dashboard.dashboards.sdscontroller import exceptions as sdsexception\n\n\nclass UploadMetricModule(forms.SelfHandlingForm):\n metric_module_file = forms.FileField(label=_(\"File\"),\n required=True,\n allow_empty_file=False)\n\n class_name = forms.CharField(max_length=255,\n label=_(\"Class Name\"),\n help_text=_(\"The main class of the metric module to be created.\"),\n widget=forms.TextInput(\n attrs={\"ng-model\": \"name\", \"not-blank\": \"\"}\n ))\n\n out_flow = forms.BooleanField(required=False)\n\n in_flow = forms.BooleanField(required=False)\n\n execution_server = forms.ChoiceField(\n label=_('Execution Server'),\n choices=[\n ('proxy', _('Proxy Server')),\n ('object', _('Object Storage Servers'))\n ],\n widget=forms.Select(attrs={\n 'class': 'switchable',\n 'data-slug': 'source'\n })\n )\n\n enabled = forms.BooleanField(label=_(\"Enable Workload Metric\"),\n required=False)\n\n def __init__(self, request, *args, **kwargs):\n super(UploadMetricModule, self).__init__(request, *args, **kwargs)\n\n @staticmethod\n def handle(request, data):\n metric_module_file = data['metric_module_file']\n del data['metric_module_file']\n\n try:\n response = api.mtr_add_metric_module_metadata(request, data, metric_module_file)\n if 200 <= response.status_code < 300:\n messages.success(request, _('Successfully metric module creation and upload.'))\n return data\n else:\n raise sdsexception.SdsException(response.text)\n except Exception as ex:\n redirect = reverse(\"horizon:sdscontroller:administration:index\")\n error_message = \"Unable to create metric module.\\t %s\" % ex.message\n exceptions.handle(request, _(error_message), redirect=redirect)\n\n\nclass UpdateMetricModule(forms.SelfHandlingForm):\n class_name = forms.CharField(max_length=255,\n label=_(\"Class Name\"),\n help_text=_(\"The main class of the metric module to be created.\"))\n\n out_flow = forms.BooleanField(required=False)\n in_flow = forms.BooleanField(required=False)\n\n execution_server = forms.ChoiceField(\n label=_('Execution Server'),\n choices=[\n ('proxy', _('Proxy Server')),\n ('object', _('Object Storage Servers'))\n ],\n widget=forms.Select(attrs={\n 'class': 'switchable',\n 'data-slug': 'source'\n })\n )\n\n enabled = forms.BooleanField(label=_(\"Enable Workload Metric\"),\n required=False)\n\n def __init__(self, request, *args, **kwargs):\n super(UpdateMetricModule, self).__init__(request, *args, **kwargs)\n\n failure_url = 'horizon:sdscontroller:administration:index'\n\n def handle(self, request, data):\n try:\n metric_module_id = self.initial['id']\n # print \"\\n#################\\n\", request, \"\\n#################\\n\", data, \"\\n#################\\n\"\n response = api.mtr_update_metric_module(request, metric_module_id, data)\n if 200 <= response.status_code < 300:\n messages.success(request, _('Successfully metric module updated.'))\n return data\n else:\n raise sdsexception.SdsException(response.text)\n except Exception as ex:\n redirect = reverse(\"horizon:sdscontroller:administration:index\")\n error_message = \"Unable to update metric module.\\t %s\" % ex.message\n exceptions.handle(request, _(error_message), redirect=redirect)\n", "sub_path": "openstack_dashboard/dashboards/sdscontroller/administration/metric_modules/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 4189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "horizon.forms.SelfHandlingForm", "line_number": 13, "usage_type": "attribute"}, {"api_name": "horizon.forms", "line_number": 13, "usage_type": "name"}, {"api_name": "horizon.forms.FileField", "line_number": 14, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 14, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 14, "usage_type": "call"}, {"api_name": "horizon.forms.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 18, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 19, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 20, "usage_type": "call"}, {"api_name": "horizon.forms.TextInput", "line_number": 21, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 21, "usage_type": "name"}, {"api_name": "horizon.forms.BooleanField", "line_number": 25, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 25, "usage_type": "name"}, {"api_name": "horizon.forms.BooleanField", "line_number": 27, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 27, "usage_type": "name"}, {"api_name": "horizon.forms.ChoiceField", "line_number": 29, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 29, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 30, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 32, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 33, "usage_type": "call"}, {"api_name": "horizon.forms.Select", "line_number": 35, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 35, "usage_type": "name"}, {"api_name": "horizon.forms.BooleanField", "line_number": 41, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 41, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 41, "usage_type": "call"}, {"api_name": "openstack_dashboard.api.sds_controller.mtr_add_metric_module_metadata", "line_number": 53, "usage_type": "call"}, {"api_name": "openstack_dashboard.api.sds_controller", "line_number": 53, "usage_type": "name"}, {"api_name": "horizon.messages.success", "line_number": 55, "usage_type": "call"}, {"api_name": "horizon.messages", "line_number": 55, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 55, "usage_type": "call"}, {"api_name": "openstack_dashboard.dashboards.sdscontroller.exceptions.SdsException", "line_number": 58, "usage_type": "call"}, {"api_name": "openstack_dashboard.dashboards.sdscontroller.exceptions", "line_number": 58, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 60, "usage_type": "call"}, {"api_name": "horizon.exceptions.handle", "line_number": 62, "usage_type": "call"}, {"api_name": "horizon.exceptions", "line_number": 62, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 62, "usage_type": "call"}, {"api_name": "horizon.forms.SelfHandlingForm", "line_number": 65, "usage_type": "attribute"}, {"api_name": "horizon.forms", "line_number": 65, "usage_type": "name"}, {"api_name": "horizon.forms.CharField", "line_number": 66, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 66, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 67, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 68, "usage_type": "call"}, {"api_name": "horizon.forms.BooleanField", "line_number": 70, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 70, "usage_type": "name"}, {"api_name": "horizon.forms.BooleanField", "line_number": 71, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 71, "usage_type": "name"}, {"api_name": "horizon.forms.ChoiceField", "line_number": 73, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 73, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 74, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 76, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 77, "usage_type": "call"}, {"api_name": "horizon.forms.Select", "line_number": 79, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 79, "usage_type": "name"}, {"api_name": "horizon.forms.BooleanField", "line_number": 85, "usage_type": "call"}, {"api_name": "horizon.forms", "line_number": 85, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 85, "usage_type": "call"}, {"api_name": "openstack_dashboard.api.sds_controller.mtr_update_metric_module", "line_number": 97, "usage_type": "call"}, {"api_name": "openstack_dashboard.api.sds_controller", "line_number": 97, "usage_type": "name"}, {"api_name": "horizon.messages.success", "line_number": 99, "usage_type": "call"}, {"api_name": "horizon.messages", "line_number": 99, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 99, "usage_type": "call"}, {"api_name": "openstack_dashboard.dashboards.sdscontroller.exceptions.SdsException", "line_number": 102, "usage_type": "call"}, {"api_name": "openstack_dashboard.dashboards.sdscontroller.exceptions", "line_number": 102, "usage_type": "name"}, {"api_name": "django.core.urlresolvers.reverse", "line_number": 104, "usage_type": "call"}, {"api_name": "horizon.exceptions.handle", "line_number": 106, "usage_type": "call"}, {"api_name": "horizon.exceptions", "line_number": 106, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 106, "usage_type": "call"}]} +{"seq_id": "59776699", "text": "\"\"\"This package contains modules related to objective functions, optimizations, and network architectures.\n\nTo add a custom model class called 'dummy', you need to add a file called 'dummy_model.py' and define a subclass DummyModel inherited from BaseModel.\nYou need to implement the following five functions:\n -- <__init__>: initialize the class; first call BaseModel.__init__(self, opt).\n -- : unpack data from dataset and apply preprocessing.\n -- : produce intermediate results.\n -- : calculate loss, gradients, and update network weights.\n -- : (optionally) add model-specific options and set default options.\n\nIn the function <__init__>, you need to define four lists:\n -- self.loss_names (str list): specify the training losses that you want to plot and save.\n -- self.model_names (str list): define networks used in our training.\n -- self.visual_names (str list): specify the images that you want to display and save.\n -- self.optimizers (optimizer list): define and initialize optimizers. You can define one optimizer for each network. If two networks are updated at the same time, you can use itertools.chain to group them. See cycle_gan_model.py for an usage.\n\nNow you can use the model class by specifying flag '--model dummy'.\nSee our template model class 'template_model.py' for more details.\n\"\"\"\n\nimport importlib\nfrom models.base_model import BaseModel\nfrom torch.optim import lr_scheduler\nfrom torch.optim import optimizer\nimport torch\n\ndef find_model_using_name(model_name):\n \"\"\"Import the module \"models/[model_name]_model.py\".\n\n In the file, the class called DatasetNameModel() will\n be instantiated. It has to be a subclass of BaseModel,\n and it is case-insensitive.\n \"\"\"\n model_filename = \"models.\" + model_name + \"_model\"\n modellib = importlib.import_module(model_filename)\n model = None\n target_model_name = model_name.replace('_', '') + 'model'\n for name, cls in modellib.__dict__.items():\n if name.lower() == target_model_name.lower() \\\n and issubclass(cls, BaseModel):\n model = cls\n\n if model is None:\n print(\"In %s.py, there should be a subclass of BaseModel with class name that matches %s in lowercase.\" % (model_filename, target_model_name))\n exit(0)\n\n return model\n\n\ndef get_option_setter(model_name):\n \"\"\"Return the static method of the model class.\"\"\"\n model_class = find_model_using_name(model_name)\n return model_class.modify_commandline_options\n\n\ndef create_model(opt, specific_model=None):\n \"\"\"Create a model given the option.\n\n This function warps the class CustomDatasetDataLoader.\n This is the main interface between this package and 'train.py'/'test.py'\n\n Example:\n >>> from models import create_model\n >>> model = create_model(opt)\n \"\"\" \n if specific_model is None:\n specific_model = opt.model\n model = find_model_using_name(specific_model)\n instance = model(opt)\n print(\"model [%s] was created\" % type(instance).__name__)\n return instance\n\ndef get_scheduler(optimizer, opt):\n \"\"\"Return a learning rate scheduler\n\n Parameters:\n optimizer -- the optimizer of the network\n opt (option class) -- stores all the experiment flags; needs to be a subclass of BaseOptions. \n opt.lr_policy is the name of learning rate policy: linear | step | plateau | cosine\n\n For 'linear', we keep the same learning rate for the first epochs\n and linearly decay the rate to zero over the next epochs.\n For other schedulers (step, plateau, and cosine), we use the default PyTorch schedulers.\n See https://pytorch.org/docs/stable/optim.html for more details.\n \"\"\"\n if opt.lr_policy == 'linear':\n def lambda_rule(epoch):\n lr_l = 1.0 - max(0, epoch + opt.epoch_count - opt.niter) / float(opt.niter_decay + 1)\n return lr_l\n scheduler = lr_scheduler.LambdaLR(optimizer, lr_lambda=lambda_rule)\n elif opt.lr_policy == 'step':\n scheduler = lr_scheduler.StepLR(optimizer, step_size=opt.lr_decay_iters, gamma=0.1)\n elif opt.lr_policy == 'plateau':\n scheduler = lr_scheduler.ReduceLROnPlateau(optimizer, mode='min', factor=opt.lr_plateau_factor, threshold=0.01, patience=opt.lr_plateau_patience, verbose=True, min_lr=1e-7)\n elif opt.lr_policy == 'cosine':\n scheduler = lr_scheduler.CosineAnnealingLR(optimizer, T_max=opt.niter, eta_min=0)\n elif opt.lr_policy == 'multi':\n milestones = list(map(int, opt.lr_scheduler_milestones.split(',')))\n scheduler = lr_scheduler.MultiStepLR(optimizer, milestones)\n else:\n return NotImplementedError('learning rate policy [%s] is not implemented', opt.lr_policy)\n return scheduler\n\ndef get_optimizer(opt, params):\n if opt.optimizer == 'SGD':\n return torch.optim.SGD(params, opt.lr, opt.momentum, weight_decay=opt.weight_decay)\n elif opt.optimizer == 'rmsprop':\n return torch.optim.RMSprop(params, opt.lr, weight_decay=opt.weight_decay)\n elif opt.optimizer == 'adam':\n return torch.optim.Adam(params, opt.lr, weight_decay=opt.weight_decay)\n raise ValueError('Chosen optimizer is not supported, please choose from (SGD | adam | rmsprop)')", "sub_path": "models/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 5488, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "importlib.import_module", "line_number": 35, "usage_type": "call"}, {"api_name": "models.base_model.BaseModel", "line_number": 40, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler.LambdaLR", "line_number": 90, "usage_type": "call"}, {"api_name": "torch.optim.optimizer", "line_number": 90, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 90, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.StepLR", "line_number": 92, "usage_type": "call"}, {"api_name": "torch.optim.optimizer", "line_number": 92, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 92, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.ReduceLROnPlateau", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.optim.optimizer", "line_number": 94, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.CosineAnnealingLR", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.optim.optimizer", "line_number": 96, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 96, "usage_type": "name"}, {"api_name": "torch.optim.lr_scheduler.MultiStepLR", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.optim.optimizer", "line_number": 99, "usage_type": "argument"}, {"api_name": "torch.optim.lr_scheduler", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.optim.SGD", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 106, "usage_type": "attribute"}, {"api_name": "torch.optim.RMSprop", "line_number": 108, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 108, "usage_type": "attribute"}, {"api_name": "torch.optim.Adam", "line_number": 110, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 110, "usage_type": "attribute"}]} +{"seq_id": "586236158", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport json_field.fields\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('dashboard', '0003_auto_20160811_0319'),\n ]\n\n operations = [\n migrations.RemoveField(\n model_name='metric',\n name='condition',\n ),\n migrations.RemoveField(\n model_name='metric',\n name='design',\n ),\n migrations.RemoveField(\n model_name='metric',\n name='minimum',\n ),\n migrations.RemoveField(\n model_name='metric',\n name='stretch',\n ),\n migrations.RemoveField(\n model_name='metric',\n name='units',\n ),\n migrations.RemoveField(\n model_name='metric',\n name='user',\n ),\n migrations.AddField(\n model_name='job',\n name='blobs',\n field=json_field.fields.JSONField(help_text='Data blobs produced by the job.', default=None, null=True, blank=True),\n ),\n migrations.AddField(\n model_name='measurement',\n name='metadata',\n field=json_field.fields.JSONField(help_text='Measurement metadata', default=None, null=True, blank=True),\n ),\n migrations.AddField(\n model_name='metric',\n name='operator',\n field=models.CharField(help_text='Operator used to test measurementvalue against metric specification', max_length=2, default='<'),\n ),\n migrations.AddField(\n model_name='metric',\n name='parameters',\n field=json_field.fields.JSONField(help_text='Parameters used to define the metric', default=None, null=True, blank=True),\n ),\n migrations.AddField(\n model_name='metric',\n name='reference',\n field=json_field.fields.JSONField(help_text='Metric reference', default=None, null=True, blank=True),\n ),\n migrations.AddField(\n model_name='metric',\n name='specs',\n field=json_field.fields.JSONField(help_text='Array of metric specification', default=None, null=True, blank=True),\n ),\n migrations.AddField(\n model_name='metric',\n name='unit',\n field=models.CharField(help_text='Metric unit, astropy compatible string', max_length=16, default='', null=True, blank=True),\n ),\n migrations.AlterField(\n model_name='job',\n name='ci_name',\n field=models.CharField(help_text='Name of the Jenkins project,e.g. validate_drp', max_length=32),\n ),\n migrations.AlterField(\n model_name='measurement',\n name='value',\n field=models.FloatField(help_text='Metric scalar measurement'),\n ),\n migrations.AlterField(\n model_name='metric',\n name='description',\n field=models.TextField(help_text='Metric description'),\n ),\n migrations.AlterField(\n model_name='metric',\n name='metric',\n field=models.CharField(help_text='Metric name', max_length=16, primary_key=True, serialize=False),\n ),\n ]\n", "sub_path": "squash/dashboard/migrations/0004_auto_20170216_0009.py", "file_name": "0004_auto_20170216_0009.py", "file_ext": "py", "file_size_in_byte": 3289, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 31, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 35, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 35, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 39, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 39, "usage_type": "name"}, {"api_name": "json_field.fields.fields.JSONField", "line_number": 42, "usage_type": "call"}, {"api_name": "json_field.fields.fields", "line_number": 42, "usage_type": "attribute"}, {"api_name": "json_field.fields", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 44, "usage_type": "name"}, {"api_name": "json_field.fields.fields.JSONField", "line_number": 47, "usage_type": "call"}, {"api_name": "json_field.fields.fields", "line_number": 47, "usage_type": "attribute"}, {"api_name": "json_field.fields", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 54, "usage_type": "name"}, {"api_name": "json_field.fields.fields.JSONField", "line_number": 57, "usage_type": "call"}, {"api_name": "json_field.fields.fields", "line_number": 57, "usage_type": "attribute"}, {"api_name": "json_field.fields", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 59, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 59, "usage_type": "name"}, {"api_name": "json_field.fields.fields.JSONField", "line_number": 62, "usage_type": "call"}, {"api_name": "json_field.fields.fields", "line_number": 62, "usage_type": "attribute"}, {"api_name": "json_field.fields", "line_number": 62, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 64, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 64, "usage_type": "name"}, {"api_name": "json_field.fields.fields.JSONField", "line_number": 67, "usage_type": "call"}, {"api_name": "json_field.fields.fields", "line_number": 67, "usage_type": "attribute"}, {"api_name": "json_field.fields", "line_number": 67, "usage_type": "name"}, {"api_name": "django.db.migrations.AddField", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 74, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 74, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 79, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 79, "usage_type": "name"}, {"api_name": "django.db.models.FloatField", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 87, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 87, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 89, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 89, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 92, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 92, "usage_type": "name"}]} +{"seq_id": "203716711", "text": "from numba import njit\nimport numpy as np\n\nd = int(input())\ncs = list(map(int, input().split()))\ncs = np.array(cs, dtype=np.int64)\nsm = [list(map(int, input().split())) for _ in range(d)]\nsm = np.array(sm, dtype=np.int64)\n\n@njit('i8(i8[:], i8)', cache=True)\ndef total_satisfaction(ts, d):\n ls = np.zeros(26, dtype=np.int64)\n s = 0\n for i in range(d):\n t = ts[i]\n t -= 1\n s += sm[i][t]\n ls[t] = i + 1\n\n dv = cs * ((i+1) - ls)\n s -= dv.sum()\n return s\n\n@njit('i8(i8, i8, i8, i8, i8[:])', cache=True)\ndef differential(s, i, t, d, ls):\n t -= 1\n bk = ls[t]\n s += sm[i][t]\n ls[t] = i + 1\n\n dv = cs * ((i+1) - ls)\n s -= dv.sum()\n\n ls[t] = bk\n return s\n\nts = np.array([], dtype=np.int64)\nfor i in range(d):\n sc = 0\n if len(ts) > 0:\n tt = np.array(ts, dtype=np.int64)\n sc = total_satisfaction(tt, i)\n\n ls = np.zeros(26, np.int64)\n for i, t in enumerate(ts, 1):\n ls[t-1] = i\n\n mx = -99999999\n mxt = -1\n for t in range(1, 26+1):\n df = differential(sc, len(ts), t, i + 1, ls)\n s = sc + df\n if s > mx:\n mx = s\n mxt = t\n ts = np.append(ts, [mxt])\n print(mxt)\n", "sub_path": "Python_codes/p02618/s245709820.py", "file_name": "s245709820.py", "file_ext": "py", "file_size_in_byte": 1214, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.array", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 6, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 8, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numba.njit", "line_number": 10, "usage_type": "call"}, {"api_name": "numba.njit", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 44, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 56, "usage_type": "call"}]} +{"seq_id": "264072814", "text": "import cv2\nimport json\nimport colorsys\nimport random\n\nfont = cv2.FONT_HERSHEY_DUPLEX\n\n\ndef get_n_hls_colors(num):\n hls_colors = []\n i = 0\n step = 360.0 / num\n while i < 360:\n h = i\n s = 90 + random.random() * 10\n l = 50 + random.random() * 10\n _hlsc = [h / 360.0, l / 100.0, s / 100.0]\n hls_colors.append(_hlsc)\n i += step\n return hls_colors\n\n\ndef ncolors(num):\n \"\"\"\n 生成N种颜色\n :param num:\n :return:\n \"\"\"\n rgb_colors = []\n if num < 1:\n return rgb_colors\n hls_colors = get_n_hls_colors(num)\n for hlsc in hls_colors:\n _r, _g, _b = colorsys.hls_to_rgb(hlsc[0], hlsc[1], hlsc[2])\n r, g, b = [int(x * 255.0) for x in (_r, _g, _b)]\n rgb_colors.append([r, g, b])\n return rgb_colors\n\n\ndef bbox(points, border=30):\n x_arr = []\n y_arr = []\n for p in points:\n coord = p['coord']\n x = coord[0]\n y = coord[1]\n x_arr.append(x)\n y_arr.append(y)\n x_min = min(x_arr)\n x_max = max(x_arr)\n y_min = min(y_arr)\n y_max = max(y_arr)\n # (left, top), (right, bottom)\n border = int(1 * (y_max - y_min) / 10)\n # print((x_min - border, y_min - border), (x_max + border, y_max + border))\n return (x_min - border, y_min - border), (x_max + border, y_max + border)\n\n\ndef id_map(mapfile):\n ids = {}\n with open(mapfile, 'r') as f:\n for line in f.readlines():\n sp = line.split('\\t')\n uid = sp[0]\n fn = sp[1]\n ids[uid] = fn\n return ids\n\n\ndef plot(imagepath, json_path, map_path):\n ids = id_map(map_path)\n save_path = imagepath.replace('/label', '/plot')\n rgb_colors = ncolors(11)\n print(save_path)\n with open(json_path, 'r') as js:\n data = json.load(js)\n for d in data:\n data = d['data'][0]\n uid = data['instanceUid']\n annotation = data['annotation'][0]\n points = annotation['data']['point']\n fn = ids.get(uid)\n img = cv2.imread(imagepath + fn)\n if len(points) != 11:\n print(fn)\n print('points len =', len(points), len(points) == 11)\n # 每张片子11个节点\n b1, b2 = bbox(points)\n cv2.rectangle(img, b1, b2, (0, 0, 255), 1)\n for i in range(len(points)):\n p = points[i]\n c = rgb_colors[i]\n coord = p['coord']\n identification = p['tag']['identification']\n # print(p)\n cv2.circle(img, (coord[0], coord[1]), 1, c, 2)\n cv2.putText(img, identification, (coord[0] + 3, coord[1] + 3), font, 0.3, c, 1)\n cv2.imwrite(save_path + fn, img)\n\n\nif __name__ == '__main__':\n # id_map('debug_infos/id_mapping.txt')\n ## 特别注意 !!!\n # 181.jpg --> points len = 10 False\n plot('/home/utopia/CVDATA/lumbar/0613/lumbar_train51/label/',\n '/home/utopia/CVDATA/lumbar/0613/lumbar_train51_annotation.json',\n 'debug_infos/id_mapping.txt')\n plot('/home/utopia/CVDATA/lumbar/0613/lumbar_train150/label/',\n '/home/utopia/CVDATA/lumbar/0613/lumbar_train150_annotation.json',\n 'debug_infos/id_mapping.txt')\n", "sub_path": "pretreatment/label_utils.py", "file_name": "label_utils.py", "file_ext": "py", "file_size_in_byte": 3242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "cv2.FONT_HERSHEY_DUPLEX", "line_number": 6, "usage_type": "attribute"}, {"api_name": "random.random", "line_number": 15, "usage_type": "call"}, {"api_name": "random.random", "line_number": 16, "usage_type": "call"}, {"api_name": "colorsys.hls_to_rgb", "line_number": 34, "usage_type": "call"}, {"api_name": "json.load", "line_number": 76, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 83, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 89, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.putText", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 98, "usage_type": "call"}]} +{"seq_id": "113533147", "text": "from django.urls import path\n\nfrom PumpManager import views\n\n\nurlpatterns = [\n path('', views.index, name='index'),\n #path('profile//', views.uniqueProfile, name='uniqueProfile'),\n\n path('statistic/', views.statistic, name='statistic'),\n path('random/', views.random, name='random'),\n path('lists/', views.lists, name='lists'),\n path('songs/', views.songs, name='songs'),\n path('songs/', views.songID, name='songID'),\n\n path('mixes/', views.mixes, name='mixes'),\n path('mixes/', views.mixID, name='mixID'),\n\n #path('register/', views.RegisterFormView.as_view(), name='register'),\n path('register/', views.register, name='register'),\n path('login/', views.LoginFormView.as_view(), name='login'),\n path('logout/', views.LogoutView.as_view(), name='logout'),\n\n path('profile/', views.profile, name='profile'),\n path('profile/settings', views.profile_settings, name = 'profile_settings'),\n path('profile/add', views.profile_add, name='profile_add'),\n\n path('locations/', views.locations, name = 'locations') ,\n]\n", "sub_path": "PumpManager/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1086, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.urls.path", "line_number": 7, "usage_type": "call"}, {"api_name": "PumpManager.views.index", "line_number": 7, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 7, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 10, "usage_type": "call"}, {"api_name": "PumpManager.views.statistic", "line_number": 10, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 11, "usage_type": "call"}, {"api_name": "PumpManager.views.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 11, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 12, "usage_type": "call"}, {"api_name": "PumpManager.views.lists", "line_number": 12, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 12, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "PumpManager.views.songs", "line_number": 13, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 13, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 14, "usage_type": "call"}, {"api_name": "PumpManager.views.songID", "line_number": 14, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 14, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 16, "usage_type": "call"}, {"api_name": "PumpManager.views.mixes", "line_number": 16, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 16, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 17, "usage_type": "call"}, {"api_name": "PumpManager.views.mixID", "line_number": 17, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 17, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "PumpManager.views.register", "line_number": 20, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 20, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "PumpManager.views.LoginFormView.as_view", "line_number": 21, "usage_type": "call"}, {"api_name": "PumpManager.views.LoginFormView", "line_number": 21, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 21, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "PumpManager.views.LogoutView.as_view", "line_number": 22, "usage_type": "call"}, {"api_name": "PumpManager.views.LogoutView", "line_number": 22, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 22, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 24, "usage_type": "call"}, {"api_name": "PumpManager.views.profile", "line_number": 24, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 24, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 25, "usage_type": "call"}, {"api_name": "PumpManager.views.profile_settings", "line_number": 25, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 25, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 26, "usage_type": "call"}, {"api_name": "PumpManager.views.profile_add", "line_number": 26, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 26, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 28, "usage_type": "call"}, {"api_name": "PumpManager.views.locations", "line_number": 28, "usage_type": "attribute"}, {"api_name": "PumpManager.views", "line_number": 28, "usage_type": "name"}]} +{"seq_id": "152568940", "text": "\"MIME envelope handling\"\n\nimport base64\nfrom io import BytesIO\nimport mimetypes\nfrom email import encoders, charset\nfrom email.mime.multipart import MIMEMultipart\nfrom email.mime.text import MIMEText\nfrom email.mime.image import MIMEImage\nfrom email.mime.base import MIMEBase\nfrom email.mime.audio import MIMEAudio\nfrom email import message_from_string\nfrom email import message_from_bytes\nfrom random import Random\nimport string\nimport gzip as gzip_\n\nimport logging\nlog = logging.getLogger(__name__)\n\n# override the default handling of utf-8 (which is base64)\n# as X-road protocol 4.0 insists using 8bit encoding for SOAP envelope parts\ncharset.add_charset('utf-8', charset.SHORTEST, '8bit')\n\nclass Attachment:\n \"MIME attachment in X-road SOAP message\"\n filename = None\n content_id = None\n data = None\n use_gzip = True\n\n def __init__(self, data=None, filename=None, content_id=None, use_gzip=True):\n self.data = data\n self.filename = filename\n self.content_id = content_id\n self.use_gzip = use_gzip\n\n def gen_content_id(self):\n \"Generate identificator for an attachment\"\n self.content_id = ''.join(Random().sample(string.ascii_letters+string.digits, 32))\n return self.content_id\n\n def gzip(self):\n \"Encode\"\n return gzip(self.data)\n \ndef encode_soap(xml, attachments, embedded_newlines=True, mtom=False):\n \"\"\"Compose MIME message which contains SOAP envelope and attachments\n \"\"\"\n if attachments:\n if mtom:\n # MTOM/XOP\n payload = encode_mtom(xml, attachments)\n else:\n # SOAP with attachments\n payload = encode(xml, attachments)\n payload = payload.replace('\\n','\\r\\n')\n\n # extract HTTP header to be able\n # to use high level Http().request and add Content-Length\n headers, body = _extract_body(payload, embedded_newlines)\n body = body.encode('utf-8')\n headers['Content-Length'] = str(len(body))\n headers['SOAPAction'] = '\"\"'\n else:\n # plain message without attachments, no need for MIME\n body = xml.encode('utf-8')\n headers={\n 'Content-Type': 'text/xml; charset=\"UTF-8\"',\n 'Content-Length': str(len(body)),\n 'SOAPAction': '\"\"'\n }\n\n payload = ''\n for key in headers:\n payload += '%s: %s\\r\\n' % (key, headers[key])\n payload = payload.encode('utf-8') + b'\\r\\n' + body\n return payload, headers, body\n\ndef encode(body, attachments):\n \"\"\"\n Compose MIME message for SOAP with attachments\n \"\"\"\n msg = MIMEMultipart('related', type=\"text/xml\")\n envelope = MIMEText(body, 'xml', _charset='utf-8')\n envelope.replace_header('Content-Transfer-Encoding', '8bit')\n msg.attach(envelope)\n for attachment in attachments:\n _add_part(msg, attachment)\n return msg.as_string()\n\ndef encode_mtom(body, attachments):\n \"\"\"\n Compose MIME message for MTOM/XOP\n \"\"\"\n msg = MIMEMultipart('related', type=\"application/xop+xml\", start='', start_info='text/xml')\n envelope = MIMEBase('application', 'xop+xml', charset='utf-8', type='text/xml')\n envelope.set_payload(body)\n envelope['Content-Transfer-Encoding'] = '8bit'\n envelope['Content-ID'] = ''\n msg.attach(envelope)\n for attachment in attachments:\n _add_part(msg, attachment)\n return msg.as_string()\n\ndef _add_part(msg, attachment): \n \"Add a new attachment part to the MIME message\"\n\n ctype, encoding = mimetypes.guess_type(attachment.filename or '')\n if ctype is None or encoding is not None:\n ctype = 'application/octet-stream'\n\n maintype, subtype = ctype.split('/', 1)\n if maintype == 'text':\n part = MIMEText(attachment.data, _subtype=subtype)\n\n elif maintype == 'image':\n part = MIMEImage(attachment.data, _subtype=subtype)\n\n elif maintype == 'audio':\n part = MIMEAudio(attachment.data, _subtype=subtype)\n\n else:\n part = MIMEBase(maintype, subtype)\n data = attachment.data\n if isinstance(data, str):\n data = data.encode('utf-8')\n if attachment.use_gzip:\n part.set_payload(gzip(data))\n part.add_header('Content-Encoding', 'gzip')\n else: \n part.set_payload(data)\n encoders.encode_base64(part)\n\n part.set_charset('utf-8')\n\n if attachment.filename:\n part.add_header('Content-Disposition', 'attachment', filename=attachment.filename)\n if attachment.content_id:\n part.add_header('Content-ID', '<%s>' % attachment.content_id) \n\n msg.attach(part)\n\ndef decode(response):\n \"\"\"Parse message (MIME or plain)\n \"\"\"\n env = None\n attachments = []\n if isinstance(response, str):\n msg = message_from_string(response)\n else:\n msg = message_from_bytes(response)\n\n if msg.is_multipart():\n for part in msg._payload:\n fn = part.get_filename()\n content_id = part.get('Content-ID')\n data = part.get_payload()\n bdata = None\n ctype = part.get('Content-Type')\n if part.get('Content-Transfer-Encoding') == 'base64':\n bdata = base64.b64decode(data.encode('utf-8'))\n\n if bdata and part.get('Content-Encoding') == 'gzip':\n bdata = gunzip(bdata)\n\n if bdata:\n data = bdata\n\n if not env:\n # assume that first part is SOAP envelope\n env = data\n else:\n # others are attachments\n attachment = Attachment(data, filename=fn, content_id=content_id)\n attachments.append(attachment)\n else:\n # plain message\n env = msg.get_payload(decode=True)\n if not env and response:\n # probably XML declaration is missing\n env = response\n else:\n # remove crap (HTTPS)\n if isinstance(env, bytes):\n env = env.decode('utf-8') \n n1 = env.find('')\n if n1 > -1 and n2 > -1:\n env = env[n1:n2+1]\n \n return env, attachments\n\ndef gzip(data, compresslevel=9):\n \"\"\"\n Compresses the byte string :var:`data` with gzip using the compression level\n :var:`compresslevel`.\n \"\"\"\n stream = BytesIO()\n compressor = gzip_.GzipFile(filename=\"\", mode=\"wb\", fileobj=stream, compresslevel=compresslevel)\n compressor.write(data)\n compressor.close()\n return stream.getvalue() \n \ndef gunzip(data):\n \"\"\"\n Uncompresses the byte string :var:`data` with gzip.\n \"\"\"\n stream = BytesIO(data)\n compressor = gzip_.GzipFile(filename=\"\", mode=\"rb\", fileobj=stream)\n return compressor.read()\n\ndef _extract_body(payload, embedded_newlines):\n \"\"\"\n Extract HTTP headers and body\n \"\"\"\n headers_str, body = payload.split('\\r\\n\\r\\n',1)\n headers = {}\n for line in headers_str.splitlines():\n line = line.rstrip()\n if line.find(':') > -1:\n key, value = line.split(':',1)\n headers[key] = value\n else:\n if embedded_newlines:\n # SOAP input\n headers[key] += '\\r\\n' + line\n else:\n # SOAP output\n headers[key] += ' ' + line\n return headers, body\n", "sub_path": "pyxadapterlib/pyxadapterlib/attachment.py", "file_name": "attachment.py", "file_ext": "py", "file_size_in_byte": 7404, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "email.charset.add_charset", "line_number": 23, "usage_type": "call"}, {"api_name": "email.charset", "line_number": 23, "usage_type": "name"}, {"api_name": "email.charset.SHORTEST", "line_number": 23, "usage_type": "attribute"}, {"api_name": "random.Random", "line_number": 40, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 40, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 40, "usage_type": "attribute"}, {"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 84, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 85, "usage_type": "call"}, {"api_name": "email.mime.multipart.MIMEMultipart", "line_number": 96, "usage_type": "call"}, {"api_name": "email.mime.base.MIMEBase", "line_number": 97, "usage_type": "call"}, {"api_name": "mimetypes.guess_type", "line_number": 109, "usage_type": "call"}, {"api_name": "email.mime.text.MIMEText", "line_number": 115, "usage_type": "call"}, {"api_name": "email.mime.image.MIMEImage", "line_number": 118, "usage_type": "call"}, {"api_name": "email.mime.audio.MIMEAudio", "line_number": 121, "usage_type": "call"}, {"api_name": "email.mime.base.MIMEBase", "line_number": 124, "usage_type": "call"}, {"api_name": "email.encoders.encode_base64", "line_number": 133, "usage_type": "call"}, {"api_name": "email.encoders", "line_number": 133, "usage_type": "name"}, {"api_name": "email.message_from_string", "line_number": 150, "usage_type": "call"}, {"api_name": "email.message_from_bytes", "line_number": 152, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 162, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 199, "usage_type": "call"}, {"api_name": "gzip.GzipFile", "line_number": 200, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 209, "usage_type": "call"}, {"api_name": "gzip.GzipFile", "line_number": 210, "usage_type": "call"}]} +{"seq_id": "98885353", "text": "from pyramid.view import view_config\nfrom pyramid.httpexceptions import HTTPBadRequest, HTTPFound\n\nfrom sqlalchemy.orm import aliased\nfrom sqlalchemy.orm.exc import NoResultFound\nfrom sqlalchemy import and_, or_, null, select\nfrom sqlalchemy.sql import func\nfrom sqlalchemy.sql.expression import literal_column\n\nfrom geoalchemy2.shape import to_shape\n\n\nfrom thinkhazard_common.models import (\n DBSession,\n AdministrativeDivision,\n HazardLevel,\n HazardCategory,\n HazardCategoryAdministrativeDivisionAssociation,\n HazardType,\n ClimateChangeRecommendation,\n TechnicalRecommendation,\n FurtherResource,\n)\n\n\n# An object for the \"no data\" category type.\n_hazardlevel_nodata = HazardLevel()\n_hazardlevel_nodata.mnemonic = 'no-data'\n_hazardlevel_nodata.title = 'No data available'\n_hazardlevel_nodata.description = 'No data for this hazard type.'\n_hazardlevel_nodata.order = float('inf')\n\n\n@view_config(route_name='report_overview', renderer='templates/report.jinja2')\n@view_config(route_name='report_overview_slash',\n renderer='templates/report.jinja2')\n@view_config(route_name='report', renderer='templates/report.jinja2')\ndef report(request):\n try:\n division_code = request.matchdict.get('divisioncode')\n except:\n raise HTTPBadRequest(detail='incorrect value for parameter '\n '\"divisioncode\"')\n\n hazard = request.matchdict.get('hazardtype', None)\n\n # Get all the hazard types.\n hazardtype_query = DBSession.query(HazardType).order_by(HazardType.order)\n\n # Get the hazard categories corresponding to the administrative\n # division whose code is division_code.\n hazardcategories_query = DBSession.query(HazardCategory) \\\n .join(HazardCategory.administrativedivisions) \\\n .join(HazardType) \\\n .join(HazardLevel) \\\n .filter(AdministrativeDivision.code == division_code)\n\n # Create a dict with the categories. Keys are the hazard type mnemonic.\n hazardcategories = {d.hazardtype.mnemonic: d\n for d in hazardcategories_query}\n\n hazard_types = []\n for hazardtype in hazardtype_query:\n cat = _hazardlevel_nodata\n if hazardtype.mnemonic in hazardcategories:\n cat = hazardcategories[hazardtype.mnemonic].hazardlevel\n hazard_types.append({\n 'hazardtype': hazardtype,\n 'hazardlevel': cat\n })\n\n association = None\n technical_recommendations = None\n further_resources = None\n climate_change_recommendation = None\n\n # Get the administrative division whose code is division_code.\n _alias = aliased(AdministrativeDivision)\n division = DBSession.query(AdministrativeDivision) \\\n .outerjoin(_alias, _alias.code == AdministrativeDivision.parent_code) \\\n .filter(AdministrativeDivision.code == division_code).one()\n\n if hazard is not None:\n try:\n association = DBSession.query(\n HazardCategoryAdministrativeDivisionAssociation) \\\n .join(AdministrativeDivision) \\\n .join(HazardCategory) \\\n .join(HazardLevel) \\\n .join(HazardType) \\\n .filter(HazardType.mnemonic == hazard) \\\n .filter(AdministrativeDivision.code == division_code) \\\n .one()\n except NoResultFound:\n url = request.route_url('report_overview',\n divisioncode=division_code)\n return HTTPFound(location=url)\n\n try:\n # get the code for level 0 division\n code = division.code\n if division.leveltype_id == 2:\n code = division.parent.code\n if division.leveltype_id == 3:\n code = division.parent.parent.code\n climate_change_recommendation = DBSession.query(\n ClimateChangeRecommendation) \\\n .join(AdministrativeDivision) \\\n .join(HazardType) \\\n .filter(AdministrativeDivision.code == code) \\\n .filter(HazardType.mnemonic == hazard) \\\n .one()\n except NoResultFound:\n pass\n\n technical_recommendations = DBSession.query(TechnicalRecommendation) \\\n .join(TechnicalRecommendation.hazardcategory_associations) \\\n .join(HazardCategory) \\\n .filter(HazardCategory.id == association.hazardcategory.id) \\\n .all()\n\n further_resources = DBSession.query(FurtherResource) \\\n .join(FurtherResource.hazardcategory_associations) \\\n .join(HazardCategory) \\\n .outerjoin(AdministrativeDivision) \\\n .filter(HazardCategory.id == association.hazardcategory.id) \\\n .filter(or_(AdministrativeDivision.code == division_code,\n AdministrativeDivision.code == null())) \\\n .all()\n\n # Get the geometry for division and compute its extent\n cte = select([AdministrativeDivision.geom]) \\\n .where(AdministrativeDivision.code == division_code) \\\n .cte('bounds')\n bounds = list(DBSession.query(\n func.ST_XMIN(cte.c.geom),\n func.ST_YMIN(cte.c.geom),\n func.ST_XMAX(cte.c.geom),\n func.ST_YMAX(cte.c.geom))\n .one())\n division_bounds = bounds\n\n # compute a 0-360 version of the extent\n cte = select([\n func.ST_Shift_Longitude(AdministrativeDivision.geom).label('shift')]) \\\n .where(AdministrativeDivision.code == division_code) \\\n .cte('bounds')\n bounds_shifted = list(DBSession.query(\n func.ST_XMIN(cte.c.shift),\n func.ST_YMIN(cte.c.shift),\n func.ST_XMAX(cte.c.shift),\n func.ST_YMAX(cte.c.shift))\n .one())\n\n # Use the 0-360 if it's smaller\n if bounds_shifted[2] - bounds_shifted[0] < bounds[2] - bounds[0]:\n division_bounds = bounds_shifted\n\n parents = []\n if division.leveltype_id >= 2:\n parents.append(division.parent)\n if division.leveltype_id == 3:\n parents.append(division.parent.parent)\n\n return {'hazards': hazard_types,\n 'hazards_sorted': sorted(hazard_types,\n key=lambda a: a['hazardlevel'].order),\n 'hazard_category': association.hazardcategory\n if association else '',\n 'source': association.source\n if association else '',\n 'climate_change_recommendation': climate_change_recommendation,\n 'recommendations': technical_recommendations,\n 'resources': further_resources,\n 'division': division,\n 'bounds': division_bounds,\n 'parents': parents,\n 'parent_division': division.parent}\n\n\n@view_config(route_name='report_json', renderer='geojson')\n@view_config(route_name='report_overview_json', renderer='geojson')\ndef report_json(request):\n\n try:\n division_code = request.matchdict.get('divisioncode')\n except:\n raise HTTPBadRequest(detail='incorrect value for parameter '\n '\"divisioncode\"')\n\n try:\n resolution = float(request.params.get('resolution'))\n except:\n raise HTTPBadRequest(detail='invalid value for parameter \"resolution\"')\n\n hazard_type = request.matchdict.get('hazardtype', None)\n\n _filter = or_(AdministrativeDivision.code == division_code,\n AdministrativeDivision.parent_code == division_code)\n\n simplify = func.ST_Simplify(\n func.ST_Transform(AdministrativeDivision.geom, 3857), resolution / 2)\n\n if hazard_type is not None:\n divisions = DBSession.query(AdministrativeDivision) \\\n .add_columns(simplify, HazardLevel.mnemonic, HazardLevel.title) \\\n .outerjoin(AdministrativeDivision.hazardcategories) \\\n .join(HazardCategory) \\\n .outerjoin(HazardType)\\\n .outerjoin(HazardLevel) \\\n .filter(and_(_filter, HazardType.mnemonic == hazard_type))\n else:\n divisions = DBSession.query(AdministrativeDivision) \\\n .add_columns(simplify, literal_column(\"'None'\"),\n literal_column(\"'blah'\")) \\\n .filter(_filter)\n\n return [{\n 'type': 'Feature',\n 'geometry': to_shape(geom_simplified),\n 'properties': {\n 'name': division.name,\n 'code': division.code,\n 'hazardLevelMnemonic': hazardlevel_mnemonic,\n 'hazardLevelTitle': hazardlevel_title\n }\n } for division, geom_simplified, hazardlevel_mnemonic,\n hazardlevel_title in divisions]\n", "sub_path": "thinkhazard/views/report.py", "file_name": "report.py", "file_ext": "py", "file_size_in_byte": 8603, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "thinkhazard_common.models.HazardLevel", "line_number": 27, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 42, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 48, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 48, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 48, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardType.order", "line_number": 48, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardLevel", "line_number": 55, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 54, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 52, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 52, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 52, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardCategory.administrativedivisions", "line_number": 53, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 53, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 56, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 56, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.aliased", "line_number": 78, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 78, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 79, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 79, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 79, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.parent_code", "line_number": 80, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 80, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 81, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 81, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 90, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardLevel", "line_number": 89, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 88, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 87, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 85, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.HazardCategoryAdministrativeDivisionAssociation", "line_number": 86, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 85, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardType.mnemonic", "line_number": 91, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 91, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 92, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 92, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 94, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPFound", "line_number": 97, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 109, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 108, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 106, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.ClimateChangeRecommendation", "line_number": 107, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 106, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 110, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 110, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardType.mnemonic", "line_number": 111, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 111, "usage_type": "name"}, {"api_name": "sqlalchemy.orm.exc.NoResultFound", "line_number": 113, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 118, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 116, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.TechnicalRecommendation", "line_number": 116, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 116, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.TechnicalRecommendation.hazardcategory_associations", "line_number": 117, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.TechnicalRecommendation", "line_number": 117, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardCategory.id", "line_number": 119, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 119, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 125, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 124, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 122, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.FurtherResource", "line_number": 122, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 122, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.FurtherResource.hazardcategory_associations", "line_number": 123, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.FurtherResource", "line_number": 123, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardCategory.id", "line_number": 126, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 126, "usage_type": "name"}, {"api_name": "sqlalchemy.or_", "line_number": 127, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 127, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 127, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 128, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 128, "usage_type": "name"}, {"api_name": "sqlalchemy.null", "line_number": 128, "usage_type": "call"}, {"api_name": "sqlalchemy.select", "line_number": 132, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.geom", "line_number": 132, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 132, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 133, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 133, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 135, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 135, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_XMIN", "line_number": 136, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 136, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_YMIN", "line_number": 137, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 137, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_XMAX", "line_number": 138, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 138, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_YMAX", "line_number": 139, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 139, "usage_type": "name"}, {"api_name": "sqlalchemy.select", "line_number": 144, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func.ST_Shift_Longitude", "line_number": 145, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 145, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.geom", "line_number": 145, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 145, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 146, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 146, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 148, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 148, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_XMIN", "line_number": 149, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 149, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_YMIN", "line_number": 150, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 150, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_XMAX", "line_number": 151, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 151, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_YMAX", "line_number": 152, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 152, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 34, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 35, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 37, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 188, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPBadRequest", "line_number": 194, "usage_type": "call"}, {"api_name": "sqlalchemy.or_", "line_number": 198, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.code", "line_number": 198, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 198, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.parent_code", "line_number": 199, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 199, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_Simplify", "line_number": 201, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 201, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.func.ST_Transform", "line_number": 202, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.func", "line_number": 202, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.geom", "line_number": 202, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 202, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardLevel", "line_number": 210, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 209, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.HazardCategory", "line_number": 208, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 205, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 205, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 205, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardLevel.mnemonic", "line_number": 206, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardLevel", "line_number": 206, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.HazardLevel.title", "line_number": 206, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision.hazardcategories", "line_number": 207, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 207, "usage_type": "name"}, {"api_name": "sqlalchemy.and_", "line_number": 211, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.HazardType.mnemonic", "line_number": 211, "usage_type": "attribute"}, {"api_name": "thinkhazard_common.models.HazardType", "line_number": 211, "usage_type": "name"}, {"api_name": "thinkhazard_common.models.DBSession.query", "line_number": 213, "usage_type": "call"}, {"api_name": "thinkhazard_common.models.AdministrativeDivision", "line_number": 213, "usage_type": "argument"}, {"api_name": "thinkhazard_common.models.DBSession", "line_number": 213, "usage_type": "name"}, {"api_name": "sqlalchemy.sql.expression.literal_column", "line_number": 214, "usage_type": "call"}, {"api_name": "sqlalchemy.sql.expression.literal_column", "line_number": 215, "usage_type": "call"}, {"api_name": "geoalchemy2.shape.to_shape", "line_number": 220, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 181, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 182, "usage_type": "call"}]} +{"seq_id": "545153157", "text": "import pygame\r\n\r\n\r\nclass Window:\r\n def __init__(self, icon, name : str):\r\n pygame.init()\r\n\r\n self.width = 640\r\n self.height = 512\r\n\r\n self.MAIN = pygame.display.set_mode((self.width, self.height))\r\n self.NAME = pygame.display.set_caption(name)\r\n self.ICON = pygame.display.set_icon(icon)\r\n\r\n self.clock = pygame.time.Clock()\r\n self.fps = 64\r\n\r\n def update(self):\r\n pygame.display.update()\r\n self.clock.tick(self.fps)\r\n\r\n def write(self, text : str, font : str, size : int, pos : tuple, color : tuple, bold = False, italic = False, isCenter = False, backgroundColor = None):\r\n visual = pygame.font.SysFont(font, size, bold, italic)\r\n surface = visual.render(text, True, color, backgroundColor)\r\n\r\n rectangle = surface.get_rect()\r\n if isCenter:\r\n rectangle.center = pos\r\n else:\r\n rectangle = pos\r\n\r\n self.MAIN.blit(surface, rectangle)\r\n\r\n def button(self, x : int, width : int, y : int, height : int, text : str, font : str, color : tuple, action = None):\r\n mouse = pygame.mouse.get_pos()\r\n click = pygame.mouse.get_pressed()\r\n\r\n self.write(text, font, height - 4, (x + 4, y + 2), color)\r\n\r\n if x < mouse[0] < x + width and y < mouse[1] < y + height:\r\n if click[0] == 1:\r\n if action != None:\r\n action()\r\n\r\n\r\nWhite = (255, 255, 255)\r\nBlack = (0, 0, 0)\r\n\r\nicon = pygame.image.load_extended(\"Icon.png\")\r\n\r\nwindow = Window(icon, \"Slime Adventures\")", "sub_path": "Graphism.py", "file_name": "Graphism.py", "file_ext": "py", "file_size_in_byte": 1561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pygame.init", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.display.set_icon", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 15, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 15, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 23, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pressed", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.image.load_extended", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 49, "usage_type": "attribute"}]} +{"seq_id": "84091165", "text": "import requests, pandas as pd, numpy as np\nfrom pandas import DataFrame\nfrom io import StringIO\nimport time, json\nfrom datetime import date\nfrom statsmodels.tsa.stattools import adfuller, acf, pacf\nfrom statsmodels.tsa.arima_model import ARIMA\nfrom statsmodels.tsa.arima_model import ARMA, ARIMAResults\nfrom statsmodels.tsa.seasonal import seasonal_decompose\nfrom sklearn.metrics import mean_squared_error\nimport matplotlib\nmatplotlib.use('agg')\nimport matplotlib.pylab as plt\nfrom matplotlib.pylab import rcParams\nimport os, sklearn\nfrom statsmodels.graphics.tsaplots import plot_acf, plot_pacf\nfrom matplotlib import pyplot\nimport sys\n\n\n# 检验统计稳定性\ndef test_stationarity(timeseries):\n # Determing rolling statistics\n rolmean = pd.rolling_mean(timeseries, window=6)\n rolstd = pd.rolling_std(timeseries, window=6)\n\n # Plot rolling statistics:\n fig = plt.figure(figsize=(12, 8))\n orig = plt.plot(timeseries, color='blue', label='Original')\n mean = plt.plot(rolmean, color='red', label='Rolling Mean')\n std = plt.plot(rolstd, color='black', label='Rolling Std')\n plt.legend(loc='best')\n plt.title('Rolling Mean & Standard Deviation')\n plt.show()\n\n # # Perform Dickey-Fuller test:\n # print('Results of Dickey-Fuller Test:')\n # dftest = adfuller(timeseries, autolag='AIC')\n # dfoutput = pd.Series(dftest[0:4], index=['Test Statistic', 'p-value', '#Lags Used', 'Number of Observations Used'])\n # for key, value in dftest[4].items():\n # dfoutput['Critical Value (%s)' % key] = value\n # print(dfoutput)\n\n\ndef acf_pacf_plot(ts_log_diff):\n plot_acf(ts_log_diff, lags=40) # ARIMA,q\n plot_pacf(ts_log_diff, lags=40) # ARIMA,p\n pyplot.show()\n\n\n# 注意这里面使用的ts_log_diff是经过合适阶数的差分之后的数据,\n# 上文中提到ARIMA该开源库,不支持3阶以上的#差分。所以我们需要提前将数据差分好再传入\n# 求解最佳模型参数p,q\ndef _proper_model(ts_log_diff, maxLag):\n best_p = 0\n best_q = 0\n best_bic = sys.maxsize\n best_model = None\n for p in np.arange(maxLag):\n for q in np.arange(maxLag):\n try:\n model = ARMA(ts_log_diff, order=(p, q))\n results_ARMA = model.fit(disp=-1)\n except:\n continue\n bic = results_ARMA.bic\n # print(bic, best_bic)\n if bic < best_bic:\n best_p = p\n best_q = q\n best_bic = bic\n best_model = results_ARMA\n return best_p, best_q, best_model\n\n\nprint(os.getcwd())\n\nSin = np.load('../data/data_5features_narry/SMSIn_9999*8784.npy') # 9999 * timecount\nSout = np.load('../data/data_5features_narry/SMSOut_9999*8784.npy')\nCin = np.load('../data/data_5features_narry/CallIn_9999*8784.npy')\nCout = np.load('../data/data_5features_narry/CallOut_9999*8784.npy')\nItra = np.load('../data/data_5features_narry/InterTra_9999*8784.npy')\nprint(Sin.shape)\n\ndata_test = np.load('../data_test/test_total.npy') # (144,100,100,5)\ndata_test = data_test.reshape(144, 10000, 5)\nprint(data_test.shape)\n\ntimeseries = [Sin, Sout, Cin, Cout, Itra]\ntimeseriesname = ['Sin', 'Sout', 'Cin', 'Cout', 'Itra']\n\n\ndef pre_data(timeseries_fea, feaID, GridID):\n \"\"\"\n prepare timeSeries and timeSeeries_diff\n \"\"\"\n ts = pd.Series(timeseries_fea[GridID]) # [8784]\n ts_diff = ts - ts.shift()\n ts_diff.dropna(inplace=True)\n X = ts_diff.values\n ts_diff = X.astype('float64') # numpy array\n X = ts.values\n ts = X.astype('float64') # numpy array\n\n y_true = data_test[:, GridID, feaID] # [144]\n\n return ts, ts_diff, y_true\n\n\n# 验证一阶差分稳定性\n# # 一阶差分\n# ts_log_diff = ts_log - ts_log.shift()\n# ts_log_diff.dropna(inplace=True)\n# X = ts_log_diff.values\n# ts_log_diff = X.astype('float32')\n\n\n# test_stationarity(ts)\n# test_stationarity(ts_log)\n# test_stationarity(ts_log_diff)\n# test_stationarity(ts_diff)\n\n# acf_pacf_plot(ts_log_diff)\n# acf_pacf_plot(ts_log_diff)\n\n# monkey patch around bug in ARIMA class for save model\ndef __getnewargs__(self):\n return ((self.endog), (self.k_lags, self.k_diff, self.k_ma))\n\n\n# train model and save model\ndef train_save_arima(ts, save_path, best_p, best_q):\n a = 0\n try:\n model = ARIMA(ts, order=(best_p, 1, best_q)) # 一阶差分\n results_ARIMA = model.fit(disp=-1)\n # Save Models\n ARIMA.__getnewargs__ = __getnewargs__\n # save model\n results_ARIMA.save(save_path)\n except:\n model = ARIMA(ts, order=(best_p, 1, best_q)) # 一阶差分\n results_ARIMA = model.fit(transparams=False, disp=-1)\n # Save Models\n ARIMA.__getnewargs__ = __getnewargs__\n # save model\n results_ARIMA.save(save_path)\n a = a+1\n print('pass')\n\n return results_ARIMA,a\n\n\n# predict and eval model\ndef pred_eval_model(results_ARIMA, ts, forecast_n, y_true, fig_save_path):\n # forecast方法会自动进行差分还原,当然仅限于支持的1阶和2阶差分\n # forecast_n = 144 # 预测未来12个月走势\n forecast_ARIMA_log = results_ARIMA.forecast(forecast_n)\n forecast_ARIMA_log = forecast_ARIMA_log[0]\n # print(forecast_ARIMA_log[:144])\n\n # ValueError: Input contains NaN, infinity or a value too large for dtype('float64').\n forecast_ARIMA_log = pd.Series(forecast_ARIMA_log,\n index=np.arange(len(ts[-2016:]) + 1, len(ts[-2016:]) + len(forecast_ARIMA_log) + 1,\n 1))\n forecast_ARIMA_log.fillna(forecast_ARIMA_log.mean(), inplace=True)\n\n MSE = sklearn.metrics.mean_squared_error(y_true, forecast_ARIMA_log)\n # diff = y_true - forecast_ARIMA_log # [144]\n y_true_mean = np.mean(y_true)\n acc = MSE / y_true_mean\n\n y_true = pd.Series(y_true, index=np.arange(2016 + 1, 2016 + 144 + 1, 1))\n\n plt.plot(ts[-2016:], color=\"blue\", label='Original')\n plt.plot(y_true, color=\"navy\", label='y_true')\n plt.plot(forecast_ARIMA_log, color='red', label='Predicted')\n plt.legend(loc='best')\n plt.title('ARIMA MSE: %.4f ACC: %.4f' % (MSE, acc))\n plt.xlim([0, 2016+144])\n # show the biggest figure\n # manager = plt.get_current_fig_manager()\n # manager.window.showMaximized()\n fig = plt.gcf()\n #plt.show()\n fig.savefig(fig_save_path, bbox_inches='tight',dpi=100)\n plt.close(fig)\n\n return MSE, acc\n\n\n\n# predict model\nfor feaID in range(1,len(timeseriesname)):\n filename = open('./log_test_%s.log'%timeseriesname[feaID],'w')\n MSE_arr = []\n acc_arr = []\n for gridID in range(0,len(Sin),250):\n # Prepare train data and test data\n ts, ts_diff, y_true = pre_data(timeseries[feaID], feaID, gridID)\n\n # optimize the best model parameters\n best_p, best_q, model_ama = _proper_model(ts_diff, 10) # 对一阶差分求最优p和q\n print(best_p, best_q)\n filename.write('Grid %d predict stat Info:\\n'%gridID)\n filename.write('Best_p:%s, Best_q:%s\\n'%(str(best_p),str(best_q)))\n\n # train and save model\n save_path = './train_model_arima_save/train_model_save_%s/ARIMA_model_%s_grid%d.pkl' % (\n timeseriesname[feaID], timeseriesname[feaID], gridID)\n results_ARIMA, a = train_save_arima(ts, save_path, best_p, best_q)\n if a == 1:\n filename.write('###########################################parameters not converge very weill. \\n')\n\n fig_save_path = './predict_figure_save/predict_figure_save_%s/ARIMA_fig_%s_grid%d.png' % (timeseriesname[feaID], timeseriesname[feaID], gridID)\n MSE, acc = pred_eval_model(results_ARIMA,ts=ts, forecast_n=144,y_true=y_true,fig_save_path=fig_save_path)\n filename.write('MSE: %s, acc: %s\\n'%(str(MSE),str(acc)))\n MSE_arr.append(MSE)\n acc_arr.append(acc)\n\n MSE_mean = np.mean(MSE_arr)\n acc_mean = np.mean(acc_arr)\n print(MSE_mean)\n filename.write('MSE_mean: %s, acc_mean: %s\\n'%(str(MSE_mean),str(acc_mean)))\n filename.close()\n MSE_ts = pd.Series(MSE_arr)\n acc_ts = pd.Series(acc_arr)\n plt.plot(MSE_ts, color=\"blue\", label='MSE_timeSeries')\n plt.plot(acc_ts, color='red', label='Acc_timeSeries')\n plt.legend(loc='best')\n plt.title('%s MSE_mean: %s, acc_mean: %s\\n'%(timeseriesname[feaID],str(MSE_mean),str(acc_mean)))\n mse_acc_save_path = './MSE_ACC_%s.png'%timeseriesname[feaID]\n plt.xlim([0, len(MSE_arr)])\n # show the biggest figure\n # manager = plt.get_current_fig_manager()\n # manager.window.showMaximized()\n fig = plt.gcf()\n #plt.show()\n fig.savefig(mse_acc_save_path, bbox_inches='tight',dpi=100)\n plt.close(fig)\n\n\n\n\n\n\n\n\n\n\n# # load model\n# results_ARIMA = ARIMAResults.load('ARIMA_model.pkl')\n\n\n\n\n\n# 将预测的结果与原始图像画在一张图片上。\n# #定义获取连续时间,start是起始时间,limit是连续的天数,level可以是day,month,year\n# import arrow\n# def get_date_range(start, limit, level='month',format='YYYY-MM-DD'):\n# start = arrow.get(start, format)\n# result=(list(map(lambda dt: dt.format(format) , arrow.Arrow.range(level, start, limit=limit))))\n# dateparse2 = lambda dates:pd.datetime.strptime(dates,'%Y-%m-%d')\n# return map(dateparse2, result)\n\n\n# # 预测从1961-01-01开始,也就是我们训练数据最后一个数据的后一个日期\n# new_index = get_date_range('1961-01-01', forecast_n)\n# forecast_ARIMA_log = pd.Series(forecast_ARIMA_log, copy=True, index=new_index)\n# print(forecast_ARIMA_log.head())\n# # 直接取指数,即可恢复至原数据\n# forecast_ARIMA = np.exp(forecast_ARIMA_log)\n# print(forecast_ARIMA)\n# plt.plot(ts,label='Original',color='blue')\n# plt.plot(forecast_ARIMA, label='Forcast',color='red')\n# plt.legend(loc='best')\n# plt.title('forecast')\n# plt.show()\n", "sub_path": "models/ARIMA/ARIMA2.py", "file_name": "ARIMA2.py", "file_ext": "py", "file_size_in_byte": 9792, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "matplotlib.use", "line_number": 12, "usage_type": "call"}, {"api_name": "pandas.rolling_mean", "line_number": 24, "usage_type": "call"}, {"api_name": "pandas.rolling_std", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pylab.figure", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pylab.legend", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pylab.title", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pylab.show", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 34, "usage_type": "name"}, {"api_name": "statsmodels.graphics.tsaplots.plot_acf", "line_number": 46, "usage_type": "call"}, {"api_name": "statsmodels.graphics.tsaplots.plot_pacf", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "sys.maxsize", "line_number": 57, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 60, "usage_type": "call"}, {"api_name": "statsmodels.tsa.arima_model.ARMA", "line_number": 62, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 85, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 97, "usage_type": "call"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA", "line_number": 135, "usage_type": "call"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA.__getnewargs__", "line_number": 138, "usage_type": "attribute"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA", "line_number": 138, "usage_type": "name"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA", "line_number": 142, "usage_type": "call"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA.__getnewargs__", "line_number": 145, "usage_type": "attribute"}, {"api_name": "statsmodels.tsa.arima_model.ARIMA", "line_number": 145, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 164, "usage_type": "call"}, {"api_name": "sklearn.metrics.mean_squared_error", "line_number": 168, "usage_type": "call"}, {"api_name": "sklearn.metrics", "line_number": 168, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 170, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 173, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 175, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 175, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 176, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 176, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 177, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 177, "usage_type": "name"}, {"api_name": "matplotlib.pylab.legend", "line_number": 178, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 178, "usage_type": "name"}, {"api_name": "matplotlib.pylab.title", "line_number": 179, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 179, "usage_type": "name"}, {"api_name": "matplotlib.pylab.xlim", "line_number": 180, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 180, "usage_type": "name"}, {"api_name": "matplotlib.pylab.gcf", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pylab.close", "line_number": 187, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 187, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 221, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 222, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 226, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 227, "usage_type": "call"}, {"api_name": "matplotlib.pylab.plot", "line_number": 228, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 228, "usage_type": "name"}, {"api_name": "matplotlib.pylab.plot", "line_number": 229, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 229, "usage_type": "name"}, {"api_name": "matplotlib.pylab.legend", "line_number": 230, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 230, "usage_type": "name"}, {"api_name": "matplotlib.pylab.title", "line_number": 231, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 231, "usage_type": "name"}, {"api_name": "matplotlib.pylab.xlim", "line_number": 233, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 233, "usage_type": "name"}, {"api_name": "matplotlib.pylab.gcf", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 237, "usage_type": "name"}, {"api_name": "matplotlib.pylab.close", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pylab", "line_number": 240, "usage_type": "name"}]} +{"seq_id": "313776991", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: ./_schwiki/schwiki/models.py\n# Compiled at: 2019-11-02 09:56:54\n# Size of source mod 2**32: 14215 bytes\nimport django\nfrom django.db import models\nfrom pytigon_lib.schdjangoext.fields import *\nfrom pytigon_lib.schdjangoext.models import *\nimport pytigon_lib.schdjangoext.fields as ext_models\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.contrib import admin\nimport os, os.path, sys\nfrom pytigon_lib.schhtml.htmltools import superstrip\nfrom django.template import RequestContext, Context, Template\nimport markdown2 as markdown\nfrom pytigon_lib.schdjangoext.django_ihtml import ihtml_to_html\nfrom pytigon_lib.schtools.wiki import wikify, wiki_from_str, make_href\nfrom pytigon_lib.schtools.tools import norm_indent\nfrom django.template.loader import select_template\nfrom datetime import datetime\nfrom collections import namedtuple\ntemplate_content = '\\n{# -*- coding: utf-8 -*- #}\\n{%% load exfiltry %%}\\n{%% load exsyntax %%}\\n%s\\n'\n\ndef _get_wiki_object(page, buf, name, paragraf):\n name0 = name.split('_')[0]\n conf = None\n x = PageObjectsConf.objects.filter(name=name0)\n if len(x) > 0:\n conf = x[0]\n d = page.get_json_data()\n if name in d:\n c = d[name]\n else:\n c = ''\n inline_content = norm_indent(buf)\n if conf.inline_wiki:\n inline_content = html_from_wiki(page, inline_content)\n context = {'param':c, 'inline_content':inline_content, 'object':conf, 'page':page, 'paragraf':paragraf, 'name':name}\n if conf.view_dict:\n exec(conf.view_dict)\n context = locals()['get_view_dict'](context)\n template_name1 = (conf.app + '/' + conf.name).lower() + '_wikiobj_view.html'\n template_name2 = 'schwiki/wikiobj_view.html'\n t = select_template([template_name1, template_name2])\n return t.render(context).replace('[{', '{{').replace('}]', '}}').replace('[%', '{%').replace('%]', '%}')\n else:\n return ''\n\n\ndef _get_markdown_object(buf):\n return markdown.markdown(('\\n'.join(buf)), extras=['tables', 'codehilite'])\n\n\ndef html_from_wiki(page, wiki_str):\n document = []\n paragraf = []\n buf = []\n in_wiki_object = False\n name = ''\n paragraf_prefix = None\n paragraf_suffix = None\n section_close_elements = []\n document_close_elements = []\n\n def write_papragraf():\n nonlocal buf\n nonlocal paragraf\n if in_wiki_object:\n x = _get_wiki_object(page, buf, name, [paragraf_prefix, paragraf_suffix])\n document.append(x)\n buf = []\n else:\n if buf:\n paragraf.append((buf, True))\n buf = []\n if paragraf:\n if paragraf_prefix:\n x = paragraf_prefix\n else:\n x = ''\n for pos in paragraf:\n if pos[1]:\n x += _get_markdown_object(pos[0])\n else:\n x += pos[0]\n\n if paragraf_suffix:\n x += paragraf_suffix\n document.append(x)\n paragraf = []\n\n def write_section():\n nonlocal section_close_elements\n if section_close_elements:\n document.append(''.join(list(reversed(section_close_elements))))\n section_close_elements = []\n\n def write_document():\n nonlocal document_close_elements\n if document_close_elements:\n document.append(''.join(list(reversed(document_close_elements))))\n document_close_elements = []\n\n lines = wiki_str.replace('\\r', '').split('\\n')\n for line in lines:\n if in_wiki_object:\n if line.startswith(' ') or line.startswith('\\t') or not line:\n buf.append(line)\n continue\n else:\n x = _get_wiki_object(page, buf, name, [paragraf_prefix, paragraf_suffix])\n if x.startswith('@@@'):\n if '|||' in x:\n y = x[3:].split('|||')\n paragraf_prefix = y[0]\n paragraf_suffix = y[1]\n else:\n paragraf_prefix = x[3:]\n paragraf_suffix = ''\n else:\n if '|||' in x:\n if '||||' in x:\n y = x.split('||||')\n paragraf.append((y[0], False))\n document_close_elements.append(y[1])\n else:\n y = x.split('|||')\n paragraf.append((y[0], False))\n section_close_elements.append(y[1])\n else:\n paragraf.append((x, False))\n buf = []\n in_wiki_object = False\n else:\n if line.startswith('@'):\n if buf:\n paragraf.append((buf, True))\n buf = []\n in_wiki_object = True\n name = line.split(':')[0][1:].strip()\n else:\n if line.startswith('...') or line.startswith('+++'):\n write_papragraf()\n if line.startswith('+++'):\n write_section()\n paragraf_prefix = ''\n paragraf_suffix = ''\n else:\n buf.append(line)\n\n write_papragraf()\n write_section()\n write_document()\n return '\\n'.join(document)\n\n\npage_type_choices = (('W', 'Wiki'), ('I', 'Indent html'), ('H', 'Html'))\n\nclass PageObjectsConf(models.Model):\n\n class Meta:\n verbose_name = _('Page objects configurations')\n verbose_name_plural = _('Page objects configurations')\n default_permissions = ('add', 'change', 'delete', 'list')\n app_label = 'schwiki'\n ordering = [\n 'id']\n\n app = models.CharField('Application', null=False, blank=False, editable=True, max_length=32)\n name = models.CharField('Name', null=False, blank=False, editable=True, max_length=64)\n description = models.CharField('Description', null=True, blank=True, editable=True, max_length=128)\n inline_editing = ext_models.NullBooleanField('Inline editing', null=False, blank=False, editable=True, default=False)\n inline_wiki = ext_models.NullBooleanField('Inline wiki', null=False, blank=False, editable=True, default=False)\n edit_form = models.TextField('Edit form', null=True, blank=True, editable=False)\n load_fun = models.TextField('Load function', null=True, blank=True, editable=False)\n save_fun = models.TextField('Save function', null=True, blank=True, editable=False)\n view_dict = models.TextField('Get view dict function', null=True, blank=True, editable=False)\n doc = models.TextField('Documentaction', null=True, blank=True, editable=False)\n\n def __str__(self):\n return self.name\n\n\nadmin.site.register(PageObjectsConf)\n\nclass Page(JSONModel):\n\n class Meta:\n verbose_name = _('Page')\n verbose_name_plural = _('Page')\n default_permissions = ('add', 'change', 'delete', 'list')\n app_label = 'schwiki'\n ordering = [\n 'id']\n\n subject = models.CharField('Subject', null=False, blank=False, editable=True, max_length=64)\n name = models.CharField('Name', null=False, blank=False, editable=True, max_length=64)\n description = models.CharField('Description', null=True, blank=True, editable=True, max_length=64)\n content_src = models.TextField('Content source', null=True, blank=True, editable=False)\n content = models.TextField('Content', null=True, blank=True, editable=False)\n base_template = models.CharField('Base template', null=True, blank=True, editable=True, max_length=64)\n rights_group = models.CharField('Rights group', null=True, blank=True, editable=True, max_length=64)\n menu = models.CharField('Menu', null=True, blank=True, editable=True, max_length=64)\n operator = models.CharField('Operator', null=True, blank=True, editable=False, max_length=64)\n update_time = models.DateTimeField('Update time', null=False, blank=False, editable=False, default=(datetime.now))\n published = ext_models.NullBooleanField('Published', null=False, blank=False, editable=False, default=False)\n latest = ext_models.NullBooleanField('Latest', null=False, blank=False, editable=False, default=True)\n\n def save_from_request(self, request, view_type, param):\n if 'direct_save' in request.POST:\n super(Page, self).save()\n else:\n conf_list = WikiConf.objects.filter(subject=(self.subject))\n conf_exists = False\n if len(conf_list) > 0:\n conf = conf_list[0]\n conf_exists = True\n if conf.backup_copies > 0:\n pages = Page.objects.filter(subject=(self.subject), name=(self.name)).update(latest=False)\n obj_to_save = Page()\n obj_to_save.subject = self.subject\n obj_to_save.name = self.name\n obj_to_save.description = self.description\n obj_to_save.content_src = self.content_src\n obj_to_save.content = self.content\n obj_to_save.base_template = self.base_template\n obj_to_save.rights_group = self.rights_group\n obj_to_save.menu = self.menu\n obj_to_save.operator = self.operator\n obj_to_save.update_time = self.update_time\n obj_to_save.jsondata = self.jsondata\n obj_to_save.published = False\n obj_to_save.latest = True\n obj_to_save.operator = request.user.username\n obj_to_save.update_time = datetime.now()\n obj_to_save.save()\n pages = Page.objects.filter(subject=(self.subject), name=(self.name)).order_by('update_time')\n if len(pages) > conf.backup_copies:\n to_delete_count = len(pages) - conf.backup_copies\n to_delete = []\n for pos in pages:\n if not pos.published:\n if not pos.latest:\n to_delete.append(pos)\n to_delete_count -= 1\n if to_delete_count <= 0:\n break\n\n if to_delete:\n for pos2 in to_delete:\n pos2.delete()\n\n return\n self.operator = request.user.username\n self.update_time = datetime.now()\n self.latest = True\n if not conf_exists:\n self.published = True\n self.save()\n\n def save(self, *args, **kwargs):\n (super(Page, self).save)(*args, **kwargs)\n if self.content_src:\n content = html_from_wiki(self, self.content_src + '\\n\\n\\n ')\n else:\n content = ''\n self.content = content\n (super(Page, self).save)(*args, **kwargs)\n\n def transform_template_name(self, request, template_name):\n return 'schwiki/edit_wiki_content.html'\n\n def get_form(self, view, request, form_class, adding):\n pass\n\n def get_page_for_wiki(self, wiki_str, user=None):\n wiki_word = wiki_from_str(wiki_str)\n return Page.get_page(user, self.subject, wiki_word)\n\n def get_href(self, path=None):\n return make_href((self.description if self.description else self.name), new_win=False, section=(self.subject), path=path)\n\n @staticmethod\n def get_page(request_or_username, subject, name):\n if type(request_or_username) == str:\n username = request_or_username\n else:\n if request_or_username.user:\n username = request_or_username.user.username\n else:\n username = None\n objs = None\n if username:\n objs = Page.objects.filter(subject=subject, name=name, operator=username, latest=True)\n if not objs or len(objs) == 0:\n objs = Page.objects.filter(subject=subject, name=name, published=True)\n if not objs or len(objs) == 0:\n objs = Page.objects.filter(subject=subject, name=name)\n if len(objs) > 0:\n return objs[0]\n else:\n return\n\n def __str__(self):\n return self.name\n\n\nadmin.site.register(Page)\n\nclass WikiConf(JSONModel):\n\n class Meta:\n verbose_name = _('Wiki config')\n verbose_name_plural = _('Wiki config')\n default_permissions = ('add', 'change', 'delete', 'list')\n app_label = 'schwiki'\n ordering = [\n 'id']\n\n subject = models.CharField('Wiki subject', null=False, blank=False, editable=True, max_length=64)\n group_of_rights_to_view = models.CharField('A group of rights to view wiki', null=True, blank=True, editable=True, max_length=64)\n group_of_rights_to_edit = models.CharField('A group of rights to edit wiki', null=True, blank=True, editable=True, max_length=64)\n backup_copies = models.IntegerField('Number of backup copies', null=False, blank=False, editable=True)\n publish_fun = models.TextField('Function called after publishing', null=True, blank=True, editable=False)\n scss = models.TextField('Additional scss styles', null=True, blank=True, editable=False)\n css = models.TextField('Css styles', null=True, blank=True, editable=False)\n\n def get_css(self):\n import sass\n if self.scss:\n buf = self.scss.replace('page_class', 'wiki_' + self.subject.lower())\n style = sass.compile(string=buf, indented=True)\n return style\n else:\n return ''\n\n def save(self, *args, **kwargs):\n self.css = self.get_css()\n ret = (super().save)(*args, **kwargs)\n return ret\n\n\nadmin.site.register(WikiConf)", "sub_path": "pycfiles/pytigon-0.99-py3-none-any/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 14243, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pytigon_lib.schtools.tools.norm_indent", "line_number": 38, "usage_type": "call"}, {"api_name": "django.template.loader.select_template", "line_number": 47, "usage_type": "call"}, {"api_name": "markdown2.markdown", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models.Model", "line_number": 162, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 162, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 165, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 166, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 172, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 172, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 173, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 173, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 174, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 174, "usage_type": "name"}, {"api_name": "pytigon_lib.schdjangoext.fields.NullBooleanField", "line_number": 175, "usage_type": "call"}, {"api_name": "pytigon_lib.schdjangoext.fields", "line_number": 175, "usage_type": "name"}, {"api_name": "pytigon_lib.schdjangoext.fields.NullBooleanField", "line_number": 176, "usage_type": "call"}, {"api_name": "pytigon_lib.schdjangoext.fields", "line_number": 176, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 177, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 177, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 178, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 178, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 179, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 179, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 180, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 180, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 181, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 181, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 187, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 187, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 187, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 192, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 193, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 199, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 199, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 200, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 200, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 201, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 201, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 202, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 202, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 203, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 203, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 204, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 204, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 205, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 205, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 206, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 206, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 207, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 207, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 208, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 208, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 208, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 208, "usage_type": "name"}, {"api_name": "pytigon_lib.schdjangoext.fields.NullBooleanField", "line_number": 209, "usage_type": "call"}, {"api_name": "pytigon_lib.schdjangoext.fields", "line_number": 209, "usage_type": "name"}, {"api_name": "pytigon_lib.schdjangoext.fields.NullBooleanField", "line_number": 210, "usage_type": "call"}, {"api_name": "pytigon_lib.schdjangoext.fields", "line_number": 210, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 238, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 238, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 258, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 258, "usage_type": "name"}, {"api_name": "pytigon_lib.schtools.wiki.wiki_from_str", "line_number": 280, "usage_type": "call"}, {"api_name": "pytigon_lib.schtools.wiki.make_href", "line_number": 284, "usage_type": "call"}, {"api_name": "django.contrib.admin.site.register", "line_number": 311, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 311, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 311, "usage_type": "name"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 316, "usage_type": "call"}, {"api_name": "django.utils.translation.ugettext_lazy", "line_number": 317, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 323, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 323, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 324, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 324, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 325, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 325, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 326, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 326, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 327, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 327, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 328, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 328, "usage_type": "name"}, {"api_name": "django.db.models.TextField", "line_number": 329, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 329, "usage_type": "name"}, {"api_name": "sass.compile", "line_number": 335, "usage_type": "call"}, {"api_name": "django.contrib.admin.site.register", "line_number": 346, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 346, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 346, "usage_type": "name"}]} +{"seq_id": "267920795", "text": "from functools import lru_cache\r\n\r\ndef frog_escape(swamp):\r\n @lru_cache(maxsize=None)\r\n def jump(position, e):\r\n if position >= len(swamp): #rebni pogoj: ce je zacetna pozicija >= enaka dolzini mocvare, je zabica uspesno pobegnila\r\n return 0\r\n else:\r\n e += swamp[position] #se postavimo na dano pozicijo in dobimo swamp[position] energije, to je toliko muh, kot se nahaja na danem polju\r\n return 1 + min([jump(position + d, e - d) for d in range(1, e + 1)]) #seznam vrednosti useh moznih skokov\r\n #porabimo en skos plus minimum vseh moznosti, ko skocimo za d naprej, pri cemer d med 1 in e\r\n #obe meji vkljuceni, range(1, e+1) da [1, e+1)\r\n #pri tem se energija zmanjsa za d, trenutna pozicija pa gre na position + d\r\n return jump(0, 0) #zacnemo na zacetku z 0 energije\r\n\r\n\r\ntest1 = [2, 4, 1, 2, 1, 3, 1, 1, 5]\r\ntest2 = [4, 1, 8, 2, 11, 1, 1, 1, 1, 1]", "sub_path": "vaje/3.naloga zabica.py", "file_name": "3.naloga zabica.py", "file_ext": "py", "file_size_in_byte": 940, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "functools.lru_cache", "line_number": 4, "usage_type": "call"}]} +{"seq_id": "451900404", "text": "from tensorflow.keras.models import load_model\nimport tensorflow.keras.backend as K\nimport os, sys, gc, math, argparse\nimport tensorflow as tf\nimport pandas as pd\nimport numpy as np\nimport shap\nfrom Bio.SeqRecord import SeqRecord\nfrom Bio.Seq import Seq\nfrom Bio import SeqIO\nimport sys\nfrom sklearn import metrics\nfrom callback_ocp import *\nfrom cnn_utils import *\n\n# gc.collect()\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '3' # or any {'0', '1', '2'}\n# make sure running on GPU #\ntf.config.experimental.list_physical_devices('GPU')\n# tf.get_logger().setLevel('WARNING')\nnp.seterr(divide = 'ignore')\n\n\n\ndef print_2darray(arr):\n s = ''\n for i in range(arr.shape[0]):\n for j in range(arr.shape[1]):\n if j == 0 and i == 0: #first entry\n s = str(arr[i,j][0])\n elif j == 0: # new line\n s = s + ';' + str(arr[i,j][0])\n else: # all others\n s = s + ',' + str(arr[i,j][0])\n return s\n\n\n\ndef print_1darray(arr):\n s = ''\n for i in range(arr.shape[0]):\n s = s + arr[i][0]\n return s\n\n\n\ndef onehot2fasta(arr,ids,args):\n letters = np.array(['A','C','G','T'])[np.argmax(arr, axis=2)]\n my_seq = [SeqRecord(seq=Seq(print_1darray(letters[i,])), id=ids[i].split('_')[1], description = '',name='') for i in range(len(ids))]\n return my_seq\n\n\n\ndef predict_sequences(model_name, x, ids):\n # Creating an empty Dataframe with column names only\n # predict labels\n # combineStrands will average the logit of the for and rev DNA strands\n model = load_model(model_name, compile=False)\n y_pred_score = model.predict(x).flatten()\n y_pred_score = np.nan_to_num(y_pred_score)\n df = pd.DataFrame({'id': ids, 'y_pred_logit': np.log10(y_pred_score) - np.log10(1-y_pred_score)})\n df = df.groupby(['id']).mean()\n df['y_pred_score'] = 1 / (1 + np.exp(-df['y_pred_logit']))\n df['y_pred_class'] = df['y_pred_score'] > 0.5\n return df\n\n\n\ndef evaluate_SHAP_scores(args, fg, fg_ids, bg):\n ## create deep SHAP explainer for bg vs. fg\n minRepBg = np.min([bg.shape[0], args.numBg])\n minRepFg = np.min([fg.shape[0], args.numFg])\n np.random.seed(seed=args.seed)\n bg = bg[np.random.choice(bg.shape[0], minRepBg, replace = False)]\n indFg = np.random.choice(fg.shape[0], minRepFg, replace = False)\n fg = fg[indFg]; ids = fg_ids[indFg]\n chunks = 10\n fg_list = np.array_split(fg, fg.shape[0] // chunks)\n ## initialize deep SHAP explainer for bg vs. fg & compute fg in 100 seq chunks\n model = load_model(args.model_name, compile=False)\n explainer = shap.DeepExplainer(model, bg)\n rawShapList = []\n for ind in range(0,len(fg_list)):\n print(f'Computing SHAP values: {(ind + 1) * chunks} of {fg.shape[0]}.')\n raw_shap_explanations = explainer.shap_values(fg_list[ind], check_additivity = False)\n rawShapList.append(raw_shap_explanations)\n # combine explanations and normalize\n hyp = np.squeeze(np.concatenate( rawShapList, axis = 1 ), axis=0)\n imp = hyp*fg # mask hypothetical w/ actual sequences\n return (ids, hyp, imp, fg)\n\n\n\ndef main(args):\n \"\"\"Main function\n Args:\n args (argparse):\n \"\"\"\n # file names\n # call main functions\n if not os.path.exists(args.model_name):\n print('No model found with specified training parameters. Please train model first.')\n return\n if not os.path.exists(f'{args.out_dir}/shap'):\n os.makedirs(f'{args.out_dir}/shap')\n if args.mode == 'evaluate':\n print('In evaluation mode to compute SHAP scores.')\n print(f'Reading in positive sequences from {args.eval_fasta_pos}.')\n (x_pos, ids_pos) = encode_sequence3(args.eval_fasta_pos, size = args.seq_length, shuffleOff = True)\n x_pos = x_pos[range(len(ids_pos)//2),:]; ids_pos = ids_pos[range(len(ids_pos)//2)]\n df_pos = predict_sequences(args.model_name, x_pos, ids_pos).loc[ids_pos,:]\n\n indKeep_pos = np.where(df_pos['y_pred_class'] == 1)[0]\n \n\n #x_pos = x_pos[indKeep_pos]; ids_pos = ids_pos[indKeep_pos\n # uncomment for only positive seqs \n df_pos = df_pos.loc[ids_pos,:]\n #\n print(f'Reading in negative sequences from {args.eval_fasta_neg}.')\n (x_neg, ids_neg) = encode_sequence3(args.eval_fasta_neg, size = args.seq_length, shuffleOff = True)\n x_neg = x_neg[range(len(ids_neg)//2),:]; ids_neg = ids_neg[range(len(ids_neg)//2)]\n df_neg = predict_sequences(args.model_name, x_neg, ids_neg).loc[ids_neg,:]\n indKeep_neg = np.where(df_neg['y_pred_class'] == 0)[0]\n x_neg = x_neg[indKeep_neg]; ids_neg = ids_neg[indKeep_neg]\n df_neg = df_neg.loc[ids_neg,:]\n #\n # get hypothetical and importance scores\n print(f'Computing SHAP scores with true positives and true negatives.')\n (ids, hyp, imp, fg) = evaluate_SHAP_scores(args, fg = x_pos, fg_ids = ids_pos, bg = x_neg)\n df = df_pos.loc[ids,:]\n #\n # save hypothetical importance scores\n imp_out = pd.DataFrame({'id': df.index.str.split(pat = '_').str[1],\n 'y_pred_logit': df['y_pred_logit'],\n 'SHAP_score': [print_2darray(imp[i,:]) for i in range(imp.shape[0])]})\n imp_fn = f'{args.out_dir}/shap/{args.predict_out}.pos{args.numFg}.imp_SHAP_scores.txt'\n print(f'Writing importance scores to {imp_fn}.')\n imp_out.to_csv(imp_fn, index = False, sep = '\\t', header = False)\n #\n hyp_out = pd.DataFrame({'id': df.index.str.split(pat = '_').str[1],\n 'y_pred_logit': df['y_pred_logit'],\n 'SHAP_score': [print_2darray(hyp[i,:]) for i in range(hyp.shape[0])]})\n hyp_fn = f'{args.out_dir}/shap/{args.predict_out}.pos{args.numFg}.hyp_SHAP_scores.txt'\n print(f'Writing hypothetical importance scores to {hyp_fn}.')\n hyp_out.to_csv(hyp_fn, index = False, sep = '\\t', header = False)\n #\n fasta_out = onehot2fasta(fg,ids,args)\n fasta_fn = f'{args.out_dir}/shap/{args.predict_out}.pos{args.numFg}.fasta'\n print(f'Writing scored DNA sequences {fasta_fn}.')\n SeqIO.write(fasta_out, fasta_fn, \"fasta\")\n #\n elif args.mode == 'predict':\n print('In prediction mode to compute SHAP scores.')\n print(f'Reading in sequences from {args.predict_fasta}.')\n (x, ids) = encode_sequence3(args.predict_fasta, size = args.seq_length, shuffleOff = True)\n df = predict_sequences(args.model_name, x, ids).loc[ids,:]\n ids = np.array([ ids[i] if i < len(ids)//2 else ids[i] + '.rev' for i in range(len(ids)) ])\n df.index = ids; args.numFg = len(ids)\n #\n print(f'Reading in negative sequences from {args.eval_fasta_neg}.')\n (x_neg, ids_neg) = encode_sequence3(args.eval_fasta_neg, size = args.seq_length, shuffleOff = True)\n x_neg = x_neg[range(len(ids_neg)//2),:]; ids_neg = ids_neg[range(len(ids_neg)//2)]\n df_neg = predict_sequences(args.model_name, x_neg, ids_neg).loc[ids_neg,:]\n indKeep_neg = np.where(df_neg['y_pred_class'] == 0)[0]\n x_neg = x_neg[indKeep_neg]; ids_neg = ids_neg[indKeep_neg]\n df_neg = df_neg.loc[ids_neg,:]\n #\n # get hypothetical and importance scores\n print(f'Computing SHAP scores of sequences against true negatives.')\n (ids, hyp, imp, fg) = evaluate_SHAP_scores(args, fg = x, fg_ids = ids, bg = x_neg)\n df = df.loc[ids,:]\n #\n # save hypothetical importance scores\n imp_out = pd.DataFrame({'id': df.index.str.split(pat = '_').str[1],\n 'y_pred_logit': df['y_pred_logit'],\n 'SHAP_score': [print_2darray(imp[i,:]) for i in range(imp.shape[0])]})\n imp_fn = f'{args.out_dir}/shap/{args.predict_out}.predict.imp_SHAP_scores.txt'\n print(f'Writing importance scores to {imp_fn}.')\n imp_out.to_csv(imp_fn, index = False, sep = '\\t', header = False)\n #\n hyp_out = pd.DataFrame({'id': df.index.str.split(pat = '_').str[1],\n 'y_pred_logit': df['y_pred_logit'],\n 'SHAP_score': [print_2darray(hyp[i,:]) for i in range(hyp.shape[0])]})\n hyp_fn = f'{args.out_dir}/shap/{args.predict_out}.predict.hyp_SHAP_scores.txt'\n print(f'Writing hypothetical importance scores to {hyp_fn}.')\n hyp_out.to_csv(hyp_fn, index = False, sep = '\\t', header = False)\n #\n fasta_out = onehot2fasta(fg,ids,args)\n fasta_fn = f'{args.out_dir}/shap/{args.predict_out}.pos{args.numFg}.fasta'\n print(f'Writing scored DNA sequences {fasta_fn}.')\n SeqIO.write(fasta_out, fasta_fn, \"fasta\")\n return\n\n\n\nif __name__ == '__main__':\n #### set cnn parameters:\n parser = argparse.ArgumentParser(description='Parse CNN parameters.')\n parser.add_argument(\"--mode\", type=str, help=\"Mode to perform. Train needs all fasta. Evaluate needs validation fasta. Predict only fasta passed predict_fasta.\",\n choices=[ 'evaluate', 'predict'], default = 'evaluate', required=False)\n #\n parser.add_argument(\"--model_name\", type=str, help=\"complete model name\", required=True)\n parser.add_argument(\"--predict_out\", type=str, help=\"prediction file prefix model name\", required=True)\n parser.add_argument(\"--predict_fasta\", type=str, help=\"fasta sequence file for predictions.\", required=False)\n parser.add_argument(\"--eval_fasta_pos\", type=str, help=\"validation fasta sequence file of positives.\", required=False)\n parser.add_argument(\"--eval_fasta_neg\", type=str, help=\"validation fasta sequence file of negatives.\", required=True)\n parser.add_argument(\"--numFg\", type=int, default = 2000, help=\"number of positives to evaluate with SHAP scores. Note that all predict fasta are scored.\", required=False)\n parser.add_argument(\"--numBg\", type=int, default = 500, help=\"number of negatives in background to compute SHAP scores\", required=False)\n parser.add_argument(\"--seq_length\", type=str, default = 500, help=\"number of negatives in background to compute SHAP scores\", required=False)\n parser.add_argument(\"--seed\", type=int, default = 1, help=\"number of negatives\", required=False)\n parser.add_argument(\"--force\", help=\"Whether to overwrite previously trained model.\", action='store_true')\n parser.add_argument(\"--verbose\", type=int, default = 2, help=\"Verbosity in keras.\")\n parser.add_argument(\"--out_dir\", type=str, default = '.', help=\"path to ouputput directory, default is pwd\")\n\n # args = parser.parse_args(['--model_name=/projects/pfenninggroup/machineLearningForComputationalBiology/addiction_gwas_enrichment/celltype_specific_ml_models/models/Mo2015_EXCpos_Ctx_fold1/Mo2015_EXCpos_Ctx_fold1_OCP_NB1000_NE23_BR0.01_MR0.1_BM0.85_MM0.99_DO0.25.h5',\n # '--predict_out=Mo2015_EXCpos_Ctx_fold1',\n # '--eval_fasta_pos=FASTA_CV/Mo2015_EXCpos_Ctx_fold1_validPos.fa',\n # '--eval_fasta_neg=FASTA_CV/Mo2015_EXCpos_Ctx_fold1_validNeg10x.fa',\n # '--mode=evaluate'])\n\n # args = parser.parse_args(['--model_name=/projects/pfenninggroup/machineLearningForComputationalBiology/addiction_gwas_enrichment/celltype_specific_ml_models/models/Mo2015_EXCpos_Ctx_fold1/Mo2015_EXCpos_Ctx_fold1_OCP_NB1000_NE23_BR0.01_MR0.1_BM0.85_MM0.99_DO0.25.h5',\n # '--predict_out=top_addiction_snps_effect_allele.Mo2015_EXCpos_Ctx_fold1',\n # '--predict_fasta=FASTA/top_addiction_snps_allele_effect_501.fa',\n # '--eval_fasta_neg=FASTA_CV/Mo2015_EXCpos_Ctx_fold1_validNeg10x.fa',\n # '--mode=predict'])\n\n # args.numBg = 100\n\n args = parser.parse_args()\n main(args)\n\n\n\n'''\ncommand to run 06/15/21\npython step12a_deepshap.py --model_name /projects/pfenninggroup/machineLearningForComputationalBiology/retina/models/mouse4_v3/tmp.h5 --predict_out test2 --eval_fasta_pos /projects/pfenninggroup/machineLearningForComputationalBiology/retina/models/mouse4_v3/pos_VALIDATION.fa --eval_fasta_neg /projects/pfenninggroup/machineLearningForComputationalBiology/retina/models/mouse4_v3/neg_VALIDATION.fa\n'''\n", "sub_path": "evaluationScriptsRetinaModels/step12a_deepshap.py", "file_name": "step12a_deepshap.py", "file_ext": "py", "file_size_in_byte": 11965, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "tensorflow.config.experimental.list_physical_devices", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 19, "usage_type": "attribute"}, {"api_name": "numpy.seterr", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 48, "usage_type": "call"}, {"api_name": "Bio.SeqRecord.SeqRecord", "line_number": 49, "usage_type": "call"}, {"api_name": "Bio.Seq.Seq", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.nan_to_num", "line_number": 60, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 71, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 74, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.array_split", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 80, "usage_type": "call"}, {"api_name": "shap.DeepExplainer", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 124, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 134, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 141, "usage_type": "call"}, {"api_name": "Bio.SeqIO.write", "line_number": 151, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 151, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 165, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 175, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 182, "usage_type": "call"}, {"api_name": "Bio.SeqIO.write", "line_number": 192, "usage_type": "call"}, {"api_name": "Bio.SeqIO", "line_number": 192, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 199, "usage_type": "call"}]} +{"seq_id": "218372344", "text": "from django.conf import settings\nfrom django.contrib.auth.models import User\nimport urllib\nimport json\nfrom auth.models import FacebookProfile\nimport time\nfrom keepbrain import settings as my_settings\n\nclass FacebookBackend(object):\n def authenticate(self, request):\n code = request.GET['code']\n params = {\n 'code': code,\n 'client_id': my_settings.FACEBOOK_CREDENTIALS['KEY'],\n 'client_secret': my_settings.FACEBOOK_CREDENTIALS['SECRET'],\n 'redirect_uri': 'http://' + request.get_host() + request.path,\n }\n\n handler = urllib.urlopen('https://graph.facebook.com/oauth/access_token?%s' % urllib.urlencode(params))\n token = handler.read()\n\n handler = urllib.urlopen('https://graph.facebook.com/me?%s' % token)\n response = handler.read()\n response = json.loads(response)\n \n try:\n profile = FacebookProfile.objects.get(id=response['id'])\n user = profile.user\n except FacebookProfile.DoesNotExist:\n uid = 'facebook_%s' % str(time.time())\n user = User.objects.create(username=uid,first_name=response['first_name'],last_name=response['last_name'])\n FacebookProfile.objects.create(id=response['id'], user=user)\n \n return user\n \n def get_user(self, user_id):\n try:\n return User.objects.get(pk=user_id)\n except User.DoesNotExist:\n return None\n", "sub_path": "auth/backends.py", "file_name": "backends.py", "file_ext": "py", "file_size_in_byte": 1475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "keepbrain.settings.FACEBOOK_CREDENTIALS", "line_number": 14, "usage_type": "attribute"}, {"api_name": "keepbrain.settings", "line_number": 14, "usage_type": "name"}, {"api_name": "keepbrain.settings.FACEBOOK_CREDENTIALS", "line_number": 15, "usage_type": "attribute"}, {"api_name": "keepbrain.settings", "line_number": 15, "usage_type": "name"}, {"api_name": "urllib.urlopen", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 19, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 22, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "auth.models.FacebookProfile.objects.get", "line_number": 27, "usage_type": "call"}, {"api_name": "auth.models.FacebookProfile.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "auth.models.FacebookProfile", "line_number": 27, "usage_type": "name"}, {"api_name": "auth.models.FacebookProfile.DoesNotExist", "line_number": 29, "usage_type": "attribute"}, {"api_name": "auth.models.FacebookProfile", "line_number": 29, "usage_type": "name"}, {"api_name": "time.time", "line_number": 30, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects.create", "line_number": 31, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 31, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 31, "usage_type": "name"}, {"api_name": "auth.models.FacebookProfile.objects.create", "line_number": 32, "usage_type": "call"}, {"api_name": "auth.models.FacebookProfile.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "auth.models.FacebookProfile", "line_number": 32, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.objects.get", "line_number": 38, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 38, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User.DoesNotExist", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.contrib.auth.models.User", "line_number": 39, "usage_type": "name"}]} +{"seq_id": "237891036", "text": "#! /usr/bin/env python\n\n# plotting rover q generalized coordinates.\n\nimport matplotlib\nmatplotlib.use('Qt4Agg')\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfrom matplotlib import rc, rcParams\n\nrc('text', usetex=True)\n\nt, l1, l2, l3, l4, l5, l6 = np.loadtxt('LAMBDAS.dat', usecols=(0, 1, 4, 7, 10, 13, 16), unpack=True)\n\nfig = plt.figure(1)\n\nax = fig.add_subplot(111)\n\nax.plot(t, l1, label='Wheel FL')\nax.plot(t, l2, label='Wheel FR')\nax.plot(t, l3, label='Wheel ML')\nax.plot(t, l4, label='Wheel BL')\nax.plot(t, l5, label='Wheel MR')\nax.plot(t, l6, label='Wheel BR')\n\n#ax.set_xlim([0, 10])\n\nplt.title(r'$\\lambda_{N}$ $for$ $each$ $wheel$')\nplt.xlabel(r'$T [s]$')\nplt.ylabel(r'$\\lambda_{N}$ $[N]$')\n\nplt.legend(loc='center right')\n\nseq = [7, 4, 3, 4]\n\nplt.axvline(x=50, color='black', dashes=seq)\nplt.axvline(x=100, color='black', dashes=seq)\nplt.axvline(x=150, color='black', dashes=seq)\nplt.axvline(x=220, color='black', dashes=seq)\nplt.axvline(x=300, color='black', dashes=seq)\n\nplt.show()\n\n", "sub_path": "src/Rover3D_On_TriangleMesh/MechTests/MechTest3/lambdaNormal.py", "file_name": "lambdaNormal.py", "file_ext": "py", "file_size_in_byte": 1001, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "matplotlib.use", "line_number": 6, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axvline", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}]} +{"seq_id": "550412413", "text": "import requests\nimport os\nimport time\n\neis_url = os.environ.get(\"URL\")\n\n\ndef main_handler(event, context):\n try:\n for _ in range(3):\n r = requests.post(eis_url, json=event)\n if r.status_code in (429, 503):\n time.sleep(1)\n continue\n else:\n print(r.text)\n break\n except Exception as e:\n print(e)\n\n", "sub_path": "Python3.6-EBEISTarget/src/index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 408, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.environ.get", "line_number": 5, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 5, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 11, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "398070427", "text": "import numpy as np\nimport math\nimport matplotlib.pyplot as plt\n\narrs = np.arange(0,10,0.01)\nA = 5\n\nfff_sin = lambda k:A*np.sin(arrs)\nfsq_sin = lambda k:A*(1.0/(2*k+1))*np.sin(2*math.pi*(2*k+1)*arrs)\nfsm_sin = lambda k:A*(1.0/(k))*np.sin(2*math.pi*(k)*arrs)\n\ns=np.zeros(len(arrs))\nfor i in range(1,100):\n\ts+=fsm_sin(i)\n\nplt.plot(arrs, s)\n# plt.plot(arrs, fsm_sin(1))\n# plt.show()\nplt.savefig('home.png')\n\n# print(fsin(5,6))", "sub_path": "home.py", "file_name": "home.py", "file_ext": "py", "file_size_in_byte": 422, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.arange", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 9, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 10, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 12, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 16, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}]} +{"seq_id": "124559379", "text": "# -*- coding: utf-8 -*-\nfrom django import template\nfrom django.conf import settings\nfrom django.template.defaulttags import URLNode\nfrom django.contrib.sites.models import Site\nimport os\nimport urlparse\n\nregister = template.Library()\n\ndef _absolute_url(url):\n if url.startswith('http://') or url.startswith('https://'):\n return url\n domain = Site.objects.get_current().domain\n return 'http://%s%s' % (domain, url)\n\n@register.simple_tag\ndef media(filename, flags=''):\n flags = set(f.strip() for f in flags.split(','))\n url = urlparse.urljoin(settings.MEDIA_URL, filename)\n if 'absolute' in flags:\n url = _absolute_url(url)\n if (filename.endswith('.css') or filename.endswith('.js')) and 'no-timestamp' not in flags or \\\n 'timestamp' in flags:\n fullname = os.path.join(settings.MEDIA_ROOT, filename)\n if os.path.exists(fullname):\n url += '?%d' % os.path.getmtime(fullname)\n return url\n\n@register.simple_tag\ndef static_media(filename):\n return urlparse.urljoin(settings.STATIC_MEDIA_URL, filename)", "sub_path": "app/main/templatetags/media.py", "file_name": "media.py", "file_ext": "py", "file_size_in_byte": 1067, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.template.Library", "line_number": 9, "usage_type": "call"}, {"api_name": "django.template", "line_number": 9, "usage_type": "name"}, {"api_name": "django.contrib.sites.models.Site.objects.get_current", "line_number": 14, "usage_type": "call"}, {"api_name": "django.contrib.sites.models.Site.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "django.contrib.sites.models.Site", "line_number": 14, "usage_type": "name"}, {"api_name": "urlparse.urljoin", "line_number": 20, "usage_type": "call"}, {"api_name": "django.conf.settings.MEDIA_URL", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.conf.settings.MEDIA_ROOT", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 25, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.getmtime", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "urlparse.urljoin", "line_number": 32, "usage_type": "call"}, {"api_name": "django.conf.settings.STATIC_MEDIA_URL", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 32, "usage_type": "name"}]} +{"seq_id": "308260546", "text": "import logging\n\nimport coloredlogs\n\nfrom JanggiCoach import JanggiCoach as Coach\nfrom janggi.JanggiGame import JanggiGame as Game\nfrom janggi.pytorch.NNet import NNetWrapper as nn\nfrom utils import *\n\nimport torch\n\nimport JanggiMainConstants as JMC\n\nlog = logging.getLogger(__name__)\n\ncoloredlogs.install(level='INFO') # Change this to DEBUG to see more info.\n\nargs = dotdict({\n # 'numIters': JMC.numIters,\n # 'numEps': JMC.numEps, # Number of complete self-play games to simulate during a new iteration.\n # 'updateThreshold': JMC.updateThreshold, # During arena playoff, new neural net will be accepted if threshold or more of games are won.\n # 'arenaCompare': JMC.arenaCompare, # Number of games to play during arena play to determine if new net will be accepted.\n 'tempThreshold': JMC.tempThreshold, # \n 'maxlenOfQueue': JMC.maxlenOfQueue, # Number of game examples to train the neural networks.\n 'maxDataCount': JMC.maxDataCount,\n 'numMCTSSims': JMC.numMCTSSims, # Number of games moves for MCTS to simulate.\n 'cpuct': JMC.cpuct,\n\n # 'checkpoint': JMC.checkpoint,\n 'load_model': JMC.load_model,\n 'load_folder_file': JMC.load_folder_file,\n 'numItersForTrainExamplesHistory': JMC.numItersForTrainExamplesHistory,\n 'checkpoint_folder': JMC.checkpoint_folder,\n 'remote_checkpoint_folder': JMC.remote_checkpoint_folder,\n # 'checkpoint_folders': JMC.checkpoint_folders,\n\n 'num_gpu_procs': JMC.num_gpu_procs,\n 'num_selfplay_procs': JMC.num_selfplay_procs,\n 'gpus_to_use': JMC.gpus_to_use,\n\n 'is_training_client' : JMC.is_training_client,\n 'request_base_url': JMC.request_base_url,\n 'scp_base_url': JMC.scp_base_url,\n\n 'trainFrequency': JMC.trainFrequency, # Update the network everytime after trainFrequency selfplays are done.\n})\n\n\ndef main():\n log.info('GPU availability: %s', torch.cuda.is_available())\n log.info('Loading %s...', Game.__name__)\n g = Game(0, 0)\n\n log.info('Loading %s...', nn.__name__)\n nnet = nn(g)\n\n if args.load_model:\n log.info('Loading checkpoint \"%s/%s\"...', args.load_folder_file)\n nnet.load_checkpoint(args.load_folder_file[0], args.load_folder_file[1])\n else:\n log.warning('Not loading a checkpoint!')\n\n log.info('Loading the Coach...')\n c = Coach(g, nnet, args, JMC.selfPlaysPlayed)\n\n if args.load_model:\n log.info(\"Loading 'trainExamples' from file...\")\n c.loadTrainExamples()\n\n log.info('Starting the learning process 🎉')\n c.learn()\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "JanggiMain.py", "file_name": "JanggiMain.py", "file_ext": "py", "file_size_in_byte": 2594, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 14, "usage_type": "call"}, {"api_name": "coloredlogs.install", "line_number": 16, "usage_type": "call"}, {"api_name": "JanggiMainConstants.tempThreshold", "line_number": 23, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.maxlenOfQueue", "line_number": 24, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.maxDataCount", "line_number": 25, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.numMCTSSims", "line_number": 26, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.cpuct", "line_number": 27, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.load_model", "line_number": 30, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.load_folder_file", "line_number": 31, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.numItersForTrainExamplesHistory", "line_number": 32, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.checkpoint_folder", "line_number": 33, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.remote_checkpoint_folder", "line_number": 34, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.num_gpu_procs", "line_number": 37, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.num_selfplay_procs", "line_number": 38, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.gpus_to_use", "line_number": 39, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.is_training_client", "line_number": 41, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.request_base_url", "line_number": 42, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.scp_base_url", "line_number": 43, "usage_type": "attribute"}, {"api_name": "JanggiMainConstants.trainFrequency", "line_number": 45, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 50, "usage_type": "attribute"}, {"api_name": "janggi.JanggiGame.JanggiGame.__name__", "line_number": 51, "usage_type": "attribute"}, {"api_name": "janggi.JanggiGame.JanggiGame", "line_number": 51, "usage_type": "name"}, {"api_name": "janggi.JanggiGame.JanggiGame", "line_number": 52, "usage_type": "call"}, {"api_name": "janggi.pytorch.NNet.NNetWrapper.__name__", "line_number": 54, "usage_type": "attribute"}, {"api_name": "janggi.pytorch.NNet.NNetWrapper", "line_number": 54, "usage_type": "name"}, {"api_name": "janggi.pytorch.NNet.NNetWrapper", "line_number": 55, "usage_type": "call"}, {"api_name": "JanggiCoach.JanggiCoach", "line_number": 64, "usage_type": "call"}, {"api_name": "JanggiMainConstants.selfPlaysPlayed", "line_number": 64, "usage_type": "attribute"}]} +{"seq_id": "163423707", "text": "import os\nimport numpy as np\nimport argparse\nimport matplotlib.pyplot as plt\nimport skimage\nfrom skimage import io, transform\nimport scipy.io as sio\nfrom scipy.signal import argrelextrema\nimport glob\nimport shutil\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\nfrom mywarper import warp, plot\n\nfrom model_trainer import ae_trainer, vae_trainer\n\nfrom appearance_ae import appearance_autoencoder\nfrom landmark_ae import landmark_autoencoder\nfrom landmark_vae import landmark_VAE\nfrom appearance_vae import appearance_VAE\n\n\n# some experiment logging setup\nALL_EXPERIMENTS_DIR = \"experiments\"\nEXP_LOSS_PLOTS_DIR = \"loss_plots\"\nEXP_METRICS_DIR = \"metrics\"\nEXP_MODELS_DIR = \"models\"\nEXP_CODE_DIR = \"code\"\n\nclass ExperimentConfig:\n def __init__(self, exp_name):\n self.exp_name = exp_name\n self.exp_dir = os.path.join(ALL_EXPERIMENTS_DIR, self.exp_name)\n self.exp_loss_plots_dir = os.path.join(self.exp_dir, EXP_LOSS_PLOTS_DIR)\n self.exp_metrics_dir = os.path.join(self.exp_dir, EXP_METRICS_DIR)\n self.exp_models_dir = os.path.join(self.exp_dir, EXP_MODELS_DIR)\n self.exp_code_dir = os.path.join(self.exp_dir, EXP_CODE_DIR)\n \n self.dirs = [self.exp_dir, self.exp_loss_plots_dir, self.exp_metrics_dir, self.exp_models_dir, self.exp_code_dir]\n for dir in self.dirs:\n if not os.path.exists(dir):\n os.makedirs(dir)\n \n self.str_repr = \"ExperimentConfig for experiment {}\\n\".format(self.exp_name)\n for dir in self.dirs:\n self.str_repr += (\"\\t\" + dir + \"\\n\")\n self.str_repr += \"\\n\"\n \n # copy all code into backup folder\n for pyfile in glob.glob('*.py'):\n shutil.copy(pyfile, self.exp_code_dir)\n \n def __str__(self):\n return self.str_repr\n \ndef setup_custom_logging(exp_name=\"\"):\n import datetime\n import sys\n \n curr_date_time = datetime.datetime.now().strftime(\"%Y_%m_%d-%H_%M_%S\")\n \n # make current experiment directory\n curr_exp_dt_name = \"{}-experiment{}\".format(curr_date_time, (\"-\"+exp_name) if (exp_name != \"\") else \"\")\n exp_config = ExperimentConfig(curr_exp_dt_name)\n \n curr_exp_log = os.path.join(exp_config.exp_dir, \"log.txt\")\n outfile = open(curr_exp_log, 'w')\n \n class CustomLogging:\n def __init__(self, orig_stream):\n self.orig_stream = orig_stream\n self.fileout = outfile\n def write(self, data):\n self.orig_stream.write(data)\n self.orig_stream.flush()\n self.fileout.write(data)\n self.fileout.flush()\n def flush(self):\n self.orig_stream.flush()\n self.fileout.flush()\n \n sys.stdout = CustomLogging(sys.stdout)\n \n return exp_config\n\ndef get_args(print_args=False):\n parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n \n parser.add_argument('--seed', type=int, default=12345,\n help=\"random seed\")\n parser.add_argument('--use_cuda', action='store_true', default=False,\n help=\"attempts to enable cuda training, if cuda available\")\n parser.add_argument('--device', type=int, default=0,\n help=\"Device to use for cuda, only applicable if cuda is available and --use_cuda is set.\")\n \n parser.add_argument('--image_dir', type=str, default='images',\n help=\"location of eigenface images\")\n parser.add_argument('--landmark_dir', type=str, default='landmarks',\n help=\"location of eigenface landmarks\")\n # parser.add_argument('--cache_dir', type=str, default='cache')\n \n parser.add_argument('--appear_lr', type=float, default=7e-4,\n help=\"learning rate of appearance model.\")\n parser.add_argument('--landmark_lr', type=float, default=1e-4,\n help=\"learning rate of landmark model.\")\n parser.add_argument('--appear_latent_dim', type=int, default=50,\n help=\"number of elements in the latent vector for the appearance model\")\n parser.add_argument('--landmark_latent_dim', type=int, default=10,\n help=\"number of elements in the latent vector for the landmark model\")\n \n parser.add_argument('--epochs', type=int, default=70,\n help=\"number of epochs to train both models\")\n parser.add_argument('--batch_size', type=int, default=32,\n help=\"batch size to use in training of both models\")\n \n required_group = parser.add_argument_group('required arguments:')\n required_group.add_argument('--model', type=str, required=True, choices=('ae', 'vae'),\n help=\"type of model to train, choose from 'ae' or 'vae'\")\n required_group.add_argument('--faces', type=str, required=True, choices=('aligned', 'unaligned'),\n help=\"type of faces data to train on, choose from 'aligned' or 'unaligned'\")\n \n args = parser.parse_args()\n \n if args.use_cuda and not torch.cuda.is_available():\n args.use_cuda = False\n print(\"args.use_cuda set to False because cuda is not available\")\n \n if print_args:\n print(\"Arguments:\")\n print(args)\n print('\\n')\n \n return args\n\n\n# Read Dataset\nclass data_reader(object):\n def __init__(self, root_dir, file_str_len, origin_name, file_format):\n self.root_dir = root_dir\n self.file_str_len = file_str_len\n self.origin_name = origin_name\n self.file_format = file_format\n\n def read(self, split, read_type):\n files_len = len([name for name in os.listdir(self.root_dir) \n if os.path.isfile(os.path.join(self.root_dir, name))])\n counter = 0\n idx = counter\n dataset = []\n train_dataset = []\n test_dataset = []\n while counter < files_len:\n name = self.origin_name + str(idx)\n if len(name) > self.file_str_len:\n name = name[len(name)-self.file_str_len:]\n try:\n if read_type == 'image':\n # data = io.imread(self.root_dir + name + self.file_format)\n data = io.imread(os.path.join(self.root_dir, name + self.file_format))\n elif read_type == 'landmark':\n # mat_data = sio.loadmat(self.root_dir + name + self.file_format)\n mat_data = sio.loadmat(os.path.join(self.root_dir, name + self.file_format))\n\n data = mat_data['lms']\n dataset.append(data)\n counter += 1\n except FileNotFoundError:\n pass\n idx += 1\n train_dataset = dataset[:split]\n test_dataset = dataset[split:]\n return train_dataset, test_dataset\n\n# Construct Dataset\nclass ImgToTensor(object):\n def __call__(self, sample):\n sample = sample.transpose((2, 0, 1))\n return torch.tensor(sample, dtype=torch.float32)/255\n\nclass LandmarkToTensor(object):\n def __call__(self, sample):\n sample = sample.reshape(-1)\n return torch.tensor(sample, dtype=torch.float32)/128\n\nclass dataset_constructor(Dataset):\n def __init__(self, dataset, transform=None):\n self.dataset = dataset\n self.transform = transform\n\n def __len__(self):\n return len(self.dataset)\n\n def __getitem__(self, idx):\n sample_data = self.dataset[idx]\n if self.transform:\n sample_data = self.transform(sample_data)\n return sample_data\n\n\n\ndef train_ae_appearance_model(exp_config, latent_dim, learning_rate, num_epochs, batch_size, cuda_avail, loss_function, face_images_train_warped):\n face_train_split = face_images_train_warped[:-100]\n face_val_split = face_images_train_warped[-100:]\n # face_trainset = dataset_constructor(face_images_train_warped, transform=transforms.Compose([ImgToTensor()]))\n face_trainset = dataset_constructor(face_train_split, transform=transforms.Compose([ImgToTensor()]))\n face_valset = dataset_constructor(face_val_split, transform=transforms.Compose([ImgToTensor()]))\n\n face_trainloader = torch.utils.data.DataLoader(face_trainset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n \n face_valloader = torch.utils.data.DataLoader(face_valset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n\n # app_model = appearance_autoencoder(latent_dim_size=50)\n app_model = appearance_autoencoder(latent_dim_size=latent_dim)\n optimizer = optim.Adam(app_model.parameters(), lr=learning_rate)\n trainer = ae_trainer(optimizer=optimizer,\n use_cuda=cuda_avail,\n model=app_model, \n loss_func=loss_function, \n model_name=\"Appearance-AE\", exp_config=exp_config)\n \n trainer.train_model(num_epochs, face_trainloader, face_valloader)\n\ndef train_ae_landmark_model(exp_config, latent_dim, learning_rate, num_epochs, batch_size, cuda_avail, loss_function, landmark_train):\n landmark_train_split = landmark_train[:-100]\n landmark_val_split = landmark_train[-100:]\n # landmark_trainset = dataset_constructor(landmark_train, transform=transforms.Compose([LandmarkToTensor()]))\n landmark_trainset = dataset_constructor(landmark_train_split, transform=transforms.Compose([LandmarkToTensor()]))\n landmark_valset = dataset_constructor(landmark_val_split, transform=transforms.Compose([LandmarkToTensor()]))\n\n landmark_trainloader = torch.utils.data.DataLoader(landmark_trainset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n \n landmark_valloader = torch.utils.data.DataLoader(landmark_valset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n\n # lm_model = landmark_autoencoder(latent_dim_size=10)\n lm_model = landmark_autoencoder(latent_dim_size=latent_dim)\n optimizer = optim.Adam(lm_model.parameters(), lr=learning_rate)\n trainer = ae_trainer(optimizer=optimizer,\n use_cuda=cuda_avail,\n model=lm_model, \n loss_func=loss_function, \n model_name=\"Landmark-AE\", exp_config=exp_config)\n\n trainer.train_model(num_epochs, landmark_trainloader, landmark_valloader)\n\ndef train_vae_appearance_model(exp_config, latent_dim, learning_rate, num_epochs, batch_size, cuda_avail, loss_function, face_images_train_warped):\n face_train_split = face_images_train_warped[:-100]\n face_val_split = face_images_train_warped[-100:]\n # face_trainset = dataset_constructor(face_images_train_warped, transform=transforms.Compose([ImgToTensor()]))\n face_trainset = dataset_constructor(face_train_split, transform=transforms.Compose([ImgToTensor()]))\n face_valset = dataset_constructor(face_val_split, transform=transforms.Compose([ImgToTensor()]))\n\n face_trainloader = torch.utils.data.DataLoader(face_trainset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n \n face_valloader = torch.utils.data.DataLoader(face_valset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n\n # app_model = appearance_VAE(latent_dim_size=50, use_cuda=cuda_avail)\n app_model = appearance_VAE(latent_dim_size=latent_dim, use_cuda=cuda_avail)\n optimizer = optim.Adam(app_model.parameters(), lr=learning_rate)\n trainer = vae_trainer(optimizer=optimizer,\n use_cuda=cuda_avail,\n model=app_model, \n recon_loss_func=loss_function,\n model_name=\"Appearance-VAE\", exp_config=exp_config)\n \n trainer.train_model(num_epochs, face_trainloader, face_valloader)\n\ndef train_vae_landmark_model(exp_config, latent_dim, learning_rate, num_epochs, batch_size, cuda_avail, loss_function, landmark_train):\n landmark_train_split = landmark_train[:-100]\n landmark_val_split = landmark_train[-100:]\n # landmark_trainset = dataset_constructor(landmark_train, transform=transforms.Compose([LandmarkToTensor()]))\n landmark_trainset = dataset_constructor(landmark_train_split, transform=transforms.Compose([LandmarkToTensor()]))\n landmark_valset = dataset_constructor(landmark_val_split, transform=transforms.Compose([LandmarkToTensor()]))\n\n landmark_trainloader = torch.utils.data.DataLoader(landmark_trainset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n \n landmark_valloader = torch.utils.data.DataLoader(landmark_valset, \n batch_size=batch_size, \n shuffle=False, \n num_workers=2)\n\n # lm_model = landmark_VAE(latent_dim_size=10, use_cuda=cuda_avail)\n lm_model = landmark_VAE(latent_dim_size=latent_dim, use_cuda=cuda_avail)\n optimizer = optim.Adam(lm_model.parameters(), lr=learning_rate)\n trainer = vae_trainer(optimizer=optimizer,\n use_cuda=cuda_avail,\n model=lm_model, \n recon_loss_func=loss_function,\n model_name=\"Landmark-VAE\", exp_config=exp_config)\n\n trainer.train_model(num_epochs, landmark_trainloader, landmark_valloader)\n\n\nif __name__ == '__main__':\n args = get_args(print_args=True)\n \n exp_config = setup_custom_logging()\n print(\"args\\n{}\\n\".format(args))\n print(\"ExperimentConfig\\n{}\\n\".format(exp_config)) \n \n if args.use_cuda:\n torch.cuda.set_device(args.device)\n print('Setting torch.cuda.set_device({})'.format(args.device))\n torch.cuda.manual_seed(args.seed)\n print('Setting torch.cuda.manual_seed({})\\n'.format(args.seed))\n \n # set the model type and loss functions\n app_loss_func, landmark_loss_func = None, None\n app_train_func, landmark_train_func = None, None\n if args.model == 'ae':\n app_loss_func, landmark_loss_func = nn.MSELoss(), nn.MSELoss()\n app_train_func = train_ae_appearance_model\n landmark_train_func = train_ae_landmark_model\n elif args.model == 'vae':\n app_loss_func, landmark_loss_func = nn.BCELoss(reduction='sum'), nn.BCELoss(reduction='sum')\n app_train_func = train_vae_appearance_model\n landmark_train_func = train_vae_landmark_model\n \n # choose appropriate data to read from for faces\n if args.faces == 'aligned':\n faces_data_loc = 'all-warped-images.npy'\n elif args.faces == 'unaligned':\n faces_data_loc = 'all-raw-images.npy'\n \n # read the appropriate data, make train/test split\n all_face_images_warped = np.load(faces_data_loc)\n face_images_train_warped = all_face_images_warped[:-100]\n face_images_test_warped = all_face_images_warped[-100:]\n print(\"Read cached images from {}\".format(faces_data_loc))\n \n # always train the appearance model for both aligned and unaligned face options\n app_train_func(exp_config, args.appear_latent_dim, args.appear_lr, args.epochs, args.batch_size, args.use_cuda, app_loss_func, face_images_train_warped)\n \n # only train the landmark model if we are training the aligned version\n if args.faces == 'aligned':\n face_landmark_reader = data_reader(args.landmark_dir, 6, '000000', '.mat')\n face_landmark_train, face_landmark_test = face_landmark_reader.read(split=800, read_type='landmark')\n print(\"read landmarks\")\n\n face_landmark_train = np.asarray(face_landmark_train)\n face_landmark_test = np.asarray(face_landmark_test)\n \n landmark_train_func(exp_config, args.landmark_latent_dim, args.appear_lr, args.epochs, args.batch_size, args.use_cuda, landmark_loss_func, face_landmark_train)\n \n # face_images_reader = data_reader(args.image_dir, 6, '000000', '.jpg')\n # face_images_train, face_images_test = face_images_reader.read(split=800, read_type='image')\n # print(\"read images\")\n\n # face_images_train = np.asarray(face_images_train)\n # face_images_test = np.asarray(face_images_test)\n \n \n # Train Autoencoders\n # train_ae_appearance_model(exp_config, args.appear_lr, args.epochs, args.batch_size, args.use_cuda, nn.MSELoss(), face_images_train_warped)\n # train_ae_landmark_model(exp_config, args.landmark_lr, args.epochs, args.batch_size, args.use_cuda, nn.MSELoss(), face_landmark_train)\n\n # Train Variational Autoencoders\n # loss_func = nn.BCELoss()\n # loss_func.size_average = False\n # train_vae_appearance_model(exp_config, args.appear_lr, args.epochs, args.batch_size, args.use_cuda, loss_func, face_images_train_warped)\n # train_vae_landmark_model(exp_config, args.landmark_lr, args.epochs, args.batch_size, args.use_cuda, loss_func, face_landmark_train)\n\n", "sub_path": "experiments/2019_03_10-12_16_31-experiment/code/eigenface_train.py", "file_name": "eigenface_train.py", "file_ext": "py", "file_size_in_byte": 17832, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 46, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 54, "usage_type": "call"}, {"api_name": "shutil.copy", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 64, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path", "line_number": 70, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 86, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 91, "usage_type": "call"}, {"api_name": "argparse.ArgumentDefaultsHelpFormatter", "line_number": 91, "usage_type": "attribute"}, {"api_name": "torch.cuda.is_available", "line_number": 128, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 128, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 150, "usage_type": "call"}, {"api_name": "os.path", "line_number": 150, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 150, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 163, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 163, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 163, "usage_type": "call"}, {"api_name": "os.path", "line_number": 163, "usage_type": "attribute"}, {"api_name": "scipy.io.loadmat", "line_number": 166, "usage_type": "call"}, {"api_name": "scipy.io", "line_number": 166, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 182, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 187, "usage_type": "attribute"}, {"api_name": "torch.utils.data.Dataset", "line_number": 189, "usage_type": "name"}, {"api_name": "skimage.transform", "line_number": 192, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 209, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 209, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 210, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 210, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 212, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 217, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 217, "usage_type": "attribute"}, {"api_name": "appearance_ae.appearance_autoencoder", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 224, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 224, "usage_type": "name"}, {"api_name": "model_trainer.ae_trainer", "line_number": 225, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 237, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 237, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 238, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 238, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 240, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 240, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 245, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 245, "usage_type": "attribute"}, {"api_name": "landmark_ae.landmark_autoencoder", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 252, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 252, "usage_type": "name"}, {"api_name": "model_trainer.ae_trainer", "line_number": 253, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 265, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 265, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 266, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 266, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 268, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 268, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 273, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 273, "usage_type": "attribute"}, {"api_name": "appearance_vae.appearance_VAE", "line_number": 279, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 280, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 280, "usage_type": "name"}, {"api_name": "model_trainer.vae_trainer", "line_number": 281, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 293, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 293, "usage_type": "name"}, {"api_name": "torchvision.transforms.Compose", "line_number": 294, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 294, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 296, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 296, "usage_type": "attribute"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 301, "usage_type": "attribute"}, {"api_name": "landmark_vae.landmark_VAE", "line_number": 307, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 308, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 308, "usage_type": "name"}, {"api_name": "model_trainer.vae_trainer", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.cuda.set_device", "line_number": 326, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 326, "usage_type": "attribute"}, {"api_name": "torch.cuda.manual_seed", "line_number": 328, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 328, "usage_type": "attribute"}, {"api_name": "torch.nn.MSELoss", "line_number": 335, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 335, "usage_type": "name"}, {"api_name": "torch.nn.BCELoss", "line_number": 339, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 339, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 365, "usage_type": "call"}]} +{"seq_id": "42369945", "text": "#!/usr/bin/env python\n# coding: utf-8\n\n# # CSC-321: Data Mining and Machine Learning\n# # Xavier Quinn\n# ## Assignment 6: Classification with a neuron\n#\n# ### Part 1: Perceptron classification\n#\n# The perceptron, as we saw in class, is the simpliest form of neural network consisting of a single neuron. Because it's so simple, it can only be used for two-class classification problems.\n#\n# The perceptron is inspired by a single neural cell, called a neuron. This accepts input signals via dendrites. Similarly, the perceptron receives inputs from examples of training data, that we weight and combine in a linear equation, called the activation function.\n#\n# activation = bias + sum(weight(i) * xi)\n#\n# You should notice the similarity between this, and the linear regression and logistic regression that we've implemented so far.\n#\n# Once we've computed the activation, we then transform it into the output value, using a transfer function (such as the step transfer function below)\n#\n# prediction = 1.0 IF activation >= 0.0, ELSE 0.0\n#\n# In order for this mechanism to work, we have to estimate the weights given in the activation function. Fortunately, we know how to do that using stochastic gradient descent.\n#\n# Each epoch, the weights are updated using the equation:\n#\n# w = w + learning_rate * (expected - predicted) * x\n#\n# Where you know that (expected - predicted) is the measure of error.\n#\n# This is enough information for you to implement the following (which will be closely related to previous assignments):\n#\n# - a predict function\n# - that takes a single instance, and a list of weights, where weights[0] is the bias\n# - a stochastic gradient descent function\n# - that takes training data, learning rate and a number of epochs\n# - where the weights are first assigned zero scores, and then iteratively updated based on the formula\n# - w(i) = w(i) + learning_rate * (expected - predicted) * x(i)\n# - where you also update the bias based on the formula:\n# - bias = bias + learning_rate * (expected - predicted)\n# - a perceptron function\n# - that takes training set, test set, learning rate and epochs\n# - that learns the weights using SGD\n# - then makes predictions over the test set using these weights\n# - and returns these predictions as a list\n#\n# I've given you a contrived data set for both your predict function, and for testing your SGD function. I've included sample output below.\n#\n# Then I want you to apply your classifier to the included sonar dataset, using the parameters given, as well as running a reasonable baseline comparison algorithm. You should perform a 3 fold cross validation. You can find out more about this data set here: https://archive.ics.uci.edu/ml/datasets/Connectionist+Bench+(Sonar,+Mines+vs.+Rocks)\n#\n# The extra twist here is that the data in the sonar data set should be converted to floats EXCEPT for the class (in the last position in each instance), that we should convert to an integer that represents...what? Currently, the class is a nominal category, and we should convert it to an integer: 1 for one class and 0 for the other. Also we will not normalize this data. Why not?\n\n# In[ ]:\n\n\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# ## Author: Xavier Theo Quinn\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n\n#Imports\nimport csv\nimport matplotlib.pyplot as plt\nimport math\nimport copy\nimport random\nimport statistics as stat\nimport ast\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# ##Methods:\n# #General\n#Format print\nVERBOSE=False\n\ndef fprint(label, value) :\n\ttry: VERBOSE\n\texcept NameError: print(\"{0}: {1}\\n\".format(label, value))\n\telse:\n\t\tif VERBOSE :\n\t\t\tprint(\"{0}: {1}\\n\".format(label, value))\n\n\n#Quick Print\ndef qPrint(*args) :\n\tvar=args[len(args)-1]\n\tfor i in range(len(args)-1) :\n\t\tprint(args[i], end =\"\")\n\n\tprint(\"{0}: {1}\".format(var, repr(eval(var)))) #yeah yeah, insecure methods, but I doubt anyone will be tryig to inject code into my ML hw.\n\n\n# #CSV functions\n\n#cleans CSV\ndef column2Float(dataset,column) :\n\tdataset[column]=[float(data.strip()) for data in dataset[column]]\n\n#loads Data\ndef load_data(filename) :\n\twith open(filename, 'r') as f:\n\t\tdata=csv.reader(f)\n\t\trowList=[]\n\t\tfor row in data:\n\t\t\tcolList=[]\n\t\t\tfor column in row :\n\t\t\t\ttry :\n\t\t\t\t\tcolList.append(column)\n\t\t\t\texcept :\n\t\t\t\t\tprint(\"csv error\")\n\t\t\trowList.append(colList)\n\t\treturn rowList\n\n#imports and cleans specified CSV\ndef importAndCleanCSV(filename) :\n\tloadedData=load_data(filename)\n\n\t# Apply to the loaded Swedish auto data here\n\tfor i in range(len(loadedData)) :\n\t\tcolumn2Float(loadedData,i)\n\n\tprint(\"{0}:\\ninstances: {1} \\nfeaturs: {2}\".format(filename, len(loadedData), len(loadedData[0])))\n\treturn loadedData\n\ndef cross_validation_data(dataset, folds) :\n\t# fprint(\"dataset\", dataset)\n\t# fprint(\"folds\", folds)\n\tfoldSize=(int)(len(dataset)/folds)\n\ttoFold=copy.deepcopy(dataset)\n\tfolded=[]\n\tfor foldIndex in range(folds) :\n\t\tsublist=[]\n\t\tfor instIndex in range(foldSize) :\n\t\t\t#generate nested list\n\t\t\tsublist.append(toFold.pop(random.randint(0,len(toFold)-1)))\n\t\t#put sublist into folded list\n\t\tfolded.append(sublist)\n\n\treturn folded\n\n\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# # Evualuation Assistant functions\n#normalizes a dataset\ndef normalize(dataset, minmax) :\n\tif (len(dataset[0])!=len(minmax)) :\n\t\tprint(\"lengths do not match\")\n\t\tprint(\"{0} !={1}\".format(len(dataset[0]),len(minmax)))\n\t\texit()\n\n\tfor data in dataset :\n\t\tfor i in range(0,len(minmax)) :\n\t\t\t#fprint(\"i:{0}, data\".format(i),data[i])\n\t\t\tdata[i]= (data[i] - minmax[i][0] )/( minmax[i][1] - minmax[i][0])\n\n#returns the min and max\ndef minmax(dataset) :\n\tcols=dataset[0]\n\tcols=[[val,val] for val in cols]\n\tfor data in dataset :\n\t\tfor i in range(0,len(cols)) :\n\t\t\tcurrent=data[i]\n\t\t\tif currentcols[i][1] :\n\t\t\t\tcols[i][1]=current\n\treturn cols\n\n\n#rmse_eval\ndef rmse_eval(actual,predicted) :\n\t# fprint(\"actual list\", actual)\n\t# fprint(\"predicted list\", predicted)\n\tresult=0\n\tfor i in range(len(actual)) :\n\t\t# fprint(\"RMSE: actual\", actual[i])\n\t\t# fprint(\"RMSE: predicted\", predicted[i])\n\t\tresult=result+(predicted[i]-actual[i])**2\n\n\taverage=result/len(actual)\n\treturn average\n\n\ndef accuracy(actual, predicted) :\n\tcounter=0\n\tfprint(\"acc actual\", actual)\n\tfprint(\"acc pred \", predicted)\n\tfor i in range(0, len(actual)) :\n\t\tact=actual[i]\n\t\tpred=predicted[i]\n\t\t# fprint(\"pred \", pred)\n\t\t# fprint(\"pred \", round(pred))\n\t\tif (act==pred) :\n\t\t\tcounter=counter+1;\n\t\t# fprint(\"count\", counter)\n\tscore=100.0*(counter/len(actual))\n\tfprint(\"score\", counter)\n\treturn score\n\n\ndef zeroRC(train, test) :\n\tyList=[inst[-1:][0] for inst in train]\n\ttoReturn=[max(set(yList), key=yList.count) for i in test]\n\tfprint(\"toReturn\", toReturn)\n\treturn toReturn\n#~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n# Implement or copy your code here\n\n# - a predict function\n# - that takes a single instance, and a list of weights, where weights[0] is the bias\ndef perceptronPredict(instance, weights) :\n\tactivated=weights[0] #+ bias\n\n\tfor i in range(1, len(instance)) :\n\t\t# fprint(\"Predict: Instance\", instance[i-1])\n\t\t# fprint(\"Predict: weight \", weights[i])\n\t\tmult=weights[i]*instance[i-1]\n\t\t# fprint(\"Predict: mult \", mult)\n\t\tactivated=activated+mult;\n\n\tpredicted = 1.0 if activated >= 0.0 else 0.0\n\treturn predicted\n\n\n# - a stochastic gradient descent function\n# - that takes training data, learning rate and a number of epochs\n# - where the weights are first assigned zero scores, and then iteratively updated based on the formula\n# - w(i) = w(i) + learning_rate * (expected - predicted) * x(i)\n# - where you also update the bias based on the formula:\n# - bias = bias + learning_rate * (expected - predicted)\n# returns coeffs\ndef SGD(train, learningRate, epochs) :\n\tweights=[0.0 for i in range(len(train[0]))]\n\n\tfor i in range(0, epochs) :\n\t\ttotError=0;\n\n\t\tfor n in range(0, len(train)) :\n\t\t\tinstance=train[n]\n\t\t\texpected=instance[-1:][0]\n\t\t\tpredicted=perceptronPredict(instance,weights)\n\t\t\t# fprint(\"instance\",instance)\n\t\t\t# fprint(\"expected\",expected)\n\t\t\t# fprint(\"predicted\",predicted)\n\n\n\t\t\tinstError=expected-predicted\n\t\t\ttotError+= instError**2\n\t\t\t# fprint(\"error\",error)\n\n\n\n\t\t\tweights[0]=weights[0] + learningRate * instError #bias update\n\n\t\t\tfor j in range(1,len(instance)) :\n\t\t\t\tweights[j]=weights[j] + learningRate * instError * instance[j-1]\n\n\t\tfprint(\"Epoch\",i)\n\t\tfprint(\"Learning rate\", learningRate)\n\t\tfprint(\"Total Error\",totError)\n\t\t# fprint(\"coefficents\", weights)\n\treturn weights\n\n\n\n\n# - a perceptron function\n# - that takes training set, test set, learning rate and epochs\n# - that learns the weights using SGD\n# - then makes predictions over the test set using these weights\n# - and returns just these predictions as a list\ndef perceptron(train, test, learningRate, epochs) :\n\tweights=SGD(train, learningRate, epochs)\n\tpredictions=[]\n\tfor testInst in test :\n\t\tpredictions.append(perceptronPredict(testInst,weights))\n\treturn predictions\n\n\n#evaluate algorithm method\ndef evaluate_algorithm(dataset, algorithm, folds, metric, *args):\n\n\n\tscores = []\n\tfolded= cross_validation_data(dataset, folds)\n\n\tfor i in range(len(folded)) :\n\t\ttrainData=copy.deepcopy(folded) #these two lines seperate the fold at the index from the rest of the folds\n\t\ttest=trainData.pop(i)\n\n\t\t#the Flattening\n\t\ttmpData=[]\n\t\tfor fold in trainData :\n\t\t\tfor instance in fold :\n\t\t\t\ttmpData.append(instance)\n\t\ttrainData=copy.deepcopy(tmpData) #I can't think of a reason for why this should be a deep copy, but might as well\n\t\tactual=[]\n\t\t#Prepares test Set\n\n\t\tfor inst in test :\n\t\t\tactual.append(inst[-1:][0])\n\t\t\tinst[len(inst)-1]=None\n\n\t\tfprint(\"Eval_Alg: actual\", actual)\n\t\tfprint(\"Eval_Alg: *args\", args)\n\t\tpredicted = algorithm(trainData,test, *args)\n\t\tfprint(\"Eval_Alg: predicted\", predicted)\n\t\t# fprint(\"Eval_Alg: trainData\", trainData)\n\n\t\tresult = metric(actual,predicted)\n\t\t# fprint(\"evaluation result\", result)\n\t\tscores.append(result)\n\n\treturn scores\n\n\n# Contrived data\n# Predict should work, given the weights below\n\n\ndataset = [[2.7810836,2.550537003,0],\n\t[1.465489372,2.362125076,0],\n\t[3.396561688,4.400293529,0],\n\t[1.38807019,1.850220317,0],\n\t[3.06407232,3.005305973,0],\n\t[7.627531214,2.759262235,1],\n\t[5.332441248,2.088626775,1],\n\t[6.922596716,1.77106367,1],\n\t[8.675418651,-0.242068655,1],\n\t[7.673756466,3.508563011,1]]\n\nweights = [-0.1, 0.20653640140000007, -0.23418117710000003]\n\n\nfor row in dataset:\n\tprediction = perceptronPredict(row, weights)\n\tprint(\"Expected={0}, Predicted={1}\".format(row[-1], prediction))\n\n# fprint(\"predict test\", perceptronPredict(dataset[0], weights))\n\n# Using your SGD function with a learning rate of 0.1, and 5 epochs, should give you:\n#\n#>epoch=0, lrate=0.100, error=2.000\n#>epoch=1, lrate=0.100, error=1.000\n#>epoch=2, lrate=0.100, error=0.000\n#>epoch=3, lrate=0.100, error=0.000\n#>epoch=4, lrate=0.100, error=0.000\n#\n#[-0.1, 0.20653640140000007, -0.23418117710000003]\n\nSGD(dataset, .1, 5)\n\n\n# Parameters for learning over real data\n\nfilename = 'sonar.all-data.csv'\n\nloadedData=load_data(filename)\n\n\n#I'm sure theres a better way to do this for this dataset, but this works.\nloadedData=str(loadedData) #cast to string\nloadedData=loadedData.replace('R', '0') #replace using strings built in method\nloadedData=loadedData.replace('M', '1')\nloadedData=ast.literal_eval(loadedData) #convert back to nested list\n\n\nfor i in range(len(loadedData)) :\n\tcolumn2Float(loadedData,i)\ndataset=loadedData\nprint(\"{0}:\\ninstances: {1} \\nfeaturs: {2}\".format(filename, len(loadedData), len(loadedData[0])))\n\n\n# dataset=importAndCleanCSV(filename)\n\n\nfolds = 3\nlearning_rate = 0.01\nepochs = 500\n\n\n\nprint(\"SCORE\")\n\n# VERBOSE=True\nperceptronScore = evaluate_algorithm(dataset, perceptron, folds, rmse_eval, learning_rate, epochs)\nzeroRCScore = evaluate_algorithm(dataset, zeroRC, folds, rmse_eval)\n\n\n\n\n\nVERBOSE=True\nfprint('SGD_LOG', perceptronScore)\nfprint(' best', min(perceptronScore))\nfprint(' worst', max(perceptronScore))\nfprint(' mean', stat.mean(perceptronScore))\n\nfprint('zeroRC RMSE', zeroRCScore)\nfprint(' best', min(zeroRCScore))\nfprint(' worst', max(zeroRCScore))\nfprint(' mean', stat.mean(zeroRCScore))\n\n\nprint(\"As we would have guessed, the perceptron greatly outpreforms ZERORC. I was briefly wondering about if I should use ZeroRC or ZeroRR, but I beleive it would not cause notable difference as the values are only 1 or 0\")\n", "sub_path": "csc321/CSC-321 Assignment 6.py", "file_name": "CSC-321 Assignment 6.py", "file_ext": "py", "file_size_in_byte": 12494, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "csv.reader", "line_number": 100, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 127, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 133, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 287, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 295, "usage_type": "call"}, {"api_name": "ast.literal_eval", "line_number": 364, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 396, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 401, "usage_type": "call"}]} +{"seq_id": "258451592", "text": "import sys \nfrom pykafka import KafkaClient\nfrom pykafka.simpleconsumer import OffsetType \n \nreload(sys) \nsys.setdefaultencoding('utf8') \n\nclass PyKafka: \n consumer = None \n TOPIC = 'log_download' \n BROKER_LIST = '1.1.1.111:9092,1.1.1.11:9092' \n ZK_LIST = '1.1.1.1:2181,1.1.1.2:2181/sh-bt' \n \n server = topic = zsServer = None\n \n def __init__(self): \n self.server = self.BROKER_LIST \n self.topic = self.TOPIC \n self.zkServer= self.ZK_LIST \n \n def getConnect(self): \n client = KafkaClient(hosts=self.server) \n topic = client.topics[self.topic] \n \n self.consumer = topic.get_balanced_consumer( \n consumer_group=\"zs_download_04\",\n auto_offset_reset=OffsetType.LATEST, # 在consumer_group存在的情况下,设置此变量,表示从最新的开始取 \n # auto_offset_reset=OffsetType.EARLIEST, \n # reset_offset_on_start=True, \n # auto_commit_enable=True, \n zookeeper_connect=self.zkServer \n ) \n #self.consumer = topic.get_simple_consumer(reset_offset_on_start=False) \n self.consumer.consume() \n self.consumer.commit_offsets() \n return self.consumer \n \n def disConnect(self): \n #self.consumer.close() \n pass \n \n \n def beginConsumer(self): \n for oneLog in self.consumer: \n print(oneLog.offset) \n print(oneLog.value) \n \n \nif __name__ == '__main__': \n \n pk = PyKafka() \n pk.getConnect() \n pk.beginConsumer() \n", "sub_path": "2b/db_dao/kafka_dao.py", "file_name": "kafka_dao.py", "file_ext": "py", "file_size_in_byte": 1593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 6, "usage_type": "call"}, {"api_name": "pykafka.KafkaClient", "line_number": 22, "usage_type": "call"}, {"api_name": "pykafka.simpleconsumer.OffsetType.LATEST", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pykafka.simpleconsumer.OffsetType", "line_number": 27, "usage_type": "name"}]} +{"seq_id": "76080590", "text": "import insightconnect_plugin_runtime\nfrom .schema import AntivirusScanInput, AntivirusScanOutput, Input, Output, Component\n# Custom imports below\nfrom insightconnect_plugin_runtime.exceptions import PluginException\n\n\nclass AntivirusScan(insightconnect_plugin_runtime.Action):\n\n def __init__(self):\n super(self.__class__, self).__init__(\n name='antivirus_scan',\n description=Component.DESCRIPTION,\n input=AntivirusScanInput(),\n output=AntivirusScanOutput())\n\n def run(self, params={}):\n uuid = None\n page_key = None\n agent = params.get(Input.AGENT)\n\n for index in range(9999):\n endpoints = self.connection.client.get_endpoints(page_key=page_key)\n page_key = endpoints.get(\"pages\", {}).get(\"nextKey\", None)\n\n for e in endpoints.get(\"items\", []):\n if e.get(\"hostname\") == agent:\n uuid = e.get(\"id\")\n elif e.get(\"id\") == agent:\n uuid = e.get(\"id\")\n elif agent in e.get(\"ipv4Addresses\"):\n uuid = e.get(\"id\")\n elif agent in e.get(\"macAddresses\"):\n uuid = e.get(\"id\")\n elif agent in e.get(\"ipv6Addresses\"):\n uuid = e.get(\"id\")\n\n if page_key is None or index > endpoints.get(\"pages\", {}).get(\"total\", 0):\n break\n\n if uuid is None:\n raise PluginException(preset=PluginException.Preset.NOT_FOUND)\n\n antivirus_scan = self.connection.client.antivirus_scan(uuid)\n\n return {\n Output.ID: antivirus_scan.get(\"id\"),\n Output.STATUS: antivirus_scan.get(\"status\"),\n Output.REQUESTED_AT: antivirus_scan.get(\"requestedAt\")\n }\n", "sub_path": "sophos_central/icon_sophos_central/actions/antivirus_scan/action.py", "file_name": "action.py", "file_ext": "py", "file_size_in_byte": 1806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "insightconnect_plugin_runtime.Action", "line_number": 7, "usage_type": "attribute"}, {"api_name": "schema.Component.DESCRIPTION", "line_number": 12, "usage_type": "attribute"}, {"api_name": "schema.Component", "line_number": 12, "usage_type": "name"}, {"api_name": "schema.AntivirusScanInput", "line_number": 13, "usage_type": "call"}, {"api_name": "schema.AntivirusScanOutput", "line_number": 14, "usage_type": "call"}, {"api_name": "schema.Input.AGENT", "line_number": 19, "usage_type": "attribute"}, {"api_name": "schema.Input", "line_number": 19, "usage_type": "name"}, {"api_name": "insightconnect_plugin_runtime.exceptions.PluginException", "line_number": 41, "usage_type": "call"}, {"api_name": "insightconnect_plugin_runtime.exceptions.PluginException.Preset", "line_number": 41, "usage_type": "attribute"}, {"api_name": "schema.Output.ID", "line_number": 46, "usage_type": "attribute"}, {"api_name": "schema.Output", "line_number": 46, "usage_type": "name"}, {"api_name": "schema.Output.STATUS", "line_number": 47, "usage_type": "attribute"}, {"api_name": "schema.Output", "line_number": 47, "usage_type": "name"}, {"api_name": "schema.Output.REQUESTED_AT", "line_number": 48, "usage_type": "attribute"}, {"api_name": "schema.Output", "line_number": 48, "usage_type": "name"}]} +{"seq_id": "590742794", "text": "# Import the libaries\nimport pathlib\nimport socket\nimport termcolor\n\n# Server Information\nIP = \"127.0.0.1\"\nPORT = 8080\n\n\ndef read(FILENAME): # read() is the function read_fasta_data() from other practice\n # open and read the file\n file_contents = pathlib.Path(FILENAME).read_text().split(\"\\n\")[1:] # Split lines and skip the first one [1:]\n body = \"\".join(file_contents) # Join all the list in the same string without spaces\n return body\n\n\ndef process_client(s):\n # Receive the message\n req_raw = s.recv(2000)\n request_received = req_raw.decode()\n\n print(\"Message from CLIENT: \")\n\n lines = request_received.split('\\n') # Split the request\n first_line = lines[0] # first LINE of the request\n\n print(\"Request line: \", end=\"\")\n termcolor.cprint(first_line, \"green\")\n\n # Response message\n FOLDER = \"../P4/\"\n\n file_request = first_line.split(\" \")[1]\n # file_request is like req_line (GET /info/A HTTP/1.1) only with (/info/A) or (/info/C)\n\n if \"/info/A\" == file_request:\n FILENAME = \"A.html\"\n body = read(FOLDER + FILENAME)\n elif \"/info/C\" == file_request:\n FILENAME = \"C.html\"\n body = read(FOLDER + FILENAME)\n elif \"/info/G\" == file_request:\n FILENAME = \"G.html\"\n body = read(FOLDER + FILENAME)\n elif \"/info/T\" == file_request:\n FILENAME = \"T.html\"\n body = read(FOLDER + FILENAME)\n else:\n body = \"\"\n\n # This new contents are written in HTML language\n # Status line\n status_line = \"HTTP/1.1 200 OK\\n\" # everything is ok (200 code)\n\n # Content-Type header and Content-Length\n header = \"Content-Type: text/html\\n\"\n header += f\"Content-Length: {len(body)}\\n\"\n\n # Join all the parts\n response = status_line + header + \"\\n\" + body # Join the status_line , header and body\n cs.send(response.encode()) # send the response\n\n\n# Server\n# Listening socket\nlsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\n# Preventing the error: \"Port already in use\"\nlsocket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n\nlsocket.bind((IP, PORT)) # Socket's IP and PORT with .bind() method\n\n# -- Become a listening socket\nlsocket.listen()\n\nprint(\"SEQ Server configured!\")\n\n# --- MAIN LOOP\nwhile True:\n print(\"Waiting for clients....\")\n try:\n (cs, client_ip_port) = lsocket.accept()\n except KeyboardInterrupt:\n print(\"Server Stopped!\")\n lsocket.close()\n exit()\n else:\n\n # Service the client\n process_client(cs) # This is the function created above\n\n # Close\n cs.close()\n", "sub_path": "P4/EX4.py", "file_name": "EX4.py", "file_ext": "py", "file_size_in_byte": 2589, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pathlib.Path", "line_number": 13, "usage_type": "call"}, {"api_name": "termcolor.cprint", "line_number": 29, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 67, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 67, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 67, "usage_type": "attribute"}, {"api_name": "socket.SOL_SOCKET", "line_number": 70, "usage_type": "attribute"}, {"api_name": "socket.SO_REUSEADDR", "line_number": 70, "usage_type": "attribute"}]} +{"seq_id": "158535331", "text": "from pptx import Presentation\n\nprs = Presentation()\nbullet_slide_layout = prs.slide_layouts[1] # 确定顺序\n\nslide = prs.slides.add_slide(bullet_slide_layout) # 增加幻灯片\nshapes = slide.shapes # shapes:对幻灯片的某一区域操作\n\ntitle_shape = shapes.title\nbody_shape = shapes.placeholders[1]\n\ntitle_shape.text = \"Adding a Bullet Slide\"\ntf = body_shape.text_frame\ntf.text = \"Find the bullet slide layout\"\n\np = tf.add_paragraph() # 增加文本\np.text = \"Use __TextFrame.text for first bullet\"\np.level = 1 # 设置级别\n\np = tf.add_paragraph()\np.text = \"Use __TextFrame.add_paragraph() for subsequent bullets\"\np.level = 2\n\nprs.save(\"pptxTry_text.pptx\")\n\n'''\nshapes.add_textbox()\nshapes.add_picture()\nshapes.add_shape()\nshapes.add_table()\n'''\n", "sub_path": "pptxTry_text.py", "file_name": "pptxTry_text.py", "file_ext": "py", "file_size_in_byte": 777, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pptx.Presentation", "line_number": 3, "usage_type": "call"}]} +{"seq_id": "7256411", "text": "from django.db import models\nfrom django.contrib.auth.models import User\nfrom django.conf import settings\nfrom django.utils.text import slugify\nfrom cities_light.models import Country, Region, City\nfrom smart_selects.db_fields import ChainedForeignKey\n\nclass SchoolType(models.Model):\n \"\"\"A model for the various types of schools.\"\"\"\n\n name = models.CharField(blank=False,\n max_length=40)\n description = models.CharField(blank=True,\n max_length=500)\n\n def __str__(self):\n return self.name\n\n\nclass StudentLevel(models.Model):\n \"\"\"A model for the various levels of students.\"\"\"\n\n name = models.CharField(blank=False,\n max_length=40)\n description = models.CharField(blank=True,\n max_length=500)\n\n def __str__(self):\n return self.name\n\n\nclass SchoolReview(models.Model):\n \"\"\"A salary report for one school.\"\"\"\n\n # here are our choices for the various fields.\n\n RATINGS = (\n ('G', 'Good'),\n ('N', 'Neutral'),\n ('B', 'Bad')\n )\n\n YES = 'Y'\n NO = 'N'\n MAYBE = 'M'\n\n RECOMMEND = (\n (YES, 'Yes'),\n (MAYBE, 'Maybe'),\n (NO, 'No')\n )\n\n YES_NO_CHOICES = (\n (YES, 'Yes'),\n (NO, 'No'),\n )\n\n CONTRACT_TYPE_CHOICES = (\n ('FT', 'Full-time'),\n ('PT', 'Part-time'),\n )\n\n HOURLY_MONTHLY_CHOICES = (\n ('H', 'Hourly'),\n ('M', 'Monthly'),\n )\n\n # now on to the DB models.\n created = models.DateTimeField(auto_now_add=True, editable=False)\n school = models.ForeignKey('School')\n author = models.ForeignKey(User)\n school_experience = models.CharField(blank=False,\n max_length=7,\n choices=RATINGS,\n help_text=\"How would you rate your experience teaching at this school?\",\n default='N')\n recommend_school = models.CharField(blank=False,\n max_length=5,\n choices=RECOMMEND,\n help_text=\"Would you recommend this school to a friend?\",\n default=MAYBE)\n start_year = models.IntegerField(blank=False,\n help_text=\"In what year did the contract start?\")\n end_year = models.IntegerField(blank=False,\n help_text=\"In what year did the contract end?\")\n contract_type = models.CharField(blank=False,\n max_length=2,\n help_text=\"Full-time (FT) or Part-Time (PT) work.\",\n choices=CONTRACT_TYPE_CHOICES,\n default='FT')\n pay = models.IntegerField(blank=True,\n default=0,\n help_text=\"How much were you paid?\")\n type_of_wage = models.CharField(help_text=\"Is the pay above hourly or monthly?\",\n choices=HOURLY_MONTHLY_CHOICES,\n default='M',\n max_length=1)\n hours_per_week = models.IntegerField(blank=False,\n default=0)\n zed_visa = models.CharField(blank=True,\n choices=YES_NO_CHOICES,\n default=NO,\n max_length=1,\n help_text=\"Did the employer provide a Z visa & Residence Permit?\")\n vacation_time = models.IntegerField(blank=True,\n default=0,\n help_text=\"Days of vacation per contract (INCLUDING Chinese holidays).\")\n housing = models.CharField(blank=True,\n choices=YES_NO_CHOICES,\n default=NO,\n max_length=1,\n help_text=\"Did the employer provide either a housing allowance or an apartment?\")\n housing_amount = models.IntegerField(default=0,\n help_text=\"If money was provided for housing, how much?\",\n blank=True,)\n other_perks = models.CharField(help_text=\"Were there any other perks with the job? (Max 500 characters.)\",\n max_length=500,\n default='None.',\n blank=True)\n your_comments = models.CharField(max_length=2000,\n blank=True,\n default='None.',\n help_text=\"Anything else to say about the job? (Max 2000 characters.)\")\n\n class Meta:\n ordering = ('-created',)\n\n\nclass School(models.Model):\n\n school_name = models.CharField(blank=False, max_length=200, unique=True)\n slug = models.SlugField()\n school_country = models.ForeignKey(Country)\n school_province = ChainedForeignKey(\n Region,\n chained_field=\"school_country\",\n chained_model_field=\"country\",\n help_text=\"The school's province.\"\n )\n school_city = ChainedForeignKey(\n City,\n chained_field='school_province',\n chained_model_field='region',\n show_all=False,\n help_text=\"The school's city.\"\n )\n school_type = models.ForeignKey(SchoolType,\n help_text='Is the school public, private, or a joint venture?')\n student_level = models.ManyToManyField(help_text='What level(s) are the students?',\n to=StudentLevel)\n\n def save(self, *args, **kwargs):\n self.slug = slugify(self.school_name)\n super(School, self).save(*args, **kwargs)\n\n def __str__(self):\n return self.school_name\n\n class Meta:\n ordering = ('school_name',)\n\n", "sub_path": "school_reviews_app/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 6121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.db.models.Model", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 11, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 11, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 32, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 69, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 69, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 70, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 70, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.auth.models.User", "line_number": 71, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 71, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 72, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 72, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 77, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 77, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 82, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 84, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 86, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 86, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 91, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 91, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 94, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 94, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 98, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 98, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 100, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 100, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 105, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 105, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 108, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 108, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 113, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 113, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 116, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 116, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 120, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 120, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 129, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 129, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 131, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 131, "usage_type": "name"}, {"api_name": "django.db.models.SlugField", "line_number": 132, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 132, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 133, "usage_type": "call"}, {"api_name": "cities_light.models.Country", "line_number": 133, "usage_type": "argument"}, {"api_name": "django.db.models", "line_number": 133, "usage_type": "name"}, {"api_name": "smart_selects.db_fields.ChainedForeignKey", "line_number": 134, "usage_type": "call"}, {"api_name": "cities_light.models.Region", "line_number": 135, "usage_type": "argument"}, {"api_name": "smart_selects.db_fields.ChainedForeignKey", "line_number": 140, "usage_type": "call"}, {"api_name": "cities_light.models.City", "line_number": 141, "usage_type": "argument"}, {"api_name": "django.db.models.ForeignKey", "line_number": 147, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 147, "usage_type": "name"}, {"api_name": "django.db.models.ManyToManyField", "line_number": 149, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 149, "usage_type": "name"}, {"api_name": "django.utils.text.slugify", "line_number": 153, "usage_type": "call"}]} +{"seq_id": "583925513", "text": "from django import forms\nfrom django.contrib.auth.models import User\nfrom django.contrib.auth import authenticate, login\nfrom django.http import HttpResponse\nfrom django.shortcuts import render\n\nfrom .models import Usuario\n\n\nclass RegistrarCadastroForm(forms.Form):\n\n\n def __init__(self, *args, **kwargs):\n super(RegistrarCadastroForm, self).__init__(*args, **kwargs)\n self.sem_erro = True\n\n def is_valid(self):\n if not super(RegistrarCadastroForm, self).is_valid():\n self.adiciona_erro('Por favor, verifique os dados informados')\n\n if Usuario.objects.filter(cpf=self.data['cpf']).exists():\n self.adiciona_erro(\"Já existe um usu��rio cadastrado com este CPF.\")\n\n if Usuario.objects.filter(email=self.data['email']).count() > 0:\n self.adiciona_erro(\"Já existe um usuário cadastrado com este e-Mail.\")\n\n if self.data['senha'] != self.data['senha2']:\n self.adiciona_erro(\"Senhas não correspondem.\")\n\n return self.sem_erro\n\n def adiciona_erro(self, message):\n errors = self._errors.setdefault(forms.forms.NON_FIELD_ERRORS, forms.utils.ErrorList())\n errors.append(message)\n self.sem_erro = False\n\n\n\nclass RegistrarLoginForm(forms.Form):\n\n def __init__(self, *args, **kwargs):\n super(RegistrarLoginForm, self).__init__(*args, **kwargs)\n self.sem_erro = True\n\n def is_valid(self):\n if not super(RegistrarLoginForm, self).is_valid():\n self.adiciona_erro('Por favor, verifique os dados informados')\n\n if Usuario.objects.filter(email=self.data['email']).count() == 0:\n self.adiciona_erro(\"Email não cadastrado.\")\n\n #verificando se usuário e senha não existem no banco\n elif not Usuario.objects.filter(email=self.data['email'], password=self.data['senha']).exists():\n self.adiciona_erro(\"E-mail ou senha não correspondem.\")\n\n return self.sem_erro\n\n\n def adiciona_erro(self, message):\n errors = self._errors.setdefault(forms.forms.NON_FIELD_ERRORS, forms.utils.ErrorList())\n errors.append(message)\n self.sem_erro = False\n\n\n", "sub_path": "cadastro/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 2153, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.forms.Form", "line_number": 10, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 10, "usage_type": "name"}, {"api_name": "models.Usuario.objects.filter", "line_number": 21, "usage_type": "call"}, {"api_name": "models.Usuario.objects", "line_number": 21, "usage_type": "attribute"}, {"api_name": "models.Usuario", "line_number": 21, "usage_type": "name"}, {"api_name": "models.Usuario.objects.filter", "line_number": 24, "usage_type": "call"}, {"api_name": "models.Usuario.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.Usuario", "line_number": 24, "usage_type": "name"}, {"api_name": "django.forms.forms", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 33, "usage_type": "name"}, {"api_name": "django.forms.utils.ErrorList", "line_number": 33, "usage_type": "call"}, {"api_name": "django.forms.utils", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.forms.Form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 39, "usage_type": "name"}, {"api_name": "models.Usuario.objects.filter", "line_number": 49, "usage_type": "call"}, {"api_name": "models.Usuario.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "models.Usuario", "line_number": 49, "usage_type": "name"}, {"api_name": "models.Usuario.objects.filter", "line_number": 53, "usage_type": "call"}, {"api_name": "models.Usuario.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "models.Usuario", "line_number": 53, "usage_type": "name"}, {"api_name": "django.forms.forms", "line_number": 60, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 60, "usage_type": "name"}, {"api_name": "django.forms.utils.ErrorList", "line_number": 60, "usage_type": "call"}, {"api_name": "django.forms.utils", "line_number": 60, "usage_type": "attribute"}]} +{"seq_id": "320733442", "text": "\"\"\"Defines the main trainer model for combinatorial problems\n\nEach task must define the following functions:\n* mask_fn: can be None\n* update_fn: can be None\n* reward_fn: specifies the quality of found solutions\n* render_fn: Specifies how to plot found solutions. Can be None\n\"\"\"\n\nimport os\nimport time\nimport argparse\nimport datetime\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.utils.data import DataLoader\n\nfrom model import DRL4TSP, Encoder\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n#device = torch.device('cpu')\n\n\nclass StateCritic(nn.Module):\n \"\"\"Estimates the problem complexity.\n\n This is a basic module that just looks at the log-probabilities predicted by \n the encoder + decoder, and returns an estimate of complexity\n \"\"\"\n\n def __init__(self, static_size, dynamic_size, hidden_size):\n super(StateCritic, self).__init__()\n\n self.static_encoder = Encoder(static_size, hidden_size)\n self.dynamic_encoder = Encoder(dynamic_size, hidden_size)\n\n # Define the encoder & decoder models\n self.fc1 = nn.Conv1d(hidden_size * 2, 20, kernel_size=1)\n self.fc2 = nn.Conv1d(20, 20, kernel_size=1)\n self.fc3 = nn.Conv1d(20, 1, kernel_size=1)\n \n for p in self.parameters():\n if len(p.shape) > 1:\n nn.init.xavier_uniform_(p)\n\n def forward(self, static, dynamic):\n\n # Use the probabilities of visiting each \n static_hidden = self.static_encoder(static)\n dynamic_hidden = self.dynamic_encoder(dynamic)\n\n hidden = torch.cat((static_hidden, dynamic_hidden), 1)\n\n output = F.relu(self.fc1(hidden))\n output = F.relu(self.fc2(output))\n output = self.fc3(output).sum(dim=2)\n return output\n\n\nclass Critic(nn.Module):\n \"\"\"Estimates the problem complexity.\n\n This is a basic module that just looks at the log-probabilities predicted by \n the encoder + decoder, and returns an estimate of complexity\n \"\"\"\n\n def __init__(self, hidden_size):\n super(Critic, self).__init__()\n\n # Define the encoder & decoder models\n self.fc1 = nn.Conv1d(1, hidden_size, kernel_size=1)\n self.fc2 = nn.Conv1d(hidden_size, 20, kernel_size=1)\n self.fc3 = nn.Conv1d(20, 1, kernel_size=1)\n\n for p in self.parameters():\n if len(p.shape) > 1:\n nn.init.xavier_uniform_(p)\n\n def forward(self, input):\n\n output = F.relu(self.fc1(input.unsqueeze(1)))\n output = F.relu(self.fc2(output)).squeeze(2)\n output = self.fc3(output).sum(dim=2)\n return output\n\n\ndef validate(data_loader, actor, reward_fn, render_fn=None, save_dir='.', \n num_plot=5):\n \"\"\"Used to monitor progress on a validation set & optionally plot solution.\"\"\"\n\n actor.eval()\n\n rewards = []\n for batch_idx, batch in enumerate(data_loader):\n\n static, dynamic, x0 = batch\n\n static = static.to(device)\n dynamic = dynamic.to(device)\n x0 = x0.to(device) if len(x0) > 0 else None\n\n # Full forward pass through the dataset\n with torch.no_grad():\n\n tour_indices, _ = actor.forward(static, dynamic, x0)\n\n reward = reward_fn(static, tour_indices).detach()\n rewards.append(torch.mean(reward)) \n\n if render_fn is not None and batch_idx < num_plot:\n name = 'batch%d_%2.4f.png'%(batch_idx, reward.mean())\n path = os.path.join(save_dir, name)\n render_fn(static, tour_indices, path)\n\n actor.train()\n return np.mean(rewards)\n\n\ndef train(actor, critic, task, num_nodes, train_data, valid_data, reward_fn,\n render_fn, batch_size, actor_lr, critic_lr, max_grad_norm,\n checkpoint_every, **kwargs):\n \"\"\"Constructs the main actor & critic networks, and performs all training.\"\"\"\n\n now = '%s' % datetime.datetime.now().time()\n now = now.replace(':', '_')\n save_dir = os.path.join(task, '%d' % num_nodes, now)\n \n checkpoint_dir = os.path.join(save_dir, 'checkpoints')\n\n if not os.path.exists(checkpoint_dir):\n os.makedirs(checkpoint_dir)\n\n if kwargs['optimizer'] == 'sgd':\n optimizer = optim.SGD\n elif kwargs['optimizer'] == 'adam':\n optimizer = optim.Adam\n elif kwargs['optimizer'] == 'adagrad':\n optimizer = optim.Adagrad\n else:\n raise ValueError('Optimizer <%s> not understood'%kwargs['optimizer'])\n\n actor_optim = optimizer(actor.parameters(), lr=actor_lr)\n critic_optim = optimizer(critic.parameters(), lr=critic_lr)\n\n train_loader = DataLoader(train_data, batch_size, True, num_workers=0)\n valid_loader = DataLoader(valid_data, batch_size, False, num_workers=0)\n\n for epoch in range(20):\n\n actor.train()\n critic.train()\n\n losses, rewards, critic_rewards = [], [], []\n for batch_idx, batch in enumerate(train_loader):\n\n static, dynamic, x0 = batch\n\n static = static.to(device).requires_grad_()\n dynamic = dynamic.to(device).requires_grad_()\n x0 = x0.to(device).requires_grad_() if len(x0) > 0 else None\n\n # Full forward pass through the dataset\n tour_indices, tour_logp = actor(static, dynamic, x0)\n\n # Sum the log probabilities for each city in the tour\n reward = reward_fn(static, tour_indices)\n\n # Query the critic for an estimate of the reward\n #critic_in = torch.tensor(tour_logp.data, device=device, requires_grad=True)\n #critic_est = critic(critic_in).view(-1)\n critic_est = critic(static, dynamic).view(-1)\n\n advantage = (reward - critic_est)\n actor_loss = torch.mean(advantage.detach() * tour_logp.sum(dim=1))\n critic_loss = torch.mean(torch.pow(advantage, 2))\n\n actor_optim.zero_grad()\n actor_loss.backward()\n torch.nn.utils.clip_grad_norm_(actor.parameters(), max_grad_norm)\n actor_optim.step()\n\n critic_optim.zero_grad()\n critic_loss.backward()\n torch.nn.utils.clip_grad_norm_(critic.parameters(), max_grad_norm)\n critic_optim.step()\n\n # GOALS: TSP_20=3.97, TSP_50=6.08, TSP_100=8.44\n critic_rewards.append(torch.mean(critic_est.detach().data))\n rewards.append(torch.mean(reward.detach().data))\n losses.append(torch.mean(actor_loss.detach().data))\n\n if (batch_idx + 1) % checkpoint_every == 0:\n\n mean_loss = np.mean(losses[-checkpoint_every:])\n mean_reward = np.mean(rewards[-checkpoint_every:])\n\n prefix = 'epoch%d_batch%d_%2.4f' % (epoch, batch_idx, mean_reward)\n save_path = os.path.join(checkpoint_dir, prefix + '_actor.pt')\n torch.save(actor.state_dict(), save_path)\n\n save_path = os.path.join(checkpoint_dir, prefix + '_critic.pt')\n torch.save(critic.state_dict(), save_path)\n\n print('%d/%d, reward: %2.3f, loss: %2.4f' %\n (batch_idx, len(train_loader), mean_reward, mean_loss))\n\n mean_loss = np.mean(losses)\n mean_reward = np.mean(rewards)\n mean_valid = validate(valid_loader, actor, reward_fn, render_fn,\n save_dir, num_plot=5)\n\n print('Mean epoch loss/reward: %2.4f, %2.4f, %2.4f' % \\\n (mean_loss, mean_reward, mean_valid))\n\n\ndef train_tsp(args):\n\n # Goals:\n # TSP20, 3.82 (Optimal) - 3.97 (DRL4VRP)\n # TSP50, 5.70 (Optimal) - 6.08 (DRL4VRP)\n # TSP100, 7.77 (OptimalBS) - 8.44 (DRL4VRP)\n\n from tasks import tsp\n from tasks.tsp import TSPDataset\n\n STATIC_SIZE = 2 # (x, y)\n DYNAMIC_SIZE = 1 # dummy for compatability\n\n train_data = TSPDataset(args['num_nodes'], args['train_size'], args['seed'])\n valid_data = TSPDataset(args['num_nodes'], args['valid_size'], args['seed'])\n\n args['train_data'] = train_data\n args['valid_data'] = valid_data\n args['reward_fn'] = tsp.reward\n args['render_fn'] = tsp.render\n mask_fn = tsp.update_mask\n update_fn = None\n\n actor = DRL4TSP(STATIC_SIZE, \n DYNAMIC_SIZE, \n args['hidden_size'], \n update_fn,\n mask_fn, \n args['num_layers'], \n args['dropout']).to(device)\n\n #critic = Critic(args['hidden_size']).to(device)\n critic = StateCritic(STATIC_SIZE, DYNAMIC_SIZE, args['hidden_size']).to(device)\n \n train(actor, critic, **args)\n\n\ndef train_vrp(args):\n\n # Goals:\n # VRP10, Capacity 20: 4.65 (BS) - 4.80 (Greedy)\n # VRP20, Capacity 30: 6.34 (BS) - 6.51 (Greedy)\n # VRP50, Capacity 40: 11.08 (BS) - 11.32 (Greedy)\n # VRP100, Capacity 50: 16.86 (BS) - 17.12 (Greedy)\n\n from tasks import vrp\n from tasks.vrp import VehicleRoutingDataset\n\n # Determines the maximum amount of load for a vehicle based on num nodes\n LOAD_DICT = {10: 20, 20: 30, 50: 40, 100: 50}\n MAX_DEMAND = 9\n STATIC_SIZE = 2 # (x, y)\n DYNAMIC_SIZE = 2 # (load, demand)\n\n max_load = LOAD_DICT[args['num_nodes']]\n\n train_data = VehicleRoutingDataset(args['train_size'], \n args['num_nodes'],\n max_load, MAX_DEMAND, args['seed'])\n valid_data = VehicleRoutingDataset(args['valid_size'], \n args['num_nodes'],\n max_load, MAX_DEMAND, args['seed'])\n\n args['train_data'] = train_data\n args['valid_data'] = valid_data\n args['reward_fn'] = vrp.reward\n args['render_fn'] = vrp.render\n\n actor = DRL4TSP(STATIC_SIZE, \n DYNAMIC_SIZE, \n args['hidden_size'], \n train_data.update_dynamic,\n train_data.update_mask, \n args['num_layers'], \n args['dropout']).to(device)\n #critic = Critic(args['hidden_size']).to(device)\n critic = StateCritic(STATIC_SIZE, DYNAMIC_SIZE, args['hidden_size']).to(device)\n\n train(actor, critic, **args)\n\n '''\n # path = 'vrp/50/checkpoints/batch13499_11.5925_'\n path = 'vrp/50/checkpoints/batch499_11.6461_'\n params = torch.load(path + 'actor.pt', map_location=lambda storage, loc: storage)\n actor.load_state_dict(params)\n\n params = torch.load(path + 'critic.pt', map_location=lambda storage, loc: storage)\n critic.load_state_dict(params)\n '''\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='Combinatorial Optimization')\n parser.add_argument('--seed', dest='seed', default=1234, type=int)\n parser.add_argument('--task', dest='task', default='tsp')\n parser.add_argument('--nodes', dest='num_nodes', default=20, type=int)\n parser.add_argument('--actor_lr', dest='actor_lr', default=1e-3, type=float)\n parser.add_argument('--critic_lr', dest='critic_lr', default=1e-3,\n type=float)\n parser.add_argument('--max_grad_norm', dest='max_grad_norm', default=2.,\n type=float)\n parser.add_argument('--checkpoint', dest='checkpoint_every', default=500,\n type=int)\n parser.add_argument('--batch_size', dest='batch_size', default=128, type=int)\n parser.add_argument('--hidden', dest='hidden_size', default=128, type=int)\n parser.add_argument('--dropout', dest='dropout', default=0.1, type=float)\n parser.add_argument('--layers', dest='num_layers', default=1, type=int)\n parser.add_argument('--optimizer', dest='optimizer', default='adam')\n parser.add_argument('--train_size', dest='train_size', default=1000000,\n type=int)\n parser.add_argument('--valid_size', dest='valid_size', default=1000,\n type=int)\n\n args = vars(parser.parse_args())\n\n if args['task'] == 'tsp':\n train_tsp(args)\n elif args['task'] == 'vrp':\n train_vrp(args)\n else:\n raise ValueError('Task <%s> not understood'%args['task'])\n", "sub_path": "trainer.py", "file_name": "trainer.py", "file_ext": "py", "file_size_in_byte": 12151, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.device", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 27, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 27, "usage_type": "name"}, {"api_name": "model.Encoder", "line_number": 37, "usage_type": "call"}, {"api_name": "model.Encoder", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 41, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 47, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 57, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 63, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 74, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 74, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.init.xavier_uniform_", "line_number": 80, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 80, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 80, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 84, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 85, "usage_type": "name"}, {"api_name": "torch.no_grad", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 127, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 129, "usage_type": "call"}, {"api_name": "os.path", "line_number": 129, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 131, "usage_type": "call"}, {"api_name": "os.path", "line_number": 131, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 133, "usage_type": "call"}, {"api_name": "os.path", "line_number": 133, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 134, "usage_type": "call"}, {"api_name": "torch.optim.SGD", "line_number": 137, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 137, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 139, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 139, "usage_type": "name"}, {"api_name": "torch.optim.Adagrad", "line_number": 141, "usage_type": "attribute"}, {"api_name": "torch.optim", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 149, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.pow", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 182, "usage_type": "attribute"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 187, "usage_type": "attribute"}, {"api_name": "torch.mean", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 205, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 211, "usage_type": "call"}, {"api_name": "tasks.tsp.TSPDataset", "line_number": 232, "usage_type": "call"}, {"api_name": "tasks.tsp.TSPDataset", "line_number": 233, "usage_type": "call"}, {"api_name": "tasks.tsp.reward", "line_number": 237, "usage_type": "attribute"}, {"api_name": "tasks.tsp", "line_number": 237, "usage_type": "name"}, {"api_name": "tasks.tsp.render", "line_number": 238, "usage_type": "attribute"}, {"api_name": "tasks.tsp", "line_number": 238, "usage_type": "name"}, {"api_name": "tasks.tsp.update_mask", "line_number": 239, "usage_type": "attribute"}, {"api_name": "tasks.tsp", "line_number": 239, "usage_type": "name"}, {"api_name": "model.DRL4TSP", "line_number": 242, "usage_type": "call"}, {"api_name": "tasks.vrp.VehicleRoutingDataset", "line_number": 275, "usage_type": "call"}, {"api_name": "tasks.vrp.VehicleRoutingDataset", "line_number": 278, "usage_type": "call"}, {"api_name": "tasks.vrp.reward", "line_number": 284, "usage_type": "attribute"}, {"api_name": "tasks.vrp", "line_number": 284, "usage_type": "name"}, {"api_name": "tasks.vrp.render", "line_number": 285, "usage_type": "attribute"}, {"api_name": "tasks.vrp", "line_number": 285, "usage_type": "name"}, {"api_name": "model.DRL4TSP", "line_number": 287, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 311, "usage_type": "call"}]} +{"seq_id": "208280795", "text": "from deepar.dataset import Dataset\nimport numpy as np\nimport pandas as pd\nimport logging\nfrom sklearn.preprocessing import StandardScaler, LabelEncoder\nimport tensorflow as tf\n\nimport logging\nlogger = logging.getLogger(__name__)\n#logger.setLevel(logging.INFO)\n\npd.options.mode.chained_assignment = None # default='warn'\n\nclass TimeSeries(Dataset):\n\n def __init__(self, pandas_df, target_idx = None, timestamp_idx = None, grouping_idx=None, one_hot_indices=None, \n index_col = None, count_data=False, negative_obs = 1, val_split = 0.2, mask_value = -10000, \n integer_timestamps = False):\n \"\"\"\n :param pandas_df: df to sample time series from\n :param target_idx: index of target column, if None Error will be raised\n :param timestamp_idx: index of column containing timestamps, timestamps are parsed as # of seconds\n :param grouping_idx: index of grouping column\n :param one_hot_indices: list of indices of columns that are one hot encoded\n :param index_col: index column, if not None will be dropped\n :param count_data: boolean indicating whether data is count data (determines loss function)\n :param negative_obs: how far before beginning of time series is it possible to set t\n :param val_split: proportion of data to withhold for validation\n :param mask_value: mask to use on missing target values\n :param integer_timestamps: whether timestamp column is expressed in ints (instead of s)\n \"\"\"\n super().__init__()\n\n self.data = pandas_df\n\n # store constructor arguments as instance variables\n self.integer_timestamps = integer_timestamps\n self.one_hot_indices = one_hot_indices\n self.negative_obs = negative_obs\n col_names = list(self.data)\n if grouping_idx is None:\n self.data['category'] = 'dummy_cat'\n self.grouping_name = 'category'\n else:\n self.grouping_name = col_names[grouping_idx]\n self.timestamp_idx = timestamp_idx\n self.count_data = count_data\n self.mask_value = mask_value\n\n if self.data is None:\n raise ValueError('Must provide a Pandas df to instantiate this class')\n\n # set name of target variable\n if target_idx is None:\n raise ValueError('Must provide an index for a target column to instantiate this class')\n self.data = self.data.rename(columns={col_names[target_idx]:'target'})\n\n # delete index column if one exists (absolute position information only available through covariates)\n if index_col is not None:\n self.data = self.data.drop(col_names[index_col], axis=1)\n\n # augment dataset with covariates\n time_name = self._sort_by_timestamp(col_names)\n self._age_augmentation()\n self._datetime_augmentation(time_name)\n\n # need embeddings for cats in val, even if not in train\n # 1 extra for test cats not included in train or val\n self.unique_cats = self.data[self.grouping_name].unique()\n self.num_cats = len(self.unique_cats) + 1\n\n # convert groups to ints\n self.label_encoder = LabelEncoder()\n cat_names = self.data[self.grouping_name].append(pd.Series(['dummy_test_category']))\n self.label_encoder.fit(cat_names)\n self.data[self.grouping_name] = self.label_encoder.transform(self.data[self.grouping_name])\n\n # split into train + validation sets, create sampling dist.\n self._train_val_split(val_split)\n\n # standardize\n self._standardize(val_split)\n\n # store number of features and categorical count and target means\n self.features = self.data.shape[1] - 1 # no 1) target or 2) grouping name, add 3) prev target\n self.count_data = count_data\n\n def _sort_by_timestamp(self, col_names):\n \"\"\"\n util function\n sort df by timestamp\n \"\"\"\n if self.timestamp_idx is None:\n raise ValueError('Must provide the index of the timestamp column to instantiate this class')\n time_name = col_names[self.timestamp_idx]\n self.data = self.data.sort_values(by = time_name)\n if self.integer_timestamps:\n self.data[time_name] = pd.to_datetime(self.data[time_name] - 1, unit = 'D')\n else:\n self.data[time_name] = pd.to_datetime(self.data[time_name], unit = 's')\n return time_name\n\n def _age_augmentation(self):\n \"\"\"\n util function\n augment dataset with age covariate\n \"\"\"\n # age (timesteps from 0 for each unique time series)\n self.data['age'] = self.data.groupby(self.grouping_name).cumcount()\n self.train_set_ages = self.data.groupby(self.grouping_name)['age'].agg('max')\n self.train_set_ages['dummy_test_category'] = 0\n\n def _datetime_augmentation(self, time_name):\n \"\"\"\n util function\n augment dataset with datetime covariates\n \"\"\"\n # datetime features\n self.data['hour_of_day'] = self.data[time_name].dt.hour\n self.data['day_of_week'] = self.data[time_name].dt.dayofweek\n self.data['day_of_month'] = self.data[time_name].dt.day\n self.data['day_of_year'] = self.data[time_name].dt.dayofyear\n self.data['week_of_year'] = self.data[time_name].dt.weekofyear\n self.data['month_of_year'] = self.data[time_name].dt.month\n self.data['year'] = self.data[time_name].dt.year\n self.data = self.data.drop(time_name, axis=1)\n\n def _create_sampling_dist(self):\n \"\"\"\n util function\n create scaled sampling distribution over time series \n \"\"\"\n\n self.scale_factors = 1 + self.target_means\n #self.scale_factors = self.scale_factors.apply(lambda x: max(0, x))\n\n # softmax the distribution for sampling\n e_x = np.exp(self.scale_factors - np.max(self.scale_factors))\n self.scale_factors_softmax = e_x / e_x.sum(axis = 0)\n \n def _train_val_split(self, val_split):\n \"\"\"\n util function\n split dataset object into training and validation data frames\n \"\"\"\n # split data into training and validation sets\n assert val_split >= 0 and val_split < 1, \\\n 'Validation split must be between 0 (inclusive) and 1 (exclusive)'\n\n if val_split == 0:\n self.train_data = self.data\n else:\n nrow = int(self.data.shape[0] * val_split)\n self.train_data = self.data.head(self.data.shape[0] - nrow)\n self.val_data = self.data.tail(nrow)\n\n # store target means over training set\n self.target_means = self.train_data.groupby(self.grouping_name)['target'].agg('mean')\n self.target_mean = self.train_data['target'].dropna().mean()\n\n # create scale factor sampling dist. before adding dummy keys\n self._create_sampling_dist()\n\n # add 'dummy_test_category' as key to target means\n self.target_means[self.label_encoder.transform(['dummy_test_category'])[0]] = self.target_mean\n\n if val_split != 0:\n # if group in val doesn't exist in train, standardize by overall mean\n for group in self.val_data[self.grouping_name].unique():\n if group not in self.train_data[self.grouping_name].unique():\n self.target_means[group] = self.target_mean\n\n def _mask_missing_targets(self, df):\n \"\"\"\n util function\n mask missing target values in training and validation frames\n \"\"\"\n # mask missing target values\n for idx in pd.isnull(df)['target'].to_numpy().nonzero()[0]:\n key = df[self.grouping_name][idx]\n if key in self.missing_tgt_vals.keys():\n self.missing_tgt_vals[key].append(df['age'][idx])\n else:\n self.missing_tgt_vals[key] = [df['age'][idx]]\n\n def _standardize(self, val_split):\n \"\"\" \n util function\n standardize covariates and record locations of missing tgt values (for standardization later)\n \"\"\"\n # standardize covariates N(0,1) and 'target' col by mean\n covariate_mask = [False if col_name == 'target' or col_name == self.grouping_name\n else True for col_name in self.data.columns]\n self.scaler = StandardScaler()\n self.train_data.loc[:, covariate_mask] = self.scaler.fit_transform(self.train_data.loc[:, covariate_mask].astype('float'))\n\n # record locations of missing target values\n self.missing_tgt_vals = {}\n self._mask_missing_targets(self.train_data)\n\n if val_split != 0:\n self.val_data = self.val_data.reset_index(drop=True)\n self.val_data.loc[:, covariate_mask] = self.scaler.transform(self.val_data.loc[:, covariate_mask].astype('float'))\n\n # record locations of missing target values\n self._mask_missing_targets(self.val_data)\n \n # keep full dataset up to date\n self.data.loc[:, covariate_mask] = self.scaler.transform(self.data.loc[:, covariate_mask].astype('float'))\n\n def _one_hot_padding(self, pandas_df, padding_df):\n \"\"\"\n Util padding function\n :param padding_df:\n :param one_hot_root_list:\n :return: padding_df\n\n from https://github.com/arrigonialberto86/deepar\n \"\"\"\n for one_hot_root in self.one_hot_indices:\n one_hot_columns = [i for i in pandas_df.columns # select columns equal to 1\n if i.startswith(one_hot_root) and pandas_df[i].values[0] == 1]\n for col in one_hot_columns:\n padding_df[col] = 1\n return padding_df\n\n def _pad_ts(self, pandas_df, desired_len, padding_val=0):\n \"\"\"\n Add padding int to the time series\n :param pandas_df:\n :param desired_len: (int)\n :param padding_val: (int)\n :return: X (feature_space), y\n \n from https://github.com/arrigonialberto86/deepar\n \"\"\"\n pad_length = desired_len - pandas_df.shape[0]\n padding_df = pd.DataFrame({col: padding_val for col in pandas_df.columns}, \n index=[i for i in range(pad_length)])\n\n if self.one_hot_indices:\n padding_df = self._one_hot_padding(pandas_df, padding_df)\n\n return pd.concat([padding_df, pandas_df]).reset_index(drop=True)\n\n def _sample_ts(self, pandas_df, desired_len, padding_val = 0, negative_obs = 1, val_set = False):\n \"\"\"\n :param pandas_df: input pandas df with 'target' columns e features\n :param desired_len: desired sample length (number of rows)\n :param padding_val: default is 0\n :param negative_obs: how far before beginning of time series is it possible to set t\n :param val_set: whether we are sampling from train or validation data\n :return: a pandas df (sample)\n \n from https://github.com/arrigonialberto86/deepar\n \"\"\"\n if pandas_df.shape[0] < desired_len:\n raise ValueError('Desired sample length is greater than df row len')\n if pandas_df.shape[0] == desired_len:\n return pandas_df\n\n # do not sample negative observations in validation\n if val_set:\n negative_obs = 0\n start_index = np.random.choice([i for i in range(0 - negative_obs, pandas_df.shape[0] - desired_len)])\n\n # replace beginning of series with padded values to learn beginning\n if start_index < 0:\n return self._pad_ts(pandas_df.head(desired_len + start_index), desired_len, padding_val = padding_val)\n\n return pandas_df.iloc[start_index: start_index+desired_len, ]\n\n def _add_prev_target_col(self, df, train_df = None):\n \"\"\"\n util function\n add column with previous target value for autoregressive modeling\n \"\"\"\n\n df = df.reset_index(drop=True)\n if train_df is None:\n\n # add feature column for previous output value (z_{t-1})\n df.loc[:,'prev_target'] = pd.Series([0]).append(df['target'].iloc[:-1], ignore_index=True)\n\n # scale\n df.loc[:, 'prev_target'] = \\\n df['prev_target'] / self.target_means[df[self.grouping_name]].reset_index(drop = True)\n\n # interpolate (will only replace NA rows (first time))\n df.loc[:,'prev_target'] = df['prev_target'].interpolate()\n\n else:\n df.loc[:, 'prev_target'] = \\\n train_df['target'].tail(1).repeat(repeats = df.shape[0]).reset_index(drop = True)\n df.loc[:, 'prev_target'] = \\\n df['prev_target'] / self.target_means[df[self.grouping_name]].reset_index(drop = True)\n\n # replace target missing rows with mask\n df.loc[:, 'target'] = df['target'].fillna(self.mask_value)\n\n return df\n\n def _sample_missing_tgts(self, df, model, category, missing_tgt_vals, window_size, batch_size, include_target = True):\n \"\"\"\n util function\n sample missing target values from current model parameters\n \"\"\"\n\n # sample missing 'targets' from current model parameters (for 'prev_targets')\n if category in missing_tgt_vals.keys():\n if not set(missing_tgt_vals[category]).isdisjoint(df['age'].iloc[:-1]):\n drop_list = [self.grouping_name]\n if include_target:\n drop_list.append('target')\n continuous = df.drop(drop_list, 1).values.reshape(1, window_size, -1)\n continuous = np.repeat(continuous, batch_size, axis = 0)\n categorical = df[self.grouping_name].values.reshape(1, window_size)\n categorical = np.repeat(categorical, batch_size, axis = 0)\n preds = model([continuous, categorical], training = True)[0][0]\n\n # refill indices \n refill_indices = df.index[df['age'].isin(missing_tgt_vals[category])]\n refill_values = [preds[i] for i in [r - df.index[0] for r in refill_indices]]\n for idx, val in zip(refill_indices, refill_values):\n df['prev_target'][idx] = val\n return df\n\n def next_batch(self, model, batch_size, window_size, verbose=False, padding_value=0, val_set = False):\n \"\"\"\n :param model: model object, allows sampling for missing target obs. in training set\n :param batch_size: how many time series to be sampled in this batch (int)\n :param window_size: window of each sampled time series\n :param verbose: default false\n :param padding_value: default 0\n :param val_set: boolean, whether this generator should sample for training or validation\n :return: [X_continouous, X_categorical], C (categorical grouping variable), y\n\n bootstrapped from https://github.com/arrigonialberto86/deepar\n \"\"\"\n\n # save padding value for test object\n self.padding_value = padding_value\n\n # Generate sampling of time series according to prob dist. defined by scale factors\n if val_set:\n assert self.val_data is not None, \"Asking for validation batch, but validation split was 0 in object construction\"\n cat_samples = np.random.choice(self.val_data[self.grouping_name].unique(), batch_size)\n data = self.val_data\n else:\n cat_samples = np.random.choice(self.train_data[self.grouping_name].unique(), batch_size, \n p = self.scale_factors_softmax)\n data = self.train_data\n\n sampled = []\n for cat in cat_samples:\n cat_data = data[data[self.grouping_name] == cat]\n\n # add 'prev_target' column for this category\n if val_set:\n cat_data = self._add_prev_target_col(cat_data, self.train_data)\n else:\n cat_data = self._add_prev_target_col(cat_data)\n\n # Initial padding for each selected time series to reach window_size\n if cat_data.shape[0] < window_size:\n sampled_cat_data = self._pad_ts(pandas_df=cat_data,\n desired_len=window_size,\n padding_val=padding_value)\n # sample window from time series\n else:\n sampled_cat_data = self._sample_ts(pandas_df=cat_data,\n desired_len=window_size,\n padding_val=padding_value,\n negative_obs=self.negative_obs,\n val_set = val_set)\n\n # sample missing 'targets' from current model parameters (for 'prev_targets')\n sampled_cat_data = self._sample_missing_tgts(sampled_cat_data, model, cat, self.missing_tgt_vals, \n window_size, batch_size)\n \n sampled.append(sampled_cat_data)\n data = pd.concat(sampled)\n\n # [cont_inputs, cat_inputs], cat_labels, targets\n return ([tf.constant(data.drop(['target', self.grouping_name], 1).values.reshape(batch_size, window_size, -1), dtype = tf.float32),\n tf.constant(data[self.grouping_name].values.reshape(batch_size, window_size), dtype = tf.float32)], \n tf.constant(cat_samples.reshape(batch_size, 1), dtype = tf.int32),\n tf.constant(data['target'].values.reshape(batch_size, window_size, 1), dtype = tf.float32))\n\nclass TimeSeriesTest(TimeSeries):\n\n def __init__(self, pandas_df, train_ts_obj, target_idx = None, timestamp_idx = None, \n grouping_idx=None, one_hot_indices=None, index_col = None, mask_value = -1):\n \"\"\"\n :param pandas_df: df of test time series\n :param train_ts_obj: TimeSeries object defined on training / validation set\n :param target_idx: index of target column, if None Error will be raised\n :param timestamp_idx: index of column containing timestamps, timestamps are parsed as # of seconds\n :param grouping_idx: index of grouping column\n :param one_hot_indices: list of indices of columns that are one hot encoded\n :param index_col: index column, if not None will be dropped\n :param mask_value: mask to use on testing timestep indices > 1 (bc predictions batched)\n \"\"\"\n \n self.data = pandas_df\n\n # store constructor arguments as instance variables\n self.one_hot_indices = one_hot_indices\n col_names = list(self.data)\n if grouping_idx is None:\n self.data['category'] = 'dummy_cat'\n self.grouping_name = 'category'\n else:\n self.grouping_name = col_names[grouping_idx]\n self.timestamp_idx = timestamp_idx\n self.mask_value = mask_value\n\n # some instance variables read from train ts object\n self.train_ts_obj = train_ts_obj\n self.integer_timestamps = train_ts_obj.integer_timestamps\n self.count_data = train_ts_obj.count_data\n\n if self.data is None:\n raise ValueError('Must provide a Pandas df to instantiate this class')\n\n # delete target column if one exists (not needed for test)\n if target_idx is not None:\n self.data = self.data.drop(col_names[target_idx], axis=1)\n\n # delete index column if one exists (absolute position information only available through covariates)\n if index_col is not None:\n self.data = self.data.drop(col_names[index_col], axis=1)\n\n # sort df by timestamp\n time_name = self._sort_by_timestamp(col_names)\n\n # age (timesteps from beginning of train set for each unique time series)\n self.test_groups = self.data[self.grouping_name].unique()\n self.new_test_groups = []\n for group in self.test_groups:\n if group not in train_ts_obj.unique_cats:\n train_ts_obj.train_set_ages[group] = 0\n self.new_test_groups.append(group)\n \n # add 1 because cumcount() starts at 0\n self.data['age'] = self.data.groupby(self.grouping_name).cumcount() + 1\n self.data['age'] += train_ts_obj.train_set_ages[self.data[self.grouping_name]].reset_index(drop=True)\n\n # datetime features\n self._datetime_augmentation(time_name)\n\n # standardize covariates N(0,1)\n covariate_mask = [False if col_name == self.grouping_name else True for col_name in self.data.columns]\n self.data.loc[:, covariate_mask] = train_ts_obj.scaler.transform(self.data.loc[:, covariate_mask].astype('float'))\n\n # compute max prediction horizon\n self.horizon = self.data.groupby(self.grouping_name)['age'].count().max()\n\n # assert compatibility with training TimeSeries object)\n assert self.data.shape[1] == train_ts_obj.features, \\\n \"Number of feature columns in test object must be equal to the number in train object\"\n assert self.count_data == train_ts_obj.count_data, \\\n \"Count data boolean in test object must be equivalent to train object\"\n\n self.prepared = False\n\n def _sort_by_timestamp(self, *args, **kwargs):\n return super(TimeSeriesTest, self)._sort_by_timestamp(*args, **kwargs)\n\n def _datetime_augmentation(self, *args, **kwargs):\n super(TimeSeriesTest, self)._datetime_augmentation(*args, **kwargs)\n\n def _pad_ts(self, *args, **kwargs):\n return super(TimeSeriesTest, self)._pad_ts(*args, **kwargs)\n\n def _add_prev_target_col(self, df, train_df = None):\n \"\"\"\n util function\n add column with previous target value for autoregressive modeling\n \"\"\"\n df = df.reset_index(drop=True)\n if train_df is None:\n\n # add feature column for previous output value (z_{t-1})\n df.loc[:,'prev_target'] = pd.Series([0]).append(df['target'].iloc[:-1], ignore_index=True)\n\n # scale\n df.loc[:, 'prev_target'] = \\\n df['prev_target'] / self.train_ts_obj.target_means[df[self.grouping_name]].reset_index(drop = True)\n\n # interpolate (will only replace NA rows (first time))\n df.loc[:,'prev_target'] = df['prev_target'].interpolate()\n\n else:\n df.loc[:, 'prev_target'] = \\\n train_df['target'].tail(1).repeat(repeats = df.shape[0]).reset_index(drop = True)\n df.loc[:, 'prev_target'] = \\\n df['prev_target'] / self.train_ts_obj.target_means[df[self.grouping_name]].reset_index(drop = True)\n\n return df\n\n def _sample_missing_tgts(self, *args, **kwargs):\n return super(TimeSeriesTest, self)._sample_missing_tgts(*args, **kwargs)\n\n def _prepare_batched_test_data(self, batch_size, window_size, include_all_training = False, verbose = False):\n \"\"\"\n Split data into batches of window_size (all batches include all categories) for stateful inference\n :param batch_size: batch size\n :param window_size: window of each sampled time series\n :param include_all_training: whether to include all training data in prep of batches\n :param verbose: default false\n \"\"\"\n\n if include_all_training:\n max_train_age = self.train_ts_obj.data.groupby(self.train_ts_obj.grouping_name)['target'].count().max()\n if max_train_age % window_size != 0:\n max_train_age = (max_train_age // window_size) * window_size + window_size\n else:\n max_train_age = window_size\n\n # calculate # train and test batches\n self.train_batch_ct = max_train_age // window_size\n # last train batch produces first test output\n self.test_batch_ct = self.train_batch_ct + self.horizon\n\n data = []\n self.scale_keys = []\n for cat in self.test_groups:\n if cat in self.new_test_groups:\n train_data = pd.DataFrame({col: self.train_ts_obj.padding_value for col in self.data.columns}, \n index=[i for i in range(max_train_age)])\n \n # add 'prev_target' column for this series\n train_data = self._add_prev_target_col(train_data)\n \n else:\n enc_cat = self.train_ts_obj.label_encoder.transform([cat])[0]\n\n train_data = self.train_ts_obj.data[self.train_ts_obj.data[self.grouping_name] == enc_cat]\n\n # add 'prev_target' column for this series\n train_data = self._add_prev_target_col(train_data)\n if train_data.shape[0] < max_train_age:\n train_data = self._pad_ts(pandas_df=train_data,\n desired_len=max_train_age,\n padding_val=self.train_ts_obj.padding_value)\n else:\n train_data = train_data.tail(max_train_age)\n train_data = train_data.reset_index(drop=True)\n\n # convert groups to ints in test data\n test_data = self.data[self.data[self.grouping_name] == cat]\n if cat in self.new_test_groups:\n test_data[self.grouping_name] = 'dummy_test_category'\n test_data[self.grouping_name] = self.train_ts_obj.label_encoder.transform(test_data[self.grouping_name])\n \n # add prev target w/ same value for all rows to test and scale\n test_data = self._add_prev_target_col(test_data, train_df = train_data)\n\n # append test data, drop 'target' col from training data\n prepped_data = \\\n pd.concat([train_data.drop('target', axis=1), test_data]).reset_index(drop = True)\n data.append(prepped_data)\n \n # track scale keys for inverse scaling at inference\n self.scale_keys.append(prepped_data[self.grouping_name][0])\n\n self.prepped_data = data\n self.prepared = True\n self.batch_idx = 0\n\n # pad scale keys until batch size\n self.scale_keys.extend([self.scale_keys[0]] * (batch_size - len(self.scale_keys)))\n self.scale_keys = \\\n tf.constant(np.array(self.scale_keys).reshape(batch_size, 1), dtype = tf.int32)\n\n def _one_hot_padding(self, *args, **kwargs):\n return super(TimeSeriesTest, self)._one_hot_padding(*args, **kwargs)\n\n def next_batch(self, model, batch_size, window_size, include_all_training = False, verbose=False):\n \"\"\"\n Split data into batches of window_size (all batches include all categories) for stateful inference\n :param batch_size: batch size\n :param window_size: window of each sampled time series\n :param include_all_training: whether to include all training data in prep of batches\n :param verbose: default false\n :return [X_continouous, X_categorical], C (categorical grouping variable), prediction_horizon_index\n \"\"\"\n\n if not self.prepared:\n self._prepare_batched_test_data(batch_size, \n window_size, \n include_all_training = include_all_training, \n verbose = verbose)\n\n # return -1 if no more batches\n if self.batch_idx == self.test_batch_ct:\n return (None, None)\n \n # grab current batch\n if self.batch_idx >= self.train_batch_ct:\n batch_data = []\n start_idx = self.train_batch_ct * window_size + self.batch_idx - self.train_batch_ct\n for df in self.prepped_data:\n # mask continuous inputs so timesteps > 0 will be skipped in lstm inference\n datum = df.iloc[start_idx:start_idx + window_size, :].reset_index(drop = True)\n datum.loc[1:, datum.columns != self.grouping_name] = self.mask_value\n if window_size - datum.shape[0] > 0:\n datum = datum.append(pd.concat([datum.iloc[-1:, :]] * (window_size - datum.shape[0]), \n ignore_index = True))\n batch_data.append(datum)\n else:\n batch_data = [df.iloc[self.batch_idx * window_size:(self.batch_idx + 1) * window_size, :] \n for df in self.prepped_data] \n \n # sample missing 'targets' from current model parameters (for 'prev_targets')\n batch_data = [self._sample_missing_tgts(b_data, model, b_data[self.grouping_name].iloc[0], \n self.train_ts_obj.missing_tgt_vals, window_size, batch_size, include_target = False) \n for b_data in batch_data]\n\n batch_data = pd.concat(batch_data)\n self.batch_idx += 1\n\n # 'prev_target' in test batches will be overwritten during ancestral sampling in predict, \n # so doesn't matter if dropped here\n x_cont = batch_data.drop([self.grouping_name], 1).values.reshape(len(self.test_groups), window_size, -1)\n x_cat = batch_data[self.grouping_name].values.reshape(len(self.test_groups), window_size)\n x_cont = tf.Variable(np.append(x_cont, [x_cont[0]] * (batch_size - len(self.test_groups)), axis = 0), \n dtype = tf.float32)\n x_cat = tf.constant(np.append(x_cat, [x_cat[0]] * (batch_size - len(self.test_groups)), axis = 0), \n dtype = tf.float32)\n return ([x_cont, x_cat], self.batch_idx - self.train_batch_ct) \n\ndef train_ts_generator(model, ts_obj, batch_size, window_size, verbose = False, padding_value = 0, val_set = False):\n \"\"\"\n This is a util generator function\n :param model: model (with current parameters) to sample missing tgt values\n :param ts_obj: a TimeSeries class object that implements the 'next_batch' method\n :param batch_size: batch size\n :param window_size: window of each sampled time series\n :param verbose: default false\n :param padding_value: default 0\n :param val_set: boolean for mode, True is validation \n :yield: [X_continouous, X_categorical], C (categorical grouping variable), y\n\n bootstrapped from https://github.com/arrigonialberto86/deepar\n \"\"\"\n while 1:\n yield ts_obj.next_batch(model, \n batch_size, \n window_size, \n verbose = verbose, \n padding_value = padding_value, \n val_set = val_set)\n \ndef test_ts_generator(model, ts_obj, batch_size, window_size, include_all_training = False, verbose = False):\n \"\"\"\n This is a util generator function\n :param model: model (with current parameters) to sample missing tgt values\n :param ts_obj: a TimeSeriesTest class object that implements the 'next_batch' method\n :param batch_size: batch size\n :param window_size: window of each sampled time series\n :param include_all_training: whether to start inference at beginning of training set\n :param verbose: default false\n :yield: [X_continouous, X_categorical], C (categorical grouping variable), prediction_horizon_index\n \"\"\"\n \n while 1:\n x_test, horizon_idx = ts_obj.next_batch(model, \n batch_size, \n window_size, \n include_all_training = include_all_training,\n verbose = verbose)\n x_test, horizon_idx\n yield x_test, horizon_idx\n\n\n", "sub_path": "deepar/dataset/time_series.py", "file_name": "time_series.py", "file_ext": "py", "file_size_in_byte": 31551, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 9, "usage_type": "call"}, {"api_name": "pandas.options", "line_number": 12, "usage_type": "attribute"}, {"api_name": "deepar.dataset.Dataset", "line_number": 14, "usage_type": "name"}, {"api_name": "sklearn.preprocessing.LabelEncoder", "line_number": 73, "usage_type": "call"}, {"api_name": "pandas.Series", "line_number": 74, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 98, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 138, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 179, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.StandardScaler", "line_number": 194, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 238, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 244, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 265, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 316, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 318, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 347, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 347, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 350, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 382, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 385, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 385, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 386, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 386, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 387, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 387, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 388, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 388, "usage_type": "attribute"}, {"api_name": "pandas.Series", "line_number": 485, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 530, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 562, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 575, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 575, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 609, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 621, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 628, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 628, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 629, "usage_type": "attribute"}, {"api_name": "tensorflow.constant", "line_number": 630, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 630, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 631, "usage_type": "attribute"}]} +{"seq_id": "288095249", "text": "# -*- coding: utf-8 -*-\n\nfrom .models import Report\nfrom backtest_py2.settings import MEDIA_ROOT\nfrom django.core.files.storage import default_storage\nfrom utils import utils\nimport os\nimport fcntl\n\ndef Query():\n queryset = Report.objects.filter(status = 0)\n inprocess_queryset = Report.objects.filter(status = 1)\n in_length = len(inprocess_queryset)\n for report in queryset:\n if(in_length < 5):\n report.status = 1\n report.save()\n print(report.alpha_name, report.file)\n in_length += 1\n flag = False\n utils.unzip(report)\n if (utils.validate_files(report)):\n try:\n flag = utils.compile_alpha(report) \n except RuntimeError as e:\n report.error_message = u\"编译错误\"\n #utils.clean()\n if (flag == True):\n report.status = 2\n else:\n report.status = 3\n else:\n report.error_message = u\"解压错误或文件名错误\"\n report.status = 3\n report.save()", "sub_path": "report/query.py", "file_name": "query.py", "file_ext": "py", "file_size_in_byte": 1152, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "models.Report.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Report.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "models.Report", "line_number": 11, "usage_type": "name"}, {"api_name": "models.Report.objects.filter", "line_number": 12, "usage_type": "call"}, {"api_name": "models.Report.objects", "line_number": 12, "usage_type": "attribute"}, {"api_name": "models.Report", "line_number": 12, "usage_type": "name"}, {"api_name": "utils.utils.unzip", "line_number": 21, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 21, "usage_type": "name"}, {"api_name": "utils.utils.validate_files", "line_number": 22, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 22, "usage_type": "name"}, {"api_name": "utils.utils.compile_alpha", "line_number": 24, "usage_type": "call"}, {"api_name": "utils.utils", "line_number": 24, "usage_type": "name"}]} +{"seq_id": "346512734", "text": "import telebot\nimport keyboard\n\nimport library_db as libdb\nimport translator\nimport context\n\nimport config\nimport status\n\nbot = telebot.TeleBot(config.token)\n\n\n# start message\n@bot.message_handler(commands=[\"start\"])\ndef start_message(message):\n status.set_state(message.chat.id, config.States.S_START.value) # задание начального состояния\n bot.send_message(message.chat.id, \"Выберите действие:\\n\\n\", reply_markup=keyboard.keyboard)\n\n\n# поиск в контексте\n@bot.message_handler(func=lambda message:\n status.get_current_state(message.chat.id) == config.States.S_CONTEXT.value)\ndef get_context(message):\n text = message.text\n # отправка запроса на получение контекста (из context.py)\n list_with_contexts = context.get_context(text)\n bot.send_message(message.chat.id, \"Ниже приведены примеры использования:\")\n status.set_state(message.chat.id, config.States.S_START.value)\n # ответ возвращается в виде списков со всеми найденными контекстами\n # для возможности в дальнейшем пользователю выбрать сколько вариантов будет выводиться на экран,\n # функция возвращает все, пока что показываем только три\n try:\n bot.send_message(message.chat.id, list_with_contexts[0][1] + '\\n\\n' + list_with_contexts[1][1])\n except:\n bot.send_message(message.chat.id, \"Что-то пошло не так. Возможно, в тексте содержится ошибка, \"\n \"или запрос очень непопулярен\")\n try:\n bot.send_message(message.chat.id, list_with_contexts[0][2] + '\\n\\n' + list_with_contexts[1][2])\n except:\n status.set_state(message.chat.id, config.States.S_START.value)\n try:\n bot.send_message(message.chat.id, list_with_contexts[0][3] + '\\n\\n' + list_with_contexts[1][3],\n reply_markup=keyboard.keyboard)\n except:\n status.set_state(message.chat.id, config.States.S_START.value)\n\n\n# обработка событий при нажатии кнопок\n@bot.callback_query_handler(func=lambda call: True)\ndef callback_inline(call):\n if call.data == \"add_word_from_ru\":\n bot.send_message(call.message.chat.id,\n \"Введите текст для перевода на английский язык и добавления в словарь:\\n\")\n status.set_state(call.message.chat.id, config.States.S_ENTER_WORD_RU.value)\n elif call.data == \"add_word_from_en\":\n bot.send_message(call.message.chat.id,\n \"Введите текст для перевода на русский язык и добавления в словарь:\\n\")\n status.set_state(call.message.chat.id, config.States.S_ENTER_WORD_EN.value)\n elif call.data == \"library\":\n try:\n result = libdb.get_library(str(call.message.chat.id))\n bot.send_message(call.message.chat.id, \"\" + result, reply_markup=keyboard.keyboard_library)\n status.set_state(call.message.chat.id, config.States.S_START.value)\n except:\n bot.send_message(call.message.chat.id, \"Словарь пуст. Давай для начала добавим пару переводов:\",\n reply_markup=keyboard.keyboard_error_lib)\n elif call.data == \"delete\":\n bot.send_message(call.message.chat.id, \"Введите id перевода, который хотите удалить\")\n status.set_state(call.message.chat.id, config.States.S_DELETE.value)\n elif call.data == \"delete_last\":\n try:\n libdb.delete_last(str(call.message.chat.id))\n bot.send_message(call.message.chat.id, \"Из словаря удален последний перевод\")\n status.set_state(call.message.chat.id, config.States.S_START.value)\n except:\n bot.send_message(call.message.chat.id, \"Что-то пошло не так.\\nВозможно, словарь пуст.\",\n reply_markup=keyboard.keyboard_error_lib)\n status.set_state(call.message.chat.id, config.States.S_START.value)\n elif call.data == \"from_ru_yourself\":\n bot.send_message(call.message.chat.id, \"Введите свой вариант:\\n(только текст перевода на английском)\")\n status.set_state(call.message.chat.id, config.States.S_ENTER_YOURSELF_EN.value)\n elif call.data == \"from_en_yourself\":\n bot.send_message(call.message.chat.id, \"Введите свой вариант:\\n(только текст перевода на русском)\")\n status.set_state(call.message.chat.id, config.States.S_ENTER_YOURSELF_RU.value)\n elif call.data == \"game\":\n try:\n random_test = libdb.get_random(str(call.message.chat.id))\n status.set_state(call.message.chat.id, random_test[1])\n bot.send_message(call.message.chat.id, \"Введите перевод:\\n'{}'\".format(random_test[0]))\n # this module is still in process\n except:\n bot.send_message(call.message.chat.id, \"Что-то пошло не так.\\nВозможно, словарь пуст или перевод содержит \"\n \"одну букву/цифру.\", reply_markup=keyboard.keyboard)\n status.set_state(call.message.chat.id, config.States.S_START.value)\n elif call.data == \"exit_game\":\n status.set_state(call.message.chat.id, config.States.S_START.value)\n bot.send_message(call.message.chat.id, \"Выберите действие:\\n\\n\", reply_markup=keyboard.keyboard)\n elif call.data == \"help\":\n if status.get_state_game(call.message.chat.id):\n bot.send_message(call.message.chat.id,\n \"Правильный перевод:\\n{}\\nПродожим?\".format(\n status.get_current_state(call.message.chat.id)),\n reply_markup=keyboard.keyboard_correct_answer)\n status.set_state(call.message.chat.id, config.States.S_START.value)\n else:\n bot.send_message(call.message.chat.id, \"Видимо, кнопка была нажата вне игры. Вы отдичный тестировщик!\",\n reply_markup=keyboard.keyboard)\n status.set_state(call.message.chat.id, config.States.S_START.value)\n elif call.data == \"context\":\n status.set_state(call.message.chat.id, config.States.S_CONTEXT.value)\n bot.send_message(call.message.chat.id, \"Введите слово или текст для выполнения поиска контекста.\")\n\n\n# перевод с английского\n@bot.message_handler(\n func=lambda message: status.get_current_state(message.chat.id) == config.States.S_ENTER_WORD_EN.value)\ndef translation_from_en(message):\n status.set_state(message.chat.id, config.States.S_START.value)\n text = message.text\n result = translator.translate_en(text)\n bot.send_message(message.chat.id, \"В словарь добавлен перевод:\\n\\n{0} - {1}\\n\\nПереведено сервисом \"\n \"«Яндекс.Переводчик»\\nhttp://translate.yandex.ru\".format(message.text, result),\n reply_markup=keyboard.keyboard_after_translation_en)\n db_data = {message.text: result}\n libdb.add_translation(str(message.chat.id), db_data)\n\n\n# перевод с русского\n@bot.message_handler(\n func=lambda message: status.get_current_state(message.chat.id) == config.States.S_ENTER_WORD_RU.value)\ndef translation_from_ru(message):\n status.set_state(message.chat.id, config.States.S_START.value)\n text = message.text\n result = translator.translate_ru(text)\n bot.send_message(message.chat.id, \"В словарь добавлен перевод:\\n\\n{0} - {1}\\n\\nПереведено сервисом \"\n \"«Яндекс.Переводчик»\\nhttp://translate.yandex.ru\".format(message.text, result),\n reply_markup=keyboard.keyboard_after_translation_ru)\n db_data = {result: message.text}\n libdb.add_translation(str(message.chat.id), db_data)\n status.set_state(message.chat.id, config.States.S_START.value)\n\n\n# удаление перевода\n@bot.message_handler(func=lambda message: status.get_current_state(message.chat.id) == config.States.S_DELETE.value)\ndef delete_item(message):\n try:\n libdb.delete_one(str(message.chat.id), int(message.text))\n bot.send_message(message.chat.id, \"Перевод успешно удален из словаря\", reply_markup=keyboard.keyboard)\n status.set_state(message.chat.id, config.States.S_START.value)\n except:\n bot.send_message(message.chat.id, \"Что-то пошло не так.\\nДля получения id перевода откро�� словарь.\",\n reply_markup=keyboard.keyboard_error_delete_id)\n status.set_state(message.chat.id, config.States.S_START.value)\n\n\n# изменение перевода с английского пользователем\n@bot.message_handler(func=lambda message:\n status.get_current_state(message.chat.id) == config.States.S_ENTER_YOURSELF_EN.value)\ndef translation_yourself_en(message):\n try:\n text = message.text\n libdb.correct_ru(str(message.chat.id), text)\n bot.send_message(message.chat.id, \"Перевод заменён успешно\", reply_markup=keyboard.keyboard)\n status.set_state(message.chat.id, config.States.S_START.value)\n except:\n bot.send_message(message.chat.id, \"Что-то пошло не так.\")\n status.set_state(message.chat.id, config.States.S_START.value)\n\n\n# изменение перевода с русского пользователем\n@bot.message_handler(func=lambda message:\n status.get_current_state(message.chat.id) == config.States.S_ENTER_YOURSELF_RU.value)\ndef translation_yourself_ru(message):\n try:\n text = message.text\n libdb.correct_en(str(message.chat.id), text)\n bot.send_message(message.chat.id, \"Перевод заменён успешно\", reply_markup=keyboard.keyboard)\n status.set_state(message.chat.id, config.States.S_START.value)\n except:\n bot.send_message(message.chat.id, \"Что-то пошло не так. Возможно, кнопка нажата при пустом словаре?\",\n reply_markup=keyboard.keyboard_error_lib)\n status.set_state(message.chat.id, config.States.S_START.value)\n\n\n# запуск тренировки (временный вариант)\n@bot.message_handler(func=lambda message: status.get_state_game(message.chat.id))\ndef test_start(message):\n if message.text.lower() == status.get_current_state(message.chat.id).lower():\n bot.send_message(message.chat.id, \"Верно! Продолжим? :)\", reply_markup=keyboard.keyboard_correct_answer)\n status.set_state(message.chat.id, config.States.S_START.value)\n else:\n bot.send_message(message.chat.id, \"Что-то пошло не так, попробуй ещё раз.\",\n reply_markup=keyboard.keyboard_error_game)\n\n\n# polling\nif __name__ == '__main__':\n bot.polling(none_stop=True)\n", "sub_path": "bot.py", "file_name": "bot.py", "file_ext": "py", "file_size_in_byte": 11818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "telebot.TeleBot", "line_number": 11, "usage_type": "call"}, {"api_name": "config.token", "line_number": 11, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 17, "usage_type": "call"}, {"api_name": "config.States", "line_number": 17, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard", "line_number": 18, "usage_type": "attribute"}, {"api_name": "context.get_context", "line_number": 27, "usage_type": "call"}, {"api_name": "status.set_state", "line_number": 29, "usage_type": "call"}, {"api_name": "config.States", "line_number": 29, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 41, "usage_type": "call"}, {"api_name": "config.States", "line_number": 41, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard", "line_number": 44, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 46, "usage_type": "call"}, {"api_name": "config.States", "line_number": 46, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 23, "usage_type": "call"}, {"api_name": "config.States", "line_number": 23, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 55, "usage_type": "call"}, {"api_name": "config.States", "line_number": 55, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 59, "usage_type": "call"}, {"api_name": "config.States", "line_number": 59, "usage_type": "attribute"}, {"api_name": "library_db.get_library", "line_number": 62, "usage_type": "call"}, {"api_name": "keyboard.keyboard_library", "line_number": 63, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 64, "usage_type": "call"}, {"api_name": "config.States", "line_number": 64, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard_error_lib", "line_number": 67, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 70, "usage_type": "call"}, {"api_name": "config.States", "line_number": 70, "usage_type": "attribute"}, {"api_name": "library_db.delete_last", "line_number": 73, "usage_type": "call"}, {"api_name": "status.set_state", "line_number": 75, "usage_type": "call"}, {"api_name": "config.States", "line_number": 75, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard_error_lib", "line_number": 78, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 79, "usage_type": "call"}, {"api_name": "config.States", "line_number": 79, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 82, "usage_type": "call"}, {"api_name": "config.States", "line_number": 82, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 85, "usage_type": "call"}, {"api_name": "config.States", "line_number": 85, "usage_type": "attribute"}, {"api_name": "library_db.get_random", "line_number": 88, "usage_type": "call"}, {"api_name": "status.set_state", "line_number": 89, "usage_type": "call"}, {"api_name": "keyboard.keyboard", "line_number": 94, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 95, "usage_type": "call"}, {"api_name": "config.States", "line_number": 95, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 97, "usage_type": "call"}, {"api_name": "config.States", "line_number": 97, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard", "line_number": 98, "usage_type": "attribute"}, {"api_name": "status.get_state_game", "line_number": 100, "usage_type": "call"}, {"api_name": "status.get_current_state", "line_number": 103, "usage_type": "call"}, {"api_name": "keyboard.keyboard_correct_answer", "line_number": 104, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 105, "usage_type": "call"}, {"api_name": "config.States", "line_number": 105, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard", "line_number": 108, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 109, "usage_type": "call"}, {"api_name": "config.States", "line_number": 109, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 111, "usage_type": "call"}, {"api_name": "config.States", "line_number": 111, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 119, "usage_type": "call"}, {"api_name": "config.States", "line_number": 119, "usage_type": "attribute"}, {"api_name": "translator.translate_en", "line_number": 121, "usage_type": "call"}, {"api_name": "keyboard.keyboard_after_translation_en", "line_number": 124, "usage_type": "attribute"}, {"api_name": "library_db.add_translation", "line_number": 126, "usage_type": "call"}, {"api_name": "status.get_current_state", "line_number": 117, "usage_type": "call"}, {"api_name": "config.States", "line_number": 117, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 133, "usage_type": "call"}, {"api_name": "config.States", "line_number": 133, "usage_type": "attribute"}, {"api_name": "translator.translate_ru", "line_number": 135, "usage_type": "call"}, {"api_name": "keyboard.keyboard_after_translation_ru", "line_number": 138, "usage_type": "attribute"}, {"api_name": "library_db.add_translation", "line_number": 140, "usage_type": "call"}, {"api_name": "status.set_state", "line_number": 141, "usage_type": "call"}, {"api_name": "config.States", "line_number": 141, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 131, "usage_type": "call"}, {"api_name": "config.States", "line_number": 131, "usage_type": "attribute"}, {"api_name": "library_db.delete_one", "line_number": 148, "usage_type": "call"}, {"api_name": "keyboard.keyboard", "line_number": 149, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 150, "usage_type": "call"}, {"api_name": "config.States", "line_number": 150, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard_error_delete_id", "line_number": 153, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 154, "usage_type": "call"}, {"api_name": "config.States", "line_number": 154, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 145, "usage_type": "call"}, {"api_name": "config.States", "line_number": 145, "usage_type": "attribute"}, {"api_name": "library_db.correct_ru", "line_number": 163, "usage_type": "call"}, {"api_name": "keyboard.keyboard", "line_number": 164, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 165, "usage_type": "call"}, {"api_name": "config.States", "line_number": 165, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 168, "usage_type": "call"}, {"api_name": "config.States", "line_number": 168, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 159, "usage_type": "call"}, {"api_name": "config.States", "line_number": 159, "usage_type": "attribute"}, {"api_name": "library_db.correct_en", "line_number": 177, "usage_type": "call"}, {"api_name": "keyboard.keyboard", "line_number": 178, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 179, "usage_type": "call"}, {"api_name": "config.States", "line_number": 179, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard_error_lib", "line_number": 182, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 183, "usage_type": "call"}, {"api_name": "config.States", "line_number": 183, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 173, "usage_type": "call"}, {"api_name": "config.States", "line_number": 173, "usage_type": "attribute"}, {"api_name": "status.get_current_state", "line_number": 189, "usage_type": "call"}, {"api_name": "keyboard.keyboard_correct_answer", "line_number": 190, "usage_type": "attribute"}, {"api_name": "status.set_state", "line_number": 191, "usage_type": "call"}, {"api_name": "config.States", "line_number": 191, "usage_type": "attribute"}, {"api_name": "keyboard.keyboard_error_game", "line_number": 194, "usage_type": "attribute"}, {"api_name": "status.get_state_game", "line_number": 187, "usage_type": "call"}]} +{"seq_id": "19846731", "text": "from apscheduler.schedulers.background import BackgroundScheduler\nfrom django_apscheduler.jobstores import DjangoJobStore, register_events\nfrom ..python_project.ali_alarm.main import main as cal_kde_value\nfrom ..python_project.ali_alarm.alarm_priority_algorithm1126.SituationOperate import so_run\nfrom ..python_project.ali_alarm.alarm_priority_algorithm1126.alarm_data_regular_filter import his_model_update\nfrom ..python_project.scats_interface.condition_monitor import main as monitor\n# from ..python_project.scats_interface.runStrategicinfo_getdata import thread_creat\n# from ..python_project.scats_interface.scats_5min_volumns import RequestDynaDataFromInt\n# from ..python_project.scats_interface.Request_Data_From_Int import RequestDynaDataFromInt as get_operate\n# from proj.python_project.scats_operate_parsing.seperate_operate_record import main as operate_resolve\nfrom ..config.database import Postgres\nimport cx_Oracle\nimport pandas as pd\nimport numpy as np\n# from proj.config.log_record import LogRecord\nimport logging\nimport datetime as dt\n\nlogger = logging.getLogger('schedulerTask') # 获取settings.py配置文件中logger名称\nfrom ..python_project.scats_operate_parsing import seperate_operate_record\nfrom ..config.sql_text import SqlText\n\n\ndef clear_database():\n pg = Postgres()\n save_date = (dt.datetime.now() - dt.timedelta(days=3)).strftime(\"%Y-%m-%d %H:%M:%S\")\n # pg.execute(SqlText.sql_delete_real_phase.format(save_date))\n pg.execute(SqlText.sql_delete_kde_vaue.format(save_date))\n pg.db_close()\n print(\"数据库清理完成\")\n\ndef time_task(task):\n date = dt.datetime.now()\n print(task, \"I'm a test job1236!\", date)\n\n\nclass CONSTANT:\n IF_DATA_REPAIR = False\n S_REPAIR_DATE = '2018-10-14 20:30:00'\n E_REPAIR_DATE = '2018-10-15 15:00:00'\n request_interval = 300\n group_interval = 200\n TimeDelta = 360\n TimeDelay = 0\n OracleUser = 'enjoyor/admin@33.83.100.139/orcl'\n # OracleUser = 'SIG_OPT_ADMIN/admin@192.168.20.56/orcl'\n pg_inf = {'database': \"superpower\", 'user': \"postgres\", 'password': \"postgres\",\n 'host': \"172.20.251.98\", 'port': \"5432\"}\n\n\ndef CallOracle():\n rs1 = []\n match_records = []\n try: # 数据库连接超时即退出程序\n db = cx_Oracle.connect(CONSTANT.OracleUser)\n cr = db.cursor()\n except cx_Oracle.DatabaseError:\n print('ERROR:数据库连接超时')\n # sys.exit(0)\n else:\n try: # 表名错误或日期错误即退出\n sql1 = \" select * from INTERSECT_INFORMATION order by SITEID \"\n cr.execute(sql1)\n rs1 = cr.fetchall()\n except cx_Oracle.DatabaseError:\n print('ERROR:数据表名输入错误或不存在')\n # sys.exit(0)\n else:\n match_records = pd.DataFrame(rs1)\n # print(match_records)\n match_records.columns = ['SITEID', 'SITENAME', 'REGION']\n finally:\n cr.close()\n db.close()\n return match_records\n\n\ndef get_scats_int():\n def int_grouped(node_num, group):\n node_list = []\n for i in range(int(group)):\n try:\n select_node = node_num[i * CONSTANT.group_interval:(i + 1) * CONSTANT.group_interval]\n except Exception as e:\n select_node = node_num[i * CONSTANT.group_interval:]\n # select_node = node_num[i*100:]\n print(e)\n # print(select_node)\n node_list.append(select_node)\n # print(node_list)\n return node_list\n\n IntersectInfo = CallOracle() # 从数据库读取路口列表\n currenttime = dt.datetime.now()\n if len(IntersectInfo) > 0:\n IntersectIDlist = IntersectInfo['SITEID']\n try:\n conn = cx_Oracle.connect(CONSTANT.OracleUser) # 连接数据库\n except Exception as e:\n print('MainProcess:连接数据库失败', e)\n else:\n cr = conn.cursor() # 建立游标\n try:\n cr.execute(\"SELECT * FROM INT_STR_INPUT order by SITEID\") # 从Oracle中读取数据\n IntStrInput = cr.fetchall()\n except Exception as e:\n print(\"oracle连接失败\", e)\n else:\n conn.commit()\n # print(IntersectIDlist)\n # print(group)\n int_id = np.array(IntersectIDlist).tolist()\n int_num = len(int_id)\n print(\"请求总路口数:\", int_num)\n group = round(int_num / CONSTANT.group_interval, 0) + 1\n int_grouped_data = int_grouped(int_id, group)\n return group, int_grouped_data, IntStrInput\n\n finally:\n cr.close()\n conn.close()\n else:\n print('获取节点列表失败')\n\n\ndef create_scheduler():\n # manage = SchedulerManage()\n scheduler = BackgroundScheduler(daemonic=True)\n scheduler.add_jobstore(DjangoJobStore(), \"default\")\n date = dt.datetime.now()\n # 报警\n scheduler.add_job(cal_kde_value, \"date\", run_date=date, id='alarm_proj', args=[], replace_existing=True)\n scheduler.add_job(his_model_update, \"date\", run_date=date, id='his_model_up', args=[], replace_existing=True)\n scheduler.add_job(seperate_operate_record.main, \"date\", run_date=date, id='operate_parsing', args=[],\n replace_existing=True)\n scheduler.add_job(clear_database, 'cron', hour='16', minute='04', id='clear_database',replace_existing=True )\n # scheduler.add_job(operate_resolve, \"date\", run_date=date, id='alarm_proj', args=[], replace_existing=True)\n # scheduler.add_job(seperate_operate_record.main, \"interval\", minutes=1, id='operate_proj', args=[])\n # scheduler.add_job(time_task, \"interval\", seconds=5, id='mytask2', args=['mytask2',], replace_existing=True)\n scheduler.add_job(so_run, \"interval\", minutes=1, id='operate_match', args=[], replace_existing=True)\n scheduler.add_job(monitor, \"interval\", minutes=15, id='interface_monitor', args=[], replace_existing=True)\n\n\n try:\n group, int_list, scats_input = get_scats_int()\n except Exception as e:\n logger.error(e)\n print(e)\n else:\n logger.info(\"get scats basic inf successfully!\")\n scheduler.add_job(thread_creat, \"interval\", minutes=5, id='scats_salklist', args=[group, int_list, scats_input],\n replace_existing=True)\n scheduler.add_job(RequestDynaDataFromInt, \"interval\", minutes=5, id='scats_volumns', args=[int_list],\n replace_existing=True)\n scheduler.add_job(get_operate, \"interval\", minutes=3, id='scats_operate', args=[],\n replace_existing=True)\n scheduler.start()\n logger.info('start scheduler task')\n print(\"=======================定时任务启动==========================\")\n print(scheduler.get_jobs())\n print(scheduler.state)\n logger.info('start task register,check on admin platform!')\n register_events(scheduler)\n # scheduler.add_executor()\n # return manage\n\n\n# message_queue = queue.Queue(10)\nmanage = create_scheduler()\n\n# def getScheduler(request):\n# json_demo = {'appcode': False, 'result': []}\n# if request.GET:\n# json_demo2 = {'appcode': True, 'result': []}\n#\n# if 'TaskId' in request.GET:\n# task_id = request.GET['TaskId']\n# task_state = SchedulerManage.jobstate.get(task_id)\n# json_demo2['result'].append(task_state)\n# else:\n# json_demo2['appcode'] = False\n# # json_result = json.dumps(json_demo2, ensure_ascii=False)\n# response = JsonResponse(json_demo2, safe=False, json_dumps_params={'ensure_ascii': False})\n# return response\n# else:\n# response = JsonResponse(json_demo, safe=False, json_dumps_params={'ensure_ascii': False})\n# return response\n", "sub_path": "xjc_pyfile/proj_scats/proj/controller/schedular_task.py", "file_name": "schedular_task.py", "file_ext": "py", "file_size_in_byte": 7902, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 19, "usage_type": "call"}, {"api_name": "config.database.Postgres", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 26, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 26, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 26, "usage_type": "call"}, {"api_name": "config.sql_text.SqlText.sql_delete_kde_vaue.format", "line_number": 28, "usage_type": "call"}, {"api_name": "config.sql_text.SqlText.sql_delete_kde_vaue", "line_number": 28, "usage_type": "attribute"}, {"api_name": "config.sql_text.SqlText", "line_number": 28, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 33, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 33, "usage_type": "attribute"}, {"api_name": "cx_Oracle.connect", "line_number": 55, "usage_type": "call"}, {"api_name": "cx_Oracle.DatabaseError", "line_number": 57, "usage_type": "attribute"}, {"api_name": "cx_Oracle.DatabaseError", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 69, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 94, "usage_type": "attribute"}, {"api_name": "cx_Oracle.connect", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "apscheduler.schedulers.background.BackgroundScheduler", "line_number": 128, "usage_type": "call"}, {"api_name": "django_apscheduler.jobstores.DjangoJobStore", "line_number": 129, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 130, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 130, "usage_type": "attribute"}, {"api_name": "python_project.ali_alarm.main.main", "line_number": 132, "usage_type": "argument"}, {"api_name": "python_project.ali_alarm.alarm_priority_algorithm1126.alarm_data_regular_filter.his_model_update", "line_number": 133, "usage_type": "argument"}, {"api_name": "python_project.scats_operate_parsing.seperate_operate_record.main", "line_number": 134, "usage_type": "attribute"}, {"api_name": "python_project.scats_operate_parsing.seperate_operate_record", "line_number": 134, "usage_type": "name"}, {"api_name": "python_project.ali_alarm.alarm_priority_algorithm1126.SituationOperate.so_run", "line_number": 140, "usage_type": "argument"}, {"api_name": "python_project.scats_interface.condition_monitor.main", "line_number": 141, "usage_type": "argument"}, {"api_name": "django_apscheduler.jobstores.register_events", "line_number": 163, "usage_type": "call"}]} +{"seq_id": "400155727", "text": "from __future__ import division\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport torch\nfrom torch.jit import script,trace\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nimport csv\nimport random\nimport re\nimport os\nimport unicodedata\nimport codecs\nfrom io import open\nimport itertools\nimport math\n\nUSE_CUDA = torch.cuda.is_available()\ndevice = torch.device(\"cuda\" if USE_CUDA else \"cpu\")\n\ncorpus_name = \"cornell movie-dialogs corpus/\"\ncorpus = os.path.join('./',corpus_name)\n\n#Print a few lines of the corpus\ndef printLines(file,n=10):\n\twith open(file,'rb') as datafile:\n\t\tlines = datafile.readlines()\n\tfor line in lines[:n]:\n\t\tprint(line)\n\n#printLines(os.path.join(corpus,\"movie_lines.txt\"),10)\n\n#Load lines of text into dictionary\ndef loadLines(fileName,fields):\n\tlines_dict = {}\n\twith open(fileName,'r',encoding='iso-8859-1') as f:\n\t\tfor line in f:\n\t\t\tvalues = line.split(' +++$+++ ')\n\t\t\tlineObj = {}\n\t\t\tfor i,field in enumerate(fields):\n\t\t\t\tlineObj[field]=values[i]\n\t\t\tlines_dict[lineObj['lineID']] = lineObj\n\n\treturn lines_dict\n\n#Group fields of lines into conversations based on movie_conversations.txt\ndef loadConversations(fileName,lines,fields):\n\tconversations = []\n\twith open(fileName,'r',encoding='iso-8859-1') as f:\n\t\tfor line in f:\n\t\t\tvalues = line.split(' +++$+++ ')\n\t\t\tconvObj = {}\n\t\t\tfor i,field in enumerate(fields):\n\t\t\t\tconvObj[field]=values[i]\n\n\t\t\tlineIds = eval(convObj['utteranceIDs'])\n\t\t\tconvObj['lines'] = []\n\t\t\tfor lineId in lineIds:\n\t\t\t\tconvObj['lines'].append(lines[lineId])\n\t\t\tconversations.append(convObj) \n\n\treturn conversations\n\n#Extracting sentence pairs for training\ndef extractSentencePairs(conversations):\n\tqa_pairs = []\n\tfor conversation in conversations:\n\n\t\tfor i in range(len(conversation['lines']) - 1):\n\t\t\tinputLine = conversation['lines'][i]['text'].strip()\n\t\t\toutputLine = conversation['lines'][i+1]['text'].strip()\n\n\t\t\tif inputLine and outputLine:\n\t\t\t\tqa_pairs.append([inputLine,outputLine])\n\n\treturn qa_pairs\n\n\ndatafile = os.path.join(corpus,'formatted_movie_lines.txt')\ndelimiter = '\\t'\n#Unescaping the delimiter\ndelimiter = str(codecs.decode(delimiter,'unicode_escape'))\n\nlines = {}\nconversations = {}\n\nMOVIE_LINES_FIELDS = [\"lineID\", \"characterID\", \"movieID\", \"character\", \"text\"]\nMOVIE_CONVERSATIONS_FIELDS = [\"character1ID\", \"character2ID\", \"movieID\", \"utteranceIDs\"]\n\n'''#Preprocessing\nprint(\"Loading Corpus....\")\nlines = loadLines(os.path.join(corpus,'movie_lines.txt'),MOVIE_LINES_FIELDS)\nprint(\"Loading Conversations....\")\nconversations = loadConversations(os.path.join(corpus,'movie_conversations.txt'),lines,MOVIE_CONVERSATIONS_FIELDS)\nprint(\"Writing to csv file....\")\nwith open(datafile,'w',encoding='utf-8') as outputFile:\n\twriter = csv.writer(outputFile,delimiter=delimiter,lineterminator='\\n')\n\tfor pair in extractSentencePairs(conversations):\n\t\twriter.writerow(pair)\n\nprint(\"Printing sample line from formatted file:\")\nprintLines(datafile)'''\n\n\n\n#Creating class Voc, which adds/removes and stores vocabulary of the dataset\nPAD_TOKEN = 0\nSOS_TOKEN = 1\nEOS_TOKEN = 2\n\nclass Voc:\n\tdef __init__(self,name):\n\t\tself.name = name\n\t\tself.trimmed = False\n\t\tself.word2index = {}\n\t\tself.word2count = {}\n\t\tself.index2word = {PAD_TOKEN : 'PAD',SOS_TOKEN:'SOS',EOS_TOKEN:'EOS'}\n\t\tself.num_words = 3\n\n\tdef addSentence(self,sentence):\n\t\tfor word in sentence.split(' '):\n\t\t\tself.addWord(word)\n\n\tdef addWord(self,word):\n\t\tif word not in self.word2index:\n\t\t\tself.word2index[word] = self.num_words\n\t\t\tself.word2count[word] = 1\n\t\t\tself.index2word[self.num_words] = word\n\t\t\tself.num_words+=1\n\t\telse:\n\t\t\tself.word2count[word]+=1\n\n\tdef trim(self,min_count):\n\t\tif self.trimmed:\n\t\t\treturn\n\t\tself.trimmed=True\n\n\t\tkeep_words = []\n\n\t\tfor k,v in self.word2count.items():\n\t\t\tif v>=min_count:\n\t\t\t\tkeep_words.append(k)\n\n\t\tprint('Keep words {}/{} = {:.4f}'.format(len(keep_words),len(self.word2index),len(keep_words)/len(self.word2index)))\n\n\t\tself.word2index = {}\n\t\tnew_word2count = self.word2count\n\t\tself.index2word = {}\n\t\tself.index2word = {PAD_TOKEN : 'PAD',SOS_TOKEN:'SOS',EOS_TOKEN:'EOS'}\n\t\tself.num_words = 3\n\n\t\tfor word in keep_words:\n\t\t\tself.addWord(word)\n\n\t\tself.word2count = {}\n\t\tfor word in keep_words:\n\t\t\tself.word2count[word] = new_word2count[word]\n\n\n#Functions for preprocessing\nMAX_LENGTH=10\n\ndef unicodeToAscii(s):\n\treturn ''.join(c for c in unicodedata.normalize('NFD',s) if unicodedata.category(c)!='Mn')\n\ndef normalizeString(s):\n\ts = unicodeToAscii(s.lower().strip())\n\ts = re.sub(r\"([.!?])\",r\" \\1\",s)\n\ts = re.sub(r\"[^a-zA-Z.!?]+\", r\" \", s)\n\ts = re.sub(r\"\\s+\", r\" \", s).strip()\n\treturn s\n\n#Read query,response pairs and return a Voc object\ndef readVocs(datafile,corpus_name):\n\tprint(\"Reading lines...\")\n\tlines = open(datafile,encoding='utf-8').read().strip().split('\\n')\n\tpairs = [[normalizeString(s) for s in l.split('\\t')] for l in lines]\n\tvoc = Voc(corpus_name)\n\treturn voc,pairs\n\ndef filterPair(p):\n\treturn len(p[0].split(' ')) < MAX_LENGTH and len(p[1].split(' ')) < MAX_LENGTH\n\ndef filterPairs(pairs):\n\treturn [pair for pair in pairs if filterPair(pair)]\n\ndef loadPrepareData(corpus,corpus_name,datafile,save_dir):\n\tprint(\"Start preparing training data....\")\n\tvoc,pairs = readVocs(datafile,corpus_name)\n\tprint(\"Read {!s} sentence pairs\".format(len(pairs)))\n\tpairs = filterPairs(pairs)\n\tprint(\"Trimmed to {!s} sentence pairs\".format(len(pairs)))\n\tprint(\"Counting words...\")\n\tfor pair in pairs:\n\t\tvoc.addSentence(pair[0])\n\t\tvoc.addSentence(pair[1])\n\tprint(\"Counted words:\",voc.num_words)\n\n\treturn voc,pairs\n\nsave_dir = os.path.join('./','save')\nvoc,pairs = loadPrepareData(corpus,corpus_name,datafile,save_dir)\n#Print some pairs to validate\nprint(\"\\npairs:\")\nfor pair in pairs[:10]:\n\tprint(pair)\n\n#Trimming Rare Words\nMIN_COUNT = 3\n\ndef trimRareWords(voc,pairs,MIN_COUNT):\n\tvoc.trim(MIN_COUNT)\n\tkeep_pairs = []\n\tfor pair in pairs:\n\t\tinput_sentence = pair[0]\n\t\toutput_sentence = pair[1]\n\t\tkeep_input = True\n\t\tkeep_output = True\n\n\t\tfor word in input_sentence.split(' '):\n\t\t\tif word not in voc.word2index:\n\t\t\t\tkeep_input = False\n\t\t\t\tbreak\n\t\tfor word in output_sentence.split(' '):\n\t\t\tif word not in voc.word2index:\n\t\t\t\tkeep_input = False\n\t\t\t\tbreak\n\n\t\tif keep_input and keep_output:\n\t\t\tkeep_pairs.append(pair)\n\n\tprint(\"Trimmed from {} pairs to {}, {:.4f} of total\".format(len(pairs), len(keep_pairs), len(keep_pairs) / len(pairs)))\n\n\treturn keep_pairs\n\npairs = trimRareWords(voc,pairs,MIN_COUNT)\n\ndef indexesFromSentences(voc,sentence):\n\treturn [voc.word2index[word] for word in sentence.split(' ')] + [EOS_TOKEN]\n\ndef zeroPadding(l,fillvalue=PAD_TOKEN):\n\t# zip_longest('ABCD', 'xy', fillvalue='-') --> Ax By C- D-\n\treturn list(itertools.zip_longest(*l,fillvalue=fillvalue))\n\n#Creating binary mask matrix, 0 for pad_token else 1\ndef binaryMatrix(l,value=PAD_TOKEN):\n\tm = []\n\tfor i,seq in enumerate(l):\n\t\tm.append([])\n\t\tfor token in seq:\n\t\t\tif token == PAD_TOKEN:\n\t\t\t\tm[i].append(0)\n\t\t\telse:\n\t\t\t\tm[i].append(1)\n\n\treturn m\n\n# Returns padded input sequence tensor and lengths\ndef inputVar(l,voc):\n\tindexes_batch = [indexesFromSentences(voc,sentence) for sentence in l]\n\tlengths = torch.tensor([len(indexes) for indexes in indexes_batch])\n\tpadList = zeroPadding(indexes_batch)\n\tpadVar = torch.LongTensor(padList)\n\n\treturn padVar,lengths\n\n# Returns padded target sequence tensor, padding mask, and max target length\ndef outputVar(l,voc):\n\tindexes_batch = [indexesFromSentences(voc,sentence) for sentence in l]\n\tmax_target_len = max([len(indexes) for indexes in indexes_batch])\n\tpadList = zeroPadding(indexes_batch)\n\tmask = binaryMatrix(padList)\n\tmask = torch.ByteTensor(mask)\n\tpadVar = torch.LongTensor(padList)\n\n\treturn padVar,mask,max_target_len\n\n# Returns all items for a given batch of pairs\ndef batch2TrainData(voc,pair_batch):\n\tpair_batch.sort(key=lambda x:len(x[0].split(\" \")),reverse=True)\n\tinput_batch,output_batch = [],[]\n\tfor pair in pair_batch:\n\t\tinput_batch.append(pair[0])\n\t\toutput_batch.append(pair[1])\n\tinp,lengths = inputVar(input_batch,voc)\n\tout,mask,max_target_len = outputVar(output_batch,voc)\n\n\treturn inp,lengths,out,mask,max_target_len\n\n#Examples for validation\nsmall_batch_size = 5\nbatches = batch2TrainData(voc,[random.choice(pairs) for _ in range(small_batch_size)])\ninput_variable, lengths, target_variable, mask, max_target_len = batches\n\nprint(\"input_variable:\", input_variable)\nprint(\"lengths:\", lengths)\nprint(\"target_variable:\", target_variable)\nprint(\"mask:\", mask)\nprint(\"max_target_len:\", max_target_len)\n\n\n#Seq2Seq Model\n\n#Encoder \nclass EncoderRNN(nn.Module):\n\tdef __init__(self,hidden_size,embedding,n_layers=1,dropout=0):\n\t\tsuper(EncoderRNN,self).__init__()\n\t\tself.n_layers = n_layers\n\t\tself.hidden_size = hidden_size\n\t\tself.embedding = embedding\n\n\t\tself.gru = nn.GRU(hidden_size,hidden_size,n_layers,dropout=(0 if n_layers==1 else dropout),bidirectional=True)\n\n\tdef forward(self,input_seq,input_lengths,hidden=None):\n\t\t#Convert word indexes to embeddings\n\t\tembedded = self.embedding(input_seq)\n\t\t#Pack padded batch of sequences for RNN Module\n\t\tpacked = nn.utils.rnn.pack_padded_sequence(embedded,input_lengths)\n\t\t#Forward pass through GRU\n\t\toutputs,hidden = self.gru(packed,hidden)\n\t\t#Unpack padding\n\t\toutputs,_ = nn.utils.rnn.pad_packed_sequence(outputs)\n\t\t#Sum bidirectional GRU outputs\n\t\toutputs = outputs[:,:,:self.hidden_size] + outputs[:,:,self.hidden_size:]\n\n\t\treturn outputs,hidden\n\n#Luong Attention Layer\nclass Attn(nn.Module):\n\tdef __init__(self,method,hidden_size):\n\t\tsuper(Attn,self).__init__()\n\t\tself.method = method\n\t\tif self.method not in ['dot','general','concat']:\n\t\t\traise ValueError(self.method,\" is not an appropriate attention method\")\n\t\tself.hidden_size = hidden_size\n\t\tif self.method == 'general':\n\t\t\tself.attn = nn.Linear(self.hidden_size,hidden_size)\n\t\telif self.method == 'concat':\n\t\t\tself.attn = nn.Linear(self.hidden_size*2,hidden_size)\n\t\t\tself.v = nn.Parameter(torch.floatTensor(hidden_size))\n\n\tdef dot_score(self,hidden,encoder_output):\n\t\treturn torch.sum(hidden*encoder_output,dim=2)\n\n\tdef general_score(self,hidden,encoder_output):\n\t\tenergy = self.attn(encoder_output)\n\t\treturn torch.sum(hidden*energy,dim=2)\n\n\tdef concat_score(self,hidden,encoder_output):\n\t\tenergy = self.attn(torch.cat((hidden.expand(encoder_output.size(0),-1,-1),encoder_output),2)).tanh()\n\t\treturn torch.sum(self.v*energy,dim=2)\n\n\tdef forward(self,hidden,encoder_output):\n\t\tif self.method == 'general':\n\t\t\tattn_energies = self.general_score(hidden,encoder_output)\n\t\telif self.method == 'concat':\n\t\t\tattn_energies = self.concat_score(hidden,encoder_output)\n\t\telse:\n\t\t\tattn_energies = self.dot_score(hidden,encoder_output)\n\n\t\tattn_energies = attn_energies.t()\n\n\t\treturn F.softmax(attn_energies,dim=1).unsqueeze(1)\n\n#Decoder\nclass LuongAttnDecoderRNN(nn.Module):\n\tdef __init__(self,attn_model,embedding,hidden_size,output_size,n_layers=1,dropout=0):\n\t\tsuper(LuongAttnDecoderRNN,self).__init__()\n\t\tself.attn_model = attn_model\n\t\tself.embedding = embedding\n\t\tself.hidden_size = hidden_size\n\t\tself.output_size = output_size\n\t\tself.n_layers = n_layers\n\t\tself.dropout = dropout\n\n\t\tself.embedding_dropout = nn.Dropout(dropout)\n\t\tself.gru = nn.GRU(hidden_size,hidden_size,n_layers,dropout=(0 if n_layers==1 else dropout))\n\t\tself.concat = nn.Linear(hidden_size*2,hidden_size)\n\t\tself.out = nn.Linear(hidden_size,output_size)\n\n\t\tself.attn = Attn(attn_model,hidden_size)\n\n\tdef forward(self,input_step,last_hidden,encoder_outputs):\n\t\t#Embedding of current input word\n\t\tembedded = self.embedding(input_step)\n\t\tembedded = self.embedding_dropout(embedded)\n\n\t\trnn_output,hidden = self.gru(embedded,last_hidden)\n\n\t\tattn_weights = self.attn(rnn_output,encoder_outputs)\n\t\t# Multiply attention weights to encoder outputs to get new \"weighted sum\" context vector\n\t\tcontext = attn_weights.bmm(encoder_outputs.transpose(0,1))\n\t\t#Concatenate weighted context vector and GRU output using Luong eq 5\n\t\trnn_output = rnn_output.squeeze(0)\n\t\tcontext = context.squeeze(1)\n\t\tconcat_input = torch.cat((rnn_output,context),1)\n\t\tconcat_output = torch.tanh(self.concat(concat_input))\n\t\t#Predict next word Luong eq 6\n\t\toutput = self.out(concat_output)\n\t\toutput = F.softmax(output,dim=1)\n\n\t\treturn output,hidden\n\n\n#Loss only incorporates non-padded tokens in loss calculation, using the mask binary matrix\ndef maskedNLLLoss(inp,targets,mask):\n\tnTotal = mask.sum()\n\tcrossEntropy = -torch.log(torch.gather(inp,1,targets.view(-1,1)).squeeze(1))\n\tloss = crossEntropy.masked_select(mask).mean()\n\tloss = loss.to(device)\n\treturn loss,nTotal.item()\n\n#Training procedure\ndef train(input_variable,lengths,target_variable,mask,max_target_len,encoder,decoder,embedding,encoder_optimizer,\n\tdecoder_optimizer,batch_size,clip,teacher_forcing_ratio,max_length=MAX_LENGTH):\n\n\t#Zero gradients\n\tencoder_optimizer.zero_grad()\n\tdecoder_optimizer.zero_grad()\n\n\t#Set device options\n\tinput_variable = input_variable.to(device)\n\tlengths = lengths.to(device)\n\ttarget_variable = target_variable.to(device)\n\tmask = mask.to(device)\n\n\t#Initialize\tvariables\n\tloss = 0\n\tprint_losses = []\n\tn_totals = 0\n\n\t#Forward pass through encoder\n\tencoder_outputs,encoder_hidden = encoder(input_variable,lengths)\n\n\t#Create initial decoder input(starting with SOS_tokens)\n\tdecoder_input = torch.LongTensor([[SOS_TOKEN for _ in range(batch_size)]])\n\tdecoder_input = decoder_input.to(device)\n\n\t#Set initial decoder state to encoder's final hidden state\n\tdecoder_hidden = encoder_hidden[:decoder.n_layers]\n\n\t#Determine if we are using teacher forcing this iteration\n\tuse_teacher_forcing = True if random.random() < teacher_forcing_ratio else False\n\n\t#Forward batch of sequences through decoder one step at a time\n\tif use_teacher_forcing:\n\t\tfor t in range(max_target_len):\n\t\t\tdecoder_output,decoder_hidden = decoder(decoder_input,decoder_hidden,encoder_outputs)\n\t\t\t#Teacher forcing part, next input is current target\n\t\t\tdecoder_input = target_variable[t].view(1,-1)\n\t\t\t#Calculate and accumulate loss\n\t\t\tmask_loss,nTotal = maskedNLLLoss(decoder_output,target_variable[t],mask[t])\n\t\t\tloss+=mask_loss\n\t\t\tprint_losses.append(mask_loss.item()*nTotal)\n\t\t\tn_totals = nTotal\n\telse:\n\t\tfor t in range(max_target_len):\n\t\t\tdecoder_output,decoder_hidden = decoder(decoder_input,decoder_hidden,encoder_outputs)\n\t\t\t#Normal pass without teacher forcing\n\t\t\t_,topi = decoder_output.topk(1)\n\t\t\tdecoder_input = torch.LongTensor([[topi[i][0] for i in range(batch_size)]])\n\t\t\tdecoder_input = decoder_input.to(device)\n\t\t\tmask_loss,nTotal = maskedNLLLoss(decoder_output,target_variable[t],mask[t])\n\t\t\tloss+=mask_loss\n\t\t\tprint_losses.append(mask_loss.item()*nTotal)\n\t\t\tn_totals = nTotal\n\n\t#Backprop\n\tloss.backward()\n\n\t#Clipping gradients\n\t_ = nn.utils.clip_grad_norm_(encoder.parameters(),clip)\n\t_ = nn.utils.clip_grad_norm_(decoder.parameters(),clip)\n\n\t#Adjust model weights\n\tencoder_optimizer.step()\n\tdecoder_optimizer.step()\n\n\treturn sum(print_losses)/n_totals\n\n#Training function for n iterations\ndef trainIters(model_name,voc,pairs,encoder,decoder,encoder_optimizer,decoder_optimizer,embedding,\n\tencoder_n_layers,decoder_n_layers,save_dir,n_iteration,batch_size,print_every,save_every,clip,\n\tcorpus_name,loadFileName,teacher_forcing_ratio,checkpoint=None):\n\n\t#Load batches for each iteration\n\ttraining_batches = [batch2TrainData(voc,[random.choice(pairs) for _ in range(batch_size)]) for _ in range(n_iteration)]\n\t#Initializing\n\tprint(\"Initializing...\")\n\tstart_iteration = 1\n\tprint_loss = 0\n\tif loadFileName:\n\t\tstart_iteration = checkpoint['iteration']+1\n\n\t#Training loop\n\tprint(\"Training...\")\n\tfor iteration in range(start_iteration,n_iteration+1):\n\t\ttraining_batch = training_batches[iteration-1]\n\t\t#Extract fields from batch\n\t\tinput_variable,lengths,target_variable,mask,max_target_len = training_batch\n\n\t\t#Run a training iteration with batch\n\t\tloss = train(input_variable,lengths,target_variable,mask,max_target_len,encoder,decoder,embedding,encoder_optimizer,\n\t\t\tdecoder_optimizer,batch_size,clip,teacher_forcing_ratio)\n\n\t\tprint_loss+=loss\n\n\t\t#Print Progress\n\t\tif iteration%print_every==0:\n\t\t\tprint_loss_avg = print_loss/print_every\n\t\t\tprint(\"Iteration {}; Percent complete: {:.1f}%,Average Loss:{:.4f}\".format(iteration,iteration/n_iteration*100,print_loss_avg))\n\t\t\tprint_loss = 0\n\n\t\tif (iteration%save_every==0):\n\t\t\tdirectory = os.path.join(save_dir, model_name, corpus_name, '{}-{}_{}'.format(encoder_n_layers, decoder_n_layers, hidden_size))\n\t\t\tif not os.path.exists(directory):\n\t\t\t\tos.makedirs(directory)\n\t\t\ttorch.save({\n\t\t\t 'iteration': iteration,\n\t\t\t 'en': encoder.state_dict(),\n\t\t\t 'de': decoder.state_dict(),\n\t\t\t 'en_opt': encoder_optimizer.state_dict(),\n\t\t\t 'de_opt': decoder_optimizer.state_dict(),\n\t\t\t 'loss': loss,\n\t\t\t 'voc_dict': voc.__dict__,\n\t\t\t 'embedding': embedding.state_dict()\n\t\t\t}, os.path.join(directory, '{}_{}.tar'.format(iteration, 'checkpoint')))\n\n\n#Decoder used for inference stage\nclass GreedySearchDecoder(nn.Module):\n\tdef __init__(self,encoder,decoder):\n\t\tsuper(GreedySearchDecoder,self).__init__()\n\t\tself.encoder = encoder\n\t\tself.decoder = decoder\n\n\tdef forward(self,input_seq,input_length,max_length):\n\t\t#Forward input through encoder model\n\t\tencoder_outputs,encoder_hidden = self.encoder(input_seq,input_length)\n\t\t#Prepare encoder's final hidden layer to be first hidden input to the decoder\n\t\tdecoder_hidden = encoder_hidden[:decoder.n_layers]\n\t\t#Initialize decoder input with SOS_TOKEN\n\t\tdecoder_input = torch.ones(1,1,device=device,dtype=torch.long)*SOS_TOKEN\n\t\t#Initialize tensors to append decoder words to\n\t\tall_tokens = torch.zeros([0],device=device,dtype=torch.long)\n\t\tall_scores = torch.zeros([0],device=device)\n\t\t#Iteratively decode one word at a time\n\t\tfor _ in range(max_length):\n\t\t\t#Forward pass through decoder\n\t\t\tdecoder_output,decoder_hidden = self.decoder(decoder_input,decoder_hidden,encoder_outputs)\n\t\t\t#Obtain most likely word token and its softmax score\n\t\t\tdecoder_scores,decoder_input = torch.max(decoder_output,dim=1)\n\t\t\t#Record tokens and score\n\t\t\tall_tokens = torch.cat((all_tokens,decoder_input),dim=0)\n\t\t\tall_scores = torch.cat((all_scores,decoder_scores),dim=0)\n\n\t\t\t#Prepare current token to be the next input (add a dimension)\n\t\t\tdecoder_input = torch.unsqueeze(decoder_input,0)\n\n\t\t#Return collections of word tokens and scores\n\t\treturn all_tokens,all_scores\n\ndef evaluate(encoder,decoder,searcher,voc,sentence,max_length=MAX_LENGTH):\n\t#Formatting input sentence as a batch\n\t#words->indexes\n\n\tindexes_batch = [indexesFromSentences(voc,sentence)]\n\t#Create lengths tensor\n\tlengths = torch.Tensor([len(indexes) for indexes in indexes_batch])\n\t#Transpose dimensions of batch to match models' expectations\n\tinput_batch = torch.LongTensor(indexes_batch).transpose(0,1)\n\t#Use appropriate device\n\tinput_batch = input_batch.to(device)\n\t#Decode sentences with searcher\n\ttokens,scores = searcher(input_batch,lengths,max_length)\n\t#indexes->words\n\tdecoded_words = [voc.index2word[token.item()] for token in tokens]\n\n\treturn decoded_words\n\ndef evaluateInput(encoder,decoder,searcher,voc):\n\tinput_sentence = ''\n\twhile(1):\n\t\ttry:\n\t\t\tinput_sentence = input('> ')\n\t\t\tif input_sentence == 'q' or input_sentence == 'quit': break\n\t\t\tinput_sentence = normalizeString(input_sentence)\n\t\t\toutput_words = evaluate(encoder,decoder,searcher,voc,input_sentence)\n\t\t\toutput_words[:] = [x for x in output_words if not (x=='EOS' or x=='PAD')]\n\t\t\tprint(\"Bot:\",' '.join(output_words))\n\n\t\texcept KeyError:\n\t\t\tprint(\"Error: Encountered unknown word\")\n\n\n#Configure model\nmodel_name = 'cb_model'\nattn_model = 'dot'\nhidden_size = 500\nencoder_n_layers = 2\ndecoder_n_layers = 2\ndropout = 0.1\nbatch_size = 64\n\n#Set checkpoint to load from;set to None if starting from Scratch\n#loadFileName = None\ncheckpoint_iter = 4000\nloadFileName = os.path.join(save_dir, model_name, corpus_name,\n '{}-{}_{}'.format(encoder_n_layers, decoder_n_layers, hidden_size),\n '{}_checkpoint.tar'.format(checkpoint_iter))\n\n#Load model if a loadFileName is provided\nif loadFileName:\n\t# If loading on same machine the model was trained on\n\tcheckpoint = torch.load(loadFileName)\n\t# If loading a model trained on GPU to CPU\n #checkpoint = torch.load(loadFilename, map_location=torch.device('cpu'))\n\tencoder['sd'] = checkpoint['en']\n\tdecoder['sd'] = checkpoint['de']\n\tencoder_optimizer = checkpoint['en_opt']\n\tdecoder_optimizer = checkpoint['de_opt']\n\tembedding_sd = checkpoint['embedding']\n\tvoc.__dict__ = checkpoint['voc_dict']\n\nprint('Building encoder and decoder...')\n#Initializing word embeddings\nembedding = nn.Embedding(voc.num_words,hidden_size)\nif loadFileName:\n\tembedding.load_state_dict(embedding_sd)\n#Initializing encoder and decoder models\nencoder = EncoderRNN(hidden_size,embedding,encoder_n_layers,dropout)\ndecoder = LuongAttnDecoderRNN(attn_model,embedding,hidden_size,voc.num_words,decoder_n_layers,dropout)\nif loadFileName:\n\tencoder.load_state_dict(encoder_sd)\n\tdecoder.load_state_dict(decoder_sd)\n\nencoder = encoder.to(device)\ndecoder = decoder.to(device)\nprint(\"Models built and ready to go...\")\n\n\n#Configure training/optimization\nclip=50.0\nteacher_forcing_ratio = 1.0\nlearning_rate = 0.0001\ndecoder_learning_ratio = 5.0\nn_iteration = 4000\nprint_every = 100\nsave_every = 500\n\n#Ensure dropout layers are in train mode\nencoder.train()\ndecoder.train()\n\n#Initializing optimizers\nprint(\"Building optimizers...\")\nencoder_optimizer = optim.Adam(encoder.parameters(),lr=learning_rate)\ndecoder_optimizer = optim.Adam(decoder.parameters(),lr=learning_rate*decoder_learning_ratio)\nif loadFileName:\n\tencoder_optimizer.load_state_dict(encoder_optimizer_sd)\n\tdecoder_optimizer.load_state_dict(decoder_optimizer_sd)\n\nprint(\"Started training!\")\nif not loadFileName:\n\tcheckpoint=None\ntrainIters(model_name,voc,pairs,encoder,decoder,encoder_optimizer,decoder_optimizer,embedding,\n\tencoder_n_layers,decoder_n_layers,save_dir,n_iteration,batch_size,print_every,save_every,clip,corpus_name,\n\tloadFileName,teacher_forcing_ratio,checkpoint)\n\n\n# Set dropout layers to eval mode\nencoder.eval()\ndecoder.eval()\n\n# Initialize search module\nsearcher = GreedySearchDecoder(encoder, decoder)\n\n# Begin chatting (uncomment and run the following line to begin)\nevaluateInput(encoder, decoder, searcher, voc)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n", "sub_path": "chatbot_tut.py", "file_name": "chatbot_tut.py", "file_ext": "py", "file_size_in_byte": 22183, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.cuda.is_available", "line_number": 21, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 21, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "io.open", "line_number": 29, "usage_type": "call"}, {"api_name": "io.open", "line_number": 39, "usage_type": "call"}, {"api_name": "io.open", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "codecs.decode", "line_number": 85, "usage_type": "call"}, {"api_name": "unicodedata.normalize", "line_number": 167, "usage_type": "call"}, {"api_name": "unicodedata.category", "line_number": 167, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 171, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 172, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 173, "usage_type": "call"}, {"api_name": "io.open", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}, {"api_name": "itertools.zip_longest", "line_number": 246, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 266, "usage_type": "call"}, {"api_name": "torch.ByteTensor", "line_number": 276, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 277, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 308, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 308, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 315, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 315, "usage_type": "name"}, {"api_name": "torch.nn.utils.rnn.pack_padded_sequence", "line_number": 321, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 321, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 321, "usage_type": "name"}, {"api_name": "torch.nn.utils.rnn.pad_packed_sequence", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 325, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 325, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 332, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 332, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 340, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 340, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 342, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 342, "usage_type": "name"}, {"api_name": "torch.nn.Parameter", "line_number": 343, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 343, "usage_type": "name"}, {"api_name": "torch.floatTensor", "line_number": 343, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 346, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 350, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 353, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 354, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 366, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 366, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 369, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 369, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 379, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 379, "usage_type": "name"}, {"api_name": "torch.nn.GRU", "line_number": 380, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 380, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 381, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 381, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 382, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 382, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 399, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 400, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 403, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 403, "usage_type": "name"}, {"api_name": "torch.log", "line_number": 411, "usage_type": "call"}, {"api_name": "torch.gather", "line_number": 411, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 439, "usage_type": "call"}, {"api_name": "random.random", "line_number": 446, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 464, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 475, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 475, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 475, "usage_type": "name"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 476, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 476, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 476, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 490, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 518, "usage_type": "call"}, {"api_name": "os.path", "line_number": 518, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 519, "usage_type": "call"}, {"api_name": "os.path", "line_number": 519, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 520, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 521, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 530, "usage_type": "call"}, {"api_name": "os.path", "line_number": 530, "usage_type": "attribute"}, {"api_name": "torch.nn.Module", "line_number": 534, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 534, "usage_type": "name"}, {"api_name": "torch.ones", "line_number": 546, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 546, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 548, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 548, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 549, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 555, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 557, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 558, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 561, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 572, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 574, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 611, "usage_type": "call"}, {"api_name": "os.path", "line_number": 611, "usage_type": "attribute"}, {"api_name": "torch.load", "line_number": 618, "usage_type": "call"}, {"api_name": "torch.nn.Embedding", "line_number": 630, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 630, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 660, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 660, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 661, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 661, "usage_type": "name"}]} +{"seq_id": "199797296", "text": "import numpy as np\nimport matplotlib.pyplot as plt \nimport matplotlib.patches as patches\nfrom matplotlib import style \nimport seaborn as sns\nimport pandas as pd\nimport csv\n\nstyle.use('ggplot')\n#sensors_1536745135536\nxaxis,x, y, z, x2, y2, z2, latitude, longitude, speed, anomaly = np.loadtxt('Data4\\sensors_1539148085752.csv', delimiter = ';',unpack=True)\n\nthreshcheck = 5.6\nthreshdiff = 5.5\nthreshstd = 2\nDetectedLat = []\nDetectedLong = []\n\ndef graph():\n i=0\n plt.figure(1)\n plt.ylim(-9.8, 9.8)\n \n \n \n plt.plot(xaxis,z, 'b',label='Transformed Z-axis Linear Acceleration')\n\n plt.legend(loc = 2) \n plt.ylabel('m/s2')\n plt.xlabel('timestamps')\n\n \n\n #while i < len(xaxis):\n #plt.axvline(xaxis[i],color= 'g', linestyle= 'dashed')\n #i= i+200\n #plt.axvline(xaxis[len(xaxis)-1],color= 'g', linestyle= 'dashed')\n\n plt.title('Linear Acceleration Data Collected with Road Surface Data Collection Application')\n #plt.show()\n\ndef GraphLocations():\n plt.figure(2)\n plt.ylabel('Longitude')\n plt.xlabel('Latitude')\n plt.title('Location of Road Anomalies')\n\n i=0\n plt.scatter(-33.958141, 18.464571,color='b')\n plt.scatter(-33.958080, 18.464573,color='b')\n plt.scatter(-33.957740, 18.464612,color='b')\n plt.scatter( -33.956755, 18.464706,color='b')\n plt.scatter(-33.955637, 18.464686,color='b')\n plt.scatter(-33.955687, 18.464627,color='r')\n \n\n while i < len(DetectedLat):\n if(DetectedLat[i]!=0 and i !=6 ):\n plt.scatter(DetectedLat[i],DetectedLong[i],color='r')\n print(DetectedLat[i], \" \",DetectedLong[i])\n i= i+1\n plt.show()\n\ndef checkthresh():\n detected1=0\n i=0\n z2 = []\n sec=0\n found =0\n graph() #Graphing Function\n while i < len(z): \n if(abs(z[i])>=threshcheck): #Performs check if Acceleration above Threshold\n z2.append(z[i]) \n print(z[i])\n plt.scatter(xaxis[i],z[i],color='r')\n detected1 = detected1 + 1\n if sec == 0: #Method to Plot a 1 second window around detected anomaly\n found = 1\n DetectedLat.append(latitude[i])\n DetectedLong.append(longitude[i])\n plt.axvline(xaxis[i],color= 'g', linestyle= 'dashed')\n sec=sec+1\n else:\n z2.append(None)\n if found == 1:\n if sec == 200:\n plt.axvline(xaxis[i],color= 'g', linestyle= 'dashed')\n sec = 0\n found = 0 \n else: \n sec=sec+1 \n i=i+1\n plt.plot(xaxis,z2,color='r') # Plots the parts of the grpah where the acceleration exceeds threshold.\n plt.axhline(y=threshcheck,color= 'r', linestyle= 'dashed') # Provides the threshold level on the graph.\n plt.axhline(y=-threshcheck,color= 'r', linestyle= 'dashed')\n plt.show()\n GraphLocations()\n\n\ndef diffcheck():\n detected2=0\n k=0\n c=0\n z3 = []\n sec=0\n found =0\n graph() #Graphing Function\n\n while k < len(z)-1:\n diff = abs(z[k]-z[k+1]) # Computes the Difference between consecutive readings.\n if(diff>=threshdiff): #Performs check if difference above Difference Threshold.\n detected2 = detected2+1\n if sec == 0: #Method to Plot a 1 second window around detected anomaly\n DetectedLat.append(latitude[k])\n DetectedLong.append(longitude[k])\n found = 1\n plt.axvline(xaxis[k],color= 'g', linestyle= 'dashed')\n sec=sec+1\n if z[k] not in z3:\n z3.append(z[k])\n if z[k+1] not in z3:\n z3.append(z[k+1]) # Appending the readings if the difference exceeds Difference Threshold.\n elif z[k] not in z3:\n z3.append(None)\n if found == 1:\n if sec == 200:\n plt.axvline(xaxis[k],color= 'g', linestyle= 'dashed')\n sec = 0\n found = 0 \n else: \n sec=sec+1 \n k += 1\n \n while len(z3) != len(xaxis):\n z3.append(None) \n\n plt.plot(xaxis,z3,color='r') #Plots the readings that exceeds difference threshold in red. \n plt.show()\n GraphLocations()\n\n\n\ndef stdDevcheck(): \n j=0\n detected3=0\n z4 =[]\n o=0\n graph()\n while j < len(xaxis):\n b = z[j:j+40].std() #Calculate the standard deviation for the specified window size.\n #plt.axvline(xaxis[j],color= 'r', linestyle= 'dashed') #Plotting the window barriers.\n if(abs(b)>=threshstd): #Check if standard deviation for the window exceeds the threshold.\n z4.extend(z[j:j+40]) #Add the window to a separate array.\n detected3 = detected3 + 1\n DetectedLat.append(latitude[j])\n DetectedLong.append(longitude[j])\n #print(b)\n else:\n o=0\n while o < 40 and len(z4) random point\r\n #type = 1 -> dataset point\r\n if type == 0:\r\n return math.sqrt(math.pow(self.x - other.x, 2) + math.pow(self.y - other.y, 2))\r\n elif type == 1:\r\n s = 0\r\n for i in range(len(self.attributes)):\r\n s += (self.attributes[i]-other.attributes[i])**2\r\n return s**0.5\r\n\r\n\r\n #def distance_based_similairty(self,other):\r\n # d = sum([(self.attributes[i]-other.attributes[i])**2 for i in range(len(self.attributes))])**0.5\r\n #\r\n # return sim * 2 - 1\r\n\r\n def similairty(self,other,attribute_ranges):\r\n #ratio_list = [min(self.attributes[i],other.attributes[i])/max(self.attributes[i],other.attributes[i]) for i in range(len(self.attributes))]\r\n ratio_list = [abs(self.attributes[i]-other.attributes[i]) / attribute_ranges[i] for i in range(len(self.attributes))]\r\n ratio_list = [(-2)*v+1 for v in ratio_list]\r\n print(\"ratio: \",ratio_list)\r\n #sim = sum(ratio_list)/len(ratio_list)\r\n sim = sum(ratio_list)/len(ratio_list)\r\n return sim\r\n\r\n def __hash__(self):\r\n return self.id\r\n\r\n def __eq__(self, other):\r\n\r\n if isinstance(other,self.__class__) and other.id == self.id:\r\n return True\r\n return False\r\n\r\n\r\n def __ne__(self, other):\r\n return not self.__eq__(other)", "sub_path": "EntityPoint.py", "file_name": "EntityPoint.py", "file_ext": "py", "file_size_in_byte": 2990, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PyQt5.QtWidgets.QGraphicsItem", "line_number": 8, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QGraphicsItem.ItemIsSelectable", "line_number": 11, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QGraphicsItem", "line_number": 11, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 12, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 17, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.black", "line_number": 17, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 17, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QPen", "line_number": 18, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt.black", "line_number": 18, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 18, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 43, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 47, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 57, "usage_type": "call"}, {"api_name": "math.pow", "line_number": 57, "usage_type": "call"}]} +{"seq_id": "384797628", "text": "import requests\nimport pandas as pd\nimport glob\nimport re\nimport numpy as np\nimport os\nfrom requests import get\nfrom requests.exceptions import RequestException\nfrom contextlib import closing\nfrom bs4 import BeautifulSoup\nfrom multiprocessing import Pool\nimport datetime\nfrom sqlalchemy import MetaData\nfrom sqlalchemy.dialects.postgresql import insert\nfrom sqlalchemy import create_engine\nimport sqlite3\n\n\nengine = create_engine('postgresql+psycopg2://postgres:postgres@localhost:5432/hydrodatahub',\n echo=False, use_batch_mode=True)\n\n\nshp_path = r'H:\\Projets_communs\\2019\\SH-XX-XX-XX-HYD-CRU-FREQ-LaGrande\\01_Intrants\\06_Données_physio\\shp'\n\n\ndef hydat_daily2(df, get_flow, to_plot):\n # Deux ou plus stations\n # station_number = \"'02TC001' OR STATION_NUMBER='02TC002'\"\n # station1 = \"'02TC001'\"\n # station2 = \"'02TC002'\"\n # qstr = \"SELECT * FROM DLY_FLOWS WHERE STATION_NUMBER=%s OR STATION_NUMBER=%s\\n\" %(station1, station2)\n # qstr = SELECT * FROM DLY_FLOWS WHERE STATION_NUMBER='02TC001' OR STATION_NUMBER='02TC002'\n # ATTENTION PEUT ETRE TRES LONG\n\n # print(df)\n # if get_flow == True:\n # header = \"^FLOW\\\\d+\"\n # else:\n # header = \"^LEVEL\\\\d+\"\n header = \"^FLOW\\\\d+\" if get_flow == True else \"^LEVEL\\\\d+\"\n\n dly = df[[\"STATION_NUMBER\", \"YEAR\", \"MONTH\"]]\n dly.columns = [\"STATION_NUMBER\", \"YEAR\", \"MONTH\"]\n dly\n # value.cols = df.columns[df.filter(regex=\"^FLOW\\\\d+\")]\n # filter sur les FLOW\n value = df.filter(regex=header)\n valuecols = value.columns\n # print(dlydata.shape)\n # now melt the data frame for data and flags\n dlydata = pd.melt(df, id_vars=[\"STATION_NUMBER\", \"YEAR\", \"MONTH\"], value_vars=valuecols)\n\n if get_flow is True:\n dlydata[\"DAY\"] = dlydata['variable'].apply(lambda x: np.int8(x[4:]))\n else:\n dlydata[\"DAY\"] = dlydata['variable'].apply(lambda x: np.int8(x[5:]))\n # flowvariable = dlydata[\"variable\"]\n # days = [x[4:6] for x in flowvariable]\n # dlydata[\"DAY\"] = list(map(int, days))\n # censor ambiguous dates (e.g., 31st day for Apr, Jun, Sept, Nov)\n d = dlydata.loc[dlydata[\"MONTH\"].isin([4, 6, 9, 11]) & (dlydata[\"DAY\"] > 30)]\n d30 = d\n # print(d.index[:])\n # print(len(d))#\n\n if len(d) > 0:\n dlydata = dlydata.drop(d.index).reset_index(drop=True)\n # print(dlydata.shape)\n\n d = dlydata.loc[(dlydata[\"MONTH\"].isin([2]) &\n pd.to_datetime(dlydata[\"YEAR\"], format='%Y').dt.is_leap_year &\n (dlydata[\"DAY\"] > 29))]\n if len(d) > 0:\n dlydata = dlydata.drop(d.index).reset_index(drop=True)\n d29 = d\n # print(dlydata.shape)\n\n d = dlydata.loc[(dlydata[\"MONTH\"].isin([2]) &\n ~pd.to_datetime(dlydata[\"YEAR\"], format='%Y').dt.is_leap_year.values &\n (dlydata[\"DAY\"] > 28))]\n # print(d)\n if len(d) > 0:\n dlydata = dlydata.drop(d.index).reset_index(drop=True)\n d28 = d\n # print(dlydata.shape)\n # print(valuecols)\n\n # ----------------------------------SYMBOL--------------------------------------------------\n header_sym = \"^FLOW_SYMBOL\\\\d+\" if get_flow == True else \"^LEVEL_SYMBOL\\\\d+\"\n flag = df.filter(regex=header_sym)\n flagcols = flag.columns\n # print(flagcols)\n # ordonner les flag dans un dataframe\n dlyflags = pd.melt(df, id_vars=[\"STATION_NUMBER\", \"YEAR\", \"MONTH\"], value_vars=flagcols)\n\n if len(d30) > 0:\n dlyflags = dlyflags.drop(d30.index).reset_index(drop=True)\n # print(dlyflags.shape)\n\n if len(d29) > 0:\n dlyflags = dlyflags.drop(d29.index).reset_index(drop=True)\n # print(dlyflags.shape)\n\n if len(d28) > 0:\n dlyflags = dlyflags.drop(d28.index).reset_index(drop=True)\n # print(dlyflags.shape)\n # -----------------------------------END SYMBOL---------------------------------------------\n\n # transform date\n dlydata.insert(loc=1, column='DATE', value=pd.to_datetime(dlydata[['YEAR', 'MONTH', 'DAY']]))\n # ---------------------------------plot the dataframe--------------------------------------\n dlytoplot = dlydata[['DATE', 'value']].set_index('DATE')\n dlydata = dlydata.drop(['YEAR', 'MONTH', 'DAY', 'variable'], axis=1)\n print(dlydata.shape)\n\n if to_plot == 1:\n dlytoplot.plot()\n return dlytoplot\n else:\n return dlydata\n\n\ndef parse_data_from_hydat_files(source,\n regions_list=['QC', 'ON', 'NB', 'NL']):\n \"\"\"\n Get list of all available stations from cehq with station's number and name as value\n \"\"\"\n cnx = sqlite3.connect(source)\n\n df1 = pd.read_sql_query(\"SELECT * FROM STATIONS\", cnx)\n\n df1 = df1[df1['PROV_TERR_STATE_LOC'].isin(regions_list)]\n df1 = df1[['STATION_NUMBER', 'STATION_NAME', 'PROV_TERR_STATE_LOC', 'DRAINAGE_AREA_GROSS', 'LATITUDE', 'LONGITUDE']]\n df1.columns = ['station_number', 'station_name', 'province', 'drainage_area', 'latitude', 'longitude']\n df1.insert(loc=0, column='id_point', value=df1['station_number'])\n df1['id_point'] = df1['station_number'].astype(str)\n df1.insert(loc=3, column='data_type', value='Flow')\n\n df2 = pd.read_sql_query(\"SELECT * FROM STN_REGULATION\", cnx)\n df = pd.merge(df1, df2, how='left', left_on=['station_number'], right_on=['STATION_NUMBER'])\n df.insert(loc=5, column='regulation', value=df['REGULATED'].map({0: 'Naturel', 1: 'Influencé', np.nan: 'Indisponible'}))\n df_sup1 = df.drop(columns=['YEAR_FROM', 'YEAR_TO', 'REGULATED', 'station_number'])\n\n meta_sta_hydro = df_sup1.drop(columns=['id_point', 'data_type'])\n meta_sta_hydro = meta_sta_hydro.drop_duplicates()\n meta_sta_hydro.insert(loc=1, column='equivalent_name', value=np.nan)\n meta_sta_hydro.insert(loc=0, column='station_number', value=meta_sta_hydro['STATION_NUMBER'])\n meta_sta_hydro = meta_sta_hydro.drop(columns=['STATION_NUMBER'])\n meta_sta_hydro.insert(loc=0, column='id_point', value=range(2000, 2000 + meta_sta_hydro.shape[0], 1))\n\n meta_ts = df_sup1.drop(columns=['station_name', 'regulation', 'drainage_area', 'latitude', 'longitude'])\n meta_ts['id_point'] = range(2000, 2000 + meta_sta_hydro.shape[0], 1)\n meta_ts.insert(loc=3, column='time_step', value='1_J')\n meta_ts.insert(loc=4, column='aggregation', value='moy')\n meta_ts.insert(loc=5, column='units', value='m3/s')\n meta_ts.insert(loc=7, column='source', value='HYDAT')\n\n meta_ts = pd.merge(meta_ts, meta_sta_hydro[['station_number']],\n left_on='STATION_NUMBER', right_on='station_number', how='left').drop(columns=['station_number'])\n\n cols = meta_ts.columns.tolist()\n cols = cols[-1:] + cols[:-1]\n meta_ts = meta_ts[cols]\n\n meta_ts.insert(loc=0, column='id_time_serie', value=range(2000, 2000 + meta_ts.shape[0], 1))\n print(meta_ts.columns)\n print(df.columns)\n\n # df = pd.merge(df, meta_ts[['id_time_serie', 'STATION_NUMBER']],\n # left_on='id_point', right_on='STATION_NUMBER', how='left').drop(columns=['id_point'])\n # cols = df.columns.tolist()\n # cols = cols[-1:] + cols[:-1]\n # df = df[cols]\n\n meta_ts = meta_ts.drop(columns=['STATION_NUMBER','province'])\n\n meta_ts['start_date'] = np.nan\n meta_ts['end_date'] = np.nan\n\n\n # Débit\n list_df = []\n for idx, row in meta_sta_hydro.iterrows():\n numero_station = row['station_number']\n\n sql = \"\"\"\n SELECT *\n FROM DLY_FLOWS\n WHERE STATION_NUMBER in\n (\"%s\"\n )\n \"\"\" % (numero_station)\n chunk = pd.read_sql_query(sql, cnx)\n daily_station = hydat_daily2(chunk, True, False)\n daily_station.columns = ['id_time_serie', 'date', 'value']\n daily_station = daily_station.set_index([\"date\"])\n daily_station.index = pd.to_datetime(daily_station.index)\n daily_station.index = daily_station.index.tz_localize(\"America/Montreal\", ambiguous='infer',\n nonexistent='shift_forward')\n print(row['id_point'])\n daily_station['id_time_serie'] = meta_ts[meta_ts['id_point'] == row['id_point']]['id_time_serie'].values[0]\n daily_station.reset_index(level=0, inplace=True)\n if daily_station['date'].shape[0]>0:\n meta_ts.loc[meta_ts['id_point'] == row['id_point'], 'start_date'] = daily_station.dropna()['date'].iloc[0]\n meta_ts.loc[meta_ts['id_point'] == row['id_point'], 'end_date'] = daily_station.dropna()['date'].iloc[-1]\n list_df.append(daily_station)\n else:\n meta_ts = meta_ts[meta_ts['id_point'] != row['id_point']]\n meta_sta_hydro = meta_sta_hydro[meta_sta_hydro['id_point'] != row['id_point']]\n df = pd.concat(list_df)\n\n print(meta_sta_hydro.head())\n print(meta_ts.head())\n print(df.head())\n return [meta_sta_hydro, meta_ts, df]\n\n\ndef load_files_from_hydat(stations_list,\n store='tmp'):\n \"\"\"\n\n \"\"\"\n\n\ndef delete_files_in_store(store='tmp'):\n import os\n import shutil\n for root, dirs, files in os.walk(store):\n for f in files:\n os.unlink(os.path.join(root, f))\n for d in dirs:\n shutil.rmtree(os.path.join(root, d))\n\n\ndef df_update_index_basins(df=None):\n \"\"\"\n\n \"\"\"\n # query on station_number field (unique field for CEHQ only)\n df_db = pd.read_sql(\"\"\"\n SELECT id_point, station_number FROM basins \n \"\"\", con=engine, index_col='station_number')\n print(df_db.shape[0])\n if df_db.shape[0] > 0:\n id_point_new = df_db['id_point'].max() + 1\n # replace df's index with the one from database\n df = df.set_index('station_number')\n print(df[['id_point']].head())\n print(df_db[['id_point']].head())\n df_index = pd.merge(df[['id_point']], df_db[['id_point']], left_index=True, right_index=True)\n df_index.to_csv('/home/slanglois/PycharmProjects/flask-hydrodatahub/test.csv')\n dict_index = df_index[['id_point_x', 'id_point_y']].set_index('id_point_x')\n df['id_point'].update(df_db['id_point'])\n\n df.loc[~df.index.isin(df_db.index)]['id_point'] = range(id_point_new, id_point_new +\n df.loc[~df.index.isin(df_db.index)]['id_point'].shape[0], 1)\n df = df.reset_index()\n return df, dict_index\n\n\ndef df_update_index_ts(df=None):\n \"\"\"\n\n \"\"\"\n # query on station_number field (unique field for CEHQ only)\n df_db = pd.read_sql(\"\"\"\n SELECT * FROM meta_ts \n \"\"\", con=engine, index_col=['id_point', 'data_type', 'time_step',\n 'aggregation', 'units', 'source'])\n print(df_db.shape[0])\n if df_db.shape[0] > 0:\n id_point_new = df_db['id_time_serie'].max() + 1\n # replace df's index with the one from database\n df = df.set_index(['id_point', 'data_type', 'time_step',\n 'aggregation', 'units', 'source'])\n df_index = pd.merge(df[['id_time_serie']], df_db[['id_time_serie']], left_index=True, right_index=True)\n dict_index = df_index[['id_time_serie_x', 'id_time_serie_y']].set_index('id_time_serie_x')\n df['id_time_serie'].update(df_db['id_time_serie'])\n print(dict)\n\n df.loc[~df.index.isin(df_db.index)]['id_time_serie'] = range(id_point_new, id_point_new +\n df.loc[~df.index.isin(df_db.index)]\n ['id_time_serie'].shape[0], 1)\n df = df.reset_index()\n return df, dict_index\n\n\n\ndef df_to_sql(all_dfs, n=200000):\n \"\"\"\n\n \"\"\"\n # Basins metadata\n meta_sta_hydro, meta_ts, df = all_dfs\n meta_sta_hydro.columns = meta_sta_hydro.columns.str.lower()\n meta_ts.columns = meta_ts.columns.str.lower()\n df.columns = df.columns.str.lower()\n meta = MetaData(bind=engine)\n meta.reflect(bind=engine)\n meta_sta_hydro, basins_index = df_update_index_basins(meta_sta_hydro)\n insrt_vals = meta_sta_hydro.to_dict(orient='records')\n table = meta.tables['basins']\n insrt_stmnt = insert(table).values(insrt_vals)\n update_dict = {\n c.name: c\n for c in insrt_stmnt.excluded\n if not c.primary_key\n }\n do_nothing_stmt = insrt_stmnt.on_conflict_do_update(index_elements=['id_point'],\n set_=update_dict)\n engine.execute(do_nothing_stmt)\n\n # time series metadata\n meta_ts['id_point'].update(basins_index['id_point_y'])\n meta_ts, ts_index = df_update_index_ts(meta_ts)\n insrt_vals = meta_ts.to_dict(orient='records')\n table = meta.tables['meta_ts']\n insrt_stmnt = insert(table).values(insrt_vals)\n update_dict = {\n c.name: c\n for c in insrt_stmnt.excluded\n if not c.primary_key\n }\n do_nothing_stmt = insrt_stmnt.on_conflict_do_update(index_elements=['id_time_serie'],\n set_=update_dict)\n engine.execute(do_nothing_stmt)\n\n\n list_df = [df[i:i + n] for i in range(0, df.shape[0], n)]\n table = meta.tables['don_ts']\n constraint = table.primary_key.columns.keys()\n for idx, chunked_df in enumerate(list_df):\n chunked_df['id_time_serie'] = chunked_df['id_time_serie'].replace(ts_index.index,\n ts_index['id_time_serie_y'])\n try:\n print(str(idx*n))\n # chunked_df['id_time_serie'].update(ts_index['id_time_serie_y'])\n insrt_vals = chunked_df.drop_duplicates().to_dict(orient='records')\n insrt_stmnt = insert(table).values(insrt_vals)\n update_dict = {\n c.name: c\n for c in insrt_stmnt.excluded\n if not c.primary_key\n }\n do_nothing_stmt = insrt_stmnt.on_conflict_do_update(index_elements=constraint,\n set_=update_dict)\n engine.execute(do_nothing_stmt)\n except :\n print('chunk # {} was not inserted correctly in database'.format(idx))\n\n\nif __name__ == '__main__':\n\n\n # load_files_from_cehq(stations)\n\n\n all_dfs = parse_data_from_hydat_files('/home/slanglois/Downloads/Hydat_sqlite3_20191016/Hydat.sqlite3',\n regions_list=['QC', 'ON', 'NB', 'NL'])\n df_to_sql(all_dfs)\n #delete_files_in_store()\n\n", "sub_path": "app/tasks/hydat.py", "file_name": "hydat.py", "file_ext": "py", "file_size_in_byte": 14484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sqlalchemy.create_engine", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas.melt", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.int8", "line_number": 56, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 71, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 79, "usage_type": "call"}, {"api_name": "pandas.melt", "line_number": 94, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 110, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 128, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 130, "usage_type": "call"}, {"api_name": "pandas.read_sql_query", "line_number": 139, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 141, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 146, "usage_type": "attribute"}, {"api_name": "pandas.merge", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 177, "usage_type": "attribute"}, {"api_name": "numpy.nan", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pandas.read_sql_query", "line_number": 193, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 197, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 210, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 228, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 230, "usage_type": "call"}, {"api_name": "os.path", "line_number": 230, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 232, "usage_type": "call"}, {"api_name": "os.path", "line_number": 232, "usage_type": "attribute"}, {"api_name": "pandas.read_sql", "line_number": 240, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 250, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 266, "usage_type": "call"}, {"api_name": "pandas.merge", "line_number": 276, "usage_type": "call"}, {"api_name": "sqlalchemy.MetaData", "line_number": 298, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.insert", "line_number": 303, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.insert", "line_number": 318, "usage_type": "call"}, {"api_name": "sqlalchemy.dialects.postgresql.insert", "line_number": 339, "usage_type": "call"}]} +{"seq_id": "26043735", "text": "#Author: Matt Wolf\n#Date:3/16/19,\n#3-19 --notes\n#working on adding the logic for game over. needs scores. balloons to come. start screen, instructions\n#pause, level up options etc\n#very rough between classes (or during)\n#Description:\n#Attempt a simple balloon color match game\n\n\n#works. need to add logic for score and up coming balls\n#also need to add logic for balls next to balls\n\n#import modules\nimport pygame\nimport math\nimport random\n\n#####################################\n#init pygame and window variables\n#####################################\npygame.init()\n#refresh clock\nclock = pygame.time.Clock()\n#screen size\nscreen_Width = 750\nscreen_Height = 500\n#define win as our window\nwin = pygame.display.set_mode(( screen_Width, screen_Height))\npygame.display.set_caption('Balloon Breaker')\n#assign all pictures variables\nbg100 = pygame.image.load('backGround100.png')\nbg80 = pygame.image.load('backGround80.png')\nbg60 = pygame.image.load('backGround60.png')\nbg40 = pygame.image.load('backGround40.png')\nbg40F = pygame.image.load('backGround40Flash.png')\nbg20 = pygame.image.load('backGround20.png')\nbg20F = pygame.image.load('backGround20Flash.png')\ngameover = pygame.image.load('game over.png')\nbg = bg100\n###################################################\n# init all variables for game play\n###################################################\n#border width ## not setable constant\nBORDER = 4\nlevel = 4\nballoons = [] #init the balloon list\nrun = True\nplayerStep = 64 #based on the size of the picture\nplayerBall = True \nrestart = True\n#images for the balloon\nballoonImage = [pygame.image.load(\"R1.png\"),pygame.image.load(\"O1.png\"),pygame.image.load(\"Y1.png\"),pygame.image.load(\"B1.png\"),\n pygame.image.load(\"P1.png\"),pygame.image.load(\"G1.png\"),pygame.image.load(\"M1.png\")]\n#variable for which balloons in the list can be called\nballoonRange = 5\n#player clock for timing game eventually\nplayerClock = 0\n#############################################\n#this is a logger to show the position of all\n#ballons in the sequence after each move\n#############################################\nfileout = open(\"outfile.txt\", 'a')\nfileout.write(\"new run\" + '_'*48)\nfileout.close()\n#logger to check balloon positions\ndef logBalloons(balloons):\n #takes the balloon list as input\n #outputs the file outfile.txt\n #each run of the game is \"new run\"\n try:\n fileout = open(\"outfile.txt\" , 'a')\n fileout.write('_' *48 + '\\n')\n for item in balloons:\n fileout.write(str(item.x) + ', ' + str(item.y) + '\\n')\n fileout.close()\n except IOError:\n print('There has been an IOError')\n except Error as err:\n print('There has been a file error: ', err)\n\n#balloon for game\nclass balloon(object):\n\n def __init__(self, x, y):\n self.x = x\n self.y = y\n self.balloonImage = balloonImage[random.randint(0, balloonRange)]\n self.isFlying = False\n self.stop = 0\n self.isLive = True\n def draw(self, win):\n win.blit(self.balloonImage, (self.x , self.y))\n\n\n#setup the init board function\ndef setupBoard(level):\n for x in range(BORDER, 500, 64):\n for y in range(BORDER, 64 * level + BORDER , 64):\n balloons.append(balloon(x , y ))\n for item in balloons:\n item.draw(win)\n pygame.display.update()\n\n\n#redraw the game window on refresh\ndef reDrawGameWindow():\n win.blit(bg, (0,0))\n player.draw(win)\n for item in balloons:\n item.draw(win)\n if not restart:\n win.blit(gameover, (57,50))\n pygame.display.update()\n\n\n\ndef moveLeft():\n allowMove = True\n for item in balloons:\n if (item.x + playerStep == player.x) and (item.y + playerStep > player.y):\n allowMove = False\n if allowMove:\n player.x -= playerStep\n\ndef moveRight():\n allowMove = True\n for item in balloons:\n if (item.x == player.x + playerStep) and (item.y + playerStep > player.y):\n allowMove = False\n if allowMove:\n player.x += playerStep\n\n################################################\n#setup functions\n################################################\nsetupBoard(level) #run the setup function\nplayer = balloon( 260, 400 + BORDER)\nplayer.draw(win)\npygame.display.update()\n\n\n################################################\n#main game loop\n################################################\nwhile run:\n clock.tick(15)\n playerClock += 1\n\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n run = False\n \n if not (playerBall):\n #print(\"here\")\n player.balloonImage = balloonImage[random.randint(0, balloonRange)]\n player.x = 260\n player.y = 388 + BORDER\n playerBall = True\n #print(playerClock , bg)\n \n #set bg image attempted to refactor into a function that would take input of playerClock and\n #background did not update. and game play was extremely laggy. leave this as part of the\n #main game loop! 3/24/19\n if playerClock >= 0 and playerClock < 15:\n bg = bg100\n elif playerClock > 15 and playerClock < 30: \n bg = bg80 \n elif playerClock > 30 and playerClock < 45: \n bg = bg60\n elif playerClock > 45 and playerClock < 60: \n bg = bg40\n elif playerClock > 60 and playerClock < 80:\n if playerClock % 5 == 0:\n bg = bg20F\n else:\n bg = bg20\n elif playerClock > 75 :\n bg = bg100\n playerClock = 0\n player.isFlying = True\n # leave the above as part of the main game loop\n \n #get keys pressed \n keys = pygame.key.get_pressed()\n #if player can move they do\n if keys[pygame.K_LEFT] and player.x > 0+ BORDER:\n moveLeft()\n elif keys[pygame.K_RIGHT] and player.x < 520- playerStep*2:\n moveRight() \n if not (player.isFlying): \n if keys[pygame.K_SPACE]:\n player.isFlying = True \n\n if player.isFlying:\n player.stop = BORDER\n largestY = BORDER\n for item in balloons:\n if item.x == player.x:\n if item.y >= largestY:\n largestY = item.y\n player.stop = largestY + playerStep\n \n print(largestY, ' ', player.stop) \n if player.y > player.stop:\n player.y -= playerStep//8\n elif player.y < player.stop:\n player.y = player.stop\n \n else: \n player.isFlying = False\n playerBall = False\n balloons.append(balloon(player.x,player.y))\n balloons[-1].balloonImage = player.balloonImage\n aboveIndex = -1\n yVal = 0\n i = 0\n while i < len(balloons)-1:\n if round(balloons[i].x) == round(balloons[-1].x):\n if balloons[i].y > yVal:\n yVal = balloons[i].y\n print(yVal)\n aboveIndex = i\n \n \n i +=1\n print(aboveIndex)\n if not aboveIndex == -1:\n if balloons[-1].balloonImage == balloons[aboveIndex].balloonImage:\n del balloons[-1]\n del balloons[aboveIndex]\n elif yVal > 323: \n restart = False \n while not restart:\n #print('gameover')\n \n reDrawGameWindow()\n logBalloons(balloons) \n \n \n reDrawGameWindow()\npygame.quit()\n \n", "sub_path": "balloonGame.py", "file_name": "balloonGame.py", "file_ext": "py", "file_size_in_byte": 7648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.init", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 29, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 29, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 30, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 36, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 38, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 53, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 54, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 54, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 88, "usage_type": "call"}, {"api_name": "pygame.display.update", "line_number": 103, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 103, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 114, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 140, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 140, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 150, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 150, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 151, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 156, "usage_type": "call"}, {"api_name": "pygame.key.get_pressed", "line_number": 185, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 185, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 187, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 189, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 242, "usage_type": "call"}]} +{"seq_id": "551073319", "text": "# File containing validators for forms - data validation is a part of security so we must use our own data validators\n\nfrom . import db # Import database for use in validation\n\nfrom . import sanitise # Some data may have to be sanitised if validating with database\n\nfrom wtforms import ValidationError # Raise wtforms ValidationError whenever data is invalid\n\nfrom collections.abc import Iterable # Used to check if data is Iterable for various validators\nfrom string import ascii_letters, digits # Used to check if data has alphabetic and/or numeric characters\nimport re\n\n\nclass NotInTableColumn(object):\n def __init__(self, table, column_name, message=None):\n self.table = str(table)\n self.column_name = str(column_name)\n if message:\n self.message = message\n else:\n self.message = u'Field is invalid'\n\n def __call__(self, form, field):\n if field.data:\n raw_sql = 'SELECT * FROM ' + self.table + ' WHERE ' + self.column_name + '=\"{}\"'.format(sanitise.all(field.data))\n existing_user = db.session.execute(raw_sql).first()\n if existing_user:\n raise ValidationError(self.message)\n\n\nclass Required(object):\n def __init__(self, message=None):\n if message:\n self.message = message\n else:\n self.message = u'Field is required'\n\n def __call__(self, form, field):\n if not field.data:\n raise ValidationError(self.message)\n\n\nclass EqualTo(object):\n def __init__(self, field_name, message=None):\n self.field_name = field_name\n if message:\n self.message = message\n else:\n self.message = u'Field must be the same as {}'.format(self.field_name)\n\n def __call__(self, form, field):\n if form[self.field_name].data != field.data:\n raise ValidationError(self.message)\n\n\nclass DifferentFrom(object):\n def __init__(self, field_name, message=None):\n self.field_name = field_name\n if message:\n self.message = message\n else:\n self.message = u'Field must not be the same as {}'.format(self.field_name)\n\n def __call__(self, form, field):\n if form[self.field_name].data == field.data:\n raise ValidationError(self.message)\n\n\nclass Length(object):\n def __init__(self, min_length=None, max_length=None, message=None):\n if not min_length and not max_length:\n raise ValueError('Length validator does not have a minimum or maximum length to validate against')\n self.min = min_length\n self.max = max_length\n if message:\n self.message = message\n else:\n if self.min and self.max:\n self.message = u'Field must contain at least {} but no more than {} characters'.format(self.min, self.max)\n elif self.min:\n self.message = u'Field must contain at least {} characters'.format(self.min)\n elif self.max:\n self.message = u'Field must contain no more than {} characters'.format(self.max)\n\n def __call__(self, form, field):\n data_length = 0 # Handle no data input (data length default to 0)\n if field.data:\n data_length = len(field.data)\n if (self.min and data_length < self.min) or (self.max and data_length > self.max):\n raise ValidationError(self.message)\n\n\nclass AlphaNumeric(object):\n def __init__(self, message=None):\n if message:\n self.message = message\n else:\n self.message = u'Field must contain both letters and numbers'\n\n def __call__(self, form, field):\n alphabetic = False\n numeric = False\n for letter in ascii_letters:\n if letter in field.data:\n alphabetic = True\n break\n for digit in digits:\n if digit in field.data:\n numeric = True\n break\n if not alphabetic or not numeric:\n raise ValidationError(self.message)\n\n\nclass NotContainAny(object):\n def __init__(self, check, message=None):\n if not isinstance(check, Iterable):\n raise TypeError('NotContainAny validator requires validation data to be Iterable')\n self.check = check\n if message:\n self.message = message\n else:\n self.message = u'Field must not contain any of the following terms: {}'.format(self.check)\n\n def __call__(self, form, field):\n for i in self.check:\n if str(i) in str(field.data):\n raise ValidationError(self.message)\n\n\nclass MustContain(object):\n def __init__(self, check, message=None):\n if not isinstance(check, Iterable):\n raise TypeError('MustContain validator requires validation data to be Iterable')\n self.check = check\n if message:\n self.message = message\n else:\n self.message = u'Field must contain the following terms: {}'.format(self.check)\n\n def __call__(self, form, field):\n for i in self.check:\n if str(i) not in str(field.data):\n raise ValidationError(self.message)\n\n\nclass MustContainRegex(object):\n def __init__(self, message=None):\n if message:\n self.message = message\n else:\n self.message = u'Invalid Email Address'.format()\n\n def __call__(self, form, field):\n regex = '^\\w+([\\.-]?\\w+)*@\\w+([\\.-]?\\w+)*(\\.\\w{2,3})+$'\n if not re.search(regex, field.data):\n raise ValidationError(self.message)\n\n\n\n\n", "sub_path": "blogsite/validators.py", "file_name": "validators.py", "file_ext": "py", "file_size_in_byte": 5586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "wtforms.ValidationError", "line_number": 28, "usage_type": "call"}, {"api_name": "wtforms.ValidationError", "line_number": 40, "usage_type": "call"}, {"api_name": "wtforms.ValidationError", "line_number": 53, "usage_type": "call"}, {"api_name": "wtforms.ValidationError", "line_number": 66, "usage_type": "call"}, {"api_name": "wtforms.ValidationError", "line_number": 90, "usage_type": "call"}, {"api_name": "string.ascii_letters", "line_number": 103, "usage_type": "name"}, {"api_name": "string.digits", "line_number": 107, "usage_type": "name"}, {"api_name": "wtforms.ValidationError", "line_number": 112, "usage_type": "call"}, {"api_name": "collections.abc.Iterable", "line_number": 117, "usage_type": "argument"}, {"api_name": "wtforms.ValidationError", "line_number": 128, "usage_type": "call"}, {"api_name": "collections.abc.Iterable", "line_number": 133, "usage_type": "argument"}, {"api_name": "wtforms.ValidationError", "line_number": 144, "usage_type": "call"}, {"api_name": "re.search", "line_number": 156, "usage_type": "call"}, {"api_name": "wtforms.ValidationError", "line_number": 157, "usage_type": "call"}]} +{"seq_id": "144718499", "text": "\"\"\"\n This file contains repetitive procedural scripts that perform certain sequence of actions\n that are necessary for a unittest but too ugly to be in those files \n\"\"\"\n\nimport time\nfrom selenium.webdriver.support.ui import Select\n\n\ndef admin_create_new_user_logout(obj):\n \"\"\"\n Login to admin\n Create a new random user using information in the unitest object\n Logout\n \"\"\"\n\n # Admin login\n driver = obj.driver\n driver.get(obj.base_url + \"/login\")\n username_input = driver.find_element(\"xpath\", \"//div[@id='login-tabpane-admin']/form/div/input\")\n placeholder = username_input.get_attribute('placeholder')\n username_input.send_keys(\"admin\")\n submit_button = driver.find_element(\"xpath\", \"//div[@id='login-tabpane-admin']/form/button\")\n submit_button.click()\n # Add a new user\n create_new_user_button = driver.find_element_by_id(\"create_new_user_button\")\n create_new_user_button.click()\n create_modal = obj.selenium.is_element_exist(\"id\", \"create_modal\")\n create_first_name = driver.find_element_by_id(\"create_first_name\")\n create_last_name = driver.find_element_by_id(\"create_last_name\")\n create_emal = driver.find_element_by_id(\"create_email\")\n create_first_name.send_keys(obj.user_first_name)\n create_last_name.send_keys(obj.user_last_name)\n create_emal.send_keys(obj.user_email)\n confirm_create_new_user_button = driver.find_element_by_id(\"confirm_create_new_user_button\")\n confirm_create_new_user_button.click()\n # Log out from admin\n logout_button = driver.find_element_by_id(\"logout\")\n time.sleep(1)\n logout_button.click()\n\ndef login_as_user(obj):\n \"\"\"\n Just ... login as an user. Using information in the unittest object.\n \"\"\"\n obj.driver.get(obj.base_url + \"/login\")\n obj.driver.find_element_by_link_text(\"User\").click()\n obj.driver.find_element(\"xpath\", \"//div[@id='login-tabpane-user']/form/div/input\").send_keys(obj.user_email)\n obj.driver.find_element(\"xpath\", \"//div[@id='login-tabpane-user']/form/button\").click()\n\n\ndef user_add_project(obj, another_project_name=None):\n \"\"\"\n Create a project\n \"\"\"\n driver = obj.driver\n time.sleep(1)\n # Click create new user button to show modal\n create_project_button = driver.find_element_by_id(\"create_project_button\")\n create_project_button.click()\n # Check CreateModal is pop up\n create_modal = obj.selenium.is_element_exist(\"id\", \"create_project_modal\")\n # Type in project name\n create_project_name = driver.find_element_by_id(\"create_project_name\")\n create_project_name.clear()\n if another_project_name == None:\n create_project_name.send_keys(obj.project_name)\n else:\n create_project_name.send_keys(another_project_name)\n # click \"create\" button\n confirm_create_project_button = driver.find_element_by_id(\"confirm_create_project_button\")\n confirm_create_project_button.click()\n time.sleep(1)\n\n\ndef pull_out_create_pomodoro_modal(obj):\n \"\"\"\n The name tells it\n \"\"\"\n driver = obj.driver\n # Click on \"Pomodoro\" tab\n driver.find_element_by_id(\"sidebar_user_Pomodoro\").click()\n # Click on \"Create a new pomodoro\" button\n driver.find_element_by_id(\"create_pomodoro_button\").click()\n\ndef create_session_for_project(obj):\n \"\"\"\n Make a trivial session for given project\n \"\"\"\n driver = obj.driver\n # Type in 1 second pomodoro\n pull_out_create_pomodoro_modal(obj)\n driver.find_element_by_id(\"create_second\").send_keys(1)\n # Select given project\n Select(driver.find_element_by_id(\"project_select_box\")).select_by_visible_text(obj.project_name)\n # Confirm create pomodoro\n driver.find_element_by_id(\"create_confirm_button\").click()\n # Wait for continue modal then click \"No\"\n obj.selenium.element_wait(\"id\", \"decide_stop\")\n driver.find_element_by_id(\"decide_stop\").click()\n time.sleep(1)\n # Go back to project tab\n driver.find_element_by_id(\"sidebar_user_Projects\").click()\n time.sleep(1)\n\ndef start_associated_pomodoro(obj):\n \"\"\"\n The name tells it.\n \"\"\"\n pull_out_create_pomodoro_modal(obj)\n\n driver = obj.driver\n # # Type in 10 second pomodoro\n driver.find_element_by_id(\"create_second\").send_keys(10)\n # Select given project\n Select(driver.find_element_by_id(\"project_select_box\")).select_by_visible_text(obj.project_name)\n # Confirm create pomodoro\n driver.find_element_by_id(\"create_confirm_button\").click()\n\ndef start_not_associated_pomodoro(obj):\n \"\"\"\n The name tells it.\n \"\"\"\n pull_out_create_pomodoro_modal(obj)\n\n driver = obj.driver\n # Type in 10 second pomodoro\n driver.find_element_by_id(\"create_second\").send_keys(10)\n # Select given project\n Select(driver.find_element_by_id(\"project_select_box\")).select_by_visible_text('No association')\n # Confirm create pomodoro\n driver.find_element_by_id(\"create_confirm_button\").click()\n\n\ndef make_a_session_switch_to_report(obj):\n \"\"\"\n Create a session related to a project with just 1 pomodoro\n \"\"\"\n pull_out_create_pomodoro_modal(obj)\n\n driver = obj.driver\n driver.find_element_by_id(\"create_second\").send_keys(1)\n Select(driver.find_element_by_id(\"project_select_box\")).select_by_visible_text(obj.project_name)\n driver.find_element_by_id(\"create_confirm_button\").click()\n st = time.time()\n\n obj.selenium.element_wait(\"id\", \"decide_stop\")\n driver.find_element_by_id(\"decide_stop\").click()\n et = time.time()\n\n time.sleep(1)\n driver.find_element_by_id(\"sidebar_user_Report\").click()\n\ndef clean_up(obj):\n \"\"\"\n Switch to admin and delete the user just created in tearDown\n \"\"\"\n driver = obj.driver\n driver.get(obj.base_url + \"/login\")\n driver.find_element_by_link_text(\"Admin\").click()\n driver.find_element(\"xpath\", \"//div[@id='login-tabpane-admin']/form/div/input\").send_keys('admin')\n driver.find_element(\"xpath\", \"//div[@id='login-tabpane-admin']/form/button\").click()\n time.sleep(1)\n\n for i in [\"email\", \"user_email\"]:\n if hasattr(obj, i):\n j = getattr(obj, i)\n\n driver.find_element_by_id(\"general-search\").send_keys(j)\n is_user = obj.selenium.is_element_exist(\"xpath\", \"//tr[td[contains(text(), '{}')]]//i[contains(@class, 'pe-7s-trash')]\".format(j))\n if is_user:\n driver.find_element_by_xpath(\"//tr[td[contains(text(), '{}')]]//i[contains(@class, 'pe-7s-trash')]\".format(j)).click()\n is_delete_modal = obj.selenium.is_element_exist(\"id\", \"delete_modal\")\n if is_delete_modal:\n confirm_delete_button = driver.find_element_by_id(\"confirm_delete\")\n confirm_delete_button.click()\n time.sleep(1)\n \n ", "sub_path": "test/utils/scripts.py", "file_name": "scripts.py", "file_ext": "py", "file_size_in_byte": 6791, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "time.sleep", "line_number": 39, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 73, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.Select", "line_number": 95, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 101, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 104, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.Select", "line_number": 116, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.Select", "line_number": 130, "usage_type": "call"}, {"api_name": "selenium.webdriver.support.ui.Select", "line_number": 143, "usage_type": "call"}, {"api_name": "time.time", "line_number": 145, "usage_type": "call"}, {"api_name": "time.time", "line_number": 149, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 151, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 163, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 177, "usage_type": "call"}]} +{"seq_id": "561765443", "text": "from Webwork_bot import *\r\n\r\nimport itertools\r\n\r\nimport time\r\n\r\nfrom login_info import j_pass, j_user\r\n\r\nfrom random import *\r\n\r\n# A3 url\r\n\r\na3_url = 'https://webwork.math.mcgill.ca/webwork2/MATH240_WINTER2020/Assignment_3/7/'\r\n\r\nanswers = ['A', 'B', 'C', 'D', 'E']\r\n\r\npossible_answer_list = itertools.permutations(answers)\r\n\r\n# A4\r\n\r\na4_url = 'https://webwork.math.mcgill.ca/webwork2/MATH240_WINTER2020/Assignment_4/2/'\r\n\r\n# Question 1, 2 creating all combinations\r\n\r\nanswers = \"YNY\"\r\n\r\nanswers_2 = 'NYN'\r\n\r\npossible_combinations_one = itertools.combinations_with_replacement(answers, 3)\r\n\r\npossible_combinations_two = itertools.combinations_with_replacement(answers_2, 3)\r\n\r\nall_combinations = []\r\n\r\nfor combination in possible_combinations_one:\r\n\r\n if combination not in all_combinations:\r\n all_combinations.append(combination)\r\n\r\nfor combination in possible_combinations_two:\r\n\r\n if combination not in all_combinations:\r\n all_combinations.append(combination)\r\n\r\n# combining combinations from the two lists (no duplicates)\r\n\r\ncombinations_2D = []\r\n\r\nfor end_combination in all_combinations:\r\n\r\n listCombinations = list(end_combination)\r\n\r\n if ['Y', 'Y', 'Y'] == listCombinations:\r\n\r\n listCombinations.append('Y')\r\n\r\n combinations_2D.append(listCombinations)\r\n\r\n else:\r\n\r\n listCombinations.append('N')\r\n\r\n combinations_2D.append(listCombinations)\r\n\r\nfor combo in combinations_2D:\r\n print(combo)\r\n\r\n#\r\n\r\n#\r\n\r\n#\r\n\r\n#\r\n\r\n#\r\n\r\n#\r\n\r\n#\r\n\r\nrun_it = False\r\n\r\n# web work bot entering combinations A3\r\n\r\nif run_it:\r\n\r\n web_work_bot = WebWorkBot()\r\n\r\n web_work_bot.login(a3_url, j_user, j_pass)\r\n\r\n sleep_time = .01\r\n\r\n for possible_answers in possible_answer_list:\r\n web_work_bot.erase_answers_A3()\r\n\r\n time.sleep(sleep_time)\r\n\r\n web_work_bot.enter_answers_A3(possible_answers)\r\n\r\n time.sleep(sleep_time)\r\n\r\n web_work_bot.submit_answers()\r\n\r\n# randomize list\r\nnew_combinations = []\r\nwhile len(new_combinations) < 8:\r\n random_combination = choice(combinations_2D)\r\n if random_combination not in new_combinations:\r\n new_combinations.append(random_combination)\r\n\r\n# run bot\r\n\r\nrun_it_A4 = True\r\n\r\n# web work bot entering combinations A4\r\n\r\nif run_it_A4:\r\n\r\n sleep_time = .000001\r\n\r\n web_work_bot = WebWorkBot()\r\n\r\n web_work_bot.login(a4_url, j_user, j_pass)\r\n # boolean variables\r\n checked_r2 = False\r\n checked_r3 = False\r\n checked_r4 = False\r\n\r\n for i in range(len(new_combinations)):\r\n answer_r1 = new_combinations[i]\r\n if checked_r2:\r\n answer_r2 = choice(new_combinations)\r\n answer_r3 = choice(new_combinations)\r\n answer_r4 = choice(new_combinations)\r\n web_work_bot.erase_answers_A4()\r\n time.sleep(sleep_time)\r\n web_work_bot.enter_answers_A4(answer_r1, answer_r2, answer_r3, answer_r4)\r\n time.sleep(sleep_time)\r\n web_work_bot.submit_answers()\r\n else:\r\n for j in range(len(new_combinations)):\r\n answer_r2 = new_combinations[j]\r\n if checked_r3:\r\n answer_r3 = choice(new_combinations)\r\n answer_r4 = choice(new_combinations)\r\n web_work_bot.erase_answers_A4()\r\n time.sleep(sleep_time)\r\n web_work_bot.enter_answers_A4(answer_r1, answer_r2, answer_r3, answer_r4)\r\n time.sleep(sleep_time)\r\n web_work_bot.submit_answers()\r\n else:\r\n for k in range(len(new_combinations)):\r\n answer_r3 = new_combinations[k]\r\n if checked_r4:\r\n answer_r4 = choice(new_combinations)\r\n web_work_bot.erase_answers_A4()\r\n time.sleep(sleep_time)\r\n web_work_bot.enter_answers_A4(answer_r1, answer_r2, answer_r3, answer_r4)\r\n time.sleep(sleep_time)\r\n web_work_bot.submit_answers()\r\n else:\r\n for l in range(len(new_combinations)):\r\n answer_r4 = new_combinations[l]\r\n web_work_bot.erase_answers_A4()\r\n time.sleep(sleep_time)\r\n web_work_bot.enter_answers_A4(answer_r1, answer_r2, answer_r3, answer_r4)\r\n time.sleep(sleep_time)\r\n web_work_bot.submit_answers()\r\n checked_r4 = True\r\n checked_r3 = True\r\n checked_r2 = True\r\n", "sub_path": "Webwork 240.py", "file_name": "Webwork 240.py", "file_ext": "py", "file_size_in_byte": 4734, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "itertools.permutations", "line_number": 17, "usage_type": "call"}, {"api_name": "itertools.combinations_with_replacement", "line_number": 29, "usage_type": "call"}, {"api_name": "itertools.combinations_with_replacement", "line_number": 31, "usage_type": "call"}, {"api_name": "login_info.j_user", "line_number": 90, "usage_type": "argument"}, {"api_name": "login_info.j_pass", "line_number": 90, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 97, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 101, "usage_type": "call"}, {"api_name": "login_info.j_user", "line_number": 124, "usage_type": "argument"}, {"api_name": "login_info.j_pass", "line_number": 124, "usage_type": "argument"}, {"api_name": "time.sleep", "line_number": 137, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 139, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 148, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 150, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 158, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 160, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 166, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 168, "usage_type": "call"}]} +{"seq_id": "388753262", "text": "from chatbot_simplebot import simplebot\nimport os\nimport requests\nimport json\nimport argparse\nimport uuid\nimport dialogflow\nfrom hashlib import sha256\nimport webphone_info\nfrom dialogflow_parameters import para_to_dict\nfrom chatbot_simplebot.simplebot import SimpleHandler\n\nclass EricTestingBotHandler(simplebot.SimpleHandler):\n\n @staticmethod\n def getEnvironment(param):\n # param.update({'path_root': os.path.dirname(os.path.abspath(__file__))})\n env = simplebot.SimpleHandler.getEnvironment(param)\n return env\n\n def process_post(self, resultObj):\n \"\"\"\n Customized processing function for producing a result.\n \"\"\"\n project_id = \"erictestagent\"\n language_code = \"en-US\"\n session_id = sha256(str.encode(resultObj['from'])).hexdigest()\n text = resultObj['message']\n\n session_client = dialogflow.SessionsClient()\n\n session = session_client.session_path(project_id, session_id)\n print('Session path: {}\\n'.format(session))\n\n text_input = dialogflow.types.TextInput(text=text, language_code=language_code)\n\n query_input = dialogflow.types.QueryInput(text=text_input)\n\n response = session_client.detect_intent(\n session=session, query_input=query_input)\n\n ericTestingBot_json = open('erictestingbot.json', 'r').read()\n ericTestingBot_json_dict = json.loads(ericTestingBot_json)\n\n print('=' * 20)\n print('Query text: {}'.format(response.query_result.query_text))\n print('Detected intent: {} (confidence: {})\\n'.format(\n response.query_result.intent.display_name,\n response.query_result.intent_detection_confidence))\n print('Fulfillment text: {}\\n'.format(\n response.query_result.fulfillment_text))\n\n# fulfillment_dict = dict(response.query_result.fulfillment_text)\n# print(type(fulfillment_dict))\n# print(fulfillment_dict)\n ###\n# print('Parameter text: {}\\n'.format(\n# response.query_result.parameters))\n print('Parameter text: {}\\n'.format(\n para_to_dict(response.query_result.parameters)))\n para_dict = para_to_dict(response.query_result.parameters)\n\n print(response.query_result.parameters)\n print('Output context: {}\\n'.format(\n response.query_result.output_contexts))\n ericTestingBot_intent = response.query_result.intent.display_name\n\n if response.query_result.intent.display_name == \"getUserInfo\":\n try:\n resultObj['result'] = \"
\"+webphone_info.get_profile_table(para_dict['attuid']) + \"
View report chain\"\n except:\n resultObj['result'] = \"User not found\"\n elif response.query_result.intent.display_name == \"getReportChain\" or response.query_result.intent.display_name == \"info_getReportChain\":\n try:\n resultObj['result'] = \"
\"+webphone_info.get_report_chain(para_dict['attuid'])\n except:\n resultObj['result'] = \"User not found\"\n elif ericTestingBot_intent in ericTestingBot_json_dict:\n ### find answer from json_data\n resultObj['result'] = ericTestingBot_json_dict.get(ericTestingBot_intent)\n elif response.query_result.fulfillment_text:\n resultObj['result'] = response.query_result.fulfillment_text\n else:\n resultObj['result'] = \"Sorry, I did not understand you. Please ask your question in a different way.\"\n print(resultObj['result'])\n print(resultObj)\n ###\n# resultObj['result'] = response.query_result.fulfillment_text\n return resultObj\n\n\nif __name__ == \"__main__\":\n import sys\n sys.path.append(os.path.dirname(os.path.abspath(__file__))) # add path\n defArgs = {'port':9988, 'verbose':True} # default args\n if len(sys.argv)>2: # add user/password\n defArgs['bot_name'] = sys.argv[1]\n defArgs['bot_secret'] = sys.argv[2]\n defArgs['json_response']= True\n\n print(\"Launching console-interactive server... [credentials? run {:} ]\".format(sys.argv[0]))\n print()\n EricTestingBotHandler.serve_forever(defArgs, EricTestingBotHandler)\n", "sub_path": "erictestingbot0.py", "file_name": "erictestingbot0.py", "file_ext": "py", "file_size_in_byte": 4337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "chatbot_simplebot.simplebot.SimpleHandler", "line_number": 13, "usage_type": "attribute"}, {"api_name": "chatbot_simplebot.simplebot", "line_number": 13, "usage_type": "name"}, {"api_name": "chatbot_simplebot.simplebot.SimpleHandler.getEnvironment", "line_number": 18, "usage_type": "call"}, {"api_name": "chatbot_simplebot.simplebot.SimpleHandler", "line_number": 18, "usage_type": "attribute"}, {"api_name": "chatbot_simplebot.simplebot", "line_number": 18, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 27, "usage_type": "call"}, {"api_name": "dialogflow.SessionsClient", "line_number": 30, "usage_type": "call"}, {"api_name": "dialogflow.types.TextInput", "line_number": 35, "usage_type": "call"}, {"api_name": "dialogflow.types", "line_number": 35, "usage_type": "attribute"}, {"api_name": "dialogflow.types.QueryInput", "line_number": 37, "usage_type": "call"}, {"api_name": "dialogflow.types", "line_number": 37, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 43, "usage_type": "call"}, {"api_name": "dialogflow_parameters.para_to_dict", "line_number": 60, "usage_type": "call"}, {"api_name": "dialogflow_parameters.para_to_dict", "line_number": 61, "usage_type": "call"}, {"api_name": "webphone_info.get_profile_table", "line_number": 70, "usage_type": "call"}, {"api_name": "webphone_info.get_report_chain", "line_number": 75, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 94, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 96, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 97, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 98, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 101, "usage_type": "attribute"}]} +{"seq_id": "388360715", "text": "import os\nfrom flask import Flask\nfrom . import db\nfrom . import auth\n\n#create and configure the application\ndef create_app(test_config=None):\n #__name__ -> current Python module. Where the app is located\n #instance_relative_config -> the config files are relative to the instance folder\n #instance folder -> located outside the package. Can hold local data that\n # shouldn't be commited to version control. Ex: db, config files,others..\n app=Flask(__name__,instance_relative_config=True)\n\n #sets default config\n #SECRET_KEY -> keep data safe should be random\n #DATABASE -> path to the db file\n #app.instance_path -> oath to the instance folder\n app.config.from_mapping(\n SECRET_KEY='dev',\n DATABASE=os.path.join(app.instance_path,'flask-skeleton.sqlite'),\n )\n\n if test_config is None:\n #load the instance config, if it exists, when not testing\n #overrides the default config with values taken from the config file\n app.config.from_pyfile('config.py',silent=True)\n else:\n #load the test config if passed in\n #to pass test config independentlyof any development values\n app.config.from_mapping(test_config)\n\n #ensure the instance folder exists\n try:\n #Flask doesn't create the instance folder automatically\n #but it needs to be created to store files\n os.makedirs(app.instance_path)\n except OSError:\n pass\n\n #a simple page that says hello\n @app.route('/hello')\n def hello():\n return \"Hello David\"\n\n #Initialize Database\n db.init_app(app)\n\n #register auth blueprint in the application\n app.register_blueprint(auth.bp)\n\n return app\n", "sub_path": "denunciapp/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1694, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 36, "usage_type": "call"}]} +{"seq_id": "496911168", "text": "# coding: utf-8\nfrom openpyxl import load_workbook\nfrom openpyxl.cell import get_column_letter\nimport numpy as np\n#from numarray import *\nfrom array import *\n#from numeric import *\n\ndef read_xlsx2(filename):\n Data_Only=True\n a=[]\n b=[]\n #wb=load_workbook('200nmbeadlsmpsfvalues.xlsx',data_only=True)\n wb=load_workbook(filename,data_only=True)\n #print wb.get_sheet_names()\n ws=wb['Sheet1']\n #a=ws['A1':'A197']\n for col_idx in xrange(4, 6):\n col = get_column_letter(col_idx)\n for row in xrange(1, 11):\n #ws.cell(’%s%s’%(col, row)).value = ’%s%s’ % (col, row)\n a.append(ws.cell(\"%s%s\" % (col, row)).value)\n #print ws.cell(col, row).value\n b.extend(a[10:])\n a=a[:10]\n #print a[:196],a[196:]\n #print [a,b]\n return [a,b]\n\n", "sub_path": "read_xlsx2.py", "file_name": "read_xlsx2.py", "file_ext": "py", "file_size_in_byte": 736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "openpyxl.load_workbook", "line_number": 14, "usage_type": "call"}, {"api_name": "openpyxl.cell.get_column_letter", "line_number": 19, "usage_type": "call"}]} +{"seq_id": "292822187", "text": "#%%\nimport json\nimport colorsys\nfrom pathlib import Path\nimport numpy as np\nfrom pprint import pprint\nfrom colour import Color\nimport copy\n\n# %%\n\ns = json.loads(\n Path(\"/Users/chaichontat/Downloads/vscode-theme-darcula/themes/darculaori.json\").read_text()\n)\nori = copy.deepcopy(s)\n\nout = \"\"\n\n\ndef run(d):\n global out\n if isinstance(d, list):\n for k in d:\n if isinstance(k, dict):\n run(k)\n\n else:\n for k, v in d.items():\n if isinstance(v, dict) or isinstance(v, list):\n run(v)\n if isinstance(v, str) and v.startswith(\"#\"):\n # v = v.lstrip(\"#\")\n c = Color(v[:7])\n out += c.get_hex_l()\n if 0 < c.saturation < 1:\n if c.luminance > 0.3:\n c.luminance = np.clip(c.luminance + 0.1, 0, 1)\n if c.saturation > 0:\n c.saturation = np.clip(c.saturation + 0.2, 0, 1)\n elif c.saturation == 0:\n if c.luminance > 0.5:\n c.luminance = np.clip(c.luminance + 0.2, 0, 1)\n else:\n ...\n # c.luminance = np.clip(c.luminance + 0.05, 0, 1)\n\n # out = (\n # \"#\"\n # + \"\".join([format(int(v), \"02x\") for v in colorsys.hls_to_rgb(*u)])\n # + v[6:]\n # )\n out += \"\\t\" + c.get_hex_l() + \"\\n\"\n d[k] = c.get_hex_l() + v[7:]\n # print(v, out)\n # d[k] = colorsys.hsv_to_rgb(*u)\n\n\n# pprint(s)\nrun(s)\nPath(\"/Users/chaichontat/Downloads/vscode-theme-darcula/themes/darcula.json\").write_text(\n json.dumps(s)\n)\nPath(\"/Users/chaichontat/Downloads/vscode-theme-darcula/themes/comp.txt\").write_text(out)\n# %%\ns\n# %%\n", "sub_path": "color_transform.py", "file_name": "color_transform.py", "file_ext": "py", "file_size_in_byte": 1865, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "json.loads", "line_number": 12, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 13, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 15, "usage_type": "call"}, {"api_name": "colour.Color", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 42, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 60, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 61, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 63, "usage_type": "call"}]} +{"seq_id": "12795923", "text": "from django.contrib import admin\nfrom .models import Empleado, Habilidades\n# Register your models here.\n\n\nclass EmpleadoAdmin(admin.ModelAdmin):\n # El display que se hace con toda la información\n list_display = (\n \"id\",\n \"first_name\",\n \"last_name\",\n \"departamento\",\n \"job\",\n \"fullname\",\n \"avatar\",\n )\n\n # fullname no existe en la classe Emppleado de Models\n def fullname(self, obj):\n # Cualquier operación que necesite. Obj es lo que regresa Empleado en __str__. Si se pinta en \n # consola con print, itera sobre cada valor guardado ya en el admi de Django.\n return obj.first_name + \" \" + obj.last_name\n\n\n search_fields = ('first_name',)\n list_filter = ('departamento', 'job', 'habilidades')\n # Este solo funciona para relaciones muchos a muchos\n filter_horizontal = ('habilidades',)\nadmin.site.register(Empleado, EmpleadoAdmin)\nadmin.site.register(Habilidades)", "sub_path": "applications/persona/admin.py", "file_name": "admin.py", "file_ext": "py", "file_size_in_byte": 960, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.contrib.admin.ModelAdmin", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 6, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 29, "usage_type": "call"}, {"api_name": "models.Empleado", "line_number": 29, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 29, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 29, "usage_type": "name"}, {"api_name": "django.contrib.admin.site.register", "line_number": 30, "usage_type": "call"}, {"api_name": "models.Habilidades", "line_number": 30, "usage_type": "argument"}, {"api_name": "django.contrib.admin.site", "line_number": 30, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 30, "usage_type": "name"}]} +{"seq_id": "403460398", "text": "from django.conf.urls import url, include\nfrom django.contrib import admin\nfrom home.views import *\n\nurlpatterns = [\n\n url(r'^$', home_view, name=\"main\"),\n\n url(r'^admin/', admin.site.urls),\n\n url(r'^contact/', contact_view, name=\"contact\"),\n\n url(r'^yonetici/', yonet_view, name=\"yonetici\"),\n\n url(r'^mezun/', mezun_view, name=\"mezun\"),\n\n\n]\n\n", "sub_path": "blog/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 9, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}]} +{"seq_id": "344890971", "text": "from datetime import datetime\nfrom lxml import etree\nimport os\nimport time\n\n\ndef get_order_from_xml(xml):\n \"\"\"\n Recolhe valor da tag lot_no (Ordem de fabrico) do xml.\n :param xml: path do ficheiro xml.\n :return: valor da tag lot_no ou None\n \"\"\"\n tree = etree.parse(xml)\n order_element = tree.find('./lot_no')\n return order_element.text if order_element is not None else None\n\n\ndef parse(order_folder, xml_10, xml_11):\n tree_10 = etree.parse(xml_10)\n tree_11 = etree.parse(xml_11)\n ok_path = os.path.join(order_folder, '{}\\\\OK'.format(order_folder))\n nok_path = os.path.join(order_folder, '{}\\\\NOK'.format(order_folder))\n modules_per_panel = int(tree_11.find('./pcbs_in_panel').text)\n module_status = {}\n for module in range(modules_per_panel):\n module_sn = tree_11.find('serial_pcb_{}'.format(module + 1)).text\n status_top = tree_11.find('status_pcb_{}'.format(module + 1)).text\n status_bot = tree_10.find('status_pcb_{}'.format(module + 1)).text\n module_status[module_sn] = [status_top, status_bot]\n if not module_sn or len(module_sn) != 14:\n raise Exception('Campo \"serial_pcb_\" não é válido.')\n ng_modules = []\n for key, value in module_status.items():\n if 'NG' in value:\n ng_modules.append(key)\n filename = xml_11.split('\\\\')[-1].split('_')[0][2:]\n\n if ng_modules:\n file_path = os.path.join(nok_path, '{}.xml'.format(filename))\n else:\n file_path = os.path.join(ok_path, '{}.xml'.format(filename))\n\n root = etree.Element(\"result_file\")\n header = etree.SubElement(root, \"header\")\n etree.SubElement(header, \"creation_date\").text = str(datetime.now().strftime('%Y-%m-%d %H:%M:%S'))\n etree.SubElement(header, \"supplier_name\").text = \"Uartronica\"\n etree.SubElement(header, \"supplier_SAP_number\").text = \"Codigo_SAP_atribuido_Uartronica\"\n etree.SubElement(header, \"supplier_location\").text = \"Portugal\"\n etree.SubElement(header, \"last_vtpid\").text = \"10\"\n etree.SubElement(header, \"file_name\").text = filename\n etree.SubElement(header, \"file_length\").text = \"0\"\n etree.SubElement(header, 'part_number').text = ''.join(tree_11.find('./program').text.split('_')[:3])\n\n module_no = 1\n for sn, status in module_status.items():\n module_info = etree.SubElement(root, \"module_info\")\n etree.SubElement(module_info, 'name').text = str(module_no)\n etree.SubElement(module_info, 'serial_number').text = sn\n\n if not 'NG' in status:\n etree.SubElement(module_info, 'result').text = '1'\n else:\n etree.SubElement(module_info, 'result').text = '0'\n\n module_no += 1\n\n temp_file = os.path.join('temp\\\\', str(time.time()) + '.xml')\n\n tree = etree.ElementTree(root)\n tree.write(temp_file, encoding='utf-8', xml_declaration=True, pretty_print=True)\n\n # reescreve o ficheiro com total de bytes atualizado\n new_value = os.path.getsize(temp_file)\n tree = etree.parse(temp_file)\n xml_length = tree.find('.//header//file_length')\n old_value = xml_length.text\n new_value += len(str(new_value)) - len(old_value)\n xml_length.text = str(new_value)\n\n # remove duplicado se existir\n if filename + '.xml' in os.listdir(ok_path):\n os.remove(os.path.join(ok_path, filename + '.xml'))\n if filename + '.xml' in os.listdir(nok_path):\n os.remove(os.path.join(nok_path, filename + '.xml'))\n\n tree.write(file_path, encoding='utf-8', xml_declaration=True, pretty_print=True) # escreve xml\n os.remove(temp_file)\n", "sub_path": "file_parser.py", "file_name": "file_parser.py", "file_ext": "py", "file_size_in_byte": 3569, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "lxml.etree.parse", "line_number": 13, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 13, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 19, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 19, "usage_type": "name"}, {"api_name": "lxml.etree.parse", "line_number": 20, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "lxml.etree.Element", "line_number": 43, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 43, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 44, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 44, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 45, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 45, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 45, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 45, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 46, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 46, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 47, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 47, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 48, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 48, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 49, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 49, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 50, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 50, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 51, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 51, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 52, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 52, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 56, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 56, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 57, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 57, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 58, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 58, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 61, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 61, "usage_type": "name"}, {"api_name": "lxml.etree.SubElement", "line_number": 63, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 63, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 67, "usage_type": "call"}, {"api_name": "os.path", "line_number": 67, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 67, "usage_type": "call"}, {"api_name": "lxml.etree.ElementTree", "line_number": 69, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 69, "usage_type": "name"}, {"api_name": "os.path.getsize", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path", "line_number": 73, "usage_type": "attribute"}, {"api_name": "lxml.etree.parse", "line_number": 74, "usage_type": "call"}, {"api_name": "lxml.etree", "line_number": 74, "usage_type": "name"}, {"api_name": "os.listdir", "line_number": 81, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 83, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path", "line_number": 84, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 87, "usage_type": "call"}]} +{"seq_id": "404402253", "text": "import pygame\nimport random\n\n# dimension of each tiles\nTILE_SIZE = 32\n\n# texture of colors\nYELLOW = (255, 255, 0)\nRED = (255, 0, 0)\nBLUE = (0 , 0, 255)\nGREEN = (0, 255, 0)\nBROWN = (160, 82, 45)\n\ndef create_texture(color):\n image = pygame.Surface((TILE_SIZE, TILE_SIZE))\n image.fill(color)\n return image\n\n# 0x0 -> grass\n# 0xb -> dirt\ntextures = {\n 0x0 : create_texture(GREEN),\n 0xb : create_texture(BROWN)\n}\n\ntiles = [0x0, 0xb]\n\n# generate with tiles randomly\ndef generate_map(width, height, tilesize = TILE_SIZE):\n map_data = []\n for i in range(height // tilesize):\n map_data.append([])\n for j in range(width // tilesize):\n rand_index = random.randint(0,1)\n # convert to hex from string value\n tile = int(hex(tiles[rand_index]), 16)\n map_data[i].append(tile)\n return map_data\n\n\ndef draw_map(screen, map_data):\n MAP_HEIGHT = len(map_data) \n MAP_WIDTH = len(map_data[0])\n for row in range(MAP_HEIGHT):\n for col in range(MAP_WIDTH):\n screen.blit(textures[map_data[row][col]],\n (col*TILE_SIZE, row*TILE_SIZE)) \n", "sub_path": "004-Generating Tile Map/tilemap.py", "file_name": "tilemap.py", "file_ext": "py", "file_size_in_byte": 1158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pygame.Surface", "line_number": 15, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 34, "usage_type": "call"}]} +{"seq_id": "219260470", "text": "import os, numpy as np\nfrom .chunk_array import chunk_array\nfrom .gdown import download_file_from_google_drive\n#install dependencies\n# os.system('python3 -m pip install --upgrade pip')\n# os.system('python3 -m pip install gdown')\n#suppose get_txt.sh has run\n# os.system('chmod +x get_txt.sh')\n# os.system(f'./get_txt.sh {txt_id}')\n\ndef run_downloader(gid):\n\ttxt_ic_fn='ic/ic1800x1800.npz'\n\tdestination=txt_ic_fn\n\tretval=download_file_from_google_drive(gid, destination)\n\treturn None\n\ndef get_gid(txt_id):\n\timport random\n\tdef decision(probability):\n\t return random.random() < probability\n\t#two gid's per texture lowers the load on google drive servers\n\tif txt_id==0:#at time, 1210\n\t\tif decision(0.5):\n\t\t\tgid='1OYtQNnp5KnGfKMkskk7GeDQSCe3Mo7Gu'\n\t\telse:\n\t\t\tgid='1LTQxE9sacdb3BidFYeefqKUzEA_HiSOu'\n\tif txt_id==1:#at time, 2020\n\t\tif decision(0.5):\n\t\t\tgid='1td_6aQHFWzvunt1kU14ViW5DZ69rUhMD'\n\t\telse:\n\t\t\tgid='1qf2-Cf5Bbfjos5QDxp2FZtyJoL3FU4zO'\n\tif txt_id==2:#at time, 2830\n\t\tif decision(0.5):\n\t\t\tgid='12dLQ_YFwSAvuuZc1lhNsKPcv4QXZB86u'\n\t\telse:\n\t\t\tgid='1MCM6hVxC0Ch73ZnK97PKhjPChHI0PRxx'\n\tif txt_id==3:#at time, 3640\n\t\tif decision(0.5):\n\t\t\tgid='14SipoA-gemvfyuA5v9tAUQRP3Firmu8G'\n\t\telse:\n\t\t\tgid='1vmeI5SyyveaZ00qEeiqb04MVDssw9s5p'\n\treturn gid\n\ndef download_txt(txt_id,worker_dir):\n\t'''returns the first gdrive download file found in the directory, worker_dir.'''\n\tos.chdir(worker_dir)\n\tif not os.path.exists('ic'):\n\t\tos.mkdir('ic')\n\tgid=get_gid(txt_id)\n\trun_downloader(gid=gid)\n\t# cmd=f'gdown https://drive.google.com/uc?id={gid} -O ic/ic1800x1800.npz'\n\t# os.system(cmd)#at time, 1210\n\t# if txt_id==0:\n\t# \t# run_downloader(gid='1OYtQNnp5KnGfKMkskk7GeDQSCe3Mo7Gu')\n\t# \tos.system('gdown https://drive.google.com/uc?id=1OYtQNnp5KnGfKMkskk7GeDQSCe3Mo7Gu -O ic/ic1800x1800.npz')#at time, 1210\n\t# if txt_id==1:\n\t# \t# run_downloader(gid='1td_6aQHFWzvunt1kU14ViW5DZ69rUhMD')\n\t# \tos.system('gdown https://drive.google.com/uc?id=1td_6aQHFWzvunt1kU14ViW5DZ69rUhMD -O ic/ic1800x1800.npz')#at time, 1210\n\t# if txt_id==2:\n\t# \t# run_downloader(gid='12dLQ_YFwSAvuuZc1lhNsKPcv4QXZB86u')\n\t# \tos.system('gdown https://drive.google.com/uc?id=12dLQ_YFwSAvuuZc1lhNsKPcv4QXZB86u -O ic/ic1800x1800.npz')#at time, 1210\n\t# if txt_id==3:\n\t# \t# run_downloader(gid='14SipoA-gemvfyuA5v9tAUQRP3Firmu8G')\n\tos.chdir(worker_dir)\n\ttxt=load_buffer('ic/ic1800x1800.npz')[0]#,allow_pickle=True)\n\t# txt=load_buffer('ic/ic1800x1800.npz')[0]#,allow_pickle=True)\n\treturn txt\n\ndef get_txt_lst(txt_id1,width,height,worker_dir):\n\ttxt_in=download_txt(txt_id1,worker_dir)\n\tarray_lst = chunk_array(txt_in, width, height, typeout='float64')\n\treturn array_lst\n\ndef get_txt(txt_id1,txt_id2,width,height,worker_dir):\n\tarray_lst=get_txt_lst(txt_id1,width,height,worker_dir)\n\t# N=len(array_lst)\n\ttry:\n\t\ttxt=array_lst[txt_id2]\n\texcept IndexError as e:\n\t\timport random\n\t\tprint (f'IndexError for {(width,txt_id1,txt_id2)}')\n\t\tprint ( e )\n\t\ttxt_id2=random.randint(0,len(array_lst)-1)\n\t\tprint (f'Choosing txt_id2={txt_id2}...')\n\t\ttxt=array_lst[txt_id2]#-1]\n\n\treturn txt\n\ndef load_buffer(data_dir,**kwargs):\n\tif data_dir[-4:]=='.npy':\n\t\ttxt = np.load(data_dir)\n\t\treturn txt\n\telif data_dir[-4:]=='.npz':\n\t\ttxt = np.load(data_dir,**kwargs)\n\t\ttxt = txt[txt.files[0]] #only take the first buffer because there's typically one\n\t\treturn txt\n\telse:\n\t\tprint(f\"Warning: file format not supported {data_dir}.\")\n\t\traise Exception(f\"Warning: file format not supported {data_dir}.\")\n\n\nif __name__=='__main__':\n\tos.get_cwd()\n\tfor input_fn in sys.argv[1:]:\n\t\tdownload_txt(txt_id,worker_dir)", "sub_path": "notebooks/lib/utils/get_txt_npz.py", "file_name": "get_txt_npz.py", "file_ext": "py", "file_size_in_byte": 3510, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "gdown.download_file_from_google_drive", "line_number": 14, "usage_type": "call"}, {"api_name": "random.random", "line_number": 20, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 48, "usage_type": "call"}, {"api_name": "os.chdir", "line_number": 64, "usage_type": "call"}, {"api_name": "chunk_array.chunk_array", "line_number": 71, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 94, "usage_type": "call"}, {"api_name": "os.get_cwd", "line_number": 103, "usage_type": "call"}]} +{"seq_id": "55263482", "text": "#!/usr/bin/env python2\n\n\"\"\" Utility script that finds .coverage files under the current directory and combines them into a single file, deleting them as it goes.\n\"\"\"\n\nfrom __future__ import print_function\nimport time, sys, os.path\nimport coverage\n\ndef main(args):\n\tif args:\n\t\tprint(\"Run this tool from the directory you wish to crawl for .coverage files. \")\n\t\tprint(\"Has no arguments. Produces\")\n\t\treturn\n\t\n\tprint(\"Searching for .coverage* files under %s\"%os.path.normpath('.'))\n\tcov = []\n\tfor (dirpath, dirnames, filenames) in os.walk('.'):\n\t\tfor f in filenames:\n\t\t\tif f == '.coverage.combined': continue\n\t\t\tif f.startswith('.coverage'): cov.append(os.path.join(dirpath, f))\n\t\n\tif cov:\n\t\tprint('Found %d coverage file(s)'%(len(cov)))\n\t\tdest = os.path.abspath('./.coverage.combined')\n\t\tc = coverage.Coverage(dest)\n\t\t# nb: combine automatically deletes all the files (!)\n\t\tc.combine(cov)\n\t\tc.save() \n\t\tprint('Saved combined coverage to: %s'%dest)\n\telse:\n\t\tprint(\"Nothing to do - no coverage files found\")\n\t\t\nif __name__ == \"__main__\":\n\tmain(sys.argv[1:])\n", "sub_path": "tests/combine_coverage.py", "file_name": "combine_coverage.py", "file_ext": "py", "file_size_in_byte": 1054, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.path.normpath", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 16, "usage_type": "name"}, {"api_name": "os.path.walk", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 21, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 21, "usage_type": "name"}, {"api_name": "os.path.path.abspath", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 25, "usage_type": "name"}, {"api_name": "coverage.Coverage", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 35, "usage_type": "attribute"}]} +{"seq_id": "620009", "text": "import json\nfrom collections import defaultdict\nfrom gzip import decompress\nfrom urllib import request\n\nwith request.urlopen('http://eshop-checker.xyz/games.json', timeout=3) as f:\n # we only need the price of each game\n # each element in `data` looks like {'CA': 123, 'ZA': 234} <= {AREA: PRICE}\n data = map(lambda info: info['price'],\n json.loads(decompress(f.read()))['list'])\n\nwith request.urlopen('http://eshop-checker.xyz/beta/statics/conv_rate.json', timeout=3) as f:\n rate = json.loads(decompress(f.read()))['rates']\n\n# 2 char country code to 3 char currency code\n# https://www.nationsonline.org/oneworld/country_code_list.htm\n# https://www.iban.com/currency-codes\ncountry_to_currency_conversion = {'CA': 'CAD',\n 'MX': 'MXN',\n 'US': 'USD',\n 'DK': 'DKK',\n 'EE': 'EUR',\n 'FI': 'EUR',\n 'IE': 'EUR',\n 'LV': 'EUR',\n 'LT': 'EUR',\n 'NO': 'NOK',\n 'SE': 'SEK',\n 'GB': 'GBP',\n 'HR': 'EUR',\n 'CY': 'EUR',\n 'GR': 'EUR',\n 'IT': 'EUR',\n 'MT': 'EUR',\n 'PT': 'EUR',\n 'SI': 'EUR',\n 'ES': 'EUR',\n 'BG': 'EUR',\n 'CZ': 'CZK',\n 'HU': 'EUR',\n 'PL': 'PLN',\n 'RO': 'EUR',\n 'RU': 'RUB',\n 'SK': 'EUR',\n 'AT': 'EUR',\n 'BE': 'EUR',\n 'FR': 'EUR',\n 'DE': 'EUR',\n 'LU': 'EUR',\n 'NL': 'EUR',\n 'CH': 'CHF',\n 'AU': 'AUD',\n 'NZ': 'NZD',\n 'ZA': 'ZAR',\n 'JP': 'JPY',\n }\n\n\nresult = defaultdict(dict)\nfor game in data:\n for country in game:\n result[country]['count'] = result[country].get('count', 0) + 1\n\n normalized_price =\\\n ((area, price / rate[country_to_currency_conversion[area]])\n for area, price in game.items())\n min_ = min(normalized_price, key=lambda info: info[1])\n country = min_[0]\n result[country]['win'] = result[country].get('win', 0) + 1\n\nmost_count = max(result, key=lambda country: result[country].get('win', 0))\nprint('{country} has {count} games with good prices'\n .format(country=most_count, count=result[most_count]['win']))\n\nmost_rate = max(result, key=lambda country: result[country].get('win', 0) / result[country]['count'])\nprint('{country} has {percent}% of games with good prices'\n .format(country=most_rate, percent=round(result[most_rate]['win'] / result[most_rate]['count'] * 100, 2)))\n\nprint(result)\n", "sub_path": "find.py", "file_name": "find.py", "file_ext": "py", "file_size_in_byte": 3431, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "urllib.request.urlopen", "line_number": 6, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 6, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "gzip.decompress", "line_number": 10, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 12, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 12, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 13, "usage_type": "call"}, {"api_name": "gzip.decompress", "line_number": 13, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 59, "usage_type": "call"}]} +{"seq_id": "170746439", "text": "from mxnet import autograd, nd\nfrom mxnet.gluon import data as gdata\nfrom mxnet.gluon import nn\nfrom mxnet import init\nfrom mxnet.gluon import loss as gloss\nfrom mxnet import gluon\n\nnum_inputs = 2\nnum_examples = 1000\ntrue_w = [2, -3.4]\ntrue_b = 4.2\nfeatures = nd.random.normal(scale=1, shape=(num_examples, num_inputs))\nlabels = true_w[0] * features[:, 0] + true_w[1] * features[:, 1] + true_b\nlabels += nd.random.normal(scale=0.01, shape=labels.shape)\n\nbatch_size = 10\ndataset = gdata.ArrayDataset(features, labels)\n#随机读取小批量\ndata_iter = gdata.DataLoader(dataset, batch_size, shuffle=True)\n\nnet = nn.Sequential()\nnet.add(nn.Dense(1)) #Dense定义该层输出个数为 1。全连接:Dense\n\n#初始化模型参数\nnet.initialize(init.Normal(sigma=0.01)) #指定权重参数每个元素将在初始化时随机采样于均值为 0 标准差为 0.01 的正态分布。\n\n#定义损失函数\nloss = gloss.L2Loss() # 平⽅损失⼜称 L2 范数损失。\n\n#定义优化算法 学习率的数值一般设置为1/batch_size\ntrainer = gluon.Trainer(net.collect_params(), 'sgd', {'learning_rate': 0.03}) #指定学习率为 0.03 的小批量随机梯度下降(sgd)为优化算法 些参数可以通过 collect_params 函数获取\n\n#训练模型\nnum_epochs = 3\nfor epoch in range(1, num_epochs + 1):\n for X, y in data_iter:\n with autograd.record():\n l = loss(net(X), y)\n l.backward()\n trainer.step(batch_size) #迭代模型参数 指明批量⼤小,从而对批量中样本梯度求平均\n l = loss(net(features), labels)\n print('epoch %d, loss: %f' % (epoch, l.mean().asnumpy()))\n\n#获取模型参数 与构造数据集的权重偏差作对比\ndense = net[0]\nprint(true_w, dense.weight.data())\n\nprint(true_b, dense.bias.data())\n#查看梯度\nprint(dense.weight.grad())", "sub_path": "regression/linearRegressionSimple.py", "file_name": "linearRegressionSimple.py", "file_ext": "py", "file_size_in_byte": 1823, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "mxnet.nd.random.normal", "line_number": 12, "usage_type": "call"}, {"api_name": "mxnet.nd.random", "line_number": 12, "usage_type": "attribute"}, {"api_name": "mxnet.nd", "line_number": 12, "usage_type": "name"}, {"api_name": "mxnet.nd.random.normal", "line_number": 14, "usage_type": "call"}, {"api_name": "mxnet.nd.random", "line_number": 14, "usage_type": "attribute"}, {"api_name": "mxnet.nd", "line_number": 14, "usage_type": "name"}, {"api_name": "mxnet.gluon.data.ArrayDataset", "line_number": 17, "usage_type": "call"}, {"api_name": "mxnet.gluon.data", "line_number": 17, "usage_type": "name"}, {"api_name": "mxnet.gluon.data.DataLoader", "line_number": 19, "usage_type": "call"}, {"api_name": "mxnet.gluon.data", "line_number": 19, "usage_type": "name"}, {"api_name": "mxnet.gluon.nn.Sequential", "line_number": 21, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn", "line_number": 21, "usage_type": "name"}, {"api_name": "mxnet.gluon.nn.Dense", "line_number": 22, "usage_type": "call"}, {"api_name": "mxnet.gluon.nn", "line_number": 22, "usage_type": "name"}, {"api_name": "mxnet.init.Normal", "line_number": 25, "usage_type": "call"}, {"api_name": "mxnet.init", "line_number": 25, "usage_type": "name"}, {"api_name": "mxnet.gluon.loss.L2Loss", "line_number": 28, "usage_type": "call"}, {"api_name": "mxnet.gluon.loss", "line_number": 28, "usage_type": "name"}, {"api_name": "mxnet.gluon.Trainer", "line_number": 31, "usage_type": "call"}, {"api_name": "mxnet.gluon", "line_number": 31, "usage_type": "name"}, {"api_name": "mxnet.autograd.record", "line_number": 37, "usage_type": "call"}, {"api_name": "mxnet.autograd", "line_number": 37, "usage_type": "name"}]} +{"seq_id": "398909482", "text": "from urllib.request import urlopen, Request\r\nfrom bs4 import BeautifulSoup\r\nimport warnings, os\r\nwarnings.filterwarnings('ignore')\r\n\r\n\r\nif __name__ == \"__main__\":\r\n\r\n\r\n url = \"http://the-japan-news.com/\"\r\n url_response = urlopen(url)\r\n news_html = BeautifulSoup(url_response)\r\n #print(news_html)\r\n\r\n update_url_list = []\r\n for news_block in news_html.find_all(\"a\", itemprop=\"url\"):\r\n #print(news_block[\"href\"])\r\n news_url = \"http://the-japan-news.com\" + news_block[\"href\"]\r\n #print(page_url)\r\n update_url_list.append(news_url)\r\n #print(update_url_list)\r\n\r\n\r\n old_url_list = [] # 紀錄之前爬過的新聞網址\r\n # 開啟紀錄全部新聞網址的檔案\r\n if os.path.exists(\"./yomiuri_news_url_tmp.txt\"):\r\n with open(\"./yomiuri_news_url_tmp.txt\", \"r\", encoding=\"utf-8\") as f:\r\n old_url_list = f.read().split(\"\\n\")\r\n old_url_list.remove(\"\")\r\n # print(\"old_url_list:\", len(old_url_list))\r\n\r\n url_list = [] # 紀錄更新的新聞網址\r\n # 不記錄重複的新聞網址\r\n for url in update_url_list:\r\n if not url in old_url_list:\r\n url_list.append(url)\r\n # print(\"update:\", len(url_list))\r\n\r\n if not url_list == []:\r\n url_list.extend(old_url_list)\r\n # print(url_list)\r\n\r\n with open(\"./yomiuri_news_url_tmp.txt\", \"w\", encoding=\"utf-8\") as f:\r\n for url in url_list:\r\n f.write(str(url + \"\\n\"))\r\n\r\n else:\r\n with open(\"./yomiuri_news_url_tmp.txt\", \"w\", encoding=\"utf-8\") as f:\r\n for url in update_url_list:\r\n f.write(str(url + \"\\n\"))", "sub_path": "yomiuri_Url.py", "file_name": "yomiuri_Url.py", "file_ext": "py", "file_size_in_byte": 1690, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "warnings.filterwarnings", "line_number": 4, "usage_type": "call"}, {"api_name": "urllib.request.urlopen", "line_number": 11, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}]} +{"seq_id": "409557667", "text": "from django.db import models\nfrom wagtail.core.fields import StreamField\nfrom wagtail.core import blocks\nfrom wagtail.core.models import Page\nfrom wagtail.core.fields import RichTextField\nfrom wagtail.admin.edit_handlers import FieldPanel, InlinePanel, StreamFieldPanel, MultiFieldPanel, FieldRowPanel\nfrom wagtail.images.blocks import ImageChooserBlock\nfrom wagtail.core.blocks import BlockQuoteBlock\nfrom django import template\n\n\nfrom modelcluster.fields import ParentalKey\nfrom wagtail.admin.edit_handlers import (\n FieldPanel, FieldRowPanel,\n InlinePanel, MultiFieldPanel, PageChooserPanel\n)\nfrom wagtail.core.fields import RichTextField\nfrom wagtail.contrib.forms.models import AbstractEmailForm, AbstractFormField\nfrom wagtail.contrib.settings.models import BaseSetting, register_setting\n\n# Create your models here.\n\n\nclass HomePage(Page):\n body = RichTextField(blank=True)\n content_panels = Page.content_panels + [\n FieldPanel('body', classname=\"full\"),\n ]\n\n def get_context(self, request):\n context = super().get_context(request)\n components = self.get_children()\n context['featured'] = components.type(FeaturedIndexPage)\n context['teams'] = components.type(TeamIndexPage)\n context['activities'] = components.type(ActivityIndexPage)\n context['staticPage'] = components.type(StaticPage)\n return context\n\n\n@register_setting\nclass Footer(BaseSetting):\n title_left = models.CharField(max_length=250)\n list_item1_left = models.CharField(max_length=250, blank=True, null=True)\n list_item2_left = models.CharField(max_length=250, blank=True, null=True)\n list_item3_left = models.CharField(max_length=250, blank=True, null=True)\n list_item4_left = models.CharField(max_length=250, blank=True, null=True)\n list_item5_left = models.CharField(max_length=250, blank=True, null=True)\n list_item6_left = models.CharField(max_length=250, blank=True, null=True)\n logo_left = models.CharField(max_length=250, blank=True, null=True)\n\n title_right = models.CharField(max_length=250)\n list_item1_right = models.CharField(max_length=250, blank=True, null=True)\n list_item2_right = models.CharField(max_length=250, blank=True, null=True)\n list_item3_right = models.CharField(max_length=250, blank=True, null=True)\n list_item4_right = models.CharField(max_length=250, blank=True, null=True)\n list_item5_right = models.CharField(max_length=250, blank=True, null=True)\n list_item6_right = models.CharField(max_length=250, blank=True, null=True)\n logo_right = models.ImageField(\n upload_to='images/', default='images/areto.jpg')\n\n content_panels = Page.content_panels + [\n FieldPanel('title_left '),\n FieldPanel('list_item1_left'),\n FieldPanel('list_item2_left'),\n FieldPanel('list_item3_left'),\n FieldPanel('list_item4_left'),\n FieldPanel('list_item5_left'),\n\n FieldPanel('title_right'),\n FieldPanel('list_item1_right'),\n FieldPanel('list_item2_right'),\n FieldPanel('list_item3_right'),\n FieldPanel('list_item4_right'),\n FieldPanel('list_item5_right'),\n ]\n\n class Meta:\n verbose_name = 'footer settings'\n\n\n@register_setting\nclass NewsletterCustomSettings(BaseSetting):\n newsletter_form_page = models.ForeignKey(\n 'wagtailcore.Page', null=True, on_delete=models.SET_NULL)\n content_panels = [\n # note the page type declared within the pagechooserpanel\n PageChooserPanel('newsletter_form_page', [\n 'web.NewsletterFormPage']),\n ]\n\n\n# featured area\n\n\nclass FeaturedIndexPage(Page):\n intro = RichTextField(blank=True)\n content_panels = Page.content_panels + [\n FieldPanel('intro', classname='full')\n ]\n\n\nclass FeaturedPage(Page):\n body = StreamField([\n ('image', ImageChooserBlock(blank=True)),\n ('button', blocks.CharBlock(blank=True)),\n ('paragraph', blocks.CharBlock(blank=True)),\n ])\n\n content_panels = Page.content_panels + [\n StreamFieldPanel('body'),\n ]\n\n# team area\n\n\nclass TeamIndexPage(Page):\n intro = RichTextField(blank=True)\n content_panels = Page.content_panels + [\n FieldPanel('intro', classname='full')\n ]\n\n\nclass TeamMemberPage(Page):\n\n body = StreamField([\n ('image', ImageChooserBlock(blank=True)),\n ('name', blocks.CharBlock(blank=True)),\n ('paragraph', blocks.CharBlock(blank=True)),\n ('button', blocks.CharBlock(blank=True)),\n ])\n\n content_panels = Page.content_panels + [\n StreamFieldPanel('body'),\n ]\n\n\nclass ActivityIndexPage(Page):\n intro = RichTextField(blank=True)\n content_panels = Page.content_panels + [\n FieldPanel('intro', classname='full')\n ]\n\n\nclass ActivityPage(Page):\n body = StreamField([\n ('image', ImageChooserBlock(blank=True)),\n ('name', blocks.CharBlock(blank=True)),\n ])\n\n content_panels = Page.content_panels + [\n StreamFieldPanel('body'),\n ]\n\n\nclass StaticPage(Page):\n body = StreamField([\n ('title', blocks.CharBlock(blank=True)),\n ('lead', blocks.CharBlock(blank=True)),\n ('paragraph', blocks.RichTextBlock()),\n ('image', ImageChooserBlock(blank=True)),\n ('quote', BlockQuoteBlock(blank=True)),\n ])\n\n content_panels = Page.content_panels + [\n StreamFieldPanel('body'),\n ]\n\n\nclass FormField(AbstractFormField):\n newsletter_form_page = ParentalKey('NewsletterFormPage', on_delete=models.CASCADE,\n related_name='form_fields')\n\n\nclass NewsletterFormPage(AbstractEmailForm):\n intro = RichTextField(blank=True)\n thank_you_title = models.CharField(max_length=250, blank=True, null=True)\n thank_you_text = RichTextField(blank=True)\n\n content_panels = AbstractEmailForm.content_panels + [\n FieldPanel('intro', classname=\"full\"),\n InlinePanel('form_fields', label=\"Form fields\"),\n FieldPanel('thank_you_title'),\n FieldPanel('thank_you_text', classname=\"full\"),\n MultiFieldPanel([\n FieldRowPanel([\n FieldPanel('from_address', classname=\"col6\"),\n FieldPanel('to_address', classname=\"col6\"),\n ]),\n FieldPanel('subject'),\n ], \"Email\"),\n ]\n", "sub_path": "holidaydream_project/web/models.py", "file_name": "models.py", "file_ext": "py", "file_size_in_byte": 6316, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "wagtail.core.models.Page", "line_number": 24, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 25, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 26, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 26, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 27, "usage_type": "call"}, {"api_name": "wagtail.contrib.settings.models.BaseSetting", "line_number": 41, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 42, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 42, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 43, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 43, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 44, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 44, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 45, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 45, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 46, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 46, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 47, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 47, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 48, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 48, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 49, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 49, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 51, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 51, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 52, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 52, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 53, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 53, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 54, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 54, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 55, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 55, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 56, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 56, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 57, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 57, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 58, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 58, "usage_type": "name"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 61, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 61, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 62, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 63, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 64, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 65, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 66, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 67, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 69, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 70, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 71, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 72, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 73, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 74, "usage_type": "call"}, {"api_name": "wagtail.contrib.settings.models.register_setting", "line_number": 40, "usage_type": "name"}, {"api_name": "wagtail.contrib.settings.models.BaseSetting", "line_number": 82, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 83, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 83, "usage_type": "name"}, {"api_name": "django.db.models.SET_NULL", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 84, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.PageChooserPanel", "line_number": 87, "usage_type": "call"}, {"api_name": "wagtail.contrib.settings.models.register_setting", "line_number": 81, "usage_type": "name"}, {"api_name": "wagtail.core.models.Page", "line_number": 95, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 96, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 97, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 97, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 98, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 102, "usage_type": "name"}, {"api_name": "wagtail.core.fields.StreamField", "line_number": 103, "usage_type": "call"}, {"api_name": "wagtail.images.blocks.ImageChooserBlock", "line_number": 104, "usage_type": "call"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 105, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 105, "usage_type": "name"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 106, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 106, "usage_type": "name"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 109, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 109, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.StreamFieldPanel", "line_number": 110, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 116, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 117, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 118, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 118, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 119, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 123, "usage_type": "name"}, {"api_name": "wagtail.core.fields.StreamField", "line_number": 125, "usage_type": "call"}, {"api_name": "wagtail.images.blocks.ImageChooserBlock", "line_number": 126, "usage_type": "call"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 127, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 127, "usage_type": "name"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 128, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 128, "usage_type": "name"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 129, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 129, "usage_type": "name"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 132, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 132, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.StreamFieldPanel", "line_number": 133, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 137, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 138, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 139, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 139, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 140, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 144, "usage_type": "name"}, {"api_name": "wagtail.core.fields.StreamField", "line_number": 145, "usage_type": "call"}, {"api_name": "wagtail.images.blocks.ImageChooserBlock", "line_number": 146, "usage_type": "call"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 147, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 147, "usage_type": "name"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 150, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 150, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.StreamFieldPanel", "line_number": 151, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page", "line_number": 155, "usage_type": "name"}, {"api_name": "wagtail.core.fields.StreamField", "line_number": 156, "usage_type": "call"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 157, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 157, "usage_type": "name"}, {"api_name": "wagtail.core.blocks.CharBlock", "line_number": 158, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 158, "usage_type": "name"}, {"api_name": "wagtail.core.blocks.RichTextBlock", "line_number": 159, "usage_type": "call"}, {"api_name": "wagtail.core.blocks", "line_number": 159, "usage_type": "name"}, {"api_name": "wagtail.images.blocks.ImageChooserBlock", "line_number": 160, "usage_type": "call"}, {"api_name": "wagtail.core.blocks.BlockQuoteBlock", "line_number": 161, "usage_type": "call"}, {"api_name": "wagtail.core.models.Page.content_panels", "line_number": 164, "usage_type": "attribute"}, {"api_name": "wagtail.core.models.Page", "line_number": 164, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.StreamFieldPanel", "line_number": 165, "usage_type": "call"}, {"api_name": "wagtail.contrib.forms.models.AbstractFormField", "line_number": 169, "usage_type": "name"}, {"api_name": "modelcluster.fields.ParentalKey", "line_number": 170, "usage_type": "call"}, {"api_name": "django.db.models.CASCADE", "line_number": 170, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 170, "usage_type": "name"}, {"api_name": "wagtail.contrib.forms.models.AbstractEmailForm", "line_number": 174, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 175, "usage_type": "call"}, {"api_name": "django.db.models.CharField", "line_number": 176, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 176, "usage_type": "name"}, {"api_name": "wagtail.core.fields.RichTextField", "line_number": 177, "usage_type": "call"}, {"api_name": "wagtail.contrib.forms.models.AbstractEmailForm.content_panels", "line_number": 179, "usage_type": "attribute"}, {"api_name": "wagtail.contrib.forms.models.AbstractEmailForm", "line_number": 179, "usage_type": "name"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 180, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.InlinePanel", "line_number": 181, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 182, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 183, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.MultiFieldPanel", "line_number": 184, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldRowPanel", "line_number": 185, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 186, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 187, "usage_type": "call"}, {"api_name": "wagtail.admin.edit_handlers.FieldPanel", "line_number": 189, "usage_type": "call"}]} +{"seq_id": "323389052", "text": "\"\"\"\nProcess syllables and save data.\n\n\"\"\"\n__author__ = \"Jack Goffinet\"\n__date__ = \"December 2018 - April 2019\"\n\n\nimport os\nimport numpy as np\nimport h5py\nfrom datetime import datetime, timedelta\nfrom scipy.io import wavfile, loadmat\nfrom scipy.signal import stft\n\nfrom time import strptime, mktime, localtime\n\nfrom skimage.transform import resize\n\nfrom scipy.interpolate import interp1d\n\nimport matplotlib.pyplot as plt\nplt.switch_backend('agg')\n\n\n\n# Constants\nEPSILON = 1e-12\n\n\n\ndef process_sylls(load_dir, save_dir, p, noise_detector=None):\n \"\"\"\n Main method: process files in and save to .\n\n Parameters\n ----------\n load_dir : string\n Directory to load files from\n save_fir : string\n Directory to save files to\n p : dictionary\n Parameters\n noise_detector : noise_detection.NoiseDetector or None\n Throws away bad syllables\n\n Returns\n -------\n\n Notes\n -----\n \"\"\"\n\n if save_dir[-1] != '/':\n save_dir += '/'\n sylls_per_file = p['sylls_per_file']\n num_freq_bins = p['num_freq_bins']\n num_time_bins = p['num_time_bins']\n if not os.path.exists(save_dir):\n os.makedirs(save_dir)\n filenames = [os.path.join(load_dir, i) for i in os.listdir(load_dir) if i[-4:] in ['.wav', '.mat']]\n np.random.shuffle(filenames)\n if p['seg_params']['algorithm'] == get_onsets_offsets_from_file:\n filenames = [i for i in filenames if os.path.exists('.'.join(i.split('.')[:-1]) + '.txt')]\n write_file_num = 0\n syll_data = {\n 'specs':[],\n 'times':[],\n 'file_times':[],\n 'durations':[],\n 'filenames':[],\n }\n print(\"Processing audio files in\", load_dir)\n for load_filename in filenames:\n start_time = time_from_filename(load_filename)\n # Get a spectrogram.\n spec, f, dt = get_spec(load_filename, p)\n # Collect syllable onsets and offsets.\n if 'f' not in p['seg_params']:\n p['seg_params']['f'] = f\n # Get onsets and offsets.\n if p['seg_params']['algorithm'] == get_onsets_offsets_from_file:\n t_onsets, t_offsets = get_onsets_offsets_from_file(load_filename, dt)\n else:\n t_onsets, t_offsets = p['seg_params']['algorithm'](spec, dt, p['seg_params'])\n t_durations = [(b-a+1)*dt for a,b in zip(t_onsets, t_offsets)]\n # Retrieve spectrograms and start times for each detected syllable.\n t_specs, t_times = get_syll_specs(t_onsets, t_offsets, spec, start_time, dt, p)\n # Find noise and expunge it.\n if noise_detector is not None:\n mask = noise_detector.batch_classify(t_specs, threshold=0.5)\n for i in range(len(mask))[::-1]:\n if not mask[i]:\n del t_durations[i]\n del t_specs[i]\n del t_times[i]\n # Add the remaining syllables to .\n syll_data['durations'] += t_durations\n syll_data['specs'] += t_specs\n syll_data['times'] += t_times\n syll_data['file_times'] += [i - start_time for i in t_times]\n syll_data['filenames'] += len(t_durations)*[load_filename.split('/')[-1]]\n # Write a file when we have enough syllables.\n while len(syll_data['durations']) >= sylls_per_file:\n save_filename = save_dir + \"syllables_\"\n save_filename += str(write_file_num).zfill(3) + '.hdf5'\n with h5py.File(save_filename, \"w\") as f:\n # Zero-pad the spectrograms and add them to the file.\n temp = np.zeros((sylls_per_file, num_freq_bins, num_time_bins),\n dtype='float')\n syll_specs = syll_data['specs']\n for i in range(sylls_per_file):\n gap = max(0, (num_time_bins - syll_specs[i].shape[1]) // 2)\n try:\n temp[i,:,gap:gap+syll_specs[i].shape[1]] = syll_specs[i][:,:num_time_bins]\n except:\n print(\"caught in process_sylls\")\n print(temp.shape)\n print(gap)\n print(syll_specs[i].shape)\n quit()\n f.create_dataset('specs', data=temp)\n # Then add the rest.\n for k in ['durations', 'times', 'file_times']:\n f.create_dataset(k, data=np.array(syll_data[k][:sylls_per_file]))\n temp = [save_dir + i for i in syll_data['filenames'][:sylls_per_file]]\n f.create_dataset('filenames', data=np.array(temp).astype('S'))\n # Remove the written data from temporary storage.\n for k in syll_data:\n syll_data[k] = syll_data[k][sylls_per_file:]\n write_file_num += 1\n if 'max_num_syllables' in p and write_file_num*sylls_per_file >= p['max_num_syllables']:\n return\n\n\ndef get_syll_specs(onsets, offsets, spec, start_time, dt, p, audio):\n \"\"\"\n Return a list of spectrograms, one for each syllable.\n \"\"\"\n syll_specs, syll_times, syll_audio = [], [], []\n # For each syllable...\n for t1, t2 in zip(onsets, offsets):\n # Take a slice of audio\n temp_audio = audio[t1:t2+1]\n # Take a slice of the spectrogram.\n temp_spec = spec[:,t1:t2+1]\n # Within-syllable normalization.\n temp_spec -= np.percentile(temp_spec, 10.0)\n temp_spec[temp_spec<0.0] = 0.0\n temp_spec /= np.max(temp_spec)\n # Switch to square root duration.\n if p['time_stretch']:\n new_dur = int(round(temp_spec.shape[1]**0.5 * p['num_time_bins']**0.5))\n temp_spec = resize(temp_spec, (temp_spec.shape[0], new_dur), anti_aliasing=True, mode='reflect')\n # Collect spectrogram, duration, & onset time.\n syll_specs.append(temp_spec)\n syll_times.append(start_time + t1*dt) # in seconds\n syll_audio.append(temp_audio)\n return syll_specs, syll_times,syll_audio\n\n\ndef get_audio(filename, p, start_index=None, stop_index=None):\n \"\"\"Get a waveform given a filename.\"\"\"\n # Make sure the samplerate is correct and the audio is mono.\n if filename[-4:] == '.wav':\n temp_fs, audio = wavfile.read(filename)\n elif filename[-4:] == '.mat':\n d = loadmat(filename)\n audio = d['spike2Chunk'].reshape(-1)\n temp_fs = d['fs'][0,0]\n assert temp_fs == p['fs'], \"found fs: \"+str(temp_fs)+\", expected: \"+str(p['fs'])\n if len(audio.shape) > 1:\n audio = audio[0,:]\n if start_index is not None and stop_index is not None:\n start_index = max(start_index, 0)\n audio = audio[start_index:stop_index]\n return audio\n\n\ndef get_spec(filename, p, start_index=None, stop_index=None):\n \"\"\"Get a spectrogram.\"\"\"\n audio = get_audio(filename, p, start_index=start_index, stop_index=stop_index)\n f, t, spec = stft(audio, fs=p['fs'], nperseg=p['nperseg'], noverlap=p['noverlap'])\n spec = np.log(np.abs(spec) + EPSILON)\n spec -= p['seg_params']['spec_thresh']\n spec[spec < 0.0] = 0.0\n # Switch to mel frequency spacing.\n if p['mel']:\n new_f = np.linspace(mel(p['min_freq']), mel(p['max_freq']), p['num_freq_bins'], endpoint=True)\n new_f = inv_mel(new_f)\n new_f[0] = f[0] # Correct for numerical errors.\n new_f[-1] = f[-1]\n else:\n new_f = np.linspace(p['min_freq'], p['max_freq'], p['num_freq_bins'], endpoint=True)\n new_spec = np.zeros((p['num_freq_bins'], spec.shape[1]), dtype='float')\n for j in range(spec.shape[1]):\n interp = interp1d(f, spec[:,j], kind='linear', assume_sorted=True)\n new_spec[:,j] = interp(new_f)\n spec = new_spec\n f = new_f\n return spec, f, t[1] - t[0], audio\n\n\n# # Old function.\n# def get_spec(filename, p, start_index=None, stop_index=None):\n# \"\"\"Get a spectrogram given a filename.\"\"\"\n# audio = get_audio(filename, p, start_index=start_index, stop_index=stop_index)\n# # Convert to a magnitude-only spectrogram.\n# f, t, Zxx = stft(audio, fs=p['fs'], nperseg=p['nperseg'],\n# noverlap=p['noverlap'])\n# i1 = np.searchsorted(f, p['min_freq'])\n# i2 = np.searchsorted(f, p['max_freq'])\n# f = f[i1:i2]\n# spec = np.log(np.abs(Zxx[i1:i2,:]) + EPSILON)\n# # Denoise.\n# spec -= p['seg_params']['spec_thresh']\n# spec[spec<0.0] = 0.0\n# # Switch to mel frequency spacing.\n# if p['mel']:\n# new_f = np.linspace(mel(f[0]), mel(f[-1]), p['num_freq_bins'], endpoint=True)\n# new_f = inv_mel(new_f)\n# new_f[0] = f[0] # Correct for numerical errors.\n# new_f[-1] = f[-1]\n# else:\n# new_f = np.linspace(f[0], f[-1], p['num_freq_bins'], endpoint=True)\n# new_spec = np.zeros((p['num_freq_bins'], spec.shape[1]), dtype='float')\n# for j in range(spec.shape[1]):\n# interp = interp1d(f, spec[:,j], kind='cubic')\n# new_spec[:,j] = interp(new_f)\n# spec = new_spec\n# f = new_f\n# return spec, f, t[1]-t[0], i1, i2\n\n\ndef tune_segmenting_params(load_dirs, p):\n \"\"\"Tune params by visualizing segmenting decisions.\"\"\"\n print('entered tune_segmenting_params')\n print(load_dirs)\n fs = p['fs']\n seg_params = p['seg_params']\n filenames = []\n for load_dir in load_dirs:\n print(load_dir)\n filenames += [os.path.join(load_dir, i) for i in os.listdir(load_dir) if i[-4:] in ['.wav', '.mat']]\n if len(filenames) == 0:\n print(\"Found no audio files!\")\n return\n if p['seg_params']['algorithm'] == get_onsets_offsets_from_file:\n filenames = [i for i in filenames if os.path.exists('.'.join(i.split('.')[:-1]) + '.txt')]\n filenames = np.array(filenames)\n filenames = np.random.choice(filenames, min(1000, len(filenames)), replace=False)\n file_lens = [get_wav_len(filename) for filename in filenames]\n dur_seconds = 2.0 * seg_params['max_dur']\n dur_samples = int(dur_seconds * fs)\n filenames, file_lens = np.array(filenames), np.array(file_lens, dtype='int')\n filenames = filenames[file_lens > 3 * dur_samples]\n file_lens = file_lens[file_lens > 3 * dur_samples]\n assert len(filenames) >= 1\n # Keep tuning params...\n while True:\n for key in seg_params:\n # Skip non-tunable parameters.\n if key in ['num_time_bins', 'num_freq_bins'] or not is_number(seg_params[key]):\n continue\n temp = 'not a valid option'\n while not is_number_or_empty(temp):\n temp = input('Set value for '+key+': ['+str(seg_params[key])+ '] ')\n if temp != '':\n seg_params[key] = float(temp)\n # Visualize segmenting decisions.\n temp = ''\n while temp != 'q' and temp != 'r':\n file_index = np.random.randint(len(filenames))\n filename = filenames[file_index]\n # Get spec & onsets/offsets.\n start_index = np.random.randint(file_lens[file_index] - 3*dur_samples)\n stop_index = start_index + 3*dur_samples\n spec, f, dt = get_spec(filename, p, start_index=start_index, stop_index=stop_index)\n\n if 'f' not in seg_params:\n seg_params['f'] = f\n if seg_params['algorithm'] == get_onsets_offsets_from_file:\n onsets, offsets = get_onsets_offsets_from_file(filename, dt)\n traces = []\n temp = int(round(start_index/fs/dt))\n onsets = [i-temp for i in onsets]\n offsets = [i-temp for i in offsets]\n else:\n onsets, offsets, traces = seg_params['algorithm'](spec, dt, seg_params=seg_params, return_traces=True)\n dur_t_bins = int(dur_seconds / dt)\n\n # Plot.\n i1 = dur_t_bins\n i2 = 2 * dur_t_bins\n t1, t2 = i1 * dt, i2 * dt\n _, axarr = plt.subplots(2,1, sharex=True)\n axarr[0].set_title(filename)\n axarr[0].imshow(spec[:,i1:i2], origin='lower', \\\n aspect='auto', \\\n extent=[t1, t2, f[0], f[-1]])\n for j in range(len(onsets)):\n if onsets[j] >= i1 and onsets[j] < i2:\n time = onsets[j] * dt\n for k in [0,1]:\n axarr[k].axvline(x=time, c='b', lw=0.5)\n if offsets[j] >= i1 and offsets[j] < i2:\n time = offsets[j] * dt\n for k in [0,1]:\n axarr[k].axvline(x=time, c='r', lw=0.5)\n for key in ['th_1', 'th_2', 'th_3']:\n if key in seg_params:\n axarr[1].axhline(y=seg_params[key], lw=0.5, c='b')\n xvals = np.linspace(t1, t2, i2-i1)\n for trace in traces:\n axarr[1].plot(xvals, trace[i1:i2])\n plt.savefig('temp.pdf')\n plt.close('all')\n if len([j for j in onsets if j>i1 and j 0:\n temp = input('Continue? [y] or [q]uit or [r]etune params: ')\n else:\n print(\"searching\")\n temp = 'y'\n if temp == 'q':\n return seg_params\n\n\ndef get_onsets_offsets_from_file(audio_filename, dt):\n onsets = []\n offsets = []\n filename = '.'.join(audio_filename.split('.')[:-1]) + '.txt'\n try:\n d = np.loadtxt(filename).reshape(-1,3)\n except:\n return onsets, offsets\n for i in range(len(d)):\n try:\n onsets.append(int(np.floor(d[i,1]/dt)))\n except:\n print(\"caught\")\n print(d)\n return onsets, offsets\n offsets.append(int(np.ceil(d[i,2]/dt))+1)\n if offsets[-1] - onsets[-1] >= 128:\n onsets = onsets[:-1]\n offsets = offsets[:-1]\n return onsets, offsets\n\n\ndef mel(a):\n return 1127 * np.log(1 + a / 700)\n\n\ndef inv_mel(a):\n return 700 * (np.exp(a / 1127) - 1)\n\n\ndef is_number_or_empty(s):\n if s == '':\n return True\n try:\n float(s)\n return True\n except:\n return False\n\n\ndef is_number(s):\n return type(s) == type(4) or type(s) == type(4.0)\n\n\ndef get_wav_len(filename):\n if filename[-4:] == '.wav':\n _, audio = wavfile.read(filename)\n elif filename[-4:] == '.mat':\n audio = loadmat(filename)['spike2Chunk'].reshape(-1)\n else:\n raise NotImplementedError\n return len(audio)\n\n\ndef time_from_filename(filename):\n \"\"\"Return time in seconds.\"\"\"\n \n try:\n# anchor = mktime(strptime(\"1899 12 29 19\", \"%Y %m %d %H\")) #SAP anchor time\n\n anchor = datetime(1899, 12, 29, 19) #SAP anchor time\n\n temp = filename.split('/')[-1].split('_')[1].split('.')\n\n day = float(temp[0])\n\n millisecond = float(temp[1])\n\n time = anchor + timedelta(days = day) + timedelta(milliseconds=millisecond)\n\n unix_start_time = datetime(1970,1,1);\n time_duration = time - unix_start_time\n time = time_duration.total_seconds()\n return time\n except:\n return 0.0\n\n\nif __name__ == '__main__':\n pass\n\n\n###\n", "sub_path": "preprocessing/preprocessing.py", "file_name": "preprocessing.py", "file_ext": "py", "file_size_in_byte": 15056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.pyplot.switch_backend", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "h5py.File", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 125, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.percentile", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 150, "usage_type": "call"}, {"api_name": "skimage.transform.resize", "line_number": 154, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.read", "line_number": 166, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 166, "usage_type": "name"}, {"api_name": "scipy.io.loadmat", "line_number": 168, "usage_type": "call"}, {"api_name": "scipy.signal.stft", "line_number": 183, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 184, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 194, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 195, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 197, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path", "line_number": 244, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 249, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 250, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 251, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 251, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 273, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 273, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 276, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 313, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 316, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 316, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 317, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 317, "usage_type": "name"}, {"api_name": "numpy.loadtxt", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 337, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 342, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 354, "usage_type": "call"}, {"api_name": "scipy.io.wavfile.read", "line_number": 373, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 373, "usage_type": "name"}, {"api_name": "scipy.io.loadmat", "line_number": 375, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 387, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 395, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 397, "usage_type": "call"}]} +{"seq_id": "381973207", "text": "#!/usr/bin/env python\n\nimport argparse\nimport configparser\nimport os\nimport re\nfrom batotodownloader import BatotoDownloader\nfrom mangahere_downloader import MangahereDownloader\n\ndef setup_arguments():\n parser = argparse.ArgumentParser()\n parser.add_argument('-u', '--url', help = 'Url of any page within a chapter', default='')\n parser.add_argument('-o', '--output', help = 'Archive output', default = \"/tmp/comics.cbz\")\n parser.add_argument('-c', '--chapters', type = int, help = 'How many chapters should be downloaded', default=1)\n parser.add_argument('-i', '--initialchapter', type = int, help = 'Initial number for Chapter sequence', default=1)\n parser.add_argument('-l', '--load', action=\"store_true\", help = 'Use previous option')\n return parser.parse_args()\n\ndef merge_arguments(args):\n config_path = os.path.expanduser(\"~/.mangadl.rc\")\n config = configparser.ConfigParser()\n config.read(config_path)\n\n if 'history' not in config:\n return\n\n history = config['history']\n\n if 'lastchapter' in history and args.initialchapter == 1:\n args.initialchapter = int (history['lastchapter'])\n\n if 'nexturl' in history and not args.url:\n args.url = history['nexturl']\n\n if 'chapters' in history and args.chapters == 1:\n args.chapters = int (history['chapters'])\n\n return args\n\nif __name__ == '__main__':\n args = setup_arguments()\n if (args.load):\n args = merge_arguments(args)\n\n if args.url:\n if re.search(r'bato(\\.)*to', args.url):\n downloader = BatotoDownloader(args.url, args.output, args.chapters, args.initialchapter)\n elif re.search(r'mangahere', args.url):\n downloader = MangahereDownloader(args.url, args.output, args.chapters, args.initialchapter)\n else:\n print (\"No suitable downloader\")\n downloader.download()\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1871, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "configparser.ConfigParser", "line_number": 21, "usage_type": "call"}, {"api_name": "re.search", "line_number": 46, "usage_type": "call"}, {"api_name": "batotodownloader.BatotoDownloader", "line_number": 47, "usage_type": "call"}, {"api_name": "re.search", "line_number": 48, "usage_type": "call"}, {"api_name": "mangahere_downloader.MangahereDownloader", "line_number": 49, "usage_type": "call"}]} +{"seq_id": "462142300", "text": "from django.urls import include, path\nfrom rest_framework import routers\n\nfrom application_form.api import views\n\napp_name = \"application_form\"\n\nrouter = routers.DefaultRouter()\nrouter.register(r\"haso_application\", views.HasoApplicationViewSet)\nrouter.register(r\"hitas_application\", views.HitasApplicationViewSet)\n\nurlpatterns = [\n path(\"\", include(router.urls)),\n]\n", "sub_path": "application_form/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "rest_framework.routers.DefaultRouter", "line_number": 8, "usage_type": "call"}, {"api_name": "rest_framework.routers", "line_number": 8, "usage_type": "name"}, {"api_name": "application_form.api.views.HasoApplicationViewSet", "line_number": 9, "usage_type": "attribute"}, {"api_name": "application_form.api.views", "line_number": 9, "usage_type": "name"}, {"api_name": "application_form.api.views.HitasApplicationViewSet", "line_number": 10, "usage_type": "attribute"}, {"api_name": "application_form.api.views", "line_number": 10, "usage_type": "name"}, {"api_name": "django.urls.path", "line_number": 13, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "581491598", "text": "\"\"\"\nAssembling together radiation data for the 2016-04-26 testing round\n\nNotes\nThe count rate sometimes goes negative because for the sake of simplicity data\nfiles have been concatenated immediately after resampling before performing \ncount subtraction by column shifting to obtain the count rate (CPM).\n\nThese spikes occur when the counter register gets reset and new counts end up\nbeing for a sample or two lower than they previously were. This is mainly a\ncosmetic issue.\n\nI verified manually that the average count rate for Amptek data during background\noperation seems to be about 52-55 CPM, or about 1/3 that of Ortec Quad Counter\ndata (~150 CPM).\n\"\"\"\n\nimport sys\nsys.path.append(\"../spectrautil\")\n\nimport sqlite3\nimport os.path\nimport glob\nimport multiprocessing as mp\nfrom datetime import datetime\n\nimport matplotlib.pyplot as plt\nimport matplotlib.dates as mdates\nimport pandas\n\n# A custom module for reading MCA files\nimport readmca\n\n# A custom module for creating a csv file containing CPM\nimport UCScounter_plot\n\n\nif __name__ == \"__main__\":\n mcadir = \"d:\\\\Downloads\\\\GS5-3\\\\Radiation\\\\GS5-3_xspec\\\\AHK\"\n mcahdffile = os.path.join(mcadir, \"AHK_CdTe_all.hdf\")\n mcafiles = glob.glob(os.path.join(mcadir, \"*.mca\"))\n \n ortecdir = \"D:\\\\Downloads\\\\GS5-3\\\\Radiation\\\\GS5-3_TracorNorthern_TN-7200-Counter\\\\source\"\n ortechdffile = os.path.join(ortecdir, \"Ortec_011-015.hdf\")\n ortecfiles = [\"Counter_Dump 009 - UTC.csv\",\n \"Counter_Dump 010 - UTC - FIXED.csv\", \"Counter_Dump 011 - UTC.csv\",\n \"Counter_Dump 012 - UTC.csv\", \"Counter_Dump 013 - UTC.csv\",\n \"Counter_Dump 014 - UTC.csv\", \"Counter_Dump 015 - UTC.csv\"] \n ortecfiles = [os.path.join(ortecdir, file) for file in ortecfiles]\n \n ucsdir = \"D:\\\\Downloads\\\\GS5-3\\\\Radiation\\\\GS5-3_spec\\\\spu-pressure-test\"\n ucscsvfile = os.path.join(ucsdir, \"ucs_pressure_all.csv\")\n ucsfiles = glob.glob(os.path.join(ucsdir, \"*.spu\"))\n \n pwrdir = \"d:\\\\Downloads\\\\GS5-3\\\\\"\n pwrfiles = [\"GS5-3_pwr015.dbf\"]\n pwrfiles = [os.path.join(pwrdir, file) for file in pwrfiles]\n \n \n if not os.path.exists(mcahdffile):\n print(\"HDF for CdTe data missing, rebuilding...\") \n \n # Read all files in a list. Despite being a multiprocessing operation, this\n # still appends them sequentially as entered by the previous glob function.\n with mp.Pool(mp.cpu_count()) as p: \n mcas = p.map(readmca.Mca, mcafiles)\n \n # Create a dataframe using a list made of selected data from the above list.\n # This will concatenate them all in a single pandas DataFrame.\n df = pandas.DataFrame([[mca.endtime, mca.totalcount] for mca in mcas])\n \n # Rename the columns and set the index (important for resampling)\n df.columns = [\"Timestamp\", \"Total Count\"]\n df.set_index([\"Timestamp\"], inplace=True)\n \n # Make sure that data is sampled to 1 minute bins. This will also fill missing\n # timestamps with NaD (Not a Date).\n df = df.resample('60s')\n \n # Finally write into an HDF file\n df.to_hdf(mcahdffile, \"cdte\")\n \n \n # Same as above\n ortec_resample_seconds = 30 \n if not os.path.exists(ortechdffile):\n print(\"HDF for Ortec data missing, rebuilding...\")\n \n # Reading and concatenating all Ortec csv files into a single DataFrame\n df = pandas.concat([pandas.read_csv(file, index_col=['Timestamp'], parse_dates=['Timestamp']) for file in ortecfiles])\n\n # Resampling into 1 minute bins\n df = df.resample('{}s'.format(ortec_resample_seconds))\n \n # Finally write into an HDF File\n df.to_hdf(ortechdffile, \"ortec\")\n \n \n # Again\n if not os.path.exists(ucscsvfile):\n print(\"CSV file for UCS-40 spectrometer missing, rebuilding...\")\n \n # This will perform all changes in the selected directory. It's a bit slow.\n UCScounter_plot.makecsv(ucsfiles, ucscsvfile)\n \n\n\n # Read data from previously saved HDF files\n mcahdf = pandas.read_hdf(mcahdffile, \"cdte\")\n ortechdf = pandas.read_hdf(ortechdffile, \"ortec\")\n \n # Read UCS CPM from the previously saved CSV file\n ucscsv = pandas.read_csv(ucscsvfile, index_col=[\"Timestamp\"], parse_dates=[\"Timestamp\"])\n \n # Read power data\n con = sqlite3.connect(pwrfiles[0])\n pwr = pandas.read_sql(\"SELECT Taken, StdResults_Vrms, StdResults_Arms, StdResults_Watts FROM MeasResultSet;\",\n con, parse_dates=[\"Taken\"], index_col=[\"Taken\"])\n \n # Shift columns to obtain CPM where applicable\n mcahdf[\"CPM\"] = mcahdf[\"Total Count\"] - mcahdf[\"Total Count\"].shift(1)\n ortechdf[\"CPM\"] = (ortechdf[\"Counter Register\"] - ortechdf[\"Counter Register\"].shift(1)) * (60 / ortec_resample_seconds)\n\n # Filter out negative values resulted from the above shift with unplottable NaN\n mcahdf[mcahdf[\"CPM\"] < 0] = pandas.np.NaN\n ortechdf[ortechdf[\"CPM\"] < 0] = pandas.np.NaN\n\n\n ##### Plotting section #####\n\n plt.style.use('bmh')\n fig = plt.figure(figsize=(24, 12))\n fig.suptitle(\"Radiation data\", fontsize=16)\n \n ax1 = fig.add_subplot(1, 1, 1)\n ax1.grid(which='minor', color='#aaaaaa')\n ax1.minorticks_on()\n ax1.set_ylabel('CPM')\n ax1.set_xlabel('UTC Time') \n \n# ax1.set_ylim([0, 50000]) # Wide range to show Amptek CdTe peaks\n ax1.set_ylim([0, 500]) # Narrow range to show Ortec and UCS peaks\n \n date_fmt = '(%m-%d)\\n%H:%M'\n hour_interval = 1\n ax1.xaxis.set_major_formatter(mdates.DateFormatter(date_fmt))\n ax1.xaxis.set_major_locator(mdates.HourLocator(interval=hour_interval))\n# datelimit = [datetime(2016, 4, 26, 18), datetime(2016, 4, 27, 7)]\n datelimit = [datetime(2016, 4, 26, 21), datetime(2016, 4, 27, 2)]\n\n\n ax1.set_xlim(datelimit)\n \n ax1.plot(mcahdf.index, (mcahdf[\"Total Count\"] - mcahdf[\"Total Count\"].shift(1))/100, marker='.', lw=1, alpha=1, label=\"Amptek CdTe CPM/100 (60 second integration time)\")\n\n ax1.plot(ortechdf.index, ortechdf[\"CPM\"],\n label=\"Ortec Quad Counter CPM ({} second integration time)\".format(ortec_resample_seconds),\n marker='.', lw=1, alpha=0.75)\n\n ax1.plot(ortechdf.index, pandas.rolling_mean(ortechdf[\"CPM\"], 60, center=True),\n label=\"Ortec Quad Counter CPM ({} second integration time) - 60 samples rolling mean\".format(ortec_resample_seconds),\n marker='.', lw=0, markersize=3, alpha=1, color='cyan')\n\n ax1.plot(ucscsv.index, ucscsv[\"CPM\"], label=\"UCS-30 spectrometer CPM (10 minute integration time)\",\n marker='.', lw=2, alpha=1, color=\"yellow\")\n\n legend = ax1.legend(loc=2)\n \n \n ##### Plotting Power data #####\n \n ax2 = ax1.twinx()\n ax2.grid(False)\n ax2.set_xlim(datelimit)\n ax2.set_ylim([0, 2500])\n ax2.set_ylabel('Input Power (W)')\n ax2.plot(pwr.index, pwr[\"StdResults_Watts\"], label=\"Input power\", color=\"black\", lw=0.75)\n ax2.legend(loc=1)\n", "sub_path": "gs53/CdTeMCA_countrate_1minute_00.py", "file_name": "CdTeMCA_countrate_1minute_00.py", "file_ext": "py", "file_size_in_byte": 7042, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.path.append", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 40, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 41, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 44, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 49, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 52, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 53, "usage_type": "name"}, {"api_name": "os.path.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.path.exists", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 60, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 60, "usage_type": "name"}, {"api_name": "multiprocessing.Pool", "line_number": 65, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 65, "usage_type": "call"}, {"api_name": "readmca.Mca", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.path.exists", "line_number": 86, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 86, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 86, "usage_type": "name"}, {"api_name": "pandas.concat", "line_number": 90, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 90, "usage_type": "call"}, {"api_name": "os.path.path.exists", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 100, "usage_type": "name"}, {"api_name": "UCScounter_plot.makecsv", "line_number": 104, "usage_type": "call"}, {"api_name": "pandas.read_hdf", "line_number": 109, "usage_type": "call"}, {"api_name": "pandas.read_hdf", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 113, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 116, "usage_type": "call"}, {"api_name": "pandas.read_sql", "line_number": 117, "usage_type": "call"}, {"api_name": "pandas.np", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pandas.np", "line_number": 126, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 131, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.dates.DateFormatter", "line_number": 146, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 146, "usage_type": "name"}, {"api_name": "matplotlib.dates.HourLocator", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.dates", "line_number": 147, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 149, "usage_type": "call"}, {"api_name": "pandas.rolling_mean", "line_number": 160, "usage_type": "call"}]} +{"seq_id": "285239628", "text": "#!/usr/bin/env python3\n# Copyright (c) 2020 Graphcore Ltd. All rights reserved.\n\nimport os # pylint: disable=unused-import\nimport unittest.mock\nimport poptorch\nimport torch\nimport torchvision.models as models\nimport helpers\n\n\n@unittest.mock.patch.dict(\"os.environ\", helpers.disableSmallModel())\ndef test_resnet():\n torch.manual_seed(42)\n\n image_input = torch.randn([1, 3, 224, 224]).half()\n t1 = torch.tensor([1.]).long()\n # We are running on a dummy input so it doesn't matter if the weights are trained.\n model = models.resnet18(pretrained=False)\n model.train()\n model.half()\n\n training_model = helpers.trainingModelWithLoss(model,\n loss=torch.nn.NLLLoss())\n\n # Run on IPU.\n poptorch_out, loss = training_model(image_input, t1)\n\n assert poptorch_out.dtype == torch.half\n assert loss.dtype == torch.half\n\n\ndef test_model_with_weights():\n model = torch.nn.Linear(1, 10).half()\n t1 = torch.tensor([1.]).half()\n\n inference_model = poptorch.inferenceModel(model)\n out = inference_model(t1)\n\n assert out.dtype == torch.half\n\n # For running on host.\n model = model.float()\n t1 = t1.float()\n\n torch.testing.assert_allclose(model(t1),\n out.float(),\n rtol=0.001,\n atol=1e-04)\n\n\ndef test_simple_model():\n class SimpleAdder(torch.nn.Module):\n def forward(self, x, y, z, w):\n return x + y + 5, z + w + 5\n\n model = SimpleAdder()\n inference_model = poptorch.inferenceModel(model)\n\n t1 = torch.tensor([1.]).half()\n t2 = torch.tensor([2.]).half()\n\n t3 = torch.tensor([3.])\n t4 = torch.tensor([4.])\n\n outHalf, outFloat = inference_model(t1, t2, t3, t4)\n\n assert outHalf.dtype == torch.half\n assert outHalf.float() == 8.0\n\n assert outFloat.dtype == torch.float\n assert outFloat == 12.0\n\n\ndef test_lstm():\n torch.manual_seed(42)\n numHidden = 5\n inputSize = 3\n lstm = torch.nn.LSTM(3, numHidden)\n lstm.half()\n ipuLstm = poptorch.inferenceModel(lstm)\n inputs = [torch.randn(1, inputSize).half() for _ in range(5)]\n # Add the extra 2nd dimension\n inputs = torch.cat(inputs).view(len(inputs), 1, -1)\n hidden = (torch.randn(1, 1,\n numHidden).half(), torch.randn(1, 1,\n numHidden).half())\n ipuOut = ipuLstm(inputs, hidden)\n assert isinstance(ipuOut[0], torch.HalfTensor)\n\n\ndef test_ipu_print_tensor():\n class SimplePrinter(torch.nn.Module):\n def forward(self, x):\n return poptorch.ipu_print_tensor(x)\n\n t1 = torch.tensor([1.], dtype=torch.float16)\n inference_model = poptorch.inferenceModel(SimplePrinter())\n out = inference_model(t1)\n assert out == 1.0\n assert out.dtype == torch.float16\n", "sub_path": "tests/half_test.py", "file_name": "half_test.py", "file_ext": "py", "file_size_in_byte": 2885, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.manual_seed", "line_number": 14, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 16, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 17, "usage_type": "call"}, {"api_name": "torchvision.models.resnet18", "line_number": 19, "usage_type": "call"}, {"api_name": "torchvision.models", "line_number": 19, "usage_type": "name"}, {"api_name": "helpers.trainingModelWithLoss", "line_number": 23, "usage_type": "call"}, {"api_name": "torch.nn.NLLLoss", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.half", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.half", "line_number": 30, "usage_type": "attribute"}, {"api_name": "unittest.mock.mock.patch.dict", "line_number": 12, "usage_type": "call"}, {"api_name": "unittest.mock.mock", "line_number": 12, "usage_type": "attribute"}, {"api_name": "unittest.mock", "line_number": 12, "usage_type": "name"}, {"api_name": "helpers.disableSmallModel", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.tensor", "line_number": 35, "usage_type": "call"}, {"api_name": "poptorch.inferenceModel", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.half", "line_number": 40, "usage_type": "attribute"}, {"api_name": "torch.testing.assert_allclose", "line_number": 46, "usage_type": "call"}, {"api_name": "torch.testing", "line_number": 46, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "attribute"}, {"api_name": "poptorch.inferenceModel", "line_number": 58, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.half", "line_number": 68, "usage_type": "attribute"}, {"api_name": "torch.float", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.manual_seed", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.LSTM", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 79, "usage_type": "attribute"}, {"api_name": "poptorch.inferenceModel", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 82, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 84, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.HalfTensor", "line_number": 89, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 93, "usage_type": "attribute"}, {"api_name": "poptorch.ipu_print_tensor", "line_number": 95, "usage_type": "call"}, {"api_name": "torch.tensor", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.float16", "line_number": 97, "usage_type": "attribute"}, {"api_name": "poptorch.inferenceModel", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.float16", "line_number": 101, "usage_type": "attribute"}]} +{"seq_id": "2118167", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy.spatial.distance import cdist\n\ndef gaussian(x, z, sigma):\n return np.exp(-(x-z)**2/sigma**2)\n\ndef plot_spectrum(w, sigma=0.001, x=None, fig=None, ax=None, **kwargs):\n\n w = w[w.real > 0].real\n \n # Data stuff:\n if x is None:\n x = np.linspace(0, max(w)*1.15, 1000)\n y = np.sum(gaussian(x, w[:, np.newaxis].real, sigma), axis=0)\n\n # Figure stuff:\n if fig is None:\n fig, ax = plt.subplots()\n \n ax.plot(x, y, **kwargs)\n \n return fig, ax\n\ndef feature_histogram(features, fig=None, ax=None):\n \"\"\"\n Plots a histogram of the distance between features:\n\n Arguments:\n -- features: np.array (n, d)\n \"\"\"\n\n dists = cdist(features, features)\n\n\n if fig is None:\n fig, ax = plt.subplots()\n \n ax.hist(dists)\n\n ax.set_xlabel('Feat. dist [?]')\n ax.set_ylabel('Counts')\n\n plt.tight_layout()\n\n return fig, ax\n\ndef error_plot(history, reference, plot=True):\n \"\"\"\n Plots the error as a function of steps. \n \"\"\"\n\n fig, ax = plt.subplots()\n\n error = np.abs(history-reference)\n\n for i in range(error.shape[1]):\n ax.plot(error[:, i], label='{}'.format(i), linewidth=2)\n\n ax.set_xlim([0, error.shape[0]-1])\n ax.set_ylim([0, 500])\n ax.set_xlabel('Iteration [#]', fontsize=12)\n ax.set_ylabel(r'Absolute Error [cm$^{-1}$]', fontsize=12)\n \n\n plt.legend()\n plt.tight_layout() \n if plot:\n plt.show()\n\n return fig, ax\n\n\n\n\n \n \n", "sub_path": "plotters.py", "file_name": "plotters.py", "file_ext": "py", "file_size_in_byte": 1515, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.exp", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 15, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.cdist", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "numpy.abs", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}]} +{"seq_id": "75278124", "text": "# -*- coding: utf-8 -*-\nfrom flask import current_app\nfrom flask_mongoengine import Document\nfrom mongoengine import CASCADE, signals\nfrom mongoengine.fields import StringField, BooleanField, DictField, LazyReferenceField\nfrom mpcontribs.api.projects.document import Projects\nfrom mpcontribs.api import validate_data\n\n\nclass Contributions(Document):\n project = LazyReferenceField(\n Projects, required=True, passthrough=True, reverse_delete_rule=CASCADE\n )\n identifier = StringField(required=True, help_text=\"material/composition identifier\")\n formula = StringField(help_text=\"formula (set dynamically)\")\n is_public = BooleanField(\n required=True, default=False, help_text=\"public/private contribution\"\n )\n data = DictField(help_text=\"free-form data to be shown in Contribution Card\")\n meta = {\n \"collection\": \"contributions\",\n \"indexes\": [\"project\", \"identifier\", \"formula\", \"is_public\"],\n }\n\n @classmethod\n def pre_save_post_validation(cls, sender, document, **kwargs):\n document.data = validate_data(\n document.data, sender=sender, project=document.project\n )\n if hasattr(document, \"formula\"):\n formulae = current_app.config[\"FORMULAE\"]\n document.formula = formulae.get(document.identifier, document.identifier)\n\n @classmethod\n def post_save(cls, sender, document, **kwargs):\n # avoid circular import\n from mpcontribs.api.notebooks.document import Notebooks\n from mpcontribs.api.cards.document import Cards\n\n Notebooks.objects(pk=document.id).delete()\n Cards.objects(pk=document.id).delete()\n\n\nsignals.pre_save_post_validation.connect(\n Contributions.pre_save_post_validation, sender=Contributions\n)\nsignals.post_save.connect(Contributions.post_save, sender=Contributions)\n", "sub_path": "mpcontribs-api/mpcontribs/api/contributions/document.py", "file_name": "document.py", "file_ext": "py", "file_size_in_byte": 1834, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask_mongoengine.Document", "line_number": 10, "usage_type": "name"}, {"api_name": "mongoengine.fields.LazyReferenceField", "line_number": 11, "usage_type": "call"}, {"api_name": "mpcontribs.api.projects.document.Projects", "line_number": 12, "usage_type": "argument"}, {"api_name": "mongoengine.CASCADE", "line_number": 12, "usage_type": "name"}, {"api_name": "mongoengine.fields.StringField", "line_number": 14, "usage_type": "call"}, {"api_name": "mongoengine.fields.StringField", "line_number": 15, "usage_type": "call"}, {"api_name": "mongoengine.fields.BooleanField", "line_number": 16, "usage_type": "call"}, {"api_name": "mongoengine.fields.DictField", "line_number": 19, "usage_type": "call"}, {"api_name": "mpcontribs.api.validate_data", "line_number": 27, "usage_type": "call"}, {"api_name": "flask.current_app.config", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.current_app", "line_number": 31, "usage_type": "name"}, {"api_name": "mpcontribs.api.notebooks.document.Notebooks.objects", "line_number": 40, "usage_type": "call"}, {"api_name": "mpcontribs.api.notebooks.document.Notebooks", "line_number": 40, "usage_type": "name"}, {"api_name": "mpcontribs.api.cards.document.Cards.objects", "line_number": 41, "usage_type": "call"}, {"api_name": "mpcontribs.api.cards.document.Cards", "line_number": 41, "usage_type": "name"}, {"api_name": "mongoengine.signals.pre_save_post_validation.connect", "line_number": 44, "usage_type": "call"}, {"api_name": "mongoengine.signals.pre_save_post_validation", "line_number": 44, "usage_type": "attribute"}, {"api_name": "mongoengine.signals", "line_number": 44, "usage_type": "name"}, {"api_name": "{'Notebooks': 'mpcontribs.api.notebooks.document.Notebooks', 'Cards': 'mpcontribs.api.cards.document.Cards'}.pre_save_post_validation", "line_number": 45, "usage_type": "attribute"}, {"api_name": "mongoengine.signals.post_save.connect", "line_number": 47, "usage_type": "call"}, {"api_name": "mongoengine.signals.post_save", "line_number": 47, "usage_type": "attribute"}, {"api_name": "mongoengine.signals", "line_number": 47, "usage_type": "name"}, {"api_name": "{'Notebooks': 'mpcontribs.api.notebooks.document.Notebooks', 'Cards': 'mpcontribs.api.cards.document.Cards'}.post_save", "line_number": 47, "usage_type": "attribute"}]} +{"seq_id": "209962357", "text": "\"\"\"\nModule contains application's Main Window\n\"\"\"\n\nfrom PyQt5.QtCore import Qt, QEvent, pyqtSlot, pyqtSignal\nfrom PyQt5.QtGui import QIcon, QCloseEvent, QKeySequence\nfrom PyQt5.QtWidgets import QMainWindow, QAction, QMessageBox, qApp, QShortcut, QSystemTrayIcon\n\nfrom gui.uic.mainwindow import Ui_MainWindow\nfrom storage.res_manager import ResManager, Resources\nfrom storage.settings_manager import Settings, SettingsManager\n\n\nclass MainWindow(QMainWindow):\n \"\"\"\n Application's Main Window\n \"\"\"\n\n closed = pyqtSignal()\n raised = pyqtSignal()\n trayed = pyqtSignal()\n\n def __init__(self):\n super().__init__()\n\n # setup UI\n self.ui = Ui_MainWindow()\n self.ui.setupUi(self)\n\n self.setWindowTitle(qApp.applicationName())\n self.setWindowFlag(Qt.WindowContextHelpButtonHint, True)\n self.setWindowIcon(ResManager.get(Resources.ICONS_APP_ICON))\n\n # init\n self._createMenuActions()\n self._createShortcuts()\n self._readSettings()\n\n @pyqtSlot()\n def closeEvent(self, event: QCloseEvent):\n if SettingsManager.get(Settings.TRAY_CLOSE_TO):\n self.toTray()\n event.ignore()\n else:\n self.closed.emit()\n event.accept()\n\n @pyqtSlot()\n def changeEvent(self, event: QEvent):\n super().changeEvent(event)\n if event.type() == QEvent.WindowStateChange:\n if self.windowState() & Qt.WindowMinimized and SettingsManager.get(Settings.TRAY_MINIMIZE_TO):\n self.toTray()\n\n @pyqtSlot()\n def raiseWindow(self):\n \"\"\"\n Bring application to the front (from tray, taskbar etc)\n :return:\n \"\"\"\n self.showNormal()\n self.raise_()\n self.activateWindow()\n\n self.raised.emit()\n\n @pyqtSlot()\n def toTray(self, showInformation=False):\n self.hide()\n if showInformation:\n self._tray.showMessage(qApp.applicationName(), self.tr(\"Application was minimized to tray\"),\n QSystemTrayIcon.Information, 1500)\n\n self.trayed.emit()\n\n @pyqtSlot()\n def toggleWindow(self):\n self.toTray() if self.isVisible() else self.raiseWindow()\n\n @pyqtSlot()\n def restoreTrayState(self):\n if SettingsManager.get(Settings.MAINWINDOW_IN_TRAY):\n self.toTray()\n else:\n self.show()\n\n def _createShortcuts(self):\n \"\"\"\n create shortcuts\n :return:\n \"\"\"\n\n find = QShortcut(QKeySequence.Find, self.ui.searchLine, self.ui.searchLine.setFocus)\n find.setContext(Qt.ApplicationShortcut)\n\n def _createMenuActions(self):\n \"\"\"\n Initialize menu actions, that cannot be added in .ui form\n :return:\n \"\"\"\n # Exit action\n self.ui.actionExit.setShortcut(QKeySequence.Quit)\n self.ui.actionExit.setIcon(\n QIcon.fromTheme(\"application-exit\", ResManager.get(Resources.ICONS_EXIT)))\n self.ui.actionExit.triggered.connect(self.exit)\n\n # toggle dockable widgets actions\n act: QAction = self.ui.infoWidget.toggleViewAction()\n self.ui.menuView.addAction(act)\n act.setShortcut(QKeySequence(\"Ctrl+I\"))\n\n act = self.ui.statusWidget.toggleViewAction()\n self.ui.menuView.addAction(act)\n act.setShortcut(QKeySequence(\"Ctrl+S\"))\n\n # About actions\n self.ui.actionAbout.triggered.connect(self._onAboutAction)\n self.ui.actionAboutQt.triggered.connect(self._onAboutQtAction)\n\n # Settings action\n self.ui.actionSettings.triggered.connect(self._onSettingsAction)\n self.ui.actionSettings.setShortcut(\"Ctrl+O\")\n\n # Slots\n @pyqtSlot()\n def exit(self):\n \"\"\"\n Exit action slot\n :return: None\n \"\"\"\n\n def finalize():\n self._writeSettings()\n # TODO stop downloads etc\n\n finalize()\n qApp.exit()\n\n @pyqtSlot()\n def _onSettingsAction(self):\n \"\"\"\n Settings action slot\n :return: None\n \"\"\"\n from gui.settingsdialog import SettingsDialog\n dialog = SettingsDialog(self)\n code = dialog.exec()\n\n @pyqtSlot()\n def _onAboutQtAction(self):\n \"\"\"\n About Qt MessageBox\n :return:\n \"\"\"\n QMessageBox.aboutQt(self)\n\n @pyqtSlot()\n def _onAboutAction(self):\n \"\"\"\n About MessageBox\n :return:\n \"\"\"\n import app_info\n name = self.tr(\"About Video to Audio Converter\")\n description = self.tr(\"

Video to Audio Converter {0}

\"\n \"

Video to Audio converter is a small free graphical application that \"\n \"allows you to download any video and convert it to audio file \"\n \"as well as re-upload video to other popular video hosting sites.\"\n \"

Website: \"\n \"http://github.com/sanyarnd/vid2audio\") \\\n .format(app_info.VERSION)\n QMessageBox.about(self, name, description)\n\n # Settings\n def _writeSettings(self):\n \"\"\"\n Write current state to settings\n :return: None\n \"\"\"\n SettingsManager.set(Settings.MAINWINDOW_GEOMETRY, self.saveGeometry())\n SettingsManager.set(Settings.MAINWINDOW_STATE, self.saveState())\n SettingsManager.set(Settings.MAINWINDOW_IN_TRAY, not self.isVisible())\n\n def _readSettings(self):\n \"\"\"\n Read stored state from settings and apply\n :return: None\n \"\"\"\n self.restoreGeometry(SettingsManager.get(Settings.MAINWINDOW_GEOMETRY))\n self.restoreState(SettingsManager.get(Settings.MAINWINDOW_STATE))\n", "sub_path": "vid2audio/gui/mainwindow.py", "file_name": "mainwindow.py", "file_ext": "py", "file_size_in_byte": 5807, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "PyQt5.QtWidgets.QMainWindow", "line_number": 14, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 19, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 20, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSignal", "line_number": 21, "usage_type": "call"}, {"api_name": "gui.uic.mainwindow.Ui_MainWindow", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp.applicationName", "line_number": 30, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 30, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.WindowContextHelpButtonHint", "line_number": 31, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 31, "usage_type": "name"}, {"api_name": "storage.res_manager.ResManager.get", "line_number": 32, "usage_type": "call"}, {"api_name": "storage.res_manager.ResManager", "line_number": 32, "usage_type": "name"}, {"api_name": "storage.res_manager.Resources.ICONS_APP_ICON", "line_number": 32, "usage_type": "attribute"}, {"api_name": "storage.res_manager.Resources", "line_number": 32, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QCloseEvent", "line_number": 40, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.get", "line_number": 41, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 41, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.TRAY_CLOSE_TO", "line_number": 41, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 41, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 39, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QEvent", "line_number": 49, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QEvent.WindowStateChange", "line_number": 51, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QEvent", "line_number": 51, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.WindowMinimized", "line_number": 52, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 52, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.get", "line_number": 52, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 52, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.TRAY_MINIMIZE_TO", "line_number": 52, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 52, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 55, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp.applicationName", "line_number": 71, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 71, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QSystemTrayIcon.Information", "line_number": 72, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QSystemTrayIcon", "line_number": 72, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 67, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 76, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager.get", "line_number": 82, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 82, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_IN_TRAY", "line_number": 82, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 82, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 80, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QShortcut", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QKeySequence.Find", "line_number": 93, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 93, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.ApplicationShortcut", "line_number": 94, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 94, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QKeySequence.Quit", "line_number": 102, "usage_type": "attribute"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 102, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QIcon.fromTheme", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QIcon", "line_number": 104, "usage_type": "name"}, {"api_name": "storage.res_manager.ResManager.get", "line_number": 104, "usage_type": "call"}, {"api_name": "storage.res_manager.ResManager", "line_number": 104, "usage_type": "name"}, {"api_name": "storage.res_manager.Resources.ICONS_EXIT", "line_number": 104, "usage_type": "attribute"}, {"api_name": "storage.res_manager.Resources", "line_number": 104, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QAction", "line_number": 108, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 110, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 114, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp.exit", "line_number": 137, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.qApp", "line_number": 137, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 125, "usage_type": "call"}, {"api_name": "gui.settingsdialog.SettingsDialog", "line_number": 146, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 139, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.aboutQt", "line_number": 155, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 155, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 149, "usage_type": "call"}, {"api_name": "app_info.VERSION", "line_number": 171, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.about", "line_number": 172, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 172, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 157, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager.set", "line_number": 180, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 180, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_GEOMETRY", "line_number": 180, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 180, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.set", "line_number": 181, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 181, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_STATE", "line_number": 181, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 181, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.set", "line_number": 182, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 182, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_IN_TRAY", "line_number": 182, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 182, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.get", "line_number": 189, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 189, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_GEOMETRY", "line_number": 189, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 189, "usage_type": "name"}, {"api_name": "storage.settings_manager.SettingsManager.get", "line_number": 190, "usage_type": "call"}, {"api_name": "storage.settings_manager.SettingsManager", "line_number": 190, "usage_type": "name"}, {"api_name": "storage.settings_manager.Settings.MAINWINDOW_STATE", "line_number": 190, "usage_type": "attribute"}, {"api_name": "storage.settings_manager.Settings", "line_number": 190, "usage_type": "name"}]} +{"seq_id": "385578228", "text": "import json\nimport os\n\nimport six\nimport trafaret as t\n\nfrom datarobot.client import get_client, staticproperty\nfrom datarobot.models.api_object import APIObject\nfrom datarobot.models.sharing import SharingAccess\nfrom datarobot.utils.pagination import unpaginate\nfrom datarobot.utils.waiters import wait_for_async_resolution\n\nfrom .. import errors\nfrom ..utils import encode_utf8_if_py2\n\n\nclass CalendarFile(APIObject):\n \"\"\" Represents the data for a calendar file.\n\n For more information about calendar files, see the\n :ref:`calendar documentation `.\n\n Attributes\n ----------\n id : str\n The id of the calendar file.\n calendar_start_date : str\n The earliest date in the calendar.\n calendar_end_date : str\n The last date in the calendar.\n created : str\n The date this calendar was created, i.e. uploaded to DR.\n name : str\n The name of the calendar.\n num_event_types : int\n The number of different event types.\n num_events : int\n The number of events this calendar has.\n project_ids : list of strings\n A list containing the projectIds of the projects using this calendar.\n multiseries_id_columns: list of str or None\n A list of columns in calendar which uniquely identify events for different series.\n Currently, only one column is supported.\n If multiseries id columns are not provided, calendar is considered to be single series.\n role : str\n The access role the user has for this calendar.\n \"\"\"\n\n _base_url = \"calendars/\"\n\n _upload_url = _base_url + \"fileUpload/\"\n _calendar_url = _base_url + \"{}/\"\n _access_control_url = _calendar_url + \"accessControl/\"\n _from_country_code_url = _base_url + \"fromCountryCode/\"\n _allowed_countries_list_url = \"calendarCountryCodes/\"\n\n _client = staticproperty(get_client)\n\n _converter = t.Dict(\n {\n t.Key(\"calendar_end_date\"): t.String,\n t.Key(\"calendar_start_date\"): t.String,\n t.Key(\"created\"): t.String,\n t.Key(\"id\"): t.String,\n t.Key(\"name\"): t.String,\n t.Key(\"num_event_types\"): t.Int,\n t.Key(\"num_events\"): t.Int,\n t.Key(\"project_ids\"): t.List(t.String),\n t.Key(\"role\"): t.String,\n t.Key(\"multiseries_id_columns\", optional=True): t.Or(t.List(t.String), t.Null),\n }\n ).ignore_extra(\"*\")\n\n def __init__(\n self,\n calendar_end_date=None,\n calendar_start_date=None,\n created=None,\n id=None,\n name=None,\n num_event_types=None,\n num_events=None,\n project_ids=None,\n role=None,\n multiseries_id_columns=None,\n ):\n self.calendar_end_date = calendar_end_date\n self.calendar_start_date = calendar_start_date\n self.created = created\n self.id = id\n self.name = name\n self.num_event_types = num_event_types\n self.num_events = num_events\n self.project_ids = project_ids\n self.role = role\n self.multiseries_id_columns = multiseries_id_columns\n\n def __repr__(self):\n return encode_utf8_if_py2(u\"{}({})\".format(self.__class__.__name__, self.id))\n\n def __eq__(self, other):\n \"\"\" Compares the relevant fields of two calendars to assess equality \"\"\"\n vars1 = [self.__getattribute__(x) for x in [k.name for k in self._converter.keys]]\n vars2 = [other.__getattribute__(x) for x in [k.name for k in other._converter.keys]]\n return vars1 == vars2\n\n @classmethod\n def create(cls, file_path, calendar_name=None, multiseries_id_columns=None):\n \"\"\"\n Creates a calendar using the given file. For information about calendar files, see the\n :ref:`calendar documentation `\n\n The provided file must be a CSV in the format:\n\n .. code-block:: text\n\n Date, Event, Series ID\n , , \n , ,\n\n A header row is required, and the \"Series ID\" column is optional.\n\n Once the CalendarFile has been created, pass its ID with\n the :class:`DatetimePartitioningSpecification `\n when setting the target for a time series project in order to use it.\n\n Parameters\n ----------\n file_path : string\n A string representing a path to a local csv file.\n calendar_name : string, optional\n A name to assign to the calendar. Defaults to the name of the file if not provided.\n multiseries_id_columns : list of str or None\n a list of the names of multiseries id columns to define which series an event\n belongs to. Currently only one multiseries id column is supported.\n\n Returns\n -------\n calendar_file : CalendarFile\n Instance with initialized data.\n\n Raises\n ------\n AsyncProcessUnsuccessfulError\n Raised if there was an error processing the provided calendar file.\n\n Examples\n --------\n .. code-block:: python\n\n # Creating a calendar with a specified name\n cal = dr.CalendarFile.create('/home/calendars/somecalendar.csv',\n calendar_name='Some Calendar Name')\n cal.id\n >>> 5c1d4904211c0a061bc93013\n cal.name\n >>> Some Calendar Name\n\n # Creating a calendar without specifying a name\n cal = dr.CalendarFile.create('/home/calendars/somecalendar.csv')\n cal.id\n >>> 5c1d4904211c0a061bc93012\n cal.name\n >>> somecalendar.csv\n\n # Creating a calendar with multiseries id columns\n cal = dr.CalendarFile.create('/home/calendars/somemultiseriescalendar.csv',\n calendar_name='Some Multiseries Calendar Name',\n multiseries_id_columns=['series_id'])\n cal.id\n >>> 5da9bb21962d746f97e4daee\n cal.name\n >>> Some Multiseries Calendar Name\n cal.multiseries_id_columns\n >>> ['series_id']\n \"\"\"\n\n # make sure it's a valid filename, and set the calendar name if not provided\n if isinstance(file_path, six.string_types) and os.path.isfile(file_path):\n if not calendar_name:\n calendar_name = os.path.basename(file_path)\n else:\n raise ValueError(u\"The provided file does not exist: {}\".format(file_path))\n try:\n calendar_name.encode(\"ascii\")\n # Which exception we get here depends on whether the input was string or unicode\n # (we allow both).\n except (UnicodeEncodeError, UnicodeDecodeError):\n raise errors.IllegalFileName\n\n form_data = None\n if multiseries_id_columns:\n if not isinstance(multiseries_id_columns, (list, tuple)):\n raise ValueError(\n \"Expected list of str for multiseries_id_columns, got: {}\".format(\n multiseries_id_columns\n )\n )\n form_data = {\"multiseries_id_columns\": (None, json.dumps(multiseries_id_columns))}\n\n upload_response = cls._client.build_request_with_file(\n method=\"post\",\n url=cls._upload_url,\n fname=calendar_name,\n file_path=file_path,\n form_data=form_data,\n )\n new_calendar_url = wait_for_async_resolution(\n cls._client, upload_response.headers[\"Location\"]\n )\n\n return cls.from_location(new_calendar_url)\n\n @classmethod\n def create_calendar_from_country_code(cls, country_code, start_date, end_date):\n \"\"\"\n Generates a calendar based on the provided country code and dataset start date and end\n dates. The provided country code should be uppercase and 2-3 characters long. See\n :meth:`CalendarFile.get_allowed_country_codes\n ` for a list of allowed country codes.\n\n Parameters\n ----------\n country_code : string\n The country code for the country to use for generating the calendar.\n start_date : datetime.datetime\n The earliest date to include in the generated calendar.\n end_date : datetime.datetime\n The latest date to include in the generated calendar.\n\n Returns\n -------\n calendar_file : CalendarFile\n Instance with initialized data.\n \"\"\"\n payload = {\n \"countryCode\": country_code,\n \"startDate\": start_date,\n \"endDate\": end_date,\n }\n generation_response = cls._client.post(\n cls._from_country_code_url, data=payload, headers={\"Content-Type\": \"application/json\"}\n )\n generated_calendar_url = wait_for_async_resolution(\n cls._client, generation_response.headers[\"Location\"]\n )\n return cls.from_location(generated_calendar_url)\n\n @classmethod\n def get_allowed_country_codes(cls, offset=None, limit=None):\n \"\"\"\n Retrieves the list of allowed country codes that can be used for generating the preloaded\n calendars.\n\n Parameters\n ----------\n offset : int\n Optional, defaults to 0. This many results will be skipped.\n limit : int\n Optional, defaults to 100, maximum 1000. At most this many results are returned.\n\n Returns\n -------\n list\n A list dicts, each of which represents an allowed country codes. Each item has the\n following structure:\n\n * ``name`` : (str) The name of the country.\n * ``code`` : (str) The code for this country. This is the value that should be supplied\n to :meth:`CalendarFile.create_calendar_from_country_code\n `.\n \"\"\"\n params = {}\n if limit is not None:\n params[\"limit\"] = limit\n if offset is not None:\n params[\"offset\"] = offset\n return list(unpaginate(cls._allowed_countries_list_url, params, cls._client))\n\n @classmethod\n def get(cls, calendar_id):\n \"\"\"\n Gets the details of a calendar, given the id.\n\n Parameters\n ----------\n calendar_id : str\n The identifier of the calendar.\n\n Returns\n -------\n calendar_file : CalendarFile\n The requested calendar.\n\n Raises\n ------\n DataError\n Raised if the calendar_id is invalid, i.e. the specified CalendarFile does not exist.\n\n Examples\n --------\n .. code-block:: python\n\n cal = dr.CalendarFile.get(some_calendar_id)\n cal.id\n >>> some_calendar_id\n \"\"\"\n return cls.from_location(cls._calendar_url.format(calendar_id))\n\n @classmethod\n def list(cls, project_id=None, batch_size=None):\n \"\"\"\n Gets the details of all calendars this user has view access for.\n\n Parameters\n ----------\n project_id : str, optional\n If provided, will filter for calendars associated only with the specified project.\n batch_size : int, optional\n The number of calendars to retrieve in a single API call. If specified, the client may\n make multiple calls to retrieve the full list of calendars. If not specified, an\n appropriate default will be chosen by the server.\n\n Returns\n -------\n calendar_list : list of :class:`CalendarFile `\n A list of CalendarFile objects.\n\n Examples\n --------\n .. code-block:: python\n\n calendars = dr.CalendarFile.list()\n len(calendars)\n >>> 10\n \"\"\"\n if project_id is not None:\n list_url = cls._base_url + \"?projectId={}\".format(project_id)\n else:\n list_url = cls._base_url\n\n params = {}\n if batch_size is not None:\n params[\"limit\"] = batch_size\n\n return [cls.from_server_data(entry) for entry in unpaginate(list_url, params, cls._client)]\n\n @classmethod\n def delete(cls, calendar_id):\n \"\"\"\n Deletes the calendar specified by calendar_id.\n\n Parameters\n ----------\n calendar_id : str\n The id of the calendar to delete.\n The requester must have OWNER access for this calendar.\n\n Raises\n ------\n ClientError\n Raised if an invalid calendar_id is provided.\n\n Examples\n --------\n .. code-block:: python\n\n # Deleting with a valid calendar_id\n status_code = dr.CalendarFile.delete(some_calendar_id)\n status_code\n >>> 204\n dr.CalendarFile.get(some_calendar_id)\n >>> ClientError: Item not found\n \"\"\"\n cls._client.delete(cls._calendar_url.format(calendar_id))\n\n @classmethod\n def update_name(cls, calendar_id, new_calendar_name):\n \"\"\"\n Changes the name of the specified calendar to the specified name.\n The requester must have at least READ_WRITE permissions on the calendar.\n\n Parameters\n ----------\n calendar_id : str\n The id of the calendar to update.\n new_calendar_name : str\n The new name to set for the specified calendar.\n\n Returns\n -------\n status_code : int\n 200 for success\n\n Raises\n ------\n ClientError\n Raised if an invalid calendar_id is provided.\n\n Examples\n --------\n .. code-block:: python\n\n response = dr.CalendarFile.update_name(some_calendar_id, some_new_name)\n response\n >>> 200\n cal = dr.CalendarFile.get(some_calendar_id)\n cal.name\n >>> some_new_name\n\n \"\"\"\n try:\n new_calendar_name.encode(\"ascii\")\n # Which exception we get here depends on whether the input was string or unicode\n # (we allow both).\n except (UnicodeEncodeError, UnicodeDecodeError):\n raise errors.IllegalFileName\n\n response_data = cls._client.patch(\n cls._calendar_url.format(calendar_id), data={\"name\": new_calendar_name}\n )\n return response_data.status_code\n\n @classmethod\n def share(cls, calendar_id, access_list):\n \"\"\"\n Shares the calendar with the specified users, assigning the specified roles.\n\n Parameters\n ----------\n calendar_id : str\n The id of the calendar to update\n access_list:\n A list of dr.SharingAccess objects. Specify `None` for the role to delete a user's\n access from the specified CalendarFile. For more information on specific access levels,\n see the :ref:`sharing ` documentation.\n\n Returns\n -------\n status_code : int\n 200 for success\n\n Raises\n ------\n ClientError\n Raised if unable to update permissions for a user.\n AssertionError\n Raised if access_list is invalid.\n\n Examples\n --------\n .. code-block:: python\n\n # assuming some_user is a valid user, share this calendar with some_user\n sharing_list = [dr.SharingAccess(some_user_username,\n dr.enums.SHARING_ROLE.READ_WRITE)]\n response = dr.CalendarFile.share(some_calendar_id, sharing_list)\n response.status_code\n >>> 200\n\n # delete some_user from this calendar, assuming they have access of some kind already\n delete_sharing_list = [dr.SharingAccess(some_user_username,\n None)]\n response = dr.CalendarFile.share(some_calendar_id, delete_sharing_list)\n response.status_code\n >>> 200\n\n # Attempt to add an invalid user to a calendar\n invalid_sharing_list = [dr.SharingAccess(invalid_username,\n dr.enums.SHARING_ROLE.READ_WRITE)]\n dr.CalendarFile.share(some_calendar_id, invalid_sharing_list)\n >>> ClientError: Unable to update access for this calendar\n\n \"\"\"\n # ensure access_list is a list\n assert isinstance(access_list, list), \"access_list must be a list\"\n # ensure each item in access_list is a SharingAccess object\n assert all(\n isinstance(access, SharingAccess) for access in access_list\n ), \"access_list must be a list of dr.SharingAccess objects\"\n\n payload = {\"users\": [access.collect_payload() for access in access_list]}\n response_data = cls._client.patch(\n cls._access_control_url.format(calendar_id), data=payload, keep_attrs={\"role\"}\n )\n return response_data.status_code\n\n @classmethod\n def get_access_list(cls, calendar_id, batch_size=None):\n \"\"\"\n Retrieve a list of users that have access to this calendar.\n\n Parameters\n ----------\n calendar_id : str\n The id of the calendar to retrieve the access list for.\n batch_size : int, optional\n The number of access records to retrieve in a single API call. If specified, the client\n may make multiple calls to retrieve the full list of calendars. If not specified, an\n appropriate default will be chosen by the server.\n\n Returns\n -------\n access_control_list : list of :class:`SharingAccess `\n A list of :class:`SharingAccess ` objects.\n\n Raises\n ------\n ClientError\n Raised if user does not have access to calendar or calendar does not exist.\n \"\"\"\n\n params = {}\n if batch_size is not None:\n params[\"limit\"] = batch_size\n url = cls._access_control_url.format(calendar_id)\n return [\n SharingAccess.from_server_data(datum) for datum in unpaginate(url, params, cls._client)\n ]\n", "sub_path": "datarobot/models/calendar_file.py", "file_name": "calendar_file.py", "file_ext": "py", "file_size_in_byte": 18441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "datarobot.models.api_object.APIObject", "line_number": 17, "usage_type": "name"}, {"api_name": "datarobot.client.staticproperty", "line_number": 57, "usage_type": "call"}, {"api_name": "datarobot.client.get_client", "line_number": 57, "usage_type": "argument"}, {"api_name": "trafaret.Dict", "line_number": 59, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 61, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 62, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 63, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 64, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 65, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 66, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 67, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 68, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 69, "usage_type": "call"}, {"api_name": "trafaret.Key", "line_number": 70, "usage_type": "call"}, {"api_name": "trafaret.String", "line_number": 61, "usage_type": "attribute"}, {"api_name": "trafaret.String", "line_number": 62, "usage_type": "attribute"}, {"api_name": "trafaret.String", "line_number": 63, "usage_type": "attribute"}, {"api_name": "trafaret.String", "line_number": 64, "usage_type": "attribute"}, {"api_name": "trafaret.String", "line_number": 65, "usage_type": "attribute"}, {"api_name": "trafaret.Int", "line_number": 66, "usage_type": "attribute"}, {"api_name": "trafaret.Int", "line_number": 67, "usage_type": "attribute"}, {"api_name": "trafaret.List", "line_number": 68, "usage_type": "call"}, {"api_name": "trafaret.String", "line_number": 68, "usage_type": "attribute"}, {"api_name": "trafaret.String", "line_number": 69, "usage_type": "attribute"}, {"api_name": "trafaret.Or", "line_number": 70, "usage_type": "call"}, {"api_name": "trafaret.List", "line_number": 70, "usage_type": "call"}, {"api_name": "trafaret.String", "line_number": 70, "usage_type": "attribute"}, {"api_name": "trafaret.Null", "line_number": 70, "usage_type": "attribute"}, {"api_name": "utils.encode_utf8_if_py2", "line_number": 99, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 181, "usage_type": "call"}, {"api_name": "os.path", "line_number": 181, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 199, "usage_type": "call"}, {"api_name": "datarobot.utils.waiters.wait_for_async_resolution", "line_number": 208, "usage_type": "call"}, {"api_name": "datarobot.utils.waiters.wait_for_async_resolution", "line_number": 244, "usage_type": "call"}, {"api_name": "datarobot.utils.pagination.unpaginate", "line_number": 278, "usage_type": "call"}, {"api_name": "datarobot.utils.pagination.unpaginate", "line_number": 346, "usage_type": "call"}, {"api_name": "datarobot.models.sharing.SharingAccess", "line_number": 479, "usage_type": "argument"}, {"api_name": "datarobot.models.sharing.SharingAccess.from_server_data", "line_number": 518, "usage_type": "call"}, {"api_name": "datarobot.models.sharing.SharingAccess", "line_number": 518, "usage_type": "name"}, {"api_name": "datarobot.utils.pagination.unpaginate", "line_number": 518, "usage_type": "call"}]} +{"seq_id": "336363191", "text": "import logging\nimport pymongo\n\nfrom config import Config\n\nlogging.basicConfig(level=logging.DEBUG)\n_MONGOCLIENT = pymongo.MongoClient(Config.MONGOCLIENT)\n\nclass MongoHandler(object):\n\n def __init__(self,db,collection):\n self.db = db\n self.collection = collection\n self.mdb = _MONGOCLIENT[self.db]\n self.mcollection = self.mdb[self.collection]\n\n def insertIntoMongo(self,data):\n '''\n :param db: str -> name of mongo db (e.g. bestsellers)\n :param collection: str -> name of collection in db (e.g. outdoors)\n :param data: dict\n :return: None\n '''\n filter = {\"start_date\": {\"$gte\": data['start_date']}, \"end_date\": {\"$lte\": data['end_date']}}\n if self.mcollection.count_documents(filter):\n logging.info(f\"data already exists for {self.collection} on start_date {data['start_date']}\")\n else:\n self.mcollection.insert_one(data)\n logging.info('data loaded to : {db}|{col}'.format(db=self.db,col=self.collection))\n\n def getFromMongoByDateRange(self,start_date,end_date):\n d = {}\n filter = {\"start_date\":{\"$gte\":start_date},\"end_date\":{\"$lte\":end_date}}\n if self.mcollection.count_documents(filter):\n docs = self.mcollection.find(filter)\n for doc in docs:\n d[doc['start_date']] = doc['html']\n return d\n\ndef main():\n m = MongoHandler('futures','sp500')\n d = m.getFromMongoByDateRange('2017-12-01','2019-12-31')\n\nif __name__==\"__main__\":\n main()", "sub_path": "app/utils/mongo_handler.py", "file_name": "mongo_handler.py", "file_ext": "py", "file_size_in_byte": 1545, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.basicConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pymongo.MongoClient", "line_number": 7, "usage_type": "call"}, {"api_name": "config.Config.MONGOCLIENT", "line_number": 7, "usage_type": "attribute"}, {"api_name": "config.Config", "line_number": 7, "usage_type": "name"}, {"api_name": "logging.info", "line_number": 26, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 29, "usage_type": "call"}]} +{"seq_id": "586947664", "text": "import os\r\nimport cv2\r\nimport numpy as np\r\nfrom PIL import Image\r\n\r\nrecog = cv2.createLBPHFaceRecognizer()\r\npath = 'Faces/'\r\n\r\ndef getimgpaths(path):\r\n paths = os.listdir(path)\r\n faces = []\r\n id1 = []\r\n for i in paths:\r\n img = Image.open(path+i).convert('L')\r\n img_to_np = np.array(img, 'uint8')\r\n faces.append(img_to_np)\r\n ids = i.split('.')[1]\r\n id1.append(int(ids))\r\n cv2.imshow('Wait, I am Training...',img_to_np)\r\n cv2.waitKey(10)\r\n return np.array(id1), faces\r\n\r\nid1,faces = getimgpaths(path)\r\nrecog.train(faces,id1)\r\nrecog.save('recog/recognized.yml')\r\nprint(\"Yay! Training is Complete!\")\r\ncv2.destroyAllWindows()\r\n", "sub_path": "trainer.py", "file_name": "trainer.py", "file_ext": "py", "file_size_in_byte": 686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "cv2.createLBPHFaceRecognizer", "line_number": 6, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 10, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 14, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 14, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "286722309", "text": "import bs4,requests\r\n\r\ndef get_url_song_list(name):\r\n\tname = name.replace(\" \",\"+\")\r\n\tlink = \"http://www.lyricsfreak.com/search.php?a=search&type=band&q=\"\r\n\tresponse = requests.get(link + name)\r\n\ttry:\r\n\t\tresponse.raise_for_status()\r\n\texcept:\r\n\t\tprint (\"Check your internet connection and try again\")\r\n\t\treturn -1\r\n\r\n\thtml = bs4.BeautifulSoup(response.text,\"html.parser\")\r\n\tsearch_result = html.select(\"td.colfirst\")\r\n\tif (len(search_result) == 0):\r\n\t\tprint (\"No artist found for the given query\")\r\n\t\treturn 0\r\n\r\n\thtml_url = search_result[0].find_all('a',href=True) #list\r\n\turl = html_url[0]['href']\r\n\treturn get_url_songs(url)\r\n\r\ndef get_url_songs(url):\r\n\tlink = \"http://www.lyricsfreak.com\"\r\n\tresponse = requests.get(link + url)\r\n\ttry:\r\n\t\tresponse.raise_for_status()\r\n\texcept:\r\n\t\tprint (\"Check your internet connection and try again\")\r\n\t\treturn -1\r\n\r\n\thtml = bs4.BeautifulSoup(response.text,\"html.parser\")\r\n\tsearch_result1 = html.select(\"div.colortable.green.mtop20\")\r\n\tsearch_result = search_result1[0].select('td.colfirst')\r\n\tif (len(search_result) == 0):\r\n\t\tprint (\"No song found for the given artist\")\r\n\t\treturn 0\r\n\r\n\ttotal_count = 0\r\n\tfor song in search_result:\r\n\t\tsong_link = song.find_all('a',href=True)[0]['href']\r\n\t\ttotal_count += lyrics_count(song_link)\r\n\r\n\treturn total_count\r\n\r\ndef lyrics_count(song_link):\r\n\tlink = \"http://www.lyricsfreak.com\"\r\n\tresponse = requests.get(link + song_link)\r\n\tresponse.raise_for_status()\r\n\thtml = bs4.BeautifulSoup(response.text,\"html.parser\")\r\n\tlyrics_html = html.select(\"div#content_h > br\")\r\n\tif (len(lyrics_html) == 0):\r\n\t\treturn 0\r\n\tlyrics = str(lyrics_html[0]).replace(\"
\",\" \").replace(\"
\",\" \")\r\n\r\n\tword_count = lyrics.count(word)\r\n\treturn word_count\r\n\r\nfile = input(\"Input the file location : \")\r\ndata = open(file)\r\nartists = {}\r\nj = -2\r\nfor i,name in enumerate(data):\r\n\tover = False\r\n\tif (len(name) == 1):\r\n\t\tj = i\r\n\t\tcontinue\r\n\telif (i == j+1):\r\n\t\tword = name\r\n\t\tbreak\r\n\tname = name.strip('\\n')\r\n\tartists[name] = 0\r\ndata.close()\r\nj = 1\r\nfor name in artists:\r\n\tartists[name] = get_url_song_list(name)\r\n\tprint (\"Word Count for %s artists have been done. Wait for some more time\" %(j))\r\n\tj += 1\r\n\r\nprint (\"Counting Words done. Now printing the ascending order\")\r\n\r\norder = sorted(artists, key=artists.__getitem__,reverse=True)\r\nfor artist in order:\r\n\tprint (artist)\r\n", "sub_path": "Problem 1.py", "file_name": "Problem 1.py", "file_ext": "py", "file_size_in_byte": 2324, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "requests.get", "line_number": 6, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 32, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 48, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "511895850", "text": "\"\"\"\n\nInitialization module\n\"\"\"\n\n\n\nimport torch\nimport numpy as np\nimport torch.nn as nn\nimport torch.nn.functional as F\n\n\ndef init_bias(net, data):\n layers = list(net._modules.keys())\n # initialize layer in order\n\n for layer in layers[:-1]:\n # set bias as projection's mean\n projection = net(data, input_=layers[0], layer=layer + '_projection')\n if 'fc' in layer:\n net._modules[layer].bias = torch.nn.Parameter(\n projection.mean(dim=0), requires_grad=False)\n elif 'conv' in layer:\n net._modules[layer].bias = torch.nn.Parameter(\n projection.transpose_(0, 1).reshape((projection.size(0), -1)).mean(dim=1),\n requires_grad=False)\n del projection\n return None\n\n\ndef init_bias_last_layer(net, data, layer, criterion, target, dtype, input_=None):\n p2 = net(data, input_=input_, layer=layer + '_projection')\n unique_p2 = torch.unique(p2, sorted=True).to(dtype=dtype)\n temp_bias = -1.0 * (\n unique_p2 + torch.cat([unique_p2[0:1] - 0.1, unique_p2[:-1]])) / 2\n\n new_projection = p2 + temp_bias.reshape((1, -1))\n yp = net(new_projection, input_=layer + '_ap')\n loss_group = criterion(yp, target)\n best_index = loss_group.argmin()\n best_bias = temp_bias[best_index]\n net._modules[layer].bias.data.fill_(best_bias)\n\n return loss_group[best_index].item(), best_bias", "sub_path": "core/init_module.py", "file_name": "init_module.py", "file_ext": "py", "file_size_in_byte": 1406, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "torch.nn.Parameter", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 22, "usage_type": "attribute"}, {"api_name": "torch.nn.Parameter", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 25, "usage_type": "attribute"}, {"api_name": "torch.unique", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 36, "usage_type": "call"}]} +{"seq_id": "452205781", "text": "from flask import Flask, render_template, request, redirect, url_for, flash, jsonify\napp = Flask(__name__)\n\n@app.route(\"/\")\n@app.route(\"/index\")\ndef hello():\n return render_template('main.html')\n\n@app.route(\"/test\",methods = ['GET','POST'])\ndef on_test():\n\tif request.method == 'POST':\n\t\tprint('message!')\n\t\tprint(request)\n\t\treturn 'YES!'\n\telse:\n\t\tprint('mmm')\n\t\treturn 'sup'\n\nif __name__ == '__main__':\n\tapp.debug = True\n\tapp.run(host='0.0.0.0', port=5000)\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 487, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.Flask", "line_number": 2, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 7, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 11, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 11, "usage_type": "name"}, {"api_name": "flask.request", "line_number": 13, "usage_type": "argument"}]} +{"seq_id": "198323036", "text": "def verify_humanity(g_recaptcha_response=None):\n from django.conf import settings\n import requests, json\n\n response = requests.post(settings.RECAPTCHA_URL, data={\n 'secret': settings.RECAPTCHA_SECRET,\n 'response': g_recaptcha_response\n })\n\n return json.loads(response.text)['success']\n\ndef upload_image(image, upload_from_path=True):\n from django.conf import settings\n from imgurpython import ImgurClient\n from imgurpython.helpers.error import ImgurClientError\n\n client = ImgurClient(settings.IMGUR_CLIENT_ID, settings.IMGUR_CLIENT_SECRET)\n\n try:\n if upload_from_path:\n imgur_image = client.upload_from_path(image.path)\n else:\n imgur_image = client.upload_from_url(image)\n\n if imgur_image and imgur_image['link']:\n return imgur_image['link']\n\n except ImgurClientError:\n return False\n\ndef resize_image(imgur_url, new_size=\"medium_thumbnail\"):\n SIZES = {\n \"small_square\" : \"s\", # 90x90 *\n \"big_square\" : \"b\", # 160x160 *\n \"small_thumbnail\" : \"t\", # 160x160\n \"medium_thumbnail\" : \"m\", # 320x320\n \"large_thumbnail\" : \"l\", # 640x640\n \"huge_thumbnail\" : \"h\" # 1024x1024\n\n # * does not keep image proportions\n }\n\n thumb_arr = imgur_url.split(\".\")\n imgur_code = thumb_arr.pop(-2)\n thumb_arr.insert(-1, (imgur_code+SIZES[new_size]))\n return \".\".join(thumb_arr)\n\ndef generate_pin():\n import random\n return random.randint(1001, 9998)\n\ndef generate_listing_code():\n import random\n from myapp.models import Listing\n\n def generate_key():\n alpha = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ'\n numbers = '0123456789'\n firstPart = random.choice(alpha) + random.choice(alpha) + random.choice(alpha)\n secondPart = random.choice(numbers) + random.choice(numbers) + random.choice(numbers)\n code = firstPart + secondPart\n return code\n\n code = generate_key()\n\n if len(Listing.objects.filter(code=code)) > 0:\n return generate_listing_code()\n\n return code\n", "sub_path": "myapp/helpers.py", "file_name": "helpers.py", "file_ext": "py", "file_size_in_byte": 1931, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "requests.post", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.settings.RECAPTCHA_URL", "line_number": 5, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 5, "usage_type": "name"}, {"api_name": "django.conf.settings.RECAPTCHA_SECRET", "line_number": 6, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 6, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 10, "usage_type": "call"}, {"api_name": "imgurpython.ImgurClient", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.settings.IMGUR_CLIENT_ID", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 17, "usage_type": "name"}, {"api_name": "django.conf.settings.IMGUR_CLIENT_SECRET", "line_number": 17, "usage_type": "attribute"}, {"api_name": "imgurpython.helpers.error.ImgurClientError", "line_number": 28, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 50, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 59, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 60, "usage_type": "call"}, {"api_name": "myapp.models.Listing.objects.filter", "line_number": 66, "usage_type": "call"}, {"api_name": "myapp.models.Listing.objects", "line_number": 66, "usage_type": "attribute"}, {"api_name": "myapp.models.Listing", "line_number": 66, "usage_type": "name"}]} +{"seq_id": "419113737", "text": "from selenium import webdriver\nfrom PIL import Image\nimport time\n\n# take screenshot\ndriver = webdriver.Chrome();\n\ndriver.implicitly_wait(9)\n\ndriver.get('https://wiki.python.org');\nelement = driver.find_element_by_id(\"logo\");\nlocation = element.location;\nsize = element.size;\ndriver.save_screenshot(\"pageImage.png\");\n\n# crop image\nx = location['x'];\ny = location['y'];\nwidth = location['x']+size['width'];\nheight = location['y']+size['height'];\nim = Image.open('pageImage.png')\nim = im.crop((int(x), int(y), int(width), int(height)))\nim.save('element.png')\n\ntime.sleep(9)\n\ndriver.quit()\n", "sub_path": "ScreenShot.py", "file_name": "ScreenShot.py", "file_ext": "py", "file_size_in_byte": 586, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "selenium.webdriver.Chrome", "line_number": 6, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 6, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 21, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 21, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "42691842", "text": "\"\"\"\nTrainer module tests\n\"\"\"\n\nimport os\nimport unittest\nimport tempfile\n\nimport torch\n\nfrom transformers import AutoTokenizer, AutoModelForSequenceClassification\n\nfrom txtai.pipeline import HFTrainer, Labels, Questions\n\n\nclass TestTrainer(unittest.TestCase):\n \"\"\"\n Trainer tests\n \"\"\"\n\n @classmethod\n def setUpClass(cls):\n \"\"\"\n Create default datasets\n \"\"\"\n\n cls.data = [{\"text\": \"Dogs\", \"label\": 0}, {\"text\": \"dog\", \"label\": 0}, {\"text\": \"Cats\", \"label\": 1}, {\"text\": \"cat\", \"label\": 1}] * 100\n\n def testBasic(self):\n \"\"\"\n Test training a model with basic parameters\n \"\"\"\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\"google/bert_uncased_L-2_H-128_A-2\", self.data)\n\n labels = Labels((model, tokenizer), dynamic=False)\n self.assertEqual(labels(\"cat\")[0][0], 1)\n\n def testCustom(self):\n \"\"\"\n Test training a model with custom parameters\n \"\"\"\n\n # pylint: disable=E1120\n model = AutoModelForSequenceClassification.from_pretrained(\"google/bert_uncased_L-2_H-128_A-2\")\n tokenizer = AutoTokenizer.from_pretrained(\"google/bert_uncased_L-2_H-128_A-2\")\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\n (model, tokenizer),\n self.data,\n self.data,\n columns=(\"text\", \"label\"),\n do_eval=True,\n output_dir=os.path.join(tempfile.gettempdir(), \"trainer\"),\n )\n\n labels = Labels((model, tokenizer), dynamic=False)\n self.assertEqual(labels(\"cat\")[0][0], 1)\n\n def testDataframe(self):\n \"\"\"\n Test training a model with a mock pandas DataFrame\n \"\"\"\n\n # pylint: disable=W0613\n def to_dict(orient):\n return self.data\n\n df = unittest.mock.Mock(spec=[\"to_dict\"])\n df.to_dict = to_dict\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\"google/bert_uncased_L-2_H-128_A-2\", df)\n\n labels = Labels((model, tokenizer), dynamic=False)\n self.assertEqual(labels(\"cat\")[0][0], 1)\n\n def testDataset(self):\n \"\"\"\n Test training a model with a mock Hugging Face Dataset\n \"\"\"\n\n class TestDataset(torch.utils.data.Dataset):\n \"\"\"\n Test Dataset\n \"\"\"\n\n def __init__(self, data):\n self.data = data\n self.unique = lambda _: [0, 1]\n\n def __len__(self):\n return len(self.data)\n\n def __getitem__(self, index):\n return self.data[index]\n\n def column_names(self):\n \"\"\"\n Returns column names for this dataset\n\n Returns:\n list of columns\n \"\"\"\n\n return [\"text\", \"label\"]\n\n # pylint: disable=W0613\n def map(self, fn, batched, remove_columns):\n \"\"\"\n Map each dataset row using fn.\n\n Args:\n fn: function\n batched: batch records\n\n Returns:\n updated Dataset\n \"\"\"\n\n self.data = [fn(x) for x in self.data]\n return self\n\n ds = TestDataset(self.data)\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\"google/bert_uncased_L-2_H-128_A-2\", ds)\n\n labels = Labels((model, tokenizer), dynamic=False)\n self.assertEqual(labels(\"cat\")[0][0], 1)\n\n def testQA(self):\n \"\"\"\n Tests training a QA model.\n \"\"\"\n\n # Training data\n data = [\n {\"question\": \"What ingredient?\", \"context\": \"1 can whole tomatoes\", \"answers\": \"tomatoes\"},\n {\"question\": \"What ingredient?\", \"context\": \"1 yellow onion\", \"answers\": \"onion\"},\n {\"question\": \"What ingredient?\", \"context\": \"1 red pepper\", \"answers\": \"pepper\"},\n {\"question\": \"What ingredient?\", \"context\": \"1 clove garlic\", \"answers\": \"garlic\"},\n {\"question\": \"What ingredient?\", \"context\": \"1/2 lb beef\", \"answers\": \"beef\"},\n {\"question\": \"What ingredient?\", \"context\": \"a \" * 500 + \"1/2 lb beef\", \"answers\": \"beef\"},\n {\"question\": \"What ingredient?\", \"context\": \"Forest through the trees\", \"answers\": None},\n ]\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\"google/bert_uncased_L-2_H-128_A-2\", data, data, task=\"question-answering\", num_train_epochs=10)\n\n questions = Questions((model, tokenizer), gpu=True)\n self.assertEqual(questions([\"What ingredient?\"], [\"Peel 1 onion\"])[0], \"onion\")\n\n def testRegression(self):\n \"\"\"\n Tests training a model with a regression (continuous) output.\n \"\"\"\n\n data = []\n for x in self.data:\n x[\"label\"] = float(x[\"label\"])\n data.append(x)\n\n trainer = HFTrainer()\n model, tokenizer = trainer(\"google/bert_uncased_L-2_H-128_A-2\", data)\n\n labels = Labels((model, tokenizer), dynamic=False)\n\n # Regression tasks return a single entry with the regression output\n self.assertGreater(labels(\"cat\")[0][1], 0.5)\n", "sub_path": "test/python/testtrainer.py", "file_name": "testtrainer.py", "file_ext": "py", "file_size_in_byte": 5174, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 16, "usage_type": "attribute"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 34, "usage_type": "call"}, {"api_name": "txtai.pipeline.Labels", "line_number": 37, "usage_type": "call"}, {"api_name": "transformers.AutoModelForSequenceClassification.from_pretrained", "line_number": 46, "usage_type": "call"}, {"api_name": "transformers.AutoModelForSequenceClassification", "line_number": 46, "usage_type": "name"}, {"api_name": "transformers.AutoTokenizer.from_pretrained", "line_number": 47, "usage_type": "call"}, {"api_name": "transformers.AutoTokenizer", "line_number": 47, "usage_type": "name"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "tempfile.gettempdir", "line_number": 56, "usage_type": "call"}, {"api_name": "txtai.pipeline.Labels", "line_number": 59, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 71, "usage_type": "call"}, {"api_name": "unittest.mock", "line_number": 71, "usage_type": "attribute"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 74, "usage_type": "call"}, {"api_name": "txtai.pipeline.Labels", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 128, "usage_type": "call"}, {"api_name": "txtai.pipeline.Labels", "line_number": 131, "usage_type": "call"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 150, "usage_type": "call"}, {"api_name": "txtai.pipeline.Questions", "line_number": 153, "usage_type": "call"}, {"api_name": "txtai.pipeline.HFTrainer", "line_number": 166, "usage_type": "call"}, {"api_name": "txtai.pipeline.Labels", "line_number": 169, "usage_type": "call"}]} +{"seq_id": "74047155", "text": "from django.conf.urls import url\n\nfrom . import views\n\nurlpatterns = [\n url(r'^$', views.PostListView.as_view(), name='index'),\n url(r'^p/(?P.+)/(?P.+)/$', views.PostDetailView.as_view(),\n name=\"detail\"),\n url(r'^create/video/$', views.PostCreateView.as_view(),\n name=\"create\"), \n url(r'^parse/$', views.CheckInfoYoutubeView.as_view(),\n name=\"parse\"),\n\n]\n", "sub_path": "apps/posts/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "294205438", "text": "# -*- coding: utf-8 -*-\n\nfrom tools.city import City\nimport csv \nimport sys\n\nclass DataLoader(object):\n '''Methods for reading city data and returning a list of City objects.\n \n Methods:\n get_cities_tsv -- read tab-separated geonames file. \n \n '''\n \n @classmethod\n def get_cities_tsv(cls, fileName):\n '''Returns a list of city objects from a geonames tsv file.\n \n The file is expected to be tab-separated, with the first row being the\n field names. The field names should contain: name, id, alt_name, lat, long,\n and country. The field alt_name gives the alternative names of the \n city, comma-separated. The character encoding is UTF-8.\n \n Arguments:\n fileName -- path to the TSV file.\n \n Returns:\n A list of city objects corresponding to cities in the input file. \n \n Raises:\n NoneUniqueIDException -- raised if file contains two cities with the\n same id field.\n \n '''\n cities = []\n # Hash set of city IDs, to ensure IDs are unique. \n IDs = set()\n \n csv.field_size_limit(sys.maxsize)\n \n with open(fileName, 'rt') as f:\n reader = csv.DictReader(f, delimiter='\\t',\n quoting=csv.QUOTE_NONE)\n for row in reader: \n altnames = row['alt_name'].split(',')\n ID = int(row['id'])\n \n if(ID in IDs):\n raise NoneUniqueIDException('ID ' + row['id'] + ' duplicated')\n \n IDs.add(ID)\n cities.append(City(ID, row['name'], altnames, \n float(row['lat']), float(row['long']), \n row['country']))\n \n \n return cities \n \n\nclass NoneUniqueIDException(Exception):\n '''Exception raised if two cities in a file have the same ID.'''\n pass ", "sub_path": "tools/dataloader.py", "file_name": "dataloader.py", "file_ext": "py", "file_size_in_byte": 2052, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "csv.field_size_limit", "line_number": 39, "usage_type": "call"}, {"api_name": "sys.maxsize", "line_number": 39, "usage_type": "attribute"}, {"api_name": "csv.DictReader", "line_number": 42, "usage_type": "call"}, {"api_name": "csv.QUOTE_NONE", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tools.city.City", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "86898138", "text": "import datetime\nimport os\nimport tarfile\nimport time\nimport sys\n\nimport numpy as np\nimport pandas as pd\nimport shapefile\nfrom osgeo import gdal\nfrom osgeo import gdalnumeric\nfrom shapely.geometry import Point # Point class\nfrom shapely.geometry import shape # shape() is a function to convert geo objects through the interface\nimport pickle\n\ndef pixelToEarth(pt,gt):\n pi = pt[0] + 0.5\n pj = pt[1] + 0.5\n latitude = gt[3] + pj * gt[4] + pi * gt[5] \n longitude = gt[0] + pj * gt[1] + pi * gt[2]\n return longitude, latitude\n\ndef findCountry(point):\n inquiringPoint = Point(point)\n for i in range(len(shape_boundary)):\n if inquiringPoint.within(shape_boundary[i]):\n return all_records[i][0]\n return 0\n\nif __name__ == '__main__':\n \n sf = shapefile.Reader('Countries_WGS84/Countries_WGS84.shp')\n all_shapes = sf.shapes()\n all_records = sf.records()\n\n shape_boundary = []\n for i in range(len(all_shapes)):\n shape_boundary.append(shape(all_shapes[i]))\n \n pixelOwners4km = []; \n \n Tile_types = [\"00N060E\",\"00N060W\",\"00N180W\",\"75N060E\",\"75N060W\",\"75N180W\"]\n for k, tilename in enumerate(Tile_types):\n print(\"[ %d ]\" % (k))\n _tif_filename = \"SVDNB_npp_20180501-20180531_\" + tilename + \"_vcmcfg_v10_c201806061100.avg_rade9h.tif.4km.tif\"\n ds = gdal.Open(\"GeoFence4km/\" + _tif_filename)\n width = ds.RasterXSize\n height = ds.RasterYSize\n gt = ds.GetGeoTransform()\n pixelOwner = np.zeros([height,width]);\n for i in range(height):\n print(\"%d%%\"%(i*100/height))\n sys.stdout.flush()\n log ,lat = pixelToEarth((i,0), gt)\n if abs(lat)<=60:\n for j in range(width):\n log,lat = pixelToEarth((i,j), gt)\n pixelOwner[i][j] = findCountry((log,lat))\n \n pixelOwners4km.append(pixelOwner)\n del ds\n\n # Saving the objects:\n np.save(\"pixel4km.npy\", pixelOwners4km)\n\n", "sub_path": "satellite_B_countries_4km.py", "file_name": "satellite_B_countries_4km.py", "file_ext": "py", "file_size_in_byte": 2013, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "shapely.geometry.Point", "line_number": 24, "usage_type": "call"}, {"api_name": "shapefile.Reader", "line_number": 32, "usage_type": "call"}, {"api_name": "shapely.geometry.shape", "line_number": 38, "usage_type": "call"}, {"api_name": "osgeo.gdal.Open", "line_number": 46, "usage_type": "call"}, {"api_name": "osgeo.gdal", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.stdout.flush", "line_number": 53, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 64, "usage_type": "call"}]} +{"seq_id": "634347521", "text": "from datetime import datetime\nfrom os import listdir, symlink, system\nfrom os.path import dirname, expanduser, isdir\n\nfrom click import command, confirm\n\nfrom kraft import print_header_in_terminal, remove_path, run_command\n\nfrom .CONDA_YAML_EXTENSION import CONDA_YAML_EXTENSION\nfrom .get_project_directory_path import get_project_directory_path\nfrom .THIS_DIRECTORY_PATH import THIS_DIRECTORY_PATH\n\n\n@command()\ndef enter():\n \"\"\"* Enter project environment.\"\"\"\n\n environment_directory_path = \"{}/environment\".format(get_project_directory_path())\n\n for name in listdir(path=environment_directory_path):\n\n if name.endswith(CONDA_YAML_EXTENSION):\n\n conda_yaml_file_path = \"{}/{}\".format(environment_directory_path, name)\n\n conda_directory_path = conda_yaml_file_path[: -len(CONDA_YAML_EXTENSION)]\n\n run_command(\n \"conda env create --file {} --prefix {} --force\".format(\n conda_yaml_file_path, conda_directory_path\n )\n )\n\n essence_directory_path = \"{}/essence\".format(environment_directory_path)\n\n symlink(\n run_command(\"which spro\").stdout.strip(),\n \"{}/bin/spro\".format(essence_directory_path),\n )\n\n conda_activate_commands = [\"conda activate {}\".format(essence_directory_path)]\n\n for name in sorted(listdir(path=environment_directory_path))[::-1]:\n\n path = \"{}/{}\".format(environment_directory_path, name)\n\n if isdir(path) and path != essence_directory_path:\n\n conda_activate_commands.append(\"conda activate --stack {}\".format(path))\n\n with open(\"{}/spro_enter_template.sh\".format(THIS_DIRECTORY_PATH)) as io:\n\n spro_enter_str = io.read()\n\n for str_0, str_1 in (\n (\n \"$SPRO_CONDA_BASE_DIRECTORY_PATH\",\n dirname(dirname(run_command(\"which conda\").stdout.strip())),\n ),\n (\"$SPRO_CONDA_ACTIVATE_COMMAND\", \"\\n\\n\".join(conda_activate_commands)),\n (\"$SPRO_ENVIRONMENT_DIRECTORY_PATH\", environment_directory_path),\n ):\n\n spro_enter_str = spro_enter_str.replace(str_0, str_1)\n\n spro_enter_sh_file_path = \"{}/.spro_enter.{:%Y-%m-%d-%H-%M-%S}.sh\".format(\n expanduser(\"~\"), datetime.now()\n )\n\n with open(spro_enter_sh_file_path, mode=\"w\") as io:\n\n io.write(spro_enter_str)\n\n print_header_in_terminal(\"Entered project environment. Exit by `exit`\")\n\n system(\"bash --rcfile {}\".format(spro_enter_sh_file_path))\n\n remove_path(spro_enter_sh_file_path)\n\n if confirm(\"Do you want to export environment?\"):\n\n for name in listdir(path=environment_directory_path):\n\n path = \"{}/{}\".format(environment_directory_path, name)\n\n if isdir(path):\n\n run_command(\n \"conda env export --prefix {0} > {0}{1}\".format(\n path, CONDA_YAML_EXTENSION\n )\n )\n", "sub_path": "spro/enter.py", "file_name": "enter.py", "file_ext": "py", "file_size_in_byte": 2909, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "get_project_directory_path.get_project_directory_path", "line_number": 18, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 20, "usage_type": "call"}, {"api_name": "CONDA_YAML_EXTENSION.CONDA_YAML_EXTENSION", "line_number": 22, "usage_type": "argument"}, {"api_name": "CONDA_YAML_EXTENSION.CONDA_YAML_EXTENSION", "line_number": 26, "usage_type": "argument"}, {"api_name": "kraft.run_command", "line_number": 28, "usage_type": "call"}, {"api_name": "os.symlink", "line_number": 36, "usage_type": "call"}, {"api_name": "kraft.run_command", "line_number": 37, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 47, "usage_type": "call"}, {"api_name": "THIS_DIRECTORY_PATH.THIS_DIRECTORY_PATH", "line_number": 51, "usage_type": "argument"}, {"api_name": "os.path.dirname", "line_number": 58, "usage_type": "call"}, {"api_name": "kraft.run_command", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 67, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 67, "usage_type": "name"}, {"api_name": "kraft.print_header_in_terminal", "line_number": 74, "usage_type": "call"}, {"api_name": "os.system", "line_number": 76, "usage_type": "call"}, {"api_name": "kraft.remove_path", "line_number": 78, "usage_type": "call"}, {"api_name": "click.confirm", "line_number": 80, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 86, "usage_type": "call"}, {"api_name": "kraft.run_command", "line_number": 88, "usage_type": "call"}, {"api_name": "CONDA_YAML_EXTENSION.CONDA_YAML_EXTENSION", "line_number": 90, "usage_type": "argument"}, {"api_name": "click.command", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "233379209", "text": "from bigdata import *\nimport json\nimport sys\n\nparse_data = parse_letter(parse_site(\"2017년 축제\", 100))\ncnt = len(parse_data)\ngood_word = {}\nbad_word = {}\nnormal_word = {}\n\ngood_text = \"\"\n\nbad_text = \"\"\nnormal_text = \"\"\nword_count = 0\n\ndef analysis(sentence):\n try:\n sentence = sentence.replace(\"\\n\", \" \")\n global word_count\n word_count += len(sentence.split(\" \"))\n #s = requests.session()\n response = requests.get('http://api.datamixi.com/datamixiApi/tms?query=%s&lang=kor&analysis=om' % (sentence))\n j = json.loads(response.text)\n return j[\"return_object\"][\"sentence\"][0]\n except:\n pass\n \nfor i in range(cnt):\n data = analysis(parse_data[i])\n if data == None:\n continue\n try:\n if(data['sa']['polarity'] == 1):\n for a in data['morp']:\n if(a['type'] == 'NNG' or a['type'] == 'NNP'):\n try:\n good_word[a['lemma']] += 1\n except:\n good_word[a['lemma']] = 1\n \n elif(data['sa']['polarity'] == -1):\n for a in data['morp']:\n if(a['type'] == 'NNG' or a['type'] == 'NNP'):\n try:\n bad_word[a['lemma']] += 1\n except:\n bad_word[a['lemma']] = 1\n except:\n for a in data['morp']:\n if(a[\"type\"] == \"NNG\" or a[\"type\"] == \"NNP\"):\n try:\n normal_word[a[\"lemma\"]] += 1\n except:\n normal_word[a[\"lemma\"]] = 1\n \n #pprint(analysis(parse_data[i]))\n\nfor c in good_word.items():\n good_text += \"%-5d %s\\n\"% (c[1], c[0])\n\nfor c in bad_word.items():\n bad_text += \"%-5d %s\\n\"% (c[1], c[0])\n\nfor c in normal_word.items():\n normal_text += \"%-5d %s\\n\"% (c[1], c[0]) \n\nwith open(\"good.txt\", \"wb\") as f:\n f.write(good_text.encode(\"utf-8\"))\n\nwith open(\"bad.txt\", \"wb\") as f:\n f.write(bad_text.encode(\"utf-8\"))\n\nwith open(\"normal.txt\", \"wb\") as f:\n f.write(normal_text.encode(\"utf-8\")) \n", "sub_path": "final/adam.py", "file_name": "adam.py", "file_ext": "py", "file_size_in_byte": 2108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "json.loads", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "497579336", "text": "import json\r\nfrom json_worker import JsonObj\r\nfrom bottle import request\r\nfrom datetime import datetime\r\n\r\nwith open('tickets_updated.json', 'r+') as file:\r\n tickets_list_up = json.load(file)\r\n\r\nclass TicketsSorter(object):\r\n def __init__(self, tickets_list):\r\n \"\"\"\r\n Gerador de objetos\r\n \"\"\"\r\n self.tickets_list = tickets_list\r\n\r\n def created_interval(self, interval1 = '', interval2 = '' ):\r\n\r\n format = \"%Y-%m-%d %H:%M:%S\"\r\n if interval1 == '':\r\n interval1 = '1900-01-01 00:00:00'\r\n if interval2 == '':\r\n interval2 = '2100-01-01 00:00:00' \r\n\r\n datetime_object1 = datetime.strptime(interval1, format)\r\n datetime_object2 = datetime.strptime(interval2, format)\r\n\r\n tickets = []\r\n for item in self.tickets_list:\r\n created_str = item.DateCreate\r\n created = datetime.strptime(created_str, format)\r\n if created >= datetime_object1 and created <= datetime_object2:\r\n tickets.append(item)\r\n else:\r\n pass\r\n\r\n return tickets\r\n\r\n def sort_by_ID(self, intervalo1=\"\", intervalo2=\"\"):\r\n \"\"\"\r\n Organiza por ordem de ID\r\n \"\"\"\r\n tickets_list = self.created_interval(intervalo1, intervalo2)\r\n organizado = sorted(tickets_list, key=lambda ticket: ticket.TicketID, reverse=False)\r\n result = []\r\n for ticket in organizado:\r\n result.append(ticket.__dict__)\r\n return result\r\n\r\n def sort_by_rating(self, intervalo1=\"\", intervalo2=\"\"):\r\n \"\"\"\r\n Organiza por ordem de Rating\r\n \"\"\"\r\n tickets_list = self.created_interval(intervalo1, intervalo2)\r\n organizado = sorted(tickets_list, key=lambda ticket: ticket.Rating, reverse=True)\r\n result = []\r\n for ticket in organizado:\r\n result.append(ticket.__dict__)\r\n return result\r\n\r\n def sort_by_created_at(self, intervalo1=\"\", intervalo2=\"\"):\r\n \"\"\"\r\n Organiza por ordem de data de criação\r\n \"\"\"\r\n tickets_list = self.created_interval(intervalo1, intervalo2)\r\n organizado = sorted(tickets_list, key=lambda ticket: ticket.created_at(), reverse=False)\r\n result = []\r\n for ticket in organizado:\r\n result.append(ticket.__dict__)\r\n return result\r\n\r\n def sort_by_updated_at(self, intervalo1=\"\", intervalo2=\"\"):\r\n \"\"\"\r\n Organiza por ordem de data de update\r\n \"\"\"\r\n tickets_list = self.created_interval(intervalo1, intervalo2)\r\n organizado = sorted(tickets_list, key=lambda ticket: ticket.updated_at(), reverse=False)\r\n result = []\r\n for ticket in organizado:\r\n result.append(ticket.__dict__)\r\n return result\r\n\r\n\r\n\r\ndef order_filter(tickets_list, filter_by='', order_by='ID', page=0, interval1='', interval2=''):\r\n \"\"\"\r\n Gera os filtros e ordens necessárias para a API (Mergir para a classe TicketsSorter)\r\n \"\"\"\r\n \r\n lista = JsonObj(tickets_list).json_obj_updated()\r\n order_by_ID = TicketsSorter(lista).sort_by_ID(interval1, interval2)\r\n order_by_rating = TicketsSorter(lista).sort_by_rating(interval1, interval2)\r\n order_by_created = TicketsSorter(lista).sort_by_created_at(interval1, interval2)\r\n order_by_updated = TicketsSorter(lista).sort_by_updated_at(interval1, interval2)\r\n\r\n if order_by == 'rating':\r\n lista = order_by_rating\r\n elif order_by == 'created':\r\n lista = order_by_created\r\n elif order_by == 'updated':\r\n lista = order_by_updated\r\n else:\r\n lista = order_by_ID\r\n\r\n if request.query.page == '':\r\n page = 0\r\n else:\r\n page = int(request.query.page)\r\n\r\n if filter_by != '':\r\n if page > 0:\r\n tickets_requested = [ticket for ticket in lista if ticket['Priority'] == filter_by ]\r\n return { 'tickets' : tickets_requested[((page-1)*5):page*5] }\r\n else:\r\n tickets_requested = [ticket for ticket in lista if ticket['Priority'] == filter_by ]\r\n return { 'tickets' : tickets_requested }\r\n else:\r\n if page > 0:\r\n return { 'tickets' : lista[((page-1)*5):page*5] }\r\n else:\r\n return { 'tickets' : lista }", "sub_path": "obj_filters.py", "file_name": "obj_filters.py", "file_ext": "py", "file_size_in_byte": 4273, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "json.load", "line_number": 7, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "name"}, {"api_name": "json_worker.JsonObj", "line_number": 89, "usage_type": "call"}, {"api_name": "bottle.request.query", "line_number": 104, "usage_type": "attribute"}, {"api_name": "bottle.request", "line_number": 104, "usage_type": "name"}, {"api_name": "bottle.request.query", "line_number": 107, "usage_type": "attribute"}, {"api_name": "bottle.request", "line_number": 107, "usage_type": "name"}]} +{"seq_id": "170753351", "text": "from sqlite3 import connect\nfrom sys import argv\n\n\ndef create_table(db, table):\n curs = db.cursor()\n curs.execute('CREATE TABLE IF NOT EXISTS ' + table + ' (id INTEGER, word TEXT)')\n\n db.commit()\n \n \ndef upload(db, table , wordlist):\n template = 'INSERT INTO ' + table + ' VALUES(\"%s\", \"%s\")'\n curs = db.cursor()\n\n with open(wordlist, 'r') as infile:\n for line in infile:\n line = line.strip('\\n').split(\"\\t\")\n curs.execute(template % (line[0], line[1]))\n\n db.commit()\n \n\ndef main():\n db = connect('wordlist.db')\n table = argv[1]\n wordlist = argv[2]\n\n create_table(db, table)\n upload(db, table, wordlist)\n\n\nif __name__ == '__main__':\n main()\n\n", "sub_path": "upload.py", "file_name": "upload.py", "file_ext": "py", "file_size_in_byte": 723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sqlite3.connect", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 26, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "name"}]} +{"seq_id": "468214036", "text": "import os\nimport shutil\nimport logging\nimport pickle\nimport json\nfrom sklearn.externals import joblib\nimport tensorflow as tf\n\n\ndef export_feature_transform(fb_path, conf_path, feat_builder, conf):\n with open(fb_path, 'wb') as f:\n joblib.dump(feat_builder, f)\n with open(conf_path, 'wb') as f:\n joblib.dump(conf, f)\n\n\ndef export_model_pickle(path, model):\n with open(path, 'wb') as f:\n pickle.dump(model, f)\n\n\ndef export_model_joblib(path, model):\n with open(path, 'wb') as f:\n joblib.dump(model, f, compress=True)\n\n\ndef export_model_tf(model, model_name, version, simple_save=False):\n model_base_path = os.path.realpath(\"..\")\n export_path = os.path.join(model_base_path, \"serving\", \"models\", model_name, version)\n if os.path.isdir(export_path):\n logging.warning(\"\\tModel path \\\"%s\\\" already exists, removing...\" % export_path)\n shutil.rmtree(export_path)\n if simple_save:\n print(\"simple_save is deprecated, it will be removed in tensorflow xxx...\")\n tf.saved_model.simple_save(model.sess, export_path,\n inputs={'fi': model.feature_indices,\n 'fv': model.feature_values},\n outputs={'y_prob': model.y_prob})\n else:\n builder = tf.saved_model.builder.SavedModelBuilder(export_path)\n input_fi = tf.saved_model.utils.build_tensor_info(model.feature_indices)\n input_fv = tf.saved_model.utils.build_tensor_info(model.feature_values)\n # input_label = tf.saved_model.utils.build_tensor_info(self.labels)\n input_y = tf.saved_model.utils.build_tensor_info(model.y_prob)\n\n prediction_signature = (\n tf.saved_model.signature_def_utils.build_signature_def(\n inputs={'fi': input_fi,\n 'fv': input_fv},\n outputs={'y_prob': input_y},\n method_name=tf.saved_model.signature_constants.PREDICT_METHOD_NAME))\n\n builder.add_meta_graph_and_variables(\n model.sess, [tf.saved_model.tag_constants.SERVING],\n signature_def_map={'predict': prediction_signature}\n # tf.saved_model.signature_constants.DEFAULT_SERVING_SIGNATURE_DEF_KEY: prediction_signature},\n # main_op=tf.tables_initializer(),\n # strip_default_attrs=True\n )\n\n builder.save()\n logging.warning('\\tDone exporting!')\n\n\ndef export_TFRecord():\n example = tf.parse_single_example()\n\n\n\n\n", "sub_path": "libreco/utils/serialization.py", "file_name": "serialization.py", "file_ext": "py", "file_size_in_byte": 2520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sklearn.externals.joblib.dump", "line_number": 12, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 12, "usage_type": "name"}, {"api_name": "sklearn.externals.joblib.dump", "line_number": 14, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 14, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.dump", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 24, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 31, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.saved_model.simple_save", "line_number": 35, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.builder.SavedModelBuilder", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.utils.build_tensor_info", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.utils.build_tensor_info", "line_number": 42, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.utils.build_tensor_info", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model.signature_def_utils.build_signature_def", "line_number": 47, "usage_type": "call"}, {"api_name": "tensorflow.saved_model", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tensorflow.saved_model", "line_number": 54, "usage_type": "attribute"}, {"api_name": "logging.warning", "line_number": 62, "usage_type": "call"}, {"api_name": "tensorflow.parse_single_example", "line_number": 66, "usage_type": "call"}]} +{"seq_id": "380550845", "text": "from django.shortcuts import render, redirect, HttpResponseRedirect, get_object_or_404\nfrom pagina.forms import LoginForm\nfrom django.contrib.auth import authenticate,views\nfrom django.contrib.auth.forms import AuthenticationForm\nfrom django.contrib.auth import login as auth_login\nfrom django.http import HttpRequest, HttpResponse, JsonResponse\nfrom django.contrib.auth.models import User\nfrom pagina.models import Animales, Loteria, Agencia, Sorteo, Ticket, Ticke_item, Horas, AnimalGanador\nfrom django.contrib.auth.models import User\nfrom django_ajax.decorators import ajax\nfrom django.contrib.auth.decorators import login_required\nimport secrets,json, time\nfrom django.conf import settings \nfrom django.core import serializers\nfrom datetime import datetime, timedelta, date\nfrom django.db.models import Sum\nfrom django.db import connection\nfrom collections import namedtuple\nfrom reportlab.pdfgen import canvas\nfrom reportlab.lib.units import cm\nfrom easy_pdf.views import PDFTemplateView \nfrom easy_pdf.rendering import render_to_pdf\n# Create your views here.\n\ndef login(request):\n if request.user.is_authenticated():\n return HttpResponseRedirect ('/home/')\n if request.method == 'POST':\n form = AuthenticationForm(data=request.POST)\n if form.is_valid():\n username = request.POST.get('username')\n password = request.POST.get('password')\n acceso = authenticate (username=username, password=password)\n if acceso is not None:\n if acceso.is_active:\n auth_login(request,acceso)\n return HttpResponseRedirect('/home/')\n else:\n form = AuthenticationForm()\n script = \"alert('Usuario no activo');\"\n print ('no valido')\n return render(request, 'login.html', locals())\n\n else:\n form = AuthenticationForm()\n print ('asdasdas')\n return render(request, 'login.html', locals())\n\n\ndef home (request):\n us = request.user.id\n agencia_aso = Agencia.objects.filter(usuario=us)\n loterias = Loteria.objects.filter(agencia=agencia_aso).values('idl','nombre_lot')\n\n return render (request, 'index.html', {'loterias': loterias})\n\n\ndef lotoanimal (request, loteriaid):\n lotid = loteriaid \n animalescaracas = Animales.objects.filter(idl=lotid).values('numero','nombre')\n us = request.user\n apuestaminima = Loteria.objects.filter(idl=lotid).values('agencia__apuesta_min')\n valor = apuestaminima[0]['agencia__apuesta_min']\n horas = Sorteo.objects.all()\n loteria = Loteria.objects.filter(idl=lotid).values('agencia__nombrea')\n nombreloteria = loteria[0]['agencia__nombrea']\n \n contexto = {'lista' : animalescaracas, 'mini' : valor, 'horas': horas, 'loteria':nombreloteria, 'id_loteria':lotid,}\n return render (request, 'lotoanimal.html', contexto)\n\n\n@ajax\ndef ticket(request): \n horas = request.POST.getlist('horas[]')\n todos = json.loads(request.POST.get('animales'))\n total = sum(todos.values())\n loteria = request.POST.get('loteria')\n idlote = request.POST.get('id_loteria')\n ci = request.POST.get('ci')\n us = request.user.id\n loteria = Loteria.objects.get(pk=idlote)\n agenciar = Agencia.objects.get(usuario_id=us)\n tokena = secrets.token_hex(16)\n #Crear ticket general\n ticket = Ticket (cedula=ci, ida=agenciar, idl=loteria, total=total,token=tokena)\n ticket.save()\n idticket = ticket.id_ticket\n #Crear los tickets items y los agrego al ticket general\n for k,v in todos.items():\n ticketItem = Ticke_item(id_ticket=ticket, id_animal=Animales.objects.get(nombre=k, idl=loteria), monto_apu=v)\n ticketItem.save()\n #enlazar las horas con el ticket\n horasAm=[]\n for h in horas:\n timeObj = datetime.strptime(h,'%H:%M')\n timeFix = datetime.strftime(timeObj,'%H:%M:%S')\n horat = Horas(ticket=ticket, horas=Sorteo.objects.get(hora=timeFix))\n horat.save()\n horasAm.append(datetime.strftime(timeObj,'%I:%M%p'))\n\n return {'horas': horasAm, 'total': total, 'animales': todos,'loteria': loteria, 'ci': ci, 'agencia': agenciar, 'token': tokena, 'idtk':idticket,}\n\n@ajax\ndef search (request):\n ticketbuscado = request.POST.get('busquedatk')\n ticketb = Ticket.objects.filter(id_ticket=ticketbuscado).values('id_ticket','total','token')\n serializeFecha = serializers.serialize(\"json\", Ticket.objects.filter(id_ticket=ticketbuscado), fields=('fecha'))\n fechatk = json.loads(serializeFecha)\n us = request.user.id\n agencia = Agencia.objects.get(usuario_id=us)\n idlote = request.POST.get('idloteria')\n loteria = Loteria.objects.get(pk=idlote)\n sorteos = Horas.objects.filter(ticket_id=ticketbuscado)\n sorteosAm = []\n \n if ticketb.count() > 0:\n ticketitems = Ticke_item.objects.filter (id_ticket=ticketbuscado).values('id_animal__nombre', 'monto_apu')\n print('si hay tickets')\n return {'ticketitem': ticketitems,'loteria': loteria, 'ticketb': ticketb,'sorteos': sorteos,'agencia': agencia, 'fecha': fechatk[0]['fields']['fecha'], 'tkbuscado': ticketbuscado}\n else:\n print('no hay ticket')\n \n\ndef taquilla (request):\n hoy = datetime.today()\n ticketsHoy = Ticket.objects.filter(fecha=hoy).values('id_ticket')\n ticketDiarios = Ticket.objects.filter(fecha=hoy).count()\n\n if ticketDiarios == 0:\n totalVentas = {'total__sum': 0}\n gananciaBank = 0\n else:\n totalVentas = Ticket.objects.filter(fecha=hoy).aggregate(Sum('total'))\n gananciaBank = totalVentas['total__sum']*10/100\n \n tkitemsHoy = Ticke_item.objects.filter(id_ticket__fecha=hoy).values('id_animal','id_ticket')\n animalesGanadores = AnimalGanador.objects.filter(fecha=hoy).values('animal__id_animal', 'hora')\n\n with connection.cursor() as cursor:\n cursor.execute(\"SELECT COUNT(distinct ticket_id) FROM pagina_ticke_item inner join pagina_animalganador ON (pagina_ticke_item.id_animal_id = pagina_animalGanador.animal_id AND pagina_animalganador.fecha = CURDATE()) inner join pagina_ticket ON (pagina_ticket.id_ticket = pagina_ticke_item.id_ticket_id) inner join pagina_horas ON (pagina_horas.ticket_id = pagina_ticket.id_ticket AND pagina_horas.horas_id=pagina_animalganador.hora_id) WHERE pagina_ticket.fecha = CURDATE()\")\n row = cursor.fetchall()\n \n contexto = {'tikets_diarios':ticketDiarios,'bankGanacia':gananciaBank, 'totalventas':totalVentas['total__sum'],'totalTkPremiado':row[0][0]}\n \n return render (request, 'taquilla.html', contexto)\n\n@ajax\ndef ticketpre (request):\n with connection.cursor() as cursor:\n cursor.execute(\"SELECT id_ticket_id,id_animal_id,monto_apu FROM pagina_ticke_item inner join pagina_animalganador ON (pagina_ticke_item.id_animal_id = pagina_animalGanador.animal_id AND pagina_animalganador.fecha = CURDATE()) inner join pagina_ticket ON (pagina_ticket.id_ticket = pagina_ticke_item.id_ticket_id) inner join pagina_horas ON (pagina_horas.ticket_id = pagina_ticket.id_ticket AND pagina_horas.horas_id=pagina_animalganador.hora_id) WHERE pagina_ticket.fecha = CURDATE()\")\n row = cursor.fetchall()\n \n lista = list(({\"idTicket\": t[0],\"animalid\":t[1],\"valor\":t[2]}) for t in row)\n \n if len(row) > 0:\n return {'items':lista}\n else:\n print ('No hay Tickets Premiados')\n return {'itemsPre':(0)}\n\n\ndef pdftk (request):\n ticketnum = 331\n contenido = {'titulo': 'Animalitos',}\n serializeFecha = serializers.serialize(\"json\", Ticket.objects.filter(id_ticket=ticketnum), fields=('fecha', 'token','total','ida','idl'))\n fechat = json.loads(serializeFecha)\n ticketitems = Ticke_item.objects.filter(id_ticket=ticketnum).values()\n sorteos = Horas.objects.filter(ticket_id=ticketnum)\n response = HttpResponse(content_type='application/pdf')\n response['Content-Disposition'] = 'attachment; filename=hello.pdf'\n p = canvas.Canvas(response, pagesize=(5*cm,20*cm))\n p.setFont(\"Times-Roman\", 11)\n serializeSorteo = serializers.serialize(\"json\",Horas.objects.filter(ticket_id=ticketnum))\n \n p.drawString(0.5*cm ,19.5*cm, \".:Animalitos Loteria:.\")\n p.drawString(0.5*cm ,19.1*cm, \"==================\")\n p.drawString(0.5*cm ,18.8*cm, \"Fecha:\")\n p.drawString(0.5*cm ,18.5*cm, \"Sorteos:\")\n p.drawString(0.5*cm ,18.2*cm, \"__________________\")\n p.drawString(0.5*cm ,17.7*cm, \"Animal - - - - Apuesta\")\n for k in sorteos:\n h=0.5\n p.drawString(0.5*cm, 17.4*cm,\"\"+ str(k) +\",\" )\n h=h+0.5\n\n p.drawString(0.5*cm, 16*cm, \"Codigo:\")\n p.showPage()\n p.save()\n return response\n\n\nclass HelloPDFView(PDFTemplateView):\n \n template_name = 'weasy.html'\n base_url = 'file://' + settings.STATIC_ROOT\n download_filename = 'hello.pdf'\n \n def get_context_data(self, **kwargs):\n # code = request.POST.get('token')\n print('se ejecuta la vista PDF')\n token = self.args[0]\n sor = Horas.objects.filter(ticket_id__token=token)\n sorformated= []\n serializeFecha = serializers.serialize(\"json\", Ticket.objects.filter(token=token), fields=('fecha', 'token','total','ida','idl',))\n fechat = json.loads(serializeFecha)\n ticketitems = Ticke_item.objects.filter (id_ticket__token=token).values('id_animal__nombre', 'monto_apu')\n #Largo del PDF Ticket\n largotk=len(ticketitems)*0.4+6\n if len(sor) <= 3:\n largotk = largotk+0.3\n elif len(sor)<=6:\n largotk = largotk+0.6\n elif len(sor)<=9:\n largotk = largotk+0.9\n elif len(sor)<=12:\n largotk = largotk+1.2\n\n\n return super(HelloPDFView, self).get_context_data(\n pagesize='A4',\n title='Hi there!',\n sor = sor,\n numero = fechat[0]['pk'],\n fecha = fechat[0]['fields']['fecha'],\n token=fechat[0]['fields']['token'],\n items=ticketitems,\n largotk=largotk,\n **kwargs\n )", "sub_path": "pagina/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 10045, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 29, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 33, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 36, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponseRedirect", "line_number": 37, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 39, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "django.contrib.auth.forms.AuthenticationForm", "line_number": 45, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 47, "usage_type": "call"}, {"api_name": "pagina.models.Agencia.objects.filter", "line_number": 52, "usage_type": "call"}, {"api_name": "pagina.models.Agencia.objects", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pagina.models.Agencia", "line_number": 52, "usage_type": "name"}, {"api_name": "pagina.models.Loteria.objects.filter", "line_number": 53, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "pagina.models.Loteria", "line_number": 53, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 55, "usage_type": "call"}, {"api_name": "pagina.models.Animales.objects.filter", "line_number": 60, "usage_type": "call"}, {"api_name": "pagina.models.Animales.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "pagina.models.Animales", "line_number": 60, "usage_type": "name"}, {"api_name": "pagina.models.Loteria.objects.filter", "line_number": 62, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pagina.models.Loteria", "line_number": 62, "usage_type": "name"}, {"api_name": "pagina.models.Sorteo.objects.all", "line_number": 64, "usage_type": "call"}, {"api_name": "pagina.models.Sorteo.objects", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pagina.models.Sorteo", "line_number": 64, "usage_type": "name"}, {"api_name": "pagina.models.Loteria.objects.filter", "line_number": 65, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "pagina.models.Loteria", "line_number": 65, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 69, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 75, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects.get", "line_number": 81, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "pagina.models.Loteria", "line_number": 81, "usage_type": "name"}, {"api_name": "pagina.models.Agencia.objects.get", "line_number": 82, "usage_type": "call"}, {"api_name": "pagina.models.Agencia.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pagina.models.Agencia", "line_number": 82, "usage_type": "name"}, {"api_name": "secrets.token_hex", "line_number": 83, "usage_type": "call"}, {"api_name": "pagina.models.Ticket", "line_number": 85, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item", "line_number": 90, "usage_type": "call"}, {"api_name": "pagina.models.Animales.objects.get", "line_number": 90, "usage_type": "call"}, {"api_name": "pagina.models.Animales.objects", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pagina.models.Animales", "line_number": 90, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 95, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 95, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 96, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 96, "usage_type": "name"}, {"api_name": "pagina.models.Horas", "line_number": 97, "usage_type": "call"}, {"api_name": "pagina.models.Sorteo.objects.get", "line_number": 97, "usage_type": "call"}, {"api_name": "pagina.models.Sorteo.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "pagina.models.Sorteo", "line_number": 97, "usage_type": "name"}, {"api_name": "datetime.datetime.strftime", "line_number": 99, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 99, "usage_type": "name"}, {"api_name": "django_ajax.decorators.ajax", "line_number": 72, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 106, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 106, "usage_type": "name"}, {"api_name": "django.core.serializers.serialize", "line_number": 107, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 107, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 107, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 107, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 108, "usage_type": "call"}, {"api_name": "pagina.models.Agencia.objects.get", "line_number": 110, "usage_type": "call"}, {"api_name": "pagina.models.Agencia.objects", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pagina.models.Agencia", "line_number": 110, "usage_type": "name"}, {"api_name": "pagina.models.Loteria.objects.get", "line_number": 112, "usage_type": "call"}, {"api_name": "pagina.models.Loteria.objects", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pagina.models.Loteria", "line_number": 112, "usage_type": "name"}, {"api_name": "pagina.models.Horas.objects.filter", "line_number": 113, "usage_type": "call"}, {"api_name": "pagina.models.Horas.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pagina.models.Horas", "line_number": 113, "usage_type": "name"}, {"api_name": "pagina.models.Ticke_item.objects.filter", "line_number": 117, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticke_item", "line_number": 117, "usage_type": "name"}, {"api_name": "django_ajax.decorators.ajax", "line_number": 103, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 125, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 125, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 126, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 126, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 127, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 127, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 133, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 133, "usage_type": "name"}, {"api_name": "django.db.models.Sum", "line_number": 133, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects.filter", "line_number": 136, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects", "line_number": 136, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticke_item", "line_number": 136, "usage_type": "name"}, {"api_name": "pagina.models.AnimalGanador.objects.filter", "line_number": 137, "usage_type": "call"}, {"api_name": "pagina.models.AnimalGanador.objects", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pagina.models.AnimalGanador", "line_number": 137, "usage_type": "name"}, {"api_name": "django.db.connection.cursor", "line_number": 139, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 139, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 145, "usage_type": "call"}, {"api_name": "django.db.connection.cursor", "line_number": 149, "usage_type": "call"}, {"api_name": "django.db.connection", "line_number": 149, "usage_type": "name"}, {"api_name": "django_ajax.decorators.ajax", "line_number": 147, "usage_type": "name"}, {"api_name": "django.core.serializers.serialize", "line_number": 165, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 165, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 165, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 165, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 165, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 166, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects.filter", "line_number": 167, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects", "line_number": 167, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticke_item", "line_number": 167, "usage_type": "name"}, {"api_name": "pagina.models.Horas.objects.filter", "line_number": 168, "usage_type": "call"}, {"api_name": "pagina.models.Horas.objects", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pagina.models.Horas", "line_number": 168, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 169, "usage_type": "call"}, {"api_name": "reportlab.pdfgen.canvas.Canvas", "line_number": 171, "usage_type": "call"}, {"api_name": "reportlab.pdfgen.canvas", "line_number": 171, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 171, "usage_type": "name"}, {"api_name": "django.core.serializers.serialize", "line_number": 173, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 173, "usage_type": "name"}, {"api_name": "pagina.models.Horas.objects.filter", "line_number": 173, "usage_type": "call"}, {"api_name": "pagina.models.Horas.objects", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pagina.models.Horas", "line_number": 173, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 175, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 176, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 177, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 178, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 179, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 180, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 183, "usage_type": "name"}, {"api_name": "reportlab.lib.units.cm", "line_number": 186, "usage_type": "name"}, {"api_name": "easy_pdf.views.PDFTemplateView", "line_number": 192, "usage_type": "name"}, {"api_name": "django.conf.settings.STATIC_ROOT", "line_number": 195, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 195, "usage_type": "name"}, {"api_name": "pagina.models.Horas.objects.filter", "line_number": 202, "usage_type": "call"}, {"api_name": "pagina.models.Horas.objects", "line_number": 202, "usage_type": "attribute"}, {"api_name": "pagina.models.Horas", "line_number": 202, "usage_type": "name"}, {"api_name": "django.core.serializers.serialize", "line_number": 204, "usage_type": "call"}, {"api_name": "django.core.serializers", "line_number": 204, "usage_type": "name"}, {"api_name": "pagina.models.Ticket.objects.filter", "line_number": 204, "usage_type": "call"}, {"api_name": "pagina.models.Ticket.objects", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticket", "line_number": 204, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 205, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects.filter", "line_number": 206, "usage_type": "call"}, {"api_name": "pagina.models.Ticke_item.objects", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pagina.models.Ticke_item", "line_number": 206, "usage_type": "name"}]} +{"seq_id": "479809393", "text": "import sys\nimport logging as LOG\nfrom os import path\nfrom datetime import datetime, timedelta\n\nfrom airflow.models import Variable\nfrom airflow.decorators import dag, task\nfrom airflow.operators.python import get_current_context\nfrom airflow.providers.elasticsearch.hooks.elasticsearch import ElasticsearchHook\n\n# HACK: Fix for loading relative modules.\nsys.path.append(path.dirname(path.realpath(__file__)))\n\nfrom query import ESQueryPeers\nfrom postgres import PGDatabase\n\n# These are applied to all Operators via `kwargs[\"dag\"]`:\nARGS = {\n 'owner': 'jakubgs',\n 'depends_on_past': False,\n 'start_date': datetime(2021, 4, 13),\n 'email': ['jakub@status.im'],\n 'email_on_failure': False,\n 'email_on_retry': False,\n 'retries': 0,\n 'retry_delay': timedelta(minutes=10),\n}\n\n# These are passed to all Operators via `get_current_context()['params']`:\nPARAMS = {\n 'index_pattern': 'logstash-202*',\n 'field_name': 'peer_id',\n 'fleet_name': 'eth.prod',\n 'program': 'docker/statusd-whisper-node',\n}\n\n# ElasticSearch Logs Cluster\nesq = ESQueryPeers(conn_id='es_logs_cluster')\n# Citus PostgreSQL Database\npsg = PGDatabase(conn_id='citus_db_peers')\n\n\n@task\ndef query_indices(**kwargs):\n # This passes arguments given via Web UI when triggering a DAG.\n params = get_current_context()['params']\n\n days = psg.get_present_days()\n present_indices = [('logstash-%s' % d.replace('-', '.')) for d in days]\n\n LOG.info('Querying ES cluster for peers...')\n indices_to_query = []\n for index_name in esq.get_indices(params['index_pattern']):\n LOG.debug('Found Index: %s', index_name)\n\n # skip already injected indices\n if index_name in present_indices:\n LOG.debug('Skipping existing index: %s', index_name)\n continue\n # skip current day as it's incomplete\n if index_name == datetime.now().strftime('logstash-%Y.%m.%d'):\n LOG.warning('Skipping incomplete current day.')\n continue\n\n indices_to_query.append(index_name)\n\n return list(indices_to_query)\n\n@task\ndef query_peers(indices: list):\n # This passes arguments given via Web UI when triggering a DAG.\n params = get_current_context()['params']\n\n peers = []\n for index_name in indices:\n rval = esq.get_peers(\n index=index_name,\n field=params['field_name'],\n fleet=params['fleet_name'],\n program=params['program'],\n )\n if len(rval) == 0:\n LOG.warning('%s - No entries found!', index_name)\n continue\n\n LOG.info('%s - Found: %s', index_name, len(rval))\n peers.extend(rval)\n\n return peers\n\n@task\ndef inject_peers(peers: list):\n if len(peers) == 0:\n LOG.warning('Nothing to insert into database.')\n return\n\n LOG.info('Injecting peers data into database...')\n psg.inject_peers(peers)\n\n\n# Main definition of the DAG, needs to be global.\n@dag('es_export_peers', schedule_interval='@daily', default_args=ARGS, params=PARAMS)\ndef es_export_peers():\n inject_peers(query_peers(query_indices()))\n\ndag = es_export_peers()\n", "sub_path": "es_export_peers/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3119, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 21, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 26, "usage_type": "call"}, {"api_name": "query.ESQueryPeers", "line_number": 38, "usage_type": "call"}, {"api_name": "postgres.PGDatabase", "line_number": 40, "usage_type": "call"}, {"api_name": "airflow.operators.python.get_current_context", "line_number": 46, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 51, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 58, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "logging.warning", "line_number": 62, "usage_type": "call"}, {"api_name": "airflow.decorators.task", "line_number": 43, "usage_type": "name"}, {"api_name": "airflow.operators.python.get_current_context", "line_number": 72, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 83, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 86, "usage_type": "call"}, {"api_name": "airflow.decorators.task", "line_number": 69, "usage_type": "name"}, {"api_name": "logging.warning", "line_number": 94, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 97, "usage_type": "call"}, {"api_name": "airflow.decorators.task", "line_number": 91, "usage_type": "name"}, {"api_name": "airflow.decorators.dag", "line_number": 102, "usage_type": "call"}, {"api_name": "airflow.decorators.dag", "line_number": 106, "usage_type": "name"}]} +{"seq_id": "498359875", "text": "# -*- coding: utf-8 -*-\nimport numpy as np\nimport img_tool as imt\nimport matplotlib.pyplot as plt\nimport csv\n\nhdr = list(csv.reader(open('hist_comp.csv')))[0]\nary = np.loadtxt('hist_comp.csv', delimiter=',', skiprows=1)\n\n# hdr[0]はヒストグラムのbin。hdr[1:]がデータのラベル\nx = ary[:, 0]\ni = 1\nfor name in hdr[1:]:\n y = ary[:, i]\n plt.plot(x, y, label=name)\n i = i + 1\n\nplt.xlim(0, 255)\nplt.ylim(0,)\nplt.legend()\nplt.savefig('hist_comp.bmp')\nplt.show()\n", "sub_path": "test_function.py", "file_name": "test_function.py", "file_ext": "py", "file_size_in_byte": 479, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "csv.reader", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}]} +{"seq_id": "241356975", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('users', '0011_auto_20150609_0350'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Log',\n fields=[\n ('id', models.AutoField(serialize=False, auto_created=True, primary_key=True, verbose_name='ID')),\n ('key', models.CharField(max_length=255)),\n ('type', models.IntegerField(\n choices=[(0, 'Другое'), (1, 'Справочник: добавлена пробирка'), (2, 'Справочник: изменена пробирка'),\n (3, 'Справочник: добавлен анализ'), (4, 'Справочник: изменен анализ'),\n (5, 'Справочник: добавлена группа направления'),\n (6, 'Справочник: изменена группа направления'), (7, 'Направления: создано направление'),\n (8, 'Взятие материала: открыто направление'), (9, 'Взятие материала: пробирка взята'),\n (10, 'Взятие материала: напечатан акт приема-передачи'),\n (11, 'Прием материала: материал принят'), (12, 'Прием материала: материал не принят'),\n (13, 'Ввод результатов: результат сохранен'),\n (14, 'Ввод результатов: результат подтвержден'),\n (15, 'Ввод результатов: результаты для направления напечатаны'),\n (16, 'Администрирование: создан пользователь'),\n (17, 'Администрирование: создано подразделение'),\n (18, 'Пользователи: вход пользователя')])),\n ('body', models.CharField(max_length=1023)),\n ('time', models.DateTimeField(auto_now=True)),\n ('user', models.ForeignKey(to='users.DoctorProfile')),\n ],\n ),\n ]\n", "sub_path": "slog/migrations/0001_initial.py", "file_name": "0001_initial.py", "file_ext": "py", "file_size_in_byte": 2557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 13, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 13, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 16, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 16, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 32, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 32, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 34, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "152695669", "text": "import cv2\nimport numpy as np\nimport os\nimport argparse\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"file\",help=\"file with paths\")\n\nargs = parser.parse_args()\n\nf = open(args.file,\"r\")\n\nscenes = f.readlines()\n\nfor i,scene in enumerate(scenes):\n scene = scene.strip()\n img = np.zeros((256,512,3),np.uint8)\n if (scene.find(\"input\") != -1):\n# base = \"/vulcan/scratch/jushen/train_generated/\" + scene[0:scene.index(\"input\")]\n base = \"train_data_3/\"+scene[0:scene.index(\"input\")]\n print(base)\n tr = str(base+\"truth.png\")\n #img[:,0:256,:] = cv2.imread(tr)\n img[:,0:256,:] = cv2.imread(tr)\n img[:,256:512,:] = cv2.imread(str(base + \"input.png\"))\n cv2.imwrite(base+\"combined_2.png\",img)\n print(base+\"combined.png\")\n\n#img = np.zeros((256,512,3),np.uint8)\n#img[:,0:256,:] = cv2.imread('book_l0_yax_315deg_l1_yax_180deg_down.pbrt.png')\n#img[:,256:512,:] = cv2.imread('book_l0_yax_315deg_l1_yax_180deg_truth.pbrt.png')\n#cv2.imwrite('combined.png',img)\n#2B_C2_4L_C6_truth.png\n", "sub_path": "tools/combine.py", "file_name": "combine.py", "file_ext": "py", "file_size_in_byte": 1047, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 25, "usage_type": "call"}, {"api_name": "cv2.imwrite", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "335398478", "text": "from django.shortcuts import render, redirect\nfrom oauth2client.client import flow_from_clientsecrets\nfrom oauth2client import client\nfrom apiclient.discovery import build\nimport httplib2\nfrom django.http import HttpResponse\nfrom .models import Chart, Metric\nimport json\nfrom .forms import AddChart\n\nflow = flow_from_clientsecrets('client_secret.json',\n scope='https://www.googleapis.com/auth/analytics.readonly',\n redirect_uri='http://gaservice.pw/reg')\n\n\nclass Site():\n def __init__(self, view_id, url):\n self.view_id = view_id\n self.url = url\n\n\ndef front(request):\n return render(request, 'WebService/front.html', {'user': request.session.get('user_id')})\n\n\ndef reg(request):\n if not request.GET.get('code'):\n auth_uri = flow.step1_get_authorize_url()\n return redirect(auth_uri)\n else:\n auth_code = request.GET.get('code')\n credentials = flow.step2_exchange(auth_code)\n service = build('analytics', 'v3', http=credentials.authorize(httplib2.Http()))\n accounts = service.management().accounts().list().execute()\n item = accounts['items'][0]\n request.session['user_id'] = int(item['id'])\n request.session['credentials'] = credentials.to_json()\n\n if not request.GET.get('red'):\n return redirect('/')\n else:\n return redirect(request.GET.get('red'))\n\n\ndef sign_out(request):\n request.session.clear() # очистить сессию\n return render(request, 'WebService/out.html')\n\n\ndef analitic(request):\n user_id = request.session.get('user_id')\n sites = []\n if not user_id:\n return redirect('/reg?red=/analitic')\n try:\n credentials = client.OAuth2Credentials.from_json(request.session['credentials'])\n service = build('analytics', 'v3', http=credentials.authorize(httplib2.Http()))\n properties = service.management().webproperties().list(accountId=str(user_id)).execute()\n if properties.get('items'):\n for item in properties.get('items'):\n # Get the first property id.\n property = item.get('id')\n\n # Get a list of all views (profiles) for the first property.\n profiles = service.management().profiles().list(accountId=user_id,webPropertyId=property).execute()\n\n if profiles.get('items'):\n sites.append(Site(profiles.get('items')[0].get('id'), profiles.get('items')[0].get('websiteUrl')))\n except:\n return redirect('/reg?red=/analitic')\n view_id = request.GET.get('viewId')\n url = None\n for site in sites:\n if site.view_id == view_id:\n url = site.url\n break\n if view_id and url:\n charts = Chart.objects.filter(viewId=int(view_id))\n if charts:\n return render(request, 'WebService/analitic.html', {'user': user_id,\n 'sites': sites,\n 'charts': charts,\n 'url': url,\n 'view_id': view_id})\n else:\n return redirect('/chart?num=0&viewId='+view_id)\n\n return render(request, 'WebService/analitic.html', {'user': user_id,\n 'sites': sites,\n 'notadd': True})\n\n\ndef chart(request):\n sites = []\n user_id = request.session.get('user_id')\n if not user_id:\n return redirect('/reg?red=/analitic')\n if request.method == \"POST\":\n chart_num = int(request.POST.get('num'))\n view_id = int(request.POST.get('viewId'))\n loc_chart = Chart.objects.get(viewId=view_id, numb=chart_num)\n loc_chart.metric = Metric.objects.get(value=request.POST.get('metric'))\n loc_chart.startDate = request.POST.get('startDate')\n loc_chart.endDate = request.POST.get('endDate')\n loc_chart.max_count = request.POST.get('max_count')\n loc_chart.width = request.POST.get('width')\n loc_chart.height = request.POST.get('height')\n loc_chart.save()\n return redirect('/chart?viewId='+str(view_id)+'&num='+str(chart_num))\n chart_num = request.GET.get('num')\n view_id = request.GET.get('viewId')\n if not view_id:\n return redirect('/analitic')\n if not chart_num:\n return redirect('/analitic')\n elif chart_num == '0':\n loc_chart = Chart()\n loc_chart.metric = Metric.objects.get(value='browser')\n loc_chart.startDate = '2017-05-17'\n loc_chart.endDate = '2017-05-24'\n loc_chart.viewId = int(view_id)\n charts = list(Chart.objects.filter(viewId=int(view_id)))\n chart_num = len(charts) + 1\n loc_chart.numb = chart_num\n loc_chart.save()\n loc_chart = Chart.objects.get(viewId=int(view_id), numb=chart_num)\n\n try:\n credentials = client.OAuth2Credentials.from_json(request.session['credentials'])\n service = build('analytics', 'v3', http=credentials.authorize(httplib2.Http()))\n properties = service.management().webproperties().list(accountId=str(user_id)).execute()\n if properties.get('items'):\n for item in properties.get('items'):\n # Get the first property id.\n property = item.get('id')\n\n # Get a list of all views (profiles) for the first property.\n profiles = service.management().profiles().list(accountId=user_id,webPropertyId=property).execute()\n\n if profiles.get('items'):\n sites.append(Site(profiles.get('items')[0].get('id'), profiles.get('items')[0].get('websiteUrl')))\n except:\n return redirect('/reg?red=/analitic')\n metrics = list(Metric.objects.all())\n metrics.remove(loc_chart.metric)\n return render(request, 'WebService/chart.html', {'user': user_id,\n 'sites': sites,\n 'metrics': metrics,\n 'loc_chart': loc_chart})\n\ndef ajax_json(request):\n numb = request.GET.get('numb')\n json_data = []\n if request.session.get('user_id') and request.GET.get('viewId'):\n view_id = request.GET.get('viewId')\n credentials = client.OAuth2Credentials.from_json(request.session['credentials'])\n if numb:\n charts = [Chart.objects.get(viewId=int(view_id), numb=numb)]\n else:\n charts = Chart.objects.filter(viewId=int(view_id))\n for chart in charts:\n data = build('analytics', 'v3', http=credentials.authorize(httplib2.Http()),\n discoveryServiceUrl=('https://analyticsreporting.googleapis.com/$discovery/rest')).reports()\n body = {\n 'reportRequests': []\n }\n body['reportRequests'].append({\n 'viewId': str(view_id),\n 'dateRanges': [{'startDate': chart.startDate, 'endDate': chart.endDate}],\n 'metrics': [{'expression': 'ga:sessions'}],\n 'dimensions': [{\"name\": 'ga:'+chart.metric.value}],\n 'orderBys': [{\"fieldName\": \"ga:sessions\", \"sortOrder\": \"DESCENDING\"}],\n 'pageSize': str(chart.max_count)\n })\n report = data.batchGet(body=body).execute().get('reports')\n api_data = report[0].get('data', {}).get('rows', [])\n labels = []\n data = []\n backgroundColor = []\n borderColor = []\n for point in api_data:\n labels.append(point[\"dimensions\"][0])\n metric = point[\"metrics\"][0]\n data.append(metric[\"values\"][0])\n backgroundColor.append('rgba(54, 162, 235, 0.2)')\n borderColor.append('rgba(54, 162, 235, 1)')\n json_data.append({\n 'height': chart.height,\n 'width': chart.width,\n 'data':\n {\n\n 'type': 'bar',\n 'data': {\n 'labels': labels,\n 'datasets': [\n {\n 'label': chart.metric.value,\n 'data': data,\n 'backgroundColor': backgroundColor,\n 'borderColor': borderColor,\n 'borderWidth': 1\n }\n ]\n },\n 'options': {\n 'scales': {\n 'yAxes': [\n {\n 'ticks': {\n 'beginAtZero': True\n }\n }\n ]\n }\n }\n }\n })\n return HttpResponse(json.dumps(json_data))\n else:\n return 0\n", "sub_path": "WebService/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 9009, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "oauth2client.client.flow_from_clientsecrets", "line_number": 11, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 23, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 29, "usage_type": "call"}, {"api_name": "apiclient.discovery.build", "line_number": 33, "usage_type": "call"}, {"api_name": "httplib2.Http", "line_number": 33, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 47, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 54, "usage_type": "call"}, {"api_name": "oauth2client.client.OAuth2Credentials.from_json", "line_number": 56, "usage_type": "call"}, {"api_name": "oauth2client.client.OAuth2Credentials", "line_number": 56, "usage_type": "attribute"}, {"api_name": "oauth2client.client", "line_number": 56, "usage_type": "name"}, {"api_name": "apiclient.discovery.build", "line_number": 57, "usage_type": "call"}, {"api_name": "httplib2.Http", "line_number": 57, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "models.Chart.objects.filter", "line_number": 78, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 78, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 78, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 80, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 86, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 88, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 97, "usage_type": "call"}, {"api_name": "models.Chart.objects.get", "line_number": 101, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 101, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 101, "usage_type": "name"}, {"api_name": "models.Metric.objects.get", "line_number": 102, "usage_type": "call"}, {"api_name": "models.Metric.objects", "line_number": 102, "usage_type": "attribute"}, {"api_name": "models.Metric", "line_number": 102, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 109, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 113, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 115, "usage_type": "call"}, {"api_name": "models.Chart", "line_number": 117, "usage_type": "call"}, {"api_name": "models.Metric.objects.get", "line_number": 118, "usage_type": "call"}, {"api_name": "models.Metric.objects", "line_number": 118, "usage_type": "attribute"}, {"api_name": "models.Metric", "line_number": 118, "usage_type": "name"}, {"api_name": "models.Chart.objects.filter", "line_number": 122, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 122, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 122, "usage_type": "name"}, {"api_name": "models.Chart.objects.get", "line_number": 126, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 126, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 126, "usage_type": "name"}, {"api_name": "oauth2client.client.OAuth2Credentials.from_json", "line_number": 129, "usage_type": "call"}, {"api_name": "oauth2client.client.OAuth2Credentials", "line_number": 129, "usage_type": "attribute"}, {"api_name": "oauth2client.client", "line_number": 129, "usage_type": "name"}, {"api_name": "apiclient.discovery.build", "line_number": 130, "usage_type": "call"}, {"api_name": "httplib2.Http", "line_number": 130, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 143, "usage_type": "call"}, {"api_name": "models.Metric.objects.all", "line_number": 144, "usage_type": "call"}, {"api_name": "models.Metric.objects", "line_number": 144, "usage_type": "attribute"}, {"api_name": "models.Metric", "line_number": 144, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 146, "usage_type": "call"}, {"api_name": "oauth2client.client.OAuth2Credentials.from_json", "line_number": 156, "usage_type": "call"}, {"api_name": "oauth2client.client.OAuth2Credentials", "line_number": 156, "usage_type": "attribute"}, {"api_name": "oauth2client.client", "line_number": 156, "usage_type": "name"}, {"api_name": "models.Chart.objects.get", "line_number": 158, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 158, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 158, "usage_type": "name"}, {"api_name": "models.Chart.objects.filter", "line_number": 160, "usage_type": "call"}, {"api_name": "models.Chart.objects", "line_number": 160, "usage_type": "attribute"}, {"api_name": "models.Chart", "line_number": 160, "usage_type": "name"}, {"api_name": "apiclient.discovery.build", "line_number": 162, "usage_type": "call"}, {"api_name": "httplib2.Http", "line_number": 162, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 219, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 219, "usage_type": "call"}]} +{"seq_id": "129139241", "text": "from velociraptor.observations.objects import ObservationalData\n\nimport unyt\nimport numpy as np\n\nimport os\nimport sys\n\n# Exec the master cosmology file passed as first argument\nwith open(sys.argv[1], \"r\") as handle:\n exec(handle.read())\n\n# Cosmology\nh_sim = cosmology.h\n\ninput_filename = \"../raw/Graham2023.txt\"\ndelimiter = \" \"\n\noutput_filenames = [\"Graham2023_S.hdf5\", \"Graham2023_ESS0.hdf5\", \"Graham2023_E.hdf5\"]\ngalaxy_types = [\"S\", \"ES/S0\", \"E\"]\noutput_directory = \"../\"\n\n# conversion for Mstar from_Kroupa (2002) to Chabrier (2003) IMF\n# (table 2, Bernardi et al, 2010, 2010MNRAS.404.2087B)\nlog_M_offset = 0.05\n\nlog_M_bh, log_M_bh_err, log_M_star, log_M_star_err, Gal_type = [], [], [], [], []\nwith open(input_filename, \"r\") as file:\n rows = file.readlines()\n for row in rows:\n try:\n elements = row.split(\" \")\n gal_type, bh_mass_and_err, stellar_mass_and_err = (\n elements[1],\n elements[6],\n elements[10],\n )\n bh_mass, bh_mass_err = (float(x) for x in bh_mass_and_err.split(\"±\"))\n stellar_mass, stellar_mass_err = (\n float(x) for x in stellar_mass_and_err.split(\"±\")\n )\n\n log_M_bh.append(bh_mass)\n log_M_bh_err.append(bh_mass_err)\n log_M_star.append(stellar_mass + log_M_offset)\n log_M_star_err.append(stellar_mass_err)\n Gal_type.append(gal_type)\n except ValueError:\n pass\n\nGal_type = np.array(Gal_type)\nlog_M_bh, log_M_star = np.array(log_M_bh), np.array(log_M_star)\nlog_M_bh_err, log_M_star_err = np.array(log_M_bh_err), np.array(log_M_star_err)\n\nM_bh = unyt.unyt_array(np.power(10.0, log_M_bh), units=\"Msun\")\nM_star = unyt.unyt_array(np.power(10.0, log_M_star), units=\"Msun\")\n\nM_bh_lower = np.power(10.0, log_M_bh) - np.power(10.0, log_M_bh - log_M_bh_err)\nM_bh_upper = np.power(10.0, log_M_bh + log_M_bh_err) - np.power(10.0, log_M_bh)\n\nM_star_lower = np.power(10.0, log_M_star) - np.power(10.0, log_M_star - log_M_star_err)\nM_star_upper = np.power(10.0, log_M_star + log_M_star_err) - np.power(10.0, log_M_star)\n\nif not os.path.exists(output_directory):\n os.mkdir(output_directory)\n\nfor galaxy_type, output_filename in zip(galaxy_types, output_filenames):\n\n mask = Gal_type == galaxy_type\n\n # Meta-data\n comment = (\n f\"A (black hole mass)-(galaxy stellar mass) relation based on colour-dependent stellar mass-to-light ratios \"\n f\" for {galaxy_type} galaxies. Converted from the Kroupa (2002) to Chabrier (2003) IMF.\"\n f\" The whole sample consists of 73 ETGs plus 31 LTGs, coming from the larger sample of 84 ETGs \"\n f\" (Sahu et al. 2019) and 43 LTGs (Davis et al. 2019).\"\n )\n citation = f\"Graham & Sahu (2023) ({galaxy_type})\"\n bibcode = \"2023MNRAS.518.2177G\"\n name = f\"Black hole mass - stellar mass relation ({galaxy_type} galaxies)\"\n plot_as = \"points\"\n # We purposely make this data show up not only a z=0 but also at higher z\n redshift_lower, redshift_upper = -0.1, 3.1\n redshift = 0.0\n h = h_sim\n\n M_bh_scatter = unyt.unyt_array([M_bh_lower[mask], M_bh_upper[mask]], units=\"Msun\")\n M_star_scatter = unyt.unyt_array(\n [M_star_lower[mask], M_star_upper[mask]], units=\"Msun\"\n )\n\n # Write everything\n processed = ObservationalData()\n processed.associate_x(\n M_star[mask],\n scatter=M_star_scatter,\n comoving=True,\n description=\"Galaxy Stellar Mass\",\n )\n processed.associate_y(\n M_bh[mask],\n scatter=M_bh_scatter,\n comoving=True,\n description=\"Black Hole Mass\",\n )\n processed.associate_citation(citation, bibcode)\n processed.associate_name(name)\n processed.associate_comment(comment)\n processed.associate_redshift(redshift, redshift_lower, redshift_upper)\n processed.associate_plot_as(plot_as)\n processed.associate_cosmology(cosmology)\n\n output_path = f\"{output_directory}/{output_filename}\"\n\n if os.path.exists(output_path):\n os.remove(output_path)\n\n processed.write(filename=output_path)\n", "sub_path": "data/GalaxyStellarMassBlackHoleMass/conversion/convertGraham2023.py", "file_name": "convertGraham2023.py", "file_ext": "py", "file_size_in_byte": 4108, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.argv", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "unyt.unyt_array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 55, "usage_type": "call"}, {"api_name": "unyt.unyt_array", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 56, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 65, "usage_type": "call"}, {"api_name": "unyt.unyt_array", "line_number": 87, "usage_type": "call"}, {"api_name": "unyt.unyt_array", "line_number": 88, "usage_type": "call"}, {"api_name": "velociraptor.observations.objects.ObservationalData", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 115, "usage_type": "call"}, {"api_name": "os.path", "line_number": 115, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 116, "usage_type": "call"}]} +{"seq_id": "391726788", "text": "import pandas as pd\nimport numpy as np\nimport pgeocode\nfrom pypostalcode import PostalCodeDatabase\nimport folium # map rendering library\nfrom geopy.geocoders import Nominatim # convert an address into latitude and longitude values\nimport requests # library to handle requests\nfrom pandas.io.json import json_normalize # tranform JSON file into a pandas dataframe\nfrom sklearn.cluster import KMeans # import k-means from clustering stage\nimport matplotlib.cm as cm # Matplotlib and associated plotting modules\nimport matplotlib.colors as colors\nimport os\nimport webbrowser\n\n\n# EXERCISE PART 1: Creating the dataframe and transforming the data\n# -----------------------------------------------------------------\n\nd = pd.read_html(\"http://www.geonames.org/postalcode-search.html?q=&country=BE\")\ndf = d[2]\ndf.columns = ['SequenceNr', 'City', 'PostalCode','Country', 'Region', 'Province', 'MajorCity']\n\n# Drop rows where Borough is \"Not assgined\"\ndf = df.replace('Not assigned', np.nan)\ndf = df.dropna(subset=['SequenceNr'])\ndf = df.drop(columns='SequenceNr')\ndf = df.drop(columns='Country')\ndf = df.drop(columns= 'Region')\n\ndf_antwerp = df[df.MajorCity == \"Antwerpen\"]\n\n# EXERCISE PART 2: Adding latitude & longitude to the dataframe\n# -------------------------------------------------------------\n\nnomi = pgeocode.Nominatim('be')\n\n# Function to search for latitude based on postal code\ndef searchlatitude(x):\n try:\n t_postalcodeinfo = nomi.query_postal_code(x)\n latitude = t_postalcodeinfo[-3]\n return latitude\n except:\n return \"Not found\"\n\n# Function to search for longitude based on postal code\ndef searchlongitude(x):\n try:\n t_postalcodeinfo = nomi.query_postal_code(x)\n longitude = t_postalcodeinfo[-2]\n return longitude\n except:\n return \"Not found\"\n\n\n# Add columns Latitude and Longitude\ndf_antwerp['Latitude'] = df_antwerp.apply(lambda row: searchlatitude(row.PostalCode), axis = 1)\ndf_antwerp['Longitude'] = df_antwerp.apply(lambda row: searchlongitude(row.PostalCode), axis = 1)\n# print(df_antwerp)\n\n# Drop the rows for which the postal code was not found\ndf_antwerp = df_antwerp.replace('Not found', np.nan)\ndf_antwerp = df_antwerp.dropna(subset=['Latitude'])\n\n\n# EXERCISE PART 3: Exploring & clustering the neighborhoods of Toronto\n# --------------------------------------------------------------------\n# Get location of Antwerp\naddress = 'Antwerp'\ngeolocator = Nominatim(user_agent=\"antwerp_explorer\")\nlocation = geolocator.geocode(address)\nlatitude = location.latitude\nlongitude = location.longitude\nprint('The geograpical coordinate of Antwerp are {}, {}.'.format(latitude, longitude))\n\n# Create map of New York using latitude and longitude values\nmap_antwerp = folium.Map(location=[latitude, longitude], zoom_start=10)\n\n# Add markers to map\nfor lat, lng, postalcode, city in zip(df_antwerp['Latitude'], df_antwerp['Longitude'], df_antwerp['PostalCode'], df_antwerp['City']):\n label = '{}, {}'.format(postalcode, city)\n label = folium.Popup(label, parse_html=True)\n folium.CircleMarker(\n [lat, lng],\n radius=5,\n popup=label,\n color='blue',\n fill=True,\n fill_color='#3186cc',\n fill_opacity=0.7,\n parse_html=False).add_to(map_antwerp)\n\n# Show map (in Jupyter Notebook)\nmap_antwerp\n\n\n# Foursquare credentials\nCLIENT_ID = 'JZNEUC4UMXDUSRH140GO1MW1BXMSJXC14DLPZYWVDR5UJ5P1' # Foursquare ID\nCLIENT_SECRET = 'QDNPZM1Q0KPYYQTP21HWJSHXPRGOG4412PDTDYFYXNEJ3BTR' # Foursquare Secret\nVERSION = '20180605' # Foursquare API version\n\n# Get latitude and longitude for first neighborhood\nneighborhood_latitude = df_antwerp['Latitude'].iloc[0]\nneighborhood_longitude = df_antwerp['Longitude'].iloc[0]\nneighborhood_name = df_antwerp['City'].iloc[0]\n\n# Call the Foursquare API\nLIMIT = 100 # limit of number of venues returned by Foursquare API\nradius = 500 # define radius\n#\nurl = 'https://api.foursquare.com/v2/venues/explore?&client_id={}&client_secret={}&v={}&ll={},{}&radius={}&limit={}'.format(\n CLIENT_ID,\n CLIENT_SECRET,\n VERSION,\n neighborhood_latitude,\n neighborhood_longitude,\n radius,\n LIMIT)\n\nresults = requests.get(url).json()\n\n\n# Function that extracts the category of the venue\ndef get_category_type(row):\n try:\n categories_list = row['categories']\n except:\n categories_list = row['venue.categories']\n if len(categories_list) == 0:\n return None\n else:\n return categories_list[0]['name']\n\n# Clean the json and structure it into a pandas dataframe\nvenues = results['response']['groups'][0]['items']\nnearby_venues = json_normalize(venues) # flatten JSON\nfiltered_columns = ['venue.name', 'venue.categories', 'venue.location.lat', 'venue.location.lng'] # filter columns\nnearby_venues = nearby_venues.loc[:, filtered_columns]\nnearby_venues['venue.categories'] = nearby_venues.apply(get_category_type, axis=1) # filter the category for each row\nnearby_venues.columns = [col.split(\".\")[-1] for col in nearby_venues.columns] # clean columns\nprint(nearby_venues.head())\nprint('{} venues were returned by Foursquare.'.format(nearby_venues.shape[0]))\n\n# Function to repeat the same process to all the neighborhoods\ndef getNearbyVenues(names, latitudes, longitudes, radius=500):\n venues_list = []\n for name, lat, lng in zip(names, latitudes, longitudes):\n # create the API request URL\n url = 'https://api.foursquare.com/v2/venues/explore?&client_id={}&client_secret={}&v={}&ll={},{}&radius={}&limit={}'.format(\n CLIENT_ID,\n CLIENT_SECRET,\n VERSION,\n lat,\n lng,\n radius,\n LIMIT)\n\n # make the GET request\n results = requests.get(url).json()[\"response\"]['groups'][0]['items']\n\n # return only relevant information for each nearby venue\n venues_list.append([(\n name,\n lat,\n lng,\n v['venue']['name'],\n v['venue']['location']['lat'],\n v['venue']['location']['lng'],\n v['venue']['categories'][0]['name']) for v in results])\n\n nearby_venues = pd.DataFrame([item for venue_list in venues_list for item in venue_list])\n nearby_venues.columns = ['Neighborhood',\n 'Neighborhood Latitude',\n 'Neighborhood Longitude',\n 'Venue',\n 'Venue Latitude',\n 'Venue Longitude',\n 'Venue Category']\n\n return (nearby_venues)\n\n# Apply the function to the neighborhoods of Toronto\nantwerp_venues = getNearbyVenues(names=df_antwerp['City'],\n latitudes=df_antwerp['Latitude'],\n longitudes=df_antwerp['Longitude']\n )\n\n# One hot encoding\nantwerp_onehot = pd.get_dummies(antwerp_venues[['Venue Category']], prefix=\"\", prefix_sep=\"\")\nantwerp_onehot['Neighborhood'] = antwerp_venues['Neighborhood'] # Add neighborhood column back to dataframe\nfixed_columns = [antwerp_onehot.columns[-1]] + list(antwerp_onehot.columns[:-1]) # move neighborhood column to the first column\nantwerp_onehot = antwerp_onehot[fixed_columns]\n\n# Group by Neighborhood\nantwerp_grouped = antwerp_onehot.groupby('Neighborhood').mean().reset_index()\n\n# Function to sort venues in descending order\ndef return_most_common_venues(row, num_top_venues):\n row_categories = row.iloc[1:]\n row_categories_sorted = row_categories.sort_values(ascending=False)\n\n return row_categories_sorted.index.values[0:num_top_venues]\n\n# Create new dataframe and display the top 10 venues for each neighborhood\nnum_top_venues = 10\nindicators = ['st', 'nd', 'rd']\ncolumns = ['Neighborhood'] # Create columns according to number of top venues\nfor ind in np.arange(num_top_venues):\n try:\n columns.append('{}{} Most Common Venue'.format(ind+1, indicators[ind]))\n except:\n columns.append('{}th Most Common Venue'.format(ind+1))\nneighborhoods_venues_sorted = pd.DataFrame(columns=columns) # Create a new dataframe\nneighborhoods_venues_sorted['Neighborhood'] = antwerp_grouped['Neighborhood']\nfor ind in np.arange(antwerp_grouped.shape[0]):\n neighborhoods_venues_sorted.iloc[ind, 1:] = return_most_common_venues(antwerp_grouped.iloc[ind, :], num_top_venues)\n\n# k-means cluster\nkclusters = 5 # set number of clusters\nantwerp_grouped_clustering = antwerp_grouped.drop('Neighborhood', 1)\nkmeans = KMeans(n_clusters=kclusters, random_state=0).fit(antwerp_grouped_clustering) # run k-means clustering\nkmeans.labels_[0:10] # Check cluster labels generated for each row in the dataframe\n\n\n# Create a new dataframe that includes the cluster as well as the top 10 venues for each neighborhood\nneighborhoods_venues_sorted.insert(0, 'Cluster Labels', kmeans.labels_) # Add clustering labels\n\nantwerp_merged = df_antwerp\nantwerp_merged.columns = ['Neighborhood', 'PostalCode', 'Province', 'MajorCity', 'Latitude', 'Longitude']\n\nantwerp_merged = antwerp_merged.join(neighborhoods_venues_sorted.set_index('Neighborhood'), on='Neighborhood')\n\n# Visualise the resulting clusters\nmap_clusters = folium.Map(location=[latitude, longitude], zoom_start=11) # create map\nx = np.arange(kclusters) # set color scheme for the clusters\nys = [i + x + (i * x) ** 2 for i in range(kclusters)]\ncolors_array = cm.rainbow(np.linspace(0, 1, len(ys)))\nrainbow = [colors.rgb2hex(i) for i in colors_array]\nmarkers_colors = [] # add markers to the map\nfor lat, lon, poi, cluster in zip(antwerp_merged['Latitude'], antwerp_merged['Longitude'],\n antwerp_merged['Neighborhood'], antwerp_merged['Cluster Labels']):\n label = folium.Popup(str(poi) + ' Cluster ' + str(cluster), parse_html=True)\n folium.CircleMarker(\n [lat, lon],\n radius=5,\n popup=label,\n color=rainbow[cluster - 1],\n fill=True,\n fill_color=rainbow[cluster - 1],\n fill_opacity=0.7).add_to(map_clusters)\n\nmap_clusters\n\nfilepath = r'''C:\\Users\\rc01828\\PycharmProjects\\map.html'''\nmap_clusters.save(filepath)\nwebbrowser.open('file://' + filepath)\niframe = map_clusters._repr_html_()\n\n", "sub_path": "Capstoneproject_Antwerp.py", "file_name": "Capstoneproject_Antwerp.py", "file_ext": "py", "file_size_in_byte": 10223, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_html", "line_number": 19, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pgeocode.Nominatim", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.nan", "line_number": 62, "usage_type": "attribute"}, {"api_name": "geopy.geocoders.Nominatim", "line_number": 70, "usage_type": "call"}, {"api_name": "folium.Map", "line_number": 77, "usage_type": "call"}, {"api_name": "folium.Popup", "line_number": 82, "usage_type": "call"}, {"api_name": "folium.CircleMarker", "line_number": 83, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 120, "usage_type": "call"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 136, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 159, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 171, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 208, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 213, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 215, "usage_type": "call"}, {"api_name": "sklearn.cluster.KMeans", "line_number": 221, "usage_type": "call"}, {"api_name": "folium.Map", "line_number": 234, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 235, "usage_type": "call"}, {"api_name": "matplotlib.cm.rainbow", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.cm", "line_number": 237, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 237, "usage_type": "call"}, {"api_name": "matplotlib.colors.rgb2hex", "line_number": 238, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 238, "usage_type": "name"}, {"api_name": "folium.Popup", "line_number": 242, "usage_type": "call"}, {"api_name": "folium.CircleMarker", "line_number": 243, "usage_type": "call"}, {"api_name": "webbrowser.open", "line_number": 256, "usage_type": "call"}]} +{"seq_id": "407947920", "text": "import Cust_Functions as F\nimport autorization as aut\n# import comtypes.client\nfrom docxtpl import DocxTemplate\nimport os\n\n\ndef set_rabotn(self):\n if self.windowTitle() == \"Расчет КПЭ\":\n return\n if self.ui.cmb_rabotn.currentText() == \"\":\n return\n dolgn = ' '.join(self.ui.cmb_rabotn.currentText().split(' ')[3:])\n name = self.ui.l_period.text()\n if F.nalich_file(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name + F.sep() + name + \"$\"\n + self.ui.cmb_rabotn.currentText() + '.pickle'):\n spis = F.otkr_f(\n F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name + F.sep() + name + \"$\" +\n self.ui.cmb_rabotn.currentText() + '.pickle',\n pickl=True)\n else:\n if not F.nalich_file(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + dolgn + '.pickle'):\n self.showdialog(f'Не найден шаблон {dolgn}')\n self.ui.tbl_kpi_sotr.clear()\n return\n spis = F.otkr_f(\n F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + dolgn + '.pickle',\n pickl=True)\n vstav_pol = 8\n spis[0].insert(vstav_pol, 'Факт. вып.')\n spis[1].insert(vstav_pol, '-')\n for i in range(2, len(spis)):\n spis[i].insert(vstav_pol, '')\n vstav_pol = 9\n spis[0].insert(vstav_pol, 'Подытог,%')\n spis[1].insert(vstav_pol, '-')\n for i in range(2, len(spis)):\n spis[i].insert(vstav_pol, '')\n zapolit_tabl_kpi(self, spis)\n\n\ndef zapolit_tabl_kpi(self, spis):\n edit = {F.nom_kol_po_im_v_shap(spis, 'Факт. вып.')}\n F.zapoln_wtabl(self, spis, self.ui.tbl_kpi_sotr, 0, edit, (), (), 200, True, '')\n self.ui.tbl_kpi_sotr.setColumnWidth(1, 350)\n self.ui.tbl_kpi_sotr.setColumnWidth(0, 200)\n self.ui.tbl_kpi_sotr.setColumnWidth(7, 150)\n self.ui.tbl_kpi_sotr.setColumnWidth(9, 100)\n self.ui.tbl_kpi_sotr.horizontalHeader().setStretchLastSection(True)\n F.cvet_cell_wtabl(self.ui.tbl_kpi_sotr, 'Факт. вып.', '*', '', inventir=True)\n F.dob_color_wtab(self.ui.tbl_kpi_sotr, 0, F.nom_kol_po_im_v_shap(spis, 'Факт. вып.'), 20, 20, 20)\n for i in range(0, self.ui.tbl_kpi_sotr.columnCount()):\n F.dob_color_wtab(self.ui.tbl_kpi_sotr, 0, i, 30, 30, 30)\n\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Цель'), 12)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Наименование КПЭ'), 10)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Методика расчета'), 10)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Примечание'), 10)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Уров.вып.№1'), 18)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Уров.вып.№2'), 18)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Уров.вып.№3'), 18)\n F.font_size(self.ui.tbl_kpi_sotr, F.nom_kol_po_im_v_shap(spis, 'Факт. вып.'), 20)\n self.ui.tbl_kpi_sotr.resizeRowsToContents()\n\n\ndef save_sotr(self):\n if self.windowTitle() == \"Расчет КПЭ\":\n return\n if not proverka_dannih(self):\n return\n spis = F.spisok_iz_wtabl(self.ui.tbl_kpi_sotr, '', True)\n name = self.ui.l_period.text()\n if not F.nalich_file(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name):\n F.sozd_dir(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name)\n F.zap_f(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name + F.sep() + name + \"$\" +\n self.ui.cmb_rabotn.currentText() + '.pickle',\n spis, pickl=True)\n self.showdialog(\n f'КПЭ на {self.ui.cmb_rabotn.currentText()} успешно сохранен')\n aut.load_combo_sotr(self, self.ui.cmb_rabotn.currentIndex())\n\n\ndef proverka_dannih(self, spis=()):\n if spis == ():\n spis = F.spisok_iz_wtabl(self.ui.tbl_kpi_sotr, '', True)\n kol = F.nom_kol_po_im_v_shap(spis, 'Факт. вып.')\n kol_vid = self.shapka_shablonkpi[0].index('Тип КПЭ')\n kol_pred1 = self.shapka_shablonkpi[0].index(\"Уров.вып.№1\")\n kol_pred2 = self.shapka_shablonkpi[0].index(\"Уров.вып.№3\")\n spis[1][kol] = \"-\"\n if len(spis) < 3:\n self.showdialog(f'Таблица не может быть не заполнена')\n return False\n for i in range(2, len(spis)):\n if spis[i][kol_vid] != self.KPITIPS[0]:\n if spis[i][kol] != '' and spis[i][kol] != '0':\n spis[i][kol] = \"1\"\n else:\n spis[i][kol] = \"0\"\n if spis[i][kol] == \"\":\n self.showdialog(f'Факт. вып., строка {i} не заполнена')\n F.migat(self, self.ui.tbl_kpi_sotr, i - 1, kol)\n return False\n if spis[i][kol_vid] == self.KPITIPS[0] and not F.is_numeric(spis[i][kol]):\n self.showdialog(f'Факт. вып., строка {i} не число')\n F.migat(self, self.ui.tbl_kpi_sotr, i - 1, kol)\n return False\n\n if spis[i][kol_vid] == self.KPITIPS[0]:\n if int(spis[i][kol_pred2]) > int(spis[i][kol_pred1]):\n if int(spis[i][kol]) < int(spis[i][kol_pred1]) or int(spis[i][kol]) > int(spis[i][kol_pred2]):\n self.showdialog(f'Факт. вып., строка {i} не в пределах уров.вып.')\n F.migat(self, self.ui.tbl_kpi_sotr, i - 1, kol)\n return False\n if int(spis[i][kol_pred2]) < int(spis[i][kol_pred1]):\n if int(spis[i][kol]) > int(spis[i][kol_pred1]) or int(spis[i][kol]) < int(spis[i][kol_pred2]):\n self.showdialog(f'Факт. вып., строка {i} не в пределах уров.вып.')\n F.migat(self, self.ui.tbl_kpi_sotr, i - 1, kol)\n return False\n if int(spis[i][kol]) < 0:\n self.showdialog(f'Факт. вып., строка {i} не может быть меньше 0')\n F.migat(self, self.ui.tbl_kpi_sotr, i - 1, kol)\n return False\n return spis\n\n\ndef rasschet_sotr(self):\n if self.windowTitle() == \"Расчет КПЭ\":\n return\n if not proverka_dannih(self):\n return\n spis = proverka_dannih(self)\n summ = 0\n kol_fact = F.nom_kol_po_im_v_shap(spis, 'Факт. вып.')\n kol_tip = F.nom_kol_po_im_v_shap(spis, \"Тип КПЭ\")\n kol_z1 = F.nom_kol_po_im_v_shap(spis, \"Уров.вып.№1\")\n kol_z2 = F.nom_kol_po_im_v_shap(spis, \"Уров.вып.№2\")\n kol_z3 = F.nom_kol_po_im_v_shap(spis, \"Уров.вып.№3\")\n kol_ves = F.nom_kol_po_im_v_shap(spis, \"Вес КПЭ\")\n kol_podit = F.nom_kol_po_im_v_shap(spis, \"Подытог,%\")\n flag_otsek = False\n for i in range(2, len(spis)):\n if spis[i][kol_tip] == self.KPITIPS[0]:\n summ += rassch_nepr(spis, i, kol_fact, kol_z1, kol_z2, kol_z3) * int(spis[i][kol_ves]) / 100\n spis[i][kol_podit] = str(\n round(rassch_nepr(spis, i, kol_fact, kol_z1, kol_z2, kol_z3) * int(spis[i][kol_ves]) / 100, 1))\n if spis[i][kol_tip] == self.KPITIPS[1]:\n summ -= int(spis[i][kol_fact]) * int(spis[i][kol_ves])\n spis[i][kol_podit] = str(int(spis[i][kol_fact]) * int(spis[i][kol_ves]) * -1)\n if spis[i][kol_tip] == self.KPITIPS[2]:\n if spis[i][kol_fact] == '1':\n spis[i][kol_podit] = '*0'\n # summ -= summ * int(spis[i][kol_fact])\n flag_otsek = True\n else:\n spis[i][kol_podit] = '0'\n if flag_otsek:\n summ = 0\n self.ui.l_raschet.setText(f\"Итого: {str(round(summ, 1))}\")\n zapolit_tabl_kpi(self, spis)\n\n\ndef rassch_nepr(spis, i, kol_fact, kol_z1, kol_z2, kol_z3):\n fact = int(spis[i][kol_fact])\n z1 = int(spis[i][kol_z1])\n z2 = int(spis[i][kol_z2])\n z3 = int(spis[i][kol_z3])\n y1 = int(spis[1][kol_z1])\n y2 = int(spis[1][kol_z2])\n y3 = int(spis[1][kol_z3])\n if z3 > z1:\n if fact < z2:\n proc = (fact - z1) / (z2 - z1)\n return (y2 - y1) * proc + y1\n else:\n proc = (fact - z2) / (z3 - z2)\n return (y3 - y2) * proc + y2\n else:\n if fact > z2:\n proc = (z1 - fact) / (z1 - z2)\n return (y2 - y1) * proc + y1\n else:\n proc = (z2 - fact) / (z2 - z3)\n return (y3 - y2) * proc + y2\n\n\n# no 4194304\ndef del_kpi_sotr(self):\n if self.windowTitle() == \"Расчет КПЭ\":\n return\n rez = self.showdialogYN(f'Будет удален КПЭ для {self.ui.cmb_rabotn.currentText()} на {self.ui.l_period.text()}')\n if rez == 1024:\n name = self.ui.l_period.text()\n F.udal_file(F.scfg(\n 'strukt') + F.sep() + self.windowTitle() + F.sep() + name + F.sep() + name + \"$\" +\n self.ui.cmb_rabotn.currentText() + '.pickle')\n set_rabotn(self)\n self.showdialog(f'КПЭ для {self.ui.cmb_rabotn.currentText()} успешно удален.\\n'\n f'его не вернуть.\\n'\n f'никак.')\n aut.load_combo_sotr(self, self.ui.cmb_rabotn.currentIndex())\n\n\ndef export(self):\n if self.windowTitle() == \"Расчет КПЭ\":\n return\n if F.nalich_file(os.path.join(\"icons\", \"шаблон.docx\")) == False:\n self.showdialog(\"шаблон не найден\")\n return\n rasschet_sotr(self)\n msg = \"\"\n sch = 0\n spis = F.spisok_iz_wtabl(self.ui.tbl_kpi_sotr, '', True)\n kol_cel = F.nom_kol_po_im_v_shap(spis, \"Цель\")\n kol_naim = F.nom_kol_po_im_v_shap(spis, \"Наименование КПЭ\")\n kol_ed = F.nom_kol_po_im_v_shap(spis, \"Ед. изм.\")\n kol_fact = F.nom_kol_po_im_v_shap(spis, 'Факт. вып.')\n i = 0\n msg += f'{vpisat(\"№\", 2)}|{vpisat(spis[i][kol_cel], 22)}|{vpisat(spis[i][kol_naim], 60)}|{vpisat(spis[i][kol_ed], 10)}|{vpisat(spis[i][kol_fact], 10)}\\n'\n msg += f'{vpisat(\"-\", 2, znac=\"-\")}|{vpisat(\"-\", 22, znac=\"-\")}|' \\\n f'{vpisat(\"-\", 60, znac=\"-\")}|{vpisat(\"-\", 10, znac=\"-\")}|' \\\n f'{vpisat(\"-\", 10, znac=\"-\")}\\n'\n\n for i in range(2, len(spis)):\n sch += 1\n msg += f'{vpisat(str(sch), 2)}|{vpisat(spis[i][kol_cel], 22)}|{vpisat(spis[i][kol_naim], 60)}|{vpisat(spis[i][kol_ed], 10)}|{vpisat(spis[i][kol_fact], 10)}\\n'\n\n doc = DocxTemplate(os.path.join(\"icons\", \"шаблон.docx\"))\n context = {'emploe': self.ui.cmb_rabotn.currentText(), 'period': self.ui.l_period.text(), 'kpi': msg,\n 'itog': self.ui.l_raschet.text(), 'now': F.now()}\n\n doc.render(context)\n if F.nalich_file(F.put_po_umolch() + os.sep + 'КПЭ' + os.sep) == False:\n F.sozd_dir(F.put_po_umolch() + os.sep + 'КПЭ' + os.sep)\n putf = f'{F.put_po_umolch()}{os.sep}КПЭ{os.sep}{self.fio(self.ui.cmb_rabotn.currentText())}${self.ui.l_period.text()}.docx'\n doc.save(putf)\n F.zapyst_file(putf)\n return\n # wdFormatPDF = 17\n #\n # in_file = os.path.abspath(\"final.docx\")\n # out_file = os.path.abspath(\"final.pdf\")\n #\n # word = comtypes.client.CreateObject('Word.Application')\n # doc = word.Documents.Open(in_file)\n # doc.SaveAs(out_file, FileFormat=wdFormatPDF)\n # doc.Close()\n # word.Quit()\n\n\ndef vpisat(text, dl, orient=0, znac=\" \"):\n text = str(text)\n text = text.strip().replace('\\n', '')\n text = text[:dl]\n\n if orient == 1:\n if (dl - len(text)) % 2 > 0:\n itog = f'{znac * (((dl - len(text)) // 2) + 1)}{text}{znac * ((dl - len(text)) // 2)}'\n else:\n itog = f'{znac * (((dl - len(text)) / 2) + 1)}{text}{znac * ((dl - len(text)) / 2)}'\n if orient == 0:\n itog = f'{text}{znac * (dl - len(text))}'\n if orient == 2:\n itog = f'{znac * (dl - len(text))}{text}'\n return itog\n", "sub_path": "kpi_sotr.py", "file_name": "kpi_sotr.py", "file_ext": "py", "file_size_in_byte": 12136, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "Cust_Functions.nalich_file", "line_number": 15, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 15, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 16, "usage_type": "call"}, {"api_name": "Cust_Functions.otkr_f", "line_number": 18, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 19, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 20, "usage_type": "call"}, {"api_name": "Cust_Functions.nalich_file", "line_number": 24, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 24, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 25, "usage_type": "call"}, {"api_name": "Cust_Functions.otkr_f", "line_number": 29, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 30, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 31, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 47, "usage_type": "call"}, {"api_name": "Cust_Functions.zapoln_wtabl", "line_number": 48, "usage_type": "call"}, {"api_name": "Cust_Functions.cvet_cell_wtabl", "line_number": 54, "usage_type": "call"}, {"api_name": "Cust_Functions.dob_color_wtab", "line_number": 55, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 55, "usage_type": "call"}, {"api_name": "Cust_Functions.dob_color_wtab", "line_number": 57, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 59, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 59, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 60, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 60, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 61, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 61, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 62, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 62, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 63, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 63, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 64, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 64, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 65, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 65, "usage_type": "call"}, {"api_name": "Cust_Functions.font_size", "line_number": 66, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 66, "usage_type": "call"}, {"api_name": "Cust_Functions.spisok_iz_wtabl", "line_number": 75, "usage_type": "call"}, {"api_name": "Cust_Functions.nalich_file", "line_number": 77, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 77, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 78, "usage_type": "call"}, {"api_name": "Cust_Functions.sozd_dir", "line_number": 79, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 79, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 80, "usage_type": "call"}, {"api_name": "Cust_Functions.zap_f", "line_number": 81, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 81, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 82, "usage_type": "call"}, {"api_name": "autorization.load_combo_sotr", "line_number": 87, "usage_type": "call"}, {"api_name": "Cust_Functions.spisok_iz_wtabl", "line_number": 92, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 93, "usage_type": "call"}, {"api_name": "Cust_Functions.migat", "line_number": 109, "usage_type": "call"}, {"api_name": "Cust_Functions.is_numeric", "line_number": 111, "usage_type": "call"}, {"api_name": "Cust_Functions.migat", "line_number": 113, "usage_type": "call"}, {"api_name": "Cust_Functions.migat", "line_number": 120, "usage_type": "call"}, {"api_name": "Cust_Functions.migat", "line_number": 125, "usage_type": "call"}, {"api_name": "Cust_Functions.migat", "line_number": 129, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 141, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 142, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 143, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 144, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 145, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 146, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 147, "usage_type": "call"}, {"api_name": "Cust_Functions.udal_file", "line_number": 201, "usage_type": "call"}, {"api_name": "Cust_Functions.scfg", "line_number": 201, "usage_type": "call"}, {"api_name": "Cust_Functions.sep", "line_number": 202, "usage_type": "call"}, {"api_name": "autorization.load_combo_sotr", "line_number": 208, "usage_type": "call"}, {"api_name": "Cust_Functions.nalich_file", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 214, "usage_type": "call"}, {"api_name": "os.path", "line_number": 214, "usage_type": "attribute"}, {"api_name": "Cust_Functions.spisok_iz_wtabl", "line_number": 220, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 221, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 222, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 223, "usage_type": "call"}, {"api_name": "Cust_Functions.nom_kol_po_im_v_shap", "line_number": 224, "usage_type": "call"}, {"api_name": "docxtpl.DocxTemplate", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 235, "usage_type": "call"}, {"api_name": "os.path", "line_number": 235, "usage_type": "attribute"}, {"api_name": "Cust_Functions.now", "line_number": 237, "usage_type": "call"}, {"api_name": "Cust_Functions.nalich_file", "line_number": 240, "usage_type": "call"}, {"api_name": "Cust_Functions.put_po_umolch", "line_number": 240, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 240, "usage_type": "attribute"}, {"api_name": "Cust_Functions.sozd_dir", "line_number": 241, "usage_type": "call"}, {"api_name": "Cust_Functions.put_po_umolch", "line_number": 241, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 241, "usage_type": "attribute"}, {"api_name": "Cust_Functions.put_po_umolch", "line_number": 242, "usage_type": "call"}, {"api_name": "os.sep", "line_number": 242, "usage_type": "attribute"}, {"api_name": "Cust_Functions.zapyst_file", "line_number": 244, "usage_type": "call"}]} +{"seq_id": "577724759", "text": "from flask_restful import Resource,reqparse\nfrom Loadbalance.Loadbalance import Loadbalance\nfrom flask import jsonify\nfrom bin.Generator.Generator import Generator\n\nclass Branches(Resource):\n def get(self):\n loadbalance = Loadbalance()\n query_string = \"select * from branch\"\n print (jsonify(loadbalance.exec(query_string)))\n return jsonify(loadbalance.exec(query_string))\n\n def post(self):\n loadbalance = Loadbalance()\n generator = Generator()\n parser = reqparse.RequestParser()\n # parser.add_argument('branch_id',type=str,help='branch id cannot be null value',required=True)\n parser.add_argument('branch_name',type=str,help='branch number cannot be null value',required=True)\n parser.add_argument('addr',type=str,help='addr cannot be null value',required=True)\n args = parser.parse_args()\n query_string = \"insert into branch(branch_id,branch_name,addr) values({},{},{});\".format(generator.getBranch_id,args[\"branch_name\"],args[\"addr\"])\n print(query_string)\n print(loadbalance.execOne(query_string))\n return jsonify({\"message\":\"branch has been created\"})\n\n def put(self):\n next()\n\n def delete(self):\n next()\n", "sub_path": "Database-service/Resource/Branches.py", "file_name": "Branches.py", "file_ext": "py", "file_size_in_byte": 1239, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask_restful.Resource", "line_number": 6, "usage_type": "name"}, {"api_name": "Loadbalance.Loadbalance.Loadbalance", "line_number": 8, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 10, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 11, "usage_type": "call"}, {"api_name": "Loadbalance.Loadbalance.Loadbalance", "line_number": 14, "usage_type": "call"}, {"api_name": "bin.Generator.Generator.Generator", "line_number": 15, "usage_type": "call"}, {"api_name": "flask_restful.reqparse.RequestParser", "line_number": 16, "usage_type": "call"}, {"api_name": "flask_restful.reqparse", "line_number": 16, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "292270958", "text": "from typing import Any, ClassVar\n\nfrom django.db.aggregates import (\n Avg as Avg,\n Count as Count,\n Max as Max,\n Min as Min,\n StdDev as StdDev,\n Sum as Sum,\n Variance as Variance,\n)\nfrom django.db.models.fields import DateTimeField as DateTimeField\n\n\nclass Model:\n id: int = ...\n objects: ClassVar[Any] = ...\n DoesNotExist: Any\n\n def save(self) -> None:\n ...\n", "sub_path": "stubs/3/django/db/models/__init__.pyi", "file_name": "__init__.pyi", "file_ext": "pyi", "file_size_in_byte": 399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.ClassVar", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 18, "usage_type": "name"}]} +{"seq_id": "517224561", "text": "from django.shortcuts import render\nfrom django.views.generic.list import ListView\nfrom django.views.generic import UpdateView,CreateView\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponse\nfrom django.contrib.auth.models import User\n\nfrom .models import (CharacterTemplate,CQuestion,\n RCTemplateCQuestions,Slams,Slam,\n SlamChart,Answer)\nimport json\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.urls import reverse_lazy\nfrom django.http import HttpResponseRedirect\n\n\n@login_required\ndef test_ajax(request):\n return render(request,'testajax.html')\n \n\n# Create your views here.\n\n@login_required\ndef Questions(request):\n context={}\n return render(request,\"addquestion.html\",context)\n\n@login_required\ndef base_t(request):\n context={\"display_text\":\"This is base Slam book page\"}\n return render(request,\"base.html\",context)\n\n\n\ndef main_t(request):\n return render(request,\"main.html\")\n\n\n#CHARACTER TEMPLATE RELATED VIEWS\n@login_required\ndef index_t(request):\n \n if request.method =='GET' and 'id' in request.GET:\n value_t=request.GET['id']\n q = CharacterTemplate.objects.get(pk=value_t)\n #context={\"display_text\":\"This is base Slam book page\"}\n context={'template':q.cq_template}\n else:\n context={}\n return render(request,\"index.html\",context)\n\n\nclass character_tlist(ListView):\n template_name=\"char_tlist.html\"\n context_object_name=\"clist\"\n model=CharacterTemplate\n def get_queryset(self):\n return CharacterTemplate.objects.filter(user=self.request.user)\n \n\nclass EditTemplateName(UpdateView):\n model=CharacterTemplate\n fields = ['cq_template']\n success_url =reverse_lazy('charactertlist')\n\n\nclass CreateTemplate(CreateView):\n model=CharacterTemplate\n fields = ['cq_template']\n def form_valid(self, form):\n obj = form.save(commit=False)\n obj.user = self.request.user\n obj.save() \n return HttpResponseRedirect(self.get_success_url())\n def get_success_url(self):\n return reverse_lazy('charactertlist')\n\n@login_required\n@csrf_exempt\ndef delete_char_t(request):\n candidate = CharacterTemplate.objects.get(pk = int(request.POST['id']))\n candidate.delete()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n\n#QUESTION RELATED VIEWS\n# @login_required\n# def index_t(request):\n \n# if request.method =='GET' and 'id' in request.GET:\n# value_t=request.GET['id']\n# q = CharacterTemplate.objects.get(pk=value_t)\n# #context={\"display_text\":\"This is base Slam book page\"}\n# context={'template':q.cq_template}\n# else:\n# context={}\n# return render(request,\"index.html\",context)\n\n\nclass cquestion_tlist(ListView):\n template_name=\"cquestion_list.html\"\n context_object_name=\"clist\"\n model=CQuestion\n def get_queryset(self):\n # if 'pk' in self.kwargs:\n # tid=self.kwargs['pk']\n if self.request.user==User(pk=1):\n queryset = { \n 'normal': CQuestion.objects.filter(user=self.request.user)}\n else:\n queryset = {'admin': CQuestion.objects.filter(user=User(pk=1)), \n 'normal': CQuestion.objects.filter(user=self.request.user)}\n \n return queryset\n \n\nclass EditCQuestion(UpdateView):\n model=CQuestion\n fields = ['cquestion']\n success_url =reverse_lazy('listcquestion')\n\n\nclass CreateCQuestion(CreateView):\n model=CQuestion\n fields = ['cquestion']\n def form_valid(self, form):\n obj = form.save(commit=False)\n obj.user = self.request.user\n obj.save() \n return HttpResponseRedirect(self.get_success_url())\n def get_success_url(self):\n return reverse_lazy('listcquestion')\n\n@login_required\n@csrf_exempt\ndef delete_cquestion_t(request):\n candidate = CQuestion.objects.get(pk = int(request.POST['id']))\n candidate.delete()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n \n\n#RCQT RELATED VIEWS\n\nclass RCQT_tlist(ListView):\n template_name=\"rcqt_list.html\"\n context_object_name=\"clist\"\n model=RCTemplateCQuestions\n def get_queryset(self):\n tid=self.kwargs['pk']\n return RCTemplateCQuestions.objects.filter(user=self.request.user,ctemplate=tid)\n \n\n@login_required\n@csrf_exempt\ndef delete_RCQT_t(request):\n candidate = RCTemplateCQuestions.objects.get(pk = int(request.POST['id']))\n candidate.delete()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n\n@login_required\n@csrf_exempt\ndef add_RCQT_t(request):\n print(request.POST)\n c=request.POST.getlist('id[]',0)\n k=request.POST.getlist('pk',0)\n if c and k:\n for i in c: \n RCTemplateCQuestions.objects.get_or_create(user=request.user,ctemplate=CharacterTemplate(pk=k[0]),cquestion=CQuestion(pk=i))\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n \n \n#SLAMS RELATED VIEWS\nclass slams_list(ListView):\n template_name=\"slams_list.html\"\n context_object_name=\"slamslist\"\n model=Slams\n def get_queryset(self):\n return Slams.objects.filter(user=self.request.user)\n \n\nclass EditSlamsName(UpdateView):\n model=Slams\n fields = ['slam_name']\n success_url =reverse_lazy('listslams')\n\n\nclass CreateSlams(CreateView):\n model=Slams\n fields = ['slam_name']\n def form_valid(self, form):\n obj = form.save(commit=False)\n obj.user = self.request.user\n obj.save() \n return HttpResponseRedirect(self.get_success_url())\n def get_success_url(self):\n return reverse_lazy('listslams')\n \n@login_required\ndef generate_slam(request):\n if request.method =='GET' and 'id' in request.GET:\n value_t=request.GET['id']\n q = CharacterTemplate.objects.get(pk=value_t)\n #context={\"display_text\":\"This is base Slam book page\"}\n context={'template':q}\n elif request.method=='POST':\n sl=Slams(user=request.user,slam_name=request.POST['slamname'])\n sl.save()\n t=RCTemplateCQuestions.objects.filter(user=request.user,ctemplate=request.POST['templateid'])\n for e in t:\n Slam.objects.get_or_create(user=request.user,slam=sl,cquestion=e.cquestion,typ=1) \n context={'generate':sl.pk}\n print(context)\n else:\n context={}\n return render(request,'generateslam.html',context)\n\n@login_required\n@csrf_exempt\ndef delete_slams(request):\n candidate = Slams.objects.get(pk = int(request.POST['id']))\n candidate.delete()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n#SLAM RELATED VIEWS-RELATION\n \nclass list_slam(ListView):\n template_name=\"list_slam.html\"\n context_object_name=\"clist\"\n model=Slam\n def get_queryset(self):\n tid=self.kwargs['pk']\n return Slam.objects.filter(user=self.request.user,slam=tid)\n \n\n@login_required\n@csrf_exempt\ndef delete_slam(request):\n candidate = Slam.objects.get(pk = int(request.POST['id']))\n candidate.delete()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n\n@login_required\n@csrf_exempt\ndef add_slam(request):\n print(request.POST)\n c=request.POST.getlist('id[]',0)\n k=request.POST.getlist('pk',0)\n if c and k:\n for i in c: \n Slam.objects.get_or_create(user=request.user,slam=Slams(pk=k[0]),cquestion=CQuestion(pk=i),typ=1)\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n#User\nclass list_user(ListView):\n template_name=\"list_user.html\"\n context_object_name=\"clist\"\n model=User\n \n \n@login_required\n@csrf_exempt\ndef send_slam(request):\n print(request.POST)\n c=request.POST.getlist('id[]',0)\n k=request.POST.getlist('pk',0)\n txt=request.POST.getlist('mess',0)\n if c and k:\n for i in c:\n SlamChart.objects.get_or_create(fr=request.user,to=User(pk=i),slam=Slams(pk=k[0]),mess=txt[0])\n \n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n#INBOX SEND RESPONSES\nclass Inbox(ListView):\n template_name=\"inbox.html\"\n context_object_name=\"clist\"\n model=SlamChart\n def get_queryset(self):\n return SlamChart.objects.filter(to=self.request.user,is_to=True).order_by('-date_time')\n \nclass Sent(ListView):\n template_name=\"sent.html\"\n context_object_name=\"clist\"\n model=SlamChart\n def get_queryset(self):\n return SlamChart.objects.filter(fr=self.request.user,is_fr=True).order_by('-date_time')\n \n\nclass Response(ListView):\n template_name=\"response.html\"\n context_object_name=\"clist\"\n model=SlamChart\n def get_queryset(self):\n return SlamChart.objects.filter(fr=self.request.user,response=False).order_by('-date_time')\n \n\n\n@login_required\n@csrf_exempt\ndef delete_sent(request):\n candidate = SlamChart.objects.get(pk = int(request.POST['id']))\n if candidate.is_to==False: \n candidate.delete()\n else:\n candidate.is_fr=False\n candidate.save()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n@login_required\n@csrf_exempt\ndef delete_inbox(request):\n candidate = SlamChart.objects.get(pk = int(request.POST['id']))\n if candidate.is_fr==False: \n candidate.delete()\n else:\n candidate.is_to=False\n candidate.save()\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\n\n\n \nclass view_slam(ListView):\n template_name=\"view_slam.html\"\n context_object_name=\"clist\"\n model=Slam\n def get_queryset(self):\n tid=self.kwargs['pk']\n k=SlamChart.objects.get(pk=tid)\n l=Slams(pk=k.slam)\n m=Slam.objects.filter(slam=l.pk)\n return m\n \n \nclass edit_slam(ListView):\n template_name=\"edit_slam.html\"\n context_object_name=\"clist\"\n model=Slam\n def get_queryset(self):\n tid=self.kwargs['pk']\n k=SlamChart.objects.get(pk=tid)\n l=Slams(pk=k.slam)\n if k.response:\n queryset = {'chart': SlamChart.objects.get(pk=tid), \n 'slam':Slam.objects.filter(slam=l.pk) }\n else:\n queryset=Answer.objects.filter(slamchart=k.pk)\n return queryset\n\n@login_required\n@csrf_exempt\ndef response_slam(request):\n print(request.POST)\n c=request.POST.getlist('id[]',0)\n s=request.POST.getlist('ans[]',0)\n k=request.POST.getlist('pk',0)\n txt=request.POST.getlist('mess',0)\n if not txt:\n txt=[]\n txt.append(\"\")\n if c and k and s:\n for i in range(len(c)):\n print(c[i])\n sl=Slam.objects.get(pk=c[i])\n #sq=CQuestion(pk=sl.cquestion)\n #print(sq)\n sc=SlamChart.objects.get(pk=k[0])\n sc.response=False\n sc.rmess=txt[0]\n sc.save()\n Answer.objects.get_or_create(cquestion=sl.cquestion,slamchart=SlamChart(pk=k[0]),ans=s[i])\n payload = {'success': True}\n return HttpResponse(json.dumps(payload), content_type='application/json')\n\nclass view_response(ListView):\n template_name=\"view_response.html\"\n context_object_name=\"clist\"\n model=Answer\n def get_queryset(self):\n tid=self.kwargs['pk']\n k=SlamChart.objects.get(pk=tid)\n return Answer.objects.filter(slamchart=k.pk)", "sub_path": "slambook/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 11786, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.shortcuts.render", "line_number": 19, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 17, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 27, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 24, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 32, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 29, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 37, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects.get", "line_number": 46, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects", "line_number": 46, "usage_type": "attribute"}, {"api_name": "models.CharacterTemplate", "line_number": 46, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 51, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 41, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 54, "usage_type": "name"}, {"api_name": "models.CharacterTemplate", "line_number": 57, "usage_type": "name"}, {"api_name": "models.CharacterTemplate.objects.filter", "line_number": 59, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects", "line_number": 59, "usage_type": "attribute"}, {"api_name": "models.CharacterTemplate", "line_number": 59, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 62, "usage_type": "name"}, {"api_name": "models.CharacterTemplate", "line_number": 63, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 65, "usage_type": "call"}, {"api_name": "django.views.generic.CreateView", "line_number": 68, "usage_type": "name"}, {"api_name": "models.CharacterTemplate", "line_number": 69, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 75, "usage_type": "call"}, {"api_name": "django.urls.reverse_lazy", "line_number": 77, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects.get", "line_number": 82, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects", "line_number": 82, "usage_type": "attribute"}, {"api_name": "models.CharacterTemplate", "line_number": 82, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 85, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 85, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 79, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 80, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 102, "usage_type": "name"}, {"api_name": "models.CQuestion", "line_number": 105, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 109, "usage_type": "call"}, {"api_name": "models.CQuestion.objects.filter", "line_number": 111, "usage_type": "call"}, {"api_name": "models.CQuestion.objects", "line_number": 111, "usage_type": "attribute"}, {"api_name": "models.CQuestion", "line_number": 111, "usage_type": "name"}, {"api_name": "models.CQuestion.objects.filter", "line_number": 113, "usage_type": "call"}, {"api_name": "models.CQuestion.objects", "line_number": 113, "usage_type": "attribute"}, {"api_name": "models.CQuestion", "line_number": 113, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 113, "usage_type": "call"}, {"api_name": "models.CQuestion.objects.filter", "line_number": 114, "usage_type": "call"}, {"api_name": "models.CQuestion.objects", "line_number": 114, "usage_type": "attribute"}, {"api_name": "models.CQuestion", "line_number": 114, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 119, "usage_type": "name"}, {"api_name": "models.CQuestion", "line_number": 120, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 122, "usage_type": "call"}, {"api_name": "django.views.generic.CreateView", "line_number": 125, "usage_type": "name"}, {"api_name": "models.CQuestion", "line_number": 126, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 132, "usage_type": "call"}, {"api_name": "django.urls.reverse_lazy", "line_number": 134, "usage_type": "call"}, {"api_name": "models.CQuestion.objects.get", "line_number": 139, "usage_type": "call"}, {"api_name": "models.CQuestion.objects", "line_number": 139, "usage_type": "attribute"}, {"api_name": "models.CQuestion", "line_number": 139, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 142, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 142, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 136, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 137, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 147, "usage_type": "name"}, {"api_name": "models.RCTemplateCQuestions", "line_number": 150, "usage_type": "name"}, {"api_name": "models.RCTemplateCQuestions.objects.filter", "line_number": 153, "usage_type": "call"}, {"api_name": "models.RCTemplateCQuestions.objects", "line_number": 153, "usage_type": "attribute"}, {"api_name": "models.RCTemplateCQuestions", "line_number": 153, "usage_type": "name"}, {"api_name": "models.RCTemplateCQuestions.objects.get", "line_number": 159, "usage_type": "call"}, {"api_name": "models.RCTemplateCQuestions.objects", "line_number": 159, "usage_type": "attribute"}, {"api_name": "models.RCTemplateCQuestions", "line_number": 159, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 162, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 162, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 156, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 157, "usage_type": "name"}, {"api_name": "models.RCTemplateCQuestions.objects.get_or_create", "line_number": 173, "usage_type": "call"}, {"api_name": "models.RCTemplateCQuestions.objects", "line_number": 173, "usage_type": "attribute"}, {"api_name": "models.RCTemplateCQuestions", "line_number": 173, "usage_type": "name"}, {"api_name": "models.CharacterTemplate", "line_number": 173, "usage_type": "call"}, {"api_name": "models.CQuestion", "line_number": 173, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 175, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 175, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 165, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 166, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 179, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 182, "usage_type": "name"}, {"api_name": "models.Slams.objects.filter", "line_number": 184, "usage_type": "call"}, {"api_name": "models.Slams.objects", "line_number": 184, "usage_type": "attribute"}, {"api_name": "models.Slams", "line_number": 184, "usage_type": "name"}, {"api_name": "django.views.generic.UpdateView", "line_number": 187, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 188, "usage_type": "name"}, {"api_name": "django.urls.reverse_lazy", "line_number": 190, "usage_type": "call"}, {"api_name": "django.views.generic.CreateView", "line_number": 193, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 194, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 200, "usage_type": "call"}, {"api_name": "django.urls.reverse_lazy", "line_number": 202, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects.get", "line_number": 208, "usage_type": "call"}, {"api_name": "models.CharacterTemplate.objects", "line_number": 208, "usage_type": "attribute"}, {"api_name": "models.CharacterTemplate", "line_number": 208, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 212, "usage_type": "call"}, {"api_name": "models.RCTemplateCQuestions.objects.filter", "line_number": 214, "usage_type": "call"}, {"api_name": "models.RCTemplateCQuestions.objects", "line_number": 214, "usage_type": "attribute"}, {"api_name": "models.RCTemplateCQuestions", "line_number": 214, "usage_type": "name"}, {"api_name": "models.Slam.objects.get_or_create", "line_number": 216, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 216, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 216, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 221, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 204, "usage_type": "name"}, {"api_name": "models.Slams.objects.get", "line_number": 226, "usage_type": "call"}, {"api_name": "models.Slams.objects", "line_number": 226, "usage_type": "attribute"}, {"api_name": "models.Slams", "line_number": 226, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 229, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 229, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 223, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 224, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 233, "usage_type": "name"}, {"api_name": "models.Slam", "line_number": 236, "usage_type": "name"}, {"api_name": "models.Slam.objects.filter", "line_number": 239, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 239, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 239, "usage_type": "name"}, {"api_name": "models.Slam.objects.get", "line_number": 245, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 245, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 245, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 248, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 248, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 242, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 243, "usage_type": "name"}, {"api_name": "models.Slam.objects.get_or_create", "line_number": 259, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 259, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 259, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 259, "usage_type": "call"}, {"api_name": "models.CQuestion", "line_number": 259, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 261, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 261, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 251, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 252, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 264, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 267, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get_or_create", "line_number": 279, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 279, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 279, "usage_type": "name"}, {"api_name": "django.contrib.auth.models.User", "line_number": 279, "usage_type": "call"}, {"api_name": "models.Slams", "line_number": 279, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 282, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 282, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 270, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 271, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 285, "usage_type": "name"}, {"api_name": "models.SlamChart", "line_number": 288, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.filter", "line_number": 290, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 290, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 290, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 292, "usage_type": "name"}, {"api_name": "models.SlamChart", "line_number": 295, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.filter", "line_number": 297, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 297, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 297, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 300, "usage_type": "name"}, {"api_name": "models.SlamChart", "line_number": 303, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.filter", "line_number": 305, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 305, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 305, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 312, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 312, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 312, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 319, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 319, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 309, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 310, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 324, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 324, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 324, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 331, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 331, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 321, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 322, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 336, "usage_type": "name"}, {"api_name": "models.Slam", "line_number": 339, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 342, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 342, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 342, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 343, "usage_type": "call"}, {"api_name": "models.Slam.objects.filter", "line_number": 344, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 344, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 344, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 348, "usage_type": "name"}, {"api_name": "models.Slam", "line_number": 351, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 354, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 354, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 354, "usage_type": "name"}, {"api_name": "models.Slams", "line_number": 355, "usage_type": "call"}, {"api_name": "models.SlamChart.objects.get", "line_number": 357, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 357, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 357, "usage_type": "name"}, {"api_name": "models.Slam.objects.filter", "line_number": 358, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 358, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 358, "usage_type": "name"}, {"api_name": "models.Answer.objects.filter", "line_number": 360, "usage_type": "call"}, {"api_name": "models.Answer.objects", "line_number": 360, "usage_type": "attribute"}, {"api_name": "models.Answer", "line_number": 360, "usage_type": "name"}, {"api_name": "models.Slam.objects.get", "line_number": 377, "usage_type": "call"}, {"api_name": "models.Slam.objects", "line_number": 377, "usage_type": "attribute"}, {"api_name": "models.Slam", "line_number": 377, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 380, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 380, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 380, "usage_type": "name"}, {"api_name": "models.Answer.objects.get_or_create", "line_number": 384, "usage_type": "call"}, {"api_name": "models.Answer.objects", "line_number": 384, "usage_type": "attribute"}, {"api_name": "models.Answer", "line_number": 384, "usage_type": "name"}, {"api_name": "models.SlamChart", "line_number": 384, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 386, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 386, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 363, "usage_type": "name"}, {"api_name": "django.views.decorators.csrf.csrf_exempt", "line_number": 364, "usage_type": "name"}, {"api_name": "django.views.generic.list.ListView", "line_number": 388, "usage_type": "name"}, {"api_name": "models.Answer", "line_number": 391, "usage_type": "name"}, {"api_name": "models.SlamChart.objects.get", "line_number": 394, "usage_type": "call"}, {"api_name": "models.SlamChart.objects", "line_number": 394, "usage_type": "attribute"}, {"api_name": "models.SlamChart", "line_number": 394, "usage_type": "name"}, {"api_name": "models.Answer.objects.filter", "line_number": 395, "usage_type": "call"}, {"api_name": "models.Answer.objects", "line_number": 395, "usage_type": "attribute"}, {"api_name": "models.Answer", "line_number": 395, "usage_type": "name"}]} +{"seq_id": "245792820", "text": "# Copyright 2019 Robert Csordas. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# ==============================================================================\n\nimport torch\nimport torch.nn.functional as F\nimport math\n\nclass Graph:\n def __init__(self, batch_size, state_size, device):\n if batch_size is None:\n return\n self.batch_size = batch_size\n self.device = device\n\n self.nodes = torch.zeros(0, state_size, dtype=torch.float32, device=device)\n self.node_types = torch.zeros(0, dtype=torch.uint8, device=device)\n self.edge_source = torch.zeros(0, dtype=torch.long, device=device)\n self.edge_dest = torch.zeros(0, dtype=torch.long, device=device)\n self.edge_features = torch.zeros(0, state_size, dtype=torch.float, device=device)\n self.edge_types = torch.zeros(0, dtype=torch.uint8, device=device)\n self.owner_masks = torch.zeros(batch_size, 0, dtype=torch.uint8, device=device)\n self.last_inserted_node = torch.zeros(batch_size, dtype=torch.long, device=device)\n\n self.running = torch.ones(batch_size, device=device, dtype=torch.uint8)\n\n def get_final_graph(self, device=torch.device(\"cpu\")):\n needed = [\"node_types\", \"edge_source\", \"edge_dest\", \"edge_types\", \"owner_masks\"]\n res = Graph(self.batch_size, 1, device)\n for k, v in self.__dict__.items(): \n if torch.is_tensor(v):\n res.__dict__[k] = v.to(device) if k in needed else None\n return res\n\ndef sample_softmax(tensor, dim=-1):\n eps=1e-20\n\n # Built in gumbel softmax could end up with lots of nans. Do it manually here.\n noise = -torch.log(-torch.log(torch.rand_like(tensor)+eps) + eps)\n res = F.softmax(tensor + noise, dim=-1)\n _, res = res.max(dim=dim)\n return res\n\ndef mask_softmax_input(tensor, mask):\n return torch.where(mask, tensor, torch.full([1], float(\"-inf\"), dtype=tensor.dtype, device=tensor.device))\n\ndef masked_softmax(tensor, mask):\n tensor = mask_softmax_input(tensor, mask)\n return sample_softmax(tensor)\n\ndef loss_running_gate(l, running):\n return torch.where(running, l, torch.zeros([1], dtype=l.dtype, device=l.device)).mean()\n\ndef masked_cross_entropy_loss(tensor, mask, target, enabled):\n tensor = mask_softmax_input(tensor, mask) if mask is not None else tensor\n l = F.cross_entropy(tensor, target.long(), reduction=\"none\")\n return loss_running_gate(l, enabled)\n\ndef remap_pad(t, pad_char, transform = lambda x: x+1):\n return torch.where(t != pad_char, transform(t), torch.zeros(1, dtype=t.dtype, device=t.device))\n\ndef masked_bce_loss(tensor, target, enabled):\n l = F.binary_cross_entropy_with_logits(tensor, target.float(), reduction=\"none\")\n return loss_running_gate(l, enabled)\n\ndef sample_binary(tensor):\n tensor = torch.sigmoid(tensor)\n return torch.rand_like(tensor) < tensor\n\ndef xavier_init(layer, scale, n_inputs=None, n_outputs=None):\n n_inputs = n_inputs if n_inputs is not None else layer.weight.shape[1]\n n_outputs = n_outputs if n_outputs is not None else layer.weight.shape[0]\n limits = scale * math.sqrt(6.0 / (n_inputs + n_outputs))\n layer.weight.data.uniform_(-limits, limits)\n\n if layer.bias is not None:\n torch.nn.init.normal_(layer.bias)\n\n\nclass Aggregator(torch.nn.Module):\n def __init__(self, state_size, aggregated_size, dropout, bias_if_empty=False):\n super().__init__()\n\n self.transform = torch.nn.Linear(state_size, aggregated_size)\n self.gate = torch.nn.Sequential(\n torch.nn.Linear(state_size, aggregated_size),\n torch.nn.Sigmoid()\n )\n\n self.bias_if_empty = torch.nn.Parameter(torch.Tensor(1,aggregated_size)) if bias_if_empty else None\n\n self.drop = torch.nn.Dropout(dropout)\n\n self.aggregated_size = aggregated_size\n self._reset_parameters()\n\n def forward(self, graph: Graph):\n if graph.nodes.shape[0]==0:\n if self.bias_if_empty is not None:\n return self.bias_if_empty.expand(graph.batch_size, -1)\n else:\n return torch.zeros(graph.batch_size, self.aggregated_size, dtype=torch.float32, device=graph.device)\n\n gates = self.gate(graph.nodes)\n data = self.transform(graph.nodes)\n\n fmask = graph.owner_masks.float()\n res = torch.mm(fmask, data * gates)\n\n # Normalize the result with the number of nodes.\n return self.drop(res)\n\n def _reset_parameters(self):\n xavier_init(self.transform, 1)\n xavier_init(self.gate[0], 1)\n self.gate[0].bias.data.fill_(1)\n if self.bias_if_empty is not None:\n torch.nn.init.normal_(self.bias_if_empty)\n\n\nclass Propagator(torch.nn.Module):\n def __init__(self, state_size, dropout):\n super().__init__()\n\n self.message_size = state_size * 2\n\n self.node_update_fn = torch.nn.GRUCell(self.message_size, state_size)\n\n # The first layer of message function (fe) can be decomposed to 3 parts, which makes it easier to\n # claculate\n self.message_node = torch.nn.Linear(state_size, self.message_size, bias=False)\n self.message_features = torch.nn.Linear(state_size, self.message_size, bias=False)\n\n self.message_layer_2 = torch.nn.Sequential(\n torch.nn.Tanh(),\n torch.nn.Linear(self.message_size, self.message_size)\n )\n\n self.dropout = torch.nn.Dropout(dropout)\n self._reset_parameters(state_size)\n\n @staticmethod\n def _node_update_mask(graph: Graph, mask_override: torch.ByteTensor):\n return graph.owner_masks[graph.running if mask_override is None else mask_override].sum(0)>0\n\n def forward(self, graph: Graph, mask_override: torch.ByteTensor = None):\n if graph.nodes.shape[0]==0 or graph.edge_features.shape[0]==0:\n return graph\n\n edge_features = self.message_features(graph.edge_features)\n node_features = self.message_node(graph.nodes)\n\n e1 = node_features.index_select(dim=0, index=graph.edge_source)\n e2 = node_features.index_select(dim=0, index=graph.edge_dest)\n\n messages = e1 + e2 + edge_features\n messages = self.message_layer_2(messages)\n messages = self.dropout(messages)\n\n # Sum the messages for each node\n inputs = torch.zeros(graph.nodes.shape[0], self.message_size, device=graph.nodes.device,\n dtype=graph.nodes.dtype).index_add_(0, graph.edge_dest, messages).\\\n index_add_(0, graph.edge_source, messages)\n\n inputs = self.dropout(inputs)\n\n # Transform node state of running nodes\n new_nodes = self.node_update_fn(inputs, graph.nodes)\n\n graph.nodes = torch.where(self._node_update_mask(graph, mask_override).unsqueeze(-1), new_nodes, graph.nodes)\n return graph\n\n def _reset_parameters(self, state_size):\n # msg_gain = 1\n msg_gain = torch.nn.init.calculate_gain(\"tanh\")\n xavier_init(self.message_node, msg_gain, state_size * 3, self.message_size)\n xavier_init(self.message_features, msg_gain, state_size * 3, self.message_size)\n xavier_init(self.message_layer_2[1], 1)\n \n self.node_update_fn.bias_hh.data.fill_(0)\n self.node_update_fn.bias_ih.data.fill_(0)\n self.node_update_fn.bias_hh[:state_size].data.fill_(1)\n\n\nclass MultilayerPropagator(torch.nn.Module):\n def __init__(self, state_size, n_steps, dropout):\n super().__init__()\n self.propagators = torch.nn.ModuleList([Propagator(state_size, dropout) for i in range(n_steps)])\n\n def forward(self, graph: Graph, *args, **kwargs):\n for p in self.propagators:\n graph = p(graph, *args, **kwargs)\n return graph\n\n\nclass NodeAdder(torch.nn.Module):\n def __init__(self, state_size, aggregated_size, propagate_steps, n_node_types, pad_char, dropout):\n super().__init__()\n\n self.pad_char = pad_char\n\n self.propagator = MultilayerPropagator(state_size, propagate_steps, dropout)\n self.decision_aggregator = Aggregator(state_size, aggregated_size, dropout, bias_if_empty=True)\n self.init_aggregator = Aggregator(state_size, aggregated_size, dropout, bias_if_empty=True)\n\n self.node_type_decision = torch.nn.Linear(aggregated_size, n_node_types+1)\n\n self.node_type_embedding = torch.nn.Parameter(torch.Tensor(n_node_types, state_size))\n\n self.f_init_part1 = torch.nn.Linear(state_size, state_size)\n self.f_init_part2 = torch.nn.Linear(aggregated_size, state_size, bias=False)\n self._reset_parameters(state_size, aggregated_size)\n\n def forward(self, graph: Graph, reference: torch.ByteTensor):\n loss = 0\n graph = self.propagator(graph)\n\n new_node_types = self.node_type_decision(self.decision_aggregator(graph))\n if reference is not None:\n selected_type = remap_pad(reference, self.pad_char)\n loss = loss + masked_cross_entropy_loss(new_node_types, None, selected_type, graph.running)\n else:\n # Prevent generating empty graph. Set termination probability to 0 if generating the first element.\n if graph.nodes.shape[0]==0:\n new_node_types[:, 0]=float(\"-inf\")\n selected_type = sample_softmax(new_node_types)\n\n # Update running flags. If no new node is generated, the graph generation is stopped\n graph.running = (selected_type != 0) & graph.running\n if graph.running.any():\n # Initialize new nodes\n new_type_embedding = self.node_type_embedding.index_select(0, (selected_type.long() - 1).clamp(min=0))\n init_features = self.init_aggregator(graph)\n\n new_features = self.f_init_part1(new_type_embedding) + self.f_init_part2(init_features)\n\n # Add the new nodes\n mask = graph.running\n index_seq = torch.arange(mask.long().sum(), device = graph.device, dtype = torch.long) + \\\n (graph.nodes.shape[0] if graph.nodes is not None else 0)\n last_nodes = torch.zeros(graph.batch_size, device = graph.device, dtype = torch.long)\n last_nodes[mask] = index_seq\n\n # Select last node if updated\n graph.last_inserted_node = torch.where(mask, last_nodes, graph.last_inserted_node)\n\n # Merge new nodes to the node list\n new_nodes = new_features[mask]\n owner_masks = F.one_hot(mask.nonzero().squeeze(-1), graph.batch_size).transpose(0,1).byte()\n\n graph.nodes = torch.cat((graph.nodes, new_nodes), dim=0)\n graph.owner_masks = torch.cat((graph.owner_masks, owner_masks), dim=1)\n graph.node_types = torch.cat((graph.node_types, selected_type[mask].byte()-1), dim=0)\n\n return graph, loss\n\n def _reset_parameters(self, state_size, aggregated_size):\n torch.nn.init.normal_(self.node_type_embedding)\n xavier_init(self.f_init_part1, 1, state_size + aggregated_size, state_size)\n xavier_init(self.f_init_part2, 1, state_size + aggregated_size, state_size)\n xavier_init(self.node_type_decision, 1)\n\n\nclass EdgeAdder(torch.nn.Module):\n def __init__(self, state_size, aggregated_size, n_edge_dtypes, pad_char, propagate_steps, n_max_edges, dropout):\n super().__init__()\n\n self.pad_char = pad_char\n self.n_edge_dtypes = n_edge_dtypes\n self.n_max_edges = n_max_edges\n\n self.propagator = MultilayerPropagator(state_size, propagate_steps, dropout)\n\n self.edge_decision_aggregator = Aggregator(state_size, aggregated_size, dropout)\n self.edge_init = torch.nn.Parameter(torch.Tensor(n_edge_dtypes, state_size))\n self.edge_init_aggregator = Aggregator(state_size, aggregated_size, dropout)\n\n self.f_addedge_aggregated = torch.nn.Linear(aggregated_size, 1)\n self.f_addedge_new = torch.nn.Linear(state_size, 1, bias=False)\n\n self.fs_layer1_target = torch.nn.Linear(state_size, n_edge_dtypes)\n self.fs_layer1_new = torch.nn.Linear(state_size, n_edge_dtypes, bias=False)\n\n # self.fs_layer1_target = torch.nn.Linear(state_size, (state_size+n_edge_dtypes)//2)\n # self.fs_layer1_new = torch.nn.Linear(state_size, (state_size+n_edge_dtypes)//2, bias=False)\n\n # self.fs_rest = torch.nn.Sequential(\n # torch.nn.Tanh(),\n # torch.nn.Linear((state_size+n_edge_dtypes)//2, n_edge_dtypes)\n # )\n\n self._reset_paramters(state_size, aggregated_size, n_edge_dtypes)\n\n def forward(self, graph: Graph, reference):\n # Decide whether to add an edge.\n loss = 0\n running = graph.running\n\n if reference is not None and not reference:\n return graph, loss\n\n add_index = 0\n\n new_nodes = graph.nodes.index_select(0, graph.last_inserted_node)\n\n while True:\n graph = self.propagator(graph, running)\n new_edge_needed = (self.f_addedge_aggregated(self.edge_decision_aggregator(graph)) +\n self.f_addedge_new(new_nodes)).squeeze(-1)\n\n if reference is not None:\n assert self.n_max_edges is None or add_index < self.n_max_edges\n need_to_add = reference[add_index][1] != self.pad_char\n loss = loss + masked_bce_loss(new_edge_needed, need_to_add, running)\n else:\n need_to_add = sample_binary(new_edge_needed)\n\n # Force termination when the limit is reached.\n if self.n_max_edges is not None and add_index >= self.n_max_edges:\n need_to_add = torch.zeros_like(need_to_add)\n\n # Stop if there are no more edges added\n running = running & need_to_add\n if not running.any():\n break\n\n # Decide where to add\n # The transform is fs(new_node, all_other_nodes). First layer of this can be decomposed to\n # fs_layer1_target(all_other_nodes) + fs_layer1_new(new_node).\n\n logits = self.fs_layer1_target(graph.nodes).unsqueeze(0) + self.fs_layer1_new(new_nodes).unsqueeze(1)\n logits = logits.view(logits.shape[0], -1)\n\n # Logits is a [batch_size, n_nodes * n_edge_types] tensor. A softmax over all of this is done, and\n # then sampled.\n owner_mask_expanded = graph.owner_masks.unsqueeze(-1).expand(-1,-1, self.n_edge_dtypes).contiguous().\\\n view(graph.batch_size,-1)\n\n if reference is not None:\n selected_edge = reference[add_index][0].long() * self.n_edge_dtypes + \\\n remap_pad(reference[add_index][1].long(), self.pad_char, lambda x: x)\n\n loss = loss + masked_cross_entropy_loss(logits, owner_mask_expanded, selected_edge, running)\n else:\n selected_edge = masked_softmax(logits, owner_mask_expanded)\n\n selected_type = selected_edge % self.n_edge_dtypes\n selected_other = selected_edge / self.n_edge_dtypes\n\n # Add the new edges.\n selected_src = graph.last_inserted_node[running]\n selected_other = selected_other[running]\n type = selected_type[running]\n\n feature = self.edge_init.index_select(0, (type.long()-1).clamp(min=0))\n\n type = type.byte()\n\n graph.edge_dest = torch.cat((graph.edge_dest, selected_src), 0)\n graph.edge_source = torch.cat((graph.edge_source, selected_other), 0)\n graph.edge_features = torch.cat((graph.edge_features, feature), 0)\n graph.edge_types = torch.cat((graph.edge_types, type), 0)\n\n add_index += 1\n\n return graph, loss\n\n def _reset_paramters(self, state_size, aggregated_size, n_edge_dtypes):\n torch.nn.init.normal_(self.edge_init)\n xavier_init(self.f_addedge_aggregated, 1, state_size + aggregated_size, 1)\n xavier_init(self.f_addedge_new, 1, state_size + aggregated_size, 1)\n xavier_init(self.fs_layer1_target, 1, state_size * 2, n_edge_dtypes)\n xavier_init(self.fs_layer1_new, 1, state_size * 2, n_edge_dtypes)\n\n\nclass GraphGen(torch.nn.Module):\n def __init__(self, n_node_types, n_edge_types, state_size, pad_char=255, propagate_steps=2,\n n_max_nodes=None, n_max_edges=None, dropout=0.2):\n super().__init__()\n\n self.aggregated_size = state_size * 2\n self.state_size = state_size\n\n self.n_max_nodes = n_max_nodes\n\n self.edge_adder = EdgeAdder(state_size, self.aggregated_size, n_edge_types, pad_char, propagate_steps, n_max_edges, dropout)\n self.node_adder = NodeAdder(state_size, self.aggregated_size, propagate_steps, n_node_types, pad_char, dropout)\n\n def forward(self, ref_output, batch_size=None, device=None):\n assert ((ref_output is None) and (batch_size is not None and device is not None)) or \\\n ((ref_output is not None) and (batch_size is None and device is None)), \\\n \"To generate, pass batch_size and device, to train, pass ref_output only.\"\n\n n_batch = ref_output[0].shape[0] if batch_size is None else batch_size\n device = ref_output[0].device if device is None else device\n\n loss = 0\n\n graph = Graph(n_batch, self.state_size, device)\n\n i = 0\n while True:\n if self.n_max_nodes is not None and self.n_max_nodes <= i//2:\n break\n\n graph, l_node = self.node_adder(graph, ref_output[i] if ref_output is not None else None)\n loss = loss + l_node\n\n if not graph.running.any():\n break\n\n graph, l_edge = self.edge_adder(graph, ref_output[i+1] if ref_output is not None else None)\n loss = loss + l_edge\n\n i+=2\n\n return graph, loss\n\n def generate(self, batch_size: int, device: torch.device):\n return self(None, batch_size, device)[0]\n", "sub_path": "graphgen.py", "file_name": "graphgen.py", "file_ext": "py", "file_size_in_byte": 18506, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.zeros", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 28, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 30, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 30, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 31, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 31, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 32, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 33, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.ones", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.uint8", "line_number": 37, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.is_tensor", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 52, "usage_type": "name"}, {"api_name": "torch.where", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.full", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn.functional.cross_entropy", "line_number": 68, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 68, "usage_type": "name"}, {"api_name": "torch.where", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.functional.binary_cross_entropy_with_logits", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 75, "usage_type": "name"}, {"api_name": "torch.sigmoid", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 80, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 85, "usage_type": "call"}, {"api_name": "torch.nn.init.normal_", "line_number": 89, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 89, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 92, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 96, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 96, "usage_type": "attribute"}, {"api_name": "torch.nn.Sequential", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 98, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 98, "usage_type": "attribute"}, {"api_name": "torch.nn.Sigmoid", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "attribute"}, {"api_name": "torch.nn.Parameter", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn.Dropout", "line_number": 104, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 104, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 114, "usage_type": "call"}, {"api_name": "torch.float32", "line_number": 114, "usage_type": "attribute"}, {"api_name": "torch.mm", "line_number": 120, "usage_type": "call"}, {"api_name": "torch.nn.init.normal_", "line_number": 130, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 130, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "attribute"}, {"api_name": "torch.nn.GRUCell", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 139, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 143, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 143, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 144, "usage_type": "attribute"}, {"api_name": "torch.nn.Sequential", "line_number": 146, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 146, "usage_type": "attribute"}, {"api_name": "torch.nn.Tanh", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 147, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 148, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 148, "usage_type": "attribute"}, {"api_name": "torch.nn.Dropout", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 151, "usage_type": "attribute"}, {"api_name": "torch.ByteTensor", "line_number": 155, "usage_type": "attribute"}, {"api_name": "torch.ByteTensor", "line_number": 158, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.where", "line_number": 182, "usage_type": "call"}, {"api_name": "torch.nn.init.calculate_gain", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 187, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "attribute"}, {"api_name": "torch.nn.ModuleList", "line_number": 200, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 200, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 218, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 218, "usage_type": "attribute"}, {"api_name": "torch.nn.Parameter", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 220, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 220, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 222, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 223, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 223, "usage_type": "attribute"}, {"api_name": "torch.ByteTensor", "line_number": 226, "usage_type": "attribute"}, {"api_name": "torch.arange", "line_number": 251, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 251, "usage_type": "attribute"}, {"api_name": "torch.zeros", "line_number": 253, "usage_type": "call"}, {"api_name": "torch.long", "line_number": 253, "usage_type": "attribute"}, {"api_name": "torch.where", "line_number": 257, "usage_type": "call"}, {"api_name": "torch.nn.functional.one_hot", "line_number": 261, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 261, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 263, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.nn.init.normal_", "line_number": 270, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 270, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 276, "usage_type": "attribute"}, {"api_name": "torch.nn.Parameter", "line_number": 287, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 287, "usage_type": "attribute"}, {"api_name": "torch.Tensor", "line_number": 287, "usage_type": "call"}, {"api_name": "torch.nn.Linear", "line_number": 290, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 290, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 291, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 291, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 293, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 293, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 294, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 294, "usage_type": "attribute"}, {"api_name": "torch.zeros_like", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 371, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 372, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 373, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 374, "usage_type": "call"}, {"api_name": "torch.nn.init.normal_", "line_number": 381, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 381, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 388, "usage_type": "attribute"}, {"api_name": "torch.device", "line_number": 431, "usage_type": "attribute"}]} +{"seq_id": "121587674", "text": "# coding: utf-8\n# Copyright 2013 The Font Bakery Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# See AUTHORS.txt for the list of Authors and LICENSE.txt for the License.\n#pylint:disable-msg=W0612\nimport logging\nimport os\nimport os.path as op\n\nfrom flask import Flask, request, render_template, g, session\nimport logging.handlers\n\napp = Flask(__name__, static_folder=os.path.join(\n os.path.dirname(__file__), '..', 'static'))\napp.config.from_object('bakery.config')\napp.config.from_pyfile(os.path.realpath(op.join(op.dirname(__file__), 'local.cfg')), silent=True)\n\nfrom flask.ext.sqlalchemy import SQLAlchemy\ndb = SQLAlchemy(app)\n\nfrom flask.ext.mail import Mail\nmail = Mail(app)\n\nfrom rauth.service import OAuth2Service\n\ngithub = OAuth2Service(\n name='github',\n base_url='https://api.github.com/',\n access_token_url='https://github.com/login/oauth/access_token',\n authorize_url='https://github.com/login/oauth/authorize',\n client_id=app.config['GITHUB_CONSUMER_KEY'],\n client_secret=app.config['GITHUB_CONSUMER_SECRET']\n)\n\nfrom flask_flatpages import FlatPages\npages = FlatPages(app)\n\nfrom flask.ext.rq import RQ\nrq = RQ(app)\n\nfrom flask.ext.babel import Babel\nbabel = Babel(app)\n\n\nclass SMTPHandler(logging.handlers.SMTPHandler):\n\n def emit(self, record):\n if not app.config.get('MANDRILL_KEY'):\n return\n from flask import request\n message = render_template('exception.txt',\n request=request,\n stacktrace=self.format(record),\n current_user=g.user)\n send_mail(self.getSubject(record), message)\n\n\ndef linebreaks(value):\n \"\"\"Converts newlines into

and
s.\"\"\"\n import re\n from jinja2 import Markup\n value = re.sub(r'(\\r\\n)|\\r|\\n', '\\n', value)\n paras = re.split('\\n{2,}', value)\n paras = [u'

%s

' % p.replace('\\n', '
') for p in paras]\n paras = u'\\n\\n'.join(paras)\n return Markup(paras)\n\n\ndef send_mail(subject, message, recipients=[\"hash.3g@gmail.com\"]):\n from flask import current_app\n import mandrill\n with current_app.test_request_context('/'):\n request_msg = {\n \"html\": linebreaks(message),\n \"subject\": subject,\n \"from_email\": 'hash.3g@gmail.com',\n \"from_name\": \"Fontbakery\",\n \"to\": map(lambda x: {'email': x}, recipients),\n \"track_opens\": True,\n \"track_clicks\": True\n }\n\n m = mandrill.Mandrill(app.config['MANDRILL_KEY'])\n m.messages.send(request_msg)\n\n\ngm = SMTPHandler(\n (\"smtp.gmail.com\", 587),\n 'hash.3g@gmail.com', ['hash.3g@gmail.com'],\n '[ERROR] FontBakery has been crashed!')\ngm.setLevel(logging.ERROR)\n\napp.logger.addHandler(gm)\n\n\n@app.errorhandler(403)\ndef forbidden_page(error):\n return render_template(\"misc/403.html\"), 403\n\n\n@app.errorhandler(404)\ndef page_not_found(error):\n return render_template(\"misc/404.html\"), 404\n\n\n@app.errorhandler(500)\ndef server_error_page(error):\n return render_template(\"misc/500.html\"), 500\n\n\n@app.before_request\ndef guser():\n g.user = None\n if 'user_id' in session:\n if session['user_id']:\n #pylint:disable-msg=E1101\n from gitauth.models import User\n user = User.query.get(session['user_id'])\n if user:\n g.user = user\n else:\n del session['user_id']\n\n\n@app.before_request\ndef gdebug():\n if app.debug:\n g.debug = True\n else:\n g.debug = False\n\n\n@babel.localeselector\ndef get_locale():\n if g.user:\n if hasattr(g.user, 'ui_lang'):\n return g.user.ui_lang\n\n accept_languages = app.config.get('ACCEPT_LANGUAGES')\n return request.accept_languages.best_match(accept_languages)\n\n# iohandler = StreamHandler()\n# iohandler.setLevel(logging.WARNING)\n# app.logger.addHandler(iohandler)\n\n\ndef register_blueprints(app):\n from .gitauth import gitauth\n from .frontend import frontend\n from .realtime import realtime\n from .api import api\n from .settings import settings\n from .project import project\n app.register_blueprint(gitauth)\n app.register_blueprint(frontend)\n app.register_blueprint(realtime)\n app.register_blueprint(settings)\n app.register_blueprint(api)\n app.register_blueprint(project)\n\n\ndef register_filters(app):\n # Additional Jinja filters\n from utils import pretty_date, signify\n\n app.jinja_env.filters['pretty_date'] = pretty_date\n app.jinja_env.filters['signify'] = signify\n app.jinja_env.globals['app_version'] = git_info\n app.jinja_env.add_extension('jinja2.ext.do')\n\n\ndef git_info():\n \"\"\" If application is under git then return commit's hash\n and timestamp of the version running.\n\n Return None if application is not under git.\"\"\"\n from .tasks import prun\n import json\n params = \"git log -n1\"\n fmt = \"\"\" --pretty=format:'{\"hash\":\"%h\", \"commit\":\"%H\",\"date\":\"%cd\"}'\"\"\"\n log = prun(params + fmt, cwd=app.config['ROOT'])\n try:\n return json.loads(log)\n except ValueError:\n return None\n\nregister_filters(app)\n", "sub_path": "bakery/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 5673, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.realpath", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path", "line_number": 28, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.ext.sqlalchemy.SQLAlchemy", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.ext.mail.Mail", "line_number": 34, "usage_type": "call"}, {"api_name": "rauth.service.OAuth2Service", "line_number": 38, "usage_type": "call"}, {"api_name": "flask_flatpages.FlatPages", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.ext.rq.RQ", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.ext.babel.Babel", "line_number": 54, "usage_type": "call"}, {"api_name": "logging.handlers", "line_number": 57, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 63, "usage_type": "call"}, {"api_name": "flask.request", "line_number": 64, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 66, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 66, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 74, "usage_type": "call"}, {"api_name": "re.split", "line_number": 75, "usage_type": "call"}, {"api_name": "jinja2.Markup", "line_number": 78, "usage_type": "call"}, {"api_name": "flask.current_app.test_request_context", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.current_app", "line_number": 84, "usage_type": "name"}, {"api_name": "mandrill.Mandrill", "line_number": 95, "usage_type": "call"}, {"api_name": "{'request': 'flask.request'}", "line_number": 99, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 103, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 110, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 115, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.g.user", "line_number": 125, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 125, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 126, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 127, "usage_type": "name"}, {"api_name": "gitauth.models.User.query.get", "line_number": 130, "usage_type": "call"}, {"api_name": "gitauth.models.User.query", "line_number": 130, "usage_type": "attribute"}, {"api_name": "gitauth.models.User", "line_number": 130, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 130, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 132, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 134, "usage_type": "name"}, {"api_name": "flask.g.debug", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.g.debug", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 147, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 148, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.g.user", "line_number": 149, "usage_type": "attribute"}, {"api_name": "flask.g", "line_number": 149, "usage_type": "name"}, {"api_name": "flask.request.accept_languages.best_match", "line_number": 152, "usage_type": "call"}, {"api_name": "flask.request.accept_languages", "line_number": 152, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 152, "usage_type": "name"}, {"api_name": "gitauth.gitauth", "line_number": 166, "usage_type": "argument"}, {"api_name": "frontend.frontend", "line_number": 167, "usage_type": "argument"}, {"api_name": "realtime.realtime", "line_number": 168, "usage_type": "argument"}, {"api_name": "settings.settings", "line_number": 169, "usage_type": "argument"}, {"api_name": "api.api", "line_number": 170, "usage_type": "argument"}, {"api_name": "project.project", "line_number": 171, "usage_type": "argument"}, {"api_name": "utils.pretty_date", "line_number": 178, "usage_type": "name"}, {"api_name": "utils.signify", "line_number": 179, "usage_type": "name"}, {"api_name": "tasks.prun", "line_number": 193, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 195, "usage_type": "call"}]} +{"seq_id": "367046373", "text": "\nfrom collections import defaultdict\nfrom dataclasses import dataclass\nfrom datetime import datetime\nfrom email.policy import default\nfrom io import BytesIO\nfrom pathlib import Path\nfrom typing import Dict, List\nfrom zipfile import ZipFile\nimport requests\nimport os\nimport sys\n\n@dataclass\nclass Sample:\n name: str\n input: str\n answer: str\n\ndef main_file_content(name:str) -> str:\n return f\"\"\"#{name}\nfrom collections import *\nfrom typing import *\nimport io, os, sys\n_fast_readline = io.BytesIO(os.read(0, 2147483647)).readline\n_readline = lambda: _fast_readline().decode()\nread_line = lambda: _readline().strip()\nread_list = lambda: _readline().strip().split(' ')\nread_int = lambda: int(_readline())\nread_int_list = lambda: [int(x) for x in read_list()]\nwrite = lambda *msg: sys.stdout.write(' '.join(list(map(str, msg))))\nwriteln = lambda *msg: sys.stdout.write(' '.join(list(map(str, msg))) + '\\\\n')\n\ndef main() -> None:\n for line in sys.stdin:\n value = int(line)\n\nif __name__ == '__main__':\n main()\"\"\"\n\n\ndef test_file_content(name:str, test_cases:str):\n return f\"\"\"from pathlib import Path\nfrom subprocess import Popen, PIPE\nimport os\nimport unittest\n\nclass {name.capitalize()}Tests(unittest.TestCase):\n def _run_test_file(self, in_data: str) -> str:\n p = Popen(['python3', Path(os.path.dirname(os.path.abspath(__file__))) / Path('main.py')], stdout=PIPE, stdin=PIPE, stderr=PIPE)\n stdout, stderr = list(map(lambda x: x.decode().replace('\\\\r\\\\n', '\\\\n').strip(), p.communicate(input=in_data.encode())))\n if (stderr):\n print(stderr)\n return stdout\n {test_cases}\nif __name__ == '__main__':\n unittest.main()\"\"\"\n\ndef write_file(filename:str, content:str) -> None:\n with open(filename, mode='w+', encoding='utf-8') as file:\n file.write(content)\n\ndef test_cases_content(samples: List[Sample]) -> str:\n test_cases_code = \"\"\n for sample in samples:\n test_cases_code += f\"\"\"\n def test_{sample.name}(self) -> None:\n in_data = '''\\n{sample.input}'''.strip()\n ans_data = '''\\n{sample.answer}'''.strip()\n result = self._run_test_file(in_data)\n self.assertEqual(ans_data, result, '\\\\nExpected:\\\\n' + ans_data + '\\\\n\\\\nActual:\\\\n' + result)\"\"\"\n return test_cases_code\n\ndef download_samples(problem_name:str) -> List[Sample]:\n sample_url = f\"https://open.kattis.com/problems/{problem_name}/file/statement/samples.zip\"\n response = requests.get(sample_url).content\n test_cases = defaultdict(dict)\n with ZipFile(BytesIO(response)) as zipfile:\n for i in zipfile.filelist:\n data = zipfile.read(i.filename).decode(\"utf-8\")\n test_cases[i.filename.rsplit('.', 1)[0]][i.filename.rsplit('.', 1)[1]] = data\n return [Sample(kvp[0], kvp[1]['in'], kvp[1]['ans']) for kvp in test_cases.items()]\n\ndef write_samples_to_files(problem_name: str, samples: List[Sample]) -> None:\n for sample in samples:\n with open(Path(problem_name) / Path(sample.name + '.in'), 'w+') as file:\n file.write(sample.input)\n with open(Path(problem_name) / Path(sample.name + '.ans'), 'w+') as file:\n file.write(sample.answer)\n\n\ndef setup_problem(problem_name: str) -> None:\n os.makedirs(problem_name, exist_ok=True)\n\n samples = download_samples(problem_name)\n write_samples_to_files(problem_name, samples)\n write_file(Path(problem_name) / Path('tests.py'), test_file_content(problem_name, test_cases_content(samples)))\n main_path = Path(problem_name) / Path('main.py')\n if not main_path.is_file():\n write_file(main_path, main_file_content(problem_name))\n\n print(problem_name)\n\ndef main(args: List[str]) -> None:\n for problem in args[1:]:\n setup_problem(problem)\n\nif __name__ == '__main__':\n main(sys.argv)", "sub_path": "kattis/new.py", "file_name": "new.py", "file_ext": "py", "file_size_in_byte": 3808, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "dataclasses.dataclass", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 63, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 76, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 77, "usage_type": "call"}, {"api_name": "zipfile.ZipFile", "line_number": 78, "usage_type": "call"}, {"api_name": "io.BytesIO", "line_number": 78, "usage_type": "call"}, {"api_name": "zipfile.filelist", "line_number": 79, "usage_type": "attribute"}, {"api_name": "zipfile.read", "line_number": 80, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 74, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 84, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 86, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 88, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 93, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 97, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 98, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 104, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 109, "usage_type": "attribute"}]} +{"seq_id": "59982482", "text": "'''\n`ReFPloT/refplotepic.py`:\n\nThis Plotting tool takes in a GLImER receiver function `.mat` file as input and\noutputs a a figure that bins and plots receiver functions as a function of \nepicentral distance. Sliders change colorbar and the Gaussian filters standard \ndeviation. \n\nAuthor: Lucas Sawade\nDate: Mar 21, 2019\n\n'''\n\nimport numpy as np\nfrom scipy.io import loadmat\nfrom scipy.interpolate import interp2d\nfrom scipy import signal\nfrom scipy.ndimage.filters import convolve,gaussian_filter\nimport matplotlib.pyplot as plt\nimport matplotlib.gridspec as Gridspec\nfrom matplotlib.widgets import Slider, Button, RadioButtons\nimport argparse\n\n\n\ndef main(filename, colormap, windowi, windowf, lf, hf):\n \"\"\"\n This function takes in a `.mat` file of GLImER Receiver Functions and plots\n them as a function of epicentral distance.\n :param filename: string defining the RF filepath\n :param colormap: string defines the colormap\n :param window: 2 Element list defining the time-window of interest\n :param lf: float, low-cut frequency\n :param hf: float, high-cut frequency\n\n \"\"\"\n\n # ----- Define file to be loaded -----\n RF = loadmat(filename)\n\n # ----- Loading the necessary variables -----\n rf = RF['rf'] # Receiver Functions\n delta = RF['rdelta'][0] # Epicentral distance\n baz = RF['rbaz'][0] # Back-azimuth\n dt = RF['dt'][0] # Sampling Time\n\n\n # ----- Filtering the RFs -----\n def bandpass(x, dt, lf, hf):\n \"\"\"\n Takes in a signal in form of numpy array (multiple rows of signals if it is a matrix) and filters it. The filter\n makes use of the new sos output type to reduce numerical error associated with the traditional polynomial\n representation (b,a).\n\n :param x: numpy row matrix of timeseries\n :param dt: sampling interval\n :param lf: low frequency cut off\n :param hf: high frequency cut off\n :return: numpy array of same size as x, but filtered.\n\n \"\"\"\n\n # Filter Design\n nyq = 0.5 / dt\n wn = [lf / nyq, hf / nyq] # Critical normalized frequencies.\n sos = signal.butter(2, wn, btype='bandpass', output='sos')\n\n # Filtering\n m,n = x.shape\n y = np.zeros((m, n))\n\n for k in range(m):\n y[k, :] = signal.sosfiltfilt(sos, x[k, :])\n\n return y\n\n rf = bandpass(rf, dt, lf, hf)\n\n\n # ----- Parameters -----\n # Number of RFs in array\n N = len(delta)\n\n # Min/Max edpicentral distance\n mindelta = np.min(delta)\n maxdelta = np.max(delta)\n\n # Create time Vector with dt\n t0 = -30 # P-wave arrival is set to 30sec\n tf = 120\n t = np.arange(t0, tf, dt)\n\n # ----- Binning -----\n # Setup bins\n delta_bin_width = 5 # in degrees\n delta_edges = np.arange(mindelta, maxdelta + delta_bin_width, delta_bin_width)\n delta_n_bins = len(delta_edges) - 1\n delta_centres = delta_edges[:-1] + np.diff(delta_edges)/2\n\n # Bin\n delta_binned = np.digitize(delta, delta_edges)-1 # -1 one because the bins are numbered form 1 to N and not 0 to N-1\n delta_hist, _ = np.histogram(delta, bins=delta_edges)\n\n\n # Preallocated summed RFs\n delta_RFs = np.zeros((delta_n_bins, 1500))\n\n # Sum RFs within a bin\n for k in range(len(delta_binned)):\n delta_RFs[delta_binned[k], :] = delta_RFs[delta_binned[k], :] + rf[k, :]\n\n\n # Normalize the bins\n for k in range(delta_n_bins):\n delta_RFs[k, :] = delta_RFs[k, :]/delta_hist[k]\n\n\n # ----- Interpolation -----\n # Interpolation grid\n delta_new = np.linspace(np.min(delta_centres), np.max(delta_centres), 300)\n\n # Actual interpolation\n RF_interpolator = interp2d(delta_centres,t, np.transpose(delta_RFs), kind='linear')\n RF_raw = RF_interpolator(delta_new, t)\n\n\n # ----- Spatial Filter -----\n def disk_filter(r):\n \"\"\"\n Creates a diskfilter matrix as a function of r matrix entries\n :param r: integer\n :return:\n \"\"\"\n\n y, x = np.ogrid[-r: r + 1, -r: r + 1]\n disk = x ** 2 + y ** 2 <= r ** 2\n return disk.astype(float)\n\n # RF_PLOT = signal.convolve(RF_raw, disk_filter(0), mode='same')\n sigma = 2\n RF_PLOT = gaussian_filter(RF_raw, sigma, mode='constant')\n\n # ----- Plotting -----\n\n # Plot Input\n colormap_name = colormap\n time_window = np.array([windowi, windowf])\n\n # Pre-Plot Setup\n plt.rc('text', usetex=True)\n plt.rc('font', family='serif')\n\n # MinMax of stuff\n ampMax = np.max(np.abs(RF_PLOT)) # Needed for colorbar\n delta_plot_max = np.max(delta_centres)\n delta_plot_min = np.min(delta_centres)\n\n # Setting up the Figure\n fig = plt.figure(figsize=(7, 9))\n\n # Subplot layout\n gs = Gridspec.GridSpec(nrows=4, ncols=20, left=0.15, right=0.9, bottom=0.2, top= 0.9, wspace=0.1, figure=fig)\n\n hist_ax = fig.add_subplot(gs[0, 0:19])\n surf_ax = fig.add_subplot(gs[1:20, 0:19])\n col_ax = fig.add_subplot(gs[1:20, 19])\n\n Bbox = surf_ax.get_position()\n axpos = Bbox.bounds\n\n # Histogramsetup\n hist_ax.set_xlim([delta_plot_min, delta_plot_max])\n hist_ax.hist(delta, delta_edges)\n hist_ax.set_ylabel('Counts')\n\n # Surface axis setup\n surf_ax.axis([delta_plot_min, delta_plot_max, time_window[0], time_window[1]])\n surf_ax.invert_yaxis()\n surf_ax.set_xlabel('Epicentral Distance $\\Delta$ in [$^\\circ$]')\n s = surf_ax.pcolormesh(delta_new, t, RF_PLOT, cmap=colormap_name)\n\n\n # Colorbar\n colbar = plt.colorbar(s, cax=col_ax)\n\n # Create Slider position and Style:\n sliderc_color = 'lightgoldenrodyellow'\n slidercmin = plt.axes([axpos[0]+axpos[2]-0.25, 0.11, 0.25, 0.03], facecolor=sliderc_color)\n slidercmax = plt.axes([0.15, 0.11, 0.25, 0.03], facecolor=sliderc_color)\n slidercsym = plt.axes([0.15, 0.06, 0.25, 0.03], facecolor=sliderc_color)\n sliderfilt = plt.axes([0.15, 0.01, axpos[2], 0.03], facecolor=sliderc_color)\n\n\n # Create Slider low lim, upper lim, Initial Val, ValueFormat\n vmin, vmax = s.get_clim()\n scolormin = Slider(slidercmin, 'ColorMin', -3 * np.max(np.abs(RF_PLOT)), 0, valinit=vmin, valfmt='%1.2E')\n scolormax = Slider(slidercmax, 'ColorMax', 0, 3*np.max(np.abs(RF_PLOT)), valinit=vmax, valfmt='%1.2E')\n scolorsym = Slider(slidercsym, 'ColorSym', 0, 3 * np.max(np.abs(RF_PLOT)), valinit=(vmax-vmin)/2, valfmt='%1.2E')\n sfilt = Slider(sliderfilt, 'Gauss: $\\sigma$', 0, 50, valinit=0, valfmt='%2.1f')\n\n\n def update_min(val):\n # Set new min color\n _cmin = scolormin.val\n _, _cmax = s.get_clim()\n\n # Update figure\n s.set_clim([_cmin, _cmax])\n fig.canvas.draw_idle()\n\n def update_max(val):\n # Set new min color\n _cmin, _ = s.get_clim()\n _cmax = scolormax.val\n\n # Update figure\n s.set_clim([_cmin, _cmax])\n fig.canvas.draw_idle()\n\n def update_sym(val):\n # Set new min color\n _cmin = -scolorsym.val\n _cmax = scolorsym.val\n\n # Update the min/max sliders\n scolormin.set_val(_cmin)\n scolormax.set_val(_cmax)\n\n # Update figure\n s.set_clim([_cmin, _cmax])\n fig.canvas.draw_idle()\n\n def update_filt(val):\n # Find number of RF round so the value is not decimal\n sigma = sfilt.val-1\n\n # Filtering\n if sigma == 0:\n s.set_array(RF_raw[:-1, :-1].ravel())\n else:\n s.set_array(gaussian_filter(RF_raw, sigma, mode='constant')[:-1, :-1].ravel())\n\n # Update Figure\n fig.canvas.draw_idle()\n\n scolormin.on_changed(update_min)\n scolormax.on_changed(update_max)\n scolorsym.on_changed(update_sym)\n sfilt.on_changed(update_filt)\n\n\n resetax = plt.axes([axpos[0]+axpos[2]-0.1, 0.06, 0.1, 0.03])\n button = Button(resetax, 'Reset', color=sliderc_color, hovercolor='0.975')\n\n def reset(event):\n scolormax.reset()\n scolormin.reset()\n scolorsym.reset()\n\n button.on_clicked(reset)\n\n # Possible colors for plot\n colors = ('jet', 'seismic')\n\n # get index of color entry\n index = [i for i, x in enumerate([y == colormap for y in colors]) if x]\n\n # Create color radio\n button_color = sliderc_color\n rax = plt.axes([0.015, 0.25, 0.09, 0.1], facecolor=button_color)\n radio = RadioButtons(rax, colors, active=index[0])\n\n def colorfunc(label):\n s.set_cmap(label)\n fig.canvas.draw_idle()\n radio.on_clicked(colorfunc)\n\n plt.show()\n\nif __name__ == \"__main__\":\n\n parser = argparse.ArgumentParser()\n parser.add_argument(\"filename\", help=\"Filename of the file of to be loaded and displayed\",type=str)\n parser.add_argument(\"-c\", \"--color\", help=\"Colormap name, e.g. \"\"jet\"\"\",\n type=str, default='jet')\n parser.add_argument(\"-wi\", \"--WindowStart\",\n help=\"Time window of interest,e.g., 0\",\n type=float, default=0)\n parser.add_argument(\"-wf\", \"--WindowEnd\",\n help=\"Timew indow of interest,e.g., 30\",\n type=float, default=80)\n parser.add_argument(\"-lf\", \"--LowCut\", help=\"The low-cut frequency in Hz\",\n type=float, default=0)\n parser.add_argument(\"-hf\", \"--HighCut\", help=\"The high-cut frequency in Hz\",\n type=float, default=1.5)\n args = parser.parse_args()\n\n main(args.filename, args.color, args.WindowStart, args.WindowEnd, args.LowCut, args.HighCut)\n", "sub_path": "refplotepic.py", "file_name": "refplotepic.py", "file_ext": "py", "file_size_in_byte": 9484, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "scipy.io.loadmat", "line_number": 39, "usage_type": "call"}, {"api_name": "scipy.signal.butter", "line_number": 66, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 66, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 70, "usage_type": "call"}, {"api_name": "scipy.signal.sosfiltfilt", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.signal", "line_number": 73, "usage_type": "name"}, {"api_name": "numpy.min", "line_number": 85, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 96, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.digitize", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.histogram", "line_number": 102, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 120, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp2d", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.ogrid", "line_number": 135, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.filters.gaussian_filter", "line_number": 141, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 147, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 150, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 150, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 151, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 151, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 155, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 156, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 159, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 159, "usage_type": "name"}, {"api_name": "matplotlib.gridspec.GridSpec", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.gridspec", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.colorbar", "line_number": 184, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 184, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 188, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 188, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 189, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 189, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 190, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 190, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 191, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 191, "usage_type": "name"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 196, "usage_type": "call"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 197, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 197, "usage_type": "call"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 198, "usage_type": "call"}, {"api_name": "matplotlib.widgets.Slider", "line_number": 199, "usage_type": "call"}, {"api_name": "scipy.ndimage.filters.gaussian_filter", "line_number": 241, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.widgets.Button", "line_number": 253, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.axes", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}, {"api_name": "matplotlib.widgets.RadioButtons", "line_number": 271, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 278, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 278, "usage_type": "name"}, {"api_name": "argparse.ArgumentParser", "line_number": 282, "usage_type": "call"}]} +{"seq_id": "138629448", "text": "import os\n\nimport librosa\nimport numpy as np\nimport soundfile as sf\nimport torch\nfrom scipy import interpolate\nimport math\n\n\ndef crop_center(h1, h2, concat=True):\n # s_freq = (h2.shape[2] - h1.shape[2]) // 2\n # e_freq = s_freq + h1.shape[2]\n h1_shape = h1.size()\n h2_shape = h2.size()\n if h2_shape[3] < h1_shape[3]:\n raise ValueError('h2_shape[3] must be greater than h1_shape[3]')\n s_time = (h2_shape[3] - h1_shape[3]) // 2\n e_time = s_time + h1_shape[3]\n h2 = h2[:, :, :, s_time:e_time]\n if concat:\n return torch.cat([h1, h2], dim=1)\n else:\n return h2\n\n\ndef calc_spec(X, hop_length):\n n_fft = (hop_length - 1) * 2\n audio_left = np.asfortranarray(X[0])\n audio_right = np.asfortranarray(X[1])\n spec_left = librosa.stft(audio_left, n_fft, hop_length=hop_length)\n spec_right = librosa.stft(audio_right, n_fft, hop_length=hop_length)\n spec = np.asfortranarray([spec_left, spec_right])\n\n return spec\n\n\ndef mask_uninformative(mask, ref, thres=0.3, min_range=64, fade_area=32):\n if min_range < fade_area * 2:\n raise ValueError('min_range must be >= fade_area * 2')\n idx = np.where(ref.mean(axis=(0, 1)) < thres)[0]\n starts = np.insert(idx[np.where(np.diff(idx) != 1)[0] + 1], 0, idx[0])\n ends = np.append(idx[np.where(np.diff(idx) != 1)[0]], idx[-1])\n uninformative = np.where(ends - starts > min_range)[0]\n if len(uninformative) > 0:\n starts = starts[uninformative]\n ends = ends[uninformative]\n old_e = None\n for s, e in zip(starts, ends):\n if old_e is not None and s - old_e < fade_area:\n s = old_e - fade_area * 2\n elif s != 0:\n start_mask = mask[:, :, s:s + fade_area]\n np.clip(\n start_mask + np.linspace(0, 1, fade_area), 0, 1,\n out=start_mask)\n if e != mask.shape[2]:\n end_mask = mask[:, :, e - fade_area:e]\n np.clip(\n end_mask + np.linspace(1, 0, fade_area), 0, 1,\n out=end_mask)\n mask[:, :, s + fade_area:e - fade_area] = 1\n old_e = e\n\n return mask\n\ndef _lcm(x, y):\n return x * y // math.gcd(x, y)\n\ndef _resample(audio,desired_sample_count):\n # upsample by interpolate\n desired_sample_count = int(desired_sample_count)\n x_original = np.linspace(0,audio.shape[1]-1,audio.shape[1])\n left = interpolate.interp1d(x_original, audio[0], kind='cubic')\n right = interpolate.interp1d(x_original, audio[1], kind='cubic')\n\n x_desired = np.linspace(0,audio.shape[1]-1,desired_sample_count)\n resampled_left = left(x_desired)\n resampled_right = right(x_desired)\n\n return np.vstack([resampled_left,resampled_right])\n\ndef align_wave_head_and_tail(a, b, sr,clip_duration=8,sample_multiple=8,offset = 0):\n a_backup = a\n b_backup = b\n print(\"upsample...\")\n a=_resample(a,a.shape[1]*sample_multiple)\n b = _resample(b, b.shape[1] * sample_multiple)\n\n print(\"finding alignment...\")\n\n delta = 1000\n while delta>999:\n a_mono_front = a[:, sr*offset:sr * (clip_duration + offset)].sum(axis=0)\n b_mono_front = b[:, sr*offset:sr * (clip_duration + offset)].sum(axis=0)\n a_mono_front -= a_mono_front.mean()\n b_mono_front -= b_mono_front.mean()\n offset_front = len(a_mono_front) - 1\n\n back_end = min(a.shape[1], b.shape[1]) - sr * offset\n a_mono_back = a[:, back_end - (sr * clip_duration):back_end].sum(axis=0)[::-1]\n b_mono_back = b[:, back_end - (sr * clip_duration):back_end].sum(axis=0)[::-1]\n a_mono_back -= a_mono_back.mean()\n b_mono_back -= b_mono_back.mean()\n offset_back = len(a_mono_back) - 1\n delay_front = np.argmax(np.correlate(a_mono_front, b_mono_front, 'full')) - offset_front\n delay_back = offset_back - np.argmax(np.correlate(a_mono_back, b_mono_back, 'full'))\n print(\"delay_front:%d,delay_back:%d\"%(delay_front,delay_back))\n delta = delay_back - delay_front\n if delta >1000:\n clip_duration += 1\n offset += 1\n\n if delay_front > 0:\n a = a[:, delay_front:]\n else:\n b = b[:, np.abs(delay_front):]\n\n print(\"aligning...\")\n # adjust speed to match\n if abs(delta) == 1:\n pass\n else:\n if delta > 0:\n # make a shorter\n a = _resample(a , a.shape[1] - delta)\n # b = _resample(b,b.shape[1])\n\n pass\n else:\n # make b shorter\n b = _resample(b, b.shape[1] + delta)\n # a = _resample(a, a.shape[1] )\n pass\n\n if a.shape[1] < b.shape[1]:\n b = b[:, :a.shape[1]]\n else:\n a = a[:, :b.shape[1]]\n\n print(\"downsample...\")\n a=_resample(a,a.shape[1]/sample_multiple)\n b = _resample(b, b.shape[1] / sample_multiple)\n return a, b\n\n\ndef cache_or_load(mix_path, inst_path, sr, hop_length):\n _, mix_ext = os.path.splitext(mix_path)\n _, inst_ext = os.path.splitext(inst_path)\n spec_mix_path = mix_path.replace(mix_ext, '.npy')\n spec_inst_path = inst_path.replace(inst_ext, '.npy')\n\n if os.path.exists(spec_mix_path) and os.path.exists(spec_inst_path):\n X = np.load(spec_mix_path)\n y = np.load(spec_inst_path)\n else:\n X, _ = librosa.load(\n mix_path, sr, False, dtype=np.float32, res_type='kaiser_fast')\n y, _ = librosa.load(\n inst_path, sr, False, dtype=np.float32, res_type='kaiser_fast')\n X, _ = librosa.effects.trim(X)\n y, _ = librosa.effects.trim(y)\n X, y = align_wave_head_and_tail(X, y, sr)\n\n X = np.abs(calc_spec(X, hop_length))\n y = np.abs(calc_spec(y, hop_length))\n\n _, ext = os.path.splitext(mix_path)\n np.save(spec_mix_path, X)\n np.save(spec_inst_path, y)\n\n return X, y\n\ndef spec_to_wav(mag, phase, hop_length):\n spec = mag * phase\n spec_left = np.asfortranarray(spec[0])\n spec_right = np.asfortranarray(spec[1])\n wav_left = librosa.istft(spec_left, hop_length=hop_length)\n wav_right = librosa.istft(spec_right, hop_length=hop_length)\n wav = np.asfortranarray([wav_left, wav_right])\n\n return wav\n\n\ndef batch_generate_dataset(sr=44100,sour='preprocess_input',dest='preprocess_output'):\n flist = []\n for i in os.listdir(os.path.join(sour,'mix')):\n if i.lower().split('.')[-1] in ('wav','flac','mp3'):\n flist.append(i)\n\n for i in flist:\n print(\"processing %s\" % i)\n mix, _ = librosa.load(\n os.path.join(sour,\"mix\",i), sr, False, dtype=np.float64, res_type='kaiser_best')\n inst, _ = librosa.load(\n os.path.join(sour,\"instrument\",i), sr, False, dtype=np.float64, res_type='kaiser_best')\n mix, _ = librosa.effects.trim(mix)\n inst, _ = librosa.effects.trim(inst)\n inst_done, mix_done = align_wave_head_and_tail(inst, mix, sr)\n\n sf.write(os.path.join(dest,'instrument','%s.wav'%i.split(\".\")[0]), inst_done.T, sr)\n sf.write(os.path.join(dest,'mix','%s.wav'%i.split(\".\")[0]), mix_done.T, sr)\n sf.write(os.path.join(dest,'vocal','%s.wav'%i.split(\".\")[0]), (inst_done - mix_done).T, sr)\n pass\n\nif __name__ == \"__main__\":\n batch_generate_dataset(sys.argv[1],sys.argv[2],sys.argv[3]) # sample rate, source folder path, destination folder path \n # both of the folders should contain three folders: instrument, mix, vocal\n", "sub_path": "dummy-vocal-extractor.py", "file_name": "dummy-vocal-extractor.py", "file_ext": "py", "file_size_in_byte": 7453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "torch.cat", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 30, "usage_type": "call"}, {"api_name": "librosa.stft", "line_number": 31, "usage_type": "call"}, {"api_name": "librosa.stft", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 60, "usage_type": "call"}, {"api_name": "math.gcd", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 73, "usage_type": "call"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 74, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 74, "usage_type": "name"}, {"api_name": "scipy.interpolate.interp1d", "line_number": 75, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 75, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.correlate", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.correlate", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path", "line_number": 148, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 154, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 155, "usage_type": "call"}, {"api_name": "librosa.load", "line_number": 157, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 158, "usage_type": "attribute"}, {"api_name": "librosa.load", "line_number": 159, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 160, "usage_type": "attribute"}, {"api_name": "librosa.effects.trim", "line_number": 161, "usage_type": "call"}, {"api_name": "librosa.effects", "line_number": 161, "usage_type": "attribute"}, {"api_name": "librosa.effects.trim", "line_number": 162, "usage_type": "call"}, {"api_name": "librosa.effects", "line_number": 162, "usage_type": "attribute"}, {"api_name": "numpy.abs", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 168, "usage_type": "call"}, {"api_name": "os.path", "line_number": 168, "usage_type": "attribute"}, {"api_name": "numpy.save", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 170, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 177, "usage_type": "call"}, {"api_name": "librosa.istft", "line_number": 178, "usage_type": "call"}, {"api_name": "librosa.istft", "line_number": 179, "usage_type": "call"}, {"api_name": "numpy.asfortranarray", "line_number": 180, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "librosa.load", "line_number": 193, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 194, "usage_type": "call"}, {"api_name": "os.path", "line_number": 194, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 194, "usage_type": "attribute"}, {"api_name": "librosa.load", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 196, "usage_type": "call"}, {"api_name": "os.path", "line_number": 196, "usage_type": "attribute"}, {"api_name": "numpy.float64", "line_number": 196, "usage_type": "attribute"}, {"api_name": "librosa.effects.trim", "line_number": 197, "usage_type": "call"}, {"api_name": "librosa.effects", "line_number": 197, "usage_type": "attribute"}, {"api_name": "librosa.effects.trim", "line_number": 198, "usage_type": "call"}, {"api_name": "librosa.effects", "line_number": 198, "usage_type": "attribute"}, {"api_name": "soundfile.write", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 201, "usage_type": "call"}, {"api_name": "os.path", "line_number": 201, "usage_type": "attribute"}, {"api_name": "soundfile.write", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 202, "usage_type": "call"}, {"api_name": "os.path", "line_number": 202, "usage_type": "attribute"}, {"api_name": "soundfile.write", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 203, "usage_type": "call"}, {"api_name": "os.path", "line_number": 203, "usage_type": "attribute"}]} +{"seq_id": "461712998", "text": "import PySimpleGUI as sg\nfrom _manager import DbExcelFileManager as xlmng, PhpFileManager as phmng\nfrom _facade import BuilderFacade\n\"\"\"\nPySinpleGUIモジュール\n\"\"\"\n#\n# GUIコンポーネントの定義\n#\nlayout = [ \n [sg.Input(key='file_select', enable_events=True, disabled=True, size=(40,3), default_text='対象ファイルを選択してください'), sg.FileBrowse(button_text='参照…')],\n [sg.Checkbox(key='is_all', text='全シート出力', enable_events=True, disabled=True)],\n [sg.Input(key='sheet_search', enable_events=True, size=(48,3), background_color='#b2fcff', disabled=True)],\n [sg.Listbox(key='sheet_names', values=[], size=(48, 10), background_color='#ecfcff')],\n [sg.Submit( key='run', button_text='実行'), sg.Text(key='result', text='', size=(36,0), text_color='yellow', background_color='gray', justification='center')] \n]\n\n#\n# アプリ起動\n#\nwindow = sg.Window('pymigr', layout)\n\n# グローバルスコープに保持\nxl_manager = None # エクセルマネージャ\nsheet_names = [] # シート名一覧\ntemp_sheet_names = [] # 一時退避用シート名一覧\n\nwhile True: \n event, values = window.Read() \n \n if event == None: \n print('none')\n break\n\n if event == 'file_select':\n\n if values['file_select'] == '対象ファイルを選択してください':\n continue\n\n filepath = values['file_select']\n\n # エクセルファイル読み込み\n xl_manager = xlmng(filepath)\n sheet_names = xl_manager.get_sheet_name_list()\n temp_sheet_names = sheet_names\n\n # 読み込んだファイルのシートを一覧表示\n window.element('sheet_names').Update(sheet_names)\n\n # 入力解禁\n window.element('is_all').Update(disabled=False)\n window.element('sheet_search').Update(disabled=False)\n window.element('result').Update(background_color='#2f416d')\n\n if event == 'sheet_search':\n \n search_condition = values['sheet_search']\n # 前方一致検索で絞り込む\n if search_condition == '':\n window.element('sheet_names').Update(sheet_names)\n else:\n filterd_sheet_names = [s for s in sheet_names if s.startswith(search_condition)]\n window.element('sheet_names').Update(filterd_sheet_names)\n\n temp_sheet_names = window.element('sheet_names').GetListValues()\n\n if event == 'run':\n\n if len(values['sheet_names']) == 0:\n sg.Popup('エラー', 'ファイルを選択してください。')\n continue\n\n if (values['is_all'] is False) and (len(window.element('sheet_names').Widget.curselection()) == 0):\n sg.Popup('エラー', 'シートを選択するか、「全シート出力」をチェックしてから実行してください。')\n continue\n\n if values['is_all'] is True:\n # 全てのシートに対してファイル生成処理を実行\n\n for sheet_name in sheet_names:\n\n print(sheet_name)\n # 参照先のシートを設定\n xl_manager.set_source_sheet(sheet_name)\n target_sheet = xl_manager.get_sheet(sheet_name)\n\n # ソースコード生成\n source_code = BuilderFacade(xl_manager)\\\n .make_migration_file(sheet_name)\n\n php_manager = phmng()\n php_manager.output(sheet_name, source_code)\n\n # 処理結果\n window.element('result').Update(value='[success] all')\n else:\n # 個別のシートに対してファイル生成処理を実行\n\n # PySimpleGUIには選択中の値を取得するAPIが無い?\n # 暫定対応(https://github.com/PySimpleGUI/PySimpleGUI/issues/1633)\n # 選択中のindexを取得\n selected_index = window.element('sheet_names').Widget.curselection()[0]\n\n\n # 選択しているシート名を取得\n selected_sheet_name = temp_sheet_names[selected_index]\n\n # 参照先のシートを設定\n xl_manager.set_source_sheet(selected_sheet_name)\n target_sheet = xl_manager.get_sheet(selected_sheet_name)\n\n # ソースコード生成\n source_code = BuilderFacade(xl_manager)\\\n .make_migration_file(selected_sheet_name)\n\n php_manager = phmng()\n php_manager.output(selected_sheet_name, source_code)\n\n window.element('result').Update(value=f'[success] {selected_sheet_name}')\n\n\n if event == 'is_all':\n\n if (values['is_all']) is True:\n temp_sheet_names = window.element('sheet_names').GetListValues()\n window.element('sheet_names').Update(sheet_names)\n window.element('sheet_names').Update(disabled=True) # NOTE: disableのListboxにValueのUpdate不可\n window.element('sheet_search').Update(disabled=True)\n \n else:\n\n window.element('sheet_names').Update(disabled=False)\n window.element('sheet_names').Update(temp_sheet_names)\n window.element('sheet_search').Update(disabled=False)\n\n\n \n", "sub_path": "src/_layout.py", "file_name": "_layout.py", "file_ext": "py", "file_size_in_byte": 5302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PySimpleGUI.Input", "line_number": 11, "usage_type": "call"}, {"api_name": "PySimpleGUI.FileBrowse", "line_number": 11, "usage_type": "call"}, {"api_name": "PySimpleGUI.Checkbox", "line_number": 12, "usage_type": "call"}, {"api_name": "PySimpleGUI.Input", "line_number": 13, "usage_type": "call"}, {"api_name": "PySimpleGUI.Listbox", "line_number": 14, "usage_type": "call"}, {"api_name": "PySimpleGUI.Submit", "line_number": 15, "usage_type": "call"}, {"api_name": "PySimpleGUI.Text", "line_number": 15, "usage_type": "call"}, {"api_name": "PySimpleGUI.Window", "line_number": 21, "usage_type": "call"}, {"api_name": "_manager.DbExcelFileManager", "line_number": 43, "usage_type": "call"}, {"api_name": "PySimpleGUI.Popup", "line_number": 70, "usage_type": "call"}, {"api_name": "PySimpleGUI.Popup", "line_number": 74, "usage_type": "call"}, {"api_name": "_facade.BuilderFacade", "line_number": 88, "usage_type": "call"}, {"api_name": "_manager.PhpFileManager", "line_number": 91, "usage_type": "call"}, {"api_name": "_facade.BuilderFacade", "line_number": 113, "usage_type": "call"}, {"api_name": "_manager.PhpFileManager", "line_number": 116, "usage_type": "call"}]} +{"seq_id": "432509677", "text": "import pandas as pd\r\nimport matplotlib.pyplot as plt\r\nplt.rcParams[\"animation.convert_path\"] = \"C:\\Program Files\\ImageMagick-7.0.9-Q16\\magick.exe\" \r\nfrom matplotlib import animation\r\nimport random as rnd, numpy as np\r\nimport seaborn as sns;sns.set();\r\nnp.set_printoptions(suppress=True)\r\n\r\nclass kmeans_classifier:\r\n def __init__(self,df,k=None):\r\n self.df = df\r\n self.k, self.length = k, len(df)\r\n self.X = df.values\r\n self.dimensions = len(self.X[0])\r\n self.centroids = np.copy(self.X[np.random.choice(self.X.shape[0], \\\r\n self.k, replace=False), :])\r\n \r\n self.distances = np.zeros((2,self.length))\r\n \r\n self.objective_funtion = []\r\n self.fig = plt.figure()\r\n self.colors = []\r\n self.offsets = []\r\n def get_ssd(self,pi,p):\r\n ssd = 0\r\n for i in range(self.dimensions):\r\n ssd += (pi[i]- p[i])**2\r\n return ssd\r\n \r\n def get_distance(self):\r\n self.distances[0].fill(9999999999)\r\n self.distances[1].fill(0)\r\n obj = 0\r\n for i in range(self.k):\r\n for j in range(self.length):\r\n ssd = self.get_ssd(self.X[j],self.centroids[i])\r\n \r\n if self.distances[0][j]>ssd:\r\n self.distances[0][j] = ssd\r\n self.distances[1][j] = i\r\n \r\n self.objective_funtion.append(np.sum(self.distances[0]))\r\n return self.distances \r\n\r\n def get_centroids(self):\r\n cen_c = np.zeros(self.k,int)\r\n self.centroids.fill(0)\r\n for i in range(self.length):\r\n for j in range(self.dimensions):\r\n self.centroids[int(self.distances[1][i])][j] += self.X[i][j]\r\n cen_c[int(self.distances[1][i])] += 1\r\n \r\n for i in range(self.k):\r\n for j in range(self.dimensions):\r\n self.centroids[i][j] /= cen_c[i]\r\n return self.centroids\r\n \r\n def plotObjectiveFunction(self):\r\n index = [i for i in range(len(self.objective_funtion))]\r\n sns.scatterplot(index,self.objective_funtion,s = 50,edgecolor='k',**{'color':'r'})\r\n sns.lineplot(index,self.objective_funtion)\r\n plt.show()\r\n \r\n def clustering(self):\r\n self.get_distance()\r\n \r\n limit = 0\r\n while(limit!=50):\r\n limit += 1\r\n self.colors.append(np.copy(self.distances[1]))\r\n self.offsets.append(np.copy(self.centroids))\r\n self.get_centroids() \r\n self.get_distance()\r\n if self.objective_funtion[-1]==self.objective_funtion[-2]:\r\n break\r\n def setup_plot(self):\r\n self.scat = plt.scatter(x=self.X[:,0],y=self.X[:,1],s=50,\\\r\n c=self.colors[0],cmap='plasma',edgecolor='k')\r\n self.scat2 = plt.scatter(x=self.offsets[0][:,0],y=self.offsets[0][:,1],\\\r\n c='w',s=70,marker='X',edgecolor='k')\r\n \r\n return self.scat2,self.scat\r\n \r\n def update(self, i):\r\n self.scat.set_array(self.colors[i])\r\n self.scat2.set_offsets(self.offsets[i])\r\n \r\n return self.scat,self.scat2\r\n\r\ndef main():\r\n df = pd.read_csv(\"Iris.csv\")\r\n df = df[df.columns[:2]]\r\n \r\n clf = kmeans_classifier(df,3)\r\n clf.clustering()\r\n ani = animation.FuncAnimation(clf.fig,clf.update,frames=range(len(clf.colors)),\r\n init_func=clf.setup_plot,interval=500,repeat=True,blit=True)\r\n\r\n #plt.show()\r\n ani.save(\"test.gif\",writer=\"imagemagick\", extra_args=\"convert\")\r\n \r\n \r\n \r\n \r\nmain()\r\nprint('end')\r\n", "sub_path": "Kmeans/KmeansAnimate.py", "file_name": "KmeansAnimate.py", "file_ext": "py", "file_size_in_byte": 3676, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "matplotlib.pyplot.rcParams", "line_number": 3, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 3, "usage_type": "name"}, {"api_name": "seaborn.set", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.set_printoptions", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 15, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 46, "usage_type": "call"}, {"api_name": "seaborn.scatterplot", "line_number": 60, "usage_type": "call"}, {"api_name": "seaborn.lineplot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.copy", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.animation.FuncAnimation", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.animation", "line_number": 96, "usage_type": "name"}]} +{"seq_id": "631548608", "text": "import requests\nimport urllib.parse\n\n\nAPI_KEY = '9fafa6282710f17c0956691d1d6ac16a'\nmain_api = \"https://api.themoviedb.org/3/\"\n\n'''\nSearch results are in data['results'] as a list, iterate using:\nfor result in data['results']\nresult is a dict\n'''\n\n\ndef search(search):\n\n # https://api.themoviedb.org/3/search/multi?api_key=9fafa6282710f17c0956691d1d6ac16a&query=avengers\n\n url = main_api + 'search/' + 'multi?' + \\\n urllib.parse.urlencode({'api_key': API_KEY, 'query': search})\n response = requests.get(url)\n data = response.json()\n return data\n\n\ndef find_imdb(imdb_id):\n '''\n Takes in IMDb ID and returns JSON\n '''\n # https://api.themoviedb.org/3/find/tt0182576?api_key=9fafa6282710f17c0956691d1d6ac16a&external_source=imdb_id\n\n url = main_api + 'find/' + str(imdb_id) + '?' + urllib.parse.urlencode(\n {'api_key': API_KEY, 'external_source': 'imdb_id'})\n response = requests.get(url)\n data = response.json()\n return data\n\n\ndef view_results(data):\n '''\n Takes in data as a JSON object (dictionary) and returns the title or name\n '''\n for result in data['results']:\n if result['media_type'] == 'movie':\n print(result['title']) # for movies\n elif result['media_type'] == 'tv' or result['media_type'] == 'person':\n print(result['name'])\n\n\ndef view_details(media_type, media_id):\n '''\n Takes in media type ('tv', 'movie', ...) and returns JSON\n '''\n # https://api.themoviedb.org/3/movie/671?api_key=9fafa6282710f17c0956691d1d6ac16a\n\n url = main_api + media_type + '/' + \\\n str(media_id) + '?' + urllib.parse.urlencode({'api_key': API_KEY})\n response = requests.get(url)\n data = response.json()\n return data\n", "sub_path": "tmdb_api.py", "file_name": "tmdb_api.py", "file_ext": "py", "file_size_in_byte": 1736, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "urllib.parse.parse.urlencode", "line_number": 20, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 20, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 20, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 21, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 32, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 32, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 32, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 34, "usage_type": "call"}, {"api_name": "urllib.parse.parse.urlencode", "line_number": 57, "usage_type": "call"}, {"api_name": "urllib.parse.parse", "line_number": 57, "usage_type": "attribute"}, {"api_name": "urllib.parse", "line_number": 57, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 58, "usage_type": "call"}]} +{"seq_id": "349289901", "text": "from django.shortcuts import render, redirect\nfrom django.http import HttpResponse\nfrom todolist_app.models import TaskList\nfrom todolist_app.forms import TaskForm\nfrom django.contrib import messages\nfrom django.core.paginator import Paginator\n\n\n# Create your views here.\n\n\n\ndef index_function(request):\n\n context ={\n 'welcome_text': \"Welcome to the home page\",\n 'title_heading': \"Home page\",\n }\n\n return render(request, 'index.html', context)\n\n\n\ndef todolist_function(request):\n\n\n if request.method == \"POST\":\n form = TaskForm(request.POST or None)\n if form.is_valid():\n form.save()\n\n # alert message\n message_text = \"New task added successfully\"\n messages.success(request, (message_text))\n\n return redirect('todolist_link')\n else:\n all_task = TaskList.objects.all()\n '''value_limit_per_page = 5\n paginator = Paginator(all_task, value_limit_per_page)\n page = request.GET.get('pg')\n all_task = paginator.get_page(page)'''\n\n context = {\n 'welcome_text': \"welcome to todo list page\",\n 'title_heading': \"todo list\",\n 'all_task': all_task\n }\n return render(request, 'todolist_app.html', context)\n\n\ndef delete_task(request, task_id):\n task = TaskList.objects.get(pk=task_id)\n if task.delete():\n message_text_delete = \"delete successfully\"\n messages.success(request, (message_text_delete))\n\n return redirect('todolist_link')\n\n\n\n\ndef edit_task(request, task_id):\n if request.method == \"POST\":\n task = TaskList.objects.get(pk=task_id)\n form = TaskForm(request.POST or None,instance=task)\n if form.is_valid():\n form.save()\n\n massage_text = \"Task Edited Successfully\"\n messages.success(request, (massage_text))\n\n return redirect('todolist_link')\n else:\n edit_object = TaskList.objects.get(pk=task_id)\n context = {\n 'welcome_text': \"This is edit section\",\n 'title_heading': \"Edit list\",\n 'edit_object': edit_object\n }\n\n return render(request, 'edit.html', context)\n\ndef complete_task(request,task_id):\n task = TaskList.objects.get(pk=task_id)\n task.done = True\n task.save()\n\n return redirect('todolist_link')\n\ndef pending_task(request,task_id):\n task = TaskList.objects.get(pk=task_id)\n task.done = False\n task.save()\n\n return redirect('todolist_link')\n\n\n\ndef about_section(request):\n return render(request,'about_section.html')\n", "sub_path": "todolist_app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 2460, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.shortcuts.render", "line_number": 20, "usage_type": "call"}, {"api_name": "todolist_app.forms.TaskForm", "line_number": 28, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 34, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 34, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 36, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.all", "line_number": 38, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 38, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 38, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 49, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.get", "line_number": 53, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 53, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 53, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 56, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 56, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 58, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.get", "line_number": 65, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 65, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 65, "usage_type": "name"}, {"api_name": "todolist_app.forms.TaskForm", "line_number": 66, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 71, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 71, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 73, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.get", "line_number": 75, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 75, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 82, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.get", "line_number": 85, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 85, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 85, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 89, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects.get", "line_number": 92, "usage_type": "call"}, {"api_name": "todolist_app.models.TaskList.objects", "line_number": 92, "usage_type": "attribute"}, {"api_name": "todolist_app.models.TaskList", "line_number": 92, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 96, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 101, "usage_type": "call"}]} +{"seq_id": "67962543", "text": "import numpy as np\nfrom olds import dataLoader\nfrom torchvision import transforms\nfrom tqdm import tqdm\nimport pickle as pkl\nimport torch\nimport pickle\nimport time\n\ndevice = torch.device('cuda:6')\n\n\ndef evalHot(y, pred):\n \"\"\"\n 评估效果\n :param y:真实值的独热编码\n :param pred: 预测值的输出\n :return: 正确的个数\n \"\"\"\n _y = torch.argmax(y, dim=-1)\n _pred = torch.argmax(pred, dim=-1)\n N = np.sum((_y == _pred).cpu().numpy())\n return N\n\n\ndef KMeansRepeatX(X, repeat, train=True):\n \"\"\"\n :param X:Raw data \\\\in R^{batch_size X n_dim}\n :param repeat:重复的次数、采样数\n :return: 加了偏置项和重复数据的样本 维度[batch_size,repeat,n_dum+1]\n \"\"\"\n X = X.reshape(len(X), -1)\n if train:\n repeatX = torch.cat([X] * repeat, dim=0).to(device)\n one_shape = tuple(repeatX.shape[:-1]) + (1,)\n one = torch.ones(size=one_shape, dtype=torch.float).to(device)\n return torch.cat([repeatX, one], dim=-1)\n else:\n one = torch.ones(tuple(X.shape[:-1]) + (1,), dtype=torch.float).to(device)\n return torch.cat([X, one], dim=-1)\n\n\ndef OneHotLabel(Y, n):\n \"\"\"\n :param Y:序列型标签\n :param n: 标签数目\n :return: 标签的独热编码\n \"\"\"\n y = torch.zeros([len(Y), n]).to(device)\n y[torch.arange(0, len(Y)), Y] = 1\n return y\n\n\ndef KMeansRepeatY(Y, repeat):\n # print(Y.shape)\n repeatY = torch.cat([Y] * repeat, dim=0)\n return repeatY\n\n\nclass Activation:\n \"\"\"\n 包含激活函数\n \"\"\"\n\n @staticmethod\n def logistic(z):\n return 1 / (1 + torch.exp(-z))\n\n @staticmethod\n def softmax(z):\n stable_exps = torch.exp(z)\n return stable_exps / stable_exps.sum(dim=-1, keepdim=True)\n\n @staticmethod\n def threshold(z):\n z[z < 0] = 0\n return torch.sign(z)\n\n @staticmethod\n def relu(z):\n z[z < 0] = 0\n return z\n\n\ndef CELoss(Y, T):\n \"\"\"\n :param Y:模型输出\n :param T: 样本标签\n :return: 交叉熵损失\n \"\"\"\n return -(T * torch.log(Y)).sum(dim=-1)\n\n\nclass Layer:\n def __init__(self, n_input, n_output, sigma, activation):\n \"\"\"\n :param n_input:输入维度\n :param n_output: 输出维度\n :param sigma: 方差\n :param activation: 激活函数\n \"\"\"\n self.w = torch.randn(size=[n_input, n_output]).to(device) # 多出来的是bias\n self.w *= (2 / self.w.shape[0] ** 0.5)\n self.sigma = sigma\n self.n_input = n_input\n self.n_output = n_output\n self.input = None\n self.output = None\n self.noise = None\n self.activation = activation\n self.bp_grad = None\n self.lr_grad = None\n self.batch_bp_grad = None\n self.batch_lr_grad = None\n\n def get_params(self):\n return self.w\n\n def forward(self, x, train=False, BP=False):\n self.input = x\n if BP:\n # print(self.input.shape)\n # print(self.w.shape)\n self.output = self.input.matmul(self.w)\n self.noise = torch.randn([len(self.input), self.n_output]) * self.sigma\n self.noise = self.noise.to(device)\n self.output += self.noise\n if self.activation:\n self.output = self.activation(self.output)\n return self.output\n else:\n if not train:\n self.output = self.input.matmul(self.w)\n if self.activation:\n self.output = self.activation(self.output)\n return self.output\n else:\n self.noise = torch.randn([len(self.input), self.n_output]) * self.sigma\n self.noise = self.noise.to(device)\n self.output = self.input.matmul(self.w) + self.noise\n if self.activation:\n self.output = self.activation(self.output)\n return self.output\n\n def backward(self, target, BP=True):\n \"\"\"\n :param target: BP训练模式下,target是残差;LR训练模式下,target是损失\n :param BP: 是否为BP训练\n :return: BP训练模式下,返回残差;LR训练模式下,返回损失\n \"\"\"\n if BP:\n eta = target\n if self.activation == Activation.softmax:\n eta = self.output - eta\n elif self.activation == Activation.logistic:\n eta = self.output * (1 - self.output) * eta\n else:\n print('尚未注册!\\n')\n exit()\n batch_size = len(self.input)\n grad = self.input.T.matmul(eta)\n self.bp_grad = grad / batch_size\n return torch.einsum('ij,kj->ik', eta, self.w)\n else:\n term = self.input * target[:, np.newaxis]\n batch_grad = torch.einsum('ni, nj->nij', term, self.noise)\n batch_grad /= self.sigma ** 2\n batch_grad = torch.mean(batch_grad, dim=0)\n self.lr_grad = batch_grad\n return target\n\n def update_params(self, learning_rate, BP=True):\n if BP:\n self.w -= learning_rate * self.bp_grad\n else:\n self.w -= learning_rate * self.lr_grad\n\n\nclass Network(object):\n def __init__(self, n_input, units_per_layers: list, activation_per_layers: list, sigma):\n assert len(units_per_layers) == len(activation_per_layers)\n self.n_layers = len(units_per_layers)\n self.params = [(n_input, units_per_layers[0], sigma, activation_per_layers[0])]\n for i in range(self.n_layers - 1):\n self.params.append(\n (units_per_layers[i], units_per_layers[i + 1], sigma,\n activation_per_layers[i + 1]))\n self.layers = [Layer(*self.params[i]) for i in range(self.n_layers)]\n print('模型层数为:{} 各层及对应的激活函数为:{}'.format(len(self.layers),\n [(units_per_layers[i], activation_per_layers[i]) for i in\n range(self.n_layers)]))\n\n def forward(self, X, train=True, BP=False):\n z = X\n for layer in self.layers:\n # print(BP)\n z = layer.forward(z, train, BP)\n return z\n\n def backward(self, target, BP=True):\n \"\"\"\n :param target:BP训练方式下target是标签 LR训练方式下target是损失\n :param BP: 是否为BP模式\n :return: None\n \"\"\"\n if BP:\n for i in range(self.n_layers - 1, -1, -1):\n target = self.layers[i].backward(target, BP)\n else:\n for layer in self.layers:\n layer.backward(target, BP)\n\n def update_params(self, learning_rate, BP=True):\n for layer in self.layers:\n layer.update_params(learning_rate, BP)\n\n\n\n\nif __name__ == \"__main__\":\n mnist = 'MNIST'\n task = mnist\n train_loss = []\n test_loss = []\n acc = []\n time_list = []\n epoch_train_estimation_relative_error = []\n epoch_test_estimation_relative_error = []\n repeat_n = 1\n net_arc = [50, 10]\n learning_rate = 1e-1\n sigma = 2.0\n alg_start = 0.\n alg_end = 0.\n net_act = [Activation.logistic, Activation.softmax]\n assert len(net_arc) == len(net_act)\n n_layers = len(net_arc)\n if task == mnist:\n print('run mnist')\n batch_size = 128\n n_input = 28 * 28 + 1\n n_output = 10\n seed = None\n epoches = 10\n loss = 0.\n num_classes = 10\n reuse = False\n BP_train = True\n print('BP+ train on ')\n transform = transforms.Compose([transforms.ToTensor()])\n start_epoch = 0\n train_dataloader, test_dataloader = dataLoader.LoadMNIST('../data/MNIST', transform, batch_size, False)\n net = Network(n_input, net_arc, net_act, sigma)\n # train_img, train_label = loadMNIST_RAM(train_dataloader, repeat_n)\n print('数据准备完成!')\n trainLoss = 0.\n testLoss = 0.\n print('epoch to run:{} learning rate:{}'.format(epoches, learning_rate))\n start = time.time()\n print('模型信息:\\narc:{}\\nact:{}\\nK:{}'.format(net_arc, net_act, repeat_n))\n alg_start = time.time()\n for epoch in range(start_epoch, start_epoch + epoches):\n loss = 0.\n nbatch = 0.\n N = 0.\n n = 0.\n trainLoss = 0.\n train_estimation_relative_error = 0\n for batch, [trainX, trainY] in enumerate(tqdm(train_dataloader, ncols=10)):\n # break\n\n nbatch += 1\n trainX = trainX.to(device)\n trainY = trainY.to(device)\n trainY = OneHotLabel(trainY, num_classes)\n batch_train_repeatX, batch_train_repeatY = KMeansRepeatX(trainX, repeat_n), KMeansRepeatY(trainY,\n repeat_n)\n pre = net.forward(batch_train_repeatX, train=True, BP=BP_train)\n\n loss = CELoss(pre, batch_train_repeatY)\n trainLoss += torch.mean(loss).detach().cpu().numpy()\n if BP_train:\n net.backward(batch_train_repeatY, BP_train)\n net.update_params(learning_rate, BP_train)\n else:\n net.backward(loss, BP_train)\n net.update_params(learning_rate, BP_train)\n trainLoss /= nbatch\n train_loss.append(trainLoss)\n epoch_train_estimation_relative_error.append(train_estimation_relative_error / nbatch)\n # trainAcc = n / N\n print('train epoch:{} loss:{}'.format(epoch, trainLoss))\n if ((epoch + 1) % 10 == 0):\n learning_rate *= 0.8\n print('学习率衰减至{}'.format(learning_rate))\n loss = 0.\n N = 0.\n n = 0.\n nbatch = 0.\n test_estimation_relative_error = 0\n for batch, [testX, testY] in enumerate(tqdm(test_dataloader, ncols=10)):\n nbatch += 1\n testX = testX.to(device)\n testY = testY.to(device)\n testX = KMeansRepeatX(testX, 1, False)\n testY = OneHotLabel(testY, num_classes)\n\n pre = net.forward(testX, train=False)\n testLoss += torch.mean(CELoss(pre, testY)).detach().cpu().numpy()\n N += len(testX)\n n += evalHot(testY, pre)\n testLoss /= nbatch\n test_loss.append(testLoss)\n testAcc = n / N\n acc.append(testAcc)\n epoch_test_estimation_relative_error.append(test_estimation_relative_error / nbatch)\n print('test epoch:{} loss:{} acc:{}'.format(epoch, testLoss, n / N))\n time_list.append(time.time() - start)\n net.save_weights(os.path.join('./models_mnist', model_path), epoch, learning_rate)\n alg_end = time.time()\n print('train_loss:{}\\n test_loss:{}\\n acc:{}'.format(train_loss, test_loss, acc))\n print('time:{}'.format(time_list))\n with open('BPplus_{}.pkl'.format(sigma), 'wb') as file:\n pkl.dump(\n [train_loss, test_loss, acc], file)\n print(alg_end-alg_start)", "sub_path": "MNIST&Fashion/train/BPplus.py", "file_name": "BPplus.py", "file_ext": "py", "file_size_in_byte": 11326, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "torch.device", "line_number": 10, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.ones", "line_number": 39, "usage_type": "call"}, {"api_name": "torch.float", "line_number": 39, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 40, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.arange", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 56, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 67, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.sign", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.log", "line_number": 91, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 125, "usage_type": "call"}, {"api_name": "torch.randn", "line_number": 138, "usage_type": "call"}, {"api_name": "torch.einsum", "line_number": 163, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 165, "usage_type": "attribute"}, {"api_name": "torch.einsum", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 168, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 250, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 250, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 250, "usage_type": "call"}, {"api_name": "olds.dataLoader.LoadMNIST", "line_number": 252, "usage_type": "call"}, {"api_name": "olds.dataLoader", "line_number": 252, "usage_type": "name"}, {"api_name": "time.time", "line_number": 259, "usage_type": "call"}, {"api_name": "time.time", "line_number": 261, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 269, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 281, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 301, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 309, "usage_type": "call"}, {"api_name": "time.time", "line_number": 318, "usage_type": "call"}, {"api_name": "time.time", "line_number": 320, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 324, "usage_type": "call"}]} +{"seq_id": "509162647", "text": "from collections import defaultdict\nfrom regulus.topo import Partition\n\n\nclass Merge(object):\n def __init__(self, level, is_max, src, dest):\n self.level = level\n self.is_max = is_max\n self.src = src\n self.dest = dest\n\n\nclass PartitionNode(object):\n _id_generator = -1\n\n @staticmethod\n def gen_id():\n PartitionNode._id_generator += 1\n return PartitionNode._id_generator\n\n @staticmethod\n def reset():\n PartitionNode._id_generator = -1\n\n def __init__(self, persistence, base_pts=None, min_idx=None, max_idx=None, from_partition=None, is_max=None):\n self.id = PartitionNode.gen_id()\n self.persistence = persistence\n self.span = []\n self.parent = None\n self.children = []\n\n self.extrema = []\n self.base_pts = base_pts if base_pts is not None else []\n self.min_idx = min_idx\n self.max_idx = max_idx\n self.max_merge = is_max\n\n if from_partition is not None:\n self.min_idx = from_partition.min_idx\n self.max_idx = from_partition.max_idx\n self.children.append(from_partition)\n from_partition.parent = self\n\n def add_child(self, child):\n child.parent = self\n self.children.append(child)\n # if child.min_idx != self.min_idx and child.max_idx != self.max_idx:\n # print(\"ERROR: child {} [{} {}] merged into parent {} [{} {}] without a matching extrema\".format(child.id,\n # child.min_idx, child.max_idx, self.id, self.min_idx, self.max_idx))\n\n\nclass Builder(object):\n def __init__(self, debug=False):\n self.base = None\n self.merges = []\n self.min_map = defaultdict(set)\n self.max_map = defaultdict(set)\n self.active = set()\n self.root = None\n self.pts = []\n self.original_pts = set()\n self.debug = debug\n self.mapping = dict()\n self.unique = set()\n self.all = dict()\n self.data_pts = []\n self.single = 0\n\n def data(self, pts):\n self.data_pts = pts\n return self\n\n def msc(self, base, hierarchy):\n self.base = base\n for entry in hierarchy:\n row = entry.split(',')\n self.merges.append(Merge(float(row[1]), row[0] == 'Maxima', int(row[2]), int(row[3])))\n return self\n\n def build(self):\n self.prepare()\n self.merge()\n\n # get root\n if len(self.active) != 1:\n raise RuntimeError('Error: found {} roots'.format(len(self.active)))\n self.root = self.active.pop()\n\n self.single = 0\n idx = self.build_idx(self.root, 0)\n print('found {} singles'.format(self.single))\n print('len(idx)=', idx)\n\n self.test_uniques()\n\n self.pts.extend([self.root.min_idx, self.root.max_idx])\n self.rename(self.root, 0)\n return self\n\n # internal\n\n def merge(self):\n for record in self.merges:\n # print(record.level, record.is_max, record.src, record.dest)\n if record.src == record.dest:\n continue\n\n # merge.dest may have been merged already (same persistence level: degenerate case)\n dest = self.current(record.dest)\n src = self.current(record.src)\n\n if src == dest:\n continue\n\n record.dest = dest\n record.src = src\n self.mapping[record.src] = record.dest\n\n if record.is_max:\n self.collapse(record, self.max_map, lambda item: item.min_idx)\n else:\n self.collapse(record, self.min_map, lambda item: item.max_idx)\n\n def prepare(self):\n PartitionNode.reset()\n for key, value in self.base.items():\n m, x = [int(s) for s in key.split(',')]\n p = PartitionNode(0, list(value), m, x)\n self.add(p)\n\n # self.find_unique()\n self.remove_non_unique()\n\n self.merges.sort(key=lambda m: (m.level, m.src))\n high = self.merges[-1].level\n for merge in self.merges:\n merge.level /= high\n\n if self.debug:\n for partition in self.active:\n self.check_partition(partition)\n\n def collapse(self, merge, idx_map, idx):\n add_partitions = []\n remove_partitions = set()\n\n for d in idx_map[merge.dest]:\n new_partition = None\n remove_src = set()\n for s in idx_map[merge.src]:\n if idx(s) == idx(d):\n if s.persistence != merge.level:\n if new_partition is None:\n new_partition = PartitionNode(merge.level, from_partition=d, is_max=merge.is_max)\n remove_partitions.add(d) # can't be removed during the iterations\n add_partitions.append(new_partition)\n new_partition.add_child(s)\n else:\n # s is an intermediate and should be absorbed\n if len(s.children) == 0:\n # s is a base partition\n d.base_pts.extend(s.base_pts)\n else:\n for child in s.children:\n d.add_child(child)\n if len(s.extrema) > 0:\n d.extrema.extend(s.extrema)\n remove_src.add(s) # can't be removed during the iterations\n for s in remove_src:\n self.remove(s)\n\n for s in idx_map[merge.src]:\n # create a new partition with a single child because the max value has changed\n new_partition = PartitionNode(merge.level, from_partition=s)\n if merge.is_max:\n new_partition.max_idx = merge.dest\n else:\n new_partition.min_idx = merge.dest\n add_partitions.append(new_partition)\n\n for r in remove_partitions | idx_map[merge.src]:\n self.remove(r)\n\n # assign the eliminated extrema as an extra internal point to the first new partition\n if merge.src not in self.unique:\n if len(add_partitions) > 0:\n target = add_partitions[0]\n else:\n target = next(iter(idx_map[merge.dest]))\n target.extrema.append(merge.src)\n\n for new_partition in add_partitions:\n self.add(new_partition)\n\n # consistency checks\n\n def check_partition(self, p):\n min_v = self.data_pts[p.min_idx]\n max_v = self.data_pts[p.max_idx]\n if min_v > max_v:\n print('*** min > max', min_v, max_v)\n for pt_idx in p.base_pts:\n if self.data_pts[pt_idx] < min_v:\n print('*** Partition id:{} min:{} at {} found min:{} at {}'.format(p.id, min_v, p.min_idx, self.data_pts[pt_idx], pt_idx))\n if self.data_pts[pt_idx] > max_v:\n print('*** Partition id:{} max:{} at {} found max:{} at {}'.format(p.id, max_v, p.max_idx, self.data_pts[pt_idx], pt_idx))\n\n #\n # helpers\n #\n\n def current(self, partition):\n while partition in self.mapping:\n partition = self.mapping[partition]\n return partition\n\n def find_loop(self, dest):\n loop = [dest]\n while dest in self.mapping:\n dest = self.mapping[dest]\n loop.append(dest)\n return loop\n\n def find_unique(self):\n count = defaultdict(int)\n for p in self.active:\n count[p.min_idx] += 1\n count[p.max_idx] += 1\n self.unique = {k for k, v in count.items() if v == 1}\n print(' unique:', self.unique)\n self.all = count\n\n def remove_non_unique(self):\n for p in self.active:\n for idx in [p.min_idx, p.max_idx]:\n if idx not in self.unique:\n p.base_pts.remove(idx)\n else:\n print(idx, 'not removed becuase it is unique')\n\n def add(self, n):\n self.min_map[n.min_idx].add(n)\n self.max_map[n.max_idx].add(n)\n self.active.add(n)\n\n def remove(self, p):\n self.max_map[p.max_idx].discard(p)\n self.min_map[p.min_idx].discard(p)\n self.active.remove(p)\n\n def build_idx(self, partition, idx):\n first = idx\n if len(partition.children) == 0:\n if partition.min_idx in partition.base_pts and partition.min_idx not in self.unique:\n print('*** WARNING: min_idx {} in partition {}'.format(partition.min_idx, partition.id))\n if partition.max_idx in partition.base_pts and partition.max_idx not in self.unique:\n print('*** WARNING: max_idx {} in partition {}'.format(partition.max_idx, partition.id))\n\n n = len(partition.base_pts)\n if n > 0:\n self.pts.extend(partition.base_pts)\n idx += n\n else:\n if len(partition.children) == 1:\n self.single += 1\n\n if len(partition.extrema) > 0:\n self.pts.extend(partition.extrema)\n idx += len(partition.extrema)\n\n for child in partition.children:\n idx = self.build_idx(child, idx)\n\n partition.span = (first, idx)\n return idx\n\n def test_uniques(self):\n for u in self.unique:\n self.test_unique(u, self.root, 0)\n\n def test_unique(self, u, node, lvl):\n if u == node.min_idx:\n print('unique {} is min_idx for node {} lvl {}'.format(u, node.id, lvl))\n if u == node.max_idx:\n print('unique {} is max_idx for node {} lvl {}'.format(u, node.id, lvl))\n if u in node.extrema:\n print('unique {} in extrema for node {} lvl {}'.format(u, node.id, lvl))\n if node.span[0] <= u < node.span[1]:\n print('unique {} in span for node {} lvl {}'.format(u, node.id, lvl))\n for child in node.children:\n self.test_unique(u, child, lvl+1)\n\n def rename(self, node, idx):\n node.id = idx\n idx += 1\n if node.persistence > 0:\n for child in node.children:\n idx = self.rename(child, idx)\n return idx\n\n #\n # save\n #\n\n def visit(self, p, visitor):\n visitor(p)\n for child in p.children:\n self.visit(child, visitor)\n\n def get_tree(self, name, params=''):\n partitions = []\n self.collect_partitions(self.root, partitions)\n tree = {\n 'partitions': partitions,\n 'pts_idx': self.pts\n }\n return tree\n\n def collect_partitions(self, node, array):\n array.append({\n 'id': node.id,\n 'lvl': node.persistence,\n 'span': [node.span[0], node.span[1]],\n 'minmax_idx': [node.min_idx, node.max_idx],\n 'merge': 'max' if node.is_max_merge else 'min',\n 'parent': node.parent.id if node.parent is not None else None,\n 'children': [child.id for child in node.children] if node.persistence > 0 else []\n })\n\n self.check_partition(node)\n\n if node.persistence > 0:\n if len(node.children) > 2:\n print('\\t{} has {} children at level {}'.format(node.id, len(node.children), node.persistence))\n for child in node.children:\n self.collect_partitions(child, array)\n\n #\n # verify\n #\n\n def verify(self):\n if self.debug:\n self.statistics()\n return self\n\n def statistics(self):\n levels = defaultdict(list)\n self.stat(self.root, levels)\n n = 0\n b = 0\n for level in levels.keys():\n if level > 0:\n n += len(levels[level])\n else:\n b = len(levels[level])\n print('\\tstatistics: {} levels {} base, {} new'.format(len(levels), b, n))\n # for level in sorted(levels.keys()):\n # print(\"{:.2g} {}\".format(level, len(levels[level])))\n\n def stat(self, node, levels):\n levels[node.persistence].append(node)\n if node.persistence > 0:\n for child in node.children:\n self.stat(child, levels)\n", "sub_path": "regulus/topo/builder.py", "file_name": "builder.py", "file_ext": "py", "file_size_in_byte": 12269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "collections.defaultdict", "line_number": 56, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 57, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 225, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 348, "usage_type": "call"}]} +{"seq_id": "403918136", "text": "#!/usr/bin/env python3\n# Guy Serbin, EOanalytics Ltd.\n# Talent Garden Dublin, Claremont Ave. Glasnevin, Dublin 11, Ireland\n# email: guyserbin eoanalytics ie\n\n# version 1.3.1\n\n# This script will create and update a geopackage layer of all available Landsat TM/ETM+/OLI-TIRS scenes, including available metadata\n# Changes:\n# 23 May 2018: XML functionality deprecated in favor of JSON queries, as the former is no longer available or efficient\n# 25 March 2019: This script will now read configuration data from ieo.ini\n# 14 August 2019: This now creates and updates a layer within a geopackage, and will migrate data from an old shapefile to a new one\n# 12 January 2021: Modified to support Landsat Collection 2\n\nimport os, sys, urllib.error, datetime, shutil, glob, argparse, json, getpass, requests, math #, ieo\nfrom osgeo import ogr, osr\n#import xml.etree.ElementTree as ET\nfrom PIL import Image\n\ntry: # This is included as the module may not properly install in Anaconda.\n import ieo\nexcept:\n print('Error: IEO failed to load. Please input the location of the directory containing the IEO installation files.')\n ieodir = input('IEO installation path: ')\n if os.path.isfile(os.path.join(ieodir, 'ieo.py')):\n sys.path.append(ieodir)\n import ieo\n else:\n print('Error: that is not a valid path for the IEO module. Exiting.')\n sys.exit()\n\nif sys.version_info[0] == 2:\n from urllib import urlretrieve\n from urllib2 import urlopen, URLError\nelse:\n from urllib.request import urlopen, urlretrieve\n from urllib.error import URLError\n\nglobal pathrows, errorsfound\n\nconfig = ieo.config\n\npathrowvals = config['Landsat']['pathrowvals'] # this is a comma-delimited string containing multiples of four values: start path, end path, start row, end row. It is designed to query rectangular path/row combinations, in order to avoid scenes that don't touch landmasses or are not of interest.\nuseWRS2 = config['Landsat']['useWRS2'] # Setting this parameter to \"Yes\" in updateshp.ini will query WRS-2 Path/ Row field values from ieo.WRS2, and may result in a great increase in the number of queries to USGS servers\n\nparser = argparse.ArgumentParser('This script imports LEDAPS-processed scenes into the local library. It stacks images and converts them to the locally defined projection in IEO, and adds ENVI metadata.')\nparser.add_argument('-u','--username', type = str, default = None, help = 'USGS/EROS Registration System (ERS) username.')\nparser.add_argument('-p', '--password', type = str, default = None, help = 'USGS/EROS Registration System (ERS) password.')\nparser.add_argument('-c', '--catalogID', type = str, default = 'EE', help = 'USGS/EROS Catalog ID (default = \"EE\").')\nparser.add_argument('-v', '--version', type = str, default = \"1.4.1\", help = 'JSON version, default = 1.4.1.')\nparser.add_argument('--startdate', type = str, default = \"1982-07-16\", help = 'Start date for query in YYYY-MM-DD format. (Default = 1982-07-16, e.g., Landsat 4 launch date).')\nparser.add_argument('--enddate', type = str, default = None, help = \"End date for query in YYYY-MM-DD format. (Default = today's date).\")\nparser.add_argument('-m', '--MBR', type = str, default = None, help = 'Minimum Bounding Rectangle (MBR) coordinates in decimal degrees in the following format (comma delimited, no spaces): lower left latitude, lower left longitude, upper right latitude, upper right longitude. If not supplied, these will be determined from WRS-2 Paths and Rows in updateshp.ini.')\nparser.add_argument('-b', '--baseURL', type = str, default = 'https://earthexplorer.usgs.gov/inventory/json/v/', help = 'Base URL to use excluding JSON version (Default = \"https://earthexplorer.usgs.gov/inventory/json/v/\").')\nparser.add_argument('--maxResults', type = int, default = 50000, help = 'Maximum number of results to return (1 - 50000, default = 50000).')\nparser.add_argument('--overwrite', type = bool, default = False, help = 'Overwrite existing files.')\nparser.add_argument('--thumbnails', type = bool, default = True, help = 'Download thumbnails (default = True).')\nparser.add_argument('--savequeries', action = 'store_true', help = 'Save queries.')\nparser.add_argument('--usesaved', action = 'store_true', help = 'Use any saved queries on disk, rather than online.')\nparser.add_argument('--migrate', type = bool, default = False, help = 'Force migration of Landsat shapefile data to catalog geopackage.')\nparser.add_argument('--verbose', type = bool, default = False, help = 'Display more messages during migration..')\nparser.add_argument('-t', '--tiledir', type = str, default = os.path.dirname(ieo.srdir), help = 'Directory path for tile subdirectories.')\n\nargs = parser.parse_args()\n\nif not (args.username and args.password):\n if not args.username:\n args.username = input('USGS/ERS username: ')\n if not args.password:\n args.password = getpass.getpass('USGS/ERS password: ')\n\nsubpathrow = []\n\ningestdir = os.path.join(ieo.ingestdir, 'Metadata')\ndirname = os.path.join(ieo.catdir, 'Landsat')\nlogdir = ieo.logdir\njpgdir = os.path.join(ieo.catdir, 'Landsat', 'Thumbnails')\nitmdir = ieo.srdir\nshapefile = ieo.landsatshp # This is a layer in a geopackage, not a shapefile any longer\nlayername = ieo.landsatshp # os.path.basename(shapefile)[:-4] # assumes a shapefile ending in '.shp'\nshapefilepath = os.path.join(ieo.catdir, 'Landsat', '{}.shp'.format(shapefile))\nerrorlist = []\nscenelist = []\nif not args.enddate:\n today = datetime.datetime.today()\n args.enddate = today.strftime('%Y-%m-%d')\n\nerrorfile = os.path.join(logdir, 'Landsat_inventory_download_errors.csv')\nerrorsfound = False\n\npathrowstrs = [] # list of strings containing WRS-2 Path/ Row combinations\npaths = [] # list containing WRS-2 Paths\nrows = [] # List containing WRS-2 Rows\n\nif useWRS2.lower() == 'yes':\n print('Getting WRS-2 Path/Row combinations from geopackage: {}'.format(ieo.WRS2))\n driver = ogr.GetDriverByName(\"GPKG\")\n print('WRS-2 = {}'.format(ieo.WRS2))\n ds = driver.Open(ieo.ieogpkg, 0)\n layer = ds.GetLayer(ieo.WRS2)\n for feature in layer:\n path = feature.GetField('PATH')\n if not path in paths:\n paths.append(path)\n row = feature.GetField('ROW')\n if not row in rows:\n rows.append(row)\n pathrowstrs.append('{:03d}{:03d}'.format(path, row))\n ds = None\nelse:\n print('Using WRS-2 Path/Row combinations from INI file.')\n pathrowvals = pathrowvals.split(',')\n iterations = int(len(pathrowvals) / 4)\n for i in range(iterations):\n for j in range(int(pathrowvals[i * 4]), int(pathrowvals[i * 4 + 1]) + 1):\n if not j in paths:\n paths.append(j)\n for k in range(int(pathrowvals[i * 4 + 2]), int(pathrowvals[i * 4 + 3]) + 1):\n pathrowstrs.append('{:03d}{:03d}'.format(j, k))\n if not k in rows:\n rows.append(k)\n\n## JSON functions\n\ndef getapiKey():\n # This function gets the apiKey used for all queries to the USGS/EROS servers\n URL = '{}{}/login'.format(args.baseURL, args.version)\n print('Logging in to: {}'.format(URL))\n data = json.dumps({'username': args.username, 'password': args.password, 'catalog_ID': args.catalogID})\n response = requests.post(URL, data = {'jsonRequest':data}) #\n json_data = json.loads(response.text)\n apiKey = json_data['data']\n return apiKey\n\ndef getMBR():\n # This creates the Minimum Bounding Rectangle (MBR) for JSON queries\n URL = '{}{}/grid2ll'.format(args.baseURL, args.version)\n prs = [[min(paths), min(rows)], [min(paths), max(rows)], [max(paths), max(rows)], [max(paths), min(rows)]]\n Xcoords = []\n Ycoords = []\n for pr in prs:\n print('Requesting coordinates for WRS-2 Path {} Row {}.'.format(pr[0], pr[1]))\n jsonRequest = json.dumps({\"gridType\" : \"WRS2\", \"responseShape\" : \"point\", \"path\" : str(pr[0]), \"row\" : str(pr[1])}).replace(' ','')\n requestURL = '{}?jsonRequest={}'.format(URL, jsonRequest)\n response = requests.post(requestURL) # URL, data = {'jsonRequest': jsonRequest}\n json_data = json.loads(response.text)\n # print(response.text)\n Xcoords.append(float(json_data[\"data\"][\"coordinates\"][0][\"longitude\"]))\n Ycoords.append(float(json_data[\"data\"][\"coordinates\"][0][\"latitude\"]))\n return [min(Ycoords), min(Xcoords), max(Ycoords), max(Xcoords)]\n\ndef scenesearch(apiKey, scenelist, updatemissing, badgeom, lastmodifiedDate):\n # This searches the USGS archive for scene metadata, and checks it against local metadata. New scenes will be queried for metadata.\n RequestURL = '{}{}/search'.format(args.baseURL, args.version)\n QueryURL = '{}{}/metadata'.format(args.baseURL, args.version)\n datasetNames = {'landsat_ot_c2_l2' : '2013-02-11', 'landsat_etm_c2_l2' : '1999-04-15', 'landsat_tm_c2_l2' : '1982-07-16'}\n scenedict = {}\n# js = {'LL': 0, 'UL': 1, 'UR': 2, 'LR': 3}\n for datasetName in datasetNames.keys():\n print('Querying collection: {}'.format(datasetName))\n if lastmodifiedDate and not (len(updatemissing) > 0 or len(badgeom) > 0):\n startdate = lastmodifiedDate\n else:\n startdate = args.startdate\n if '/' in startdate:\n startdate = startdate.replace('/', '-')\n datetuple = datetime.datetime.strptime(startdate, '%Y-%m-%d')\n sensorstarttuple = datetime.datetime.strptime(datasetNames[datasetName], '%Y-%m-%d') # restrict searches to times from which sensor was in orbit\n if datetuple < sensorstarttuple:\n datetuple = sensorstarttuple\n enddatetuple = datetime.datetime.strptime(args.enddate, '%Y-%m-%d')\n if datasetName == 'landsat_tm_c2_l2':\n l5enddatetuple = datetime.datetime.strptime('2013-06-05', '%Y-%m-%d') # end of Landsat 5 mission\n if l5enddatetuple < enddatetuple:\n enddatetuple = l5enddatetuple\n while datetuple < enddatetuple:\n edatetuple = datetuple + datetime.timedelta(days = 365) # iterate by year\n if edatetuple > enddatetuple:\n edatetuple = enddatetuple\n startdate = datetuple.strftime('%Y-%m-%d')\n enddate = edatetuple.strftime('%Y-%m-%d')\n print('Now searching for scene data from collection {} from {} through {}.'.format(datasetName, startdate, enddate))\n searchparams = json.dumps({\"apiKey\": apiKey,\n \"datasetName\": datasetName,\n \"spatialFilter\":{\"filterType\": \"mbr\",\n \"lowerLeft\":{\"latitude\": args.MBR[0],\n \"longitude\": args.MBR[1]},\n \"upperRight\":{\"latitude\": args.MBR[2],\n \"longitude\": args.MBR[3]}},\n \"temporalFilter\":{\"startDate\": startdate,\n \"endDate\": enddate},\n \"includeUnknownCloudCover\":False,\n \"maxCloudCover\": 100,\n \"maxResults\": args.maxResults,\n \"sortOrder\": \"ASC\"})\n response = requests.post(RequestURL, data = {'jsonRequest': searchparams})\n json_data = json.loads(response.text)\n querylist = []\n # print(response.text)\n for i in range(len(json_data['data']['results'])):\n sceneID = json_data['data']['results'][i]['entityId']\n if sceneID[3:9] in pathrowstrs and (not sceneID in scenelist or sceneID in updatemissing or sceneID in badgeom):\n querylist.append(sceneID)\n scenedict[sceneID] = {'Landsat Product Identifier': json_data['data']['results'][i][\"displayId\"],\n \"browseUrl\": json_data['data']['results'][i][\"browseUrl\"],\n \"dataAccessUrl\": json_data['data']['results'][i][\"dataAccessUrl\"],\n \"downloadUrl\": json_data['data']['results'][i][\"downloadUrl\"],\n \"metadataUrl\": json_data['data']['results'][i][\"metadataUrl\"],\n \"fgdcMetadataUrl\": json_data['data']['results'][i][\"fgdcMetadataUrl\"],\n # 'modifiedDate': datetime.datetime.strptime(json_data['data']['results'][i][\"modifiedDate\"], '%Y-%m-%d'),\n \"orderUrl\": json_data['data']['results'][i][\"orderUrl\"],\n 'Dataset Identifier': datasetName,\n 'updatemodifiedDate': False,\n 'updategeom': False}\n if json_data['data']['results'][i][\"modifiedDate\"] == 'Unknown':\n scenedict[sceneID]['modifiedDate'] = datetime.datetime.strptime(json_data['data']['results'][i][\"acquisitionDate\"], '%Y-%m-%d')\n elif ' ' in json_data['data']['results'][i][\"modifiedDate\"]:\n print(json_data['data']['results'][i][\"modifiedDate\"])\n space = json_data['data']['results'][i][\"modifiedDate\"].find(' ')\n scenedict[sceneID]['modifiedDate'] = datetime.datetime.strptime(json_data['data']['results'][i][\"modifiedDate\"][:space], '%Y-%m-%d')\n else:\n scenedict[sceneID]['modifiedDate'] = datetime.datetime.strptime(json_data['data']['results'][i][\"modifiedDate\"], '%Y-%m-%d')\n \n if len(querylist) > 0:\n print('{} new scenes have been found or require updating, querying metadata.'.format(len(querylist)))\n iterations = math.ceil(len(querylist) / 100) # break up queries into blocks of 100 or less scenes\n total = 0\n # iterations = 1 # temporary limitation\n for iteration in range(iterations):\n startval = iteration * 100\n if iteration * 100 > len(querylist):\n endval = len(querylist) - startval - 1\n else:\n endval = startval + 99\n total += endval + 1\n print('Now querying {} scenes, query {}/{}.'.format((endval - startval + 1), iteration + 1, iterations))\n querystr = ''\n \n for sceneID in querylist[startval: endval]:\n querystr += ',{}'.format(sceneID)\n querystr = querystr[1:]\n queryparams = json.dumps({\"apiKey\":apiKey,\n \"datasetName\":datasetName,\n 'entityIds': querystr})\n try:\n query = requests.post(QueryURL, data = {'jsonRequest':queryparams})\n # if endval == 99:\n \n if args.savequeries:\n now = datetime.datetime.now()\n outfile = os.path.join(ieo.ingestdir, 'query_{}_{}.txt'.format(datasetName, now.strftime('%Y%m%d-%H%M%S')))\n with open(outfile, 'w') as output:\n output.write(query.text)\n querydict = json.loads(query.text)\n if len(querydict['data']) > 0:\n \n for item in querydict['data']:\n if len(item['metadataFields']) > 0:\n if item['metadataFields'][1]['fieldName'] == 'Landsat Scene Identifier':\n sceneID = item['metadataFields'][1]['value']\n else:\n for subitem in item['metadataFields']:\n if subitem['fieldName'] == 'Landsat Scene Identifier':\n sceneID = subitem['value']\n break\n for subitem in item['metadataFields']:\n fieldname = subitem['fieldName'].rstrip().lstrip().replace('L-1', 'L1')\n if fieldname in queryfieldnames and not fieldname in scenedict[sceneID].keys() and fieldname != 'Landsat Scene Identifier':\n value = subitem['value']\n if value:\n i = queryfieldnames.index(fieldname)\n if fieldvaluelist[i][3] == ogr.OFTDate or fieldname.endswith('Date'):\n if 'Time' in fieldname:\n value = datetime.datetime.strptime(value[:-1], '%Y:%j:%H:%M:%S.%f')\n elif '/' in value:\n value = datetime.datetime.strptime(value, '%Y/%m/%d')\n else:\n value = datetime.datetime.strptime(value, '%Y-%m-%d')\n elif fieldvaluelist[i][3] == ogr.OFTReal:\n value = float(value)\n elif fieldvaluelist[i][3] == ogr.OFTInteger:\n try:\n value = int(value)\n except:\n print('Error: fieldname {} has a value of {}, changing to -9999.'.format(fieldname, value))\n value = -9999\n elif fieldname == 'browseUrl':\n if value:\n if value.lower() != 'null':\n scenedict[sceneID]['browse'] = 'Y'\n else:\n scenedict[sceneID]['browse'] = 'N'\n elif fieldname == 'Data Type Level-1':\n j = value.rfind('_') + 1\n value = value[j:]\n scenedict[sceneID][fieldname] = value\n if sceneID in badgeom or sceneID in updatemissing:\n scenedict[sceneID]['updatemodifiedDate'] = True \n else: \n scenedict[sceneID]['updatemodifiedDate'] = False \n if sceneID in badgeom:\n scenedict[sceneID]['updategeom'] = True\n else: \n scenedict[sceneID]['updategeom'] = False\n scenedict[sceneID]['coords'] = item['spatialFootprint']['coordinates'][0]\n scenedict[sceneID]['modifiedDate'] = item['modifiedDate']\n except Exception as e:\n print('ERROR: {e}')\n ieo.logerror(QueryURL, e)\n \n # if not 'Spacecraft Identifier' in scenedict[sceneID].keys():\n # scenedict[sceneID]['Spacecraft Identifier'] = 'LANDSAT_{}'.format(sceneID[2:3])\n # if 'Scan Gap Interpolation' in scenedict[sceneID].keys():\n # if isinstance(scenedict[sceneID]['Scan Gap Interpolation'], float):\n # scenedict[sceneID]['Scan Gap Interpolation'] = int(scenedict[sceneID]['Scan Gap Interpolation'])\n datetuple = edatetuple + datetime.timedelta(days = 1)\n return scenedict\n\ndef findlocalfiles(sceneID, fielddict, scenedict):\n tilebase = '{}_{}'.format(sceneID[:3], sceneID[9:16])\n for fieldname in fielddict:\n tilelist = glob.glob(os.path.join(fielddict[fieldname]['dirname'], '{}*.dat'.format(tilebase)))\n tiles = []\n tilestr = None\n if len(tilelist) > 0:\n for f in tilelist:\n parentrasters = ieo.readenvihdr(f.replace('.dat', '.hdr'))['parent rasters']\n if sceneID in parentrasters:\n basename = os.path.basename(f)\n i = basename.rfind('_') + 1\n j = f.find('.')\n tiles.append(basename[i:j])\n if len(tiles) > 0:\n tilestr = tiles[0]\n if len(tiles) > 1:\n for i in range(1, len(tiles)):\n tilestr += ',{}'.format(tiles[i])\n scenedict[fieldname] = tilestr\n if fieldname == 'Pixel_QA_tiles':\n scenedict['MaskType'] = 'Pixel_QA'\n elif fieldname == 'Fmask_tiles':\n scenedict['MaskType'] = 'FMask'\n \n# srstr = feature.GetField('Surface_Reflectance_tiles')\n# if isinstance(srstr, str):\n# srlist = srstr.split(',')\n# tilebase = feature.GetField('Tile_filename_base')\n# itm = os.path.join(itmdir, '{}_{}.dat'.format(tilebase, srlist[0]))\n# else:\n# itm = ''\n# if not os.path.isfile(itm): # Populate 'SR_path' field if surface reflectance data are present in library\n# \n# if len(itmlist) > 0:\n# itm = itmlist[0]\n# else:\n# itm = os.path.join(itmdir, '{}_ref_{}.dat'.format(scenedict[sceneID]['Landsat Product Identifier'], ieo.projacronym))\n# if not os.path.isfile(itm):\n# itm = None\n# if itm:\n# scenedict['Surface_Reflectance_tiles'] = srstr\n# for key in fielddict.keys():\n# value = feature.GetField(key)\n# if isinstance(value, str):\n# scenedict[key] = value\n# if isinstance(feature.GetField('Pixel_QA_tiles'), str):\n# scenedict['MaskType'] = 'Pixel_QA'\n# elif isinstance(feature.GetField('Fmask_tiles'), str):\n# scenedict['MaskType'] = 'FMask'\n return scenedict\n\n## Migration functions\n \ndef migrate(layer, shapefilepath, fieldvaluelist, *args, **kwargs):\n # added on 14 August 2019\n # This will migrate features from a shapefile to a geopackage if they have reasonable geometries.\n tiledir = kwargs.get('tiledir', os.path.dirname(ieo.srdir))\n verbose = kwargs.get('verbose', verbose)\n print('Migrating data from shapefile to geopackage.')\n fnamelist = []\n tilesearchdict = {'SR_path' : os.path.join(tiledir, os.path.basename(ieo.srdir)), \n 'BT_path' : os.path.join(tiledir, os.path.basename(ieo.btdir)), \n 'Fmask_path' : os.path.join(tiledir, os.path.basename(ieo.fmaskdir)), \n 'PixQA_path' : os.path.join(tiledir, os.path.basename(ieo.pixelqadir)), \n 'NDVI_path' : os.path.join(tiledir, os.path.basename(ieo.ndvidir)), \n 'EVI_path' : os.path.join(tiledir, os.path.basename(ieo.evidir))}\n fieldvaluedict = {}\n for item in fieldvaluelist:\n fieldvaluedict[item[0]] = item[1]\n fieldvaluedict['MaskType'] = 'Scene_mask_type'\n fieldvaluedict['Thumb_JPG'] = 'Thumbnail_filename'\n fieldvaluedict['SR_path'] = 'Surface_reflectance_tiles'\n fieldvaluedict['BT_path'] = 'Brightness_temperature_tiles'\n fieldvaluedict['Fmask_path'] = 'CFmask_tiles'\n fieldvaluedict['PixQA_path'] = 'Pixel_QA_tiles'\n fieldvaluedict['NDVI_path'] = 'NDVI_tiles'\n fieldvaluedict['EVI_path'] = 'EVI_tiles'\n fieldvaluedict['tilebase'] = 'Tile_filename_base' \n shpdriver = ogr.GetDriverByName(\"ESRI Shapefile\")\n ds = shpdriver.Open(shapefilepath, 0)\n shplayer = ds.GetLayer()\n shplayerDefinition = shplayer.GetLayerDefn()\n# layerDefinition = layer.GetLayerDefn()\n featureCount = layer.GetFeatureCount()\n sceneids = []\n if featureCount > 0:\n for feat in layer:\n sceneids.append(feat.GetField('sceneID'))\n layer.ResetReading()\n for i in range(shplayerDefinition.GetFieldCount()):\n fnamelist.append(shplayerDefinition.GetFieldDefn(i).GetName())\n for feature in shplayer:\n sceneid = feature.GetField('sceneID')\n if (not ieo.checkscenegeometry(feature, verbose = verbose)) and (sceneid in sceneids):\n print('Migrating feature for SceneID {} and associated metadata.'.format(sceneid))\n outfeature = ogr.Feature(layer.GetLayerDefn())\n tilebase = feature.GetField('tilebase')\n for field in fnamelist:\n if field in tilesearchdict.keys():\n tilestr = ''\n filelist = glob.glob(os.path.join(tilesearchdict[field], '{}_*.dat'.format(tilebase)))\n if len(filelist) > 0:\n for f in filelist:\n basename = os.path.basename(f)\n j = basename.find('.dat')\n tilestr += ',{}'.format(basename[12:j])\n outfeature.SetField(fieldvaluedict[field], tilestr[1:])\n elif field in fieldvaluedict.keys():\n if field == 'Thumbnail_filename':\n value = feature.GetField(field)\n if not os.path.isfile(value):\n value = os.path.join(jpgdir, feature.GetField('LandsatPID'))\n if not os.path.isfile(value):\n value = os.path.join(jpgdir, feature.GetField('sceneID'))\n if os.path.isfile(value):\n outfeature.SetField(fieldvaluedict[field], os.path.basename(value)) # from now on, only base filenames will be included\n else:\n outfeature.SetField(fieldvaluedict[field], feature.GetField(field))\n geom = feature.GetGeometryRef()\n outfeature.SetGeometry(geom)\n layer.SetFeature(outfeature)\n outfeature.Destroy()\n ds = None\n print('Feature migration complete.')\n return layer\n\n\n## Old XML functions, deprecated\n\ndef dlxmls(startdate, enddate, xmls, ingestdir, *args, **kwargs): # This downloads queried XML files\n global errorsfound\n tries = 1\n downloaded = False\n for x, p in zip(xmls, pathrows):\n\n print('Downloading {} to: {}'.format(x, ingestdir))\n xml = os.path.join(ingestdir, x)\n if os.access(xml, os.F_OK):\n print('Backing up current xml file.')\n shutil.move(xml, '{}.{}.bak'.format(xml, today.strftime('%Y%m%d-%H%M%S')))\n urlname = 'http://earthexplorer.usgs.gov/EE/InventoryStream/pathrow?start_path={}&end_path={}&start_row={}&end_row={}&sensor_name=LANDSAT_COMBINED_C1&start_date={}&end_date={}'.format(p[0], p[1], p[2], p[3], startdate, enddate) #&cloud_cover = 100&seasonal = False&aoi_entry=path_row&output_type=unknown\n tries = 1\n downloaded = False\n while not downloaded and tries < 6:\n print('Download attempt {} of 5.'.format(tries))\n try:\n urlretrieve(urlname, xml) # filename=xml\n downloaded = True\n except URLError as e:\n print(e.reason)\n ieo.logerror(urlname, e.reason, errorfile = errorfile)\n errorsfound = True\n tries += 1\n if tries == 6:\n ieo.logerror(xml, 'Download error.', errorfile = errorfile)\n print('Download failure: {}'.format(x))\n errorsfound = True\n else:\n return 'Success!'\n\n\n\n\n## Other functions\n\ndef dlthumb(dlurl, jpg, *args, **kwargs): # This downloads thumbnails from the USGS\n global errorsfound\n jpgdir, basename = os.path.split(jpg)\n # f = os.path.join(jpgdir, basename)\n r = requests.get(dlurl, allow_redirects=True)\n open(jpg, 'wb').write(r.content)\n # tries = 1\n # downloaded = False\n # print('Downloading {} to {}'.format(basename, jpgdir))\n # while not downloaded and tries < 6:\n # print('Download attempt {} of 5.'.format(tries))\n # try:\n \n # # url = urlopen(dlurl)\n # # urlretrieve(dlurl, filename = f)\n # # if url.length == os.stat(f).st_size:\n # # downloaded = True\n # # else:\n # # print('Error downloading, retrying.')\n # # tries += 1\n # # except urllib.error.URLError as e:\n # except Exception as e:\n # print(e) # e.reason\n # ieo.logerror(dlurl, e, errorfile = errorfile) # e.reason\n # errorsfound = True\n # if tries == 6:\n # ieo.logerror(f, 'Download error.', errorfile = errorfile)\n # print('Download failure: {}'.format(basename))\n # errorsfound = True\n # else:\n return 'Success!'\n\ndef makeworldfile(jpg, geom): # This attempts to make a worldfile for thumbnails so they can be displayed in a GIS\n img = Image.open(jpg)\n basename = os.path.basename(jpg)\n width, height = img.size\n width = float(width)\n height = float(height)\n minX, maxX, minY, maxY = geom.GetEnvelope()\n if basename[:3] == 'LE7':\n wkt = geom.ExportToWkt()\n start = wkt.find('(') + 2\n end = wkt.find(')')\n vals = wkt[start:end]\n vals = vals.split(',')\n corners = []\n for val in vals:\n val = val.split()\n for v in val:\n corners.append(float(v))\n A = (maxX - corners[0]) / width\n B = (corners[0] - minX) / height\n C = corners[0]\n D = (maxY - corners[3]) / width\n E = (corners[3] - minY) / height\n F = corners[1]\n else:\n A = (maxX - minX) / width\n B = 0.0\n C = minX\n D = (maxY - minY) / height\n E = 0.0\n F = maxY\n jpw = jpg.replace('.jpg', '.jpw')\n if os.access(jpw, os.F_OK):\n bak = jpw.replace('.jpw', '.jpw.{}.bak'.format(today.strftime('%Y%m%d-%H%M%S')))\n shutil.move(jpw, bak)\n with open(jpw, 'w') as file:\n file.write('{}\\n-{}\\n-{}\\n-{}\\n{}\\n{}\\n'.format(A, D, B, E, C, F))\n del img\n\ndef reporthook(blocknum, blocksize, totalsize):\n # This makes a progress bar. I did not originally write it, nor do I remember from where I found the code.\n readsofar = blocknum * blocksize\n if totalsize > 0:\n percent = readsofar * 1e2 / totalsize\n s = \"\\r%5.1f%% %*d / %d\" % (\n percent, len(str(totalsize)), readsofar, totalsize)\n sys.stderr.write(s)\n if readsofar >= totalsize: # near the end\n sys.stderr.write(\"\\n\")\n else: # total size is unknown\n sys.stderr.write(\"read %d\\n\" % (readsofar,))\n\nif args.MBR: # define MBR for scene queries\n args.MBR = args.MBR.split(',')\n if len(args.MBR) != 4:\n ieo.logerror('--MBR', 'Total number of coordinates does not equal four.', errorfile = errorfile)\n print('Error: Improper number of coordinates for --MBR set (must be four). Either remove this option (will use default values) or fix. Exiting.')\n sys.exit()\nelse:\n args.MBR = getMBR()\n\n# This section borrowed from https://pcjericks.github.io/py-gdalogr-cookbook/projection.html\n# Lat/ Lon WGS-84 to local projection transformation\nsource = osr.SpatialReference() # Lat/Lon WGS-64\nsource.ImportFromEPSG(4326)\n\ntarget = ieo.prj\n\ntransform = osr.CoordinateTransformation(source, target)\n\n# Create Shapefile\ndriver = ogr.GetDriverByName(\"GPKG\")\n\npolycoords = ['UL Corner Lat dec', 'UL Corner Long ec', 'UR Corner Lat dec', 'UR Corner Long dec', 'LL Corner Lat dec', 'LL Corner Long dec', 'LR Corner Lat dec', 'LR Corner Long dec']\n\nfieldvaluelist = [\n ['LandsatPID', 'LANDSAT_PRODUCT_ID', 'Landsat Product Identifier', ogr.OFTString, 40],\n ['sceneID', 'sceneID', 'Landsat Scene Identifier', ogr.OFTString, 21],\n ['SensorID', 'SensorID', 'Sensor Identifier', ogr.OFTString, 0],\n ['SatNumber', 'satelliteNumber', 'Spacecraft Identifier', ogr.OFTString, 0],\n ['acqDate', 'acquisitionDate', 'Acquisition Date', ogr.OFTDate, 0],\n ['Updated', 'dateUpdated', 'modifiedDate', ogr.OFTDate, 0],\n ['path', 'path', 'WRS Path', ogr.OFTInteger, 0],\n ['row', 'row', 'WRS Row', ogr.OFTInteger, 0],\n ['CenterLat', 'sceneCenterLatitude', 'Center Latitude dec', ogr.OFTReal, 0],\n ['CenterLong', 'sceneCenterLongitude', 'Center Longitude dec', ogr.OFTReal, 0],\n ['CC', 'cloudCover', 'Cloud Cover Truncated', ogr.OFTInteger, 0],\n ['CCFull', 'cloudCoverFull', 'Scene Cloud Cover', ogr.OFTReal, 0],\n ['CCLand', 'CLOUD_COVER_LAND', 'Land Cloud Cover', ogr.OFTReal, 0],\n ['UL_Q_CCA', 'FULL_UL_QUAD_CCA', 'Cloud Cover Quadrant Upper Left', ogr.OFTReal, 0],\n ['UR_Q_CCA', 'FULL_UR_QUAD_CCA', 'Cloud Cover Quadrant Upper Right', ogr.OFTReal, 0],\n ['LL_Q_CCA', 'FULL_LL_QUAD_CCA', 'Cloud Cover Quadrant Lower Left', ogr.OFTReal, 0],\n ['LR_Q_CCA', 'FULL_LR_QUAD_CCA', 'Cloud Cover Quadrant Lower Right', ogr.OFTReal, 0],\n ['DT_L1', 'DATA_TYPE_L1', 'Data Type Level-1', ogr.OFTString, 0],\n ['DT_L0RP', 'DATA_TYPE_L0RP', 'Data Type Level 0Rp', ogr.OFTString, 0],\n ['L1_AVAIL', 'L1_AVAILABLE', 'L1 Available', ogr.OFTString, 0],\n ['IMAGE_QUAL', 'IMAGE_QUALITY', 'Image Quality', ogr.OFTString, 0],\n ['dayOrNight', 'dayOrNight', 'Day/Night Indicator', ogr.OFTString, 0],\n ['sunEl', 'sunElevation', 'Sun Elevation L1', ogr.OFTReal, 0],\n ['sunAz', 'sunAzimuth', 'Sun Azimuth L1', ogr.OFTReal, 0],\n ['StartTime', 'sceneStartTime', 'Start Time', ogr.OFTDate, 0],\n ['StopTime', 'sceneStopTime', 'Stop Time', ogr.OFTDate, 0],\n ['UTM_ZONE', 'UTM_ZONE', 'UTM Zone', ogr.OFTInteger, 0],\n ['DATUM', 'DATUM', 'Datum', ogr.OFTString, 0],\n ['ELEVSOURCE', 'ELEVATION_SOURCE', 'Elevation Source', ogr.OFTString, 0],\n ['ELLIPSOID', 'ELLIPSOID', 'Ellipsoid', ogr.OFTString, 0],\n ['PROJ_L1', 'MAP_PROJECTION_L1', 'Map Projection Level-1', ogr.OFTString, 0],\n ['PROJ_L0RA', 'MAP_PROJECTION_L0RA', 'Map Projection L0Ra', ogr.OFTString, 0],\n ['ORIENT', 'ORIENTATION', 'Orientation', ogr.OFTString, 0],\n ['EPHEM_TYPE', 'EPHEMERIS_TYPE', 'Ephemeris Type', ogr.OFTString, 0],\n ['CPS_MODEL', 'GROUND_CONTROL_POINTS_MODEL', 'Ground Control Points Model', ogr.OFTInteger, 0],\n ['GCPSVERIFY', 'GROUND_CONTROL_POINTS_VERIFY', 'Ground Control Points Version', ogr.OFTInteger, 0],\n ['RMSE_MODEL', 'GEOMETRIC_RMSE_MODEL', 'Geometric RMSE Model (meters)', ogr.OFTReal, 0],\n ['RMSE_X', 'GEOMETRIC_RMSE_MODEL_X', 'Geometric RMSE Model X', ogr.OFTReal, 0],\n ['RMSE_Y', 'GEOMETRIC_RMSE_MODEL_Y', 'Geometric RMSE Model Y', ogr.OFTReal, 0],\n ['RMSEVERIFY', 'GEOMETRIC_RMSE_VERIFY', 'Geometric RMSE Verify', ogr.OFTReal, 0],\n ['FORMAT', 'OUTPUT_FORMAT', 'Output Format', ogr.OFTString, 0],\n ['RESAMP_OPT', 'RESAMPLING_OPTION', 'Resampling Option', ogr.OFTString, 0],\n ['LINES', 'REFLECTIVE_LINES', 'Reflective Lines', ogr.OFTInteger, 0],\n ['SAMPLES', 'REFLECTIVE_SAMPLES', 'Reflective Samples', ogr.OFTInteger, 0],\n ['TH_LINES', 'THERMAL_LINES', 'Thermal Lines', ogr.OFTInteger, 0],\n ['TH_SAMPLES', 'THERMAL_SAMPLES', 'Thermal Samples', ogr.OFTInteger, 0],\n ['PAN_LINES', 'PANCHROMATIC_LINES', 'Panchromatic Lines', ogr.OFTInteger, 0],\n ['PANSAMPLES', 'PANCHROMATIC_SAMPLES', 'Panchromatic Samples', ogr.OFTInteger, 0],\n ['GC_SIZE_R', 'GRID_CELL_SIZE_REFLECTIVE', 'Grid Cell Size Reflective', ogr.OFTInteger, 0],\n ['GC_SIZE_TH', 'GRID_CELL_SIZE_THERMAL', 'Grid Cell Size Thermal', ogr.OFTInteger, 0],\n ['GCSIZE_PAN', 'GRID_CELL_SIZE_PANCHROMATIC', 'Grid Cell Size Panchromatic', ogr.OFTInteger, 0],\n ['PROCSOFTVE', 'PROCESSING_SOFTWARE_VERSION', 'Processing Software Version', ogr.OFTString, 0],\n ['CPF_NAME', 'CPF_NAME', 'Calibration Parameter File', ogr.OFTString, 0],\n ['DATEL1_GEN', 'DATE_L1_GENERATED', 'Date L-1 Generated', ogr.OFTString, 0],\n ['GCP_Ver', 'GROUND_CONTROL_POINTS_VERSION', 'Ground Control Points Version', ogr.OFTInteger, 0],\n ['DatasetID', 'DatasetID', 'Dataset Identifier', ogr.OFTString, 0],\n ['CollectCat', 'COLLECTION_CATEGORY', 'Collection Category', ogr.OFTString, 0],\n ['CollectNum', 'COLLECTION_NUMBER', 'Collection Number', ogr.OFTString, 0],\n ['flightPath', 'flightPath', 'flightPath', ogr.OFTString, 0],\n ['RecStation', 'receivingStation', 'Station Identifier', ogr.OFTString, 0],\n ['imageQual1', 'imageQuality1', 'Image Quality 1', ogr.OFTString, 0],\n ['imageQual2', 'imageQuality2', 'Image Quality 2', ogr.OFTString, 0],\n ['gainBand1', 'gainBand1', 'Gain Band 1', ogr.OFTString, 0],\n ['gainBand2', 'gainBand2', 'Gain Band 2', ogr.OFTString, 0],\n ['gainBand3', 'gainBand3', 'Gain Band 3', ogr.OFTString, 0],\n ['gainBand4', 'gainBand4', 'Gain Band 4', ogr.OFTString, 0],\n ['gainBand5', 'gainBand5', 'Gain Band 5', ogr.OFTString, 0],\n ['gainBand6H', 'gainBand6H', 'Gain Band 6H', ogr.OFTString, 0],\n ['gainBand6L', 'gainBand6L', 'Gain Band 6L', ogr.OFTString, 0],\n ['gainBand7', 'gainBand7', 'Gain Band 7', ogr.OFTString, 0],\n ['gainBand8', 'gainBand8', 'Gain Band 8', ogr.OFTString, 0],\n ['GainChange', 'GainChange', 'Gain Change', ogr.OFTString, 0],\n ['GCBand1', 'gainChangeBand1', 'Gain Change Band 1', ogr.OFTString, 0],\n ['GCBand2', 'gainChangeBand2', 'Gain Change Band 2', ogr.OFTString, 0],\n ['GCBand3', 'gainChangeBand3', 'Gain Change Band 3', ogr.OFTString, 0],\n ['GCBand4', 'gainChangeBand4', 'Gain Change Band 4', ogr.OFTString, 0],\n ['GCBand5', 'gainChangeBand5', 'Gain Change Band 5', ogr.OFTString, 0],\n ['GCBand6H', 'gainChangeBand6H', 'Gain Change Band 6H', ogr.OFTString, 0],\n ['GCBand6L', 'gainChangeBand6L', 'Gain Change Band 6L', ogr.OFTString, 0],\n ['GCBand7', 'gainChangeBand7', 'Gain Change Band 7', ogr.OFTString, 0],\n ['GCBand8', 'gainChangeBand8', 'Gain Change Band 8', ogr.OFTString, 0],\n ['SCAN_GAP_I', 'SCAN_GAP_INTERPOLATION', 'Scan Gap Interpolation', ogr.OFTReal, 0],\n ['ROLL_ANGLE', 'ROLL_ANGLE', 'Roll Angle', ogr.OFTReal, 0],\n ['FULL_PART', 'FULL_PARTIAL_SCENE', 'Full Partial Scene', ogr.OFTString, 0],\n ['NADIR_OFFN', 'NADIR_OFFNADIR', 'Nadir/Off Nadir', ogr.OFTString, 0],\n ['RLUT_FNAME', 'RLUT_FILE_NAME', 'RLUT File Name', ogr.OFTString, 0],\n ['BPF_N_OLI', 'BPF_NAME_OLI', 'Bias Parameter File Name OLI', ogr.OFTString, 0],\n ['BPF_N_TIRS', 'BPF_NAME_TIRS', 'Bias Parameter File Name TIRS', ogr.OFTString, 0],\n ['TIRS_SSM', 'TIRS_SSM_MODEL', 'TIRS SSM Model', ogr.OFTString, 0],\n ['TargetPath', 'Target_WRS_Path', 'Target WRS Path', ogr.OFTInteger, 0],\n ['TargetRow', 'Target_WRS_Row', 'Target WRS Row', ogr.OFTInteger, 0],\n ['DataAnom', 'data_anomaly', 'Data Anomaly', ogr.OFTString, 0],\n ['GapPSource', 'gap_phase_source', 'Gap Phase Source', ogr.OFTString, 0],\n ['GapPStat', 'gap_phase_statistic', 'Gap Phase Statistic', ogr.OFTReal, 0],\n ['L7SLConoff', 'scan_line_corrector', 'Scan Line Corrector', ogr.OFTString, 0],\n ['SensorAnom', 'sensor_anomalies', 'Sensor Anomalies', ogr.OFTString, 0],\n ['SensorMode', 'sensor_mode', 'Sensor Mode', ogr.OFTString, 0],\n ['browse', 'browseAvailable', 'Browse Available', ogr.OFTString, 0],\n ['browseURL', 'browseURL', 'browseUrl', ogr.OFTString, 0],\n ['MetadatUrl', 'metadataUrl', 'metadataUrl', ogr.OFTString, 0],\n ['FGDCMetdat', 'fgdcMetadataUrl', 'fgdcMetadataUrl', ogr.OFTString, 0],\n ['dataAccess', 'dataAccess', 'dataAccessUrl', ogr.OFTString, 0],\n ['orderUrl', 'orderUrl', 'orderUrl', ogr.OFTString, 0],\n ['DownldUrl', 'downloadUrl', 'downloadUrl', ogr.OFTString, 0]]\n\nqueryfieldnames = []\nfnames = []\n\nfor element in fieldvaluelist:\n fnames.append(element[1])\n queryfieldnames.append(element[2])\n\nif not os.access(ieo.catgpkg, os.F_OK):\n # Create geopackage\n data_source = driver.CreateDataSource(ieo.catgpkg)\nelse:\n data_source = driver.Open(ieo.catgpkg, 1)\nlayerpresent = False\nlayers = data_source.GetLayerCount()\nif layers > 0:\n for i in range(layers):\n if layername == data_source.GetLayer(i).GetName():\n layerpresent = True\nif not layerpresent:\n layer = data_source.CreateLayer(layername, target, ogr.wkbPolygon)\n for element in fieldvaluelist:\n field_name = ogr.FieldDefn(element[1], element[3])\n if element[4] > 0:\n field_name.SetWidth(element[4])\n layer.CreateField(field_name)\n\n layer.CreateField(ogr.FieldDefn('MaskType', ogr.OFTString)) # 'Fmask' or 'Pixel_QA'\n layer.CreateField(ogr.FieldDefn('Thumbnail_filename', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('Surface_reflectance_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('Brightness_temperature_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('Fmask_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('Pixel_QA_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('NDVI_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('EVI_tiles', ogr.OFTString))\n layer.CreateField(ogr.FieldDefn('Tile_filename_base', ogr.OFTString))\n \n args.migrate = True \n\nif args.migrate and os.path.isfile(shapefilepath):\n layer = migrate(layer, shapefilepath, fieldvaluelist, tiledir = args.tiledir, verbose = args.verbose)\n\n\n#else:\nlastupdate = None\nlastmodifiedDate = None\nshpfnames = []\nupdatemissing = []\nbadgeom = []\nreimport = []\n# Open existing shapefile with write access\ndata_source = driver.Open(ieo.catgpkg, 1)\nlayer = data_source.GetLayer(shapefile)\nlayerDefinition = layer.GetLayerDefn()\n# Get list of field names\nfor i in range(layerDefinition.GetFieldCount()):\n shpfnames.append(layerDefinition.GetFieldDefn(i).GetName())\n# Find missing fields and create them\nfor fname in fnames:\n if not fname in shpfnames:\n i = fnames.index(fname)\n field_name = ogr.FieldDefn(fnames[i], fieldvaluelist[i][3])\n if fieldvaluelist[i][4] > 0:\n field_name.SetWidth(fieldvaluelist[i][4])\n layer.CreateField(field_name)\n\n# Iterate through features and fetch sceneID values\nerrors = {'total' : 0,\n 'metadata' : 0,\n 'date' : 0,\n 'geometry' : 0}\n\nfeatureCount = layer.GetFeatureCount()\nif featureCount > 0:\n layer.StartTransaction()\n layer.GetNextFeature()\n while feature:\n \n datetuple = None\n try:\n sceneID = feature.GetField(\"sceneID\")\n except Exception as e:\n exc_type, exc_obj, exc_tb = sys.exc_info()\n fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n if args.verbose:\n print(exc_type, fname, exc_tb.tb_lineno)\n print('ERROR: bad feature, deleting.')\n layer.DeleteFeature(feature.GetFID())\n ieo.logerror('{}/{}'.format(ieo.catgpkg, shapefile), '{} {} {}'.format(exc_type, fname, exc_tb.tb_lineno), errorfile = errorfile)\n feature = layer.GetNextFeature()\n continue\n scenelist.append(sceneID)\n if not feature.GetField('SensorID') in ['TM', 'ETM', 'OLI', 'TIRS', 'OLI_TIRS']:\n if args.verbose:\n print('ERROR: missing metadata for SceneID {}. Feature will be deleted from shapefile and reimported.'.format(sceneID))\n ieo.logerror(sceneID, 'Feature missing metadata, deleted, reimportation required.')\n reimport.append(datetime.datetime.strptime(sceneID[9:16], '%Y%j'))\n layer.DeleteFeature(feature.GetFID())\n errors['total'] += 1\n errors['metadata'] += 1\n \n else: \n try:\n mdate = feature.GetField('dateUpdated')\n datetuple = datetime.datetime.strptime(mdate, '%Y/%m/%d')\n if not lastupdate or datetuple > lastupdate:\n lastupdate = datetuple\n lastmodifiedDate = mdate\n except Exception as e:\n if args.verbose:\n exc_type, exc_obj, exc_tb = sys.exc_info()\n fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n print(exc_type, fname, exc_tb.tb_lineno)\n print('ERROR: modifiedDate information missing for SceneID {}, adding to list.'.format(sceneID))\n ieo.logerror(sceneID, 'Modification date missing.', errorfile = errorfile)\n updatemissing.append(sceneID)\n errors['total'] += 1\n errors['date'] += 1\n \n \n try:\n geom = feature.GetGeometryRef()\n env = geom.GetEnvelope()\n if env[0] == env[1] or env[2] == env[3]:\n if args.verbose:\n print('Bad geometry identified for SceneID {}, adding to the list.'.format(sceneID))\n ieo.logerror(sceneID, 'Bad/ missing geometry.')\n badgeom.append(sceneID)\n except Exception as e:\n if args.verbose:\n exc_type, exc_obj, exc_tb = sys.exc_info()\n fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n print(exc_type, fname, exc_tb.tb_lineno)\n print('Bad geometry identified for SceneID {}, adding to the list.'.format(sceneID), errorfile = errorfile)\n ieo.logerror(sceneID, 'Bad/ missing geometry.')\n badgeom.append(sceneID)\n errors['total'] += 1\n errors['geometry'] += 1\n if errors['total'] > 0 and (errors['total'] % 100 == 0):\n print('{} errors found in layer of types: metadata: {}, missing modification date: {}, missing/ bad geometry: {}.'.format(errors['total'], errors['metadata'], errors['date'], errors['geometry']))\n feature = layer.GetNextFeature()\n layer.CommitTransaction()\n\nif len(reimport) > 0 and lastupdate:\n if min(reimport) < lastupdate:\n lastmodifiedDate = datetime.datetime.strftime('%Y-%m-%d', min(reimport))\n\nfielddict = {'Brightness_temperature_tiles' : {'ext' : '_BT_{}.dat'.format(ieo.projacronym), 'dirname' : ieo.btdir},\n 'CFmask_tiles' : {'ext' : '_cfmask.dat', 'dirname' : ieo.fmaskdir},\n 'Pixel_QA_tiles' : {'ext' : '_pixel_qa.dat', 'dirname' : ieo.pixelqadir},\n 'NDVI_tiles' : {'ext' : '_NDVI.dat', 'dirname' : ieo.ndvidir},\n 'EVI_tiles' : {'ext' : '_EVI.dat', 'dirname' : ieo.evidir}}\n\nthumbnails = []\nscenes = []\nfilenum = 1\n\n# get apiKey for USGS EarthExplorer query\napiKey = getapiKey()\n\n# run query\n\nscenedict = scenesearch(apiKey, scenelist, updatemissing, badgeom, lastmodifiedDate)\nsceneIDs = scenedict.keys()\nprint('Total scenes to be added or updated to geopackage layer: {}'.format(len(sceneIDs)))\n\nif len(sceneIDs) > 0:\n for sceneID in sceneIDs:\n print('Processing {}, scene number {} of {}.'.format(sceneID, filenum, len(sceneIDs)))\n if not (scenedict[sceneID]['updategeom'] or scenedict[sceneID]['updatemodifiedDate']) and ('coords' in scenedict[sceneID].keys()):\n scenedict = findlocalfiles(sceneID, fielddict, scenedict)\n # if scenedict[sceneID]['browseUrl'].endswith('.jpg'):\n dlurl = scenedict[sceneID]['browseUrl']\n # thumbnails.append(scenedict[sceneID]['browseUrl'])\n \n print('\\nAdding {} to geopackage layer.'.format(sceneID))\n scenelist.append(sceneID)\n # create the feature\n feature = ogr.Feature(layer.GetLayerDefn())\n # Add field attributes\n feature.SetField('sceneID', sceneID)\n for key in scenedict[sceneID].keys():\n if (scenedict[sceneID][key]) and key in queryfieldnames:\n try:\n if fieldvaluelist[queryfieldnames.index(key)][3] == ogr.OFTDate:\n if isinstance(scenedict[sceneID][key], str):\n if '/' in scenedict[sceneID][key]:\n scenedict[sceneID][key] = scenedict[sceneID][key].replace('/', '-')\n scenedict[sceneID][key] = datetime.datetime.strptime(scenedict[sceneID][key], '%Y-%m-%d')\n feature.SetField(fnames[queryfieldnames.index(key)], scenedict[sceneID][key].year, scenedict[sceneID][key].month, scenedict[sceneID][key].day, scenedict[sceneID][key].hour, scenedict[sceneID][key].minute, scenedict[sceneID][key].second, 100)\n else:\n feature.SetField(fnames[queryfieldnames.index(key)], scenedict[sceneID][key])\n except Exception as e:\n if args.verbose:\n exc_type, exc_obj, exc_tb = sys.exc_info()\n fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n print(exc_type, fname, exc_tb.tb_lineno)\n print('Error with SceneID {}, fieldname = {}, value = {}: {}'.format(sceneID, fnames[queryfieldnames.index(key)], scenedict[sceneID][key], e))\n ieo.logerror(key, e, errorfile = errorfile)\n \n coords = scenedict[sceneID]['coords']\n print(coords)\n # Create ring\n ring = ogr.Geometry(ogr.wkbLinearRing)\n for coord in coords:\n ring.AddPoint(coord[1], coord[0])\n if not coord[0] == coords[0][0] and coord[1] == coords[0][1]:\n ring.AddPoint(coords[0][1], coords[0][0])\n # Create polygon\n \n poly = ogr.Geometry(ogr.wkbPolygon)\n \n poly.AddGeometry(ring)\n poly.Transform(transform) # Convert to local projection\n feature.SetGeometry(poly)\n basename = '{}.jpg'.format(scenedict[sceneID]['Landsat Product Identifier'])\n # print(dlurl)#os.path.basename(dlurl)\n jpg = os.path.join(jpgdir, basename)\n if not os.access(jpg, os.F_OK) and args.thumbnails:\n try:\n response = dlthumb(dlurl, jpg)\n if response == 'Success!':\n geom = feature.GetGeometryRef()\n print('Creating world file.')\n makeworldfile(jpg, poly)\n print('Migrating world and projection files to new directory.')\n jpw = jpg.replace('.jpg', '.jpw')\n prj = jpg.replace('.jpg', '.prj')\n else:\n print('Error with sceneID or filename, adding to error list.')\n ieo.logerror(sceneID, response, errorfile = errorfile)\n errorsfound = True\n if os.access(jpg, os.F_OK):\n feature.SetField('Thumbnail_filename', jpg)\n\n except Exception as e:\n exc_type, exc_obj, exc_tb = sys.exc_info()\n fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n print(exc_type, fname, exc_tb.tb_lineno)\n ieo.logerror(os.path.basename(jpg), e, errorfile = errorfile)\n errorsfound = True\n layer.CreateFeature(feature)\n feature.Destroy()\n else:\n layer.ResetReading()\n for feature in layer:\n if feature.GetField('sceneID') == sceneID:\n if scenedict[sceneID]['updategeom']: \n print('Updating geometry for SceneID {}.'.format(sceneID))\n coords = scenedict[sceneID]['coords']\n # Create ring\n ring = ogr.Geometry(ogr.wkbLinearRing)\n for coord in coords:\n ring.AddPoint(coord[0], coord[1])\n if not coord[0] == coords[0][0] and coord[1] == coords[0][1]:\n ring.AddPoint(coord[0][0], coord[0][1])\n # Create polygon\n \n poly = ogr.Geometry(ogr.wkbPolygon)\n \n poly.AddGeometry(ring)\n poly.Transform(transform) # Convert to local projection\n feature.SetGeometry(poly)\n \n if scenedict[sceneID]['updatemodifiedDate']:\n# try:\n print('Updating modification date for SceneID {}.'.format(sceneID))\n# if isinstance(scenedict[sceneID]['modifiedDate'], str):\n# scenedict[sceneID]['modifiedDate'] = datetime.datetime.strptime(scenedict[sceneID]['modifiedDate'], '%Y-%m-%d')\n feature.SetField('dateUpdated', scenedict[sceneID]['modifiedDate'])\n# except Exception as e:\n# exc_type, exc_obj, exc_tb = sys.exc_info()\n# fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1]\n# print(exc_type, fname, exc_tb.tb_lineno)\n# print('ERROR: modifiedDate information (\"{}\") not set for SceneID {}, adding to list.'.format(scenedict[sceneID]['modifiedDate'], sceneID))\n# ieo.logerror(sceneID, 'Error setting \"Updated\" field.')\n layer.SetFeature(feature)\n feature.Destroy()\n# print('\\n')\n filenum += 1\n \ndata_source = None\n\nprint('Processing complete.')\n\n", "sub_path": "updatelandsat.py", "file_name": "updatelandsat.py", "file_ext": "py", "file_size_in_byte": 54029, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.isfile", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 26, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 32, "usage_type": "attribute"}, {"api_name": "ieo.config", "line_number": 41, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "ieo.srdir", "line_number": 62, "usage_type": "attribute"}, {"api_name": "getpass.getpass", "line_number": 70, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path", "line_number": 74, "usage_type": "attribute"}, {"api_name": "ieo.ingestdir", "line_number": 74, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "ieo.catdir", "line_number": 75, "usage_type": "attribute"}, {"api_name": "ieo.logdir", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path", "line_number": 77, "usage_type": "attribute"}, {"api_name": "ieo.catdir", "line_number": 77, "usage_type": "attribute"}, {"api_name": "ieo.srdir", "line_number": 78, "usage_type": "attribute"}, {"api_name": "ieo.landsatshp", "line_number": 79, "usage_type": "attribute"}, {"api_name": "ieo.landsatshp", "line_number": 80, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "ieo.catdir", "line_number": 81, "usage_type": "attribute"}, {"api_name": "datetime.datetime.today", "line_number": 85, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 85, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "ieo.WRS2", "line_number": 96, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.GetDriverByName", "line_number": 97, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 97, "usage_type": "name"}, {"api_name": "ieo.WRS2", "line_number": 98, "usage_type": "attribute"}, {"api_name": "ieo.ieogpkg", "line_number": 99, "usage_type": "attribute"}, {"api_name": "ieo.WRS2", "line_number": 100, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 129, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 130, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 131, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 143, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 145, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 146, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 167, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 167, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 168, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 168, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 171, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 171, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 173, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 173, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 177, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 183, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 196, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 197, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 216, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 216, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 220, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 220, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 222, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 222, "usage_type": "attribute"}, {"api_name": "math.ceil", "line_number": 226, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 242, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 246, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 250, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 250, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 251, "usage_type": "call"}, {"api_name": "os.path", "line_number": 251, "usage_type": "attribute"}, {"api_name": "ieo.ingestdir", "line_number": 251, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 254, "usage_type": "call"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 272, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 272, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 274, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 274, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 276, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 276, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 278, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 278, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 279, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 279, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 281, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 281, "usage_type": "name"}, {"api_name": "ieo.logerror", "line_number": 309, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 316, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 322, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 322, "usage_type": "call"}, {"api_name": "os.path", "line_number": 322, "usage_type": "attribute"}, {"api_name": "ieo.readenvihdr", "line_number": 327, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path", "line_number": 329, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 376, "usage_type": "call"}, {"api_name": "os.path", "line_number": 376, "usage_type": "attribute"}, {"api_name": "ieo.srdir", "line_number": 376, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 380, "usage_type": "call"}, {"api_name": "os.path", "line_number": 380, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 380, "usage_type": "call"}, {"api_name": "ieo.srdir", "line_number": 380, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 381, "usage_type": "call"}, {"api_name": "os.path", "line_number": 381, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 381, "usage_type": "call"}, {"api_name": "ieo.btdir", "line_number": 381, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 382, "usage_type": "call"}, {"api_name": "os.path", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 382, "usage_type": "call"}, {"api_name": "ieo.fmaskdir", "line_number": 382, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 383, "usage_type": "call"}, {"api_name": "os.path", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 383, "usage_type": "call"}, {"api_name": "ieo.pixelqadir", "line_number": 383, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 384, "usage_type": "call"}, {"api_name": "os.path", "line_number": 384, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 384, "usage_type": "call"}, {"api_name": "ieo.ndvidir", "line_number": 384, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path", "line_number": 385, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 385, "usage_type": "call"}, {"api_name": "ieo.evidir", "line_number": 385, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.GetDriverByName", "line_number": 398, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 398, "usage_type": "name"}, {"api_name": "ieo.checkscenegeometry", "line_number": 413, "usage_type": "call"}, {"api_name": "osgeo.ogr.Feature", "line_number": 415, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 415, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 420, "usage_type": "call"}, {"api_name": "os.path", "line_number": 420, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 423, "usage_type": "call"}, {"api_name": "os.path", "line_number": 423, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 430, "usage_type": "call"}, {"api_name": "os.path", "line_number": 430, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 431, "usage_type": "call"}, {"api_name": "os.path", "line_number": 431, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path", "line_number": 432, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 433, "usage_type": "call"}, {"api_name": "os.path", "line_number": 433, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 434, "usage_type": "call"}, {"api_name": "os.path", "line_number": 434, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 435, "usage_type": "call"}, {"api_name": "os.path", "line_number": 435, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 456, "usage_type": "call"}, {"api_name": "os.path", "line_number": 456, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 457, "usage_type": "call"}, {"api_name": "os.F_OK", "line_number": 457, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 459, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 466, "usage_type": "call"}, {"api_name": "urllib.error.URLError", "line_number": 468, "usage_type": "name"}, {"api_name": "ieo.logerror", "line_number": 470, "usage_type": "call"}, {"api_name": "ieo.logerror", "line_number": 474, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 487, "usage_type": "call"}, {"api_name": "os.path", "line_number": 487, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 489, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 518, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 518, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 519, "usage_type": "call"}, {"api_name": "os.path", "line_number": 519, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 549, "usage_type": "call"}, {"api_name": "os.F_OK", "line_number": 549, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 551, "usage_type": "call"}, {"api_name": "sys.stderr.write", "line_number": 563, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 563, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 565, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 565, "usage_type": "attribute"}, {"api_name": "sys.stderr.write", "line_number": 567, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 567, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 572, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 574, "usage_type": "call"}, {"api_name": "osgeo.osr.SpatialReference", "line_number": 580, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 580, "usage_type": "name"}, {"api_name": "ieo.prj", "line_number": 583, "usage_type": "attribute"}, {"api_name": "osgeo.osr.CoordinateTransformation", "line_number": 585, "usage_type": "call"}, {"api_name": "osgeo.osr", "line_number": 585, "usage_type": "name"}, {"api_name": "osgeo.ogr.GetDriverByName", "line_number": 588, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 588, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 593, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 593, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 594, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 594, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 595, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 595, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 596, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 596, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 597, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 597, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 598, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 598, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 599, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 599, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 600, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 600, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 601, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 601, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 602, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 602, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 603, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 603, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 604, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 604, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 605, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 605, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 606, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 606, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 607, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 607, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 608, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 608, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 609, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 609, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 610, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 610, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 611, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 611, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 612, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 612, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 613, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 613, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 614, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 614, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 615, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 615, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 616, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 616, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 617, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 617, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 618, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 618, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 619, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 619, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 620, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 620, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 621, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 621, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 622, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 622, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 623, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 623, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 624, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 624, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 625, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 625, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 626, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 626, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 627, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 627, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 628, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 628, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 629, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 629, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 630, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 630, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 631, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 631, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 632, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 632, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 633, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 633, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 634, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 634, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 635, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 635, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 636, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 636, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 637, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 637, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 638, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 638, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 639, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 639, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 640, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 640, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 641, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 641, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 642, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 642, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 643, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 643, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 644, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 644, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 645, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 645, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 646, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 646, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 647, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 647, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 648, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 648, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 649, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 649, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 650, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 650, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 651, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 651, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 652, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 652, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 653, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 653, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 654, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 654, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 655, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 655, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 656, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 656, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 657, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 657, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 658, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 658, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 659, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 659, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 660, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 660, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 661, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 661, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 662, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 662, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 663, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 663, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 664, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 664, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 665, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 665, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 666, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 666, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 667, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 667, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 668, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 668, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 669, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 669, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 670, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 670, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 671, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 671, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 672, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 672, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 673, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 673, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 674, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 674, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 675, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 675, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 676, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 676, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 677, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 677, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 678, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 678, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 679, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 679, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 680, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 680, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 681, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 681, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 682, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 682, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTInteger", "line_number": 683, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 683, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 684, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 684, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 685, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 685, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTReal", "line_number": 686, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 686, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 687, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 687, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 688, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 688, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 689, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 689, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 690, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 690, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 691, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 691, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 692, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 692, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 693, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 693, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 694, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 694, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 695, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 695, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 696, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 696, "usage_type": "name"}, {"api_name": "os.access", "line_number": 705, "usage_type": "call"}, {"api_name": "ieo.catgpkg", "line_number": 705, "usage_type": "attribute"}, {"api_name": "os.F_OK", "line_number": 705, "usage_type": "attribute"}, {"api_name": "ieo.catgpkg", "line_number": 707, "usage_type": "attribute"}, {"api_name": "ieo.catgpkg", "line_number": 709, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.wkbPolygon", "line_number": 717, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 717, "usage_type": "name"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 719, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 719, "usage_type": "name"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 724, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 724, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 724, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 725, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 725, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 725, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 726, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 726, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 726, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 727, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 727, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 727, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 728, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 728, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 728, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 729, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 729, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 729, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 730, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 730, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 730, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 731, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 731, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 731, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 732, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 732, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTString", "line_number": 732, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 736, "usage_type": "call"}, {"api_name": "os.path", "line_number": 736, "usage_type": "attribute"}, {"api_name": "ieo.catgpkg", "line_number": 748, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.FieldDefn", "line_number": 758, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 758, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 779, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 780, "usage_type": "call"}, {"api_name": "os.path", "line_number": 780, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 785, "usage_type": "call"}, {"api_name": "ieo.catgpkg", "line_number": 785, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 792, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 793, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 793, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 801, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 801, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 807, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 808, "usage_type": "call"}, {"api_name": "os.path", "line_number": 808, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 811, "usage_type": "call"}, {"api_name": "ieo.logerror", "line_number": 823, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 827, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 828, "usage_type": "call"}, {"api_name": "os.path", "line_number": 828, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 831, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 842, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 842, "usage_type": "attribute"}, {"api_name": "ieo.projacronym", "line_number": 844, "usage_type": "attribute"}, {"api_name": "ieo.btdir", "line_number": 844, "usage_type": "attribute"}, {"api_name": "ieo.fmaskdir", "line_number": 845, "usage_type": "attribute"}, {"api_name": "ieo.pixelqadir", "line_number": 846, "usage_type": "attribute"}, {"api_name": "ieo.ndvidir", "line_number": 847, "usage_type": "attribute"}, {"api_name": "ieo.evidir", "line_number": 848, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Feature", "line_number": 875, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 875, "usage_type": "name"}, {"api_name": "osgeo.ogr.OFTDate", "line_number": 881, "usage_type": "attribute"}, {"api_name": "osgeo.ogr", "line_number": 881, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 885, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 885, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 891, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 892, "usage_type": "call"}, {"api_name": "os.path", "line_number": 892, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 895, "usage_type": "call"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 900, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 900, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLinearRing", "line_number": 900, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 907, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 907, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbPolygon", "line_number": 907, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 914, "usage_type": "call"}, {"api_name": "os.path", "line_number": 914, "usage_type": "attribute"}, {"api_name": "os.access", "line_number": 915, "usage_type": "call"}, {"api_name": "os.F_OK", "line_number": 915, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 927, "usage_type": "call"}, {"api_name": "os.access", "line_number": 929, "usage_type": "call"}, {"api_name": "os.F_OK", "line_number": 929, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 933, "usage_type": "call"}, {"api_name": "os.path.split", "line_number": 934, "usage_type": "call"}, {"api_name": "os.path", "line_number": 934, "usage_type": "attribute"}, {"api_name": "ieo.logerror", "line_number": 936, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 936, "usage_type": "call"}, {"api_name": "os.path", "line_number": 936, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 948, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 948, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbLinearRing", "line_number": 948, "usage_type": "attribute"}, {"api_name": "osgeo.ogr.Geometry", "line_number": 955, "usage_type": "call"}, {"api_name": "osgeo.ogr", "line_number": 955, "usage_type": "name"}, {"api_name": "osgeo.ogr.wkbPolygon", "line_number": 955, "usage_type": "attribute"}]} +{"seq_id": "250878235", "text": "\"\"\"redstoneninjagaming URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/1.9/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.conf.urls import url, include\n 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))\n\"\"\"\nfrom django.conf.urls import url\nfrom django.contrib import admin\n\nfrom redstoneninjagaming import settings\nfrom redstoneninjagaming.views import homePage, about, creation, server, loginUser, logoutUser, signup, videolistmembers, member, response, home\n\nfrom members.views import membersHome, videofav, videolist, videomembers, cow\n\nurlpatterns = [\n url(r'^admin/', admin.site.urls),\n url(r'^$', homePage.as_view()),\n url(r'^about/$', about.as_view()), \n url(r'^creation/$', creation.as_view()), \n url(r'^server/$', server.as_view()), \n url(r'^login/$', loginUser.as_view()),\n url(r'^logout/$', logoutUser.as_view()),\n url(r'^signup/$', signup.as_view()),\n url(r'^videolistmembers/([0-9]{1,4})/$', videolistmembers.as_view()),\n url(r'^memberonly/', member.as_view()),\n url(r'^response/', response.as_view()),\n url(r'^creation/home', home.as_view()),\n \n url(r'^members/$', membersHome.as_view()),\n url(r'^videofav/$', videofav.as_view()),\n url(r'^videolist/$', videolist.as_view()),\n url(r'^videomembers/$', videomembers.as_view()),\n url(r'^creation/cow', cow.as_view()),\n]\n", "sub_path": "redstoneninjagaming/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1779, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.conf.urls.url", "line_number": 25, "usage_type": "call"}, {"api_name": "django.contrib.admin.site", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.contrib.admin", "line_number": 25, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.homePage.as_view", "line_number": 26, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.homePage", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.about.as_view", "line_number": 27, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.about", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 28, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.creation.as_view", "line_number": 28, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.creation", "line_number": 28, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 29, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.server.as_view", "line_number": 29, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.server", "line_number": 29, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.loginUser.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.loginUser", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.logoutUser.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.logoutUser", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 32, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.signup.as_view", "line_number": 32, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.signup", "line_number": 32, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 33, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.videolistmembers.as_view", "line_number": 33, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.videolistmembers", "line_number": 33, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.member.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.member", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 35, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.response.as_view", "line_number": 35, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.response", "line_number": 35, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 36, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.home.as_view", "line_number": 36, "usage_type": "call"}, {"api_name": "redstoneninjagaming.views.home", "line_number": 36, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 38, "usage_type": "call"}, {"api_name": "members.views.membersHome.as_view", "line_number": 38, "usage_type": "call"}, {"api_name": "members.views.membersHome", "line_number": 38, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 39, "usage_type": "call"}, {"api_name": "members.views.videofav.as_view", "line_number": 39, "usage_type": "call"}, {"api_name": "members.views.videofav", "line_number": 39, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "members.views.videolist.as_view", "line_number": 40, "usage_type": "call"}, {"api_name": "members.views.videolist", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "members.views.videomembers.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "members.views.videomembers", "line_number": 41, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 42, "usage_type": "call"}, {"api_name": "members.views.cow.as_view", "line_number": 42, "usage_type": "call"}, {"api_name": "members.views.cow", "line_number": 42, "usage_type": "name"}]} +{"seq_id": "252621064", "text": "#!/usr/bin/env python\nimport datetime\nimport json\nimport os\nimport re\nimport sys\nimport time\nimport pathlib\nimport configparser\nimport stat\n\nimport requests\nimport stravalib\n\n\nDAYS_WINDOW = int(os.getenv(\"RUNTASTIC_DAYS_WINDOW\", 3))\nSTRAVA_UPLOAD = \"upload@strava.com\"\n\nconfig = configparser.ConfigParser()\nconfig_path = pathlib.Path.home() / '.runtastic2strava.conf'\nconfig_file = str(config_path)\nif not config_path.is_file():\n print('no configuration file found ({})'.format(config_file))\n sys.exit(1)\nmode = stat.S_IMODE(config_path.stat().st_mode)\nif mode != 0o600:\n print('{} mode is not 0600'.format(config_file))\n sys.exit(1)\nconfig.read(config_file)\n\nconfigobj = config['DEFAULT']\nruntastic_email = configobj['runtastic_email']\nruntastic_password = configobj['runtastic_password']\nruntastic_username = configobj['runtastic_username']\nstrava_access_token = configobj['strava_access_token']\n\nlogin = requests.post(\"https://www.runtastic.com/en/d/users/sign_in\",\n data={\"user[email]\": runtastic_email,\n \"user[password]\": runtastic_password})\n\nif login.status_code // 100 != 2:\n print(\"Error logging in Runtastic, aborting\")\n\nresp = requests.get(\"https://www.runtastic.com/en/users/%s/sport-sessions\"\n % runtastic_username,\n cookies=login.cookies)\n\nif resp.status_code // 100 != 2:\n print(\"Error doing Runtastic request, aborting\")\n sys.exit(1)\n\nmatch_data = re.search(r\"index_data = ([^;]+);\", resp.text)\nif not match_data:\n print(\"Error looking for data, aborting\")\n sys.exit(1)\n\nactivities = json.loads(match_data.group(1))\n\nlast_sync_day = (datetime.datetime.utcnow()\n - datetime.timedelta(days=DAYS_WINDOW)).strftime(\"%Y-%m-%d\")\n\nclient = stravalib.Client(access_token=strava_access_token)\n\n# Only send the last N days of activities\nfor activity in filter(lambda a: a[1] >= last_sync_day, activities):\n activity_id = activity[0]\n filename = \"%s.tcx\" % activity_id\n filealreadyexists = pathlib.Path(filename).exists()\n if not filealreadyexists:\n while True:\n resp = requests.get(\n \"https://www.runtastic.com/en/users/%s/sport-sessions/%s.tcx\"\n % (runtastic_username, activity_id),\n cookies=login.cookies)\n if resp.status_code == 403:\n print('Runtastic query failed with 403, wait for 10 minutes and try again')\n time.sleep(600) # 10 minutes\n continue\n if resp.status_code != 200:\n raise Exception('Runtastic query failed')\n break\n else:\n print('file {} already exists'.format(filename))\n if filealreadyexists:\n mode = 'r'\n else:\n mode = 'w+'\n with open(filename, mode) as f:\n if not filealreadyexists:\n # save the file\n f.write(resp.text)\n f.seek(0)\n try:\n client.upload_activity(f, data_type=\"tcx\")\n print(\"Sent activity %s from %s\" % (activity_id, activity[1]))\n except stravalib.exc.ActivityUploadFailed as e:\n print(\"Failed to upload {} from {}: {}\".format(activity_id, activity[1], str(e)))\n if not ('duplicate' in str(e)\n or 'Unrecognized file type' in str(e)\n or 'The file is empty' in str(e)):\n raise\n", "sub_path": "runtastic2strava.py", "file_name": "runtastic2strava.py", "file_ext": "py", "file_size_in_byte": 3400, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.getenv", "line_number": 16, "usage_type": "call"}, {"api_name": "configparser.ConfigParser", "line_number": 19, "usage_type": "call"}, {"api_name": "pathlib.Path.home", "line_number": 20, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 24, "usage_type": "call"}, {"api_name": "stat.S_IMODE", "line_number": 25, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 37, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 50, "usage_type": "call"}, {"api_name": "re.search", "line_number": 52, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 55, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 57, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 60, "usage_type": "call"}, {"api_name": "stravalib.Client", "line_number": 62, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 68, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 71, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 77, "usage_type": "call"}, {"api_name": "stravalib.exc", "line_number": 96, "usage_type": "attribute"}]} +{"seq_id": "440463191", "text": "\"\"\"Contains tests for the private _solvers module\"\"\"\nfrom numpy import insert, loadtxt, allclose\nimport pytest\nfrom qmpy.solvers import schroedinger\nfrom qmpy._fileio import _read_schrodinger\n\n\nPROBLEMS = ['inf_potwell', 'fin_potwell', 'double_well', 'asym_potwell',\n 'harm_osci']\n\n\n@pytest.mark.parametrize('problem', PROBLEMS)\ndef test_computing(problem):\n \"\"\"\n Tests whether the computed wavefunctions and energies match the\n reference data.\n\n \"\"\"\n path = 'tests/test_data/{}.inp'.format(problem)\n specs = _read_schrodinger(path)\n vals = dict()\n vals['mass'] = specs['mass']\n vals['xcords'] = specs['interpolxydecs'][:, 0]\n vals['potential'] = specs['interpolxydecs'][:, 1]\n vals['xopt'], kind = (specs['xopt'], specs['interpoltype'])\n\n evs = (specs['first_ev'] - 1, specs['last_ev'] - 1)\n\n comp_energies, wfuncs, pot = schroedinger(vals, interpol=True,\n interpoltype=kind,\n select_range=evs)\n\n comp_funcs = insert(wfuncs.T, 0, values=pot[:, 1].T, axis=1)\n ref_energies = loadtxt('tests/test_data/energies_{}.ref'.format(problem))\n ref_wfuncs = loadtxt('tests/test_data/wfuncs_{}.ref'.format(problem))\n\n assert allclose(ref_energies, comp_energies)\n assert allclose(ref_wfuncs, comp_funcs)\n", "sub_path": "tests/test_solvers.py", "file_name": "test_solvers.py", "file_ext": "py", "file_size_in_byte": 1348, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "qmpy._fileio._read_schrodinger", "line_number": 20, "usage_type": "call"}, {"api_name": "qmpy.solvers.schroedinger", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.insert", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.loadtxt", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 38, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 12, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 12, "usage_type": "attribute"}]} +{"seq_id": "13033117", "text": "# -*- coding: UTF-8 -*-\n#######################################################################\n # ----------------------------------------------------------------------------\n # \"THE BEER-WARE LICENSE\" (Revision 42):\n # @tantrumdev wrote this file. As long as you retain this notice you\n # can do whatever you want with this stuff. If we meet some day, and you think\n # this stuff is worth it, you can buy me a beer in return. - Muad'Dib\n # ----------------------------------------------------------------------------\n#######################################################################\n\n# Addon Name: Placenta\n# Addon id: plugin.video.placenta\n# Addon Provider: MuadDib\n\nimport re\nimport urllib\nimport urlparse\nimport json\n\nfrom resources.lib.modules import cleantitle\nfrom resources.lib.modules import client\nfrom resources.lib.modules import source_utils\nfrom resources.lib.modules import dom_parser\nfrom resources.lib.modules import tvmaze\n\n\nclass source:\n def __init__(self):\n self.priority = 1\n self.language = ['ru']\n self.domains = ['filmix.me']\n self.base_link = 'https://filmix.me'\n self.search_link = '/engine/ajax/sphinx_search.php'\n self.search_old = '/search/%s'\n self.player_link = '/api/movies/player_data'\n\n def movie(self, imdb, title, localtitle, aliases, year):\n try:\n url = self.__search([localtitle] + source_utils.aliases_to_array(aliases), year)\n if not url and title != localtitle: url = self.__search([title] + source_utils.aliases_to_array(aliases), year)\n return urllib.urlencode({'url': url}) if url else None\n except:\n return\n\n def tvshow(self, imdb, tvdb, tvshowtitle, localtvshowtitle, aliases, year):\n try:\n url = self.__search([localtvshowtitle] + source_utils.aliases_to_array(aliases), year)\n if not url and tvshowtitle != localtvshowtitle: url = self.__search([tvshowtitle] + source_utils.aliases_to_array(aliases), year)\n return urllib.urlencode({'url': url, 'tvdb': tvdb}) if url else None\n except:\n return\n\n def episode(self, url, imdb, tvdb, title, premiered, season, episode):\n try:\n if not url:\n return\n\n data = urlparse.parse_qs(url)\n data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])\n data.update({'season': season, 'episode': episode})\n return urllib.urlencode(data)\n except:\n return\n\n def sources(self, url, hostDict, hostprDict):\n sources = []\n\n try:\n if not url:\n return sources\n\n data = urlparse.parse_qs(url)\n data = dict([(i, data[i][0]) if data[i] else (i, '') for i in data])\n url = data.get('url')\n season = data.get('season')\n episode = data.get('episode')\n abs_episode = 0\n\n if season and episode:\n abs_episode = str(tvmaze.tvMaze().episodeAbsoluteNumber(data.get('tvdb'), int(season), int(episode)))\n\n url = urlparse.urljoin(self.base_link, url)\n\n r = client.request(url)\n r = r.decode('cp1251').encode('utf-8')\n\n r = dom_parser.parse_dom(r, 'div', attrs={'class': 'players'}, req='data-player')\n r = [(i.attrs['data-player'], dom_parser.parse_dom(i, 'a', req='href')) for i in r]\n r = [(i[0], i[1][0].attrs['href']) for i in r if i[1]]\n\n for post_id, play_url in r:\n i = client.request(play_url, referer=url, output='extended')\n\n headers = i[3]\n headers.update({'Cookie': i[2].get('Set-Cookie')})\n\n i = client.request(urlparse.urljoin(self.base_link, self.player_link), post={'post_id': post_id}, headers=headers, referer=i, XHR=True)\n i = json.loads(i).get('message', {}).get('translations', {}).get('flash', {})\n\n for title, link in i.iteritems():\n try:\n link = self.decode_direct_media_url(link)\n\n if link.endswith('.txt'):\n link = self.decode_direct_media_url(client.request(link))\n link = json.loads(link).get('playlist', [])\n link = [i.get('playlist', []) for i in link]\n link = [x.get('file') for i in link for x in i if (x.get('season') == season and x.get('serieId') == episode) or (x.get('season') == '0' and x.get('serieId') == abs_episode)][0]\n\n urls = [(source_utils.label_to_quality(q), self.format_direct_link(link, q)) for q in self.get_qualitys(link)]\n urls = [{'quality': x[0], 'url': x[1]} for x in urls if x[0] in ['SD', 'HD']] # filter premium\n\n for i in urls: sources.append({'source': 'CDN', 'quality': i['quality'], 'info': title, 'language': 'ru', 'url': i['url'], 'direct': True, 'debridonly': False})\n except:\n pass\n\n return sources\n except:\n return sources\n\n def resolve(self, url):\n return url\n\n def __search(self, titles, year):\n try:\n url = urlparse.urljoin(self.base_link, self.search_link)\n\n t = [cleantitle.get(i) for i in set(titles) if i]\n y = ['%s' % str(year), '%s' % str(int(year) + 1), '%s' % str(int(year) - 1), '0']\n\n post = {'story': titles[0], 'years_ot': str(int(year) - 1), 'years_do': str(int(year) + 1)}\n r = client.request(url, post=post, XHR=True)\n\n if len(r) < 1000:\n url = urlparse.urljoin(self.base_link, self.search_old % urllib.quote_plus(titles[0]))\n r = client.request(url)\n\n r = r.decode('cp1251').encode('utf-8')\n\n r = dom_parser.parse_dom(r, 'article')\n r = dom_parser.parse_dom(r, 'div', attrs={'class': 'full'})\n r = [(dom_parser.parse_dom(i, 'a', attrs={'itemprop': 'url'}, req='href'),\n dom_parser.parse_dom(i, 'h3', attrs={'class': 'name'}, req='content'),\n dom_parser.parse_dom(i, 'div', attrs={'class': 'origin-name'}, req='content'),\n dom_parser.parse_dom(i, 'div', attrs={'class': 'year'})) for i in r]\n r = [(i[0][0].attrs['href'], i[1][0].attrs['content'], i[2][0].attrs['content'], dom_parser.parse_dom(i[3], 'a', attrs={'itemprop': 'copyrightYear'})) for i in r if i[0] and i[1] and i[2]]\n r = [(i[0], i[1], i[2], i[3][0].content) for i in r if i[3]]\n r = [i[0] for i in r if (cleantitle.get(i[1]) in t or cleantitle.get(i[2]) in t) and i[3] in y][0]\n\n return source_utils.strip_domain(r)\n except:\n return\n\n ########################\n # Credits to evgen_dev #\n ########################\n\n def decode_direct_media_url(self, encoded_url):\n import base64\n codec_a = (\"l\", \"u\", \"T\", \"D\", \"Q\", \"H\", \"0\", \"3\", \"G\", \"1\", \"f\", \"M\", \"p\", \"U\", \"a\", \"I\", \"6\", \"k\", \"d\", \"s\", \"b\", \"W\", \"5\", \"e\", \"y\", \"=\")\n codec_b = (\"w\", \"g\", \"i\", \"Z\", \"c\", \"R\", \"z\", \"v\", \"x\", \"n\", \"N\", \"2\", \"8\", \"J\", \"X\", \"t\", \"9\", \"V\", \"7\", \"4\", \"B\", \"m\", \"Y\", \"o\", \"L\", \"h\")\n i = 0\n for a in codec_a:\n b = codec_b[i]\n i += 1\n encoded_url = encoded_url.replace(a, '___')\n encoded_url = encoded_url.replace(b, a)\n encoded_url = encoded_url.replace('___', b)\n\n return base64.b64decode(encoded_url)\n\n def get_qualitys(self, source_link):\n try:\n avail_quality = re.compile(\"\\[([^\\]]+)\\]\", re.S).findall(source_link)[0]\n return [i for i in avail_quality.split(',') if i]\n except:\n return '0'.split()\n\n def format_direct_link(self, source_link, q):\n regex = re.compile(\"\\[([^\\]]+)\\]\", re.IGNORECASE)\n return regex.sub(q, source_link)\n", "sub_path": "script.module.placenta/lib/resources/lib/sources/ru/filmix.py", "file_name": "filmix.py", "file_ext": "py", "file_size_in_byte": 7979, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "resources.lib.modules.source_utils.aliases_to_array", "line_number": 39, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 39, "usage_type": "name"}, {"api_name": "resources.lib.modules.source_utils.aliases_to_array", "line_number": 40, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 40, "usage_type": "name"}, {"api_name": "urllib.urlencode", "line_number": 41, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils.aliases_to_array", "line_number": 47, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 47, "usage_type": "name"}, {"api_name": "resources.lib.modules.source_utils.aliases_to_array", "line_number": 48, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 48, "usage_type": "name"}, {"api_name": "urllib.urlencode", "line_number": 49, "usage_type": "call"}, {"api_name": "urlparse.parse_qs", "line_number": 58, "usage_type": "call"}, {"api_name": "urllib.urlencode", "line_number": 61, "usage_type": "call"}, {"api_name": "urlparse.parse_qs", "line_number": 72, "usage_type": "call"}, {"api_name": "resources.lib.modules.tvmaze.tvMaze", "line_number": 80, "usage_type": "call"}, {"api_name": "resources.lib.modules.tvmaze", "line_number": 80, "usage_type": "name"}, {"api_name": "urlparse.urljoin", "line_number": 82, "usage_type": "call"}, {"api_name": "resources.lib.modules.client.request", "line_number": 84, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 84, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 87, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 87, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 88, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 88, "usage_type": "name"}, {"api_name": "resources.lib.modules.client.request", "line_number": 92, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 92, "usage_type": "name"}, {"api_name": "resources.lib.modules.client.request", "line_number": 97, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 97, "usage_type": "name"}, {"api_name": "urlparse.urljoin", "line_number": 97, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 98, "usage_type": "call"}, {"api_name": "resources.lib.modules.client.request", "line_number": 105, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 105, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 106, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils.label_to_quality", "line_number": 110, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 110, "usage_type": "name"}, {"api_name": "urlparse.urljoin", "line_number": 126, "usage_type": "call"}, {"api_name": "resources.lib.modules.cleantitle.get", "line_number": 128, "usage_type": "call"}, {"api_name": "resources.lib.modules.cleantitle", "line_number": 128, "usage_type": "name"}, {"api_name": "resources.lib.modules.client.request", "line_number": 132, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 132, "usage_type": "name"}, {"api_name": "urlparse.urljoin", "line_number": 135, "usage_type": "call"}, {"api_name": "urllib.quote_plus", "line_number": 135, "usage_type": "call"}, {"api_name": "resources.lib.modules.client.request", "line_number": 136, "usage_type": "call"}, {"api_name": "resources.lib.modules.client", "line_number": 136, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 140, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 140, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 141, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 141, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 142, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 142, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 143, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 143, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 144, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 144, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 145, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 145, "usage_type": "name"}, {"api_name": "resources.lib.modules.dom_parser.parse_dom", "line_number": 146, "usage_type": "call"}, {"api_name": "resources.lib.modules.dom_parser", "line_number": 146, "usage_type": "name"}, {"api_name": "resources.lib.modules.cleantitle.get", "line_number": 148, "usage_type": "call"}, {"api_name": "resources.lib.modules.cleantitle", "line_number": 148, "usage_type": "name"}, {"api_name": "resources.lib.modules.source_utils.strip_domain", "line_number": 150, "usage_type": "call"}, {"api_name": "resources.lib.modules.source_utils", "line_number": 150, "usage_type": "name"}, {"api_name": "base64.b64decode", "line_number": 170, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 174, "usage_type": "call"}, {"api_name": "re.S", "line_number": 174, "usage_type": "attribute"}, {"api_name": "re.compile", "line_number": 180, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 180, "usage_type": "attribute"}]} +{"seq_id": "92873300", "text": "import time\nimport argparse\nimport socket\n\nfrom numpy.lib.shape_base import split\nfrom utilities import *\nfrom models import *\nimport models as lib\nfrom data_loader import *\nimport wandb\nimport matplotlib.pyplot as plt\n\nfrom collections import Counter\nfrom sklearn.svm import LinearSVC\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.dummy import DummyClassifier\nfrom sklearn.metrics import f1_score\n\n# FIX THE RANDOM SEED FOR REPRODUCIBILITY\nSEED = 1234\ntorch.manual_seed(0)\nnp.random.seed(0)\ntorch.backends.cudnn.benchmark = False # reduces the performance\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--data', required=True, help='Data file')\nparser.add_argument('--model', default=\"mondrain_v0.1\", help='Model name to save output in file')\nparser.add_argument('--gpu', default=-1, type=int, help='ID of the gpu to run on. If set to -1 (default), the GPU with most free memory will be chosen.')\nparser.add_argument('--epochs', default=20, type=int, help='Number of epochs to train the model')\nparser.add_argument('--embedding_dim', default=128, type=int, help='Number of dimensions of the dynamic embedding')\nparser.add_argument('--train_proportion', default=0.8, type=float, help='Fraction of interactions (from the beginning) that are used for training.The next 10% are used for validation and the next 10% for testing')\nparser.add_argument('-ws', '--wandb_sync', '--wandb_sync=1', action='store_true', help='Check if the run is going to be uploaded to WandB')\nparser.add_argument('--state_change', default=False, type=bool, help='True if training with state change of users along with interaction prediction. False otherwise. By default, set to True.')\nparser.add_argument('--full_train', default=False, action='store_true', help='Enables the full train mode training the model with all the dataset to calculate the embeddings in the final interaction.')\nparser.add_argument('--tqdmdis', action='store_true', help='Enable or disable TQDM progress bar.')\nparser.add_argument('--disable_train', action='store_true', help='Enable or disable train phase.')\nparser.add_argument('--split', default=1, type=int, help='The split of the pipeline') \nparser.add_argument('--tags', help='Tags for WandB')\n\nargs = parser.parse_args()\n\n# Set the name of the data file \nargs.dataname = args.data.split('.')[0].split('/')[-1]\nif args.tags is not None:\n tags = [socket.gethostname(), args.tags, args.dataname]\nelse:\n tags = [socket.gethostname(), args.dataname]\n \nif not args.wandb_sync:\n os.environ['WANDB_MODE'] = 'dryrun'\n\nif args.full_train:\n tags.append('full_train')\n\nwandb.init(project=\"mondrian\", config=args, tags=tags)\n\noutput_fname = \"results/interaction_prediction_%s.txt\" % args.data\n\n# SET GPU\nif args.gpu == -1:\n args.gpu = select_free_gpu()\nos.environ[\"CUDA_DEVICE_ORDER\"] = \"PCI_BUS_ID\"\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = args.gpu\n\n# LOAD DATA\n(user2id, user_list_id, user_timediference_list, user_previous_actionid_list,\n action2id, action_list_id, action_timediference_list, \n timestamp_list,\n feature_list, \n head_labels, \n tail_labels) = load_data(args, tail_as_feature=True)\n\nnum_users = len(user2id)\nnum_actions = len(action2id) + 1 # If the previous action is none\nnum_interactions = len(user_list_id)\nnum_features = len(feature_list[0])\ny_true = action_list_id\nusers_edited = [False] * num_users\nprint(\"*** Network statistics:\\n %d users\\n %d action types\\n %d features\\n %d interactions\\n ***\\n\\n\" % (num_users, num_actions, num_features, num_interactions))\n\n# SET TRAINING, VALIDATION AND TESTING SPLITS\n# train_end_idx = validation_start_idx = int(num_interactions * args.train_proportion) \n# test_start_idx = int(num_interactions * (args.train_proportion + 0.1))\n# test_end_idx = int(num_interactions * (args.train_proportion + 0.2))\n\n\ntrain_end_idx = int(num_interactions * ((0.25) * args.split))\ntest_start_idx = train_end_idx + 1\ntest_end_idx = num_interactions\n\n\nif args.full_train:\n end_idx = num_interactions\nelse:\n end_idx = train_end_idx\n\n# INITIALIZE MODEL AND PARAMETERS\nmodel = Mondrian(args, num_users, num_actions, num_features).cuda()\nmodel.cuda()\nMSELoss = nn.MSELoss()\n\n# INITIALIZE MODEL\nlearning_rate = 1e-3\noptimizer = optim.Adam(model.parameters(), lr=learning_rate, weight_decay=1e-5)\n\n# INITIALIZE EMBEDDING\ninitial_user_embedding = nn.Parameter(F.normalize(torch.rand(args.embedding_dim).cuda(), dim=0))\ninitial_action_embedding = nn.Parameter(F.normalize(torch.rand(args.embedding_dim).cuda(), dim=0))\nmodel.initial_user_embedding = initial_user_embedding\nmodel.initial_action_embedding = initial_action_embedding\n\nuser_embeddings = initial_user_embedding.repeat(num_users, 1) # initialize all users to the same embedding \naction_embeddings = initial_action_embedding.repeat(num_actions, 1)\nuser_embedding_static = Variable(torch.eye(num_users).cuda()) # one-hot vectors for static embeddings\naction_embedding_static = Variable(torch.eye(num_actions).cuda())\n\n# PERFORMANCE METRICS\nvalidation_ranks = []\ntest_ranks = []\nmean_loss = []\n\n# WANDB\nwandb.watch(model)\n\nif args.full_train:\n print('********* WARNING! ---- THIS IS A FULL TRAIN EXPERIMENT ----')\n\n# RUN THE MONDRAIN MODEL\n'''\nTHE MODEL IS TRAINED FOR SEVERAL EPOCHS. IN EACH EPOCH, JODIES USES THE TRAINING SET OF INTERACTIONS TO UPDATE ITS PARAMETERS.\nAFTER EACH EPOCH THE MODEL IS TESTED WITH A VALIDATION SET TO EARLY STOP\n'''\nprint(\"*** Training the MONDRIAN model for %d epochs ***\" % args.epochs)\nif not args.disable_train:\n with trange(1, args.epochs+1, disable=args.tqdmdis) as progress_bar1:\n \n for ep in progress_bar1:\n progress_bar1.set_description('Epoch %d of %d' % (ep, args.epochs))\n \n # print(5*'*' + ' START EPOCH -- Memory allocated: ' + str(torch.cuda.max_memory_allocated()) + 5*'*')\n epoch_start_time = time.time()\n \n # INITIALIZE EMBEDDING TRAJECTORY STORAGE\n user_embeddings_timeseries = Variable(torch.Tensor(num_interactions, args.embedding_dim).cuda())\n action_embeddings_timeseries = Variable(torch.Tensor(num_interactions, args.embedding_dim).cuda())\n\n optimizer.zero_grad()\n total_loss, loss, total_interaction_count = 0, 0, 0\n \n\n model.train()\n\n with trange(end_idx, disable=args.tqdmdis) as progress_bar2:\n for j in progress_bar2:\n progress_bar2.set_description('Processed %dth interactions' % j)\n userid = user_list_id[j]\n actionid = action_list_id[j]\n feature = feature_list[j]\n previous_actionid = user_previous_actionid_list[j]\n \n users_edited[userid] = True\n\n feature_tensor = torch.Tensor(feature).cuda().unsqueeze(0)\n user_timediffs_tensor = torch.Tensor([user_timediference_list[j]]).cuda().unsqueeze(0)\n action_timediffs_tensor = torch.Tensor([action_timediference_list[j]]).cuda().unsqueeze(0)\n \n action_embedding_previous = action_embeddings[previous_actionid,:].unsqueeze(0)\n\n actual_user_embedding_static = user_embedding_static[userid,:].unsqueeze(0)\n previous_action_embedding_static = action_embedding_static[previous_actionid,:].unsqueeze(0)\n\n # PROJECT USER EMBEDDING TO CURRENT TIME\n user_embedding_input = user_embeddings[userid,:].unsqueeze(0)\n user_projected_embedding = model.forward(user_embedding_input, action_embedding_previous, timediffs=user_timediffs_tensor, features=feature_tensor, select='project')\n user_action_embedding = torch.cat([user_projected_embedding, action_embedding_previous, previous_action_embedding_static, actual_user_embedding_static], dim=1)\n\n # PREDICT NEXT ACTION EMBEDDING \n predicted_action_embedding = model.predict_action_embedding(user_action_embedding)\n \n # CALCULATE PREDICTION LOSS\n action_embedding_input = action_embeddings[actionid,:].unsqueeze(0)\n action_embedding_static_sq = action_embedding_static[actionid,:].unsqueeze(0)\n loss += MSELoss(predicted_action_embedding, torch.cat([action_embedding_input, action_embedding_static_sq], dim=1).detach())\n\n # UPDATE DYNAMIC EMBEDDINGS AFTER ACTION\n user_embedding_output = model.forward(user_embedding_input, action_embedding_input, timediffs=user_timediffs_tensor, features=feature_tensor, select='user_update')\n action_embedding_output = model.forward(user_embedding_input, action_embedding_input, timediffs=action_timediffs_tensor, features=feature_tensor, select='action_update')\n\n # CALCULATE LOSS TO MAINTAIN TEMPORAL SMOOTHNESS\n loss += MSELoss(action_embedding_output, action_embedding_input.detach())\n loss += MSELoss(user_embedding_output, user_embedding_input.detach())\n\n # loss.requires_grad = True\n # BACKPROPAGATE ERROR \n total_loss += loss.item()\n mean_loss.append(loss.item())\n loss.backward()\n optimizer.step()\n optimizer.zero_grad()\n\n user_embeddings[userid,:] = user_embedding_output\n action_embeddings[actionid,:] = action_embedding_output\n\n user_embeddings_timeseries[j,:] = user_embedding_output\n action_embeddings_timeseries[j,:] = action_embedding_output\n\n loss = 0\n action_embeddings.detach_() # Detachment is needed to prevent double propagation of gradient\n user_embeddings.detach_()\n \n action_embeddings_timeseries.detach_() \n user_embeddings_timeseries.detach_()\n\n wandb.log({'epoch': ep,\n 'total_loss_mean': np.mean(mean_loss)})\n \n # print(5*'*' + ' END TRAIN -- Memory allocated: ' + str(torch.cuda.max_memory_allocated()) + 5*'*')\n print(\"\\nLast epoch took {} minutes\".format((time.time()-epoch_start_time)/60))\n \n #END OF ONE EPOCH\n print(\"\\nTotal loss in this epoch = %f\\n\" % (total_loss))\n\n # UNTIL MEMORY MANAGEMENT\n user_embeddings_dystat = torch.cat([user_embeddings.cpu(), user_embedding_static.cpu()], dim=1)\n action_embeddings_dystat = torch.cat([action_embeddings.cpu(), action_embedding_static.cpu()], dim=1) \n \n \n if args.full_train and ep == args.epochs:\n prepare_data_folder(args)\n args.model = 'mondrain_full_v0.1'\n torch.save(user_embeddings_dystat, 'embeddings/%s/full_%s_embeddings_ep%s.pt' % (args.dataname, args.dataname, str(ep)))\n \n if ep == args.epochs:\n save_model(model, optimizer, args, ep, user_embeddings_dystat, action_embeddings_dystat, train_end_idx, user_embeddings_timeseries, action_embeddings_timeseries)\n \n\n user_embeddings = initial_user_embedding.repeat(num_users, 1)\n action_embeddings = initial_action_embedding.repeat(num_actions, 1)\n # print(5*'*' + ' END EPOCH -- Memory allocated: ' + str(torch.cuda.max_memory_allocated()) + 5*'*')\n\n # END OF ALL EPOCHS. SAVE FINAL MODEL DISK TO BE USED IN EVALUATION.\n print(\"\\n\\n*** Training complete. Saving final model. ***\\n\\n\")\n save_model(model, optimizer, args, ep, user_embeddings_dystat, action_embeddings_dystat, train_end_idx, user_embeddings_timeseries, action_embeddings_timeseries)\n\nif not args.full_train:\n ###### TEST STEP ######\n\n # INITIALIZE EMBEDDING TRAJECTORY STORAGE \n user_embeddings_timeseries = Variable(torch.Tensor(num_interactions, args.embedding_dim).cuda())\n action_embeddings_timeseries = Variable(torch.Tensor(num_interactions, args.embedding_dim).cuda())\n \n # PERFORMANCE METRICS\n validation_ranks = []\n test_ranks = []\n\n loss = 0\n # FORWARD PASS\n print(\"*** Making interaction predictions by forward pass (no t-batching) ***\")\n with trange(train_end_idx, test_end_idx) as progress_bar:\n for j in progress_bar:\n progress_bar.set_description('%dth interaction for validation and testing' % j)\n\n # LOAD INTERACTION J\n userid = user_list_id[j]\n actionid = action_list_id[j]\n feature = feature_list[j]\n user_timediff = user_timediference_list[j]\n action_timediff = action_timediference_list[j]\n # timestamp = timestamp_sequence[j]\n # if not tbatch_start_time:\n # tbatch_start_time = timestamp\n actionid_previous = user_previous_actionid_list[j]\n\n # LOAD USER AND action EMBEDDING\n user_embedding_input = user_embeddings[torch.cuda.LongTensor([userid])]\n user_embedding_static_input = user_embedding_static[torch.cuda.LongTensor([userid])]\n action_embedding_input = action_embeddings[torch.cuda.LongTensor([actionid])]\n action_embedding_static_input = action_embedding_static[torch.cuda.LongTensor([actionid])]\n feature_tensor = Variable(torch.Tensor(feature).cuda()).unsqueeze(0)\n user_timediffs_tensor = Variable(torch.Tensor([user_timediff]).cuda()).unsqueeze(0)\n action_timediffs_tensor = Variable(torch.Tensor([action_timediff]).cuda()).unsqueeze(0)\n action_embedding_previous = action_embeddings[torch.cuda.LongTensor([actionid_previous])]\n\n # PROJECT USER EMBEDDING\n user_projected_embedding = model.forward(user_embedding_input, action_embedding_previous, timediffs=user_timediffs_tensor, features=feature_tensor, select='project')\n user_action_embedding = torch.cat([user_projected_embedding, action_embedding_previous, action_embedding_static[torch.cuda.LongTensor([actionid_previous])], user_embedding_static_input], dim=1)\n\n # PREDICT ACTION EMBEDDING\n predicted_action_embedding = model.predict_action_embedding(user_action_embedding)\n\n # CALCULATE PREDICTION LOSS\n loss += MSELoss(predicted_action_embedding, torch.cat([action_embedding_input, action_embedding_static_input], dim=1).detach())\n \n # CALCULATE DISTANCE OF PREDICTED ACTION EMBEDDING TO ALL actionS \n euclidean_distances = nn.PairwiseDistance()(predicted_action_embedding.repeat(num_actions, 1), torch.cat([action_embeddings, action_embedding_static], dim=1)).squeeze(-1) \n \n # CALCULATE RANK OF THE TRUE ACTION AMONG ALL ACTIONS\n true_action_distance = euclidean_distances[actionid]\n euclidean_distances_smaller = (euclidean_distances < true_action_distance).data.cpu().numpy()\n true_action_rank = np.sum(euclidean_distances_smaller) + 1\n\n if j < test_start_idx:\n validation_ranks.append(true_action_rank)\n else:\n test_ranks.append(true_action_rank)\n\n # UPDATE USER AND action EMBEDDING\n user_embedding_output = model.forward(user_embedding_input, action_embedding_input, timediffs=user_timediffs_tensor, features=feature_tensor, select='user_update') \n action_embedding_output = model.forward(user_embedding_input, action_embedding_input, timediffs=action_timediffs_tensor, features=feature_tensor, select='action_update') \n\n # SAVE EMBEDDINGS\n action_embeddings[actionid,:] = action_embedding_output.squeeze(0) \n user_embeddings[userid,:] = user_embedding_output.squeeze(0) \n user_embeddings_timeseries[j, :] = user_embedding_output.squeeze(0)\n action_embeddings_timeseries[j, :] = action_embedding_output.squeeze(0)\n\n # CALCULATE LOSS TO MAINTAIN TEMPORAL SMOOTHNESS\n loss += MSELoss(action_embedding_output, action_embedding_input.detach())\n loss += MSELoss(user_embedding_output, user_embedding_input.detach())\n\n # CALCULATE STATE CHANGE LOSS\n if args.state_change:\n loss += calculate_state_prediction_loss(model, [j], user_embeddings_timeseries, y_true, crossEntropyLoss) \n\n # UPDATE THE MODEL IN REAL-TIME USING ERRORS MADE IN THE PAST PREDICTION\n # if timestamp - tbatch_start_time > tbatch_timespan:\n # tbatch_start_time = timestamp\n loss.backward()\n optimizer.step()\n optimizer.zero_grad()\n \n # RESET LOSS FOR NEXT T-BATCH\n loss = 0\n action_embeddings.detach_()\n user_embeddings.detach_()\n action_embeddings_timeseries.detach_() \n user_embeddings_timeseries.detach_()\n \n prepare_data_folder(args)\n torch.save(user_embeddings, 'embeddings/%s/predicted_%s_embeddings_sp%s.pt' % (args.dataname, args.dataname, args.split))\n\n ###### CLASIFYING STEP ######\n\n users_edited = [False] * num_users\n embeddings = torch.load('embeddings/%s/predicted_%s_embeddings_sp%s.pt' % (args.dataname, args.dataname, args.split))\n full_embeddings = torch.load('embeddings/%s/full_%s_embeddings_ep%s.pt' % (args.dataname, args.dataname, args.epochs))\n \n for i in range(train_end_idx):\n userid = user_list_id[i]\n users_edited[userid] = True\n\n X = []\n y = []\n x_test = []\n y_test = []\n print(len(head_labels))\n for i in range(num_users):\n if users_edited[i]:\n X.append(full_embeddings[i].tolist()[0:128])\n y.append(head_labels[user_list_id.index(i)])\n else:\n x_test.append(embeddings[i].tolist())\n y_test.append(head_labels[user_list_id.index(i)])\n\n\n print(5 * '*' + ' Data distribution ' + 5 * '*')\n print('Train: ')\n print(Counter(y))\n print('Test: ')\n print(Counter(y_test))\n print('\\n')\n # width = 0.35\n # ind = np.arange(2)\n # plot_0 = []\n # plot_1 = []\n # plot_0.append(Counter(y)[0])\n # plot_0.append(Counter(y_test)[0])\n # plot_1.append(Counter(y)[1])\n # plot_1.append(Counter(y_test)[1])\n # p1 = plt.bar(ind, plot_0, width)\n # p2 = plt.bar(ind, plot_1, width)\n\n # plt.ylabel('Quantity')\n # plt.title('User distribution')\n # plt.xticks(ind, ('Train', 'Test'))\n # plt.legend((p1[0], p2[0]), ('Malicious', 'Legit'))\n \n # wandb.log({\"User Distribution\": plt})\n \n wandb.log({'LEGIT_TRAIN': Counter(y)[0],\n 'MALICIOUS_TRAIN':Counter(y)[1],\n 'LEGIT_TEST': Counter(y_test)[0],\n 'MALICIOUS_TEST':Counter(y_test)[1] })\n \n # Linear SVC\n clf = LinearSVC(class_weight='balanced')\n clf.fit(X, y)\n predictions = clf.predict(x_test)\n svc_acc = clf.score(x_test, y_test)\n svc_f1 = f1_score(y_test, predictions, average='macro')\n # print('LinearSVC: %f' % score)\n\n\n # KNN\n clf = KNeighborsClassifier(n_neighbors=5)\n clf.fit(X, y)\n predictions = clf.predict(x_test)\n knn_acc = clf.score(x_test, y_test)\n knn_f1 = f1_score(y_test, predictions, average='macro')\n # print('KNN: %f' % score)\n\n # MLP\n clf = MLPClassifier(random_state=1, max_iter=300)\n clf.fit(X, y)\n predictions = clf.predict(x_test)\n mlp_acc = clf.score(x_test, y_test)\n mlp_f1 = f1_score(y_test, predictions, average='macro')\n # print('MLP: %f' % score)\n \n clf = RandomForestClassifier()\n clf.fit(X, y)\n predictions = clf.predict(x_test)\n rf_acc = clf.score(x_test, y_test)\n rf_f1 = f1_score(y_test, predictions, average='macro')\n \n clf = DummyClassifier(strategy='constant', random_state=SEED, constant=0)\n clf.fit(X, y)\n predictions = clf.predict(x_test)\n dummy_acc = clf.score(x_test, y_test)\n dummy_f1 = f1_score(y_test, predictions, average='macro')\n\n \n wandb.log({'SVC_ACC': svc_acc,\n 'SVC_F1': svc_f1,\n 'KNN_ACC': knn_acc,\n 'KNN_F1': knn_f1,\n 'MLP_ACC': mlp_acc,\n 'MLP_F1': mlp_f1,\n 'RF_ACC': rf_acc,\n 'RF_F1': rf_f1,\n 'DUMMY_ACC': dummy_acc,\n 'DUMMY_F1': dummy_f1})\n \n ", "sub_path": "old_jodie/pipeline_mondrian_full.py", "file_name": "pipeline_mondrian_full.py", "file_ext": "py", "file_size_in_byte": 20627, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 27, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 47, "usage_type": "call"}, {"api_name": "socket.gethostname", "line_number": 49, "usage_type": "call"}, {"api_name": "wandb.init", "line_number": 57, "usage_type": "call"}, {"api_name": "wandb.watch", "line_number": 125, "usage_type": "call"}, {"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 216, "usage_type": "call"}, {"api_name": "time.time", "line_number": 220, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 370, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 372, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 392, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 392, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 393, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 394, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 395, "usage_type": "call"}, {"api_name": "sklearn.svm.LinearSVC", "line_number": 398, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 402, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 407, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 411, "usage_type": "call"}, {"api_name": "sklearn.neural_network.MLPClassifier", "line_number": 415, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 419, "usage_type": "call"}, {"api_name": "sklearn.ensemble.RandomForestClassifier", "line_number": 422, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 426, "usage_type": "call"}, {"api_name": "sklearn.dummy.DummyClassifier", "line_number": 428, "usage_type": "call"}, {"api_name": "sklearn.metrics.f1_score", "line_number": 432, "usage_type": "call"}, {"api_name": "wandb.log", "line_number": 435, "usage_type": "call"}]} +{"seq_id": "143360486", "text": "# -*- coding: utf-8 -*-\r\n# from __future__ import unicode_literals\r\n\r\nimport os\r\nimport cStringIO\r\nimport time\r\nimport base64\r\nfrom contextlib import closing\r\nimport utils\r\nimport threading\r\nimport logger\r\nfrom config import config\r\nimport requests\r\nimport math\r\nfrom utils import fmt\r\n\r\nlog = logger.getLogger(config['NAME'], config['LOGLEVEL'])\r\n\r\n\r\nclass Screenshoter(threading.Thread):\r\n \"\"\"\r\n Делает скриншоты каждые 2с и пытается отправить на сервер.\r\n Если нет соединения с сервером, то сохраняет последние config[SAVED_IMAGES],\r\n которые будут отправлены на сервер модулем \"uploader\" при восстановлении подключения.\r\n\r\n Capturing screenshots every 2s and try to upload to server.\r\n If no connection, then save it, and will be uploaded later, when connection recover.\r\n \"\"\"\r\n\r\n FLAG = 1 # Битовый флаг запуска\r\n\r\n def __init__(self):\r\n log.info(fmt('Init daemon: {0}', __name__))\r\n threading.Thread.__init__(self)\r\n self.name = __name__\r\n self.daemon = True\r\n self.active = False\r\n\r\n self.datadir = os.path.join(config['HOME_DIR'], config['NAME'], utils.get_user_name())\r\n self.imagesdir = os.path.join(self.datadir, 'images')\r\n self.quality = config['SCR_QUALITY']\r\n self.url = config['URL'][0]\r\n self.maxRMS = config['CHGSCR_THRESHOLD']\r\n self.auth = requests.auth.HTTPDigestAuth(*config['AUTH'])\r\n self.img1_histogram, self.img2_histogram = None, None\r\n self.params = {\"username\": utils.utf(utils.get_user_name()),\r\n 'compname': utils.utf(utils.get_comp_name())}\r\n self.jreq = {'jsonrpc': '2.0', 'method': 'image', 'id': __name__, 'params': self.params}\r\n\r\n self.headers = {'user-agent': fmt(\"{NAME}/{VERSION}\", **config)}\r\n\r\n utils.makedirs(self.datadir)\r\n\r\n def stop(self):\r\n log.info(fmt('Stop daemon: {0}', self.name))\r\n self.active = False\r\n\r\n def compare_images(self):\r\n \"\"\"\r\n Сравнение изображений методом среднеквадратичного отклонения\r\n :return: среднеквадратичное отклонение. Если 0, то изображения одинаковы\r\n \"\"\"\r\n h1 = self.img1_histogram\r\n h2 = self.img2_histogram\r\n rms = math.sqrt(sum(map(lambda a, b: (a - b) ** 2, h1, h2)) / len(h1))\r\n return rms\r\n\r\n def grabImage(self):\r\n try:\r\n return self._grabImage_win32()\r\n except Exception as e:\r\n log.error(e)\r\n return self._grabImage_wx()\r\n\r\n def _grabImage_wx(self):\r\n from PIL import Image\r\n import wx\r\n bt = time.time()\r\n try:\r\n app = wx.App() # Need to create an App instance before doing anything\r\n screen = wx.ScreenDC()\r\n size = screen.GetSize()\r\n bmp = wx.Bitmap(size[0], size[1])\r\n mem = wx.MemoryDC(bmp)\r\n mem.Blit(0, 0, size[0], size[1], screen, 0, 0)\r\n myWxImage = bmp.ConvertToImage()\r\n return Image.frombytes('RGB', size.Get(), myWxImage.GetDataBuffer().tobytes())\r\n finally:\r\n log.debug(fmt(\"time of execution = {t}\", t=time.time() - bt))\r\n\r\n def _grabImage_PIL(self):\r\n \"\"\"\r\n Делает скриншот с помощью Pillow библиотеки.\r\n ImageGrab.grab() порождает большую утечку памяти, если при вызове произошла ошибка.\r\n \"\"\"\r\n from PIL import ImageGrab\r\n bt = time.time()\r\n try:\r\n return ImageGrab.grab()\r\n finally:\r\n log.debug(fmt(\"time of execution = {t}\", t=time.time() - bt))\r\n\r\n def _grabImage_win32(self):\r\n \"\"\"\r\n Делает скриншот с помощью win32 api.\r\n \"\"\"\r\n import win32gui\r\n import win32ui\r\n import win32con\r\n import win32api\r\n from PIL import Image\r\n bt = time.time()\r\n bmp = win32ui.CreateBitmap()\r\n try:\r\n CAPTUREBLT = 0x40000000\r\n hwin = win32gui.GetDesktopWindow()\r\n width = win32api.GetSystemMetrics(win32con.SM_CXVIRTUALSCREEN)\r\n height = win32api.GetSystemMetrics(win32con.SM_CYVIRTUALSCREEN)\r\n left = win32api.GetSystemMetrics(win32con.SM_XVIRTUALSCREEN)\r\n top = win32api.GetSystemMetrics(win32con.SM_YVIRTUALSCREEN)\r\n hwindc = win32gui.GetWindowDC(hwin)\r\n srcdc = win32ui.CreateDCFromHandle(hwindc)\r\n memdc = srcdc.CreateCompatibleDC()\r\n bmp.CreateCompatibleBitmap(srcdc, width, height)\r\n memdc.SelectObject(bmp)\r\n memdc.BitBlt((0, 0), (width, height), srcdc, (left, top), win32con.SRCCOPY | CAPTUREBLT)\r\n\r\n bmpinfo = bmp.GetInfo()\r\n bmpstr = bmp.GetBitmapBits(True)\r\n return Image.frombuffer(\r\n 'RGB', (bmpinfo['bmWidth'], bmpinfo['bmHeight']),\r\n bmpstr,\r\n 'raw', 'BGRX', 0, 1)\r\n finally:\r\n memdc.DeleteDC()\r\n srcdc.DeleteDC()\r\n win32gui.ReleaseDC(hwin, hwindc)\r\n win32gui.DeleteObject(bmp.GetHandle())\r\n log.debug(fmt(\"time of execution = {t}\", t=time.time() - bt))\r\n\r\n def _check_jres(self, jres):\r\n if self.jreq['id'] != jres['id']:\r\n raise ValueError('Invalid ID')\r\n if 'error' in jres:\r\n raise requests.exceptions.HTTPError(jres['error']['message'])\r\n return jres\r\n\r\n def run(self):\r\n log.info(fmt('Start daemon: {0}', self.name))\r\n\r\n self.active = True\r\n prev_timeout, timeout = 1, 2\r\n while self.active:\r\n try:\r\n utils.makedirs(self.datadir)\r\n log.debug('Try to grab image')\r\n\r\n# img = self.grabImage_win32()\r\n# img = self.grabImage_PIL()\r\n img = self.grabImage()\r\n self.img1_histogram = img.histogram()\r\n rms = self.compare_images() if self.img1_histogram and self.img2_histogram else self.maxRMS + 1\r\n\r\n log.debug(fmt(\"Root Mean Square={rms}\", rms=rms))\r\n if rms > self.maxRMS:\r\n self.img2_histogram = self.img1_histogram\r\n with closing(cStringIO.StringIO()) as data:\r\n img.save(data, \"JPEG\", quality=self.quality)\r\n self.params['data'] = base64.b64encode(data.getvalue())\r\n\r\n self.jreq['params'] = self.params\r\n self.jreq['id'] = time.time()\r\n try:\r\n log.debug('Try to upload image data')\r\n bt = time.time()\r\n r = requests.post(self.url, json=self.jreq, headers=self.headers, auth=self.auth,\r\n timeout=(3.05, 27), verify=config['CERT'])\r\n jres = self._check_jres(r.json())\r\n log.debug(fmt(\"time of request = {t}\", t=time.time() - bt))\r\n if jres['result'] != 1:\r\n raise requests.exceptions.HTTPError\r\n\r\n except Exception as e:\r\n log.debug(e)\r\n utils.makedirs(self.imagesdir)\r\n fn = os.path.join(self.imagesdir, fmt(\"{0}.jpg\", self.jreq['id']))\r\n log.debug(fmt('Try to save: {fn}', fn=fn))\r\n with open(fn, 'wb') as imfp:\r\n imfp.write(data.getvalue())\r\n for i in os.listdir(self.imagesdir)[-config['SAVED_IMAGES']::-1]:\r\n log.debug(fmt('Try to delete: {fn}', fn=os.path.join(self.imagesdir, i)))\r\n os.unlink(os.path.join(self.imagesdir, i))\r\n raise\r\n finally:\r\n prev_timeout, timeout = 1, 2\r\n\r\n except Exception as e:\r\n if timeout < 60:\r\n prev_timeout, timeout = timeout, prev_timeout + timeout\r\n\r\n if e.__class__ in requests.exceptions.__dict__.itervalues():\r\n try:\r\n ind = config['URL'].index(self.url)\r\n self.url = config['URL'][ind + 1]\r\n except Exception:\r\n self.url = config['URL'][0]\r\n log.error(e)\r\n\r\n time.sleep(timeout)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n t = Screenshoter()\r\n t.start()\r\n t.join()\r\n", "sub_path": "services/screenshoter.py", "file_name": "screenshoter.py", "file_ext": "py", "file_size_in_byte": 8938, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logger.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "config.config", "line_number": 17, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 20, "usage_type": "attribute"}, {"api_name": "utils.fmt", "line_number": 33, "usage_type": "call"}, {"api_name": "threading.Thread.__init__", "line_number": 34, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 39, "usage_type": "name"}, {"api_name": "utils.get_user_name", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path", "line_number": 40, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 41, "usage_type": "name"}, {"api_name": "config.config", "line_number": 42, "usage_type": "name"}, {"api_name": "config.config", "line_number": 43, "usage_type": "name"}, {"api_name": "requests.auth.HTTPDigestAuth", "line_number": 44, "usage_type": "call"}, {"api_name": "requests.auth", "line_number": 44, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 44, "usage_type": "name"}, {"api_name": "utils.utf", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.get_user_name", "line_number": 46, "usage_type": "call"}, {"api_name": "utils.utf", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.get_comp_name", "line_number": 47, "usage_type": "call"}, {"api_name": "utils.fmt", "line_number": 50, "usage_type": "call"}, {"api_name": "config.config", "line_number": 50, "usage_type": "name"}, {"api_name": "utils.makedirs", "line_number": 52, "usage_type": "call"}, {"api_name": "utils.fmt", "line_number": 55, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 65, "usage_type": "call"}, {"api_name": "time.time", "line_number": 78, "usage_type": "call"}, {"api_name": "wx.App", "line_number": 80, "usage_type": "call"}, {"api_name": "wx.ScreenDC", "line_number": 81, "usage_type": "call"}, {"api_name": "wx.Bitmap", "line_number": 83, "usage_type": "call"}, {"api_name": "wx.MemoryDC", "line_number": 84, "usage_type": "call"}, {"api_name": "PIL.Image.frombytes", "line_number": 87, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 87, "usage_type": "name"}, {"api_name": "utils.fmt", "line_number": 89, "usage_type": "call"}, {"api_name": "time.time", "line_number": 89, "usage_type": "call"}, {"api_name": "time.time", "line_number": 97, "usage_type": "call"}, {"api_name": "PIL.ImageGrab.grab", "line_number": 99, "usage_type": "call"}, {"api_name": "PIL.ImageGrab", "line_number": 99, "usage_type": "name"}, {"api_name": "utils.fmt", "line_number": 101, "usage_type": "call"}, {"api_name": "time.time", "line_number": 101, "usage_type": "call"}, {"api_name": "time.time", "line_number": 112, "usage_type": "call"}, {"api_name": "win32ui.CreateBitmap", "line_number": 113, "usage_type": "call"}, {"api_name": "win32gui.GetDesktopWindow", "line_number": 116, "usage_type": "call"}, {"api_name": "win32api.GetSystemMetrics", "line_number": 117, "usage_type": "call"}, {"api_name": "win32con.SM_CXVIRTUALSCREEN", "line_number": 117, "usage_type": "attribute"}, {"api_name": "win32api.GetSystemMetrics", "line_number": 118, "usage_type": "call"}, {"api_name": "win32con.SM_CYVIRTUALSCREEN", "line_number": 118, "usage_type": "attribute"}, {"api_name": "win32api.GetSystemMetrics", "line_number": 119, "usage_type": "call"}, {"api_name": "win32con.SM_XVIRTUALSCREEN", "line_number": 119, "usage_type": "attribute"}, {"api_name": "win32api.GetSystemMetrics", "line_number": 120, "usage_type": "call"}, {"api_name": "win32con.SM_YVIRTUALSCREEN", "line_number": 120, "usage_type": "attribute"}, {"api_name": "win32gui.GetWindowDC", "line_number": 121, "usage_type": "call"}, {"api_name": "win32ui.CreateDCFromHandle", "line_number": 122, "usage_type": "call"}, {"api_name": "win32con.SRCCOPY", "line_number": 126, "usage_type": "attribute"}, {"api_name": "PIL.Image.frombuffer", "line_number": 130, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 130, "usage_type": "name"}, {"api_name": "win32gui.ReleaseDC", "line_number": 137, "usage_type": "call"}, {"api_name": "win32gui.DeleteObject", "line_number": 138, "usage_type": "call"}, {"api_name": "utils.fmt", "line_number": 139, "usage_type": "call"}, {"api_name": "time.time", "line_number": 139, "usage_type": "call"}, {"api_name": "requests.exceptions.HTTPError", "line_number": 145, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 145, "usage_type": "attribute"}, {"api_name": "utils.fmt", "line_number": 149, "usage_type": "call"}, {"api_name": "utils.makedirs", "line_number": 155, "usage_type": "call"}, {"api_name": "utils.fmt", "line_number": 164, "usage_type": "call"}, {"api_name": "contextlib.closing", "line_number": 167, "usage_type": "call"}, {"api_name": "cStringIO.StringIO", "line_number": 167, "usage_type": "call"}, {"api_name": "base64.b64encode", "line_number": 169, "usage_type": "call"}, {"api_name": "time.time", "line_number": 172, "usage_type": "call"}, {"api_name": "time.time", "line_number": 175, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 176, "usage_type": "call"}, {"api_name": "config.config", "line_number": 177, "usage_type": "name"}, {"api_name": "utils.fmt", "line_number": 179, "usage_type": "call"}, {"api_name": "time.time", "line_number": 179, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 181, "usage_type": "attribute"}, {"api_name": "utils.makedirs", "line_number": 185, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 186, "usage_type": "call"}, {"api_name": "os.path", "line_number": 186, "usage_type": "attribute"}, {"api_name": "utils.fmt", "line_number": 186, "usage_type": "call"}, {"api_name": "utils.fmt", "line_number": 187, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 190, "usage_type": "call"}, {"api_name": "config.config", "line_number": 190, "usage_type": "name"}, {"api_name": "utils.fmt", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "os.unlink", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 192, "usage_type": "call"}, {"api_name": "os.path", "line_number": 192, "usage_type": "attribute"}, {"api_name": "requests.exceptions.__dict__.itervalues", "line_number": 201, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 201, "usage_type": "attribute"}, {"api_name": "config.config", "line_number": 203, "usage_type": "name"}, {"api_name": "config.config", "line_number": 204, "usage_type": "name"}, {"api_name": "config.config", "line_number": 206, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 209, "usage_type": "call"}, {"api_name": "{'Image': 'PIL.Image', 'wx': 'wx', 'ImageGrab': 'PIL.ImageGrab', 'win32gui': 'win32gui', 'win32ui': 'win32ui', 'win32con': 'win32con', 'win32api': 'win32api'}", "line_number": 213, "usage_type": "call"}]} +{"seq_id": "93172309", "text": "import xml.etree.cElementTree as ET\nfrom typing import Dict, List\nfrom tasks import Task\nfrom filegens import FileGenTask\nfrom misc import log, parse_preset, update_preset, untempl, flatten\nfrom errors import ParseError\nimport itertools\n\nimport compilers\n\nclass ObjectTask(FileGenTask): # Inherits Task\n sources: List[str]\n file: str\n\n output: List[str]\n\n compiler: compilers.Compiler\n params: Dict\n\n _preset_name: str # Due to implementation this has to be stored until self.run() is called. (only then is `project` passed to us)\n\n def __init__(self, node: ET.Element, props: Dict, **kwargs):\n\n filesets = kwargs['filesets']\n\n if 'source' in node.attrib: # some old files may have these\n raise ParseError('`source` attribute is deprecated. Switch to `src`.')\n if 'source-set' in node.attrib:\n raise ParseError('`source-set` attribute is deprecated. Switch to `src-set`.')\n\n if 'src' in node.attrib:\n self.source = [untempl(node.attrib['src'], props)]\n elif 'src-set' in node.attrib:\n self.source = filesets[node.attrib['src-set']].get_files()\n else:\n self.source = None\n\n self.file = untempl(node.attrib['file'], props) if 'file' in node.attrib else None # file overrides source if specified\n\n if self.file:\n self.output = [self.file]\n elif 'output' in node.attrib:\n self.output = [untempl(node.attrib['output'], props)]\n else:\n self.output = [path + '.o' for path in self.source]\n\n self.compiler = compilers.find(\n name=node.attrib['compiler'] if 'compiler' in node.attrib else None, # name overrides lang if specified\n lang=node.attrib['lang'] if 'lang' in node.attrib else None\n )\n\n self._preset_name = node.attrib['preset'] if 'preset' in node.attrib else None\n self.params = parse_preset(node, props)\n\n self.for_shlib = False\n\n def run(self, project):\n # Apply preset\n if self._preset_name:\n preset = project.presets[self._preset_name]\n preset = update_preset(preset, self.params)\n self.params = preset\n\n if not self.file:\n for i in range(len(self.source)):\n log(1, 'compile {} -> {}'.format(self.source[i], self.output[i]))\n self.compiler.create_object(self.output[i], self.source[i], self.params)\n\n def get_files(self):\n return self.output\n\n @property\n def for_shlib(self) -> bool:\n return self.params.get('for-shlib', False)\n @for_shlib.setter\n def for_shlib(self, val: bool):\n self.params['for-shlib'] = val\n\n def __repr__(self):\n if self.file:\n return self.file\n else:\n return 'compile(' + ', '.join(['{} -> {}'.format(self.source[i], self.output[i]) for i in range(len(self.source))]) + ')'\n\nclass BinaryTask(Task):\n output: str\n\n objects: List[ObjectTask]\n linked_libs: List[str]\n\n compiler: compilers.Compiler\n params: Dict\n\n _preset_name: str # Due to implementation this has to be stored until self.run() is called. (only then is `project` passed to us)\n\n def __init__(self, node, props: Dict, **kwargs):\n # Parse attributes\n self.output = untempl(node.attrib['output'], props) if 'output' in node.attrib else None\n\n self.compiler = compilers.find(\n name=node.attrib['compiler'] if 'compiler' in node.attrib else None, # name overrides lang if specified\n lang=node.attrib['lang'] if 'lang' in node.attrib else None\n )\n\n self._preset_name = node.attrib['preset'] if 'preset' in node.attrib else None\n self.params = parse_preset(node, props)\n\n # Parse ingredients (subtags)\n self.objects = []\n self.linked_libs = []\n\n def apply_preset(self, presets: Dict):\n if self._preset_name:\n preset = project.presets[self._preset_name]\n preset = update_preset(preset, self.params)\n self.params = preset\n\ndef get_libpath(node: ET.Element, props: Dict) -> str:\n if 'libpath' in node.attrib:\n return untempl(node.attrib['libpath'], props)\n\nclass ExecutableTask(BinaryTask):\n def __init__(self, node, props: Dict, **kwargs):\n super().__init__(node, props)\n\n for objnode in node.iterfind('object'):\n self.objects.append(ObjectTask(objnode, props, **kwargs))\n\n for lnknode in node.iterfind('link'):\n self.linked_libs.append(get_libpath(lnknode, props=props))\n\n def run(self, project):\n self.apply_preset(project.presets)\n\n # Compile objects first\n for obj in self.objects:\n obj.run(project)\n\n log(1, 'compile {} -> {}'.format(str(flatten([o.get_files() for o in self.objects])), self.output))\n self.compiler.create_executable(untempl(self.output, project.props), flatten([obj.get_files() for obj in self.objects]), self.linked_libs, self.params)\n\n def __repr__(self):\n return 'compile({} -> {})'.format(str([o.output for o in self.objects]), self.output)\n\nclass SharedLibTask(BinaryTask):\n def __init__(self, node, props: Dict, **kwargs):\n super().__init__(node, props)\n\n for objnode in node.iterfind('object'):\n obj_task = ObjectTask(objnode, props, **kwargs)\n obj_task.for_shlib = True\n self.objects.append(obj_task)\n\n for lnknode in node.iterfind('link'):\n self.linked_libs.append(get_libpath(lnknode, props=props))\n\n def run(self, project):\n self.apply_preset(project.presets)\n\n # Compile objects first\n for obj in self.objects:\n obj.run(project)\n\n obj_files = flatten([o.get_files() for o in self.objects])\n log(1, 'compile {} -> {}'.format(str(obj_files), self.output))\n self.compiler.create_shlib(untempl(self.output, project.props), obj_files, self.linked_libs, self.params)\n\n def __repr__(self):\n return 'compile({} -> {})'.format(str(flatten([o.get_files() for o in self.objects])), self.output)\n", "sub_path": "binary_tasks.py", "file_name": "binary_tasks.py", "file_ext": "py", "file_size_in_byte": 6142, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "filegens.FileGenTask", "line_number": 11, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 15, "usage_type": "name"}, {"api_name": "compilers.Compiler", "line_number": 17, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 18, "usage_type": "name"}, {"api_name": "xml.etree.cElementTree.Element", "line_number": 22, "usage_type": "attribute"}, {"api_name": "xml.etree.cElementTree", "line_number": 22, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}, {"api_name": "errors.ParseError", "line_number": 27, "usage_type": "call"}, {"api_name": "errors.ParseError", "line_number": 29, "usage_type": "call"}, {"api_name": "misc.untempl", "line_number": 32, "usage_type": "call"}, {"api_name": "misc.untempl", "line_number": 38, "usage_type": "call"}, {"api_name": "misc.untempl", "line_number": 43, "usage_type": "call"}, {"api_name": "compilers.find", "line_number": 47, "usage_type": "call"}, {"api_name": "misc.parse_preset", "line_number": 53, "usage_type": "call"}, {"api_name": "misc.update_preset", "line_number": 61, "usage_type": "call"}, {"api_name": "misc.log", "line_number": 66, "usage_type": "call"}, {"api_name": "tasks.Task", "line_number": 85, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 88, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 89, "usage_type": "name"}, {"api_name": "compilers.Compiler", "line_number": 91, "usage_type": "attribute"}, {"api_name": "typing.Dict", "line_number": 92, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 96, "usage_type": "name"}, {"api_name": "misc.untempl", "line_number": 98, "usage_type": "call"}, {"api_name": "compilers.find", "line_number": 100, "usage_type": "call"}, {"api_name": "misc.parse_preset", "line_number": 106, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 112, "usage_type": "name"}, {"api_name": "misc.update_preset", "line_number": 115, "usage_type": "call"}, {"api_name": "xml.etree.cElementTree.Element", "line_number": 118, "usage_type": "attribute"}, {"api_name": "xml.etree.cElementTree", "line_number": 118, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 118, "usage_type": "name"}, {"api_name": "misc.untempl", "line_number": 120, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 123, "usage_type": "name"}, {"api_name": "misc.log", "line_number": 139, "usage_type": "call"}, {"api_name": "misc.flatten", "line_number": 139, "usage_type": "call"}, {"api_name": "misc.untempl", "line_number": 140, "usage_type": "call"}, {"api_name": "misc.flatten", "line_number": 140, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 146, "usage_type": "name"}, {"api_name": "misc.flatten", "line_number": 164, "usage_type": "call"}, {"api_name": "misc.log", "line_number": 165, "usage_type": "call"}, {"api_name": "misc.untempl", "line_number": 166, "usage_type": "call"}, {"api_name": "misc.flatten", "line_number": 169, "usage_type": "call"}]} +{"seq_id": "501925181", "text": "'''\nstun module\n'''\n\nimport stun\nimport network_interface as net_if\n\n\nclass Stun():\n '''\n stun\n '''\n\n def __init__(self):\n self.nat_type = (\"Full Cone\", # 0\n \"Restrict NAT\", # 1\n \"Restrict Port NAT\", # 2\n \"Symmetric NAT\", # 3\n \"Unknown NAT\") # 4\n\n def get_nat_type(self):\n '''\n get External NAT Type, IP and Port\n '''\n\n nat_type, external_ip, external_port = stun.get_ip_info(stun_host = net_if.get_stun_server())\n\n print(\"\\n\\n=======================\")\n print(\"NAT Type:\", nat_type)\n print(\"External IP:\", external_ip)\n print(\"External Port:\", external_port)\n print(\"=======================\\n\\n\")\n\n return nat_type, external_ip, external_port\n\n def check_nat_type(self):\n '''\n make a decision for the corresponding NAT type\n '''\n\n nat_type, external_ip, external_port = self.get_nat_type()\n\n if nat_type == self.nat_type[0]:\n print(\"LOG: P2P mode\")\n return external_ip + \":\" + str(external_port), \"fullcone\"\n elif nat_type == self.nat_type[1]:\n print(\"LOG: P2P mode\")\n return external_ip + \":\" + str(external_port), \"restrict\"\n else:\n print(\"LOG: Relay mode\") #controller\n return external_ip + \":\" + str(external_port), \"symmetric\"\n", "sub_path": "client_dresolution/stun_procedure.py", "file_name": "stun_procedure.py", "file_ext": "py", "file_size_in_byte": 1374, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "stun.get_ip_info", "line_number": 26, "usage_type": "call"}, {"api_name": "network_interface.get_stun_server", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "632066026", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nlast layer without sigmoid\r\n@author: cdq\r\n\"\"\"\r\n\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\ndef sigmoid(x):\r\n return 1.0 / (1.0 + np.exp(-x))\r\ndef dsigmoid(x):\r\n return sigmoid(x) * (1 - sigmoid(x))\r\n\r\nclass BP(object):\r\n def __init__(self, layers, activation='sigmoid', learning_rate=0.01):\r\n self.layers = layers\r\n self.learning_rate = learning_rate\r\n self.caches = {}\r\n self.grades = {}\r\n if activation == 'sigmoid':\r\n self.activation = sigmoid\r\n self.dactivation = dsigmoid\r\n self.parameters = {}\r\n for i in range(1, len(self.layers)):\r\n self.parameters[\"w\"+str(i)] = np.random.random((self.layers[i], self.layers[i-1]))\r\n self.parameters[\"b\"+str(i)] = np.zeros((layers[i],1))\r\n \r\n def forward(self, X):\r\n a = []\r\n z = []\r\n a.append(X)\r\n z.append(X)\r\n \r\n len_layers = len(self.parameters) // 2\r\n for i in range(1, len_layers):\r\n z.append(self.parameters[\"w\"+str(i)] @ a[i-1] + self.parameters[\"b\"+str(i)])\r\n a.append(sigmoid(z[-1]))\r\n #last layer without sigmoid\r\n z.append(self.parameters[\"w\"+str(len_layers)] @ a[-1] + self.parameters[\"b\"+str(len_layers)])\r\n a.append(z[-1])\r\n \r\n self.caches['z'] = z\r\n self.caches['a'] = a\r\n \r\n return self.caches, a[-1]\r\n \r\n def backward(self, y):\r\n a = self.caches['a']\r\n m = y.shape[1]\r\n # last layer grade\r\n len_layers = len(self.parameters) // 2\r\n self.grades[\"dz\"+str(len_layers)] = a[-1]-y\r\n self.grades[\"dw\"+str(len_layers)] = self.grades[\"dz\"+str(len_layers)].dot(a[-2].T) / m\r\n self.grades[\"db\"+str(len_layers)] = np.sum(self.grades[\"dz\"+str(len_layers)], axis=1, keepdims=True) / m\r\n # compute grades\r\n for i in reversed(range(1, len_layers)):\r\n self.grades[\"dz\"+str(i)] = self.parameters[\"w\"+str(i+1)].T.dot(self.grades[\"dz\"+str(i+1)]) * dsigmoid(self.caches[\"z\"][i])\r\n self.grades[\"dw\"+str(i)] = self.grades[\"dz\"+str(i)].dot(self.caches[\"a\"][i-1].T)/m\r\n self.grades[\"db\"+str(i)] = np.sum(self.grades[\"dz\"+str(i)],axis = 1,keepdims = True) /m\r\n #update weights and bias\r\n for i in range(1, len(self.layers)):\r\n self.parameters[\"w\"+str(i)] -= self.learning_rate * self.grades[\"dw\"+str(i)]\r\n self.parameters[\"b\"+str(i)] -= self.learning_rate * self.grades[\"db\"+str(i)]\r\n \r\n def compute_loss(self, y):\r\n return np.mean(np.square(self.caches['a'][-1]-y))\r\n#%%\r\ndef test():\r\n x = np.arange(0.0,1.0,0.01)\r\n y =20* np.sin(2*np.pi*x)\r\n plt.scatter(x,y)\r\n \r\n x = x.reshape(1, 100)\r\n y = y.reshape(1, 100)\r\n \r\n bp = BP([1, 6, 1], learning_rate = 0.01)\r\n \r\n for i in range(1, 50000):\r\n caches, al = bp.forward(x)\r\n bp.backward(y)\r\n \r\n if(i%50 == 0):\r\n print(bp.compute_loss(y))\r\n plt.scatter(x, al)\r\n plt.show()\r\n \r\ntest()\r\n \r\n \r\n \r\n \r\n \r\n \r\n \r\n \r\n ", "sub_path": "BP.py", "file_name": "BP.py", "file_ext": "py", "file_size_in_byte": 3189, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.exp", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random.random", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 25, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.square", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 70, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 84, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 84, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 85, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 85, "usage_type": "name"}]} +{"seq_id": "525270398", "text": "#! /usr/bin/python\n# encoding=utf-8\n\ndef config(errorFile, params):\n \"\"\" Helper to parse logging configuration \"\"\"\n \"\"\" errorFile str - Path to log file \"\"\"\n \"\"\" params dict - configuration parameters \"\"\"\n \"\"\" stations dict - keys are stationIDs, values are instances of as_nma.config.app.AS_NMA_STATION \"\"\"\n\n \"\"\" A seperate log logger and timestamp logger will be created for each station \"\"\"\n \"\"\" Get the loggers with the log.weather, log.timestamp, and log.error modules' getLogger() functions \"\"\"\n\n if not 'wsApp' in params:\n params['wsApp'] = None\n \n import os.path\n\n # We are going to store the message file in the same folder as errorFile\n # This way, we don't have to go change all our callers.\n messageFile = os.path.join(os.path.dirname(errorFile), 'messages.txt')\n\n errorFile\n\n import logging\n import logging.config\n config = {\n 'version': 1,\n 'disable_existing_loggers': False,\n 'formatters': {\n 'message': {\n 'format': '%(message)s'\n },\n 'simple': {\n 'format': '%(asctime)s - %(name)s - %(filename)s:%(lineno)d - %(levelname)s - %(message)s',\n 'datefmt': '%Y-%m-%d %H:%M:%S'\n },\n 'brief': {\n 'format': '%(name)s - %(filename)s:%(lineno)d - %(levelname)s - %(message)s'\n },\n 'brieftime': {\n 'format': '%(asctime)s - %(filename)s:%(lineno)d - %(levelname)s - %(message)s',\n 'datefmt': '%Y-%m-%d %H:%M:%S'\n },\n 'briefnoname': {\n 'format': '%(filename)s:%(lineno)d - %(levelname)s - %(message)s'\n }\n },\n 'handlers': {\n 'console': {\n 'class': 'logging.StreamHandler',\n 'formatter': 'briefnoname',\n 'stream': 'ext://sys.stdout',\n 'level': 'INFO'\n },\n 'error_file': {\n 'class': 'logging.handlers.RotatingFileHandler',\n 'formatter': 'simple',\n 'filename': errorFile,\n 'maxBytes': 1024*1024,\n 'backupCount': 10,\n 'delay': True\n },\n 'message_file': {\n 'class': 'logging.handlers.RotatingFileHandler',\n 'formatter': 'brieftime',\n 'filename': messageFile,\n 'maxBytes': 1024*1024,\n 'backupCount': 4,\n 'delay': True\n },\n 'dbhandler': {\n 'class': 'as_weatherstation.log.dbhandler.DBHandler',\n 'formatter': 'message',\n 'level': 'INFO',\n 'delay': True,\n 'wsApp': params['wsApp']\n },\n },\n 'loggers': {\n 'as_weatherstation.log.error': {\n 'level': 'WARN',\n 'handlers': ['error_file'],\n 'propagate': False\n },\n 'as_weatherstation.log.message': {\n 'level': 'INFO',\n 'handlers': ['message_file'],\n 'propagate': False\n },\n 'as_weatherstation.write.db': {\n 'level': 'INFO',\n 'handlers': ['dbhandler'],\n 'propagate': False\n }\n },\n 'root': {\n 'handlers': ['console']\n }\n }\n\n\n if 'stations' in params:\n for sLabel in params['stations']:\n\n sID = params['stations'][sLabel].id\n logHandlerID = ('weather_log_file_%d' % sID)\n timestampHandlerID = ('weather_timestamp_file_%d' % sID)\n logLoggerID = getStationLogLoggerID(sID)\n timestampLoggerID = getStationTimestampLoggerID(sID)\n\n # Add some handlers\n config['handlers'][logHandlerID] = {\n 'class': 'logging.handlers.TimedRotatingFileHandler',\n 'level': 'INFO',\n 'formatter': 'message',\n 'filename': params['stations'][sLabel].logFile,\n 'when': 'S', # use seconds so we can have nice timestamp on backup files\n 'interval': 24*60*60, # every day\n 'backupCount': 1000000000000, # One-trillion. Essentially unlimited. You will run out of diskspace if you don't archive backups.\n 'encoding': None,\n 'delay': True,\n 'utc': False\n }\n config['handlers'][timestampHandlerID] = {\n 'class': 'logging.FileHandler',\n 'level': 'INFO',\n 'formatter': 'message',\n 'filename': params['stations'][sLabel].timestampFile,\n 'mode': 'w', # every time we write we create a new file\n 'encoding': None,\n 'delay': True\n }\n\n # Add some loggers\n config['loggers'][logLoggerID] = {\n 'level': 'DEBUG',\n 'handlers': [logHandlerID],\n 'propagate': False\n }\n config['loggers'][timestampLoggerID] = {\n 'level': 'DEBUG',\n 'handlers': [timestampHandlerID],\n 'propagate': False\n }\n\n # save names of station loggers\n params['stations'][sLabel].logLoggerID = logLoggerID\n params['stations'][sLabel].timestampLoggerID = timestampLoggerID\n\n\n logging.config.dictConfig(config)\n\n\n\ndef getStationLogLoggerID(sID):\n \n return ('as_ws.log.weather_%d' % sID)\n\n\n\ndef getStationTimestampLoggerID(sID):\n return ('as_ws.log.timestamp_%d' % sID)\n\n", "sub_path": "as_weatherstation/config/log.py", "file_name": "log.py", "file_ext": "py", "file_size_in_byte": 5733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 20, "usage_type": "name"}, {"api_name": "os.path.path.dirname", "line_number": 20, "usage_type": "call"}, {"api_name": "logging.config.dictConfig", "line_number": 151, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 151, "usage_type": "attribute"}]} +{"seq_id": "455211427", "text": "from __future__ import unicode_literals, print_function\n\nimport ffmpeg\nimport os\nimport signal\nimport sys\nimport time\n\nfrom azure.core.exceptions import AzureError\nfrom azure.storage.blob import BlobClient\n\nfrom package.utility import *\n\nPATH_TO_FILE = r\"./capture.jpg\"\n\n# Function to perform Azure Blob Storage connection and perform upload\nasync def store_blob(blob_info, file_name):\n try:\n sas_url = \"https://{}/{}/{}{}\".format(\n blob_info[\"hostName\"],\n blob_info[\"containerName\"],\n blob_info[\"blobName\"],\n blob_info[\"sasToken\"]\n )\n\n print(\"\\nUploading file: {} to Azure Storage as blob: {} in container {}\\n\".format(file_name, blob_info[\"blobName\"], blob_info[\"containerName\"]))\n print( sas_url )\n # Upload the specified file\n with BlobClient.from_blob_url(sas_url) as blob_client:\n with open(file_name, \"rb\") as f:\n result = blob_client.upload_blob(f, overwrite=True)\n return (True, result)\n\n except FileNotFoundError as ex:\n # catch file not found and add an HTTP status code to return in notification to IoT Hub\n ex.status_code = 404\n return (False, ex)\n\n except AzureError as ex:\n # catch Azure errors that might result from the upload operation\n return (False, ex)\n\n\n# Function to start FFMPEG in a streaming mode and save Webcam image with the\n# provided frame per second\n# Native ffmpeg command: \n# ffmpeg -f v4l2 -input_format mjpeg -i /dev/video0 -update 1 -r 1 -y capture.jpg \ndef start_ffmpeg(fps = 1):\n\n try:\n process = (\n ffmpeg\n .input('/dev/video0', vcodec='mjpeg', format='v4l2', nostdin=None)\n .output(PATH_TO_FILE, r=int(fps), format='image2', update='1')\n .overwrite_output()\n .run_async(quiet=True)\n )\n\n except ffmpeg.Error as e:\n print(e.stderr, file=sys.stderr)\n sys.exit(1)\n\n print(f\"FFMPEG started with {fps} FPS\")\n\n return process\n\n# Function to stop FFMPEG since FFMPEG is started asynchronously with non-blocking\n# STDIN\ndef stop_ffmpeg(process):\n\n # Suspend the FFMPEG\n os.kill(process.pid, signal.SIGKILL)\n\n print(\"FFMPEG stopped\")\n\n return True\n\n# Function to suspend FFMPEG to allow dumped frame to be uploaded\nasync def suspend_ffmpeg_and_upload(device_client, process):\n\n # Suspend the FFMPEG (Ctrl + Z)\n os.kill(process.pid, signal.SIGSTOP)\n\n # Get the storage info for the blob\n blob_name = os.path.basename(PATH_TO_FILE)\n timestr = time.strftime(\"%Y%m%d-%H%M%S\")\n blob_name = \"{0}_{2}.{1}\".format(*blob_name.split('.'), timestr)\n storage_info = await device_client.get_storage_info_for_blob(blob_name)\n\n # Upload to blob\n success, result = await store_blob(storage_info, PATH_TO_FILE)\n\n # Continue the FFMPEG (fg)\n os.kill(process.pid, signal.SIGCONT)\n\n if success == True:\n print(\"Upload succeeded. Result is: \\n\")\n print(result)\n print()\n\n await device_client.notify_blob_upload_status(\n storage_info[\"correlationId\"], True, 200, \"OK: {}\".format(PATH_TO_FILE)\n )\n\n else :\n # If the upload was not successful, the result is the exception object\n print(\"Upload failed. Exception is: \\n\")\n print(result)\n print()\n\n await device_client.notify_blob_upload_status(\n storage_info[\"correlationId\"], False, result.status_code, str(result)\n )\n\n raise FileNotFoundError(\"Missing image file\")\n\n return f'https://{storage_info[\"hostName\"]}/{storage_info[\"containerName\"]}/{storage_info[\"blobName\"]}'\n\n# Function to clean up the staging file\ndef remove_staging_file():\n os.remove(PATH_TO_FILE)\n", "sub_path": "azure-de10nano-document/uvc-telemetry-reference-design-for-azure/sw/software-code/package/library.py", "file_name": "library.py", "file_ext": "py", "file_size_in_byte": 3752, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "azure.storage.blob.BlobClient.from_blob_url", "line_number": 29, "usage_type": "call"}, {"api_name": "azure.storage.blob.BlobClient", "line_number": 29, "usage_type": "name"}, {"api_name": "azure.core.exceptions.AzureError", "line_number": 39, "usage_type": "name"}, {"api_name": "ffmpeg.input", "line_number": 52, "usage_type": "call"}, {"api_name": "ffmpeg.Error", "line_number": 59, "usage_type": "attribute"}, {"api_name": "sys.stderr", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 61, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 72, "usage_type": "call"}, {"api_name": "signal.SIGKILL", "line_number": 72, "usage_type": "attribute"}, {"api_name": "os.kill", "line_number": 82, "usage_type": "call"}, {"api_name": "signal.SIGSTOP", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "time.strftime", "line_number": 86, "usage_type": "call"}, {"api_name": "os.kill", "line_number": 94, "usage_type": "call"}, {"api_name": "signal.SIGCONT", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 121, "usage_type": "call"}]} +{"seq_id": "218076712", "text": "\"\"\"\nPrototype 4\n\nThis prototype is based entirely on technical analysis, and has been configured for multi-threading task management\nVictor Guillet\n12/17/2018\n\"\"\"\n\nfrom PhyTrade.Economic_model.Big_Data import BIGDATA\n\nfrom PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.RSI_gen import RSI\nfrom PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.SMA_gen import SMA\n\nfrom PhyTrade.Economic_model.Technical_Analysis.Amplification_signals.Volume_gen import VOLUME\nfrom PhyTrade.Economic_model.Technical_Analysis.Amplification_signals.Volatility_gen import VOLATILITY\n\nfrom PhyTrade.Economic_model.MAJOR_SPLINE_gen import MAJOR_SPLINE\nfrom PhyTrade.Economic_model.Technical_Analysis.Tools.OC_tools import OC\nfrom PhyTrade.Tools.SPLINE_tools import SPLINE\n\nfrom PhyTrade.Tools.MATH_tools import MATH_tools\n\nimport pandas as pd\nimport threading\n\n\nclass Prototype_4:\n def __init__(self, parameters):\n\n # ========================= DATA COLLECTION INITIALISATION =======================\n ticker = 'AAPL' # Ticker selected for Yahoo data collection\n # data = pull_yahoo_data(ticker) # Pull data from Yahoo\n\n path = r\"C:\\Users\\Victor Guillet\\Google Drive\\2-Programing\\Repos\\Steffegium\\Data\\NVDA_Yahoo_data.csv\".replace(\n '\\\\', '/')\n\n data = pd.read_csv(path)\n\n # ========================= ANALYSIS INITIALISATION ==============================\n data_slice_start_ind = -400\n data_slice_stop_ind = len(data)-200\n\n self.big_data = BIGDATA(data, ticker, data_slice_start_ind, data_slice_stop_ind)\n\n # ------------------ Tools initialisation\n self.oc_tools = OC()\n self.spline_tools = SPLINE(self.big_data)\n self.math_tools = MATH_tools()\n\n # ------------------ Technical_Indicators initialisation\n # -- RSI initialisation\n self.big_data.rsi_1 = RSI(self.big_data,\n timeframe=parameters[\"timeframe\"][\"rsi_1_timeframe\"],\n standard_upper_threshold=parameters[\"rsi_standard_upper_thresholds\"][\"rsi_1_standard_upper_threshold\"],\n standard_lower_threshold=parameters[\"rsi_standard_lower_thresholds\"][\"rsi_1_standard_lower_threshold\"])\n\n self.big_data.rsi_2 = RSI(self.big_data,\n timeframe=parameters[\"timeframe\"][\"rsi_2_timeframe\"],\n standard_upper_threshold=parameters[\"rsi_standard_upper_thresholds\"][\"rsi_2_standard_upper_threshold\"],\n standard_lower_threshold=parameters[\"rsi_standard_lower_thresholds\"][\"rsi_2_standard_lower_threshold\"])\n\n self.big_data.rsi_3 = RSI(self.big_data,\n timeframe=parameters[\"timeframe\"][\"rsi_3_timeframe\"],\n standard_upper_threshold=parameters[\"rsi_standard_upper_thresholds\"][\"rsi_3_standard_upper_threshold\"],\n standard_lower_threshold=parameters[\"rsi_standard_lower_thresholds\"][\"rsi_3_standard_lower_threshold\"])\n\n # -- SMA initialisation\n self.big_data.sma_1 = SMA(self.big_data,\n timeperiod_1=parameters[\"timeframe\"][\"sma_1_timeperiod_1\"],\n timeperiod_2=parameters[\"timeframe\"][\"sma_1_timeperiod_2\"])\n\n self.big_data.sma_2 = SMA(self.big_data,\n timeperiod_1=parameters[\"timeframe\"][\"sma_2_timeperiod_1\"],\n timeperiod_2=parameters[\"timeframe\"][\"sma_2_timeperiod_2\"])\n\n self.big_data.sma_3 = SMA(self.big_data,\n timeperiod_1=parameters[\"timeframe\"][\"sma_3_timeperiod_1\"],\n timeperiod_2=parameters[\"timeframe\"][\"sma_3_timeperiod_2\"])\n\n # -- Volume initialisation\n self.big_data.volume = VOLUME(self.big_data,\n amplification_factor=parameters[\"amplification_factor\"][\"volume_amplification_factor\"])\n\n # -- Volatility initialisation\n self.big_data.volatility = VOLATILITY(self.big_data,\n timeframe=parameters[\"timeframe\"][\"volatility_timeframe\"],\n amplification_factor=parameters[\"amplification_factor\"][\"volatility_amplification_factor\"])\n\n # ================================================================================\n \"\"\"\n\n\n\n\n \"\"\"\n # ========================= DATA GENERATION AND PROCESSING =======================\n # ~~~~~~~~~~~~~~~~~~ Technical_Indicators output generation\n # - RSI\n t1 = threading.Thread(target=self.big_data.rsi_1.get_output, args=(self.big_data,))\n t2 = threading.Thread(target=self.big_data.rsi_2.get_output, args=(self.big_data,))\n t3 = threading.Thread(target=self.big_data.rsi_3.get_output, args=(self.big_data,))\n\n t1.start()\n t2.start()\n t3.start()\n\n t1.join()\n t2.join()\n t3.join()\n\n # self.big_data.rsi_1.get_output(self.big_data, include_triggers_in_bb_signal=True)\n # self.big_data.rsi_2.get_output(self.big_data, include_triggers_in_bb_signal=True)\n # self.big_data.rsi_3.get_output(self.big_data, include_triggers_in_bb_signal=True)\n\n # - SMA\n t1 = threading.Thread(target=self.big_data.sma_1.get_output, args=(self.big_data,))\n t2 = threading.Thread(target=self.big_data.sma_2.get_output, args=(self.big_data,))\n t3 = threading.Thread(target=self.big_data.sma_3.get_output, args=(self.big_data,))\n\n t1.start()\n t2.start()\n t3.start()\n\n t1.join()\n t2.join()\n t3.join()\n\n # self.big_data.sma_1.get_output(self.big_data, include_triggers_in_bb_signal=False)\n # self.big_data.sma_2.get_output(self.big_data, include_triggers_in_bb_signal=False)\n # self.big_data.sma_3.get_output(self.big_data, include_triggers_in_bb_signal=False)\n\n # ~~~~~~~~~~~~~~~~~~ BB signals processing\n # -- Creating splines from signals\n # - RSI\n self.big_data.spline_rsi_1 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.rsi_1.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"rsi_1_spline_smoothing_factor\"])\n\n self.big_data.spline_rsi_2 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.rsi_2.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"rsi_2_spline_smoothing_factor\"])\n\n self.big_data.spline_rsi_3 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.rsi_3.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"rsi_3_spline_smoothing_factor\"])\n\n # - SMA\n self.big_data.spline_sma_1 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.sma_1.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"sma_1_spline_smoothing_factor\"])\n self.big_data.spline_sma_2 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.sma_2.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"sma_2_spline_smoothing_factor\"])\n self.big_data.spline_sma_3 = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.sma_3.bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"sma_3_spline_smoothing_factor\"])\n\n # - OC avg gradient\n self.big_data.spline_oc_avg_gradient = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.oc_avg_gradient_bb_signal,\n smoothing_factor=parameters[\"smoothing_factors\"][\"oc_avg_gradient_spline_smoothing_factor\"])\n\n # -- Generating amplification signals\n self.big_data.spline_volume = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.volume.amp_coef,\n smoothing_factor=parameters[\"smoothing_factors\"][\"volume_spline_smoothing_factor\"])\n\n self.big_data.spline_volatility = \\\n self.spline_tools.calc_signal_to_spline(self.big_data, self.big_data.volatility.amp_coef,\n smoothing_factor=parameters[\"smoothing_factors\"][\"volatility_spline_smoothing_factor\"])\n\n # -- Tuning separate signals\n self.big_data.spline_sma_3 = self.spline_tools.flip_spline(self.big_data.spline_sma_3)\n\n # -- Adding signals together\n self.big_data.combined_spline = \\\n self.spline_tools.combine_7_splines(self.big_data,\n self.big_data.spline_rsi_1,\n self.big_data.spline_rsi_2,\n self.big_data.spline_rsi_3,\n self.big_data.spline_sma_1,\n self.big_data.spline_sma_2,\n self.big_data.spline_sma_3,\n self.big_data.spline_oc_avg_gradient,\n weight_2=parameters[\"weights\"][\"rsi_1_spline_weight\"],\n weight_3=parameters[\"weights\"][\"rsi_2_spline_weight\"],\n weight_4=parameters[\"weights\"][\"rsi_3_spline_weight\"],\n weight_5=parameters[\"weights\"][\"sma_1_spline_weight\"],\n weight_6=parameters[\"weights\"][\"sma_2_spline_weight\"],\n weight_7=parameters[\"weights\"][\"sma_3_spline_weight\"],\n weight_1=parameters[\"weights\"][\"oc_avg_gradient_spline_weight\"])\n\n # -- Tuning combined signal\n self.big_data.combined_spline = \\\n self.spline_tools.modulate_amplitude_spline(\n self.big_data.combined_spline, self.big_data.spline_volume, std_dev_max=3)\n\n self.big_data.combined_spline = \\\n self.spline_tools.modulate_amplitude_spline(\n self.big_data.combined_spline, self.big_data.spline_volatility, std_dev_max=3)\n\n self.big_data.combined_spline = self.math_tools.normalise_minus_one_one(self.big_data.combined_spline)\n\n # ~~~~~~~~~~~~~~~~~~ Threshold determination\n # -- Creating dynamic thresholds\n upper_threshold, lower_threshold = \\\n self.spline_tools.calc_thresholds(self.big_data, self.big_data.combined_spline,\n buffer=0.05, buffer_setting=1,\n standard_upper_threshold=0.45,\n standard_lower_threshold=-0.5)\n\n # -- Modulating threshold with SMA 3 value\n # upper_threshold = self.spline_tools.modulate_amplitude_spline(\n # upper_threshold, self.math_tools.amplify(\n # self.math_tools.normalise_zero_one(self.big_data.spline_sma_3), 0.3))\n #\n # lower_threshold = self.spline_tools.modulate_amplitude_spline(\n # lower_threshold, self.math_tools.amplify(\n # self.math_tools.normalise_zero_one(self.big_data.spline_sma_3), 0.3))\n\n # ~~~~~~~~~~~~~~~~~~ Creating Major Spline/trigger values\n self.big_data.Major_spline = MAJOR_SPLINE(self.big_data, self.big_data.combined_spline,\n upper_threshold, lower_threshold)\n\n # ================================================================================\n \"\"\"\n\n\n\n\n \"\"\"\n\n # ========================= SIGNAL PLOTS =========================================\n def plot(self, plot_1=True, plot_2=True, plot_3=True):\n import matplotlib.pyplot as plt\n\n if plot_1:\n # ---------------------------------------------- Plot 1\n # ------------------ Plot Open/Close prices\n ax1 = plt.subplot(211)\n self.oc_tools.plot_oc_values(self.big_data)\n # oc.plot_trigger_values(self.big_data)\n\n # ------------------ Plot RSI\n ax2 = plt.subplot(212, sharex=ax1)\n self.big_data.rsi.plot_rsi(self.big_data)\n plt.show()\n\n if plot_2:\n # ---------------------------------------------- Plot 2\n # ------------------ Plot Open/Close prices\n ax3 = plt.subplot(211)\n self.oc_tools.plot_oc_values(self.big_data)\n # oc.plot_trigger_values(self.big_data)\n\n # ------------------ Plot SMA Signal\n ax4 = plt.subplot(212, sharex=ax3)\n self.big_data.sma_1.plot_sma(self.big_data, plot_trigger_signals=False)\n plt.show()\n\n if plot_3:\n # ---------------------------------------------- Plot 3\n # ------------------ Plot Open/Close prices\n ax5 = plt.subplot(211)\n self.oc_tools.plot_oc_values(self.big_data)\n self.oc_tools.plot_trigger_values(\n self.big_data, self.big_data.Major_spline.sell_dates, self.big_data.Major_spline.buy_dates)\n\n # ------------------ Plot bb signal(s)\n ax6 = plt.subplot(212)\n # self.spline_tools.plot_spline(\n # self.big_data, self.big_data.spline_rsi, label=\"RSI bb spline\")\n # self.spline_tools.plot_spline(\n # self.big_data, self.big_data.spline_oc_avg_gradient, label=\"OC gradient bb spline\", color='m')\n # self.spline_tools.plot_spline(\n # self.big_data, self.big_data.spline_sma_1, label=\"SMA_1 bb spline\", color='b')\n # self.spline_tools.plot_spline(\n # self.big_data, self.big_data.spline_sma_2, label=\"SMA_2 bb spline\", color='b')\n # self.spline_tools.plot_spline(\n # self.big_data, self.big_data.spline_sma_3, label=\"SMA_3 bb spline\", color='r')\n\n self.spline_tools.plot_spline(\n self.big_data, self.big_data.Major_spline.spline, label=\"Major spline\", color='y')\n\n self.spline_tools.plot_spline(\n self.big_data, self.big_data.Major_spline.upper_threshold, label=\"Upper threshold\")\n self.spline_tools.plot_spline(\n self.big_data, self.big_data.Major_spline.lower_threshold, label=\"Lower threshold\")\n\n self.spline_tools.plot_spline_trigger(\n self.big_data, self.big_data.Major_spline.spline, self.big_data.Major_spline.sell_dates,\n self.big_data.Major_spline.buy_dates)\n\n # self.spline_tools.plot_spline(self.big_data, self.big_data.spline_volume, label=\"Volume\", color='k')\n # self.spline_tools.plot_spline(self.big_data, self.big_data.spline_volatility, label=\"Volatility\", color='grey')\n plt.show()\n", "sub_path": "PhyTrade/Economic_model/Analysis_protocols/Prototype_4.py", "file_name": "Prototype_4.py", "file_ext": "py", "file_size_in_byte": 15392, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pandas.read_csv", "line_number": 37, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Big_Data.BIGDATA", "line_number": 43, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Tools.OC_tools.OC", "line_number": 46, "usage_type": "call"}, {"api_name": "PhyTrade.Tools.SPLINE_tools.SPLINE", "line_number": 47, "usage_type": "call"}, {"api_name": "PhyTrade.Tools.MATH_tools.MATH_tools", "line_number": 48, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.RSI_gen.RSI", "line_number": 52, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.RSI_gen.RSI", "line_number": 57, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.RSI_gen.RSI", "line_number": 62, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.SMA_gen.SMA", "line_number": 68, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.SMA_gen.SMA", "line_number": 72, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Technical_Indicators.SMA_gen.SMA", "line_number": 76, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Amplification_signals.Volume_gen.VOLUME", "line_number": 81, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.Technical_Analysis.Amplification_signals.Volatility_gen.VOLATILITY", "line_number": 85, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 99, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 100, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 101, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 116, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 117, "usage_type": "call"}, {"api_name": "threading.Thread", "line_number": 118, "usage_type": "call"}, {"api_name": "PhyTrade.Economic_model.MAJOR_SPLINE_gen.MAJOR_SPLINE", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 240, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 240, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 245, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 245, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 247, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 247, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 252, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 257, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 259, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 264, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 264, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 270, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 270, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 296, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 296, "usage_type": "name"}]} +{"seq_id": "89995691", "text": "#!/usr/bin/env python3\n\n'''Note this matrix must be positive semi definite\nclearly symmetric so only checking for non-negative eigvals '''\n\nimport numpy as np\nimport sys\nfrom scipy import linalg\n\ndef is_pos_sem_def(x):\n return np.all(linalg.eigvals(x) >= 0)\n\nk = int(sys.argv[1]) #52\n\noutput_file = 'theta_MVN_Lambda_0.npy'\nto_write = np.zeros((k,k))\nnp.fill_diagonal(to_write, 25)\nnp.save(output_file, to_write)\n\nprint(is_pos_sem_def(to_write), flush=True)\n", "sub_path": "regress_dist_gibbs/make_init_matrices_input3/make_theta_MVN_Lambda_0.py", "file_name": "make_theta_MVN_Lambda_0.py", "file_ext": "py", "file_size_in_byte": 461, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.all", "line_number": 11, "usage_type": "call"}, {"api_name": "scipy.linalg.eigvals", "line_number": 11, "usage_type": "call"}, {"api_name": "scipy.linalg", "line_number": 11, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 13, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 16, "usage_type": "call"}, {"api_name": "numpy.fill_diagonal", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 18, "usage_type": "call"}]} +{"seq_id": "124150847", "text": "\"\"\"\n Copyright (c) 2023, NVIDIA CORPORATION.\n \n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n\"\"\"\n\nimport hugectr\nfrom mpi4py import MPI\nimport threading\nimport sys\n\n\ndef model_test(json_file):\n solver = hugectr.CreateSolver(\n max_eval_batches=100,\n batchsize_eval=16384,\n batchsize=16384,\n vvgpu=[[0, 1], [2, 3], [4, 5], [6, 7]],\n i64_input_key=False,\n use_mixed_precision=False,\n repeat_dataset=True,\n use_cuda_graph=True,\n )\n reader = hugectr.DataReaderParams(\n data_reader_type=hugectr.DataReaderType_t.Norm,\n source=[\"./file_list.txt\"],\n eval_source=\"./file_list_test.txt\",\n check_type=hugectr.Check_t.Sum,\n )\n optimizer = hugectr.CreateOptimizer(optimizer_type=hugectr.Optimizer_t.Adam)\n model = hugectr.Model(solver, reader, optimizer)\n model.construct_from_json(graph_config_file=json_file, include_dense_network=True)\n model.summary()\n model.compile()\n model.fit(\n max_iter=10000, display=200, eval_interval=1000, snapshot=100000, snapshot_prefix=\"dcn\"\n )\n\n\nif __name__ == \"__main__\":\n json_file = sys.argv[1]\n comm = MPI.COMM_WORLD\n rank = comm.Get_rank()\n thread = threading.Thread(target=model_test, args=(json_file,), name=\"[rank-%d train]\" % rank)\n current_thread = threading.currentThread()\n print(\"[HUGECTR][INFO] %s is main thread: %s\" % (current_thread.name, MPI.Is_thread_main()))\n print(\"[HUGECTR][INFO] before: rank %d \" % (rank))\n # start the thread\n thread.start()\n # wait for terminate\n thread.join()\n print(\"[HUGECTR][INFO] after: rank %d \" % (rank))\n", "sub_path": "test/pybind_test/dcn_4node_2gpu.py", "file_name": "dcn_4node_2gpu.py", "file_ext": "py", "file_size_in_byte": 2122, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "hugectr.CreateSolver", "line_number": 24, "usage_type": "call"}, {"api_name": "hugectr.DataReaderParams", "line_number": 34, "usage_type": "call"}, {"api_name": "hugectr.DataReaderType_t", "line_number": 35, "usage_type": "attribute"}, {"api_name": "hugectr.Check_t", "line_number": 38, "usage_type": "attribute"}, {"api_name": "hugectr.CreateOptimizer", "line_number": 40, "usage_type": "call"}, {"api_name": "hugectr.Optimizer_t", "line_number": 40, "usage_type": "attribute"}, {"api_name": "hugectr.Model", "line_number": 41, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 51, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 52, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 52, "usage_type": "name"}, {"api_name": "threading.Thread", "line_number": 54, "usage_type": "call"}, {"api_name": "threading.currentThread", "line_number": 55, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Is_thread_main", "line_number": 56, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 56, "usage_type": "name"}]} +{"seq_id": "147683614", "text": "from logging import getLogger\nfrom typing import Any, Dict, List, Tuple, Union, cast\n\nfrom eth_utils import function_abi_to_4byte_selector\nfrom hexbytes import HexBytes\nfrom web3 import Web3\nfrom web3._utils.abi import (get_abi_input_names, get_abi_input_types,\n map_abi_data)\nfrom web3._utils.normalizers import BASE_RETURN_NORMALIZERS\nfrom web3.contract import Contract, ContractFunction\n\nfrom gnosis.eth.contracts import (get_safe_contract, get_safe_V0_0_1_contract,\n get_safe_V1_0_0_contract)\n\nlogger = getLogger(__name__)\n\n\nAbiType = Dict[str, Any]\n\n\nclass TxDecoderException(Exception):\n pass\n\n\nclass UnexpectedProblemDecoding(TxDecoderException):\n pass\n\n\nclass CannotDecode(TxDecoderException):\n pass\n\n\ndef get_tx_decoder() -> 'TxDecoder':\n if not hasattr(get_tx_decoder, 'instance'):\n get_tx_decoder.instance = TxDecoder()\n return get_tx_decoder.instance\n\n\nclass TxDecoder:\n \"\"\"\n Decode txs for supported contracts\n \"\"\"\n def __init__(self):\n self.dummy_w3 = Web3()\n # Order is important. If signature is the same (e.g. renaming of `baseGas`) last elements in the list\n # will take preference\n self.supported_contracts = [get_safe_V0_0_1_contract(self.dummy_w3),\n get_safe_V1_0_0_contract(self.dummy_w3),\n get_safe_contract(self.dummy_w3)]\n\n # Web3 generates possible selectors every time. We cache that and use a dict to do a fast check\n # Store selectors with abi\n self.supported_fn_selectors: Dict[bytes, ContractFunction] = {}\n for supported_contract in self.supported_contracts:\n self.supported_fn_selectors.update(self._generate_selectors_with_abis_from_contract(supported_contract))\n\n def _generate_selectors_with_abis_from_contract(self, contract: Contract) -> Dict[bytes, ContractFunction]:\n return {function_abi_to_4byte_selector(contract_fn.abi): contract_fn\n for contract_fn in contract.all_functions()}\n\n def _parse_decoded_arguments(self, decoded_value: Any) -> Any:\n \"\"\"\n Parse decoded arguments, like converting `bytes` to hexadecimal `str`\n :param decoded:\n :return: Dict[str, Any]\n \"\"\"\n if isinstance(decoded_value, bytes):\n decoded_value = HexBytes(decoded_value).hex()\n return decoded_value\n\n def decode_transaction_with_types(self, data: Union[bytes, str]) -> Tuple[str, List[Tuple[str, str, Any]]]:\n \"\"\"\n Decode tx data\n :param data: Tx data as `hex string` or `bytes`\n :return: Tuple with the `function name` and a list of dictionaries dictionary {'name', 'type', 'value'}\n :raises: CannotDecode if data cannot be decoded. You should catch this exception when using this function\n :raises: UnexpectedProblemDecoding if there's an unexpected problem decoding (it shouldn't happen)\n \"\"\"\n fn_name, parameters = self._decode_transaction(data)\n return fn_name, [{'name': name, 'type': argument_type, 'value': value}\n for name, argument_type, value in parameters]\n\n def decode_transaction(self, data: Union[bytes, str]) -> Tuple[str, Dict[str, Any]]:\n \"\"\"\n Decode tx data\n :param data: Tx data as `hex string` or `bytes`\n :return: Tuple with the `function name` and a dictionary with the arguments of the function\n :raises: CannotDecode if data cannot be decoded. You should catch this exception when using this function\n :raises: UnexpectedProblemDecoding if there's an unexpected problem decoding (it shouldn't happen)\n \"\"\"\n fn_name, parameters = self._decode_transaction(data)\n return fn_name, {name: value for name, argument_type, value in parameters}\n\n def _decode_transaction(self, data: Union[bytes, str]) -> Tuple[str, List[Tuple[str, str, Any]]]:\n \"\"\"\n Decode tx data\n :param data: Tx data as `hex string` or `bytes`\n :return: Tuple with the `function name` and a List of sorted tuples with\n the `name` of the argument, `type` and `value`\n :raises: CannotDecode if data cannot be decoded. You should catch this exception when using this function\n :raises: UnexpectedProblemDecoding if there's an unexpected problem decoding (it shouldn't happen)\n \"\"\"\n\n if not data:\n raise CannotDecode(data)\n\n data = HexBytes(data)\n selector, params = data[:4], data[4:]\n if selector not in self.supported_fn_selectors:\n raise CannotDecode(data.hex())\n\n try:\n contract_fn = self.supported_fn_selectors[selector]\n names = get_abi_input_names(contract_fn.abi)\n types = get_abi_input_types(contract_fn.abi)\n decoded = self.dummy_w3.codec.decode_abi(types, cast(HexBytes, params))\n normalized = map_abi_data(BASE_RETURN_NORMALIZERS, types, decoded)\n values = map(self._parse_decoded_arguments, normalized)\n except ValueError as exc:\n raise UnexpectedProblemDecoding from exc\n\n return contract_fn.fn_name, list(zip(names, types, values))\n", "sub_path": "safe_transaction_service/history/indexers/tx_decoder.py", "file_name": "tx_decoder.py", "file_ext": "py", "file_size_in_byte": 5233, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 15, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 18, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 18, "usage_type": "name"}, {"api_name": "web3.Web3", "line_number": 44, "usage_type": "call"}, {"api_name": "gnosis.eth.contracts.get_safe_V0_0_1_contract", "line_number": 47, "usage_type": "call"}, {"api_name": "gnosis.eth.contracts.get_safe_V1_0_0_contract", "line_number": 48, "usage_type": "call"}, {"api_name": "gnosis.eth.contracts.get_safe_contract", "line_number": 49, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 53, "usage_type": "name"}, {"api_name": "web3.contract.ContractFunction", "line_number": 53, "usage_type": "name"}, {"api_name": "web3.contract.Contract", "line_number": 57, "usage_type": "name"}, {"api_name": "eth_utils.function_abi_to_4byte_selector", "line_number": 58, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 57, "usage_type": "name"}, {"api_name": "web3.contract.ContractFunction", "line_number": 57, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 61, "usage_type": "name"}, {"api_name": "hexbytes.HexBytes", "line_number": 68, "usage_type": "call"}, {"api_name": "typing.Union", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 71, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 83, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 94, "usage_type": "name"}, {"api_name": "hexbytes.HexBytes", "line_number": 107, "usage_type": "call"}, {"api_name": "web3._utils.abi.get_abi_input_names", "line_number": 114, "usage_type": "call"}, {"api_name": "web3._utils.abi.get_abi_input_types", "line_number": 115, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 116, "usage_type": "call"}, {"api_name": "hexbytes.HexBytes", "line_number": 116, "usage_type": "argument"}, {"api_name": "web3._utils.abi.map_abi_data", "line_number": 117, "usage_type": "call"}, {"api_name": "web3._utils.normalizers.BASE_RETURN_NORMALIZERS", "line_number": 117, "usage_type": "argument"}, {"api_name": "typing.Tuple", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 94, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 94, "usage_type": "name"}]} +{"seq_id": "96696249", "text": "import sys\n\nfrom django.core.management.base import BaseCommand\nfrom django.db.models.loading import get_model\n\nfrom orchestra.contrib.orchestration import manager\n\n\nclass Command(BaseCommand):\n help = 'Runs orchestration backends.'\n \n def add_arguments(self, parser):\n parser.add_argument('model',\n help='Label of a model to execute the orchestration.')\n parser.add_argument('query', nargs='*',\n help='Query arguments for filter().')\n parser.add_argument('--noinput', action='store_false', dest='interactive', default=True,\n help='Tells Django to NOT prompt the user for input of any kind.')\n parser.add_argument('--action', action='store', dest='action',\n default='save', help='Executes action. Defaults to \"save\".')\n parser.add_argument('--dry-run', action='store_true', dest='dry', default=False,\n help='Only prints scrtipt.')\n \n def handle(self, *args, **options):\n model = get_model(*options['model'].split('.'))\n action = options.get('action')\n interactive = options.get('interactive')\n dry = options.get('dry')\n kwargs = {}\n for comp in options.get('query', []):\n comps = iter(comp.split('='))\n for arg in comps:\n kwargs[arg] = next(comps).strip().rstrip(',')\n operations = []\n operations = set()\n route_cache = {}\n for instance in model.objects.filter(**kwargs):\n manager.collect(instance, action, operations=operations, route_cache=route_cache)\n scripts, block = manager.generate(operations)\n servers = []\n # Print scripts\n for key, value in scripts.items():\n server, __ = key\n backend, operations = value\n servers.append(server.name)\n sys.stdout.write('# Execute on %s\\n' % server.name)\n for method, commands in backend.scripts:\n script = '\\n'.join(commands) + '\\n'\n script = script.encode('ascii', errors='replace')\n sys.stdout.write(script.decode('ascii'))\n if interactive:\n context = {\n 'servers': ', '.join(servers),\n }\n msg = (\"\\n\\nAre your sure to execute the previous scripts on %(servers)s (yes/no)? \" % context)\n confirm = input(msg)\n while 1:\n if confirm not in ('yes', 'no'):\n confirm = input('Please enter either \"yes\" or \"no\": ')\n continue\n if confirm == 'no':\n return\n break\n if not dry:\n logs = manager.execute(scripts, block=block)\n for log in logs:\n print(log.stdout)\n sys.stderr.write(log.stderr)\n for log in logs:\n print(log.backend, log.state)\n", "sub_path": "orchestra/contrib/orchestration/management/commands/orchestrate.py", "file_name": "orchestrate.py", "file_ext": "py", "file_size_in_byte": 2890, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.core.management.base.BaseCommand", "line_number": 9, "usage_type": "name"}, {"api_name": "django.db.models.loading.get_model", "line_number": 25, "usage_type": "call"}, {"api_name": "orchestra.contrib.orchestration.manager.collect", "line_number": 38, "usage_type": "call"}, {"api_name": "orchestra.contrib.orchestration.manager", "line_number": 38, "usage_type": "name"}, {"api_name": "orchestra.contrib.orchestration.manager.generate", "line_number": 39, "usage_type": "call"}, {"api_name": "orchestra.contrib.orchestration.manager", "line_number": 39, "usage_type": "name"}, {"api_name": "sys.stdout.write", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 46, "usage_type": "attribute"}, {"api_name": "sys.stdout.write", "line_number": 50, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 50, "usage_type": "attribute"}, {"api_name": "orchestra.contrib.orchestration.manager.execute", "line_number": 65, "usage_type": "call"}, {"api_name": "orchestra.contrib.orchestration.manager", "line_number": 65, "usage_type": "name"}, {"api_name": "sys.stderr.write", "line_number": 68, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 68, "usage_type": "attribute"}]} +{"seq_id": "511122523", "text": "#coding:utf-8\n\nfrom selenium import webdriver\nfrom time import sleep\nfrom Lib.screen import Screen\n\n\ndesired_caps = {}\ndesired_caps['platformName'] = 'Android'\ndesired_caps['platformVersion'] = '19'\ndesired_caps['deviceName'] = '110e46e6'\ndesired_caps['appPackage'] = 'com.jsfund.basket'\ndesired_caps['appActivity'] = 'com.jsfund.basket.module.setting.view.starter.SplashActivity'\n# desired_caps['appPackage'] = 'com.android.calculator2'\n# desired_caps['appActivity'] = '.Calculator'\ndriver = webdriver.Remote('http://localhost:4723/wd/hub', desired_caps)\n# driver.find_element_by_id(\"com.android.calculator2:id/digit9\").click()\n# driver.find_element_by_id(\"action_create\").click()\n# driver.find_element_by_id(\"editText_phonenumber\").send_keys(\"15818678573\")\n# driver.find_element_by_id(\"editText_password\").send_keys(\"pengye\")\n# driver.find_element_by_id(\"button_login\").click()\n# driver.find_element_by_link_text(\"英雄榜\").click()\ndr=Screen(driver)\nmap_area = ('name', '首页')\ndr.get_element(map_area).click()\ndriver.find_element_by_name(\"首页\").click()\nimageview=driver.find_elements_by_id(\"imgv_event\")\nif imageview:\n print('The number of imageview on the page is:%s ' % len(imageview))\nimageview[2].click()\n\ndriver.find_elements_by_id(\"imgv_event\").click()\ndriver.find_elements_by_id(id_)\ndriver.find_element_by_class_name(\"android.widget.GridView\")\ndriver.find_elements_by_id(\"com.jsfund.basket:id/tv_hero\").click()\n\n\n\ndriver.find_element_by_name(\"5\").click()\ndriver.quit()\n\n", "sub_path": "demo/Demo/caltest.py", "file_name": "caltest.py", "file_ext": "py", "file_size_in_byte": 1491, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "selenium.webdriver.Remote", "line_number": 16, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 16, "usage_type": "name"}, {"api_name": "Lib.screen.Screen", "line_number": 23, "usage_type": "call"}]} +{"seq_id": "166614230", "text": "from urllib.request import urlopen\nimport ssl\nfrom bs4 import BeautifulSoup\n\nctx = ssl.create_default_context()\nctx.check_hostname = False\nctx.verify_mode = ssl.CERT_NONE\n\nurl = 'http://py4e-data.dr-chuck.net/known_by_Emilia.html'\n\ni = 8\nwhile i != 0:\n print(url)\n html = urlopen(url, context = ctx).read()\n soup = BeautifulSoup(html, 'html.parser')\n\n tags = soup('a')\n links = list()\n for tag in tags:\n links.append(tag.get('href', None))\n i = i - 1\n url = links[17]\n", "sub_path": "Python-Crash-Course/Day-3/Networking/Soup_Assignment.py", "file_name": "Soup_Assignment.py", "file_ext": "py", "file_size_in_byte": 499, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "ssl.create_default_context", "line_number": 5, "usage_type": "call"}, {"api_name": "ssl.CERT_NONE", "line_number": 7, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 14, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 15, "usage_type": "call"}]} +{"seq_id": "325389401", "text": "from flask import Flask\nfrom redis import Redis\n\n# https://yeasy.gitbooks.io/docker_practice/content/compose/usage.html\n# docker-compose up\n\napp = Flask(__name__)\nredis = Redis(host='redis', port=6379)\n\n@app.route('/')\ndef hello():\n count = redis.incr('hits')\n return 'Hello World! the page has been visisted {} times\\n'.format(count)\n\nif __name__ == \"__main__\":\n app.run(host=\"0.0.0.0\", debug=True)\n\n\n", "sub_path": "arsenal/eipi10/docker/compose-example/compose-app.py", "file_name": "compose-app.py", "file_ext": "py", "file_size_in_byte": 411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 8, "usage_type": "call"}, {"api_name": "redis.incr", "line_number": 12, "usage_type": "call"}]} +{"seq_id": "446659819", "text": "#!/usr/bin/env python3\nimport socket\nimport binascii\nimport sctp\nimport time\nimport ipaddress\nfrom pycrate_asn1dir import S1AP\nfrom pycrate_mobile.NAS import *\nfrom packets import *\nfrom kamene.all import *\nfrom kamene.contrib.gtp import *\nfrom six.moves import configparser\nfrom multiprocessing import Process, Queue\nimport threading\nfrom threading import Thread\n\n#Created PDU object of S1AP\nPDU = S1AP.S1AP_PDU_Descriptions.S1AP_PDU\n\ndef mysocket():\n #Creating sctp socket\n mysocket = sctp.sctpsocket_tcp(socket.AF_INET)\n mysocket.initparams.max_instreams = 1\n mysocket.initparams.num_ostreams = 1\n mysocket.events.clear()\n mysocket.events.data_io = 1\n mysocket.connect((HOST, PORT))\n return mysocket\n\n\ndef ap_request(cell_id):\n msg = s1_setup_request(cell_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n print(\"ap cell_id={}\".format(cell_id))\n data = ap_socket.recv(1024)\n return cell_id\n #time.sleep(1)\n\ndef ue_attach_and_release(ue_cell_id,imsi,enb_id,gtp_teid_icr):\n # mme_ue_s1ap_id = 1\n # GTP_TEID = 1234\n # GTP_ICMP_ADDRESS = '172.86.2.15'\n UDP_IP_ADDRESS_SRC = '172.24.2.123'\n # UDP_IP_ADDRESS_DST = '172.23.254.34'\n # tmsi =0xc0000fef\n\n #Sending InitialUEMessage, Attach request, PDN connectivity request Packet\n #Receving DownlinkNASTransport, Identity request\n msg= initial_ue_message_attach(ue_cell_id,imsi,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n print(\"ue cell_id={}\".format(cell_id))\n print(\"imsi={}\".format(imsi))\n print(\"enb_id={}\".format(enb_id))\n print(\"gtp_teid_icr={}\".format(gtp_teid_icr))\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'DownlinkNASTransport', 'protocolIEs'])\n mme_ue_s1ap_id = IEs[0]['value']\n mme_ue_s1ap_id = mme_ue_s1ap_id[1]\n break\n except:\n pass\n else:\n print(\"no initial_ue_message_attach msg\")\n #time.sleep(1)\n\n # Sending UplinkNASTransport, Identity response\n # Receving DownlinkNASTransport, ESM Information request\n msg = uplink_nas_transport_identity_response(ue_cell_id,mme_ue_s1ap_id,imsi,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'DownlinkNASTransport', 'protocolIEs'])\n break\n except:\n pass\n else:\n print(\"no initial_ue_message_attach msg\")\n\n #time.sleep(1)\n # Sending UplinkNASTransport, ESM Information response\n # Receving InitialContextSetupRequest, Attach accept, Activate default EPS bearer context request\n msg = uplink_nas_transport_esm_response(mme_ue_s1ap_id,ue_cell_id,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'InitialContextSetupRequest', 'protocolIEs'])\n data = PDU.get_val_paths()\n tl_addr = data[21]\n tl_addr = tl_addr[-1]\n tl_addr = tl_addr[0]\n tl_addr = ipaddress.ip_address(tl_addr).__str__()\n UDP_IP_ADDRESS_DST = tl_addr\n print(UDP_IP_ADDRESS_DST)\n gtp_teid = data[22]\n gtp_teid = gtp_teid[-1]\n gtp_teid = gtp_teid.hex()\n GTP_TEID = int(gtp_teid,16)\n print(GTP_TEID)\n gtp_ip = data[23]\n gtp_ip = gtp_ip[1]\n gtp_ip = hexlify(gtp_ip)\n gtp_ip = parse_NAS_MO(unhexlify(gtp_ip))\n gtp_ip = gtp_ip.__getitem__(0)\n gtp_ip = gtp_ip.get_val()\n gtp_ip = gtp_ip.__getitem__(3)\n gtp_ip = hexlify(gtp_ip)\n gtp_ip = parse_NAS_MO(unhexlify(gtp_ip))\n gtp_ip = gtp_ip.__getitem__(0)\n tmsi = gtp_ip[6]\n gtp_ip = gtp_ip[5]\n gtp_ip = gtp_ip['ESMActDefaultEPSBearerCtxtRequest'][3]\n gtp_ip = gtp_ip.__getitem__(1)[2]\n gtp_ip = gtp_ip.get_val()\n gtp_ip = ipaddress.IPv4Address(gtp_ip)\n gtp_ip = gtp_ip.__str__()\n GTP_ICMP_ADDRESS =gtp_ip\n tmsi = tmsi['EPSID'][6]\n tmsi = tmsi.get_val()\n tmsi = tmsi\n print(GTP_ICMP_ADDRESS)\n print(tmsi)\n break\n except:\n print(\"wrong response gtp\")\n else:\n print(\"no uplink_nas_transport_esm_response\")\n\n # time.sleep(1)\n # Sending InitialContextSetupResponse\n msg = initial_context_setup_response(mme_ue_s1ap_id,host_ip,gtp_teid_icr,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n\n # time.sleep(1)\n # Sending UplinkNASTransport, Attach complete, Activate default EPS bearer context accept\n msg = uplink_nas_transport_attach_complete(mme_ue_s1ap_id,ue_cell_id,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n\n #time.sleep(1)\n msg = gtp_echo_request(UDP_IP_ADDRESS_SRC,UDP_IP_ADDRESS_DST)\n send(msg)\n\n t_end = time.time() + delay_in_second\n while time.time() < t_end:\n msg =gtp_icmp_request(UDP_IP_ADDRESS_SRC,UDP_IP_ADDRESS_DST,GTP_ICMP_ADDRESS,GTP_TEID,data_in_bytes)\n send(msg)\n\n # Sending UEContextReleaseRequest [RadioNetwork-cause=user-inactivity]\n # Recieved UEContextReleaseCommand\n msg = ue_context_release_request(mme_ue_s1ap_id,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'UEContextReleaseCommand', 'protocolIEs'])\n break\n except:\n pass\n else:\n print(\"no uplink_nas_transport_tracking_area_complete msg\")\n\n # time.sleep(1)\n # Sending UEContextReleaseComplete\n msg = ue_context_release_complete(mme_ue_s1ap_id,enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n # time.sleep(1)\n return imsi, enb_id, tmsi\n\ndef ue_location_attach_and_release(ue_cell_id,ue_imsi, ue_enb_id, ue_tmsi):\n #mme_ue_s1ap_id_2 = 0\n # Sending, InitialMessage, Tracking area update request\n # Received DownlinkNASTransport, Identity request\n msg = initial_ue_message_tracking_area_update(ue_cell_id,ue_tmsi,ue_enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'DownlinkNASTransport', 'protocolIEs'])\n mme_ue_s1ap_id_2 = IEs[0]['value']\n mme_ue_s1ap_id_2 = mme_ue_s1ap_id_2[1]\n break\n except:\n pass\n else:\n print(\"no ue_location_attach_and_release msg\")\n\n # time.sleep(1)\n # Sending UplinkNASTransport,Identity response\n # Receving DownlinkNASTransport, Tracking area upadate accept\n msg = uplink_nas_transport_identity_response_location(ue_cell_id,mme_ue_s1ap_id_2,ue_imsi,ue_enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'DownlinkNASTransport', 'protocolIEs'])\n break\n except:\n pass\n else:\n print(\"no uplink_nas_transport_identity_response_location msg\")\n\n # time.sleep(1)\n #Sending UplinkNASTransport,Tracking area update complete\n #Recieved UEContextReleaseCommand\n msg =uplink_nas_transport_tracking_area_complete(ue_cell_id,mme_ue_s1ap_id_2,ue_enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n while True:\n data = ap_socket.recv(1024)\n data = hexlify(data)\n PDU.from_aper(unhexlify(data))\n if data:\n try:\n IEs = get_val_at(PDU, ['initiatingMessage', 'value', 'UEContextReleaseCommand', 'protocolIEs'])\n break\n except:\n pass\n else:\n print(\"no uplink_nas_transport_tracking_area_complete msg\")\n\n # time.sleep(1)\n #Sending UEContextReleaseComplete\n msg = ue_location_context_release_complete(mme_ue_s1ap_id_2,ue_enb_id)\n ap_socket.sctp_send(msg,ppid= 301989888)\n\n\ndef main(loop_limit,imsi_2,enb_id_2,gtp_teid_icr_2):\n time.sleep(1)\n while loop_limit > 0:\n ue_imsi, ue_enb_id, ue_tmsi = ue_attach_and_release(ue_cell_id,imsi_2,enb_id_2,gtp_teid_icr_2)\n ue_location_attach_and_release(ue_cell_id,ue_imsi, ue_enb_id, ue_tmsi)\n loop_limit -= 1\n\nif __name__ == '__main__':\n # Getting input data for request_parameter.txt\n config = configparser.ConfigParser()\n configFilePath = r'/home/amit/Documents/simulator/apsim/parameter.txt'\n config.read(configFilePath)\n config.sections()\n server_ip = config['SERVER']['server_ip']\n host_ip = config['SERVER']['host_ip']\n server_port = config['SERVER']['server_port']\n cell_id = int(config['ATTRIBUTE']['cell_id'])\n Number_of_ap = int(config['ATTRIBUTE']['Number_of_ap'])\n Number_of_ue = int(config['ATTRIBUTE']['Number_of_ue'])\n delay_in_second = int(config['ATTRIBUTE']['delay_in_second'])\n data_in_bytes = config['ATTRIBUTE']['data_in_bytes']\n imsi = int(config['ATTRIBUTE']['imsi'])\n loop_limit = int(config['ATTRIBUTE']['loop_limit'])\n enb_id = 1\n gtp_teid_icr = 33558538\n\n HOST = server_ip\n PORT = int(server_port)\n SRC_HOST = host_ip\n\n num_processes = 1\n processes = []\n processes_1 = []\n reset_no = 0\n\n for ap in range(Number_of_ap):\n ap_socket = mysocket()\n p_1 = Process(target=ap_request, args=(cell_id,))\n processes_1.append(p_1)\n p_1.start()\n ue_cell_id = cell_id\n imsi += reset_no\n enb_id += reset_no\n gtp_teid_icr += reset_no\n imsi_2 = imsi\n enb_id_2 = enb_id\n gtp_teid_icr_2 = gtp_teid_icr\n\n for ue in range(Number_of_ue):\n process_name = \"Started Process {}\".format(num_processes)\n q = Queue()\n p = Process(target=main, args=(loop_limit,imsi_2,enb_id_2,gtp_teid_icr_2), name=process_name)\n imsi_2 += 1\n enb_id_2 += 1\n gtp_teid_icr_2 += 1\n num_processes = num_processes + 1\n # processes.append(p)\n p.start()\n print(process_name)\n\n reset_no = Number_of_ue\n cell_id += 1\n time.sleep(1)\n for p in processes_1:\n p.join()\n", "sub_path": "def_2.py", "file_name": "def_2.py", "file_ext": "py", "file_size_in_byte": 11767, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pycrate_asn1dir.S1AP.S1AP_PDU_Descriptions", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pycrate_asn1dir.S1AP", "line_number": 18, "usage_type": "name"}, {"api_name": "sctp.sctpsocket_tcp", "line_number": 22, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 22, "usage_type": "attribute"}, {"api_name": "ipaddress.ip_address", "line_number": 104, "usage_type": "call"}, {"api_name": "ipaddress.IPv4Address", "line_number": 127, "usage_type": "call"}, {"api_name": "time.time", "line_number": 155, "usage_type": "call"}, {"api_name": "time.time", "line_number": 156, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 248, "usage_type": "call"}, {"api_name": "six.moves.configparser.ConfigParser", "line_number": 256, "usage_type": "call"}, {"api_name": "six.moves.configparser", "line_number": 256, "usage_type": "name"}, {"api_name": "multiprocessing.Process", "line_number": 284, "usage_type": "call"}, {"api_name": "multiprocessing.Queue", "line_number": 297, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 298, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 309, "usage_type": "call"}]} +{"seq_id": "631141933", "text": "#!/usr/bin/env python3\n\n\"\"\"Draw an image showing usage of disk block by block\"\"\"\n\nimport math\nimport re\nimport subprocess\nimport sys\n\nfrom PIL import Image\nimport hilbert_curve\n\n\ndef build_palette():\n \"\"\"Build a palette of rgb colors\"\"\"\n palette = [\n 0x00, 0x00, 0x00, # Black image border\n 0x80, 0x80, 0x80, # Gray used blocks\n 0xff, 0xff, 0xff, # White free blocks\n 0x00, 0x00, 0xff, # Blue superblocks\n 0x00, 0xff, 0x00, # Green group descriptors\n 0xff, 0xff, 0x00, # Yellow inode tables\n ]\n while len(palette) < 768:\n palette.append(0)\n\n return palette\n\n\nPALETTE = build_palette()\n\nCOLOR_KEY = {\n 'border': 0,\n 'used_blocks': 1,\n 'free_blocks': 2,\n 'superblocks': 3,\n 'group_descriptors': 4,\n 'inode_tables': 5,\n}\n\n\ndef parse_block_list(string, group_base=0):\n \"\"\"Parse a block list string\"\"\"\n ret = []\n args = string.split(',')\n for arg in args:\n arg = arg.strip()\n if not arg:\n # Discard empty free lists\n pass\n elif '-' in arg:\n # Range of blocks\n ret.append([int(x) + group_base for x in arg.split('-')])\n else:\n # Single block\n ret.append([int(arg) + group_base, int(arg) + group_base])\n return ret\n\n\ndef parse_line(line, group_base):\n \"\"\"Parse a single line from the blockdev information\"\"\"\n ret = {}\n\n match = re.match(r'.*superblock at ([0-9]*).*', line)\n if match:\n ret['superblocks'] = parse_block_list(match[1])\n\n match = re.match(r'.*Group descriptors at ([0-9-]*).*', line)\n if match:\n ret['group_descriptors'] = parse_block_list(match[1], group_base)\n\n match = re.match(r'.*bitmap at ([0-9-]*).*', line)\n if match:\n ret['group_descriptors'] = parse_block_list(match[1], group_base)\n\n match = re.match(r'.*Inode table at ([0-9-]*).*', line)\n if match:\n ret['inode_tables'] = parse_block_list(match[1], group_base)\n\n match = re.match(r'Free blocks: ([0-9-, ]*)', line)\n if match:\n ret['free_blocks'] = parse_block_list(match[1])\n\n return ret\n\n\ndef parse_disk(blockdev):\n \"\"\"Parse the disk usage information out of a blockdev\n blockdv must be formatted as ext? filesystem\n \"\"\"\n if blockdev == '-':\n dump = sys.stdin.read()\n else:\n dump = subprocess.check_output(\n [\"sudo\", \"dumpe2fs\", blockdev]).decode(\"utf-8\")\n total_blocks = None\n ret = {\n 'free_blocks': [],\n 'superblocks': [],\n 'group_descriptors': [],\n 'inode_tables': [],\n }\n group_base = None\n for line in dump.splitlines():\n line = line.strip()\n\n if line.startswith(\"Block count:\"):\n total_blocks = int(line.split(':')[1].strip())\n\n match = re.match(r'^Group [0-9]*: \\(Blocks ([0-9]*).*', line)\n if match:\n group_base = int(match[1])\n\n parsed = parse_line(line, group_base)\n for key, value in parsed.items():\n ret[key] += value\n\n return total_blocks, ret\n\n\ndef set_pixels(data, blocks, color):\n \"\"\"Set a range of pixels in the provided bytearray to the specified color\"\"\"\n start = blocks[0]\n end = blocks[1]\n length = end - start + 1\n data[start:end] = [color]*length\n\n\ndef stretch_array(data, newlength):\n \"\"\"Stretch an array to a new length.\"\"\"\n oldlength = len(data)\n assert oldlength <= newlength, \"Can't shrink in stretch function\"\n\n factor = float(newlength) / float(oldlength)\n\n result = bytearray(newlength)\n i = 0\n offset = 0.0\n for byte in data:\n offset += factor\n while offset >= 1.0:\n result[i] = byte\n i += 1\n offset -= 1.0\n\n return result\n\n\ndef hilbert_convert(data_linear):\n \"\"\"Map the data in data_linear into a hilbert curve.\"\"\"\n\n total_blocks = len(data_linear)\n\n # Scale to contain a hilbert curve\n m = 1\n while (2**m) * 2 < total_blocks:\n m += 1\n pixels = (2**m) * 2\n\n width = int(math.sqrt(pixels))\n height = width\n\n data_linear = stretch_array(data_linear, pixels)\n\n data = bytearray(pixels)\n for i, byte in zip(range(len(data_linear)), data_linear):\n x, y = hilbert_curve.d2xy(m, i)\n index = (y * width) + x\n data[index] = byte\n\n return data, width, height\n\n\ndef gen_image(total_blocks, parsed):\n \"\"\"Generate an image representing the disk\"\"\"\n data_linear = bytearray(total_blocks)\n set_pixels(data_linear, (0, total_blocks-1), COLOR_KEY['used_blocks'])\n\n for key in parsed.keys():\n for block in parsed[key]:\n set_pixels(data_linear, block, COLOR_KEY[key])\n\n data, width, height = hilbert_convert(data_linear)\n\n image = Image.frombytes('P', [width, height], bytes(data))\n image.putpalette(PALETTE)\n\n return image\n\n\ndef main():\n \"\"\"Main\"\"\"\n if len(sys.argv) < 2:\n raise Exception(\n \"Usage: {} [partition] [(optional)filename.png]\".format(\n sys.argv[0]))\n\n total_blocks, parsed = parse_disk(sys.argv[1])\n image = gen_image(total_blocks, parsed)\n\n if len(sys.argv) > 2:\n image.save(sys.argv[2])\n else:\n image.show()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "usage.py", "file_name": "usage.py", "file_ext": "py", "file_size_in_byte": 5245, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "re.match", "line_number": 64, "usage_type": "call"}, {"api_name": "re.match", "line_number": 68, "usage_type": "call"}, {"api_name": "re.match", "line_number": 72, "usage_type": "call"}, {"api_name": "re.match", "line_number": 76, "usage_type": "call"}, {"api_name": "re.match", "line_number": 80, "usage_type": "call"}, {"api_name": "sys.stdin.read", "line_number": 92, "usage_type": "call"}, {"api_name": "sys.stdin", "line_number": 92, "usage_type": "attribute"}, {"api_name": "subprocess.check_output", "line_number": 94, "usage_type": "call"}, {"api_name": "re.match", "line_number": 110, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 160, "usage_type": "call"}, {"api_name": "hilbert_curve.d2xy", "line_number": 167, "usage_type": "call"}, {"api_name": "PIL.Image.frombytes", "line_number": 185, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 185, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 193, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 196, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 198, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 201, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 202, "usage_type": "attribute"}]} +{"seq_id": "196142356", "text": "\nimport bpy\n\nfrom bpy.types import PropertyGroup\nfrom mathutils import Vector\n\nfrom . import cfg\nfrom . import at_panel\nfrom . import at_operators\nfrom . at_calc_func import(\n local_x_axis,\n local_y_axis,\n local_z_axis,\n at_all_in_one,\n rotate_self,\n at_random\n)\n\n\ndef update_seed(self, context):\n if self.at_mode == 'ADV':\n sc_min = (self.sc_min_x, self.sc_min_y, self.sc_min_z)\n sc_max = (self.sc_max_x, self.sc_max_y, self.sc_max_z)\n at_random(self.at_seed, cfg.mtx_list, cfg.list_duplicate, self.tr_min, self.tr_max, sc_min,\n sc_max, self.rot_min, self.rot_max, self.at_is_tr, self.at_is_sc, self.at_is_rot, self.sc_all)\n else:\n vec = Vector((1.0, 1.0, 1.0))\n tr = self.tr_rand * vec\n sc = self.sc_rand * vec\n rot = self.rot_rand * vec\n at_random(self.at_seed, cfg.mtx_list, cfg.list_duplicate, -tr, tr, sc, 100*vec, -rot, rot,\n self.at_is_tr, self.at_is_sc, self.at_is_rot, False)\n\n\ndef update_rtr(self, context):\n self.tr_max = self.tr_rand * Vector((1.0, 1.0, 1.0))\n self.tr_min = self.tr_rand * Vector((-1.0, -1.0, -1.0))\n\n\ndef update_rsc(self, context):\n self.sc_max_x, self.sc_max_y, self.sc_max_z = (100.0, 100.0, 100.0)\n rand = self.sc_rand\n self.sc_min_x, self.sc_min_y, self.sc_min_z = rand, rand, rand\n\n\ndef update_rrot(self, context):\n self.rot_max = self.rot_rand * Vector((1.0, 1.0, 1.0))\n self.rot_min = self.rot_rand * Vector((-1.0, -1.0, -1.0))\n\n\n# ---------------------------- Properties ----------------------\nclass AT_props(PropertyGroup):\n \"\"\"Property for array tools\"\"\"\n\n def add_at_element(self, nb_to_add=-1):\n \"\"\"Add nb_to_add copy in the scene\"\"\"\n if nb_to_add == -1:\n nb_to_add = cfg.at_values[1] - cfg.at_values[0]\n obj = cfg.obj_ref\n print(f\"add {nb_to_add} element(s) \")\n for i in range(nb_to_add):\n objcp = obj.copy()\n array_col = bpy.data.collections.get(\"Array_collection\")\n array_col.objects.link(objcp)\n if self.is_copy:\n objcp.data = obj.data.copy()\n\n cfg.list_duplicate.append(objcp)\n\n cfg.add_matrix(nb_to_add, cfg.list_duplicate)\n\n if self.is_tr_off_last:\n self.update_offset(bpy.context)\n else:\n self.update_global(bpy.context)\n\n\n def at_del_element(self, nb_to_del=-1):\n \"\"\"Delete copy from scene and from list\"\"\"\n if nb_to_del == -1:\n nb_to_del = cfg.at_values[0] - cfg.at_values[1]\n print(f\"del {nb_to_del} element(s)\")\n for i in range(nb_to_del):\n obj = cfg.list_duplicate.pop()\n array_col = bpy.data.collections.get(\"Array_collection\")\n array_col.objects.unlink(obj)\n bpy.data.objects.remove(obj, do_unlink=True)\n\n cfg.del_matrix(nb_to_del)\n\n if self.is_tr_off_last:\n self.update_offset(bpy.context)\n else:\n self.update_global(bpy.context)\n\n\n def at_del_all(self):\n \"\"\"Delete all copies and remove objects from lists\"\"\"\n bpy.ops.object.delete({\"selected_objects\": cfg.list_duplicate})\n cfg.list_duplicate.clear()\n cfg.mtx_list.clear()\n cfg.obj_ref = None\n print(\"Del_all done!\")\n\n # ----------------------- UI update ----------------------\n # --------------------------------------------------------\n # ----------------------- count update -------------------\n def updateCount(self, context):\n \"\"\"update the number of element(s)\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n cfg.add_value(int(self.count))\n cfg.del_value()\n\n # cfg.at_values[0] always get old count\n self.old_count = cfg.at_values[0]\n\n if self.old_count < self.count:\n self.add_at_element()\n elif self.old_count > self.count:\n self.at_del_element()\n\n def up_ui_updateCount(self, val):\n \"\"\"Update the value of the property count in UI\"\"\"\n self.is_prog_change = True\n self.count = val\n\n # ----------------------- translation update --------------\n def up_ui_tr_offset(self, val):\n \"\"\"Update the value of the property tr_offset in UI\"\"\"\n self.is_prog_change = True\n self.tr_offset = val\n\n def up_ui_tr_global(self, val):\n \"\"\"Update the value of the property tr_global in UI\"\"\"\n self.is_prog_change = True\n self.tr_global = val\n\n # ----------------------- scale update --------------------\n def up_ui_sc_offset(self, val):\n \"\"\"Update the value of the property sc_offset in UI\"\"\"\n self.is_prog_change = True\n self.sc_offset = val\n\n def up_ui_sc_global(self, val):\n \"\"\"Update the value of the property sc_global in UI\"\"\"\n self.is_prog_change = True\n self.sc_global = val\n\n # ----------------------- rotation update -----------------\n def up_ui_rot_offset(self, val):\n \"\"\"Update the value of the property rot_offset in UI\n val is a float \"\"\"\n self.is_prog_change = True\n self.rot_offset = val\n\n def up_ui_rot_global(self, val):\n \"\"\"Update the value of the property rot_global in UI\"\"\"\n self.is_prog_change = True\n self.rot_global = val\n\n def up_ui_sc_min_x(self, val):\n \"\"\"Update the value of the property sc_min_x in UI\"\"\"\n self.is_prog_change = True\n self.sc_min_x = val\n\n def up_ui_sc_min_y(self, val):\n \"\"\"Update the value of the property sc_min_y in UI\"\"\"\n self.is_prog_change = True\n self.sc_min_y = val\n\n def up_ui_sc_min_z(self, val):\n \"\"\"Update the value of the property sc_min_z in UI\"\"\"\n self.is_prog_change = True\n self.sc_min_z = val\n\n def up_ui_sc_max_x(self, val):\n \"\"\"Update the value of the property sc_max_x in UI\"\"\"\n self.is_prog_change = True\n self.sc_max_x = val\n\n def up_ui_sc_max_y(self, val):\n \"\"\"Update the value of the property sc_max_y in UI\"\"\"\n self.is_prog_change = True\n self.sc_max_y = val\n\n def up_ui_sc_max_z(self, val):\n \"\"\"Update the value of the property sc_max_z in UI\"\"\"\n self.is_prog_change = True\n self.sc_max_z = val\n\n\n # ---------------------------------------------------------\n def update_offset(self, context):\n \"\"\"Update for all offsets\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n # user change offset\n self.is_tr_off_last = True\n i = 1\n\n loc_x = local_x_axis(cfg.obj_ref)\n loc_y = local_y_axis(cfg.obj_ref)\n loc_z = local_z_axis(cfg.obj_ref)\n localxyz = (loc_x, loc_y, loc_z)\n\n if self.at_pivot is not None:\n for elem in cfg.list_duplicate:\n r_off = i * Vector((self.rot_offset))\n t_off = i * self.tr_offset\n s_off = Vector((1.0, 1.0, 1.0)) - (i * (cfg.obj_ref.scale - (self.sc_offset/100)))\n at_all_in_one(cfg.obj_ref, elem, r_off, localxyz, t_off, s_off, self.at_pivot.location)\n i += 1\n else:\n for elem in cfg.list_duplicate:\n r_off = i * Vector((self.rot_offset))\n t_off = i * self.tr_offset\n s_off = Vector((1.0, 1.0, 1.0)) - (i * (cfg.obj_ref.scale-(self.sc_offset/100)))\n at_all_in_one(cfg.obj_ref, elem, (0.0, 0.0, 0.0), localxyz, t_off, s_off, cfg.obj_ref.location)\n rotate_self(elem, r_off, localxyz)\n i += 1\n self.up_ui_tr_global(t_off)\n self.up_ui_sc_global(s_off * 100)\n # global rotation need to include reference object\n self.up_ui_rot_global(r_off + Vector((self.rot_offset)))\n\n cfg.update_matrix(context)\n\n\n def update_global(self, context):\n \"\"\"Update for all globals\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n # user change global\n self.is_tr_off_last = False\n i = 1\n # local axis\n loc_x = local_x_axis(cfg.obj_ref)\n loc_y = local_y_axis(cfg.obj_ref)\n loc_z = local_z_axis(cfg.obj_ref)\n\n localxyz = (loc_x, loc_y, loc_z)\n\n translation_offset = Vector(self.tr_global) / (self.count - 1)\n scale_offset = (cfg.obj_ref.scale-(self.sc_global/100)) / (self.count - 1)\n rotation_offset = Vector((self.rot_global)) / self.count\n\n for elem in cfg.list_duplicate:\n r_off = i * Vector((rotation_offset))\n t_off = i * translation_offset\n s_off = Vector((1.0, 1.0, 1.0)) - (i*scale_offset)\n if self.at_pivot is not None:\n at_all_in_one(cfg.obj_ref, elem, r_off, localxyz, t_off, s_off, self.at_pivot.location)\n else:\n at_all_in_one(cfg.obj_ref, elem, (0.0, 0.0, 0.0), localxyz, t_off, s_off, cfg.obj_ref.location)\n rotate_self(elem, r_off, localxyz)\n i += 1\n self.up_ui_tr_offset(translation_offset)\n self.up_ui_sc_offset(Vector((100.0, 100.0, 100.0)) - (scale_offset*100))\n self.up_ui_rot_offset(rotation_offset)\n\n cfg.update_matrix(context)\n\n\n # ----------------------- is_copy update ------------------\n def up_ui_is_copy(self):\n \"\"\"Update the value of the property is_copy in UI\"\"\"\n self.is_prog_change = True\n self.is_copy = False\n\n def update_is_copy(self, context):\n \"\"\"Allow a copy or duplicate(copy link by default)\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n if self.is_copy: # no need to rebuild all\n for elem in cfg.list_duplicate:\n elem.data = cfg.obj_ref.data.copy()\n else: # since the value change (now duplicate) copies can't become duplicates, so need to rebuild\n nb_to_rebuild = len(cfg.list_duplicate)\n bpy.ops.object.delete({\"selected_objects\": cfg.list_duplicate})\n cfg.list_duplicate.clear()\n self.add_at_element(nb_to_rebuild)\n\n\n # -------------- update min and max ---------------\n # if user enter a max value < min, change min and vice versa\n def up_tr_min(self, context):\n \"\"\"Update tr_max if tr_min is higher\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n for i in range(3):\n if self.tr_min[i] > self.tr_max[i]:\n self.is_prog_change = True\n self.tr_max[i] = self.tr_min[i]\n\n def up_tr_max(self, context):\n \"\"\"Update tr_min if tr_max is lower\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n for i in range(3):\n if self.tr_min[i] > self.tr_max[i]:\n self.is_prog_change = True\n self.tr_min[i] = self.tr_max[i]\n\n def up_sc_min_x(self, context):\n \"\"\"Update sc_max_x if sc_min_x is higher\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_x > self.sc_max_x\n if test and self.sc_all:\n self.up_ui_sc_max_x(self.sc_min_x)\n # with uniform : min_x = min_y = min_z same for max\n self.up_ui_sc_min_y(self.sc_min_x)\n self.up_ui_sc_min_z(self.sc_min_x)\n self.up_ui_sc_max_y(self.sc_min_x)\n self.up_ui_sc_max_z(self.sc_min_x)\n elif self.sc_all:\n self.up_ui_sc_min_y(self.sc_min_x)\n self.up_ui_sc_min_z(self.sc_min_x)\n self.up_ui_sc_max_y(self.sc_max_x)\n self.up_ui_sc_max_z(self.sc_max_x)\n elif test:\n self.up_ui_sc_max_x(self.sc_min_x)\n\n\n def up_sc_min_y(self, context):\n \"\"\"Update sc_max_y if sc_min_y is higher\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_y > self.sc_max_y\n if test and self.sc_all:\n self.up_ui_sc_max_y(self.sc_min_y)\n # with uniform : min_x = min_y = min_z same for max\n self.up_ui_sc_min_x(self.sc_min_y)\n self.up_ui_sc_min_z(self.sc_min_y)\n self.up_ui_sc_max_x(self.sc_min_y)\n self.up_ui_sc_max_y(self.sc_min_y)\n elif self.sc_all:\n self.up_ui_sc_min_x(self.sc_min_y)\n self.up_ui_sc_min_z(self.sc_min_y)\n self.up_ui_sc_max_x(self.sc_max_y)\n self.up_ui_sc_max_z(self.sc_max_y)\n elif test:\n self.up_ui_sc_max_y(self.sc_min_y)\n\n\n def up_sc_min_z(self, context):\n \"\"\"Update sc_max_z if sc_min_z is higher\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_z > self.sc_max_z\n if test and self.sc_all:\n self.up_ui_sc_max_z(self.sc_min_z)\n # with uniform : min_x = min_y = min_z same for max\n self.up_ui_sc_min_x(self.sc_min_z)\n self.up_ui_sc_min_y(self.sc_min_z)\n self.up_ui_sc_max_x(self.sc_min_z)\n self.up_ui_sc_max_y(self.sc_min_z)\n elif self.sc_all:\n self.up_ui_sc_min_x(self.sc_min_z)\n self.up_ui_sc_min_y(self.sc_min_z)\n self.up_ui_sc_max_x(self.sc_max_z)\n self.up_ui_sc_max_y(self.sc_max_z)\n elif test:\n self.up_ui_sc_max_y(self.sc_min_z)\n\n\n def up_sc_max_x(self, context):\n \"\"\"Update sc_min_x if sc_max_x is lower\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_x > self.sc_max_x\n if test and self.sc_all:\n self.up_ui_sc_min_x(self.sc_max_x)\n # with uniform : min_x = min_y = min_z same for max\n self.up_ui_sc_max_y(self.sc_max_x)\n self.up_ui_sc_max_z(self.sc_max_x)\n self.up_ui_sc_min_y(self.sc_max_x)\n self.up_ui_sc_min_z(self.sc_max_x)\n elif self.sc_all:\n self.up_ui_sc_max_y(self.sc_max_x)\n self.up_ui_sc_max_z(self.sc_max_x)\n self.up_ui_sc_min_y(self.sc_min_x)\n self.up_ui_sc_min_z(self.sc_min_x)\n elif test:\n self.up_ui_sc_min_x(self.sc_max_x)\n\n\n def up_sc_max_y(self, context):\n \"\"\"Update sc_min_y if sc_max_y is lower\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_y > self.sc_max_y\n if test and self.sc_all:\n self.up_ui_sc_min_y(self.sc_max_y)\n # with uniform : min_x = min_y = min_z same for max\n self.up_ui_sc_max_x(self.sc_max_y)\n self.up_ui_sc_max_z(self.sc_max_y)\n self.up_ui_sc_min_x(self.sc_max_y)\n self.up_ui_sc_min_z(self.sc_max_y)\n elif self.sc_all:\n self.up_ui_sc_max_x(self.sc_max_y)\n self.up_ui_sc_max_z(self.sc_max_y)\n self.up_ui_sc_min_x(self.sc_min_y)\n self.up_ui_sc_min_z(self.sc_min_y)\n elif test:\n self.up_ui_sc_min_y(self.sc_max_y)\n\n\n def up_sc_max_z(self, context):\n \"\"\"Update sc_min_z if sc_max_z is lower\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n test = self.sc_min_z > self.sc_max_z\n if test and self.sc_all:\n self.up_ui_sc_min_z(self.sc_max_z)\n\n self.up_ui_sc_max_x(self.sc_max_z)\n self.up_ui_sc_max_y(self.sc_max_z)\n self.up_ui_sc_min_x(self.sc_max_z)\n self.up_ui_sc_min_y(self.sc_max_z)\n elif self.sc_all:\n self.up_ui_sc_max_x(self.sc_max_z)\n self.up_ui_sc_max_y(self.sc_max_z)\n self.up_ui_sc_min_x(self.sc_min_z)\n self.up_ui_sc_min_y(self.sc_min_z)\n elif test:\n self.up_ui_sc_min_z(self.sc_max_z)\n\n\n def up_rot_min(self, context):\n \"\"\"Update rot_max if rot_min is higher\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n for i in range(3):\n if self.rot_min[i] > self.rot_max[i]:\n self.is_prog_change = True\n self.rot_max[i] = self.rot_min[i]\n\n\n def up_rot_max(self, context):\n \"\"\"Update rot_min if rot_max is lower\"\"\"\n if self.is_prog_change:\n self.is_prog_change = False\n else:\n for i in range(3):\n if self.rot_min[i] > self.rot_max[i]:\n self.is_prog_change = True\n self.rot_min[i] = self.rot_max[i]\n\n\n # ----------------------- reset all properties ------------\n def up_ui_reset(self):\n \"\"\"Reset all UI properties\"\"\"\n self.up_ui_updateCount(2)\n self.up_ui_is_copy()\n self.up_ui_tr_offset(Vector((2.0, 0.0, 0.0)))\n self.up_ui_tr_global(Vector((2.0, 0.0, 0.0)))\n self.up_ui_sc_offset((100, 100, 100))\n self.up_ui_sc_global((100, 100, 100))\n self.up_ui_rot_offset(Vector((0.0, 0.0, 0.0)))\n self.up_ui_rot_global(Vector((0.0, 0.0, 0.0)))\n\n # ------------------------ property list -------------------\n count: bpy.props.IntProperty(\n name='Count',\n description=\"Number of elements, original count as one\",\n default=2,\n soft_min=2,\n update=updateCount\n )\n # keep the old count to compare later with the current\n old_count: bpy.props.IntProperty(default=2)\n\n # booleans use to know if user or prog change the value to avoid continuous loop\n is_prog_change: bpy.props.BoolProperty(default=False) # True if prog change value\n\n # which one between offset and global user call last, True is offset, False global\n is_tr_off_last: bpy.props.BoolProperty(default=True)\n\n # True if addon is initialised\n already_start: bpy.props.BoolProperty(default=False)\n\n # if the user need a single copy or a duplicate (link object)\n is_copy: bpy.props.BoolProperty(\n name=\"Copy only\",\n description=\"Duplicate or copy, default is duplicate\",\n default=False,\n update=update_is_copy\n )\n\n # translation vector offset\n tr_offset: bpy.props.FloatVectorProperty(\n name='Offset',\n description=\"Distance between elements\",\n default=(2.0, 0.0, 0.0),\n subtype='TRANSLATION',\n unit='LENGTH',\n precision=2,\n step=50,\n update=update_offset\n )\n\n # global translation distance\n tr_global: bpy.props.FloatVectorProperty(\n name='Global',\n description=\"Distance between the original and the last element\",\n default=(2.0, 0.0, 0.0),\n subtype='TRANSLATION',\n unit='LENGTH',\n precision=2,\n step=50,\n update=update_global\n )\n\n at_pivot: bpy.props.PointerProperty(\n name='Pivot',\n description=\"Object you want as pivot point. If none, pivot point is the object's origine\",\n type=bpy.types.Object\n )\n\n # scaling vector offset\n sc_offset: bpy.props.FloatVectorProperty(\n name='Offset',\n description=\"Incremental scale of the next elements\",\n default=(100.0, 100.0, 100.0),\n subtype='XYZ',\n precision=1,\n step=100,\n update=update_offset\n )\n\n # global scaling\n sc_global: bpy.props.FloatVectorProperty(\n name='Global',\n description=\"Scale of the last element\",\n default=(100.0, 100.0, 100.0),\n subtype='XYZ',\n precision=1,\n step=100,\n update=update_global\n )\n\n # rotation vector offset\n rot_offset: bpy.props.FloatVectorProperty(\n name='Offset',\n description=\"Angle between each element\",\n default=(0.0, 0.0, 0.0),\n subtype='XYZ',\n unit='ROTATION',\n step=500, # = 5\n update=update_offset\n )\n\n # global rotation\n rot_global: bpy.props.FloatVectorProperty(\n name='Global',\n description=\"Maximum angle from the reference to the last element\",\n default=(0.0, 0.0, 0.0),\n subtype='XYZ',\n unit='ROTATION',\n step=500, # = 5\n update=update_global\n )\n\n # ----------------------- random part ----------------------\n at_seed: bpy.props.IntProperty(\n name='Seed',\n description=\"Seed value for random\",\n soft_min=0,\n default=0,\n update=update_seed\n )\n\n at_mode: bpy.props.EnumProperty(\n name=\"Mode\",\n description=\"Choose between simple mode or advanced\",\n items=(('SIM', 'Simple', \"Simple mode\"),\n ('ADV', 'Advanced', \"Advanced mode\")),\n default='SIM'\n )\n\n at_is_tr: bpy.props.BoolProperty(\n name=\"Add translation\",\n description=\"Add translation in random?\",\n default=False\n )\n\n at_is_sc: bpy.props.BoolProperty(\n name=\"Add scale\",\n description=\"Add scale in random?\",\n default=False\n )\n\n at_is_rot:bpy.props.BoolProperty(\n name=\"Add rotation\",\n description=\"Add rotation in random?\",\n default=False\n )\n\n tr_min: bpy.props.FloatVectorProperty(\n name=\"min\",\n description=\"Minimum random value for translation\",\n unit='LENGTH',\n default=(0.0, 0.0, 0.0),\n update=up_tr_min\n )\n\n tr_max: bpy.props.FloatVectorProperty(\n name=\"max\",\n description=\"Maximum random value for translation\",\n unit='LENGTH',\n default=(0.0, 0.0, 0.0),\n update=up_tr_max\n )\n\n tr_rand: bpy.props.FloatProperty(\n name=\"Translation\",\n description=\"Random values for all axis\",\n unit='LENGTH',\n default=0.0,\n update=update_rtr\n )\n\n sc_all: bpy.props.BoolProperty(\n name=\"uniform scale\",\n description=\"Uniform or non uniform scale, default is non uniform.\",\n default=False\n )\n\n sc_min_x: bpy.props.IntProperty(\n name=\"min\",\n description=\"Minimum random value for x scale\",\n default=100,\n update=up_sc_min_x\n )\n\n sc_min_y: bpy.props.IntProperty(\n name=\"min\",\n description=\"Minimum random value for y scale\",\n default=100,\n update=up_sc_min_y\n )\n\n sc_min_z: bpy.props.IntProperty(\n name=\"min\",\n description=\"Minimum random value for z scale\",\n default=100,\n update=up_sc_min_z\n )\n\n sc_max_x: bpy.props.IntProperty(\n name=\"max\",\n description=\"Maximum random value for x scale\",\n default=100,\n update=up_sc_max_x\n )\n\n sc_max_y: bpy.props.IntProperty(\n name=\"max\",\n description=\"Maximum random value for y scale\",\n default=100,\n update=up_sc_max_y\n )\n\n sc_max_z: bpy.props.IntProperty(\n name=\"max\",\n description=\"Maximum random value for z scale\",\n default=100,\n update=up_sc_max_z\n )\n\n sc_rand: bpy.props.IntProperty(\n name=\"Scale\",\n description=\"Random scale value for all axis\",\n default=100,\n update=update_rsc\n )\n rot_min: bpy.props.FloatVectorProperty(\n name=\"min\",\n description=\"Minimum random value for rotation\",\n unit='ROTATION',\n default=(0.0, 0.0, 0.0),\n update=up_rot_min\n )\n\n rot_max: bpy.props.FloatVectorProperty(\n name=\"max\",\n description=\"Maximum random value for rotation\",\n unit='ROTATION',\n default=(0.0, 0.0, 0.0),\n update=up_rot_max\n )\n\n rot_rand:bpy.props.FloatProperty(\n name=\"Rotation\",\n description=\"Random rotation for all axis\",\n unit='ROTATION',\n default=0.0,\n update=update_rrot\n )\n\n", "sub_path": "at_interface.py", "file_name": "at_interface.py", "file_ext": "py", "file_size_in_byte": 24278, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "at_calc_func.at_random", "line_number": 24, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 27, "usage_type": "call"}, {"api_name": "at_calc_func.at_random", "line_number": 31, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 36, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 37, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 47, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 48, "usage_type": "call"}, {"api_name": "bpy.types.PropertyGroup", "line_number": 52, "usage_type": "name"}, {"api_name": "bpy.data.collections.get", "line_number": 63, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 63, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 73, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 75, "usage_type": "attribute"}, {"api_name": "bpy.data.collections.get", "line_number": 85, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 85, "usage_type": "attribute"}, {"api_name": "bpy.data.objects.remove", "line_number": 87, "usage_type": "call"}, {"api_name": "bpy.data", "line_number": 87, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 92, "usage_type": "attribute"}, {"api_name": "bpy.context", "line_number": 94, "usage_type": "attribute"}, {"api_name": "bpy.ops.object.delete", "line_number": 99, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 99, "usage_type": "attribute"}, {"api_name": "at_calc_func.local_x_axis", "line_number": 204, "usage_type": "call"}, {"api_name": "at_calc_func.local_y_axis", "line_number": 205, "usage_type": "call"}, {"api_name": "at_calc_func.local_z_axis", "line_number": 206, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 211, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 213, "usage_type": "call"}, {"api_name": "at_calc_func.at_all_in_one", "line_number": 214, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 218, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 220, "usage_type": "call"}, {"api_name": "at_calc_func.at_all_in_one", "line_number": 221, "usage_type": "call"}, {"api_name": "at_calc_func.rotate_self", "line_number": 222, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 227, "usage_type": "call"}, {"api_name": "at_calc_func.local_x_axis", "line_number": 241, "usage_type": "call"}, {"api_name": "at_calc_func.local_y_axis", "line_number": 242, "usage_type": "call"}, {"api_name": "at_calc_func.local_z_axis", "line_number": 243, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 247, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 249, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 252, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 254, "usage_type": "call"}, {"api_name": "at_calc_func.at_all_in_one", "line_number": 256, "usage_type": "call"}, {"api_name": "at_calc_func.at_all_in_one", "line_number": 258, "usage_type": "call"}, {"api_name": "at_calc_func.rotate_self", "line_number": 259, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 262, "usage_type": "call"}, {"api_name": "bpy.ops.object.delete", "line_number": 284, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 284, "usage_type": "attribute"}, {"api_name": "mathutils.Vector", "line_number": 470, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 471, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 474, "usage_type": "call"}, {"api_name": "mathutils.Vector", "line_number": 475, "usage_type": "call"}, {"api_name": "bpy.props.IntProperty", "line_number": 478, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 478, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 486, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 486, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 489, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 489, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 492, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 492, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 495, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 495, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 498, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 498, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 506, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 506, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 518, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 518, "usage_type": "attribute"}, {"api_name": "bpy.props.PointerProperty", "line_number": 529, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 529, "usage_type": "attribute"}, {"api_name": "bpy.types", "line_number": 532, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 536, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 536, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 547, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 547, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 558, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 558, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 569, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 569, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 580, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 580, "usage_type": "attribute"}, {"api_name": "bpy.props.EnumProperty", "line_number": 588, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 588, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 596, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 596, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 602, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 602, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 608, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 608, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 614, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 614, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 622, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 622, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatProperty", "line_number": 630, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 630, "usage_type": "attribute"}, {"api_name": "bpy.props.BoolProperty", "line_number": 638, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 638, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 644, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 644, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 651, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 651, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 658, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 658, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 665, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 665, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 672, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 672, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 679, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 679, "usage_type": "attribute"}, {"api_name": "bpy.props.IntProperty", "line_number": 686, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 686, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 692, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 692, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatVectorProperty", "line_number": 700, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 700, "usage_type": "attribute"}, {"api_name": "bpy.props.FloatProperty", "line_number": 708, "usage_type": "call"}, {"api_name": "bpy.props", "line_number": 708, "usage_type": "attribute"}]} +{"seq_id": "561808040", "text": "\"\"\"\nLeet Code Challenge\nIn a town, there are N people labelled from 1 to N. There is a rumor that one of these people is secretly the town judge.\n\nIf the town judge exists, then:\n\nThe town judge trusts nobody.\nEverybody (except for the town judge) trusts the town judge.\nThere is exactly one person that satisfies properties 1 and 2.\nYou are given trust, an array of pairs trust[i] = [a, b] representing that the person labelled a trusts the person labelled b.\n\nIf the town judge exists and can be identified, return the label of the town judge. Otherwise, return -1.\n\n\n\nExample 1:\n\nInput: N = 2, trust = [[1,2]]\nOutput: 2\nExample 2:\n\nInput: N = 3, trust = [[1,3],[2,3]]\nOutput: 3\nExample 3:\n\nInput: N = 3, trust = [[1,3],[2,3],[3,1]]\nOutput: -1\nExample 4:\n\nInput: N = 3, trust = [[1,2],[2,3]]\nOutput: -1\nExample 5:\n\nInput: N = 4, trust = [[1,3],[1,4],[2,3],[2,4],[4,3]]\nOutput: 3\n\n\nNote:\n\n1 <= N <= 1000\ntrust.length <= 10000\ntrust[i] are all different\ntrust[i][0] != trust[i][1]\n1 <= trust[i][0], trust[i][1] <= N\n\nhttps://leetcode.com/explore/challenge/card/may-leetcoding-challenge/535/week-2-may-8th-may-14th/3325/\n\"\"\"\nfrom typing import List\n\n\nclass Solution:\n def findJudge(self, N: int, trust: List[List[int]]) -> int:\n \"\"\"\n # Solution 1,\n in_degree = [0] * (N + 1)\n out_degree = [0] * (N + 1)\n for a,b in trust:\n out_degree[a] += 1\n in_degree[b] += 1\n for i in range(1, N + 1):\n if out_degree[i] == 0 and in_degree[i] == N - 1:\n return i\n return -1\n \"\"\"\n \"\"\"\n # Solution 2 - 764 ms\n # degree = [0 for _ in range(N + 1)]\n degree = [0] * (N + 1)\n for a, b in trust:\n degree[a] -= 1\n degree[b] += 1\n for i, num in enumerate(degree[1:], 1):\n if num == N - 1:\n return i\n return -1\n \"\"\"\n # Solution 3 - 734 ms\n trusters = {x[0] for x in trust}\n print(trusters)\n trusted_by = 0\n candidate = -1\n for i in range(1, N + 1):\n if i not in trusters:\n candidate = i\n\n for a, b in trust:\n if b == candidate:\n trusted_by += 1\n\n if trusted_by != N - 1:\n candidate = -1\n\n return candidate\n\n\n# Main Call\nsolution = Solution()\nN = 2\ntrust = [[1, 2]]\nprint(solution.findJudge(N, trust))\nN = 3\ntrust = [[1, 3], [2, 3]]\nprint(solution.findJudge(N, trust))\n", "sub_path": "src/arrays/findJudge.py", "file_name": "findJudge.py", "file_ext": "py", "file_size_in_byte": 2481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.List", "line_number": 52, "usage_type": "name"}]} +{"seq_id": "142889939", "text": "'''\n------------------\n\n``earthio.tif``\n~~~~~~~~~~~~~~~~~~~\n\nTools for reading GeoTiff files. Typically use the interface through\n\n - :func:`earthio.load_array`\n - :func:`earthio.`load_meta`\n\n'''\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nfrom collections import OrderedDict\nimport copy\nimport gc\nimport logging\nimport os\n\nimport numpy as np\nimport rasterio as rio\nimport xarray as xr\n\nfrom earthio.metadata_selection import match_meta\nfrom earthio.util import (geotransform_to_coords,\n geotransform_to_bounds,\n SPATIAL_KEYS,\n raster_as_2d,\n READ_ARRAY_KWARGS,\n take_geo_transform_from_meta,\n BandSpec,\n meta_strings_to_dict)\n\nfrom earthio import ElmStore\nfrom six import string_types\n\nlogger = logging.getLogger(__name__)\n\n\n__all__ = ['load_tif_meta',\n 'load_dir_of_tifs_meta',\n 'load_dir_of_tifs_array',]\n\n\ndef load_tif_meta(filename):\n '''Read the metadata of one TIF file\n\n Parameters:\n :filename: str: path and filename of TIF to read\n\n Returns:\n :file: TIF file\n :meta: Dictionary with meta data about the file, including;\n\n - **meta**: Meta attributes of the TIF file\n - **geo_transform**: transform\n - **bounds**: Bounds of the TIF\n - **height**: Hight of the TIF\n - **width**: Width of the TIF\n - **name**: The filename\n - **sub_dataset_name**: The filename\n\n '''\n r = rio.open(filename, driver='GTiff')\n if r.count != 1:\n raise ValueError('earthio.tif only reads tif files with 1 band (shape of [1, y, x]). Found {} bands'.format(r.count))\n meta = {'meta': r.meta}\n meta['geo_transform'] = r.get_transform()\n meta['bounds'] = r.bounds\n meta['height'] = r.height\n meta['width'] = r.width\n meta['name'] = meta['sub_dataset_name'] = filename\n return r, meta_strings_to_dict(meta)\n\n\ndef ls_tif_files(dir_of_tiffs):\n tifs = os.listdir(dir_of_tiffs)\n tifs = [f for f in tifs if f.lower().endswith('.tif') or f.lower().endswith('.tiff')]\n return [os.path.join(dir_of_tiffs, t) for t in tifs]\n\n\ndef array_template(r, meta, **reader_kwargs):\n dtype = getattr(np, r.dtypes[0])\n\n if not 'window' in reader_kwargs:\n if 'height' in reader_kwargs:\n height = reader_kwargs['height']\n else:\n height = meta['height']\n if 'width' in reader_kwargs:\n width = reader_kwargs['width']\n else:\n width = meta['width']\n else:\n if 'height' in reader_kwargs:\n height = reader_kwargs['height']\n else:\n height = np.diff(reader_kwargs['window'][0])[0]\n if 'width' in reader_kwargs:\n width = reader_kwargs['width']\n else:\n width = np.diff(reader_kwargs['window'][0])[0]\n return np.empty((1, height, width), dtype=dtype)\n\n\ndef load_dir_of_tifs_meta(dir_of_tiffs, band_specs=None, **meta):\n '''Load metadata from same-directory GeoTiffs representing\n different bands of the same image.\n\n Parameters:\n :dir_of_tiffs: Directory with GeoTiffs\n :band_specs: List of earthio.BandSpec objects\n :meta: included in returned metadata'''\n logger.debug('load_dir_of_tif_meta {}'.format(dir_of_tiffs))\n tifs = ls_tif_files(dir_of_tiffs)\n meta = copy.deepcopy(meta)\n band_order_info = []\n for band_idx, tif in enumerate(tifs):\n raster, band_meta = load_tif_meta(tif)\n\n if band_specs:\n for idx, band_spec in enumerate(band_specs):\n if (isinstance(band_spec, BandSpec) and match_meta(band_meta, band_spec)) or (isinstance(band_spec, string_types) and band_spec in tif):\n band_order_info.append((idx, tif, band_spec, band_meta))\n break\n\n else:\n band_name = 'band_{}'.format(band_idx)\n band_order_info.append((band_idx, tif, band_name, band_meta))\n\n if not band_order_info or (band_specs and (len(band_order_info) != len(band_specs))):\n logger.debug('len(band_order_info) {}'.format(len(band_order_info)))\n raise ValueError('Failure to find all bands specified by '\n 'band_specs with length {}.\\n'\n 'Found only {} of '\n 'them.'.format(len(band_specs), len(band_order_info)))\n # error if they do not share coords at this point\n band_order_info.sort(key=lambda x:x[0])\n meta['band_meta'] = [b[-1] for b in band_order_info]\n meta['band_order_info'] = [b[:-1] for b in band_order_info]\n return meta\n\ndef open_prefilter(filename, meta, **reader_kwargs):\n '''Placeholder for future operations on open file rasterio\n handle like resample / aggregate or setting width, height, etc\n on load. TODO see optional kwargs to rasterio.open'''\n try:\n r = rio.open(filename)\n raster = array_template(r, meta, **reader_kwargs)\n logger.debug('reader_kwargs {} raster template shape {}'.format(reader_kwargs, raster.shape))\n r.read(out=raster)\n return r, raster\n except Exception as e:\n logger.info('Failed to rasterio.open {}'.format(filename))\n raise\n\ndef load_dir_of_tifs_array(dir_of_tiffs, meta, band_specs=None):\n '''Return an ElmStore where each subdataset is a DataArray\n\n Parameters:\n :dir_of_tiffs: directory of GeoTiff files where each is a\n single band raster\n :meta: meta from earthio.load_dir_of_tifs_meta\n :band_specs: list of earthio.BandSpec objects,\n defaulting to reading all subdatasets\n as bands\n Returns:\n :X: ElmStore\n\n '''\n\n logger.debug('load_dir_of_tifs_array: {}'.format(dir_of_tiffs))\n band_order_info = meta['band_order_info']\n tifs = ls_tif_files(dir_of_tiffs)\n logger.info('Load tif files from {}'.format(dir_of_tiffs))\n\n if not len(band_order_info):\n raise ValueError('No matching bands with '\n 'band_specs {}'.format(band_specs))\n native_dims = ('y', 'x')\n elm_store_dict = OrderedDict()\n attrs = {'meta': meta}\n attrs['band_order'] = []\n for (idx, filename, band_spec), band_meta in zip(band_order_info, meta['band_meta']):\n band_name = getattr(band_spec, 'name', band_spec)\n if not isinstance(band_spec, string_types):\n reader_kwargs = {k: getattr(band_spec, k)\n for k in READ_ARRAY_KWARGS\n if getattr(band_spec, k)}\n else:\n reader_kwargs = {}\n if 'buf_xsize' in reader_kwargs:\n reader_kwargs['width'] = reader_kwargs.pop('buf_xsize')\n if 'buf_ysize' in reader_kwargs:\n reader_kwargs['height'] = reader_kwargs.pop('buf_ysize')\n if 'window' in reader_kwargs:\n reader_kwargs['window'] = tuple(map(tuple, reader_kwargs['window']))\n # TODO multx, multy should be handled here as well?\n if reader_kwargs:\n multy = band_meta['height'] / reader_kwargs.get('height', band_meta['height'])\n multx = band_meta['width'] / reader_kwargs.get('width', band_meta['width'])\n else:\n multx = multy = 1.\n band_meta.update(reader_kwargs)\n geo_transform = take_geo_transform_from_meta(band_spec, **attrs)\n handle, raster = open_prefilter(filename, band_meta, **reader_kwargs)\n raster = raster_as_2d(raster)\n if getattr(band_spec, 'stored_coords_order', ['y', 'x'])[0] == 'y':\n rows, cols = raster.shape\n else:\n rows, cols = raster.T.shape\n if geo_transform is None:\n band_meta['geo_transform'] = handle.get_transform()\n else:\n band_meta['geo_transform'] = geo_transform\n band_meta['geo_transform'][1] *= multx\n band_meta['geo_transform'][-1] *= multy\n\n coords_x, coords_y = geotransform_to_coords(cols,\n rows,\n band_meta['geo_transform'])\n elm_store_dict[band_name] = xr.DataArray(raster,\n coords=[('y', coords_y),\n ('x', coords_x),],\n dims=native_dims,\n attrs=band_meta)\n\n attrs['band_order'].append(band_name)\n gc.collect()\n return ElmStore(elm_store_dict, attrs=attrs)\n", "sub_path": "earthio/tif.py", "file_name": "tif.py", "file_ext": "py", "file_size_in_byte": 8725, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 38, "usage_type": "call"}, {"api_name": "rasterio.open", "line_number": 65, "usage_type": "call"}, {"api_name": "earthio.util.meta_strings_to_dict", "line_number": 74, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 80, "usage_type": "call"}, {"api_name": "os.path", "line_number": 80, "usage_type": "attribute"}, {"api_name": "numpy.diff", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 104, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 117, "usage_type": "call"}, {"api_name": "earthio.util.BandSpec", "line_number": 124, "usage_type": "argument"}, {"api_name": "earthio.metadata_selection.match_meta", "line_number": 124, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 124, "usage_type": "argument"}, {"api_name": "rasterio.open", "line_number": 149, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 182, "usage_type": "call"}, {"api_name": "six.string_types", "line_number": 187, "usage_type": "argument"}, {"api_name": "earthio.util.READ_ARRAY_KWARGS", "line_number": 189, "usage_type": "name"}, {"api_name": "earthio.util.take_geo_transform_from_meta", "line_number": 206, "usage_type": "call"}, {"api_name": "earthio.util.raster_as_2d", "line_number": 208, "usage_type": "call"}, {"api_name": "earthio.util.geotransform_to_coords", "line_number": 220, "usage_type": "call"}, {"api_name": "xarray.DataArray", "line_number": 223, "usage_type": "call"}, {"api_name": "gc.collect", "line_number": 230, "usage_type": "call"}, {"api_name": "earthio.ElmStore", "line_number": 231, "usage_type": "call"}]} +{"seq_id": "509040725", "text": "import datetime, calendar, re\r\nfrom ast import literal_eval as make_tuple\r\n\r\nfrom django.http import HttpResponseRedirect\r\nfrom django.views import generic\r\nfrom django.contrib.auth.mixins import LoginRequiredMixin\r\nfrom django.contrib import messages\r\nfrom django.shortcuts import render, render_to_response\r\n\r\nimport kaagtest.constants as c\r\nfrom .models import Lid, Atleet, Trainer\r\nfrom trainingen.models import Training\r\nfrom wedstrijden.models import Prestatie\r\nfrom .forms import smallAtleetForm, fullAtleetForm\r\nfrom trainingen.forms import filter_in_tijdForm, atleet_filterform\r\n\r\n# Create your views here.\r\n\r\n\r\nclass AlleLedenView(LoginRequiredMixin, generic.ListView):\r\n template_name = 'leden/totaal.html'\r\n context_object_name = 'trainerslijst'\r\n\r\n def get_context_data(self, **kwargs):\r\n context = super().get_context_data(**kwargs)\r\n context.update({\r\n 'atletenlijst': Atleet.objects.order_by('familienaam'),\r\n 'more_context': Atleet.objects.all(),\r\n })\r\n return context\r\n\r\n def get_queryset(self):\r\n return Trainer.objects.all()\r\n\r\n\r\nclass AtleetDetailView(LoginRequiredMixin, generic.DetailView):\r\n model = Atleet\r\n template_name = 'leden/Atleetdetail.html'\r\n\r\n def get_queryset(self):\r\n return Atleet.objects.all()\r\n\r\n def get_context_data(self, **kwargs):\r\n context = super().get_context_data(**kwargs)\r\n # detailview geeft in context al context['atleet'] is atleet in kwestie\r\n atleet = context['atleet']\r\n\r\n alle_prestaties = atleet.prestatie_set.all()\r\n d = c.ALLE_DISCIPLINES_DICT\r\n\r\n alle_disciplines = [d[pres.discipline]for pres in alle_prestaties]\r\n # enkel unieke waarden, zou sneller moeten zijn dan for loop en 'if not in' checken\r\n unieke_disciplines = sorted(list(set(alle_disciplines)))\r\n\r\n wedstrijden_verzameling = sorted(\r\n [[d[pres.discipline], pres.resultaat, pres.wind, pres.wedstrijd] for pres in alle_prestaties if pres.verified])\r\n onofficiele_wedstrijden_verzameling = sorted(\r\n [[d[pres.discipline], pres.resultaat, pres.wind, pres.wedstrijd] for pres in alle_prestaties if not pres.verified])\r\n\r\n if len(onofficiele_wedstrijden_verzameling) > 0:\r\n context.update({\r\n 'unverified': True,\r\n 'alle_onofficieel': onofficiele_wedstrijden_verzameling,\r\n })\r\n if atleet.emailadres:\r\n context['email'] = True\r\n\r\n context.update({\r\n 'alle_disciplines': unieke_disciplines,\r\n 'alle_prestaties': wedstrijden_verzameling,\r\n })\r\n\r\n ##Trainingen##\r\n filtered_trainingen = [training for training in Training.objects.all(\r\n ) if atleet in training.atleten.all()]\r\n\r\n context.update({\r\n 'trainingset': filtered_trainingen,\r\n })\r\n\r\n return context\r\n\r\n\r\nclass TrainerDetailView(LoginRequiredMixin, generic.DetailView):\r\n model = Trainer\r\n template_name = 'leden/Trainerdetail.html'\r\n\r\n def get_queryset(self):\r\n return Trainer.objects.order_by('-user__last_name')\r\n\r\n def get_context_data(self, **kwargs):\r\n context = super().get_context_data(**kwargs)\r\n\r\n if self.request.user.is_authenticated:\r\n user = self.request.user\r\n\r\n trainer = user.trainer\r\n context['naam'] = trainer.user.first_name+' '+trainer.user.last_name\r\n context['id'] = trainer.id\r\n atletenqueryset = trainer.atleten.all()\r\n context['numatleten'] = len(atletenqueryset)\r\n current_year = datetime.datetime.now().year\r\n context['num_trainingen_jaar'] = len(trainer.training_set.filter(datum__gte=datetime.date(current_year, 1, 1)))\r\n return context\r\n\r\n\r\nclass TrainingsgroepView(LoginRequiredMixin, generic.TemplateView):\r\n template_name = 'trainingen/generic.html'\r\n\r\n def get_context_data(self, **kwargs):\r\n\r\n context = super().get_context_data(**kwargs)\r\n \r\n ############\r\n #Trainingen#\r\n ############\r\n\r\n trainer = self.request.user.trainer\r\n form_trainingen = filter_in_tijdForm()\r\n trainingen_queryset = trainer.training_set.all().order_by('-datum')\r\n current_year = datetime.datetime.now().year\r\n\r\n context.update({\r\n 'trainingset': trainingen_queryset, #WIP TODO moet ik alles doorsturen? kan groot worden, misschien enkel ~15 sturen?\r\n #stuur ik die zelfs door? of krijgt de gebruiker enkel de html? want dan kan ik daarin gewoon slice doen\r\n 'id': trainer.id,\r\n 'naam': trainer.getnaam,\r\n 'form_trainingen': form_trainingen,\r\n 'trainer_num_atleten': len(trainer.atleten.all()),\r\n 'num_trainingen_jaar': len(trainer.training_set.filter(datum__gte=datetime.date(current_year, 1, 1))),\r\n })\r\n\r\n #########\r\n #Atleten#\r\n #########\r\n if not trainingen_queryset:\r\n trainingen_queryset = trainer.training_set.all().order_by('-datum')\r\n\r\n alle_atleten = trainer.atleten.all().order_by('familienaam')\r\n frequencytable = [[atleet, 0] for atleet in alle_atleten]\r\n for training in trainingen_queryset:\r\n for atleet in training.atleten.all():\r\n for i, referentie_atleet in enumerate(frequencytable):\r\n if atleet == referentie_atleet[0]:\r\n frequencytable[i][1] += 1\r\n sorted_atleten = [atleetfreq[0] for atleetfreq in sorted(frequencytable, key=lambda x: x[1], reverse=True)]\r\n ##sorted_atleten is een lijst van alle atleten gesorteerd volgens aantal trainingen, met interne sortering op familienaam##\r\n\r\n context = self.atleetfilter_context(sorted_atleten, context)\r\n\r\n form_atleten = atleet_filterform()\r\n\r\n context.update({\r\n 'form_atleten': form_atleten,\r\n 'form_newatleet': smallAtleetForm(),\r\n })\r\n\r\n ############\r\n #Prestaties#\r\n ############\r\n\r\n alle_prestaties = Prestatie.objects.all()\r\n d = c.ALLE_DISCIPLINES_DICT\r\n\r\n onofficiele_wedstrijden_verzameling = sorted(\r\n [[d[pres.discipline], pres] for pres in alle_prestaties if not pres.verified], key=lambda x: x[1].atleet.familienaam)\r\n\r\n if len(onofficiele_wedstrijden_verzameling) > 0:\r\n context.update({\r\n 'alle_onofficieel': onofficiele_wedstrijden_verzameling,\r\n })\r\n\r\n return context\r\n\r\n def post(self, request, **kwargs):\r\n form_trainingen = filter_in_tijdForm(request.POST)\r\n form_atleten = atleet_filterform(request.POST)\r\n form_newatleet = smallAtleetForm(request.POST)\r\n context = self.get_context_data(**kwargs)\r\n\r\n if form_trainingen.is_valid():\r\n form = form_trainingen\r\n startfilter = form.cleaned_data['startfilter']\r\n endfilter = form.cleaned_data['endfilter']\r\n preset = form.cleaned_data['preset']\r\n\r\n # preset heeft altijd voorrang\r\n if preset != 'title':\r\n current_date = datetime.datetime.now().date()\r\n #raise Exception(current_date, current_date-datetime.timedelta(days=10))\r\n if preset == 'cur_week':\r\n startfilter = current_date-datetime.timedelta(days=7)\r\n endfilter = current_date\r\n elif preset == 'past_week':\r\n startfilter = current_date-datetime.timedelta(days=14)\r\n endfilter = current_date-datetime.timedelta(days=7)\r\n elif preset == 'cur_month':\r\n startfilter = datetime.datetime(\r\n current_date.year, current_date.month, 1)\r\n endfilter = current_date\r\n elif preset == 'past_month':\r\n startfilter = datetime.datetime(\r\n current_date.year, current_date.month-1, 1)\r\n # timedelta zorgt ervoor dat ik niet moet weten hoe lang de vorige maand was\r\n endfilter = datetime.datetime(\r\n current_date.year, current_date.month, 1)-datetime.timedelta(days=1)\r\n elif preset == 'cur_year':\r\n startfilter = datetime.datetime(current_date.year, 1, 1)\r\n endfilter = current_date\r\n elif preset == 'past_year':\r\n startfilter = datetime.datetime(current_date.year-1, 1, 1)\r\n endfilter = datetime.datetime(current_date.year, 1, 1)-datetime.timedelta(days=1)\r\n else:\r\n raise Warning('Preset had an invalid value!')\r\n\r\n trainingen_queryset = context['trainingset'].filter(datum__gte=startfilter, datum__lte=endfilter)\r\n print(trainingen_queryset)\r\n context.update({\r\n 'startfilter': '%02d-%02d-%04d' % (startfilter.day, startfilter.month, startfilter.year),\r\n 'endfilter': '%02d-%02d-%04d' % (endfilter.day, endfilter.month, endfilter.year),\r\n 'trainingset': trainingen_queryset,\r\n })\r\n return render(request, self.template_name, context)\r\n elif not startfilter and not endfilter:\r\n messages.warning('Not a valid form entry: No preset selected, No date boundaries given.')\r\n return HttpResponseRedirect('#trainingen')\r\n trainingen_queryset = context['trainingset'].filter(datum__gte=startfilter, datum__lte=endfilter)\r\n context.update({\r\n 'startfilter': '%02d-%02d-%04d' % (startfilter.day, startfilter.month, startfilter.year),\r\n 'endfilter': '%02d-%02d-%04d' % (endfilter.day, endfilter.month, endfilter.year),\r\n 'trainingset': trainingen_queryset,\r\n })\r\n return render(request, self.template_name, context)\r\n\r\n elif form_atleten.is_valid():\r\n form = form_atleten\r\n preset = form.cleaned_data['preset']\r\n if preset != 'title':\r\n atletenlijst = context['alle_atleten']\r\n if preset == 'first_name':\r\n sorted_atleten = sorted(\r\n atletenlijst, key=lambda x: x.voornaam)\r\n elif preset == 'fam_name':\r\n sorted_atleten = sorted(\r\n atletenlijst, key=lambda x: x.familienaam)\r\n elif preset == 'num_train':\r\n pass # is the base sorting already\r\n elif preset == 'num_pres':\r\n dic = {}\r\n for atleet in atletenlijst:\r\n dic[atleet] = 0\r\n allp = Prestatie.objects.all()\r\n for pres in allp:\r\n if pres.atleet == atleet:\r\n dic[atleet] += 1\r\n sorted_atleten = sorted(atletenlijst, key=lambda x: dic[x], reverse=True)\r\n\r\n else:\r\n raise Warning('Preset had an invalid value!')\r\n if preset != 'num_train':\r\n context = self.atleetfilter_context(sorted_atleten, context)\r\n return render(request, self.template_name, context)\r\n elif form_newatleet.is_valid():\r\n form = form_newatleet\r\n voornaam = form.cleaned_data['voornaam']\r\n familienaam = form.cleaned_data['familienaam']\r\n if Atleet.objects.filter(voornaam__iexact=voornaam, familienaam__iexact=familienaam).exists():\r\n \r\n atleet = Atleet.objects.get(\r\n voornaam__iexact=voornaam, familienaam__iexact=familienaam)\r\n\r\n request.user.trainer.atleten.add(atleet)\r\n\r\n messages.success(request, '%s %s toegevoegd aan de groep.'%(voornaam.capitalise(), familienaam.capitalise()))\r\n return HttpResponseRedirect('#')\r\n else:\r\n messages.info(request, voornaam+';'+familienaam)\r\n return HttpResponseRedirect('/leden/trainers/unsuccessful/')\r\n else:\r\n #overloop prestatieform\r\n at_least_one_valid_key = False\r\n for key in request.POST.keys():\r\n if 'checkboxinput' in key:\r\n at_least_one_valid_key = True\r\n pres_id = re.search(r'\\d+', key).group()\r\n prestatie = Prestatie.objects.get(id=pres_id)\r\n prestatie.set_verified(True);\r\n\r\n if(at_least_one_valid_key):\r\n return render(request, self.template_name, self.get_context_data(**kwargs))\r\n else:\r\n raise Warning('Not a valid form entry.')\r\n\r\n def atleetfilter_context(self, sorted_atleten, context):\r\n aantal_atleten = len(sorted_atleten)\r\n inc = int(aantal_atleten/3)\r\n if aantal_atleten % 3 == 1:\r\n delta1 = 1\r\n delta2 = 0\r\n elif aantal_atleten % 3 == 2:\r\n delta1 = 1\r\n delta2 = 1\r\n elif aantal_atleten % 3 == 0:\r\n delta1 = 0\r\n delta2 = 0\r\n context.update({\r\n 'alle_atleten': sorted_atleten,\r\n 'alle_atletenp1': sorted_atleten[:inc+delta1], # WIP TODO\r\n 'alle_atletenp2': sorted_atleten[inc+delta1: 2*inc+delta1+delta2],\r\n 'alle_atletenp3': sorted_atleten[2*inc+delta1+delta2:],\r\n })\r\n return context\r\n\r\n\r\nclass AddAtleetTrainerView(LoginRequiredMixin, generic.TemplateView):\r\n template_name = \"leden/AtleetToevoegenTrainer.html\"\r\n\r\n def get_context_data(self, **kwargs):\r\n context = super().get_context_data(**kwargs)\r\n if self.request.user.is_authenticated:\r\n\r\n trainer = self.request.user.trainer\r\n current_year = datetime.datetime.now().year\r\n context.update({\r\n 'naam': trainer.getnaam,\r\n 'id': trainer.id,\r\n 'num_atleten_trainer': len(trainer.atleten.all()),\r\n 'num_trainingen_jaar': len(trainer.training_set.filter(datum__gte=datetime.date(current_year, 1, 1))),\r\n 'form': smallAtleetForm(),\r\n })\r\n\r\n return context\r\n # Voor FORM\r\n\r\n def post(self, request, **kwargs):\r\n form = smallAtleetForm(request.POST)\r\n # check whether it's valid:\r\n if form.is_valid():\r\n # process the data in form.cleaned_data as required\r\n voornaam = form.cleaned_data['voornaam']\r\n familienaam = form.cleaned_data['familienaam']\r\n if Atleet.objects.filter(voornaam__iexact=voornaam, familienaam__iexact=familienaam).exists():\r\n \r\n atleet = Atleet.objects.get(\r\n voornaam__iexact=voornaam, familienaam__iexact=familienaam)\r\n\r\n request.user.trainer.atleten.add(atleet)\r\n\r\n # redirect to overzicht trainingsgroep:\r\n return HttpResponseRedirect('/leden/trainers/successful/')\r\n else:\r\n messages.info(request, voornaam+';'+familienaam)\r\n return HttpResponseRedirect('/leden/trainers/unsuccessful/')\r\n\r\n\r\nclass AddNewAtleetView(LoginRequiredMixin, generic.TemplateView):\r\n template_name = \"leden/AtleetToevoegen.html\"\r\n\r\n def get_context_data(self, **kwargs):\r\n context = super().get_context_data(**kwargs)\r\n\r\n storage = messages.get_messages(self.request)\r\n if len(storage) == 1: # I only expect one message, being 'voornaam;familienaam'\r\n for message in storage: # you have to iterate over storage, this deletes the messages\r\n if message.level != 20: # if not info message, leave the data be\r\n storage.used = False\r\n break\r\n voornaam, familienaam = message.message.split(';')\r\n # check for somewhat valid entries, could be improved upon (WIP TODO)\r\n if voornaam != '' and familienaam != '':\r\n context.update({\r\n 'form_voornaam': voornaam,\r\n 'form_familienaam': familienaam,\r\n 'form_special_defaults': True,\r\n })\r\n\r\n if self.request.user.is_authenticated:\r\n trainer = self.request.user.trainer\r\n context.update({\r\n 'naam': trainer.getnaam,\r\n 'id': trainer.id,\r\n 'atleten_all': len(Atleet.objects.all()),\r\n 'trainers_all': len(Trainer.objects.all()),\r\n 'form': fullAtleetForm(),\r\n })\r\n\r\n return context\r\n\r\n def post(self, request, **kwargs):\r\n form = fullAtleetForm(request.POST)\r\n # check whether it's valid:\r\n if form.is_valid():\r\n # process the data in form.cleaned_data as required\r\n voornaam = form.cleaned_data['voornaam']\r\n familienaam = form.cleaned_data['familienaam']\r\n geboortedatum = form.cleaned_data['geboortedatum']\r\n atleettype = form.cleaned_data['wedstrijdatleet']\r\n wedstrijdnummer = form.cleaned_data['wedstrijdnummer']\r\n emailadres = form.cleaned_data['emailadres']\r\n\r\n user = request.user\r\n\r\n if not Atleet.objects.filter(voornaam__iexact=voornaam, familienaam__iexact=familienaam).exists(): #wip todo kan beter\r\n newatleet = Atleet(voornaam=voornaam, familienaam=familienaam,\r\n geboortedatum=geboortedatum, emailadres=emailadres, wedstrijdatleet=atleettype, wedstrijdnummer=wedstrijdnummer, username_aanvrager=user.username) #vervangen door fieldmapping? wip todo\r\n newatleet.save()\r\n \r\n #ip_address storage\r\n user.trainer.register_ip(c.get_ipaddress(request))\r\n\r\n messages.warning(request, 'Nieuwe Atleet Toegevoegd!')\r\n return HttpResponseRedirect('/leden/atleten/add/')\r\n messages.warning(\r\n request, 'Er bestond al een atleet met die naam. << WIP: eventueel link voor edit atleet voorzien?>>')\r\n return HttpResponseRedirect('/leden/atleten/add/')\r\n\r\n messages.warning(request, form.errors)\r\n return HttpResponseRedirect('/leden/atleten/add/')\r\n", "sub_path": "leden/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 18513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 20, "usage_type": "name"}, {"api_name": "django.views.generic.ListView", "line_number": 20, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 20, "usage_type": "name"}, {"api_name": "models.Atleet.objects.order_by", "line_number": 27, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 27, "usage_type": "name"}, {"api_name": "models.Atleet.objects.all", "line_number": 28, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 28, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 28, "usage_type": "name"}, {"api_name": "models.Trainer.objects.all", "line_number": 33, "usage_type": "call"}, {"api_name": "models.Trainer.objects", "line_number": 33, "usage_type": "attribute"}, {"api_name": "models.Trainer", "line_number": 33, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 36, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 36, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 36, "usage_type": "name"}, {"api_name": "models.Atleet", "line_number": 37, "usage_type": "name"}, {"api_name": "models.Atleet.objects.all", "line_number": 41, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 41, "usage_type": "name"}, {"api_name": "kaagtest.constants.ALLE_DISCIPLINES_DICT", "line_number": 49, "usage_type": "attribute"}, {"api_name": "kaagtest.constants", "line_number": 49, "usage_type": "name"}, {"api_name": "trainingen.models.Training.objects.all", "line_number": 74, "usage_type": "call"}, {"api_name": "trainingen.models.Training.objects", "line_number": 74, "usage_type": "attribute"}, {"api_name": "trainingen.models.Training", "line_number": 74, "usage_type": "name"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 84, "usage_type": "name"}, {"api_name": "django.views.generic.DetailView", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 84, "usage_type": "name"}, {"api_name": "models.Trainer", "line_number": 85, "usage_type": "name"}, {"api_name": "models.Trainer.objects.order_by", "line_number": 89, "usage_type": "call"}, {"api_name": "models.Trainer.objects", "line_number": 89, "usage_type": "attribute"}, {"api_name": "models.Trainer", "line_number": 89, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 102, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 102, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 103, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 107, "usage_type": "name"}, {"api_name": "django.views.generic.TemplateView", "line_number": 107, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 107, "usage_type": "name"}, {"api_name": "trainingen.forms.filter_in_tijdForm", "line_number": 119, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 121, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 121, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 130, "usage_type": "call"}, {"api_name": "trainingen.forms.atleet_filterform", "line_number": 151, "usage_type": "call"}, {"api_name": "forms.smallAtleetForm", "line_number": 155, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects.all", "line_number": 162, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects", "line_number": 162, "usage_type": "attribute"}, {"api_name": "wedstrijden.models.Prestatie", "line_number": 162, "usage_type": "name"}, {"api_name": "kaagtest.constants.ALLE_DISCIPLINES_DICT", "line_number": 163, "usage_type": "attribute"}, {"api_name": "kaagtest.constants", "line_number": 163, "usage_type": "name"}, {"api_name": "trainingen.forms.filter_in_tijdForm", "line_number": 176, "usage_type": "call"}, {"api_name": "trainingen.forms.atleet_filterform", "line_number": 177, "usage_type": "call"}, {"api_name": "forms.smallAtleetForm", "line_number": 178, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 189, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 189, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 192, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 195, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 196, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 198, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 202, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 205, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 206, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 208, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 211, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 212, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 212, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 223, "usage_type": "call"}, {"api_name": "django.contrib.messages.warning", "line_number": 225, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 225, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 226, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 233, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects.all", "line_number": 252, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects", "line_number": 252, "usage_type": "attribute"}, {"api_name": "wedstrijden.models.Prestatie", "line_number": 252, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 262, "usage_type": "call"}, {"api_name": "models.Atleet.objects.filter", "line_number": 267, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 267, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 267, "usage_type": "name"}, {"api_name": "models.Atleet.objects.get", "line_number": 269, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 269, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 269, "usage_type": "name"}, {"api_name": "django.contrib.messages.success", "line_number": 274, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 274, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 275, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 277, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 277, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 278, "usage_type": "call"}, {"api_name": "re.search", "line_number": 285, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects.get", "line_number": 286, "usage_type": "call"}, {"api_name": "wedstrijden.models.Prestatie.objects", "line_number": 286, "usage_type": "attribute"}, {"api_name": "wedstrijden.models.Prestatie", "line_number": 286, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 290, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 315, "usage_type": "name"}, {"api_name": "django.views.generic.TemplateView", "line_number": 315, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 315, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 323, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 323, "usage_type": "attribute"}, {"api_name": "datetime.date", "line_number": 328, "usage_type": "call"}, {"api_name": "forms.smallAtleetForm", "line_number": 329, "usage_type": "call"}, {"api_name": "forms.smallAtleetForm", "line_number": 336, "usage_type": "call"}, {"api_name": "models.Atleet.objects.filter", "line_number": 342, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 342, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 342, "usage_type": "name"}, {"api_name": "models.Atleet.objects.get", "line_number": 344, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 344, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 344, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 350, "usage_type": "call"}, {"api_name": "django.contrib.messages.info", "line_number": 352, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 352, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 353, "usage_type": "call"}, {"api_name": "django.contrib.auth.mixins.LoginRequiredMixin", "line_number": 356, "usage_type": "name"}, {"api_name": "django.views.generic.TemplateView", "line_number": 356, "usage_type": "attribute"}, {"api_name": "django.views.generic", "line_number": 356, "usage_type": "name"}, {"api_name": "django.contrib.messages.get_messages", "line_number": 362, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 362, "usage_type": "name"}, {"api_name": "models.Atleet.objects.all", "line_number": 382, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 382, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 382, "usage_type": "name"}, {"api_name": "models.Trainer.objects.all", "line_number": 383, "usage_type": "call"}, {"api_name": "models.Trainer.objects", "line_number": 383, "usage_type": "attribute"}, {"api_name": "models.Trainer", "line_number": 383, "usage_type": "name"}, {"api_name": "forms.fullAtleetForm", "line_number": 384, "usage_type": "call"}, {"api_name": "forms.fullAtleetForm", "line_number": 390, "usage_type": "call"}, {"api_name": "models.Atleet.objects.filter", "line_number": 403, "usage_type": "call"}, {"api_name": "models.Atleet.objects", "line_number": 403, "usage_type": "attribute"}, {"api_name": "models.Atleet", "line_number": 403, "usage_type": "name"}, {"api_name": "models.Atleet", "line_number": 404, "usage_type": "call"}, {"api_name": "kaagtest.constants.get_ipaddress", "line_number": 409, "usage_type": "call"}, {"api_name": "kaagtest.constants", "line_number": 409, "usage_type": "name"}, {"api_name": "django.contrib.messages.warning", "line_number": 411, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 411, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 412, "usage_type": "call"}, {"api_name": "django.contrib.messages.warning", "line_number": 413, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 413, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 415, "usage_type": "call"}, {"api_name": "django.contrib.messages.warning", "line_number": 417, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 417, "usage_type": "name"}, {"api_name": "django.http.HttpResponseRedirect", "line_number": 418, "usage_type": "call"}]} +{"seq_id": "22547804", "text": "import argparse\nimport logging\nimport config\nimport os\n\nlogger = logging.getLogger(__name__)\n\ndef run(app_name, worker):\n parser = argparse.ArgumentParser(description='Admin Server Help')\n parser.add_argument(\n '-c', '--config', type=str, nargs=\"*\",\n help=\"List of configuration files to import (python modules)\")\n parser.add_argument('-m', '--method', help=\"all, invoices or the like\")\n parser.add_argument('action', choices=['start', 'stop','restart', 'run'])\n cmd_args = parser.parse_args()\n\n\n config.configure(cmd_args.config or [])\n\n import logging_config\n logging_config.configure(app_name)\n\n\n daemon = worker(os.path.join(config.pid_dir, '{}.pid'.format(app_name)))\n if cmd_args.action == 'start':\n daemon.start()\n elif cmd_args.action == 'stop':\n daemon.stop()\n elif cmd_args.action == 'restart':\n daemon.restart()\n elif cmd_args.action == 'run':\n if not cmd_args.method:\n daemon.run()\n else:\n attr = getattr(daemon, cmd_args.method, None)\n if attr is None:\n logger.critical('method unknown')\n else:\n attr()\n daemon.close()\n\n\n\n", "sub_path": "cli.py", "file_name": "cli.py", "file_ext": "py", "file_size_in_byte": 1220, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 6, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 9, "usage_type": "call"}, {"api_name": "config.configure", "line_number": 18, "usage_type": "call"}, {"api_name": "logging_config.configure", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "config.pid_dir", "line_number": 24, "usage_type": "attribute"}]} +{"seq_id": "505074920", "text": "import cv2\nfrom hough_transform import *\nimport stereo\nfrom panorama import *\n\n\ndef run_disparity_ssd():\n path1 = '../data/pair2-L.png'\n path2 = '../data/pair2-R.png'\n\n # path1 = '../data/pair0-L.png'\n # path2 = '../data/pair0-R.png'\n\n l_im = cv2.imread(path1, cv2.IMREAD_GRAYSCALE)\n r_im = cv2.imread(path2, cv2.IMREAD_GRAYSCALE)\n\n d_im = stereo.disparity_ssd(l_im, r_im, W=35)\n d_im = cv2.normalize(d_im, d_im, alpha=0, beta=255, norm_type=cv2.NORM_MINMAX)\n\n cv2.imshow(\"depth\", d_im * 4)\n cv2.waitKey(0)\n\n\ndef run_hough():\n im_path0 = '../data/squares.png'\n im_path1 = '../data/coins.png'\n\n find_circles(im_path1, [20, 25], threshold=10)\n find_lines(im_path1, threshold=80)\n\n\n\npath1 = \"../data/coins.png\"\npath2 = \"../data/check.bmp\"\npath3 = \"../data/simA.jpg\"\nim_color = cv2.imread(path2)\nim = cv2.imread(path2, cv2.IMREAD_GRAYSCALE)\n\n\n\n", "sub_path": "src/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 882, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cv2.imread", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 14, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 15, "usage_type": "attribute"}, {"api_name": "stereo.disparity_ssd", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.normalize", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.NORM_MINMAX", "line_number": 18, "usage_type": "attribute"}, {"api_name": "cv2.imshow", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 36, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 37, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 37, "usage_type": "attribute"}]} +{"seq_id": "356747255", "text": "#encoding=utf-8\nimport pandas as pd\nimport pymysql\n\ndf = pd.read_csv('bookshorts.csv', sep='\\t', encoding='utf-8')\nprint(df)\n\ndf.columns = ['star', 'short', 'sentiment']\n\nstar_to_number = {\n '力荐' : 5,\n '推荐' : 4,\n '还行' : 3,\n '较差' : 2,\n '很差' : 1\n}\n\ndf['new_star'] = df['star'].map(star_to_number)\n\n\n# 存入mysql\nconn = pymysql.connect(host = 'localhost',\n port = 3306,\n user = 'mysql',\n password = 'Pwd_2020',\n database = 'testdb',\n charset = 'utf8mb4')\n\ncursor = conn.cursor()\n\nstar = df['new_star'].values.tolist()\nshort = df['short'].values.tolist()\nsentiment = df['sentiment'].values.tolist()\n\n# values = [(df['new_star'], df['short'], df['sentiment'])]\nvalues = [i for i in zip(star, short, sentiment)]\ncursor.executemany('insert into bookshorts(star, short, sentiment) values (%s, %s, %s)', values)\n\ncursor.close()\nconn.commit()\nconn.close()\n", "sub_path": "Week_06/G20190389010004/week06_0004/book/bookshorts_to_mysql.py", "file_name": "bookshorts_to_mysql.py", "file_ext": "py", "file_size_in_byte": 997, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 5, "usage_type": "call"}, {"api_name": "pymysql.connect", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "98599472", "text": "import sys\nsys.path.append('../Resolution') #permet les import provenant d'autres dossiers\nfrom resolution_optimisée import *\nfrom résolution_sous_optimal import *\nsys.path.append('..')\nfrom generation import *\nfrom base_de_grilles import *\nimport random\nfrom time import time\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n#pour rappel : \n# - la fonction resolve correspond à un résolution récursives assez brutales avec hypothèse systématique sur la valeur des cases\n# - la fonction resolution correspond à une résolution basée sur la méthode humaine de résolution de sudoku\n\n\n# Le but de se fichier est de tracer un graphique représentant l'évolution du temps de résolution \n# en fonction du nombre de cases remplies avant la résolution (pour les deux fonctions)\n\n\n\n\ndef grille_aleatoire(nb_cases_non_vides):\n \"\"\"renvoie une grille générée aléatoirement avec nb_cases_non_vides\"\"\"\n choix_aleatoire = randint(0,len(liste_grilles)-1)\n grille = liste_grilles[choix_aleatoire] #on utilise le fichier base_de_grilles.py\n nb_cases_vides = 81 - nb_cases_non_vides #qui contient des grilles valides déja enregistrées\n compteur=0\n while compteur < nb_cases_vides:\n i = randint(0,8)\n j = randint(0,8)\n if grille[i][j] != 0:\n grille[i][j]=0\n compteur+=1\n return(grille)\n\ndef temps_necessaire(nb_cases_non_vides):\n \"\"\"renvoie le temps necessaire à la résolution d'une grille aleatoire avec nb cases non vide par les deux méthodes de résolution\"\"\"\n temps_fct_resolve=[] \n temps_fct_resolution=[] \n grille = grille_aleatoire(nb_cases_non_vides) \n \n temps_initial_1=time() #utilisation du module time\n resolve(grille) #pour connaitre le temps necessaire à la résolution\n temps_final_1=time()\n temps_fct_resolve = temps_final_1 - temps_initial_1\n\n grille=int_to_string(grille) #changement de format necessaire pour utiliser la fonction résolution\n temps_initial_2=time()\n resolution(grille)\n temps_final_2=time()\n temps_fct_resolution = temps_final_2 - temps_initial_2\n return((temps_fct_resolve,temps_fct_resolution))\n\n######### enregistrment de nombreux calculs ##########\n# les calculs sont longs pour calculer les temps, on enregistre donc les temps dans deux dictionnaires stockés dans le fichier temps.txt\ndef dic_temps_vide():\n dic={}\n for i in range(23,81):\n dic[i]=[]\n return(dic)\n\ndef remplissage(nb_essais,nb_cases_non_vides):\n \"\"\"remplie les dictionnaires present dans le fichier temps.txt avec les temps de résolution des deux fonctions pour une grille avec nb_cases_non_vides\"\"\"\n with open('temps.txt','r') as fichier:\n dic_temps_resolve = eval(fichier.readline()) #.readline() renvoie une chaine de caractère contenant le dictionnaire\n dic_temps_resolution = eval(fichier.readline())\n for i in range(nb_essais):\n (temps_resolve,temps_resolution)=temps_necessaire(nb_cases_non_vides)\n dic_temps_resolve[nb_cases_non_vides].append(temps_resolve)\n dic_temps_resolution[nb_cases_non_vides].append(temps_resolution)\n with open('temps.txt','w') as fichier : \n fichier.write(str(dic_temps_resolve))\n fichier.write('\\n')\n fichier.write(str(dic_temps_resolution))\n\ndef remplissage_plage(debut,fin):\n \"\"\"rempli le dictionnaire du fichier temps.txt dans la plage [debut,fin] (fin compris)\"\"\"\n for nb_cases_non_vides in range(fin, debut-1, -1):\n remplissage(1,nb_cases_non_vides)\n print(nb_cases_non_vides)\n\n#remplissage_plage(23,80)\n\n\n############## affichage sous forme de graphique ##############\ndef graphique():\n liste_nb_cases_non_vides=[]\n liste_temps_resolve=[]\n liste_temps_resolution=[]\n with open('temps.txt','r') as fichier:\n dic_temps_resolve = eval(fichier.readline()) \n dic_temps_resolution = eval(fichier.readline())\n for k in range(23,81):\n for temps in dic_temps_resolve[k]:\n liste_nb_cases_non_vides.append(k) #création de listes utilisable pour faire un nuage de points avec plt.scatter\n liste_temps_resolve.append(temps)\n for temps in dic_temps_resolution[k]:\n liste_temps_resolution.append(temps)\n plt.subplot(2,1,1)\n plt.scatter(liste_nb_cases_non_vides, liste_temps_resolve, label = \"résolution brutale\", s=2)\n plt.scatter(liste_nb_cases_non_vides, liste_temps_resolution, label = \"résolution 'humaine'\", s=2)\n plt.ylabel('temps (en s)')\n plt.axis([23,80,0,1])\n plt.legend()\n\n plt.subplot(2,1,2)\n plt.scatter(liste_nb_cases_non_vides, liste_temps_resolve, label = \"résolution brutale\", s=2)\n plt.scatter(liste_nb_cases_non_vides, liste_temps_resolution, label = \"résolution 'humaine'\", s=2)\n plt.xlabel('nombre de cases remplies avant la resolution')\n plt.ylabel('temps (en s)')\n plt.axis([23,80,0,0.01])\n\n plt.show()\n\ngraphique()\n\n\n\n", "sub_path": "sudoku/calcul_temps_resolution/difference_temps_resolution.py", "file_name": "difference_temps_resolution.py", "file_ext": "py", "file_size_in_byte": 5098, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.path.append", "line_number": 2, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 2, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 44, "usage_type": "call"}, {"api_name": "time.time", "line_number": 46, "usage_type": "call"}, {"api_name": "time.time", "line_number": 50, "usage_type": "call"}, {"api_name": "time.time", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 104, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 104, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 105, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 105, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 106, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 108, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 109, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 109, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 115, "usage_type": "name"}]} +{"seq_id": "513769493", "text": "from PyQt5.uic import loadUi\nfrom pathlib import Path\n\nfrom PyQt5 import QtWidgets, QtCore\nfrom PyQt5.Qt import *\nfrom PyQt5.uic import loadUi\n\nfrom access.authorization_class.user_module import CL_userModule\nfrom access.promotion_class.Promotion_Add import CheckableComboBox\nfrom data_connection.h1pos import db1\nfrom datetime import datetime\n\n\nclass CL_role(QtWidgets.QDialog):\n dirname = ''\n\n def __init__(self):\n super(CL_role, self).__init__()\n cwd = Path.cwd()\n mod_path = Path(__file__).parent.parent.parent\n self.dirname = mod_path.__str__() + '/presentation/authorization_ui'\n self.conn = db1.connect()\n\n #Todo: method to load ui of copy role\n def FN_LOAD_COPY(self):\n filename = self.dirname + '/copyRole.ui'\n loadUi(filename, self)\n records = self.FN_GET_ROLES_N()\n for row in records:\n self.CMB_roleName.addItems([row[0]])\n self.CMB_roleName1.addItems([row[0]])\n self.BTN_copyRole.clicked.connect(self.FN_COPY_ROLE)\n self.CMB_roleName.currentIndexChanged.connect(self.FN_ASSIGN_ID)\n self.CMB_roleName1.currentIndexChanged.connect(self.FN_ASSIGN_ID)\n self.FN_ASSIGN_ID()\n\n #Todo: method to set text by id\n def FN_ASSIGN_ID(self):\n self.role1 = self.CMB_roleName.currentText()\n self.role2 = self.CMB_roleName1.currentText()\n self.LB_roleID.setText(self.FN_GET_ROLEID_N(self.role1))\n self.LB_roleID2.setText(self.FN_GET_ROLEID_N(self.role2))\n\n #Todo: method to copy role action\n def FN_COPY_ROLE(self):\n newRole = self.LB_roleID2.text()\n if self.role1 == self.role2:\n QtWidgets.QMessageBox.warning(self, \"Error\", \"Please enter 2 different users\")\n else:\n mycursor = self.conn.cursor()\n mycursor1 = self.conn.cursor()\n mycursor2 = self.conn.cursor()\n sql_select_query = \"select ur.FORM_ID ,ur.ACTION_ID \" \\\n \"from SYS_PRIVILEGE ur inner join SYS_ROLE u ON u.ROLE_ID = ur.ROLE_ID \" \\\n \"where u.ROLE_NAME = %s \"\n x = (self.role1,)\n mycursor.execute(sql_select_query, x)\n records = mycursor.fetchall()\n mycursor2 = self.conn.cursor()\n sql_select_query1 = \"delete from SYS_PRIVILEGE where ROLE_ID = '\" + newRole + \"'\"\n mycursor2.execute(sql_select_query1)\n db1.connectionCommit(self.conn)\n mycursor1.execute(\"SELECT max(cast(PRIV_ID AS UNSIGNED)) FROM SYS_PRIVILEGE\")\n myresult = mycursor1.fetchone()\n id = int(myresult[0]) + 1\n for row in records:\n mycursor3 = self.conn.cursor()\n sql = \"INSERT INTO SYS_PRIVILEGE VALUES ( %s, %s, %s, %s)\"\n val = (id, newRole, row[0], row[1])\n mycursor3.execute(sql, val)\n db1.connectionCommit(self.conn)\n print(mycursor3.rowcount, \"record inserted.\")\n id = id + 1\n mycursor2.close()\n mycursor1.close()\n mycursor.close()\n self.close()\n\n #Todo: method to get all role name and id\n def FN_GET_ROLES_N(self):\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT ROLE_NAME ROLE_ID FROM SYS_ROLE order by ROLE_ID asc\")\n records = mycursor.fetchall()\n mycursor.close()\n return records\n\n #Todo: method to get role id\n def FN_GET_ROLEID_N(self, role):\n mycursor = self.conn.cursor()\n sql_select_query = \"SELECT ROLE_ID FROM SYS_ROLE WHERE ROLE_NAME = %s \"\n x = (role,)\n mycursor.execute(sql_select_query, x)\n myresult = mycursor.fetchone()\n return myresult[0]\n\n #Todo: method to load ui of assignUserToRole\n def FN_ASSIGN(self):\n self.CMB_roleName = CheckableComboBox(self)\n self.CMB_roleName.setGeometry(100, 85, 179, 18)\n self.CMB_roleName.setLayoutDirection(QtCore.Qt.LeftToRight)\n self.CMB_roleName.setStyleSheet(\"background-color: rgb(198, 207, 199)\")\n filename = self.dirname + '/assignUserToRole.ui'\n loadUi(filename, self)\n self.BTN_assignRole.clicked.connect(self.FN_ASSIGN_ROLE)\n self.CMB_userRoleStatus.addItems([\"Active\", \"Inactive\"])\n self.FN_GET_USERS()\n self.FN_GET_USERID()\n self.FN_GET_ROLES()\n self.CMB_userName.currentIndexChanged.connect(self.FN_GET_USERID)\n\n #Todo: method to load ui of modify role\n def FN_LOAD_MODIFY(self):\n filename = self.dirname + '/modifyRole.ui'\n loadUi(filename, self)\n self.CMB_roleStatus.addItems([\"Active\", \"Inactive\"])\n self.FN_GET_ROLES1()\n self.FN_GET_ROLEID()\n self.FN_GET_ROLE()\n self.CMB_roleName.currentIndexChanged.connect(self.FN_GET_ROLE)\n self.BTN_modifyRole.clicked.connect(self.FN_MODIFY_ROLE)\n\n #Todo: method to load ui of create role\n def FN_LOAD_CREATE(self):\n filename = self.dirname + '/createRole.ui'\n loadUi(filename, self)\n self.BTN_createRole.clicked.connect(self.FN_CREATE_ROLE)\n self.CMB_roleStatus.addItems([\"Active\", \"Inactive\"])\n\n #Todo: method to get user id\n def FN_GET_USERID(self):\n self.user = self.CMB_userName.currentText()\n mycursor = self.conn.cursor()\n sql_select_query = \"SELECT USER_ID FROM SYS_USER WHERE USER_NAME = %s\"\n x = (self.user,)\n mycursor.execute(sql_select_query, x)\n myresult = mycursor.fetchone()\n self.LB_userID.setText(myresult[0])\n self.FN_GET_ROLES()\n\n #Todo: method to get role id\n def FN_GET_ROLEID1(self, roleNm):\n if roleNm is not None:\n self.role = roleNm\n else:\n self.role = self.CMB_roleName.currentText()\n mycursor = self.conn.cursor()\n sql_select_query = \"SELECT ROLE_ID FROM SYS_ROLE WHERE ROLE_NAME = %s\"\n x = (self.role,)\n mycursor.execute(sql_select_query, x)\n myresult = mycursor.fetchone()\n self.LB_roleID.setText(myresult[0])\n mycursor.close()\n return myresult[0]\n\n #Todo: method to get role id\n def FN_GET_ROLEID(self):\n self.role = self.CMB_roleName.currentText()\n mycursor = self.conn.cursor()\n sql_select_query = \"SELECT ROLE_ID FROM SYS_ROLE WHERE ROLE_NAME = %s\"\n x = (self.role,)\n mycursor.execute(sql_select_query, x)\n myresult = mycursor.fetchone()\n self.LB_roleID.setText(myresult[0])\n mycursor.close()\n return myresult[0]\n\n #Todo: method to get all users name\n def FN_GET_USERS(self):\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT USER_NAME FROM SYS_USER where USER_STATUS = 1 order by USER_ID asc\")\n records = mycursor.fetchall()\n for row in records:\n self.CMB_userName.addItems([row[0]])\n mycursor.close()\n\n #Todo: method to get role assigned to user and checked it\n def FN_GET_ROLES(self):\n self.CMB_roleName.clear()\n if self.LB_userID is not None:\n selectedRoles = self.FN_SELECT_USER_ROLES()\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT ROLE_NAME FROM SYS_ROLE order by ROLE_ID asc\")\n records = mycursor.fetchall()\n j = 0\n for row in records:\n self.CMB_roleName.addItems(row)\n for row1 in selectedRoles:\n if row[0] == row1[0]:\n items = self.CMB_roleName.findText(row[0])\n for item in range(items+1):\n self.CMB_roleName.setChecked(j)\n j = j + 1\n mycursor.close()\n\n #Todo: method to get all role name\n def FN_GET_ROLES1(self):\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT ROLE_NAME FROM SYS_ROLE order by ROLE_ID asc\")\n records = mycursor.fetchall()\n for row in records:\n self.CMB_roleName.addItems(row)\n mycursor.close()\n\n #Todo: method to get all role assigned to user\n def FN_SELECT_USER_ROLES(self):\n self.user = self.LB_userID.text()\n mycursor = self.conn.cursor()\n sql_select_query = \"SELECT ROLE_NAME FROM SYS_USER_ROLE INNER JOIN SYS_ROLE on SYS_ROLE.ROLE_ID= SYS_USER_ROLE.ROLE_ID where SYS_USER_ROLE.USER_ID= %s \"\n x = (self.user,)\n mycursor.execute(sql_select_query, x)\n records = mycursor.fetchall()\n return records\n\n #Todo: method to assign role to user\n def FN_ASSIGN_ROLE(self):\n self.status = self.CMB_userRoleStatus.currentText()\n self.user = self.LB_userID.text()\n self.role = self.LB_roleID.text()\n self.status = self.CMB_userRoleStatus.currentText()\n if self.status == 'Active':\n self.status = 1\n else:\n self.status = 0\n mycursor = self.conn.cursor(buffered=True)\n sql_select_query = \"delete from SYS_USER_ROLE where SYS_USER_ROLE.USER_ID= '\" + self.user + \"'\"\n mycursor.execute(sql_select_query)\n db1.connectionCommit(self.conn)\n items = self.CMB_roleName.currentData()\n x = []\n for i in range(len(items)):\n roleId = self.FN_GET_ROLEID1(str(items[i]))\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT max(cast(UR_USER_ROLE_ID AS UNSIGNED)) FROM SYS_USER_ROLE\")\n myresult = mycursor.fetchone()\n if myresult[0] == None:\n self.id = \"1\"\n else:\n self.id = int(myresult[0]) + 1\n creationDate = str(datetime.today().strftime('%Y-%m-%d-%H:%M-%S'))\n sql = \"INSERT INTO SYS_USER_ROLE (UR_USER_ROLE_ID, USER_ID, ROLE_ID, BRANCH_NO, UR_CREATED_BY, UR_CREATED_ON, UR_CHANGED_BY, UR_CHANGED_ON, UR_STATUS) \" \\\n \"VALUES ( %s, %s, %s, %s,%s, %s,%s,%s,%s)\"\n val = (self.id, self.user, roleId, '1', CL_userModule.user_name, creationDate, '', '', self.status)\n mycursor.execute(sql, val)\n mycursor.close()\n db1.connectionCommit(self.conn)\n print(mycursor.rowcount, \"record inserted.\")\n db1.connectionClose(self.conn)\n self.close()\n QtWidgets.QMessageBox.information(self, \"Success\", \"Role is assigned successfully\")\n\n def FN_GET_ROLE(self):\n self.FN_GET_ROLEID()\n self.name = self.CMB_roleName.currentText()\n mycursor = self.conn.cursor()\n sql_select_query = \"select * from SYS_ROLE where ROLE_NAME = %s \"\n x = (self.name,)\n mycursor.execute(sql_select_query, x)\n record = mycursor.fetchone()\n self.LE_name.setText(record[1])\n self.LE_DESC.setText(record[2])\n if record[7] == '1':\n self.CMB_roleStatus.setCurrentText('Active')\n else:\n self.CMB_roleStatus.setCurrentText('Inactive')\n mycursor.close()\n print(mycursor.rowcount, \"record retrieved.\")\n\n #Todo: method to modify role\n def FN_MODIFY_ROLE(self):\n self.old_name = self.CMB_roleName.currentText()\n self.name = self.LE_name.text().strip()\n self.desc = self.LE_DESC.text().strip()\n self.status = self.CMB_roleStatus.currentText()\n if self.status == 'Active':\n self.status = '1'\n else:\n self.status = '0'\n if self.name == '' or self.desc == '':\n QtWidgets.QMessageBox.warning(self, \"Error\", \"Please all required field\")\n else:\n mycursor = self.conn.cursor()\n changeDate = str(datetime.today().strftime('%Y-%m-%d-%H:%M-%S'))\n sql = \"UPDATE SYS_ROLE set ROLE_NAME= %s , ROLE_DESC= %s , ROLE_CHANGED_ON = %s , ROLE_CHANGED_BY = %s, ROLE_STATUS = %s where ROLE_NAME= %s \"\n val = (self.name, self.desc, changeDate, CL_userModule.user_name, self.status, self.old_name)\n mycursor.execute(sql, val)\n mycursor.close()\n db1.connectionCommit(self.conn)\n print(mycursor.rowcount, \"record Modified.\")\n db1.connectionClose(self)\n self.close()\n QtWidgets.QMessageBox.information(self, \"Success\", \"Role is modified successfully\")\n\n #Todo: method to get create role\n def FN_CREATE_ROLE(self):\n self.name = self.LE_name.text().strip()\n self.desc = self.LE_DESC.text().strip()\n self.status = self.CMB_roleStatus.currentText()\n if self.status == 'Active':\n self.status = '1'\n else:\n self.status = '0'\n if self.name == '' or self.desc == '':\n QtWidgets.QMessageBox.warning(self, \"Error\", \"Please all required field\")\n else:\n mycursor = self.conn.cursor()\n mycursor.execute(\"SELECT max(cast(role_ID AS UNSIGNED)) FROM SYS_ROLE\")\n myresult = mycursor.fetchone()\n if myresult[0] == None:\n self.id = \"1\"\n else:\n self.id = int(myresult[0]) + 1\n creationDate = str(datetime.today().strftime('%Y-%m-%d-%H:%M-%S'))\n sql = \"INSERT INTO SYS_ROLE (ROLE_ID, ROLE_NAME,ROLE_DESC,ROLE_CREATED_BY,ROLE_CREATED_ON, ROLE_STATUS) \" \\\n \"VALUES ('\" + str(\n self.id) + \"','\" + self.name + \"','\" + self.desc + \"', '\" + CL_userModule.user_name + \"', '\" + creationDate + \"','\" + self.status + \"')\"\n print(sql)\n mycursor.execute(sql)\n mycursor.close()\n db1.connectionCommit(self.conn)\n print(mycursor.rowcount, \"record inserted.\")\n db1.connectionClose(self.conn)\n self.close()\n QtWidgets.QMessageBox.information(self, \"Success\", \"Role is created successfully\")\n", "sub_path": "access/authorization_class/Old/Role.py", "file_name": "Role.py", "file_ext": "py", "file_size_in_byte": 13711, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 14, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 14, "usage_type": "name"}, {"api_name": "pathlib.Path.cwd", "line_number": 19, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 19, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 20, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1.connect", "line_number": 22, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 22, "usage_type": "name"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 27, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 48, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 48, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 48, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 62, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 62, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 71, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 71, "usage_type": "name"}, {"api_name": "access.promotion_class.Promotion_Add.CheckableComboBox", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 100, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 100, "usage_type": "name"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 103, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 114, "usage_type": "call"}, {"api_name": "PyQt5.uic.loadUi", "line_number": 125, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 227, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 227, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 239, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 239, "usage_type": "name"}, {"api_name": "access.authorization_class.user_module.CL_userModule.user_name", "line_number": 242, "usage_type": "attribute"}, {"api_name": "access.authorization_class.user_module.CL_userModule", "line_number": 242, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 245, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 245, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionClose", "line_number": 247, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 247, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 249, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 249, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 249, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 279, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 279, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 279, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 282, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 282, "usage_type": "name"}, {"api_name": "access.authorization_class.user_module.CL_userModule.user_name", "line_number": 284, "usage_type": "attribute"}, {"api_name": "access.authorization_class.user_module.CL_userModule", "line_number": 284, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 287, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 287, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionClose", "line_number": 289, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 289, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 291, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 291, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 291, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.warning", "line_number": 303, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 303, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 303, "usage_type": "name"}, {"api_name": "datetime.datetime.today", "line_number": 312, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 312, "usage_type": "name"}, {"api_name": "access.authorization_class.user_module.CL_userModule.user_name", "line_number": 315, "usage_type": "attribute"}, {"api_name": "access.authorization_class.user_module.CL_userModule", "line_number": 315, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionCommit", "line_number": 319, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 319, "usage_type": "name"}, {"api_name": "data_connection.h1pos.db1.connectionClose", "line_number": 321, "usage_type": "call"}, {"api_name": "data_connection.h1pos.db1", "line_number": 321, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.information", "line_number": 323, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 323, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 323, "usage_type": "name"}]} +{"seq_id": "406153871", "text": "import csv\nimport os\nfrom datetime import datetime\n\n\nif not os.path.exists('stock-list'):\n os.mkdir('stock-list')\n\n\ndates = list()\n\nfor root, dirs, files in os.walk('krx-stock'):\n for f in files:\n filename, ext = os.path.splitext(f)\n if ext == '.csv':\n dates.append(datetime.fromisoformat(filename))\n\n\n\ndates.sort()\nlatest_date = dates[-1]\n\n\n\nstock_list = list()\nwith open('krx-stock/%s.csv' % latest_date.isoformat()) as f:\n stockreader = csv.reader(f)\n for i, row in enumerate(stockreader):\n if i > 0:\n stock_list.append('%s, %06d' % (row[1], int(row[2])))\n\n\nwith open('stock-list/%s.csv' % latest_date.isoformat(), 'w') as f:\n f.write('\\n'.join(stock_list))\n", "sub_path": "stock-list.py", "file_name": "stock-list.py", "file_ext": "py", "file_size_in_byte": 719, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.exists", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 7, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromisoformat", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "csv.reader", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "265746722", "text": "from tools.geiger.TimestampHistogram import TimestampHistogram\nimport time\nimport random\n\nh = TimestampHistogram(sample_length=500, bin_size=2)\n\nh.set_start_now()\ns = time.time() * 1000\nt = 0\nprint(\"Generating random data\")\nwhile t < 90000:\n if random.random() > 0.9:\n h.plot(s + t)\n t += 1\n\nh.group_samples()\nh.display_histogram_and_fit_curve()", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 358, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "tools.geiger.TimestampHistogram.TimestampHistogram", "line_number": 5, "usage_type": "call"}, {"api_name": "time.time", "line_number": 8, "usage_type": "call"}, {"api_name": "random.random", "line_number": 12, "usage_type": "call"}]} +{"seq_id": "395719902", "text": "from flask import Flask, render_template, request, redirect, url_for, session\nfrom rauth.service import OAuth1Service, OAuth2Service\nfrom uuid import uuid1 as uuid\nimport requests\nimport urlparse\nimport json\n\nfrom webserver import app\nfrom webserver import db\nfrom webserver import twitter\nfrom webserver import facebook\n\nbaseurl = 'http://461.ansel.co'\n\n@app.route('/student/facebook', methods=['POST'])\ndef facebookCreateAccount():\n # Facebook auth is being handled on frontend and information passed to us directly\n\n values = request.json\n\n if 'access_token' not in values:\n return 'missing access_token'\n if 'name' not in values:\n return 'missing name'\n if 'facebook_id' not in values:\n return 'missing facebook_id'\n\n # Get a long lasting token\n data = {\n 'grant_type': 'fb_exchange_token',\n 'client_id': facebook['client_id'],\n 'client_secret': facebook['client_secret'],\n 'fb_exchange_token': values['access_token']\n }\n\n response = requests.get('https://graph.facebook.com/oauth/access_token', params=data)\n returnParams = urlparse.parse_qs(response.text)\n values['access_token'] = returnParams['access_token'][0]\n\n student_signin('facebook', values=values)\n return 'success'\n\n\n@app.route('/student/facebook/deauthorize')\ndef deauthFacebook():\n return 'Thanks for the heads up...'\n\n\n@app.route('//twitter')\ndef twitter_redirect(user_type):\n if user_type != 'psychic' and user_type != 'student':\n return 'Unknown user type'\n\n session['request_token'], session['request_token_secret'] = twitter.get_request_token(method='GET', oauth_callback='%s/%s/twitter/authorized' % (baseurl, user_type))\n authorize_url = twitter.get_authorize_url(session['request_token'])\n return redirect(authorize_url)\n\n\n@app.route('//twitter/authorized')\ndef twitter_authorized(user_type):\n token = request.args.get('oauth_token', None)\n verifier = request.args.get('oauth_verifier', None)\n\n if session.get('request_token') and (token and verifier):\n resp = twitter.get_access_token(\n 'POST',\n request_token=session.get('request_token'),\n request_token_secret=session.get('request_token_secret'), data={'oauth_verifier': verifier}\n )\n\n # Remove the request tokens from the session, they are no longer needed\n session.pop('request_token_secret', None)\n session.pop('request_token', None)\n\n if user_type == 'psychic':\n token, guid = p_signin_twitter(resp.content['oauth_token'], resp.content['oauth_token_secret'])\n return redirect('http://461.ansel.co/dashboard/%s?token=%s' % (guid, token))\n elif user_type == 'student':\n student_signin('twitter', resp.content['oauth_token'], resp.content['oauth_token_secret'])\n return render_template('studentSuccess.html')\n else:\n return 'Unknown user type'\n else:\n return redirect('/%s/twitter' % user_type)\n\ndef student_signin(service, access_token=None, access_token_secret=None, **kwarg):\n if service == 'twitter':\n response = twitter.get(\n 'https://api.twitter.com/1/account/verify_credentials.json',\n access_token=access_token,\n access_token_secret=access_token_secret\n )\n\n data = response.content\n handle = data['screen_name']\n serviceId = data['id_str']\n name = data['name']\n email = ''\n else:\n serviceId = kwarg['values']['facebook_id']\n handle = None\n access_token = kwarg['values']['access_token']\n name = kwarg['values']['name']\n email = kwarg['values']['email']\n\n # Determine if this user already exists in our database\n rows = db.query('SELECT * FROM student_client WHERE service = %s AND serviceId = %s', (service, serviceId))\n\n if len(rows) is 0:\n # New student, create account\n guid = str(uuid())\n db.query(\n 'INSERT INTO student_client (guid, name, email, service, handle, serviceId, oauthToken, oauthSecret) VALUES (%s, %s, %s, %s, %s, %s, %s, %s)',\n (guid, name, email, service, handle, serviceId, access_token, access_token_secret)\n )\n\n\ndef p_signin_twitter(access_token, access_token_secret):\n # Get some general information about this user\n response = twitter.get(\n 'https://api.twitter.com/1/account/verify_credentials.json',\n access_token=access_token,\n access_token_secret=access_token_secret\n )\n\n data = response.content\n name = data['name']\n twitter_id = data['id_str']\n twitter_username = data['screen_name']\n guid = ''\n\n # Determine if this user already exists in our database\n rows = db.query('SELECT * FROM psychic WHERE twitterId = %s', twitter_id)\n\n if len(rows) is 0:\n # New user, create account\n guid = str(uuid())\n db.query(\n 'INSERT INTO psychic (guid, fullName, twitterUsername, twitterId, twitterAccessToken, twitterAccessTokenSecret) VALUES (%s, %s, %s, %s, %s, %s)',\n (guid, name, twitter_username, twitter_id, access_token, access_token_secret)\n )\n else:\n # Existing user, set the guid\n guid = rows[0]['guid']\n\n # Set info so we know who is signed in\n session['access_token'] = access_token\n session['access_token_secret'] = access_token_secret\n session['psychic_guid'] = guid\n\n # Issue the user a token\n token = str(uuid())\n db.query('INSERT INTO access_token (guid, psychicGuid) VALUES (%s, %s)', (token, guid))\n\n return token, guid\n", "sub_path": "webserver/webserver/oauth.py", "file_name": "oauth.py", "file_ext": "py", "file_size_in_byte": 5617, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.request.json", "line_number": 19, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 19, "usage_type": "name"}, {"api_name": "webserver.facebook", "line_number": 31, "usage_type": "name"}, {"api_name": "webserver.facebook", "line_number": 32, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 36, "usage_type": "call"}, {"api_name": "urlparse.parse_qs", "line_number": 37, "usage_type": "call"}, {"api_name": "webserver.app.route", "line_number": 15, "usage_type": "call"}, {"api_name": "webserver.app", "line_number": 15, "usage_type": "name"}, {"api_name": "webserver.app.route", "line_number": 44, "usage_type": "call"}, {"api_name": "webserver.app", "line_number": 44, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 54, "usage_type": "name"}, {"api_name": "webserver.twitter.get_request_token", "line_number": 54, "usage_type": "call"}, {"api_name": "webserver.twitter", "line_number": 54, "usage_type": "name"}, {"api_name": "webserver.twitter.get_authorize_url", "line_number": 55, "usage_type": "call"}, {"api_name": "webserver.twitter", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 56, "usage_type": "call"}, {"api_name": "webserver.app.route", "line_number": 49, "usage_type": "call"}, {"api_name": "webserver.app", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 61, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 61, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 61, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 62, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 62, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 64, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 64, "usage_type": "name"}, {"api_name": "webserver.twitter.get_access_token", "line_number": 65, "usage_type": "call"}, {"api_name": "webserver.twitter", "line_number": 65, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 67, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 67, "usage_type": "name"}, {"api_name": "flask.session.get", "line_number": 68, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 68, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 72, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 72, "usage_type": "name"}, {"api_name": "flask.session.pop", "line_number": 73, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 73, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 84, "usage_type": "call"}, {"api_name": "webserver.app.route", "line_number": 59, "usage_type": "call"}, {"api_name": "webserver.app", "line_number": 59, "usage_type": "name"}, {"api_name": "webserver.twitter.get", "line_number": 88, "usage_type": "call"}, {"api_name": "webserver.twitter", "line_number": 88, "usage_type": "name"}, {"api_name": "webserver.db.query", "line_number": 107, "usage_type": "call"}, {"api_name": "webserver.db", "line_number": 107, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 111, "usage_type": "call"}, {"api_name": "webserver.db.query", "line_number": 112, "usage_type": "call"}, {"api_name": "webserver.db", "line_number": 112, "usage_type": "name"}, {"api_name": "webserver.twitter.get", "line_number": 120, "usage_type": "call"}, {"api_name": "webserver.twitter", "line_number": 120, "usage_type": "name"}, {"api_name": "webserver.db.query", "line_number": 133, "usage_type": "call"}, {"api_name": "webserver.db", "line_number": 133, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 137, "usage_type": "call"}, {"api_name": "webserver.db.query", "line_number": 138, "usage_type": "call"}, {"api_name": "webserver.db", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 147, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 148, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 149, "usage_type": "name"}, {"api_name": "uuid.uuid1", "line_number": 152, "usage_type": "call"}, {"api_name": "webserver.db.query", "line_number": 153, "usage_type": "call"}, {"api_name": "webserver.db", "line_number": 153, "usage_type": "name"}]} +{"seq_id": "179897930", "text": "# coding=utf-8\n# pylint: disable=too-many-lines\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is regenerated.\n# --------------------------------------------------------------------------\n\nfrom typing import Any, List, Optional, TYPE_CHECKING\n\nfrom ... import _serialization\n\nif TYPE_CHECKING:\n # pylint: disable=unused-import,ungrouped-imports\n from .. import models as _models\n\n\nclass Error(_serialization.Model):\n \"\"\"Error.\n\n :ivar status:\n :vartype status: int\n :ivar message:\n :vartype message: str\n \"\"\"\n\n _attribute_map = {\n \"status\": {\"key\": \"status\", \"type\": \"int\"},\n \"message\": {\"key\": \"message\", \"type\": \"str\"},\n }\n\n def __init__(self, *, status: Optional[int] = None, message: Optional[str] = None, **kwargs: Any) -> None:\n \"\"\"\n :keyword status:\n :paramtype status: int\n :keyword message:\n :paramtype message: str\n \"\"\"\n super().__init__(**kwargs)\n self.status = status\n self.message = message\n\n\nclass PagingResult(_serialization.Model):\n \"\"\"PagingResult.\n\n :ivar values:\n :vartype values: list[~multiapicombiner.v1.models.Product]\n :ivar next_link:\n :vartype next_link: str\n \"\"\"\n\n _attribute_map = {\n \"values\": {\"key\": \"values\", \"type\": \"[Product]\"},\n \"next_link\": {\"key\": \"nextLink\", \"type\": \"str\"},\n }\n\n def __init__(\n self, *, values: Optional[List[\"_models.Product\"]] = None, next_link: Optional[str] = None, **kwargs: Any\n ) -> None:\n \"\"\"\n :keyword values:\n :paramtype values: list[~multiapicombiner.v1.models.Product]\n :keyword next_link:\n :paramtype next_link: str\n \"\"\"\n super().__init__(**kwargs)\n self.values = values\n self.next_link = next_link\n\n\nclass Product(_serialization.Model):\n \"\"\"Product.\n\n :ivar id:\n :vartype id: int\n \"\"\"\n\n _attribute_map = {\n \"id\": {\"key\": \"id\", \"type\": \"int\"},\n }\n\n def __init__(self, *, id: Optional[int] = None, **kwargs: Any) -> None: # pylint: disable=redefined-builtin\n \"\"\"\n :keyword id:\n :paramtype id: int\n \"\"\"\n super().__init__(**kwargs)\n self.id = id\n\n\nclass TestLroAndPagingOptions(_serialization.Model):\n \"\"\"Parameter group.\n\n :ivar maxresults: Sets the maximum number of items to return in the response.\n :vartype maxresults: int\n :ivar timeout: Sets the maximum time that the server can spend processing the request, in\n seconds. The default is 30 seconds.\n :vartype timeout: int\n \"\"\"\n\n _attribute_map = {\n \"maxresults\": {\"key\": \"maxresults\", \"type\": \"int\"},\n \"timeout\": {\"key\": \"timeout\", \"type\": \"int\"},\n }\n\n def __init__(self, *, maxresults: Optional[int] = None, timeout: int = 30, **kwargs: Any) -> None:\n \"\"\"\n :keyword maxresults: Sets the maximum number of items to return in the response.\n :paramtype maxresults: int\n :keyword timeout: Sets the maximum time that the server can spend processing the request, in\n seconds. The default is 30 seconds.\n :paramtype timeout: int\n \"\"\"\n super().__init__(**kwargs)\n self.maxresults = maxresults\n self.timeout = timeout\n", "sub_path": "packages/autorest.python/test/multiapi/Expected/AcceptanceTests/multiapicombiner/multiapicombiner/v1/models/_models_py3.py", "file_name": "_models_py3.py", "file_ext": "py", "file_size_in_byte": 3554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "typing.TYPE_CHECKING", "line_number": 14, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 84, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 84, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 108, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 108, "usage_type": "name"}]} +{"seq_id": "532862634", "text": "import logging\nimport logging.config\nimport yaml\nimport os\n\n\ndef setup_logging(config_path='logging.yaml', level=logging.INFO, log_path=None):\n if os.path.exists(config_path):\n with open(config_path) as f:\n config = yaml.safe_load(f.read())\n\n if log_path:\n config['handlers']['logfile']['filename'] = log_path\n\n logging.config.dictConfig(config)\n else:\n logging.basicConfig(level=level)\n logging.warning('Path {} does not exist'.format(config_path))\n", "sub_path": "logging/python/logging_utilities.py", "file_name": "logging_utilities.py", "file_ext": "py", "file_size_in_byte": 513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.INFO", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.config.dictConfig", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.config", "line_number": 15, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 17, "usage_type": "call"}, {"api_name": "logging.warning", "line_number": 18, "usage_type": "call"}]} +{"seq_id": "392894972", "text": "# Checks if a date is valid\n# WIP\n\nimport re\nfrom datetime import datetime\n\n\ndef validate_date(date):\n check_date = re.compile(r\"(\\d\\d)/([0-1]\\d)/([1-2]\\d\\d\\d)\") # detect DD/MM/YYYY format\n day = int(check_date.search(date).group(1)) # set variables\n month = int(check_date.search(date).group(2))\n year = int(check_date.search(date).group(3))\n\n if month == 4 and day == 31:\n return \"False: too many days in the month!\"\n if month == 6 and day == 31:\n return \"False: too many days in the month!\"\n if month == 9 and day == 31:\n return \"False: too many days in the month!\"\n if month == 11 and day == 31:\n return \"False: too many days in the month!\"\n\n if month == 2:\n if day > 28 and year % 4 != 0: # the year is not leap year\n return \"False, february can't have more than 29 days in non-leap years!\"\n elif day > 28 and year % 4 == 0 and year % 100 == 0:\n return \"False: february can't have more than 28 days in leap years divisible by 100\"\n elif day > 29 and year % 4 == 0 and year % 100 == 0 and year % 400 == 0:\n return \"False: february can't have more than 29 days in leap years divisible by 100 and 400\"\n elif day > 29 and year % 4 == 0: # it is leap year\n return \"False, february can't have more than 29 days in leap years!\"\n\n return True\n\n\n# tests\nprint(validate_date('21/06/2061')) # True\nprint(validate_date('31/11/2019')) # false, no 31st of november\n\nprint(validate_date('29/02/2004')) # True, february can have 29 days in leap years\nprint(validate_date('29/02/2100')) # False, february can't have more than 28 days in leap years divisible by 100\nprint(validate_date('29/02/2000')) # True, february can have 29 days in leap years divisible by 100 and 400\n\nprint(\"\"\"\nsuperior version:\"\"\")\n\n\ndef validate(date_text):\n try:\n if date_text != datetime.strptime(date_text, \"%Y-%m-%d\").strftime('%Y-%m-%d'):\n raise ValueError\n return True\n except ValueError:\n return False\n\n\nprint(validate('2061-06-21'))\nprint(validate('2019-11-31'))\nprint(validate('2004-02-29'))\nprint(validate('2100-02-29'))\nprint(validate('2000-02-29'))\n", "sub_path": "ATBS 2nd edition/7 Date Detection.py", "file_name": "7 Date Detection.py", "file_ext": "py", "file_size_in_byte": 2201, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "re.compile", "line_number": 9, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 50, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 50, "usage_type": "name"}]} +{"seq_id": "227641334", "text": "from discord.ext import commands\nimport discord\nimport socket\nfrom pyraklib.protocol.EncapsulatedPacket import EncapsulatedPacket\nfrom pyraklib.protocol.UNCONNECTED_PING import UNCONNECTED_PING\nfrom pyraklib.protocol.UNCONNECTED_PONG import UNCONNECTED_PONG\nfrom mcstatus import MinecraftServer\nimport asyncio\n\nclass Xenon(commands.Cog):\n def __init__(self, bot):\n self.bot = bot\n \n @commands.Cog.listener()\n async def on_message(self, message):\n if \"discord.gg\" in message.content:\n if message.channel.guild.id == 593954944059572235:\n if message.channel.id in [593954944059572237, 593962666695983106, 609104954224803844, 594014175265685532, 659542937285165056, 636324104055816193]:\n try:\n await message.delete()\n await message.channel.send(embed=discord.Embed(color=discord.Color.green(), description=\"Discord server links aren't allowed here!\"))\n except Exception:\n pass\n \n @commands.command(name=\"mcstatus\")\n async def mcstatus(self, ctx):\n await ctx.trigger_typing()\n ping = UNCONNECTED_PING()\n ping.pingID = 4201\n ping.encode()\n s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n s.setblocking(0)\n try:\n s.sendto(ping.buffer, (\"172.10.17.177\", 19132))\n await asyncio.sleep(.01)\n recvData = s.recvfrom(2048)\n except BlockingIOError:\n await ctx.send(embed=discord.Embed(color=discord.Color.green(), description=\"Xenon BE is either offline or unavailable at the moment. Did you type the ip correctly?\"))\n return\n except socket.gaierror:\n await ctx.send(embed=discord.Embed(color=discord.Color.green(), description=\"Xenon BE is either offline or unavailable at the moment. Did you type the ip correctly?\"))\n return\n pong = UNCONNECTED_PONG()\n pong.buffer = recvData[0]\n pong.decode()\n sInfo = str(pong.serverName)[2:-2].split(\";\")\n pCount = sInfo[4]\n await ctx.send(embed=discord.Embed(color=discord.Color.green(), description=\"Xenon BE is online with \"+pCount+\" player(s).\"))\n await ctx.trigger_typing()\n status = MinecraftServer.lookup(\"172.10.17.177:25565\")\n try:\n status = status.status()\n await ctx.send(embed=discord.Embed(color=discord.Color.green(), description=\"Xenon JE is online with {0} player(s) and a ping of {1} ms.\".format(status.players.online, status.latency)))\n except Exception:\n await ctx.send(embed=discord.Embed(color=discord.Color.green(), description=\"Xenon JE is either offline or unavailable at the moment.\"))\n\ndef setup(bot):\n bot.add_cog(Xenon(bot))", "sub_path": "cogs/xenon.py", "file_name": "xenon.py", "file_ext": "py", "file_size_in_byte": 2800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 10, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 10, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 21, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 21, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 14, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 14, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 14, "usage_type": "name"}, {"api_name": "pyraklib.protocol.UNCONNECTED_PING.UNCONNECTED_PING", "line_number": 28, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 31, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 31, "usage_type": "attribute"}, {"api_name": "socket.SOCK_DGRAM", "line_number": 31, "usage_type": "attribute"}, {"api_name": "asyncio.sleep", "line_number": 35, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 38, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 38, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 38, "usage_type": "attribute"}, {"api_name": "socket.gaierror", "line_number": 40, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 41, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 41, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pyraklib.protocol.UNCONNECTED_PONG.UNCONNECTED_PONG", "line_number": 43, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 48, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 48, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 48, "usage_type": "attribute"}, {"api_name": "mcstatus.MinecraftServer.lookup", "line_number": 50, "usage_type": "call"}, {"api_name": "mcstatus.MinecraftServer", "line_number": 50, "usage_type": "name"}, {"api_name": "discord.Embed", "line_number": 53, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 53, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 53, "usage_type": "attribute"}, {"api_name": "discord.Embed", "line_number": 55, "usage_type": "call"}, {"api_name": "discord.Color.green", "line_number": 55, "usage_type": "call"}, {"api_name": "discord.Color", "line_number": 55, "usage_type": "attribute"}, {"api_name": "discord.ext.commands.command", "line_number": 25, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "86786576", "text": "import pandas as pd\nimport pickle\nfrom sklearn.model_selection import train_test_split\nfrom sklearn import tree\n\ndata = pd.read_csv('Covid_dataset.csv')\n\n\ndata_x = data.iloc[:,0:-1].values\ndata_y = data.iloc[:,-1].values\n#print(data_y)\n\nX_train,X_test,y_train,y_test = train_test_split(data_x,data_y,test_size=0.3,random_state=0)\n\nreg = tree.DecisionTreeClassifier()\nreg.fit(data_x,data_y)\n\nprint(\"Train Score:\", reg.score(X_train,y_train))\nprint(\"Test Score:\", reg.score(X_test,y_test))\n\n\npickle.dump(reg, open('covid.pkl','wb'))\n\nmodel = pickle.load(open('covid.pkl','rb'))\nprint(model.predict([[17,99.03,\t0 ,0 ,1\t]]))\n", "sub_path": "model.py", "file_name": "model.py", "file_ext": "py", "file_size_in_byte": 621, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 6, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 13, "usage_type": "call"}, {"api_name": "sklearn.tree.DecisionTreeClassifier", "line_number": 15, "usage_type": "call"}, {"api_name": "sklearn.tree", "line_number": 15, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 22, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "554456536", "text": "# testing cognet vs native Qnet timess\nimport sys\n\nfrom quasinet.qnet import qdistance\nfrom cognet.cognet import cognet as cg\nfrom cognet.dataFormatter import dataFormatter\nfrom cognet.model import model \n#import cognet.util\nimport pandas as pd\nimport numpy as np\n\nfrom quasinet.qnet import Qnet\nimport multiprocessing\nimport time\n\nyr = '2018'\nPOLEFILE='examples_data/polar_vectors.csv'\nQPATH='examples_data/gss_'+yr+'.joblib'\nIMMUTABLE_FILE='examples_data/immutable.csv'\nGSSDATA = 'examples_data/gss_'+yr+'.csv'\n\n# testing dataFormatter\ndata = dataFormatter(samples=GSSDATA)\n# load the sample data\n# have option for test/train split\n# make checks to ensure we will not throw errors at qnet construction \nprint(data.samples[:2])\nfeatures,samples = data.format_samples('train')\nuse_all_samples = True\nif use_all_samples:\n features,samples = data.Qnet_formatter()\n # default trains and tests using half\nprint(samples.shape)\n\n# format data for Qnet training and fitting\ndata.Qnet_formatter()\n\n# set mutable and immutable vars either from list or file\nim_vars_df = pd.read_csv(IMMUTABLE_FILE, names=['vars'])\nim_vars_list = im_vars_df.vars.to_list()\nmutable_vars, immutable_vars = data.mutable_variables(immutable_list=im_vars_list)\nmutable_vars, immutable_vars = data.mutable_variables(IMMUTABLE_FILE=IMMUTABLE_FILE)\n\n# testing model functionality\n# can either input features and samples directly, or infer from data obj\nmodel_ = model()\n# qnet construction parameters, \n# choose to either load or fit qnet from scratch\ntest_model_buildqnet = True\nif test_model_buildqnet:\n start = time.time()\n print(\"fitting\")\n model_.fit(data_obj=data,\n njobs=6)\n end = time.time()\n print(\"cognet time: \", end-start)\n print(\"fitted\")\n model_.export_dot(\"examples_results/tmp_dot_modelclass.dot\",\n generate_trees=True)\n model_.save(\"examples_data/tmp_nodelclass.joblib\")\n #model_.load(\"tmp_nodelclass.joblib\")\nelse:\n model_.load(\"examples_data/gss_2018.joblib\")\n\nfeatures,samples = data.format_samples('train')\ntest_samples = samples[:2]\n\nprint(multiprocessing.cpu_count())\nstart = time.time()\nQnet_ = Qnet(n_jobs=6, feature_names=features)\nprint(\"fitting\")\nstart = time.time()\nprint(\"samples: \", samples)\nQnet_.fit(samples)\nend = time.time()\nprint(\"Native Qnet fitting time: \", end-start)\n", "sub_path": "examples/.ipynb_checkpoints/cognet_example-checkpoint.py", "file_name": "cognet_example-checkpoint.py", "file_ext": "py", "file_size_in_byte": 2382, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cognet.dataFormatter.dataFormatter", "line_number": 23, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 39, "usage_type": "call"}, {"api_name": "cognet.model.model", "line_number": 46, "usage_type": "call"}, {"api_name": "time.time", "line_number": 51, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 68, "usage_type": "call"}, {"api_name": "time.time", "line_number": 69, "usage_type": "call"}, {"api_name": "quasinet.qnet.Qnet", "line_number": 70, "usage_type": "call"}, {"api_name": "time.time", "line_number": 72, "usage_type": "call"}, {"api_name": "time.time", "line_number": 75, "usage_type": "call"}]} +{"seq_id": "280675397", "text": "from os import listdir\nfrom os import path\nfrom collections import defaultdict\nimport pickle\ndef helpy(folder):\n books = listdir(folder)\n res = defaultdict(set)\n for book in books:\n print(book)\n with open(path.join(f'./{folder}/' + book)) as file:\n try:\n content = set(file.read().lower().replace(',',' ').replace('.',' ').split())\n for word in content:\n res[word].add(book)\n except UnicodeDecodeError:\n print('DecodeError')\n pickle_out = open('dicty.pickle','wb')\n pickle.dump(res,pickle_out)\n pickle_out.close()\n \n\ndef findbooks(args):\n pickle_in = open('dicty.pickle','rb')\n res = pickle.load(pickle_in)\n firstset = res[args[0].lower()]\n for ss in args[1:]:\n myset = res[ss.lower()]\n if myset:\n firstset &= myset\n return firstset\nif __name__ == '__main__':\n helpy('books')\n print(findbooks('Конь'))\n", "sub_path": "book_for_tk.py", "file_name": "book_for_tk.py", "file_ext": "py", "file_size_in_byte": 974, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.listdir", "line_number": 6, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 10, "usage_type": "call"}, {"api_name": "os.path", "line_number": 10, "usage_type": "name"}, {"api_name": "pickle.dump", "line_number": 18, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "614134389", "text": "#! /usr/bin/python\n# -*- coding: utf-8 -*-\n\nfrom functools import wraps\n\nfrom hashlib import md5\n\nfrom flask import session, render_template, url_for, request, redirect, abort, flash\n\nfrom peewee import DoesNotExist\n\nfrom werkzeug.exceptions import default_exceptions, BadRequest, HTTPException, NotFound\n\nfrom model import User\n\n# Авторизация\ndef auth_user(user):\n session['logged_in'] = True\n session['user_ID'] = user.user_ID\n session['userLogin'] = user.userLogin\n if session['userLogin'] == 'demo':\n session['demo'] = True\n else:\n session['demo'] = None\n session['userName'] = user.userName\n flash('Вы успешно вошли в систему как %s' % (user.userName), 'success')\n\n# Декоратор проверяет сессию.\n# Если пользователь не вошёл в систему, он перенаправляется на вид login.\ndef login_required(f):\n @wraps(f)\n def inner(*args, **kwargs):\n if not session.get('logged_in'):\n return redirect(\n url_for('login'))\n return f(*args, **kwargs)\n return inner\n\n# Получаем объект, соответствующий запросу или страницу ошибки 404.\n# Используется алиас вызова метода \"GET\" из модели, \n# который получает объект или выдаёт исключение DoesNotExist.\ndef get_object_or_404(model, **kwargs):\n try:\n return model.get(**kwargs)\n except model.DoesNotExist:\n abort(404)\n return render_template('404.jinja.html')\n\n## Обработка ошибки 400\n# def key_error(e):\n# flash('К сожалению, ваш запрос не удалось выполнить. \\\n# Попробуйте сделать это ещё раз, введя все необходимые данные.', 'danger')\n# return render_template(\n# 'index.jinja.html'), 400\n\n## Обработка ошибки 404\n@login_required\ndef page_not_found(error):\n return render_template(\n '404.jinja.html'), 404\n\n## Главная страница\n@login_required\ndef index():\n return render_template('index.jinja.html')\n\n## Вход в систему\ndef login():\n if session.get('logged_in') and session['logged_in'] == True:\n return redirect(\n url_for('index'), \n code=302)\n else:\n if request.method == 'POST' and request.form['login']:\n try:\n userlogin = request.form['userLogin']\n userpassw = md5(request.form['userPassword']).hexdigest()\n user = User.get(userLogin = userlogin, userPassword = userpassw)\n except User.DoesNotExist:\n flash('Имя или пароль введёны неправильно - попробуйте ещё раз', 'danger')\n else:\n auth_user(user)\n return redirect(\n url_for('index'))\n\n return render_template('login.jinja.html')\n\n## Выход из системы\n@login_required\ndef logout():\n session.pop('logged_in', None)\n flash('Вы вышли из системы', 'warning')\n return redirect(\n url_for('index'))\n", "sub_path": "controller/User.py", "file_name": "User.py", "file_ext": "py", "file_size_in_byte": 3297, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.session", "line_number": 18, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 19, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 22, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.session", "line_number": 25, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 33, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 33, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 34, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 35, "usage_type": "call"}, {"api_name": "functools.wraps", "line_number": 31, "usage_type": "call"}, {"api_name": "model.get", "line_number": 44, "usage_type": "call"}, {"api_name": "model.DoesNotExist", "line_number": 45, "usage_type": "attribute"}, {"api_name": "flask.abort", "line_number": 46, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 47, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 59, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 65, "usage_type": "call"}, {"api_name": "flask.session.get", "line_number": 69, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 69, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 70, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 74, "usage_type": "name"}, {"api_name": "flask.request.form", "line_number": 74, "usage_type": "attribute"}, {"api_name": "flask.request.form", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 77, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "model.User.get", "line_number": 78, "usage_type": "call"}, {"api_name": "model.User", "line_number": 78, "usage_type": "name"}, {"api_name": "model.User.DoesNotExist", "line_number": 79, "usage_type": "attribute"}, {"api_name": "model.User", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 80, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 83, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 84, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.session.pop", "line_number": 91, "usage_type": "call"}, {"api_name": "flask.session", "line_number": 91, "usage_type": "name"}, {"api_name": "flask.flash", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 93, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 94, "usage_type": "call"}]} +{"seq_id": "93496326", "text": "from .cv2 import *\nimport mediapipe as mdp\nimport math\nimport numpy as np\nimport time\nimport moviepy\n\n\n\nmdp_drawing = mdp.solutions.drawing_utils\nmdp_pose = mdp.solutions.pose\n\n# In[2]:\n\n\nvid_cap = cv2.VideoCapture('video5.mp4')\nprev_time = 0\nsteps = 0\nstep_left = 0\nstep_right = 0\nstance = None\nL = 0\nH = 0\nL1 = 0\nH1 = 0\n\nwith mdp_pose.Pose(min_detection_confidence= 0.5, min_tracking_confidence=0.5) as pose:\n while vid_cap.isOpened():\n ret, img = vid_cap.read()\n\n if ret:\n col_img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB) # converting BGR2GRAY for the sake of mediapipe\n\n pose_sols = mdp.solutions.pose.Pose().process(col_img) # mediapipe solutions for Pose detection\n landmarks_info = pose_sols.pose_landmarks # to understand the landmarks\n\n if landmarks_info:\n mdp.solutions.drawing_utils.draw_landmarks(img, landmarks_info,\n mdp.solutions.pose.POSE_CONNECTIONS)\n landmarks_info_list = []\n for idx, landmarks in enumerate(landmarks_info.landmark): # identifying\n hgt, wdt, chnl = img.shape # landmarks\n # at the their\n landmarks_info_list.append([idx, int(landmarks.x * wdt), int(landmarks.y * hgt)])\n cv2.circle(img, (int(landmarks.x * wdt), int(landmarks.y * hgt)), 3, (0, 255, 0)) # positions with a circle\n # print(landmarks_info_list)\n\n # lt_shoulder11 = [landmarks_info_list[11][1], landmarks_info_list[11][2]]\n # lt_elbow13 = [landmarks_info_list[13][1], landmarks_info_list[13][2]]\n # lt_wrist15 = [landmarks_info_list[15][1], landmarks_info_list[15][2]]\n\n # lt_hand_angle = abs(round(math.degrees(\n # math.atan2(lt_wrist15[1] - lt_elbow13[1], lt_wrist15[0] - lt_elbow13[0]) - math.atan2(\n # lt_shoulder11[1] - lt_elbow13[1], lt_shoulder11[0] - lt_elbow13[0]))))\n\n # rt_shoulder12 = [landmarks_info_list[12][1], landmarks_info_list[12][2]]\n # rt_elbow14 = [landmarks_info_list[14][1], landmarks_info_list[14][2]]\n # rt_wrist16 = [landmarks_info_list[16][1], landmarks_info_list[16][2]]\n\n # rt_hand_angle = abs(round(math.degrees(\n # math.atan2(rt_wrist16[1] - rt_elbow14[1], rt_wrist16[0] - rt_elbow14[0]) - math.atan2(\n # rt_shoulder12[1] - rt_elbow14[1], rt_shoulder12[0] - rt_elbow14[0]))))\n\n lt_hip23 = [landmarks_info_list[23][1], landmarks_info_list[23][2]]\n lt_knee25 = [landmarks_info_list[25][1], landmarks_info_list[25][2]]\n lt_ankle27 = [landmarks_info_list[27][1], landmarks_info_list[27][2]]\n lt_heel29 = [landmarks_info_list[29][1], landmarks_info_list[29][2]]\n lt_foot_index31 = [landmarks_info_list[31][1], landmarks_info_list[31][2]]\n\n lt_leg_angle = abs(180 - round(math.degrees(\n math.atan2(lt_ankle27[1] - lt_knee25[1], lt_ankle27[0] - lt_knee25[0]) - math.atan2(\n lt_hip23[1] - lt_knee25[1],\n lt_hip23[0] - lt_knee25[\n 0]))))\n\n lt_ankle_angle = abs(360 - round(math.degrees(\n math.atan2(lt_foot_index31[1] - lt_ankle27[1], lt_foot_index31[0] - lt_ankle27[0]) - math.atan2(\n lt_knee25[1] - lt_ankle27[1],\n lt_knee25[0] - lt_ankle27[0]))))\n\n lt_angle = lt_ankle_angle\n\n rt_hip24 = [landmarks_info_list[24][1], landmarks_info_list[24][2]]\n rt_knee26 = [landmarks_info_list[26][1], landmarks_info_list[26][2]]\n rt_ankle28 = [landmarks_info_list[28][1], landmarks_info_list[28][2]]\n rt_heel30 = [landmarks_info_list[30][1], landmarks_info_list[30][2]]\n rt_foot_index32 = [landmarks_info_list[32][1], landmarks_info_list[32][2]]\n\n rt_leg_angle = abs(180 - round(math.degrees(\n math.atan2(rt_ankle28[1] - rt_knee26[1], rt_ankle28[0] - rt_knee26[0]) - math.atan2(\n rt_hip24[1] - rt_knee26[1],\n rt_hip24[0] - rt_knee26[\n 0]))))\n\n if len(landmarks_info_list) != 0:\n cv2.circle(img, (rt_hip24[0], rt_hip24[1]), 5, (255, 0, 0))\n cv2.circle(img, (rt_knee26[0], rt_knee26[1]), 5, (255, 0, 0))\n cv2.circle(img, (rt_ankle28[0], rt_ankle28[1]), 5, (255, 0, 0))\n cv2.circle(img, (rt_heel30[0], rt_heel30[1]), 5, (255, 0, 0))\n cv2.circle(img, (rt_foot_index32[0], rt_foot_index32[1]), 5, (255, 0, 0))\n\n cv2.circle(img, (lt_hip23[0], lt_hip23[1]), 5, (255, 0, 0))\n cv2.circle(img, (lt_knee25[0], lt_knee25[1]), 5, (255, 0, 0))\n cv2.circle(img, (lt_ankle27[0], lt_ankle27[1]), 5, (255, 0, 0))\n cv2.circle(img, (lt_heel29[0], lt_heel29[1]), 5, (255, 0, 0))\n cv2.circle(img, (lt_foot_index31[0], lt_foot_index31[1]), 5, (255, 0, 0))\n\n #cv2.putText(img, str(int(lt_leg_angle)), (lt_knee25[0] - 20, lt_knee25[1] + 20), cv2.FONT_HERSHEY_PLAIN,\n #1.5,\n #(0, 0, 255), 2)\n # cv2.putText(img, str(int(lt_angle)), (lt_ankle27[0] - 20, lt_ankle27[1] + 20), cv2.FONT_HERSHEY_PLAIN,\n # 1.5,\n # (0, 0, 255), 2)\n # cv2.putText(img, str(int(rt_leg_angle)), (rt_knee26[0] - 20, rt_knee26[1] + 20), cv2.FONT_HERSHEY_PLAIN,\n # 1.5,\n # (0, 0, 255), 2)\n\n if (lt_leg_angle <= 15):\n L = 0\n H = 0\n\n if (lt_leg_angle >= 15 and H == 0):\n L = 1\n\n else:\n L = 0\n\n if lt_leg_angle >= 50 and L == 1:\n H = 1\n\n D = [L == 1 and H == 1]\n if True in D:\n step_left += 1\n # print(f'step_count', {step_left})\n # print(L, H)\n\n L = 0\n\n # Right leg angle:\n if (rt_leg_angle <= 10):\n L1 = 0\n H1 = 0\n\n if (rt_leg_angle >= 10 and H1 == 0):\n L1 = 1\n\n else:\n L1 = 0\n\n if rt_leg_angle >= 35 and L1 == 1:\n H1 = 1\n\n # print((rt_leg_angle))\n\n K = [L1 == 1 and H1 == 1]\n if True in K:\n step_right += 1\n # print(f'step_count_right', {step_right})\n # print(L, H)\n\n L1 = 0\n\n if True in K or True in D:\n total_steps = step_left + step_right\n\n print(f'Step no:', {total_steps})\n\n cur_time = time.time()\n fps = 1 / (cur_time - prev_time)\n prev_time = cur_time\n\n #cv2.putText(img, str(int(fps)), (70, 50), cv2.FONT_HERSHEY_PLAIN, 1.5, (255, 0, 0), 2)\n\n cv2.imshow(\"Video Frame\", img)\n\n if cv2.waitKey(10) & 0xFF == ord('q'):\n break\n else:\n break\n\n vid_cap.release()\n# cv2.destroyAllWindows()", "sub_path": "streamlit_app.py", "file_name": "streamlit_app.py", "file_ext": "py", "file_size_in_byte": 6550, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "mediapipe.solutions", "line_number": 10, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions", "line_number": 11, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 32, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2RGB", "line_number": 32, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions.pose.Pose", "line_number": 34, "usage_type": "call"}, {"api_name": "mediapipe.solutions", "line_number": 34, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions.drawing_utils.draw_landmarks", "line_number": 38, "usage_type": "call"}, {"api_name": "mediapipe.solutions", "line_number": 38, "usage_type": "attribute"}, {"api_name": "mediapipe.solutions", "line_number": 39, "usage_type": "attribute"}, {"api_name": "cv2.circle", "line_number": 45, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 70, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 71, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 76, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 77, "usage_type": "call"}, {"api_name": "math.degrees", "line_number": 89, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 90, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 96, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 98, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 99, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 100, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 103, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 104, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 105, "usage_type": "call"}, {"api_name": "cv2.circle", "line_number": 106, "usage_type": "call"}, {"api_name": "time.time", "line_number": 168, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 174, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 176, "usage_type": "call"}]} +{"seq_id": "410171367", "text": "from zipfile import ZipFile\nimport PyPDF2\nimport re\nimport os\nimport csv\nimport requests\nfrom scoreclass import Calc_Score\nimport json\n#We will need to use flask to take the correct zipfile name. \n#The \"files\" list here is basically a container for objects.\nclass PDFReader:\n def extract_resumes(self, zipname):\n with ZipFile(zipname,'r') as zip:\n zip.extractall()\n files = zip.infolist()\n filenames = []\n zipf = zipname[:zipname.find('.')]\n foldername = zipf + \"/\"\n #Check for PDF files\n for i in range(0, len(files)):\n if(files[i].filename == foldername):\n continue\n else:\n if(files[i].filename.endswith(\".pdf\")):\n read_file = files[i]\n filenames.append(read_file.filename)\n else:\n continue\n #create an array of open files for pdf reader to directly access\n openfiles=[]\n for i in range(len(filenames)):\n openfiles.append(open(filenames[i],'rb'))\n return [openfiles,filenames]\n\n def extract_skills(self, sentence, file_name, ID):\n sentence=sentence.lower()\n sentence = re.sub(r',|:|\\(|\\)', \"\", sentence)\n sentence = re.sub(r'/|\\\\', \" \", sentence)\n words=sentence.split(\" \")\n\n start_words=['skill','skills','abilities','languages']\n stop_words=['accomplishments','accomplishment','experience','education','objective','qualifications','qualification','summary',\n 'awards','hobbies','passion','highlights','research','honors','interest','interests','background','history','profile',\n 'link','internships','internship','email']\n skills=[]\n start=0\n for word in words:\n if(start==0):\n if(word in start_words):\n start=1\n else:\n if(word in stop_words):\n start=0\n else:\n if(word not in skills):\n skills.append(word)\n #print(skills)\n #print(\"\\n\")\n \n name=words[0]+\" \"+words[1]\n row=[ID,file_name,name,sentence]\n with open('cv_list.csv','a') as csvFile:\n writer=csv.writer(csvFile)\n \n writer.writerow(row)\n csvFile.close()\n \n return skills\n \n def analyze_resume(self, resume_file, file_name, ID, final_skills):\n scorer=Calc_Score()\n reader = PyPDF2.PdfFileReader(resume_file)\n sentence=' '\n words=[]\n for i in range(reader.numPages):\n text = reader.getPage(i)\n pageText = text.extractText()\n pageText=pageText.replace('\\n','')\n for word in pageText.split():\n if re.findall('\\s|\\xc2|\\xb7|[|]',word): continue\n words.append(word)\n \n sentence=sentence.join(words)\n #print(sentence)\n resume_skills = self.extract_skills(sentence, file_name, ID)\n \n with open(\"skills_data.json\", \"r\") as jsonfile:\n data = json.loads(jsonfile.read())\n jsonfile.close()\n \n #jd = ['python', 'deep_learning','machine_learning']\n #priority_skills = [7, 4, 5]\n analysis_score = scorer.skillscore_update(resume_skills, final_skills, data)\n \n return analysis_score", "sub_path": "compressed_files/pdfreader.py", "file_name": "pdfreader.py", "file_ext": "py", "file_size_in_byte": 3416, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "zipfile.ZipFile", "line_number": 13, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 37, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 38, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 63, "usage_type": "call"}, {"api_name": "scoreclass.Calc_Score", "line_number": 71, "usage_type": "call"}, {"api_name": "PyPDF2.PdfFileReader", "line_number": 72, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 80, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 88, "usage_type": "call"}]} +{"seq_id": "274447497", "text": "#-*- coding : utf-8 -*-\n\n# 从跑出来的调参结果文档中选择好的结果\n\n# 第一步:先调用画图visual.py重新画一遍,因为画图代码可能会有更新\n# 第二步:统计跑出来的结果选择最好的\n\n# Usage:\n# python select_params.py --datadir=../Result/cc_params/ --savepath=../Result/cc_params/result.txt --redraw=True\n# 将所有跑出来的结果放入到文件夹cc_params下,然后运行即可,redraw是指的重新画图并统计结果\n\nimport sys\nimport os\nimport argparse\nfrom params import Params\n\npas = Params()\n\nall_envs = [\"CP\", \"MC\", \"AB\"]\nall_algos = [\"mrtrpo_l1\", \"mrtrpo_l2\", \"mrtrpo_mu\"]\n\n\nif __name__ == '__main__':\n print(\"Usage : python {} --datadir=datadir --savepath=savepath --redraw=False\".format(sys.argv[0]))\n parse = argparse.ArgumentParser()\n parse.add_argument(\"--datadir\", help = \"All data dir\")\n parse.add_argument(\"--savepath\", help = \"Save path\")\n parse.add_argument(\"--redraw\", type = bool, help = \"Re draw pictures\", default = False)\n\n args = parse.parse_args()\n\n datadir = args.datadir\n savepath = args.savepath\n redraw = args.redraw\n\n # datadir下有很多子目录,每一个目录的结果大概为\"0124-1630-CP-just-test-max_kl=0.001-lam=...\"\n sub_dirs = os.listdir(datadir)\n\n if redraw:\n for sub_dir in sub_dirs:\n sub_path = os.path.join(datadir, sub_dir)\n\n sub_data_path = os.path.join(sub_path, \"Data/\")\n sub_figure_path = os.path.join(sub_path, \"Figure/\")\n\n # 重新画图\n cmdstr = \"python visual.py --ResultPath={} --PicPath={}\".format(sub_data_path, sub_figure_path)\n os.system(cmdstr)\n\n fw = open(savepath, \"w\", encoding = \"utf-8\")\n\n # 按照每个环境统计\n for env in all_envs:\n # 这个环境的所有结果目录\n env_dirs = [d for d in sub_dirs if env in d]\n\n # 每个算法统计\n for algo in all_algos:\n all_means = [] # 算法的平均reward\n all_last_means = [] # 最后100步的平均reward\n top2_means = [] # 最好的两个seed的平均reward\n top2_last_means = [] # 最好的两个seed最后100步的平均reward\n\n for env_d in env_dirs:\n fn = os.path.join(datadir, env_d, \"Figure\", \"stats\", \"{}-R0-statistics.txt\".format(algo.upper()))\n\n try:\n with open(fn, \"r\", encoding = \"utf-8\") as fr:\n for line in fr:\n if \"R0\" in line and \"r1\" in line:\n fields = [cell for cell in line.split(\"\\t\") if len(cell) > 0]\n print(fields)\n am = float(fields[3].split(\"+\")[0].strip())\n alm = float(fields[4].split(\"+\")[0].strip())\n t2m = float(fields[5].split(\"+\")[0].strip())\n t2lm = float(fields[6].split(\"+\")[0].strip())\n\n all_means.append((env_d, am))\n all_last_means.append((env_d, alm))\n top2_means.append((env_d, t2m))\n top2_last_means.append((env_d, t2lm))\n except FileNotFoundError:\n print(\"File not found error :\")\n print(\"\\t {}...\".format(fn))\n\n # 排序,得到每个指标最大的对应的参数\n print(\"################### {} #####################\".format(algo + \" \" + env))\n fw.write(\"Env = {}, Algo = {}:\\n\".format(env, algo))\n\n print(\"所有seed平均reward:\")\n for (d, am) in all_means:\n print(\"\\t dir = {}, all_reward_mean = {}\".format(d, am))\n all_means = sorted(all_means, key = lambda x : x[1], reverse = True)\n highest = all_means[0][1]\n highest_params = all_means[0][0]\n print(\"最好的参数:\")\n print(\"Highest reward = {}, best params = {}...\\n\".format(highest, highest_params))\n\n fw.write(\"\\t Evaluation = All seeds mean reward:\\n\")\n fw.write(\"\\t\\t Highest reward = {}, best params = {}\\n\".format(highest, highest_params))\n\n print(\"所有seed最后100条平均reward:\")\n for (d, alm) in all_last_means:\n print(\"\\t dir = {}, all_last_reward_mean = {}\".format(d, alm))\n all_last_means = sorted(all_last_means, key = lambda x : x[1], reverse = True)\n highest = all_last_means[0][1]\n highest_params = all_last_means[0][0]\n print(\"最好的参数:\")\n print(\"Highest reward = {}, best params = {}...\\n\".format(highest, highest_params))\n\n fw.write(\"\\t Evaluation = All seeds last 100 steps mean reward:\\n\")\n fw.write(\"\\t\\t Highest reward = {}, best params = {}\\n\".format(highest, highest_params))\n\n print(\"Top 2 seed平均reward:\")\n for (d, t2m) in top2_means:\n print(\"\\t dir = {}, top2_reward_mean = {}\".format(d, t2m))\n top2_means = sorted(top2_means, key = lambda x : x[1], reverse = True)\n highest = top2_means[0][1]\n highest_params = top2_means[0][0]\n print(\"最好的参数:\")\n print(\"Highest reward = {}, best params = {}...\\n\".format(highest, highest_params))\n\n fw.write(\"\\t Evaluation = Top2 seeds mean reward:\\n\") \n fw.write(\"\\t\\t Highest reward = {}, best params = {}\\n\".format(highest, highest_params))\n\n print(\"Top 2 seed最后100步平均reward:\")\n for (d, t2lm) in top2_last_means:\n print(\"\\t dir = {}, top2_last_reward_mean = {}\".format(d, t2lm))\n top2_last_means = sorted(top2_last_means, key = lambda x : x[1], reverse = True)\n highest = top2_last_means[0][1]\n highest_params = top2_last_means[0][0]\n print(\"最好的参数:\")\n print(\"Highest reward = {}, best params = {}...\\n\".format(highest, highest_params))\n\n fw.write(\"\\t Evaluation = Top2 seeds last 100 steps mean reward:\\n\")\n fw.write(\"\\t\\t Highest reward = {}, best params = {}\\n\".format(highest, highest_params))\n\n print(\"################### {} #####################\\n\".format(algo + \" \" + env))\n\n fw.close()\n\n\n", "sub_path": "select_params.py", "file_name": "select_params.py", "file_ext": "py", "file_size_in_byte": 6415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "params.Params", "line_number": 17, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 24, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 25, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.system", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}]} +{"seq_id": "372720568", "text": "import unittest\n\nimport common\nfrom l0114_flatten_binary_tree_to_linked_list import Solution\n\nclass Test(unittest.TestCase):\n \n def test_solution(self):\n tree = common.create_binary_tree([1, 2, 5, 3, 4, None, 6])\n Solution().flatten(tree)\n self.assertEqual(\n [1, None, 2, None, 3, None, 4, None, 5, None, 6],\n common.convert_binary_tree_to_list(tree))\n", "sub_path": "python3/test_l0114_flatten_binary_tree_to_linked_list.py", "file_name": "test_l0114_flatten_binary_tree_to_linked_list.py", "file_ext": "py", "file_size_in_byte": 401, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 6, "usage_type": "attribute"}, {"api_name": "common.create_binary_tree", "line_number": 9, "usage_type": "call"}, {"api_name": "l0114_flatten_binary_tree_to_linked_list.Solution", "line_number": 10, "usage_type": "call"}, {"api_name": "common.convert_binary_tree_to_list", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "129211526", "text": "import json\nimport sys\nimport glob\nimport os\nimport numpy as np\n\ndef scale_down_prob_score(keypoint):\n keypoint[:, -1] *= .03\n # print(keypoint)\n return keypoint\n\ndef edit_to_coco(keypoint):\n # nose: 0, 0\n # left_eye: 1, 15\n # right_eye: 2, 14\n # left_ear: 3, 16\n # right_ear: 4, 17\n # left_shoulder: 5, 5\n # right_shoulder: 6, 2\n # left_elbow: 7, 6\n # right_elbow: 8, 3\n # left_wrist: 9, 7\n # right_wrist: 10, 4\n # left_hip: 11, 11\n # right_hip: 12, 8\n # left_knee: 13, 12\n # right_knee: 14, 9\n # left_ankle: 15, 13\n # right_ankle: 16, 10\n openpose_to_coco = [0, 15, 14, 16, 17, 5, 2, 6, 3, 7, 4, 11, 8, 12, 9, 13, 10]\n # print(keypoint)\n new_keypoint = np.zeros((17, 3))\n for i in range(17):\n new_keypoint[i] = keypoint[openpose_to_coco[i]]\n # print(\"after\", new_keypoint)\n return new_keypoint\n\ndef save_to_npz(name, output_dir):\n files = glob.glob(name + '*.json')\n\n glob_keypoints = []\n for json_file in files:\n with open(json_file) as keypoint:\n data = json.load(keypoint)\n pose_keypoints_2d = np.asarray(data['people'][0]['pose_keypoints_2d'])\n pose_keypoints_2d = np.reshape(pose_keypoints_2d, (18, 3))\n pose_keypoints_2d = edit_to_coco(pose_keypoints_2d)\n pose_keypoints_2d = scale_down_prob_score(pose_keypoints_2d)\n glob_keypoints.append(pose_keypoints_2d)\n\n print(np.asarray([glob_keypoints]))\n print(np.asarray([glob_keypoints]).shape)\n\n dictionary_keypoints={'S1': {'Directions 1' : np.asarray([glob_keypoints])}}\n metadata = {\n 'layout_name': 'coco',\n 'num_joints': 17,\n 'keypoints_symmetry': [\n [1, 3, 5, 7, 9, 11, 13, 15],\n [2, 4, 6, 8, 10, 12, 14, 16],\n ]\n }\n np.savez(os.path.join(output_dir, \"data_2d_detections_\" + name + \".npz\"), metadata=metadata, positions_2d=dictionary_keypoints)\n\nif __name__ == \"__main__\":\n if len(sys.argv) != 3:\n print(\"python3 save_to_npz.py keypoint_file_prefix output_dir\")\n else:\n save_to_npz(sys.argv[1], sys.argv[2])\n", "sub_path": "save_to_npz.py", "file_name": "save_to_npz.py", "file_ext": "py", "file_size_in_byte": 2123, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 39, "usage_type": "call"}, {"api_name": "json.load", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.savez", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 69, "usage_type": "attribute"}]} +{"seq_id": "355922325", "text": "from django import forms\nfrom .models import Comment\n\nclass EmailPostForm(forms.Form):\n name=forms.CharField(label=\"姓名:\",max_length=30)\n email=forms.EmailField(label=\"电子邮件:\")\n to=forms.EmailField(label=\"发送至:\")\n comments=forms.CharField(label=\"推荐意见\" ,required=False\n ,widget=forms.Textarea)\n\n\nclass CommentForm(forms.ModelForm):\n class Meta:\n model=Comment\n fields=('name','email','body')\n\nclass SearchForm(forms.Form):\n query=forms.CharField()", "sub_path": "Blog/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.forms.Form", "line_number": 4, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 4, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 5, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 5, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 6, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 6, "usage_type": "name"}, {"api_name": "django.forms.EmailField", "line_number": 7, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 7, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 8, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 8, "usage_type": "name"}, {"api_name": "django.forms.Textarea", "line_number": 9, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 9, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 12, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 12, "usage_type": "name"}, {"api_name": "models.Comment", "line_number": 14, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 17, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 17, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 18, "usage_type": "name"}]} +{"seq_id": "626722667", "text": "from discord.ext import commands\nfrom discord import Embed, Member, File\nfrom discord.channel import TextChannel\nfrom discord.utils import escape_markdown\n\nimport os\nimport logging\nfrom typing import Union\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom scipy import interpolate\n\nimport core.utils.get\nimport core.utils.index\n\nfrom core.utils.db import Database\nfrom core.utils.checks import needs_database\n\nfrom datetime import datetime\n\n\nclass Leaderboard(commands.Cog):\n def __init__(self, bot):\n self.bot = bot\n self.log = logging.getLogger(__name__)\n\n \"\"\"--------------------------------------------------------------------------------------------------------------------------\"\"\"\n\n @staticmethod\n def chunks(l, n):\n \"\"\"Yield successive n-sized chunks from l.\"\"\"\n for i in range(0, len(l), n):\n yield l[i:i + n]\n\n \"\"\"--------------------------------------------------------------------------------------------------------------------------\"\"\"\n\n @commands.Cog.listener()\n async def on_ready(self):\n self.bot.readyCogs[self.__class__.__name__] = False\n\n #\n # leaderboard is updated via database trigger command\n #\n\n self.bot.readyCogs[self.__class__.__name__] = True\n\n \"\"\"--------------------------------------------------------------------------------------------------------------------------\"\"\"\n\n @commands.command()\n @needs_database\n async def leaderboard(self, ctx, channel: Union[TextChannel, Member] = None, member: Union[Member, TextChannel] = None, *, db: Database = None):\n channel, member = (channel if isinstance(channel, TextChannel) else\n member if isinstance(member, TextChannel) else\n None,\n\n member if isinstance(member, Member) else\n channel if isinstance(channel, Member) else\n None)\n \"\"\"\n get message leaderboard from the database\n the output format is\n FI MUNI Leaderboard!\n {n}. {count} {name}\n ... top10 ...\n\n Your position\n {n}. {count} {name}\n ... +-2 around you\n\n optional arguments\n #channel - get messages only in one channel\n @member - get only the Your position section\n \"\"\"\n\n bot_channel_id = os.getenv(\"BOT_CHANNEL\")\n if bot_channel_id is not None and ctx.channel.id != int(bot_channel_id):\n bot_channel = self.bot.get_channel(int(bot_channel_id))\n return await ctx.send(f\":point_right: {bot_channel.mention}\")\n\n author = ctx.message.author\n if ctx.guild is None:\n return await ctx.send(\"Not allowed in private channels\")\n\n bots = core.utils.get(ctx.guild.members, key=lambda user: user.bot)\n bots_ids = list(map(lambda bot: bot.id, bots))\n\n params = {\n \"guild_id\": ctx.guild.id,\n \"channel_id\": channel.id if channel else None,\n \"author_id\": member.id if member else author.id,\n \"ignored_ids\": tuple(bots_ids)\n }\n\n top10_SQL = f\"\"\"\n SELECT\n author_id,\n mem.name AS author,\n SUM(messages_sent) AS `count`\n FROM leaderboard AS ldb\n INNER JOIN `member` AS mem\n ON mem.id = ldb.author_id\n INNER JOIN `channel` AS chnl\n ON chnl.id = channel_id\n WHERE {'channel_id = %(channel_id)s' if channel is not None else '1'} AND guild_id = %(guild_id)s AND author_id NOT IN %(ignored_ids)s\n GROUP BY author_id\n ORDER BY `count` DESC\n LIMIT 10\n \"\"\"\n\n member_SQL = f\"\"\"\n DROP TEMPORARY TABLE IF EXISTS lookup;\n DROP TEMPORARY TABLE IF EXISTS first_table;\n DROP TEMPORARY TABLE IF EXISTS middle_table;\n DROP TEMPORARY TABLE IF EXISTS last_table;\n\n SET @desired_id = %(author_id)s;\n SET @row_number = 0;\n CREATE TEMPORARY TABLE lookup SELECT `row_number`, author_id, author, `count` FROM (\n SELECT\n (@row_number:=@row_number + 1) AS `row_number`,\n author_id,\n author,\n `count`\n FROM (\n SELECT\n author_id,\n mem.name AS author,\n SUM(messages_sent) AS `count`\n FROM leaderboard AS ldb\n INNER JOIN `member` AS mem\n ON mem.id = ldb.author_id\n INNER JOIN `channel` AS chnl\n ON chnl.id = channel_id\n WHERE {'channel_id = %(channel_id)s' if channel is not None else '1'} AND guild_id = %(guild_id)s AND author_id NOT IN %(ignored_ids)s\n GROUP BY author_id\n ORDER BY `count` DESC) AS sel\n ) AS sel;\n SET @desired_count = (SELECT `count` FROM lookup WHERE author_id = @desired_id);\n CREATE TEMPORARY TABLE first_table SELECT * FROM lookup WHERE `count` >= @desired_count AND author_id <> @desired_id ORDER BY `count` LIMIT 2;\n CREATE TEMPORARY TABLE middle_table SELECT * FROM lookup WHERE author_id = @desired_id;\n CREATE TEMPORARY TABLE last_table SELECT * FROM lookup WHERE `count` < @desired_count AND author_id <> @desired_id LIMIT 2;\n \"\"\"\n\n await db.execute(top10_SQL, params)\n rows1 = await db.fetchall()\n\n await db.execute(member_SQL, params)\n await db.execute(\"SELECT * from (SELECT * FROM first_table UNION ALL SELECT * FROM middle_table UNION ALL SELECT * FROM last_table) result ORDER BY `count` DESC;\")\n rows2 = await db.fetchall()\n\n \"\"\"\n print the leaderboard\n \"\"\"\n def right_justify(text, by=0, pad=\" \"):\n return pad * (by - len(str(text))) + str(text)\n\n def get_author(row):\n _id = author.id if not member else member.id\n if row[\"author_id\"] == _id:\n return f'**{escape_markdown(row[\"author\"])}**'\n else:\n return escape_markdown(row[\"author\"])\n\n template = \"`{index:0>2}.` {medal} `{count}` {author}\"\n\n embed = Embed(color=0x53acf2)\n if not member:\n embed.add_field(\n name=f\"FI MUNI Leaderboard! ({len(rows1)})\",\n inline=False,\n value=\"\\n\".join([\n template.format(\n index=i + 1,\n medal=self.get_medal(i + 1),\n count=right_justify(row[\"count\"], len(str(rows1[0][\"count\"])), \"\\u2063 \"),\n author=get_author(row)\n )\n for i, row in enumerate(rows1)\n ]))\n\n embed.add_field(\n name=\"Your position\",\n inline=True,\n value=\"\\n\".join([\n template.format(\n index=row[\"row_number\"],\n medal=self.get_medal(row[\"row_number\"]),\n count=right_justify(row[\"count\"], len(str(rows1[0][\"count\"])), \"\\u2063 \"),\n author=get_author(row)\n )\n for j, row in enumerate(rows2)\n ]))\n\n time_now = datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n embed.set_footer(text=f\"{str(author)} at {time_now}\", icon_url=author.avatar_url)\n await ctx.send(embed=embed)\n\n def get_medal(self, i):\n return {\n 1: core.utils.get(self.bot.emojis, name=\"gold_medal\"),\n 2: core.utils.get(self.bot.emojis, name=\"silver_medal\"),\n 3: core.utils.get(self.bot.emojis, name=\"bronze_medal\")\n }.get(i, core.utils.get(self.bot.emojis, name=\"BLANK\"))\n\n @commands.command(\"graph\")\n @commands.cooldown(1, 120, commands.BucketType.channel)\n @needs_database\n async def graph(self, ctx, *, db: Database = None):\n await db.execute(\"select extract( hour from created_at ) as hr, extract( minute from created_at ) as min, count( id ) from message group by hr, min order by hr, min\")\n rows = await db.fetchall()\n\n values = list(map(lambda x: tuple(x.values()), rows))\n xs = [t[0] * 60 * 60 + t[1] * 60 for t in values]\n ys = [t[2] for t in values]\n\n x_new = np.linspace(0, 86340, 86340)\n a_BSpline = interpolate.make_interp_spline(xs, ys)\n y_new = a_BSpline(x_new)\n\n plt.plot(x_new, y_new)\n plt.ylabel('msg/day')\n plt.savefig('assets/graph.png')\n\n await ctx.send(file=File('assets/graph.png'))\n\n\ndef setup(bot):\n bot.add_cog(Leaderboard(bot))\n", "sub_path": "cogs/leaderboard.py", "file_name": "leaderboard.py", "file_ext": "py", "file_size_in_byte": 8787, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "discord.ext.commands.Cog", "line_number": 23, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 23, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 26, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog.listener", "line_number": 38, "usage_type": "call"}, {"api_name": "discord.ext.commands.Cog", "line_number": 38, "usage_type": "attribute"}, {"api_name": "discord.ext.commands", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 52, "usage_type": "name"}, {"api_name": "discord.channel.TextChannel", "line_number": 52, "usage_type": "name"}, {"api_name": "discord.Member", "line_number": 52, "usage_type": "name"}, {"api_name": "core.utils.db.Database", "line_number": 52, "usage_type": "name"}, {"api_name": "discord.channel.TextChannel", "line_number": 53, "usage_type": "argument"}, {"api_name": "discord.channel.TextChannel", "line_number": 54, "usage_type": "argument"}, {"api_name": "discord.Member", "line_number": 57, "usage_type": "argument"}, {"api_name": "discord.Member", "line_number": 58, "usage_type": "argument"}, {"api_name": "os.getenv", "line_number": 76, "usage_type": "call"}, {"api_name": "core.utils.get.utils.get", "line_number": 85, "usage_type": "call"}, {"api_name": "core.utils.get.utils", "line_number": 85, "usage_type": "attribute"}, {"api_name": "core.utils.get", "line_number": 85, "usage_type": "name"}, {"api_name": "discord.utils.escape_markdown", "line_number": 161, "usage_type": "call"}, {"api_name": "discord.utils.escape_markdown", "line_number": 163, "usage_type": "call"}, {"api_name": "discord.Embed", "line_number": 167, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 195, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 195, "usage_type": "name"}, {"api_name": "discord.ext.commands.command", "line_number": 50, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 50, "usage_type": "name"}, {"api_name": "core.utils.checks.needs_database", "line_number": 51, "usage_type": "name"}, {"api_name": "core.utils.get.utils.get", "line_number": 201, "usage_type": "call"}, {"api_name": "core.utils.get.utils", "line_number": 201, "usage_type": "attribute"}, {"api_name": "core.utils.get", "line_number": 201, "usage_type": "name"}, {"api_name": "core.utils.get.utils.get", "line_number": 202, "usage_type": "call"}, {"api_name": "core.utils.get.utils", "line_number": 202, "usage_type": "attribute"}, {"api_name": "core.utils.get", "line_number": 202, "usage_type": "name"}, {"api_name": "core.utils.get.utils.get", "line_number": 203, "usage_type": "call"}, {"api_name": "core.utils.get.utils", "line_number": 203, "usage_type": "attribute"}, {"api_name": "core.utils.get", "line_number": 203, "usage_type": "name"}, {"api_name": "core.utils.get.utils.get", "line_number": 204, "usage_type": "call"}, {"api_name": "core.utils.get.utils", "line_number": 204, "usage_type": "attribute"}, {"api_name": "core.utils.get", "line_number": 204, "usage_type": "name"}, {"api_name": "core.utils.db.Database", "line_number": 209, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 217, "usage_type": "call"}, {"api_name": "scipy.interpolate.make_interp_spline", "line_number": 218, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 218, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 221, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 221, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 222, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 222, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 223, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 223, "usage_type": "name"}, {"api_name": "discord.File", "line_number": 225, "usage_type": "call"}, {"api_name": "discord.ext.commands.command", "line_number": 206, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 206, "usage_type": "name"}, {"api_name": "discord.ext.commands.cooldown", "line_number": 207, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 207, "usage_type": "name"}, {"api_name": "discord.ext.commands.BucketType", "line_number": 207, "usage_type": "attribute"}, {"api_name": "core.utils.checks.needs_database", "line_number": 208, "usage_type": "name"}]} +{"seq_id": "460144324", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport pygame\nimport pygame.mixer\nfrom time import sleep\nfrom scipy import signal, misc\nfrom scipy.io import wavfile\n\n# DTMF : Dual Tone Multiple Frequency\n\ndef Play_sound(filename):\n pygame.mixer.init()\n sound = pygame.mixer.Sound(filename)\n tmp = sound.play()\n while tmp.get_busy():\n pygame.time.delay(1)\n\nfreq_low = 852\nfreq_high = 1477\nA = 5000 # Amplitude\nfs = 8000 # freq_samp\ntu = 0.5 # time\n\nsample_num = int(fs*tu)\n# 합성한 파형의 Amplitude의 범위가 10000까지 늘어나므로 1/2 실시\ntone = [ ( A*np.sin(2 * np.pi * freq_low * x / fs) +\n A*np.sin(2 * np.pi * freq_high * x / fs) )/2 for x in range(sample_num) ]\ntone = np.cast['int16'](tone)\nwavfile.write(\"Tone.wav\", fs, tone)\nPlay_sound(\"Tone.wav\")", "sub_path": "MultimediaProgramming/Third/DMTF-Tone-Maker.py", "file_name": "DMTF-Tone-Maker.py", "file_ext": "py", "file_size_in_byte": 804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pygame.mixer.init", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.mixer.Sound", "line_number": 13, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 13, "usage_type": "attribute"}, {"api_name": "pygame.time.delay", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 16, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 26, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.cast", "line_number": 28, "usage_type": "attribute"}, {"api_name": "scipy.io.wavfile.write", "line_number": 29, "usage_type": "call"}, {"api_name": "scipy.io.wavfile", "line_number": 29, "usage_type": "name"}]} +{"seq_id": "41176321", "text": "import time\nimport logging\n\nimport pprint\npp = pprint.PrettyPrinter(indent=4)\n\nimport domain.model.mysql.pool\n\nfrom infrastructure.debug import timer\n\nlogger = logging.getLogger(__name__)\n\n\nclass Model():\n def __init__(self):\n self.db = domain.model.mysql.pool.Pool()\n\n @timer\n def getOwner(self, args):\n response = {}\n\n query = \"select * from owner where owner = '%s'\" % args['owner']\n record = self.db.retryQuery(query)\n if 'rows' in record:\n response = record['rows']\n #response['owner'] = record.owner\n #response['encryptkey'] = str(record.encryptkey)\n #response['note'] = record.note\n #response['date'] = int(record.date.strftime(\"%s\"))\n\n return(response)\n\n @timer\n def getOwnerEncryptKey(self, args):\n encryptkey = ''\n\n query = \"select encryptkey from owner where owner = '%s'\" % args['owner']\n record = self.db.retryQuery(query)\n if 'rows' in record:\n encryptkey = str(record['rows'][0]['encryptkey'])\n\n return(encryptkey)\n\n def getOwnerPWHash(self, args):\n response = {}\n\n query = \"select pwhash from owner where owner = '%s'\" % args['owner']\n record = self.db.retryQuery(query)\n if 'rows' in record:\n response['pwhash'] = record['rows'][0]['pwhash']\n\n return(response)\n\n def updateOwner(self, args):\n key = str(args['owner'])\n\n response = None\n\n if 'note' not in args:\n args['note'] = ''\n\n query = \"replace into owner (owner, encryptkey, pwhash, note) values ('%s','%s','%s','%s')\" % (\n args['owner'],\n args['encryptkey'],\n args['pwhash'],\n args['note']\n )\n\n record = self.db.retryQuery(query)\n #if record['count']:\n response = record\n \n return(response)\n\n def deleteOwner(self, args):\n response = None\n\n query = \"delete from owner where owner = '%s'\" % args['owner']\n record = self.db.retryQuery(query)\n #if record['rows']:\n response = record\n\n return(response)\n\n", "sub_path": "app/domain/model/mysql/owner.py", "file_name": "owner.py", "file_ext": "py", "file_size_in_byte": 2048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pprint.PrettyPrinter", "line_number": 5, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "domain.model.mysql.pool.model.mysql.pool.Pool", "line_number": 16, "usage_type": "call"}, {"api_name": "domain.model.mysql.pool.model", "line_number": 16, "usage_type": "attribute"}, {"api_name": "domain.model.mysql.pool", "line_number": 16, "usage_type": "name"}, {"api_name": "infrastructure.debug.timer", "line_number": 18, "usage_type": "name"}, {"api_name": "infrastructure.debug.timer", "line_number": 33, "usage_type": "name"}]} +{"seq_id": "615459536", "text": "from __future__ import print_function\n\nimport os\nimport argparse\n\nimport torch\nimport torch.utils.data as data\n\nfrom torchvision import transforms\nfrom data_loader import get_segmentation_dataset\nfrom models.model_zoo import get_segmentation_model\nfrom utils.score import SegmentationMetric\nfrom utils.visualize import get_color_pallete\n\nparser = argparse.ArgumentParser(\n description='Semantic Segmentation Evaluation')\nparser.add_argument('--model', type=str, default='fcn32s',\n help='model name (default: fcn32s)')\nparser.add_argument('--backbone', type=str, default='vgg16',\n help='backbone name (default: vgg16)')\nparser.add_argument('--dataset', type=str, default='pascal_voc',\n help='dataset name (default: pascal_voc, pascal_aug. choice=[pascal_voc, ade20k, citys]')\nparser.add_argument('--base-size', type=int, default=520,\n help='base image size')\nparser.add_argument('--crop-size', type=int, default=480,\n help='crop image size')\nparser.add_argument('--save-result', default=True,\n help='save the predict')\nparser.add_argument('--outdir', default='./eval', type=str,\n help='path to save the predict result')\nargs = parser.parse_args()\n\n\ndef eval(config):\n device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n # output folder\n if config.save_result:\n if not os.path.exists(config.outdir):\n os.makedirs(config.outdir)\n # image transform\n input_transform = transforms.Compose([\n transforms.ToTensor(),\n transforms.Normalize([.485, .456, .406], [.229, .224, .225]),\n ])\n\n # dataset and dataloader\n data_kwargs = {'transform': input_transform, 'base_size': args.base_size, 'crop_size': args.crop_size}\n test_dataset = get_segmentation_dataset(args.dataset, split='val', mode='val', **data_kwargs)\n\n test_loader = data.DataLoader(dataset=test_dataset,\n batch_size=1,\n shuffle=False)\n\n # create network\n model = get_segmentation_model(model=args.model, dataset=args.dataset, backbone=args.backbone,\n pretrained=True, pretrained_base=False, crop_size=args.crop_size).to(device)\n print('Finished loading model!')\n\n metric = SegmentationMetric(test_dataset.num_class)\n\n model.eval()\n for i, (image, label) in enumerate(test_loader):\n image = image.to(device)\n\n outputs = model(image)\n\n pred = torch.argmax(outputs[0], 1)\n pred = pred.cpu().data.numpy()\n label = label.numpy()\n\n metric.update(pred, label)\n pixAcc, mIoU = metric.get()\n print('Sample %d, validation pixAcc: %.3f%%, mIoU: %.3f%%' % (i + 1, pixAcc * 100, mIoU * 100))\n\n if config.save_result:\n predict = pred.squeeze(0)\n mask = get_color_pallete(predict, config.dataset)\n mask.save(os.path.join(config.outdir, 'seg_{}.png'.format(i)))\n\n\nif __name__ == '__main__':\n print('Testing model: ', args.model)\n eval(args)\n", "sub_path": "eval.py", "file_name": "eval.py", "file_ext": "py", "file_size_in_byte": 3114, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 38, "usage_type": "call"}, {"api_name": "os.path", "line_number": 38, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 39, "usage_type": "call"}, {"api_name": "torchvision.transforms.Compose", "line_number": 41, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 41, "usage_type": "name"}, {"api_name": "torchvision.transforms.ToTensor", "line_number": 42, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 42, "usage_type": "name"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 43, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 43, "usage_type": "name"}, {"api_name": "data_loader.get_segmentation_dataset", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 50, "usage_type": "call"}, {"api_name": "torch.utils.data", "line_number": 50, "usage_type": "name"}, {"api_name": "models.model_zoo.get_segmentation_model", "line_number": 55, "usage_type": "call"}, {"api_name": "utils.score.SegmentationMetric", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.argmax", "line_number": 67, "usage_type": "call"}, {"api_name": "utils.visualize.get_color_pallete", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path", "line_number": 78, "usage_type": "attribute"}]} +{"seq_id": "253150737", "text": "import logging\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.mlab\n\nlog = logging.getLogger(__name__)\n\ndef plot_accuracies(epochs, accuracies, labels, title):\n\n fig_ = plt.figure(figsize=(32, 16))\n\n for i in range(len(accuracies)):\n\n plt.plot(epochs, accuracies[i], label=labels[i])\n\n plt.ylim(0.0, 1.0)\n plt.yticks([0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.0])\n plt.grid()\n\n plt.legend(loc=2, ncol=1)\n plt.title(title)\n plt.xlabel(\"Epoch\")\n plt.ylabel(\"Accuracy (%)\")\n plt.show()", "sub_path": "utils/plotaccuracies.py", "file_name": "plotaccuracies.py", "file_ext": "py", "file_size_in_byte": 527, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 11, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 11, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 15, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 17, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 17, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "161931431", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport logging\nfrom pathlib import Path\nfrom typing import List\n\nfrom tqdm import tqdm\n\nimport numpy as np\nimport pandas as pd\nfrom datastep import Step, log_run_params\nfrom PIL import Image\n\n###############################################################################\n\nlog = logging.getLogger(__name__)\n\n###############################################################################\n\n\nclass Raw(Step):\n # You only need to have an __init__ if you aren't using the default values\n # In this case, we could get rid of it but for the purposes of this example\n # we will keep it.\n def __init__(self, direct_upstream_tasks=[], config=None):\n super().__init__(direct_upstream_tasks=direct_upstream_tasks, config=config)\n\n @log_run_params\n def run(self, n: int = 10, **kwargs) -> List[Path]:\n \"\"\"\n Generate N random images and save them to /images.\n\n Parameters\n ----------\n n: int\n Number of images to generate.\n\n Returns\n -------\n images: List[Path]\n A list of paths that point to the generated images.\n \"\"\"\n\n # Empty manifest to fill in -- add more columns for e.g. labels, metadata, etc.\n self.manifest = pd.DataFrame(index=range(n), columns=[\"filepath\"])\n\n # Subdirectory for the images\n imdir = self.step_local_staging_dir / Path(\"images\")\n imdir.mkdir(parents=True, exist_ok=True)\n\n # Set seed for reproducible random images\n np.random.seed(seed=112358)\n\n # Create images, save them, and fill in dataframe\n images = []\n for i in tqdm(range(n), desc=\"Creating and saving images\"):\n A = np.random.rand(128, 128, 4) * 255\n img = Image.fromarray(A.astype(\"uint8\")).convert(\"RGBA\")\n path = imdir / Path(f\"image_{i}.png\")\n img.save(path)\n self.manifest.at[i, \"filepath\"] = path\n images.append(path)\n\n # Save manifest as csv\n self.manifest.to_csv(\n self.step_local_staging_dir / Path(\"manifest.csv\"), index=False\n )\n\n return images\n", "sub_path": "{{cookiecutter.project_slug}}/{{cookiecutter.project_slug}}/steps/raw/raw.py", "file_name": "raw.py", "file_ext": "py", "file_size_in_byte": 2161, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "datastep.Step", "line_number": 22, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 46, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tqdm.tqdm", "line_number": 57, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 58, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 59, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 59, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 60, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 67, "usage_type": "call"}, {"api_name": "datastep.log_run_params", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 30, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 30, "usage_type": "name"}]} +{"seq_id": "594471363", "text": "from pytube import YouTube\n\n# video_list=['https://www.youtube.com/watch?v=lUCNWfwBjy0','https://www.youtube.com/watch?v=KskbFT-RAIU']\n\nvideo_list=open(\"sample.txt\",'r')\n#dv=yt.streams.get_by_itag(20)\n# print(yt.streams.all())\nfor i in video_list:\n yt=YouTube(i)\n try:\n # dv=yt.streams.first()\n dv=yt.streams.first()\n dv.download(\"E:/\")\n print(\"download completed:\",i)\n except:\n print(\"download failed:\",i)\n\n", "sub_path": "yt.py", "file_name": "yt.py", "file_ext": "py", "file_size_in_byte": 453, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pytube.YouTube", "line_number": 9, "usage_type": "call"}]} +{"seq_id": "262634280", "text": "import os, uuid, json\nimport apiai\n\nclass ApiaiModule():\n '''\n [参考]\n https://qiita.com/flatfisher/items/53923e33ce1f23b3820a\n https://qiita.com/nagase/items/67d334d8d59df5ce28eb\n '''\n def __init__(self):\n '''\n [引数]\n なし\n [戻り値]\n なし\n '''\n self.CLIENT_ACCESS_TOKEN = os.environ[\"DF_CLIENT_ACCESS_TOKEN\"]\n self.user_responses = \"\" # ユーザからの回答一覧.リストではなく「:+:」でつなぐものとして簡易に実装.\n\n def getResponse(self, text):\n '''\n [引数]\n ●text:メンチョン本文\n [戻り値]\n ●Dialogflowから渡されたリプライ文\n '''\n ai = apiai.ApiAI(self.CLIENT_ACCESS_TOKEN)\n\n request = ai.text_request()\n request.lang = \"ja\"\n request.session_id = str(uuid.uuid4()) #DialogFlow間のセッションIDをランダムなUUID型でセット\n \n self.user_responses += \":+:[{0}]\".format(text) # ユーザレスポンスを加えていく\n request.query = self.user_responses\n\n response = request.getresponse()\n # https://techacademy.jp/magazine/18987\n json_res = json.loads(response.read().decode()) # 標準出力はsjis?なので文字コードの変換を行い,JSONに変換.\n print(json_res)\n\n if not json_res[\"result\"][\"actionIncomplete\"]: # レスポンスが完全なものだったらユーザレスポンスをクリア.\n self.user_responses = \"\"\n\n return json_res[\"result\"][\"fulfillment\"][\"speech\"]", "sub_path": "plugins/modules/apiai_module.py", "file_name": "apiai_module.py", "file_ext": "py", "file_size_in_byte": 1606, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.environ", "line_number": 17, "usage_type": "attribute"}, {"api_name": "apiai.ApiAI", "line_number": 27, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 31, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 38, "usage_type": "call"}]} +{"seq_id": "147007817", "text": "'''\nCreated on 02/10/2014\n\n:author: alfred\n'''\nfrom unittest import TestCase\nfrom xml.etree.ElementTree import Element, register_namespace, QName, ElementTree\nfrom mc_be.commons.service_clients.xml import xml_encoder, xml_decoder\n\n\nclass XmlTest(TestCase):\n\n def setUp(self):\n self.ns_uri = 'http://test'\n self.ns_prefix = 'test'\n register_namespace(self.ns_prefix, self.ns_uri)\n\n def test_xml_encoder(self):\n elem_root = Element(QName(self.ns_uri, 'main'))\n elem_body = Element(QName(self.ns_uri, 'body'))\n elem_root.append(elem_body)\n\n child_el = Element(QName(self.ns_uri, 'child1'))\n child_el.text = 'test1'\n elem_body.append(child_el)\n child_el = Element(QName(self.ns_uri, 'child1'))\n child_el.text = 'test2'\n elem_body.append(child_el)\n\n child_el = Element(QName(self.ns_uri, 'child2'))\n child_el.text = '2'\n elem_body.append(child_el)\n\n tree = ElementTree(elem_root)\n self.assertEqual(xml_encoder(tree),\n ''\n 'test1test2'\n '2')\n\n def test_xml_decoder(self):\n xml_str = '' \\\n 'test1test2' \\\n '2'\n\n xml_tree = xml_decoder(xml_str)\n self.assertIsInstance(xml_tree, ElementTree)\n self.assertEqual(xml_encoder(xml_tree), xml_str)\n\n def test_xml_decoder_none(self):\n self.assertIsNone(xml_decoder(None))\n", "sub_path": "mc-pybe-release-smip-R4/tests/commons/service_clients/tests_xml.py", "file_name": "tests_xml.py", "file_ext": "py", "file_size_in_byte": 1782, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 11, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.register_namespace", "line_number": 16, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 19, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.QName", "line_number": 19, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 20, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.QName", "line_number": 20, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 23, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.QName", "line_number": 23, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 26, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.QName", "line_number": 26, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.Element", "line_number": 30, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.QName", "line_number": 30, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 34, "usage_type": "call"}, {"api_name": "mc_be.commons.service_clients.xml.xml_encoder", "line_number": 35, "usage_type": "call"}, {"api_name": "mc_be.commons.service_clients.xml.xml_decoder", "line_number": 45, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 46, "usage_type": "argument"}, {"api_name": "mc_be.commons.service_clients.xml.xml_encoder", "line_number": 47, "usage_type": "call"}, {"api_name": "mc_be.commons.service_clients.xml.xml_decoder", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "645568986", "text": "import hashlib\nimport sqlite3\nimport PIL.Image\nimport sys\nimport re\nimport io\nfrom typing import NamedTuple\n\nIM_TABLE = 'imgs'\nFILE_TABLE = 'files'\nURL_TABLE = 'links'\n\nIM_TYPES = [\n '.png',\n '.jpg',\n '.jpeg'\n]\n\n\"\"\"\n================\nCommunication Structures\n================\n\"\"\"\n\nclass LinkData(NamedTuple):\n link: str\n date: int\n author: str\n\nclass HashData(NamedTuple):\n f_hash: bytes\n length: int\n link: str\n date: int\n author: str\n\nclass RepostData(NamedTuple):\n repost_type: str\n original_date: str\n original_author: str\n last_date: str\n last_author: str\n num_reposts: int\n\n\n\"\"\"\n================\nHASH FUNCTIONS\n================\n\"\"\"\n\n\ndef _get_file_hash(fp: str, f_bytes: bytes) -> str:\n return len(f_bytes), hashlib.sha256(f_bytes).digest()\n\ndef _get_img_hash(fp: str, f_bytes: bytes) -> str:\n f_bytes = PIL.Image.open(io.BytesIO(f_bytes)).tobytes()\n return len(f_bytes), hashlib.sha256(f_bytes).digest()\n\ndef calculate_hash(fp: str, f_bytes: bytes) -> (bool, int, str):\n is_image = True\n try:\n valid_ims = [fp.endswith(im_end) for im_end in IM_TYPES]\n if True not in valid_ims: raise ValueError('Not an Image')\n\n f_len, f_hash = _get_img_hash(fp, f_bytes)\n except Exception as e:\n is_image = False\n f_len, f_hash = _get_file_hash(fp, f_bytes)\n\n return is_image, f_len, f_hash\n\n\ndef _insert_hash(conn: sqlite3.Connection, table: str, hash_data: HashData) -> None:\n insert_sql = f\"\"\" INSERT INTO {table}(hash,\n len,\n link,\n original_date,\n original_author,\n last_date,\n last_author,\n num_reposts)\n VALUES(?, ?, ?, ?, ?, ?, ?, ?)\n \"\"\"\n cur = conn.cursor()\n cur.execute(insert_sql, (hash_data.f_hash,\n hash_data.length,\n hash_data.link,\n hash_data.date,\n hash_data.author,\n hash_data.date,\n hash_data.author,\n 0))\n conn.commit()\n\n\ndef _update_hash(conn: sqlite3.Connection, table: str, hash_data: HashData, num_reposts: int) -> None:\n cur_reposts = num_reposts + 1\n insert_sql = f\"\"\" UPDATE {table}\n SET last_date=?, last_author=?, num_reposts=?\n WHERE hash=? AND len=?\n \"\"\"\n cur = conn.cursor()\n cur.execute(insert_sql, (hash_data.date, hash_data.author, cur_reposts, hash_data.f_hash, hash_data.length))\n conn.commit()\n\ndef check_hash_table(conn: sqlite3.Connection, is_img: bool, hash_data: HashData):\n if is_img:\n table = IM_TABLE\n else:\n table = FILE_TABLE\n\n sql_query = f\"\"\"SELECT\n original_date,\n original_author,\n last_date,\n last_author,\n num_reposts\n FROM {table} WHERE\n hash=? AND len=?\n \"\"\"\n cur = conn.cursor()\n cur.execute(sql_query, (hash_data.f_hash, hash_data.length))\n\n hits = cur.fetchall()\n\n is_original = len(hits) == 0\n\n if is_original:\n _insert_hash(conn, table, hash_data)\n repost = RepostData(None, None, None, None, None, None)\n\n else:\n _update_hash(conn, table, hash_data, hits[0][4])\n repost = RepostData(table,\n hits[0][0],\n hits[0][1],\n hits[0][2],\n hits[0][3],\n hits[0][4])\n\n return is_original, repost\n\n\n\"\"\"\n================\nURL FUNCTIONS\n================\n\"\"\"\ndef extract_urls(msg: str) -> list:\n regex = r\"(?i)\\b((?:https?://|www\\d{0,3}[.]|[a-z0-9.\\-]+[.][a-z]{2,4}/)(?:[^\\s()<>]+|\\(([^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\))+(?:\\(([^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\)|[^\\s`!()\\[\\]{};:'\\\".,<>?«»“”‘’]))\"\n return [url[0] for url in re.findall(regex, msg, flags=re.MULTILINE)]\n\n\ndef _insert_url(conn: sqlite3.Connection, link_data: LinkData) -> None:\n insert_sql = \"\"\" INSERT INTO links(link,\n original_date,\n original_author,\n last_date,\n last_author,\n num_reposts)\n VALUES(?, ?, ?, ?, ?, ?)\n \"\"\"\n cur = conn.cursor()\n cur.execute(insert_sql, (link_data.link,\n link_data.date,\n link_data.author,\n link_data.date,\n link_data.author,\n 0))\n conn.commit()\n\n\ndef _update_url(conn: sqlite3.Connection, link_data: LinkData, num_reposts: int) -> None:\n cur_reposts = num_reposts + 1\n insert_sql = \"\"\" UPDATE links\n SET last_date=?, last_author=?, num_reposts=?\n WHERE link=?\n \"\"\"\n cur = conn.cursor()\n cur.execute(insert_sql, (link_data.date, link_data.author, cur_reposts, link_data.link))\n conn.commit()\n\n\ndef check_url_table(conn: sqlite3.Connection, link_data: LinkData):\n sql_query = \"\"\"SELECT\n original_date,\n original_author,\n last_date,\n last_author,\n num_reposts\n FROM links WHERE\n link=?\n \"\"\"\n cur = conn.cursor()\n cur.execute(sql_query, (link_data.link,))\n\n hits = cur.fetchall()\n\n is_original = len(hits) == 0\n\n if is_original:\n _insert_url(conn, link_data)\n repost = RepostData(None, None, None, None, None, None)\n\n else:\n _update_url(conn, link_data, hits[0][4])\n repost = RepostData(URL_TABLE,\n hits[0][0],\n hits[0][1],\n hits[0][2],\n hits[0][3],\n hits[0][4])\n\n return is_original, repost", "sub_path": "scripts_preprocessing/repost/repost_utils.py", "file_name": "repost_utils.py", "file_ext": "py", "file_size_in_byte": 6317, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.NamedTuple", "line_number": 25, "usage_type": "name"}, {"api_name": "typing.NamedTuple", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.NamedTuple", "line_number": 37, "usage_type": "name"}, {"api_name": "hashlib.sha256", "line_number": 54, "usage_type": "call"}, {"api_name": "PIL.Image.Image.open", "line_number": 57, "usage_type": "call"}, {"api_name": "PIL.Image.Image", "line_number": 57, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 57, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 57, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 58, "usage_type": "call"}, {"api_name": "sqlite3.Connection", "line_number": 74, "usage_type": "attribute"}, {"api_name": "sqlite3.Connection", "line_number": 97, "usage_type": "attribute"}, {"api_name": "sqlite3.Connection", "line_number": 107, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 152, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 152, "usage_type": "attribute"}, {"api_name": "sqlite3.Connection", "line_number": 155, "usage_type": "attribute"}, {"api_name": "sqlite3.Connection", "line_number": 174, "usage_type": "attribute"}, {"api_name": "sqlite3.Connection", "line_number": 185, "usage_type": "attribute"}]} +{"seq_id": "225294706", "text": "import tkinter as tk\nfrom tkinter import ttk\nfrom PIL import Image, ImageTk\nfrom selenium import webdriver\nfrom faker import Faker\nfrom webdriver_manager.chrome import ChromeDriverManager\nimport sys\nimport requests\nimport time\nimport datetime\n\nwindow = tk.Tk()\nwindow.title('SkullBomb')\nwindow.geometry(\"500x600\")\nfake = Faker()\n\nentryOrder = []\nattack = False\n\nentrance = tk.Frame(window)\nentrance.pack(pady=20)\nbuttonFrame = tk.Frame(window)\nbuttonFrame.pack(ipady=20)\ntopframe = tk.Frame(window)\ntopframe.pack()\n\n\ndef attack():\n driver = webdriver.Chrome(ChromeDriverManager().install())\n \n target_url = url_entry.get()\n \n while True:\n\n driver.get(target_url)\n\n for point in entryOrder:\n current_point = driver.find_element_by_xpath(point[1].get())\n if(point[0][:-1] == 'param'):\n if(point[2].get() == 'name'):\n current_point.send_keys(fake.name())\n if(point[2].get() == 'email'):\n current_point.send_keys(fake.email())\n if(point[2].get() == 'text'):\n current_point.send_keys(fake.text())\n if(point[2].get() == 'date'):\n current_point.send_keys(fake.date())\n if(point[2].get() == 'address'):\n current_point.send_keys(fake.address())\n if(point[2].get() == 'phone'):\n current_point.send_keys(fake.phone_number())\n if(point[2].get() == 'bank acount'):\n current_point.send_keys(fake.iban())\n if(point[0][:-1] == 'button'):\n current_point.click()\n time.sleep(1)\n\n\t\t\ndef display_params_and_buttons(array_item):\n for item in array_item:\n entryIndex = entryOrder.index(item) + 1\n label = tk.Label(topframe, text=item[0])\n label.grid(row=entryIndex)\n item[1].grid(row=entryIndex, column=1)\n if(len(item) > 2):\n item[2].grid(row=entryIndex, column=2)\n \n\ndef add_new_parameter():\n length = len(entryOrder)\n newParameter = \"param\" + str(length + 1)\n entryOrder.append([\n newParameter,\n tk.Entry(topframe, text=newParameter),\n ttk.Combobox(\n topframe, \n values=[\n 'name',\n 'email',\n 'text',\n 'date',\n 'address',\n 'phone',\n 'bank account',\n ],\n )\n ])\n display_params_and_buttons(entryOrder)\n\ndef add_new_button():\n length = len(entryOrder)\n newParameter = \"button\" + str(length + 1)\n entryOrder.append([\n newParameter,\n tk.Entry(topframe, text=newParameter),\n ])\n display_params_and_buttons(entryOrder)\n\nadd_param = tk.Button(buttonFrame, text=\"Add input\", command=add_new_parameter)\nadd_button = tk.Button(buttonFrame, text=\"Add button\", command=add_new_button)\nattack_button = tk.Button(buttonFrame, text=\"Attack\", command=attack, fg='tomato')\n\nadd_param.grid(row=0)\nadd_button.grid(row=0, column=1)\nattack_button.grid(row=0, column=2)\n\n\nimg = Image.open(\"./assets/skull.png\")\nimg = img.resize((200, 250))\ntkimage = ImageTk.PhotoImage(img)\ntk.Label(entrance, image=tkimage, pady=10).grid()\n\n\nurl = tk.Label(topframe, text=\"Target url\", padx=5, pady=5)\nurl.grid(row=0)\nurl_entry = tk.Entry(topframe)\nurl_entry.grid(row=0, column=1, columnspan=3, sticky='ew')\n\nwindow.mainloop()", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3455, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "tkinter.Tk", "line_number": 12, "usage_type": "call"}, {"api_name": "faker.Faker", "line_number": 15, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 20, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 22, "usage_type": "call"}, {"api_name": "tkinter.Frame", "line_number": 24, "usage_type": "call"}, {"api_name": "selenium.webdriver.Chrome", "line_number": 29, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 29, "usage_type": "name"}, {"api_name": "webdriver_manager.chrome.ChromeDriverManager", "line_number": 29, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 56, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 62, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 74, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 75, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 75, "usage_type": "name"}, {"api_name": "tkinter.Entry", "line_number": 95, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 99, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 100, "usage_type": "call"}, {"api_name": "tkinter.Button", "line_number": 101, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 108, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 108, "usage_type": "name"}, {"api_name": "PIL.ImageTk.PhotoImage", "line_number": 110, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 110, "usage_type": "name"}, {"api_name": "tkinter.Label", "line_number": 111, "usage_type": "call"}, {"api_name": "tkinter.Label", "line_number": 114, "usage_type": "call"}, {"api_name": "tkinter.Entry", "line_number": 116, "usage_type": "call"}]} +{"seq_id": "277392199", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri May 10 12:50:16 2019\n\n@author: doorleyr\n\"\"\"\n\nimport pandas as pd\nimport pylogit as pl\nimport numpy as np\nimport json\nimport pickle\nimport random\nfrom collections import OrderedDict\n\ncity='Boston'\n\nNUM_ALTS=5\nsample_size=5000\n\nPUMS_HH_PATH='./'+city+'/raw/PUMS/csv_hma/ss16hma.csv'\nPUMA_POP_PATH='./'+city+'/raw/ACS/ACS_17_5YR_B01003/population.csv'\nPUMAS_INCLUDED_PATH='./'+city+'/raw/PUMS/pumas_included.json'\nPUMA_SHAPE_PATH='./'+city+'/raw/PUMS/pumas.geojson'\nFITTED_HOME_LOC_MODEL_PATH='./models/home_loc_logit.p'\n\n# get area of every PUMA\npumas_shape=json.load(open(PUMA_SHAPE_PATH))\nprint(pumas_shape['features'][0]['properties'])\npuma_land_sqm={int(f['properties']['PUMACE10']): f['properties']['ALAND10']\n for f in pumas_shape['features']}\n\n# load the PUMS data\nhh=pd.read_csv(PUMS_HH_PATH)\npumas_included=json.load(open(PUMAS_INCLUDED_PATH))\n# load the aggregate PUMA data\npuma_pop=pd.read_csv(PUMA_POP_PATH)\npuma_pop['PUMA']=puma_pop.apply(lambda row: int(row['GEO.id2'][2:]), axis=1)\npuma_pop=puma_pop.set_index('PUMA')\n\n#create subsets\nhh_vacant=hh[hh['NP']==0]\nhh_boston=hh[hh['PUMA'].isin(pumas_included)]\nhh_vacant_for_rent=hh_boston[hh_boston['VACS']==1]\nhh_rented=hh[hh['TEN']==3]\n#asking_rent=hh_for_rent['RNTP']\nrenters_recent_move=hh_rented[hh_rented['MV']==1]\n\n# create features at property level\n# normalise rent in each cateogory og bedroom number\nrenters_recent_move.loc[renters_recent_move['BDSP']>2, 'BDSP']=3\nrenters_recent_move.loc[renters_recent_move['BDSP']<1, 'BDSP']=1\nhh_vacant_for_rent.loc[hh_vacant_for_rent['BDSP']>2, 'BDSP']=3\nhh_vacant_for_rent.loc[hh_vacant_for_rent['BDSP']<1, 'BDSP']=1\nrent_mean={}\nrent_std={}\nfor beds in range(1,4):\n rent_mean[beds]=renters_recent_move.loc[renters_recent_move['BDSP']==beds, 'RNTP'].mean()\n rent_std[beds]=renters_recent_move.loc[renters_recent_move['BDSP']==beds, 'RNTP'].std()\nrenters_recent_move['norm_rent']=renters_recent_move.apply(\n lambda row: (row['RNTP']-rent_mean[row['BDSP']])/rent_std[row['BDSP']], axis=1)\nhh_vacant_for_rent['norm_rent']=hh_vacant_for_rent.apply(\n lambda row: (row['RNTP']-rent_mean[row['BDSP']])/rent_std[row['BDSP']], axis=1)\n\n# Age of building\nfor df in [renters_recent_move, hh_vacant_for_rent]:\n df['less_than_5yrs_old']=df.apply(lambda row: row['YBL']>14, axis=1)\n\n\n# build the PUMA aggregate data data frame\nmedian_income_by_puma=hh.groupby('PUMA')['HINCP'].median()\n#TODO: get more zonal attributes such as access to employment, amenities etc.\n\nall_PUMAs=list(set(hh['PUMA']))\npuma_obj=[{'PUMA':puma,\n 'med_income':median_income_by_puma.loc[puma],\n 'puma_pop_per_sqm':float(puma_pop.loc[puma]['HD01_VD01'])/puma_land_sqm[puma]\n } for puma in all_PUMAs]\npuma_df=pd.DataFrame(puma_obj)\npuma_df=puma_df.set_index('PUMA')\n# for each PUMS person, add to the long df their actual HH and N vacant HHs\n\nrandom.seed(1)\nlong_data_obj=[]\nind=0\nfor ind_actual, row_actual in renters_recent_move[:sample_size].iterrows():\n cid=1\n choiceObs={'custom_id':ind,# identify the individual\n 'choice_id':cid, # fake choice identifier- shouldn't matter if no ASC\n 'choice':1,\n 'rent':row_actual['RNTP'],\n 'norm_rent':row_actual['norm_rent'],\n 'puma':row_actual['PUMA'],\n 'less_than_5yrs_old':int(row_actual['less_than_5yrs_old']),\n 'hh_income':row_actual['HINCP']\n }\n cid+=1\n long_data_obj.append(choiceObs)\n for i in range(NUM_ALTS):\n selected=random.choice(range(len(hh_vacant_for_rent)))\n alt_obs={'custom_id':ind,# identify the individual\n 'choice_id':cid, # fake choice identifier- shouldn't matter if no ASC\n 'choice':0,\n 'rent':hh_vacant_for_rent.iloc[selected]['RNTP'],\n 'norm_rent':hh_vacant_for_rent.iloc[selected]['norm_rent'],\n 'puma':hh_vacant_for_rent.iloc[selected]['PUMA'],\n 'less_than_5yrs_old':int(hh_vacant_for_rent.iloc[selected]['less_than_5yrs_old']),\n 'hh_income':row_actual['HINCP']\n }\n cid+=1\n long_data_obj.append(alt_obs)\n ind+=1\n\n# get zonal attributes\nlong_data=pd.DataFrame(long_data_obj) \nlong_data['puma_pop_per_sqm']=long_data.apply(lambda row: puma_df.loc[row['puma']]['puma_pop_per_sqm'], axis=1)\nlong_data['income_disparity']=long_data.apply(lambda row: np.abs(row['hh_income']-puma_df.loc[row['puma']]['med_income']), axis=1)\nlong_data['log_rent']=long_data.apply(lambda row: np.log(row['rent']), axis=1)\n\n# TODO: calculate interactions\n\n# fit model\n\nbasic_specification = OrderedDict()\nbasic_names = OrderedDict()\n\nbasic_specification[\"puma_pop_per_sqm\"] = [list(set(long_data['choice_id']))]\nbasic_names[\"puma_pop_per_sqm\"] = ['puma_pop_per_sqm']\n\nbasic_specification[\"income_disparity\"] = [list(set(long_data['choice_id']))]\nbasic_names[\"income_disparity\"] = ['income_disparity']\n\nbasic_specification[\"norm_rent\"] = [list(set(long_data['choice_id']))]\nbasic_names[\"norm_rent\"] = ['norm_rent']\n\nbasic_specification[\"less_than_5yrs_old\"] = [list(set(long_data['choice_id']))]\nbasic_names[\"less_than_5yrs_old\"] = ['less_than_5yrs_old']\n\nhome_loc_mnl = pl.create_choice_model(data=long_data,\n alt_id_col='choice_id',\n obs_id_col='custom_id',\n choice_col='choice',\n specification=basic_specification,\n model_type=\"MNL\",\n names=basic_names)\n\n# Specify the initial values and method for the optimization.\nprint('Fitting Model')\nnumCoef=sum([len(basic_specification[s]) for s in basic_specification])\nhome_loc_mnl.fit_mle(np.zeros(numCoef))\n\n# Look at the estimation results\nprint(home_loc_mnl.get_statsmodels_summary())\n\npickle.dump(home_loc_mnl, FITTED_HOME_LOC_MODEL_PATH)\n\n\n\n\n\n", "sub_path": "python/home_loc_choice.py", "file_name": "home_loc_choice.py", "file_ext": "py", "file_size_in_byte": 6054, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "json.load", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 35, "usage_type": "call"}, {"api_name": "json.load", "line_number": 36, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 38, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 80, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 84, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 101, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 119, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 125, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 126, "usage_type": "call"}, {"api_name": "pylogit.create_choice_model", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 151, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 156, "usage_type": "call"}]} +{"seq_id": "474952489", "text": "\"\"\"\nSubarray Product Less Than K\nhttps://leetcode.com/explore/challenge/card/september-leetcoding-challenge/557/week-4-september-22nd-september-28th/3475/\nYour are given an array of positive integers nums.\n\nCount and print the number of (contiguous) subarrays where the product of all the elements in the subarray is less than k.\n\nExample 1:\nInput: nums = [10, 5, 2, 6], k = 100\nOutput: 8\nExplanation: The 8 subarrays that have product less than 100 are: [10], [5], [2], [6], [10, 5], [5, 2], [2, 6], [5, 2, 6].\nNote that [10, 5, 2] is not included as the product of 100 is not strictly less than k.\nNote:\n\n0 < nums.length <= 50000.\n0 < nums[i] < 1000.\n0 <= k < 10^6.\n Hide Hint #1\nFor each j, let opt(j) be the smallest i so that nums[i] * nums[i+1] * ... * nums[j] is less than k. opt is an increasing function.\n\"\"\"\nfrom typing import List\n\n\nclass Solution:\n def numSubarrayProductLessThanK(self, nums: List[int], k: int) -> int:\n # Solution 1 - 1080 ms\n \"\"\"\n # Base case\n if not nums:\n return 0\n i = 0\n res = 0\n current = 1\n for j in range(len(nums)):\n current = current * nums[j]\n # j >= i to prevent index out of range cases i.e. nums = [1,2,3], k = 0\n while current >= k and j >= i:\n current = current // nums[i]\n i += 1\n # Add size of the window to final output\n res += j - i + 1\n return res\n \"\"\"\n # Solution 2 - 1080 ms\n \"\"\"\n left, right = 0, 0\n ans = 0\n n = len(nums)\n cumProd = 1\n while right < n:\n cumProd *= nums[right]\n right += 1\n while cumProd >= k and left < right:\n cumProd /= nums[left]\n left += 1\n ans += right - left\n return ans\n \"\"\"\n # Solution 3 - 1052 ms\n # sliding window\n # move left while window product >= k\n # add (right - left + 1) if window < k\n\n if not nums or k <= 0:\n return 0\n n = len(nums)\n\n window = 1\n left = 0\n\n ret = 0\n for right in range(n):\n window *= nums[right]\n while left < right and window >= k:\n window //= nums[left]\n left += 1\n if window < k:\n ret += (right - left + 1)\n return ret\n\n\n# Main call\nsolution = Solution()\nnums = [10, 5, 2, 6]\nk = 100\nprint(solution.numSubarrayProductLessThanK(nums, k))\n", "sub_path": "src/arrays/numSubarrayProductLessThanK.py", "file_name": "numSubarrayProductLessThanK.py", "file_ext": "py", "file_size_in_byte": 2523, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.List", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "499282592", "text": "from datetime import datetime\nimport decimal\nimport logging\nfrom collections import OrderedDict\nfrom io import BytesIO, StringIO\nimport re\n\nfrom django.core.cache import cache\nfrom django.core.serializers.json import DjangoJSONEncoder\nfrom django.shortcuts import HttpResponse\nfrom django.utils import html\n\nfrom rest_framework.renderers import JSONRenderer\n\nimport csv\nimport xlsxwriter\n\n\ndef HAWCtoDateString(datetime):\n \"\"\"\n Helper function to ensure dates are consistent.\n \"\"\"\n return datetime.strftime(\"%B %d %Y, %I:%M %p\")\n\n\ndef cleanHTML(txt):\n return strip_entities(\n strip_tags(\n txt.replace('\\n', ' ')\n .replace('\\r', \"\")\n .replace('
', \"\\n\")\n .replace(\" \", \" \")))\n\n\ndef strip_entities(value):\n \"\"\"Return the given HTML with all entities (&something;) stripped.\"\"\"\n # Note: Originally in Django but removed in v1.10\n return re.sub(r'&(?:\\w+|#\\d+);', '', html.force_text(value))\n\n\ndef strip_tags(value):\n \"\"\"Return the given HTML with all tags stripped.\"\"\"\n # Note: in typical case this loop executes _strip_once once. Loop condition\n # is redundant, but helps to reduce number of executions of _strip_once.\n # Note: Originally in Django but removed in v1.10\n while '<' in value and '>' in value:\n new_value = html._strip_once(value)\n if new_value == value:\n # _strip_once was not able to detect more tags\n break\n value = new_value\n return value\n\n\ndef listToUl(list_):\n return \"
    {0}
\".format(\n \"\".join([\"
  • {0}
  • \".format(d) for d in list_]))\n\n\ndef tryParseInt(val, default=None):\n try:\n return int(val)\n except (ValueError, TypeError):\n return default\n\n\nclass HAWCDjangoJSONEncoder(DjangoJSONEncoder):\n \"\"\"\n Modified to return a float instead of a string.\n \"\"\"\n def default(self, o):\n if isinstance(o, decimal.Decimal):\n return float(o)\n else:\n return super().default(o)\n\n\nclass SerializerHelper(object):\n \"\"\"\n HAWC helper-object for getting serialized objects and setting cache.\n Sets cache names based on django models and primary keys automatically.\n Sets a cache using the serialized object, and also a JSON object.\n \"\"\"\n\n serializers = {}\n\n @classmethod\n def _get_cache_name(cls, model, id, json=True):\n name = \"{}.{}.{}\".format(model.__module__, model.__name__, id)\n if json:\n name += \".json\"\n return name\n\n @classmethod\n def get_serialized(cls, obj, json=True, from_cache=True):\n if from_cache:\n name = cls._get_cache_name(obj.__class__, obj.id, json)\n cached = cache.get(name)\n if cached:\n logging.debug('using cache: {}'.format(name))\n else:\n cached = cls._serialize_and_cache(obj, json=json)\n return cached\n else:\n return cls._serialize(obj, json=json)\n\n @classmethod\n def _serialize(cls, obj, json=False):\n serializer = cls.serializers.get(obj.__class__)\n serialized = serializer(obj).data\n if json:\n serialized = JSONRenderer().render(serialized)\n return serialized\n\n @classmethod\n def _serialize_and_cache(cls, obj, json):\n # get expected object names\n name = cls._get_cache_name(obj.__class__, obj.id, json=False)\n json_name = cls._get_cache_name(obj.__class__, obj.id, json=True)\n\n # serialize data and get json-representation\n if hasattr(obj, 'optimized_for_serialization'):\n obj = obj.optimized_for_serialization()\n serialized = cls._serialize(obj, json=False)\n json_str = JSONRenderer().render(serialized)\n serialized = OrderedDict(serialized) # for pickling\n\n logging.debug('setting cache: {}'.format(name))\n cache.set_many({name: serialized, json_name: json_str})\n\n if json:\n return json_str\n else:\n return serialized\n\n @classmethod\n def add_serializer(cls, model, serializer):\n cls.serializers[model] = serializer\n\n @classmethod\n def delete_caches(cls, model, ids):\n names = [cls._get_cache_name(model, id, json=False) for id in ids]\n names.extend([cls._get_cache_name(model, id, json=True) for id in ids])\n logging.debug(\"Removing caches: {}\".format(', '.join(names)))\n cache.delete_many(names)\n\n\nclass FlatFileExporter(object):\n \"\"\"\n Base class used to generate flat-file exports of serialized data.\n \"\"\"\n def __init__(self, queryset, export_format, **kwargs):\n self.queryset = queryset\n self.export_format = export_format\n self.kwargs = kwargs\n\n if self.export_format == \"tsv\":\n self.exporter = TSVFileBuilder(**kwargs)\n elif self.export_format == \"excel\":\n self.exporter = ExcelFileBuilder(**kwargs)\n else:\n raise ValueError(\"export_format not found: {}\".format(self.export_format))\n\n def _get_header_row(self):\n raise NotImplementedError()\n\n def _get_data_rows(self):\n raise NotImplementedError()\n\n @classmethod\n def _get_tags(cls, e):\n returnValue = \"\"\n\n if (\"effects\" in e):\n \"\"\" This element is an Outcome element with an \"effects\" field \"\"\"\n effects = [tag[\"name\"] for tag in e[\"effects\"]]\n\n if (len(effects) > 0):\n returnValue = \"|{0}|\".format(\"|\".join(effects))\n elif (\"resulttags\" in e):\n \"\"\" This element is a Result element with a \"resulttags\" field \"\"\"\n resulttags = [tag[\"name\"] for tag in e[\"resulttags\"]]\n\n if (len(resulttags) > 0):\n returnValue = \"|{0}|\".format(\"|\".join(resulttags))\n\n return returnValue\n\n def build_response(self):\n header_row = self._get_header_row()\n data_rows = self._get_data_rows()\n return self.exporter.generate_response(header_row, data_rows)\n\n\nclass FlatFile(object):\n \"\"\"\n Generic file-builder object, providing an interface for generation of\n some-type of flat-file-export.\n\n Optional initialization argument:\n\n - `filename`: String filename, without extension (default: \"download\")\n \"\"\"\n\n def __init__(self, filename=\"download\", **kwargs):\n self.filename = filename\n self.kwargs = kwargs\n\n def generate_response(self, header_row, data_rows):\n self._setup()\n self._write_header_row(header_row)\n self._write_data_rows(data_rows)\n return self._django_response()\n\n def _setup(self):\n raise NotImplementedError()\n\n def _write_header_row(self, header_row):\n # `header_row` is a list of strings\n raise NotImplementedError()\n\n def _write_data_rows(self, data_rows):\n # `data_rows` is a list of lists\n raise NotImplementedError()\n\n def _django_response(self):\n raise NotImplementedError()\n\n\nclass ExcelFileBuilder(FlatFile):\n \"\"\"\n Implementation of FlatFile to generate an Excel workbook with a single\n Excel worksheet. Has one header row with minor styles applied.\n\n Optional initialization argument:\n\n - `sheet_name`: String name of worksheet (default: \"Sheet1\")\n\n \"\"\"\n\n def _setup(self):\n self.output = BytesIO()\n self.wb = xlsxwriter.Workbook(self.output)\n self._add_worksheet(sheet_name=self.kwargs.get(\"sheet_name\", \"Sheet1\"))\n\n def _add_worksheet(self, sheet_name=\"Sheet1\"):\n \"\"\"\n Create a new blank worksheet, and make sure the worksheet name is valid:\n - Make sure the name you entered does not exceed 31 characters.\n - Make sure the name does not contain any of the following characters: : \\ / ? * [ or ]\n - Make sure you did not leave the name blank.\n http://stackoverflow.com/questions/451452/\n \"\"\"\n sheet_name = re.sub(r'[\\:\\\\/\\?\\*\\[\\]]+', r'-', sheet_name)[:31]\n self.ws = self.wb.add_worksheet(sheet_name)\n\n def _write_header_row(self, header_row):\n # set formatting and freeze panes for header-row\n header_fmt = self.wb.add_format({'bold': True})\n self.ws.freeze_panes(1, 0)\n self.ncols = len(header_row)\n\n # write header-rows\n for col, val in enumerate(header_row):\n self.ws.write(0, col, val, header_fmt)\n\n def _write_data_rows(self, data_rows):\n date_format = self.wb.add_format({'num_format': 'dd/mm/yy'})\n\n def write_cell(r, c, val):\n if type(val) is bool:\n return self.ws.write_boolean(r, c, val)\n elif type(val) is datetime:\n return self.ws.write_datetime(r, c, val.replace(tzinfo=None), date_format)\n\n try:\n val = float(val)\n except:\n pass\n\n return self.ws.write(r, c, val)\n\n r = 0\n for row in data_rows:\n r += 1\n for c, val in enumerate(row):\n write_cell(r, c, val)\n\n self.ws.autofilter(0, 0, r, self.ncols - 1)\n\n def _django_response(self):\n fn = '{}.xlsx'.format(self.filename)\n self.wb.close()\n self.output.seek(0)\n response = HttpResponse(self.output.read(), content_type='application/vnd.openxmlformats-officedocument.spreadsheetml.sheet')\n response['Content-Disposition'] = 'attachment; filename=\"{}\"'.format(fn)\n return response\n\n\nclass TSVFileBuilder(FlatFile):\n \"\"\"\n Implementation of FlatFile to generate an tab-separated value file.\n \"\"\"\n\n def _setup(self):\n self.output = StringIO()\n self.tsv = csv.writer(self.output, dialect='excel-tab')\n\n def _write_header_row(self, header_row):\n self.tsv.writerow(header_row)\n\n def _write_data_rows(self, data_rows):\n self.tsv.writerows(data_rows)\n\n def _django_response(self):\n self.output.seek(0)\n response = HttpResponse(self.output, content_type='text/tab-separated-values')\n response['Content-Disposition'] = 'attachment; filename=\"{}.tsv\"'.format(self.filename)\n return response\n", "sub_path": "project/utils/helper.py", "file_name": "helper.py", "file_ext": "py", "file_size_in_byte": 10120, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "datetime.datetime.strftime", "line_number": 23, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 23, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.html.force_text", "line_number": 38, "usage_type": "call"}, {"api_name": "django.utils.html", "line_number": 38, "usage_type": "name"}, {"api_name": "django.utils.html._strip_once", "line_number": 47, "usage_type": "call"}, {"api_name": "django.utils.html", "line_number": 47, "usage_type": "name"}, {"api_name": "django.core.serializers.json.DjangoJSONEncoder", "line_number": 67, "usage_type": "name"}, {"api_name": "decimal.Decimal", "line_number": 72, "usage_type": "attribute"}, {"api_name": "django.core.cache.cache.get", "line_number": 98, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 98, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 100, "usage_type": "call"}, {"api_name": "rest_framework.renderers.JSONRenderer", "line_number": 112, "usage_type": "call"}, {"api_name": "rest_framework.renderers.JSONRenderer", "line_number": 125, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 126, "usage_type": "call"}, {"api_name": "logging.debug", "line_number": 128, "usage_type": "call"}, {"api_name": "django.core.cache.cache.set_many", "line_number": 129, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 129, "usage_type": "name"}, {"api_name": "logging.debug", "line_number": 144, "usage_type": "call"}, {"api_name": "django.core.cache.cache.delete_many", "line_number": 145, "usage_type": "call"}, {"api_name": "django.core.cache.cache", "line_number": 145, "usage_type": "name"}, {"api_name": "io.BytesIO", "line_number": 242, "usage_type": "call"}, {"api_name": "xlsxwriter.Workbook", "line_number": 243, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 254, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 273, "usage_type": "name"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 295, "usage_type": "call"}, {"api_name": "io.StringIO", "line_number": 306, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 307, "usage_type": "call"}, {"api_name": "django.shortcuts.HttpResponse", "line_number": 317, "usage_type": "call"}]} +{"seq_id": "205897404", "text": "# ConvertChessPieces.py\n# Andrew Davie andrew@taswegian.com\n\n# This tool grabs chess piece definitions from a source image and converts to .asm source code\n# The pieces are 5 pixels wide x 8 pixels deep.\n# The source image is defined as (horizontally) blank / pawn / knight / bishop / rook / queen / king\n# with the following structure:\n# line 1: white pieces on black squares\n# line 2: white pieces on white squares\n# line 3: black pieces on black squares\n# line 4: black pieces on white squares\n# thus it's a 7 x 4 chessboard grid, with pieces on top\n# The board is drawn with an 8-colour palette - colour 0 being black, and colours 1 2 and 4\n# are the colours of three successive scanlines on the '2600. Groups of 3 scanlines form an\n# 'interleaved chronocolour' (ICC) pixel. This tool reads the pixel values and converts the palette\n# colour into on/off pixels for three successive ICC sub-lines - forming the colour pixel.\n# The upshot of this, you can't actually change the colours 3,5,6,7 - they are a result of mixing\n# of colours 1, 2, 4\n# The utility produces source .asm code in the form of 3 bytes per scanline x 24 scanlines per\n# piece. The three bytes directly correspond to PF0 PF1 and PF2 on the '2600, so no shifting is\n# required - they can be directly OR'd in to existing bitmap data for display.\n# The shifting 'across' to put the piece in the correct horizontal square is done by the tool,\n# but again, the shifting is within the 3 PF bytes, so it's just a direct OR'd in draw\n# Piece definitions are written to individual files so that they can easily be located in\n# multiple banks.\n\nfrom PIL import Image\nfrom enum import Enum\n\n\nTOOL = \"; Created by ConvertChessPieces.py\\n\"\nBANK_SIZE = 2048\n\nSQUARE_WIDTH = 5\nSQUARE_HEIGHT = 8\n\n\n\nclass PieceColours(Enum):\n WHITE = 0\n BLACK = 1\n# WHITE_MARKED = 2\n# BLACK_MARKED = 3\n\n\nclass SquareColours(Enum):\n WHITE = 1\n BLACK = 0\n\n\nclass PieceTypes(Enum):\n BLANK = 0\n PAWN = 1\n KNIGHT = 2\n BISHOP = 3\n ROOK = 4\n QUEEN = 5\n KING = 6\n MARKER = 7\n PROMOTE = 8\n\n\npixel_no_to_bit_position = [\n 1 << 20,\n 1 << 21,\n 1 << 22,\n 1 << 23,\n\n 1 << 15,\n 1 << 14,\n 1 << 13,\n 1 << 12,\n 1 << 11,\n 1 << 10,\n 1 << 9,\n 1 << 8,\n\n 1 << 0,\n 1 << 1,\n 1 << 2,\n 1 << 3,\n 1 << 4,\n 1 << 5,\n 1 << 6,\n 1 << 7\n]\n\n\ndef grab(pieces_bitmap, side_colour, square_colour, piece_type, wrapper, indexer):\n\n y_start = (side_colour.value * 2 + square_colour.value) * SQUARE_HEIGHT\n x_start = piece_type.value * SQUARE_WIDTH\n\n for square_offset in range(0, 4):\n\n name = side_colour.name + \"_\" + piece_type.name + \"_on_\" + square_colour.name + \"_SQUARE_\" + str(square_offset)\n f = open(name + '.asm', 'w')\n f.write(' OPTIONAL_PAGEBREAK \"' + name + '\", 72\\n')\n f.write(' DEF ' + name + \"\\n\")\n\n lo.append(' .byte <' + name + '\\n')\n hi.append(' .byte >' + name + '\\n')\n bank.append(' .byte BANK_' + name + '\\n')\n\n equate.append('INDEX_'+name + '=' + str(indexer) + '\\n')\n indexer += 1\n\n wrapper += 1\n if wrapper % 28 == 0:\n f_includes.write(' CHECK_BANK_SIZE \"CHESS_PIECES_' + str(int((wrap / 28 - 1))) + ' ; -- full 2K\"\\n')\n f_includes.write(' NEWBANK CHESS_PIECES_' + str(int(wrap / 28)) + '\\n')\n\n f_includes.write(\" include \\\"gfx/\" + name + \".asm\\\"\\n\")\n\n pf = [[],[],[]]\n\n for y_bitmap in range(0, SQUARE_HEIGHT):\n\n icc_scanline = [0, 0, 0]\n\n for x_bitmap in range(0, SQUARE_WIDTH):\n\n pixel_icc_colour = pieces_bitmap[x_start + x_bitmap, y_start + y_bitmap]\n #if piece_type != PieceTypes.BLANK:\n # pixel_icc_colour ^= pieces_bitmap[x_bitmap, y_start + y_bitmap]\n x_pf_pixel = x_bitmap + square_offset * SQUARE_WIDTH\n\n if (pixel_icc_colour & 4) != 0:\n icc_scanline[0] |= pixel_no_to_bit_position[x_pf_pixel]\n if (pixel_icc_colour & 2) != 0:\n icc_scanline[1] |= pixel_no_to_bit_position[x_pf_pixel]\n if (pixel_icc_colour & 1) != 0:\n icc_scanline[2] |= pixel_no_to_bit_position[x_pf_pixel]\n\n # Now output the three scanlines' playfield bytes\n # we are not worrying about minimising ROM here - just 3 bytes/definition\n\n for scanline in range(0, 3):\n\n pf[0].append((icc_scanline[scanline] >> 16) & 0xFF)\n pf[1].append((icc_scanline[scanline] >> 8) & 0xFF)\n pf[2].append((icc_scanline[scanline]) & 0xFF)\n\n # write the three 'columns' PF0, PF1, PF2\n # columns make it easier to access/draw using Y as an index to indirect pointers to columns\n\n mangled = [[],[],[]]\n for line in range(0, 24, 3):\n mangled[0].append(pf[0][line])\n mangled[1].append(pf[1][line])\n mangled[2].append(pf[2][line])\n for line in range(0, 24, 3):\n mangled[0].append(pf[0][line+1])\n mangled[1].append(pf[1][line+1])\n mangled[2].append(pf[2][line+1])\n for line in range(0, 24, 3):\n mangled[0].append(pf[0][line+2])\n mangled[1].append(pf[1][line+2])\n mangled[2].append(pf[2][line+2])\n\n mangled2 = [[],[],[]]\n for block in range(0, 3):\n for line in range(7, -1, -1):\n mangled2[0].append(mangled[0][block*8+line])\n mangled2[1].append(mangled[1][block*8+line])\n mangled2[2].append(mangled[2][block*8+line])\n\n for playfield in range(0, 3):\n f.write(' .byte ' + ','.join(f'${x:02x}' for x in mangled2[playfield])\n + ' ;PF' + str(playfield) + '\\n')\n\n f.close()\n\nprint(\"Converting chess pieces\")\n\nim = Image.open(\"pieces.gif\")\npix = im.load()\n\nf_includes = open('../piece_includes.asm', 'w')\nf_includes.write(TOOL)\n\nf_vector = open('../piece_vectors.asm', 'w')\nf_vector.write(TOOL)\n\nwrap = 0\n\nlo = []\nhi = []\nbank = []\nequate = []\n\nindex = 0\n\nfor side in PieceColours:\n for square in SquareColours:\n for piece in PieceTypes:\n grab(pix, side, square, piece, wrap, index)\n index += 4\n\nf_vector.write(' DEF PIECE_VECTOR_LO\\n')\nfor low_ptr in lo:\n f_vector.write(low_ptr)\n\nf_vector.write(' DEF PIECE_VECTOR_HI\\n')\nfor high_ptr in hi:\n f_vector.write(high_ptr)\n\nf_vector.write(' DEF PIECE_VECTOR_BANK\\n')\nfor bank_ptr in bank:\n f_vector.write(bank_ptr)\n\nf_vector.write('\\n; piece index equates...\\n')\nfor equ in equate:\n f_vector.write(equ)\n\nf_vector.close()\n", "sub_path": "gfx/ConvertChessPieces.py", "file_name": "ConvertChessPieces.py", "file_ext": "py", "file_size_in_byte": 6632, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "enum.Enum", "line_number": 39, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 46, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 51, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 176, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 176, "usage_type": "name"}]} +{"seq_id": "368609456", "text": "import errno\nimport os\nimport re\nimport socket\nimport sys\nfrom datetime import datetime\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand, CommandError\nfrom django.core.servers.basehttp import (\n WSGIServer, get_internal_wsgi_application, run,\n)\nfrom django.utils import autoreload\nfrom django.debug import MY\n\n\nnaiveip_re = re.compile(r\"\"\"^(?:\n(?P\n (?P\\d{1,3}(?:\\.\\d{1,3}){3}) | # IPv4 address\n (?P\\[[a-fA-F0-9:]+\\]) | # IPv6 address\n (?P[a-zA-Z0-9-]+(?:\\.[a-zA-Z0-9-]+)*) # FQDN\n):)?(?P\\d+)$\"\"\", re.X)\n\n\nclass Command(BaseCommand):\n \"\"\"\n @function: 继承BaseCommand, 添加自定义参数, 重写handle方法, 实现webserver\n \"\"\"\n help = \"Starts a lightweight Web server for development.\"\n\n # Validation is called explicitly each time the server is reloaded.\n requires_system_checks = False\n leave_locale_alone = True\n\n default_addr = '127.0.0.1'\n default_addr_ipv6 = '::1'\n default_port = '8000'\n protocol = 'http'\n server_cls = WSGIServer\n\n def add_arguments(self, parser):\n \"\"\"重写add_arugments方法\"\"\"\n # 指定runserver的host/port\n parser.add_argument(\n 'addrport', nargs='?',\n help='Optional port number, or ipaddr:port'\n )\n # 是否支持ipv6\n parser.add_argument(\n '--ipv6', '-6', action='store_true', dest='use_ipv6',\n help='Tells Django to use an IPv6 address.',\n )\n # 是否启用多线程(每来一个用户, 创建一个线程)\n parser.add_argument(\n '--nothreading', action='store_false', dest='use_threading',\n help='Tells Django to NOT use threading.',\n )\n # 是否自动重载\n parser.add_argument(\n '--noreload', action='store_false', dest='use_reloader',\n help='Tells Django to NOT use the auto-reloader.',\n )\n\n def execute(self, *args, **options):\n if options['no_color']:\n # We rely on the environment because it's currently the only\n # way to reach WSGIRequestHandler. This seems an acceptable\n # compromise considering `runserver` runs indefinitely.\n os.environ[\"DJANGO_COLORS\"] = \"nocolor\"\n super().execute(*args, **options)\n\n def get_handler(self, *args, **options):\n \"\"\"Return the default WSGI handler for the runner.\"\"\"\n # 返回一个\"可调用\"的实现 WSGI 协议的实例, 用于处理Request/Response\n return get_internal_wsgi_application()\n\n def handle(self, *args, **options):\n # 1 开启DEBUG或者配置ALLOW_HOSTS, 其中默认端口为8000\n if not settings.DEBUG and not settings.ALLOWED_HOSTS:\n raise CommandError('You must set settings.ALLOWED_HOSTS if DEBUG is False.')\n\n # 2 host/port配置\n self.use_ipv6 = options['use_ipv6']\n if self.use_ipv6 and not socket.has_ipv6:\n raise CommandError('Your Python does not support IPv6.')\n self._raw_ipv6 = False\n if not options['addrport']:\n self.addr = ''\n self.port = self.default_port\n else:\n m = re.match(naiveip_re, options['addrport'])\n if m is None:\n raise CommandError('\"%s\" is not a valid port number '\n 'or address:port pair.' % options['addrport'])\n self.addr, _ipv4, _ipv6, _fqdn, self.port = m.groups()\n if not self.port.isdigit():\n raise CommandError(\"%r is not a valid port number.\" % self.port)\n if self.addr:\n if _ipv6:\n self.addr = self.addr[1:-1]\n self.use_ipv6 = True\n self._raw_ipv6 = True\n elif self.use_ipv6 and not _fqdn:\n raise CommandError('\"%s\" is not a valid IPv6 address.' % self.addr)\n if not self.addr:\n self.addr = self.default_addr_ipv6 if self.use_ipv6 else self.default_addr\n self._raw_ipv6 = self.use_ipv6\n # 3 调用run方法\n self.run(**options)\n\n def run(self, **options):\n \"\"\"Run the server, using the autoreloader if needed.\"\"\"\n use_reloader = options['use_reloader']\n\n # 1 main对inner_run做了一层包装, 真正的 HTTP Server处理逻辑见inner_run函数\n if use_reloader:\n # 1.0 未设置--noreload时执行, 封装 self.inner_run, 创建一个新的子线程\n MY(4, '\\n\\tReloader:', options)\n autoreload.main(self.inner_run, None, options)\n else:\n # 1.1 inner_run\n MY(4, '\\n\\tUnloader(inner):', options)\n self.inner_run(None, **options)\n\n def inner_run(self, *args, **options):\n # 1 配置工作\n # If an exception was silenced in ManagementUtility.execute in order\n # to be raised in the child process, raise it now.\n MY('Running', '\\n\\t开启运行(child线程)')\n autoreload.raise_last_exception()\n\n # django 默认开启threading, 允许在开发服务器中使用多线程\n # 可以通过选项: --nothreading关闭\n threading = options['use_threading']\n # 'shutdown_message' is a stealth option.\n shutdown_message = options.get('shutdown_message', '')\n quit_command = 'CTRL-BREAK' if sys.platform == 'win32' else 'CONTROL-C'\n\n self.stdout.write(\"Performing system checks...\\n\\n\")\n self.check(display_num_errors=True)\n # Need to check migrations here, so can't use the\n # requires_migrations_check attribute.\n self.check_migrations()\n now = datetime.now().strftime('%B %d, %Y - %X')\n self.stdout.write(now)\n self.stdout.write((\n \"Django version %(version)s, using settings %(settings)r\\n\"\n \"Starting development server at %(protocol)s://%(addr)s:%(port)s/\\n\"\n \"Quit the server with %(quit_command)s.\\n\"\n ) % {\n \"version\": self.get_version(),\n \"settings\": settings.SETTINGS_MODULE,\n \"protocol\": self.protocol,\n \"addr\": '[%s]' % self.addr if self._raw_ipv6 else self.addr,\n \"port\": self.port,\n \"quit_command\": quit_command,\n })\n\n # 2 主要处理逻辑\n try:\n # 2.1 获取handler, 实际获取get_internal_wsgi_application 返回application\n # 类型: WSGIHandler(实现WSGI 协议的对象)\n # 功能: 处理 Request/Resonse, 见WSGIHandler类(可调用)\n # 调用: 其中handler调用: handler(environ, start_response)\n handler = self.get_handler(*args, **options)\n # 2.2 runserver启动\n # 开始正式进入core.servers.basehttp.run\n # HTTP, 其中server_cls=WSGIServer使用wsgiref模块来实现HTTP\n # 功能, 类: wsgiref.simple_server.WSGIServer\n # 最终会通过handler来处理每一个请求\n run(self.addr, int(self.port), handler,\n ipv6=self.use_ipv6, threading=threading, server_cls=self.server_cls)\n except socket.error as e:\n # Use helpful error messages instead of ugly tracebacks.\n ERRORS = {\n errno.EACCES: \"You don't have permission to access that port.\",\n errno.EADDRINUSE: \"That port is already in use.\",\n errno.EADDRNOTAVAIL: \"That IP address can't be assigned to.\",\n }\n try:\n error_text = ERRORS[e.errno]\n except KeyError:\n error_text = e\n self.stderr.write(\"Error: %s\" % error_text)\n # Need to use an OS exit because sys.exit doesn't work in a thread\n os._exit(1)\n except KeyboardInterrupt:\n if shutdown_message:\n self.stdout.write(shutdown_message)\n sys.exit(0)\n\n\n# Kept for backward compatibility\nBaseRunserverCommand = Command\n", "sub_path": "core/management/commands/runserver.py", "file_name": "runserver.py", "file_ext": "py", "file_size_in_byte": 8011, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "re.compile", "line_number": 17, "usage_type": "call"}, {"api_name": "re.X", "line_number": 22, "usage_type": "attribute"}, {"api_name": "django.core.management.base.BaseCommand", "line_number": 25, "usage_type": "name"}, {"api_name": "django.core.servers.basehttp.WSGIServer", "line_number": 39, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 69, "usage_type": "attribute"}, {"api_name": "django.core.servers.basehttp.get_internal_wsgi_application", "line_number": 75, "usage_type": "call"}, {"api_name": "django.conf.settings.DEBUG", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 79, "usage_type": "name"}, {"api_name": "django.conf.settings.ALLOWED_HOSTS", "line_number": 79, "usage_type": "attribute"}, {"api_name": "django.core.management.base.CommandError", "line_number": 80, "usage_type": "call"}, {"api_name": "socket.has_ipv6", "line_number": 84, "usage_type": "attribute"}, {"api_name": "django.core.management.base.CommandError", "line_number": 85, "usage_type": "call"}, {"api_name": "re.match", "line_number": 91, "usage_type": "call"}, {"api_name": "django.core.management.base.CommandError", "line_number": 93, "usage_type": "call"}, {"api_name": "django.core.management.base.CommandError", "line_number": 97, "usage_type": "call"}, {"api_name": "django.core.management.base.CommandError", "line_number": 104, "usage_type": "call"}, {"api_name": "django.debug.MY", "line_number": 118, "usage_type": "call"}, {"api_name": "django.utils.autoreload.main", "line_number": 119, "usage_type": "call"}, {"api_name": "django.utils.autoreload", "line_number": 119, "usage_type": "name"}, {"api_name": "django.debug.MY", "line_number": 122, "usage_type": "call"}, {"api_name": "django.debug.MY", "line_number": 129, "usage_type": "call"}, {"api_name": "django.utils.autoreload.raise_last_exception", "line_number": 130, "usage_type": "call"}, {"api_name": "django.utils.autoreload", "line_number": 130, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 137, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 144, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 144, "usage_type": "name"}, {"api_name": "django.conf.settings.SETTINGS_MODULE", "line_number": 152, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 152, "usage_type": "name"}, {"api_name": "django.core.servers.basehttp.run", "line_number": 171, "usage_type": "call"}, {"api_name": "socket.error", "line_number": 173, "usage_type": "attribute"}, {"api_name": "errno.EACCES", "line_number": 176, "usage_type": "attribute"}, {"api_name": "errno.EADDRINUSE", "line_number": 177, "usage_type": "attribute"}, {"api_name": "errno.EADDRNOTAVAIL", "line_number": 178, "usage_type": "attribute"}, {"api_name": "os._exit", "line_number": 186, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 190, "usage_type": "call"}]} +{"seq_id": "276875936", "text": "from azure.keyvault.secrets import SecretClient\nfrom azure.identity import DefaultAzureCredential, SharedTokenCacheCredential\n\ntry:\n KVUri = \"https://test-vault-234567876354.vault.azure.net\"\n credential = DefaultAzureCredential()\n client = SecretClient(vault_url=KVUri, credential=credential)\n secretName = \"test\"\n retrieved_secret = client.get_secret(secretName)\n print(retrieved_secret.value)\nexcept Exception as ex:\n print(ex)", "sub_path": "azurekeyvault-noncloud/python/application.py", "file_name": "application.py", "file_ext": "py", "file_size_in_byte": 450, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "azure.identity.DefaultAzureCredential", "line_number": 6, "usage_type": "call"}, {"api_name": "azure.keyvault.secrets.SecretClient", "line_number": 7, "usage_type": "call"}]} +{"seq_id": "119126224", "text": "import sys\r\nimport argparse\r\nimport util\r\n\r\nif __name__ == \"__main__\":\r\n print(\"riff is for functions\")\r\n\r\n print(\"Domino is FaaS Acceptance Test Suite for riff and PFS on Windows\\n\")\r\n\r\n parser = argparse.ArgumentParser()\r\n parser.add_argument(\"--pfs\", help=\"test using pfs CLI\", action=\"store_true\")\r\n parser.add_argument(\"--manifest\", help=\"the manifest to test with\", type=str)\r\n parser.add_argument(\"--skip-install\", help=\"whether to skip the system install/uninstall\", action=\"store_true\")\r\n args = parser.parse_args()\r\n\r\n util.skip_install = args.skip_install\r\n if args.pfs:\r\n util.cli = \"pfs\"\r\n else:\r\n util.cli = \"riff\"\r\n if args.manifest is None or len(args.manifest) <= 0:\r\n if args.pfs:\r\n raise Exception(\"A manifest must be provided for PFS\")\r\n util.manifest = \"stable\"\r\n else:\r\n util.manifest = args.manifest\r\n\r\n import setup, teardown, functions, eventing\r\n setup.run()\r\n functions.run()\r\n eventing.run()\r\n teardown.run()\r\n\r\n print(\"DONE!\")\r\n", "sub_path": "domino.py", "file_name": "domino.py", "file_ext": "py", "file_size_in_byte": 1057, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 10, "usage_type": "call"}, {"api_name": "util.skip_install", "line_number": 16, "usage_type": "attribute"}, {"api_name": "util.cli", "line_number": 18, "usage_type": "attribute"}, {"api_name": "util.cli", "line_number": 20, "usage_type": "attribute"}, {"api_name": "util.manifest", "line_number": 24, "usage_type": "attribute"}, {"api_name": "util.manifest", "line_number": 26, "usage_type": "attribute"}, {"api_name": "setup.run", "line_number": 29, "usage_type": "call"}, {"api_name": "functions.run", "line_number": 30, "usage_type": "call"}, {"api_name": "eventing.run", "line_number": 31, "usage_type": "call"}, {"api_name": "teardown.run", "line_number": 32, "usage_type": "call"}]} +{"seq_id": "100389381", "text": "from django.contrib.contenttypes.models import ContentType\nfrom django.test import tag\n\nfrom lego.apps.comments.models import Comment\nfrom lego.apps.events.models import Event\nfrom lego.apps.feed.feed_handlers import CommentHandler\nfrom lego.apps.feed.feeds.personal_feed import PersonalFeed\nfrom lego.apps.feed.feeds.user_feed import UserFeed\nfrom lego.apps.feed.tests.feed_test_base import FeedTestBase\nfrom lego.apps.followers.models import FollowEvent\nfrom lego.apps.users.models import User\nfrom lego.utils.content_types import instance_to_string\n\n\n@tag('feed')\nclass TestCommentHandler(FeedTestBase):\n fixtures = [\n 'test_abakus_groups.yaml', 'test_users.yaml', 'test_articles.yaml', 'test_comments.yaml',\n 'test_companies.yaml', 'test_events.yaml', 'test_followevent.yaml'\n ]\n\n def setUp(self):\n self.comments = Comment.objects.all()\n self.handler = CommentHandler()\n self.users = User.objects.all()\n\n self.comment1 = self.comments[0]\n self.comment2 = self.comments[1]\n self.feed = UserFeed(self.comment1.created_by.pk)\n\n def test_duplicate_create(self):\n self.handler.handle_create(self.comment1)\n self.assertEqual(self.activity_count(self.feed), 1)\n\n self.handler.handle_create(self.comment1)\n self.assertEqual(self.activity_count(self.feed), 1)\n\n def test_comment_delete(self):\n self.handler.handle_create(self.comment1)\n self.handler.handle_create(self.comment2)\n self.assertEqual(self.activity_count(self.feed), 2)\n\n self.handler.handle_delete(self.comment1)\n self.assertEqual(self.activity_count(self.feed), 1)\n\n self.handler.handle_delete(self.comment1)\n self.assertEqual(self.activity_count(self.feed), 1)\n\n self.handler.handle_delete(self.comment2)\n self.assertEqual(len(self.feed[:]), 0)\n\n def test_extra_context(self):\n self.handler.handle_create(self.comment1)\n\n activity = self.all_activities(self.feed)[0]\n self.assertIn('content', activity.extra_context)\n self.assertEqual(activity.extra_context['content'], self.comment1.text)\n\n def test_comment_on_followed_event(self):\n follow = FollowEvent.objects.filter(pk=2).first()\n\n comment = Comment.objects.filter(\n content_type=ContentType.objects.get_for_model(Event), object_id=follow.target.id\n ).first()\n\n follower_feed = PersonalFeed(follow.follower.id)\n creator_feed = PersonalFeed(comment.created_by.id)\n\n self.assertIsNotNone(comment)\n\n self.assertEqual(self.activity_count(creator_feed), 0)\n self.assertEqual(self.activity_count(follower_feed), 0)\n self.handler.handle_create(comment)\n self.assertEqual(self.activity_count(creator_feed), 0)\n self.assertEqual(self.activity_count(follower_feed), 1)\n activity = self.all_activities(follower_feed)[0]\n self.assertEqual(activity.object, instance_to_string(comment))\n\n self.handler.handle_delete(comment)\n self.assertEqual(self.activity_count(follower_feed), 0)\n self.assertEqual(self.activity_count(creator_feed), 0)\n", "sub_path": "lego/apps/feed/tests/test_comment_handler.py", "file_name": "test_comment_handler.py", "file_ext": "py", "file_size_in_byte": 3158, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "lego.apps.feed.tests.feed_test_base.FeedTestBase", "line_number": 16, "usage_type": "name"}, {"api_name": "lego.apps.comments.models.Comment.objects.all", "line_number": 23, "usage_type": "call"}, {"api_name": "lego.apps.comments.models.Comment.objects", "line_number": 23, "usage_type": "attribute"}, {"api_name": "lego.apps.comments.models.Comment", "line_number": 23, "usage_type": "name"}, {"api_name": "lego.apps.feed.feed_handlers.CommentHandler", "line_number": 24, "usage_type": "call"}, {"api_name": "lego.apps.users.models.User.objects.all", "line_number": 25, "usage_type": "call"}, {"api_name": "lego.apps.users.models.User.objects", "line_number": 25, "usage_type": "attribute"}, {"api_name": "lego.apps.users.models.User", "line_number": 25, "usage_type": "name"}, {"api_name": "lego.apps.feed.feeds.user_feed.UserFeed", "line_number": 29, "usage_type": "call"}, {"api_name": "lego.apps.followers.models.FollowEvent.objects.filter", "line_number": 60, "usage_type": "call"}, {"api_name": "lego.apps.followers.models.FollowEvent.objects", "line_number": 60, "usage_type": "attribute"}, {"api_name": "lego.apps.followers.models.FollowEvent", "line_number": 60, "usage_type": "name"}, {"api_name": "lego.apps.comments.models.Comment.objects.filter", "line_number": 62, "usage_type": "call"}, {"api_name": "lego.apps.comments.models.Comment.objects", "line_number": 62, "usage_type": "attribute"}, {"api_name": "lego.apps.comments.models.Comment", "line_number": 62, "usage_type": "name"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "line_number": 63, "usage_type": "call"}, {"api_name": "lego.apps.events.models.Event", "line_number": 63, "usage_type": "argument"}, {"api_name": "django.contrib.contenttypes.models.ContentType.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "django.contrib.contenttypes.models.ContentType", "line_number": 63, "usage_type": "name"}, {"api_name": "lego.apps.feed.feeds.personal_feed.PersonalFeed", "line_number": 66, "usage_type": "call"}, {"api_name": "lego.apps.feed.feeds.personal_feed.PersonalFeed", "line_number": 67, "usage_type": "call"}, {"api_name": "lego.utils.content_types.instance_to_string", "line_number": 77, "usage_type": "call"}, {"api_name": "django.test.tag", "line_number": 15, "usage_type": "call"}]} +{"seq_id": "468056094", "text": "import sys,bpy,bpy_extras\nfrom os import mkdir\nfrom os.path import join\nannotation_folder = r\"C:\\project_data\\interiors_2\\renders\\room_10\"\nbpy.context.scene.render.image_settings.file_format = 'JPEG'\nstart_frame = 520\ntotal_frames = 1000\nobject_range = []\ncamera_object = []\n\ni = 0\nfor obj in bpy.data.objects:\n print(obj.name,obj.type)\n if obj.type == 'CAMERA':\n camera_object.append(i)\n elif obj.type == 'MESH':\n object_range.append(i)\n i = i+1\noriginal_out = sys.stdout\nfor current_frame in range(start_frame-1,total_frames):\n frame_folder = join(annotation_folder,'frame_'+str(current_frame+1))\n render_path = join(frame_folder,'frame.JPEG')\n mkdir(frame_folder) \n scene = bpy.context.scene\n scene.frame_current = current_frame \n scene.render.filepath = render_path\n bpy.ops.render.render(write_still=True) \n obj = bpy.data.objects[camera_object[0]]\n for i in object_range:\n current_object = bpy.data.objects[i]\n if current_object.type == 'MESH':\n object_annotation_file = join(frame_folder,current_object.name+'_'+current_object.type+'.txt')\n with open(object_annotation_file,'w') as f:\n sys.stdout = f\n for j in range(len(current_object.data.vertices)):\n co = current_object.matrix_world @ current_object.data.vertices[j].co\n co_2d = bpy_extras.object_utils.world_to_camera_view(scene,obj,co)\n print(\"\\t\",round(co_2d.x,6),round(co_2d.y,6),round(co_2d.z,6))\nsys.stdout = original_out\nprint(\"finished rendering!\\n\")\n \n \n ", "sub_path": "frame_vertex_generator.py", "file_name": "frame_vertex_generator.py", "file_ext": "py", "file_size_in_byte": 1613, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "bpy.context", "line_number": 5, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 12, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 21, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 22, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 23, "usage_type": "call"}, {"api_name": "bpy.context", "line_number": 24, "usage_type": "attribute"}, {"api_name": "bpy.ops.render.render", "line_number": 27, "usage_type": "call"}, {"api_name": "bpy.ops", "line_number": 27, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 28, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 32, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 34, "usage_type": "attribute"}, {"api_name": "bpy_extras.object_utils.world_to_camera_view", "line_number": 37, "usage_type": "call"}, {"api_name": "bpy_extras.object_utils", "line_number": 37, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 39, "usage_type": "attribute"}]} +{"seq_id": "231461853", "text": "import csv\nimport requests\nimport time\nfrom bs4 import BeautifulSoup\nfrom soupsieve import select\nimport itertools\n\n\n# emidasにアクセス\nres = requests.get('https://www.nc-net.or.jp/search/search/?w=%E6%B5%B7%E5%A4%96&x=0&y=0&sxf=&sxt=&syf=&syt=&szf=&szt=&en=&es=&eq=&e=&pno=1')\nsoup = BeautifulSoup(res.content, 'lxml')\n\ntime.sleep(3)\n\n# ベースとなるURL\nbase_url = \"https://www.nc-net.or.jp/\"\n\n# 空の配列作成\ncomp_a = []\n\n# 企業個別リンク取得\ncomp_a.append(soup.select(\".ttl-h3-03 a[href]\"))\n\n\n# ページhref取得\nfor i in range(70):\n pages = soup.select('.nav-page-skip-01 li a[href]')\n next_page = pages[-1].get(\"href\")\n res = requests.get(base_url + next_page)\n soup = BeautifulSoup(res.content, 'lxml', from_encoding='utf-8')\n comp_a.append(soup.select(\".ttl-h3-03 a[href]\"))\n\n# # ��の配列作成\ncomp_hrefs = []\ncomp_urls = []\ncsv_datas = [['No', '企業名', '主要三品目①', '主要三品目②', '主要三品目③', '企業HP']]\n\n# # 取得した企業の個別リンク分処理を繰り返す(二次元配列→一次元化)\nfor comp_href in list(itertools.chain.from_iterable(comp_a)):\n# # # 配列にhref追加\n comp_hrefs.append(comp_href.get(\"href\"))\n\nfor comp_href in comp_hrefs:\n# # # 企業個別URL生成\n comp_urls.append(base_url + comp_href)\n\ni = 1\n\n\nfor comp_url in comp_urls:\n comp_url1 = requests.get(comp_url)\n soup = BeautifulSoup(comp_url1.content, 'lxml', from_encoding='utf-8')\n # 企業名、業種、採用情報URL取得\n comp_na = soup.select(\".tbl-data-01 td\")[0].text if soup.select(\".tbl-data-01 td\") else \"\"\n target = '('\n idx = comp_na.find(target)\n comp_name = comp_na[:idx].strip().replace(' ', '')\n # print(comp_name)\n # print(\"------------------------\")\n if soup.select(\".list-dot-01 li\"):\n main_item = soup.select(\".list-dot-01 li\") \n if len(main_item) > 0: \n main_item1 = main_item[0].text \n if len(main_item) > 1:\n main_item2 = main_item[1].text \n if len(main_item) > 2: \n main_item3 = main_item[2].text \n else:\n \"\"\n else:\n \"\"\n else:\n \"\"\n else:\n \"\"\n\n\n\n comp_a = soup.select_one(\".ttl-h1-02 a\") if soup.select_one(\".ttl-h1-02 a\") else \"\" \n comp_href = comp_a[\"href\"] if soup.select_one(\".ttl-h1-02 a\") else \"\"\n csv_datas.append([i, comp_name, main_item1, main_item2, main_item3, comp_url])\n # print(comp_name)\n i += 1\n\n\n\n# # # csvファイルを新規作成、取得した企業名等のデータを書き込む\nf = open(\"emidas_data.csv\", \"w\")\ncsvf = csv.writer(f)\nfor csv_data in csv_datas:\n csvf.writerow(csv_data)\nf.close\n\n\n\n", "sub_path": "emidas.py", "file_name": "emidas.py", "file_ext": "py", "file_size_in_byte": 2619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "requests.get", "line_number": 10, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 11, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 13, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 30, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 39, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 39, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 51, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 52, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 89, "usage_type": "call"}]} +{"seq_id": "467844738", "text": "import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt \n\n#Sigmoid function\ndef sigmoid(x):\n return 1/(1+np.exp(-x))\n\n#Change figure name\nfig = plt.gcf()\nfig.canvas.set_window_title('Logistic Regression')\n\n#Load data from dataset\ndata = pd.read_csv('Logistic Regression/dataset.csv').values\nN, d = data.shape\nx = data[:,0:d-1].reshape(-1,d-1) #Get columns 0 and 1\ny = data[:,2].reshape(-1,1) #Get column 2\n\n#Scatter data\nplt.xlabel('Salary (millions)')\nplt.ylabel('Experiences (years)')\nplt.scatter(x[:10,0], x[:10,1], c='red', edgecolors='none', s = 30, label = 'Accepted' ) #s: size of dot; edgecolor: around dot\nplt.scatter(x[10:,0], x[10:,1], c='blue', edgecolors='none', s = 30, label = 'Rejected')\nplt.legend(loc=1) #Location of label\n\n#Create x and w matrix\nx = np.hstack((np.ones((N,1)),x))\nw = np.array([0.,0.1,0.1]).reshape(-1,1)\n\nnumOfIteration = 1000\nlearning_rate = 0.01\ncost = np.zeros((numOfIteration,1))\n\nfor i in range (1, numOfIteration):\n #Compute the predict vale\n yPredict = sigmoid(np.dot(x,w))\n cost[i] = - np.sum(np.multiply(y, np.log(yPredict)) + np.multiply(1-y, np.log(1 - yPredict)))\n\n #Gradient descent\n w = w - learning_rate * np.dot(x.T, yPredict - y)\n print (cost[i])\n\n#Draw the median line\nt = 0.8\n#x from 4 to 10 => draw y based on x\nplt.plot((4,10), (-(w[0] + w[1]*4 + np.log(1/t - 1))/w[2], - (w[0] + w[1]*10 + np.log(1/t -1 ))/w[2]),color = 'green')\nplt.show()\n\n\n\n\n\n\n\n", "sub_path": "Logistic Regression/Logistic_Regression.py", "file_name": "Logistic_Regression.py", "file_ext": "py", "file_size_in_byte": 1441, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.exp", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.gcf", "line_number": 10, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 10, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "numpy.hstack", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.log", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}]} +{"seq_id": "142847009", "text": "from functools import update_wrapper\n\nfrom dagster import check\nfrom dagster.core.definitions.configurable import AnonymousConfigurableDefinition\n\nfrom .definition_config_schema import convert_user_facing_definition_config_schema\nfrom .utils import check_valid_name\n\n\nclass IntermediateStorageDefinition(AnonymousConfigurableDefinition):\n \"\"\"Defines intermediate data storage behaviors.\n\n Args:\n name (str): Name of the storage mode.\n is_persistent (bool): Whether the storage is persistent in a way that can cross process/node\n boundaries. Re-execution with, for example, the multiprocess executor, or with\n dagster-airflow, requires a persistent storage mode.\n required_resource_keys(Optional[Set[str]]): The resources that this storage needs at runtime to function.\n config_schema (Optional[ConfigSchema]): The schema for the storage's configuration.\n Configuration data passed in this schema will be made available to the\n ``intermediate_storage_creation_fn`` under ``init_context.intermediate_storage_config``.\n If not set, Dagster will accept any config provided.\n intermediate_storage_creation_fn: (Callable[[InitIntermediateStorageContext], IntermediateStorage])\n Called to construct the storage. This function should consume the init context and emit\n a :py:class:`IntermediateStorage`.\n \"\"\"\n\n def __init__(\n self,\n name,\n is_persistent,\n required_resource_keys,\n config_schema=None,\n intermediate_storage_creation_fn=None,\n description=None,\n ):\n self._name = check_valid_name(name)\n self._is_persistent = check.bool_param(is_persistent, \"is_persistent\")\n self._config_schema = convert_user_facing_definition_config_schema(config_schema)\n self._intermediate_storage_creation_fn = check.opt_callable_param(\n intermediate_storage_creation_fn, \"intermediate_storage_creation_fn\"\n )\n self._required_resource_keys = frozenset(\n check.set_param(\n required_resource_keys if required_resource_keys else set(),\n \"required_resource_keys\",\n of_type=str,\n )\n )\n self._description = check.opt_str_param(description, \"description\")\n\n @property\n def name(self):\n return self._name\n\n @property\n def description(self):\n return self._description\n\n @property\n def is_persistent(self):\n return self._is_persistent\n\n @property\n def config_schema(self):\n return self._config_schema\n\n @property\n def intermediate_storage_creation_fn(self):\n return self._intermediate_storage_creation_fn\n\n @property\n def required_resource_keys(self):\n return self._required_resource_keys\n\n def copy_for_configured(self, description, config_schema, _):\n return IntermediateStorageDefinition(\n name=self.name,\n is_persistent=self.is_persistent,\n required_resource_keys=self.required_resource_keys,\n config_schema=config_schema,\n intermediate_storage_creation_fn=self.intermediate_storage_creation_fn,\n description=description or self.description,\n )\n\n\ndef intermediate_storage(\n required_resource_keys=None, name=None, is_persistent=True, config_schema=None\n):\n \"\"\"Creates an intermediate storage definition\n\n The decorated function will be passed as the ``intermediate_storage_creation_fn`` to a\n :py:class:`IntermediateStorageDefinition`.\n\n Args:\n name (str): The name of the intermediate storage.\n is_persistent (bool): Whether the storage is persistent in a way that can cross process/node\n boundaries. Re-execution with, for example, the multiprocess executor, or with\n dagster-airflow, requires a persistent storage mode.\n required_resource_keys (Optional[Set[str]]):\n The resources that this storage needs at runtime to function.\n config_schema (Optional[ConfigSchema]): The schema for the config. Configuration data available in\n `init_context.intermediate_storage_config`.\n \"\"\"\n\n if callable(name):\n check.invariant(is_persistent is True)\n check.invariant(config_schema is None)\n check.invariant(required_resource_keys is None)\n return _IntermediateStorageDecoratorCallable()(name)\n\n return _IntermediateStorageDecoratorCallable(\n name=name,\n is_persistent=is_persistent,\n config_schema=config_schema,\n required_resource_keys=required_resource_keys,\n )\n\n\nclass _IntermediateStorageDecoratorCallable:\n def __init__(\n self, name=None, is_persistent=True, config_schema=None, required_resource_keys=None\n ):\n self.name = check.opt_str_param(name, \"name\")\n self.is_persistent = check.bool_param(is_persistent, \"is_persistent\")\n self.config_schema = config_schema # will be checked in definition\n self.required_resource_keys = check.opt_set_param(\n required_resource_keys, \"required_resource_keys\", of_type=str\n )\n\n def __call__(self, fn):\n check.callable_param(fn, \"fn\")\n\n if not self.name:\n self.name = fn.__name__\n\n storage_def = IntermediateStorageDefinition(\n name=self.name,\n is_persistent=self.is_persistent,\n config_schema=self.config_schema,\n intermediate_storage_creation_fn=fn,\n required_resource_keys=self.required_resource_keys,\n )\n\n update_wrapper(storage_def, wrapped=fn)\n\n return storage_def\n", "sub_path": "python_modules/dagster/dagster/core/definitions/intermediate_storage.py", "file_name": "intermediate_storage.py", "file_ext": "py", "file_size_in_byte": 5677, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "dagster.core.definitions.configurable.AnonymousConfigurableDefinition", "line_number": 10, "usage_type": "name"}, {"api_name": "utils.check_valid_name", "line_number": 37, "usage_type": "call"}, {"api_name": "dagster.check.bool_param", "line_number": 38, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 38, "usage_type": "name"}, {"api_name": "definition_config_schema.convert_user_facing_definition_config_schema", "line_number": 39, "usage_type": "call"}, {"api_name": "dagster.check.opt_callable_param", "line_number": 40, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 40, "usage_type": "name"}, {"api_name": "dagster.check.set_param", "line_number": 44, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 44, "usage_type": "name"}, {"api_name": "dagster.check.opt_str_param", "line_number": 50, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 50, "usage_type": "name"}, {"api_name": "dagster.check.invariant", "line_number": 107, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 107, "usage_type": "name"}, {"api_name": "dagster.check.invariant", "line_number": 108, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 108, "usage_type": "name"}, {"api_name": "dagster.check.invariant", "line_number": 109, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 109, "usage_type": "name"}, {"api_name": "dagster.check.opt_str_param", "line_number": 124, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 124, "usage_type": "name"}, {"api_name": "dagster.check.bool_param", "line_number": 125, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 125, "usage_type": "name"}, {"api_name": "dagster.check.opt_set_param", "line_number": 127, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 127, "usage_type": "name"}, {"api_name": "dagster.check.callable_param", "line_number": 132, "usage_type": "call"}, {"api_name": "dagster.check", "line_number": 132, "usage_type": "name"}, {"api_name": "functools.update_wrapper", "line_number": 145, "usage_type": "call"}]} +{"seq_id": "72768401", "text": "#/usr/bin/env python\n\nimport subprocess\nimport ctypes\nimport numpy as np\nimport itertools\nfrom functools import reduce\nimport os\nfrom timeit import default_timer as timer\n\nfrom substation.transformer import layer_norm_backward_weights\n\nfrom generator import generate_bsb\n\n\ndef ref_bsb(inp, scale, dout):\n dinp = dout * scale\n dinp2 = np.einsum(\"bji,i->bji\", dout, scale)\n assert(np.allclose(dinp, dinp2, atol=1e-2, rtol=1e-1))\n dscale, dbias = layer_norm_backward_weights(dout, inp, True, True)\n return dinp, dscale, dbias\n\ndef test_bsb():\n \n dims = { 'B': 2, 'J': 32, 'N': 8 }\n #dims = { 'B': 8, 'J': 512, 'N': 16 * 64 }\n reduce_dim = 'B'\n warp_reduce_dim = 'J'\n non_reduce_dim = 'N'\n \n base_layout = \"\".join(dims.keys())\n \n generate_bsb(dims, reduce_dim, warp_reduce_dim, 'bsb_test.so')\n \n lib = ctypes.cdll.LoadLibrary('./bsb_test.so')\n \n inp = np.ascontiguousarray(np.random.rand(*dims.values()), dtype='float16')\n scale = np.ascontiguousarray(np.random.rand(dims[non_reduce_dim]), dtype='float16')\n dout = np.ascontiguousarray(np.random.rand(*dims.values()), dtype='float16')\n dscale = np.ascontiguousarray(np.random.rand(dims[non_reduce_dim]), dtype='float16')\n dbias = np.ascontiguousarray(np.random.rand(dims[non_reduce_dim]), dtype='float16')\n \n dropout_probability = 0\n ref_din, ref_dscale, ref_dbias = ref_bsb(inp, scale, dout)\n \n for dims_permutation_in in itertools.permutations(dims):\n for dims_permutation_din in itertools.permutations(dims):\n for dims_permutation_dout in itertools.permutations(dims):\n in_label = \"\".join(dims_permutation_in)\n din_label = \"\".join(dims_permutation_din)\n dout_label = \"\".join(dims_permutation_dout)\n \n func_name = 'temp_%s_%s_%s' % (in_label, din_label, dout_label)\n \n bsb = getattr(lib, func_name)\n \n din_array_shape = list(map(lambda x: dims[x], dims_permutation_din))\n din = np.ascontiguousarray(np.zeros(din_array_shape), dtype='float16')\n \n temp_inp = np.ascontiguousarray(np.einsum(base_layout + \"->\" + in_label, inp), dtype='float16')\n temp_dout = np.ascontiguousarray(np.einsum(base_layout + \"->\" + dout_label, dout), dtype='float16')\n \n bsb(ctypes.c_void_p(temp_inp.ctypes.data),\n ctypes.c_void_p(scale.ctypes.data),\n ctypes.c_void_p(temp_dout.ctypes.data),\n ctypes.c_void_p(din.ctypes.data),\n ctypes.c_void_p(dscale.ctypes.data),\n ctypes.c_void_p(dbias.ctypes.data))\n \n din = np.einsum(din_label + \"->\" + base_layout, din)\n \n layouts = \"IN %s DIN %s DOUT %s\" % (in_label, din_label, dout_label)\n \n if not np.allclose(dbias, ref_dbias, atol=1e-2, rtol=1e-1):\n print(\"DBIAS\")\n raise Exception(layouts)\n elif not np.allclose(dscale, ref_dscale, atol=1e-2, rtol=1e-1):\n print(\"DSCALE\")\n raise Exception(layouts)\n elif not np.allclose(din, ref_din, atol=1e-2, rtol=1e-1):\n print(din)\n print('------------------------')\n print(ref_din)\n print(\"DIN\")\n raise Exception(layouts)\n else:\n print(\"OK\", layouts)\n\n\nif __name__ == '__main__':\n test_bsb()\n print(\"All tests are passed\")\n", "sub_path": "pytorch_module/test_bsb.py", "file_name": "test_bsb.py", "file_ext": "py", "file_size_in_byte": 3687, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.einsum", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 19, "usage_type": "call"}, {"api_name": "substation.transformer.layer_norm_backward_weights", "line_number": 20, "usage_type": "call"}, {"api_name": "generator.generate_bsb", "line_number": 33, "usage_type": "call"}, {"api_name": "ctypes.cdll.LoadLibrary", "line_number": 35, "usage_type": "call"}, {"api_name": "ctypes.cdll", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 37, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 39, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.ascontiguousarray", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "itertools.permutations", "line_number": 46, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 47, "usage_type": "call"}, {"api_name": "itertools.permutations", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.einsum", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.ascontiguousarray", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.einsum", "line_number": 61, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 63, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 64, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 65, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 66, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 67, "usage_type": "call"}, {"api_name": "ctypes.c_void_p", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.einsum", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 77, "usage_type": "call"}, {"api_name": "numpy.allclose", "line_number": 80, "usage_type": "call"}]} +{"seq_id": "428823770", "text": "import os\nfrom datetime import datetime\n\nfrom flask import render_template, request\nfrom flask_login import login_required\nfrom peewee import RawQuery\n\nfrom app import app, logger\nfrom app.models import Mail\n\n\n@app.route('/mails')\n@login_required\ndef mails_index():\n dt_start = request.args.get('dt_start') or datetime.now().strftime('%Y-%m-%d')\n dt_end = request.args.get('dt_end') or datetime.now().strftime('%Y-%m-%d')\n senders = [sender.mail_from for sender in Mail.select(Mail.mail_from).distinct() if sender.mail_from]\n mail_reports = {}\n summary = {}\n\n for sender in senders:\n operators = RawQuery(model=Mail, query='''\n SELECT\n count(distinct name), operator\n FROM \"mail\"\n where dt between '{dt_start}' and '{dt_end}' \n and mail_from = '{mail_from}'\n group by operator;'''.format(dt_start=dt_start, dt_end=dt_end, mail_from=sender)).execute()\n summary[sender] = 0\n\n for operator in operators:\n summary[sender] += operator.count\n\n mail_reports[sender] = operators\n\n senders.insert(0, 'Все ящики')\n\n operators = RawQuery(model=Mail, query='''\n SELECT\n count(distinct name), operator\n FROM \"mail\"\n where dt between '{dt_start}' and '{dt_end}' \n group by operator;'''.format(dt_start=dt_start, dt_end=dt_end)).execute()\n summary['Все ящики'] = 0\n\n for operator in operators:\n summary['Все ящики'] += operator.count\n\n mail_reports['Все ящики'] = operators\n\n return render_template('mails/index.haml', filter=request.args,\n mail_reports=mail_reports, summary=summary, senders=senders)\n", "sub_path": "app/views/mails.py", "file_name": "mails.py", "file_ext": "py", "file_size_in_byte": 1776, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.request.args.get", "line_number": 15, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 15, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 15, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 15, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 16, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "app.models.Mail.select", "line_number": 17, "usage_type": "call"}, {"api_name": "app.models.Mail", "line_number": 17, "usage_type": "name"}, {"api_name": "app.models.Mail.mail_from", "line_number": 17, "usage_type": "attribute"}, {"api_name": "peewee.RawQuery", "line_number": 22, "usage_type": "call"}, {"api_name": "app.models.Mail", "line_number": 22, "usage_type": "name"}, {"api_name": "peewee.RawQuery", "line_number": 38, "usage_type": "call"}, {"api_name": "app.models.Mail", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 51, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "app.app.route", "line_number": 12, "usage_type": "call"}, {"api_name": "app.app", "line_number": 12, "usage_type": "name"}, {"api_name": "flask_login.login_required", "line_number": 13, "usage_type": "name"}]} +{"seq_id": "174844077", "text": "import sys\nimport unittest\nimport numpy as np\n\nfrom unittest.mock import MagicMock\n\nfrom PyQt5 import QtGui, QtCore, QtWidgets\nfrom PyQt5.QtTest import QTest, QSignalSpy\n\n# set up import paths\nimport path_prepare\n\nimport sas.qtgui.Utilities.ObjectLibrary as ObjectLibrary\nimport sas.qtgui.Utilities.GuiUtils as GuiUtils\nfrom sas.qtgui.Perspectives.Fitting import FittingUtilities\nfrom sas.qtgui.Plotting.PlotterData import Data1D\n\n# Local\nfrom sas.qtgui.Perspectives.Fitting.ConstraintWidget import ConstraintWidget\nfrom sas.qtgui.Perspectives.Fitting.Constraint import Constraint\nfrom sas.qtgui.Perspectives.Fitting.FittingPerspective import FittingWindow\nfrom sas.qtgui.Perspectives.Fitting.FittingWidget import FittingWidget\n\nif not QtWidgets.QApplication.instance():\n app = QtWidgets.QApplication(sys.argv)\n\nclass ConstraintWidgetTest(unittest.TestCase):\n '''Test the ConstraintWidget dialog'''\n def setUp(self):\n '''Create ConstraintWidget dialog'''\n class dummy_manager(object):\n def communicator(self):\n return GuiUtils.Communicate()\n communicate = GuiUtils.Communicate()\n\n def __init__(self):\n self._perspective = dummy_perspective()\n\n def perspective(self):\n return self._perspective\n\n class dummy_perspective(object):\n\n def __init__(self):\n self.symbol_dict = {}\n self.constraint_list = []\n self.constraint_tab = None\n\n def getActiveConstraintList(self):\n return self.constraint_list\n\n def getSymbolDictForConstraints(self):\n return self.symbol_dict\n\n def getConstraintTab(self):\n return self.constraint_tab\n\n '''Create the perspective'''\n self.perspective = FittingWindow(dummy_manager())\n ConstraintWidget.updateSignalsFromTab = MagicMock()\n\n self.widget = ConstraintWidget(parent=self.perspective)\n\n # Example constraint object\n self.constraint1 = Constraint(parent=None, param=\"scale\", value=\"7.0\",\n min=\"0.0\", max=\"inf\", func=\"M1.sld\",\n value_ex=\"M1.scale\")\n self.constraint2 = Constraint(parent=None, param=\"poop\", value=\"7.0\", min=\"0.0\", max=\"inf\", func=\"7.0\")\n\n def tearDown(self):\n '''Destroy the GUI'''\n self.widget.close()\n self.widget = None\n\n def testDefaults(self):\n '''Test the GUI in its default state'''\n self.assertIsInstance(self.widget, QtWidgets.QWidget)\n # Default title\n self.assertEqual(self.widget.windowTitle(), \"Constrained and Simultaneous Fit\")\n # Dicts\n self.assertIsInstance(self.widget.available_constraints, dict)\n self.assertIsInstance(self.widget.available_tabs, dict)\n # TableWidgets\n self.assertEqual(self.widget.tblTabList.columnCount(), 4)\n self.assertEqual(self.widget.tblConstraints.columnCount(), 1)\n # Data accept \n self.assertFalse(self.widget.acceptsData())\n # By default, the constraint table is disabled\n self.assertFalse(self.widget.tblConstraints.isEnabled())\n\n def testOnFitTypeChange(self):\n ''' test the single/batch fit switch '''\n self.widget.initializeFitList = MagicMock()\n # Assure current type is Single\n self.assertEqual(self.widget.currentType, \"FitPage\")\n # click on \"batch\"\n QTest.mouseClick(self.widget.btnBatch, QtCore.Qt.LeftButton)\n QtWidgets.QApplication.processEvents()\n # See what the current type is now\n self.assertEqual(self.widget.currentType, \"BatchPage\")\n # See if the list is getting initialized\n self.assertTrue(self.widget.initializeFitList.called)\n # Go back to single fit\n QTest.mouseClick(self.widget.btnSingle, QtCore.Qt.LeftButton)\n QtWidgets.QApplication.processEvents()\n # See what the current type is now\n self.assertEqual(self.widget.currentType, \"FitPage\")\n\n def testGetTabsForFit(self):\n ''' Test the fitting tab list '''\n self.assertEqual(self.widget.getTabsForFit(), [])\n # add one tab\n self.widget.tabs_for_fitting = {\"foo\": True}\n self.assertEqual(self.widget.getTabsForFit(), ['foo'])\n # add two tabs\n self.widget.tabs_for_fitting = {\"foo\": True, \"bar\": True}\n self.assertEqual(self.widget.getTabsForFit(), ['foo', 'bar'])\n # disable one tab\n self.widget.tabs_for_fitting = {\"foo\": False, \"bar\": True}\n self.assertEqual(self.widget.getTabsForFit(), ['bar'])\n\n def testIsTabImportable(self):\n ''' tab checks for consistency '''\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = None\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n\n self.assertFalse(self.widget.isTabImportable(None))\n self.assertFalse(self.widget.isTabImportable(\"BatchTab1\"))\n self.widget.currentType = \"Batch\"\n self.assertFalse(self.widget.isTabImportable(\"BatchTab\"))\n self.widget.currentType = \"test\"\n self.assertFalse(self.widget.isTabImportable(\"test_tab\"))\n test_tab.data_is_loaded = True\n self.assertTrue(self.widget.isTabImportable(\"test_tab\"))\n\n def testOnTabCellEdit(self):\n ''' test what happens on monicker edit '''\n # Mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = MagicMock()\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n self.widget.updateFitLine(\"test_tab\")\n\n # disable the tab\n self.widget.tblTabList.item(0, 0).setCheckState(0)\n self.assertEqual(self.widget.tabs_for_fitting[\"test_tab\"], False)\n self.assertFalse(self.widget.cmdFit.isEnabled())\n # enable the tab\n self.widget.tblTabList.item(0, 0).setCheckState(2)\n self.assertEqual(self.widget.tabs_for_fitting[\"test_tab\"], True)\n self.assertTrue(self.widget.cmdFit.isEnabled())\n\n def testUpdateFitLine(self):\n ''' See if the fit table row can be updated '''\n # mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = MagicMock()\n test_tab.kernel_module.name = \"M1\"\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n\n # Add a tab without an constraint\n self.widget.updateFitLine(\"test_tab\")\n self.assertEqual(self.widget.tblTabList.rowCount(), 1)\n # Constraint tab should be empty\n self.assertEqual(self.widget.tblConstraints.rowCount(), 0)\n\n # Add a second tab with an active constraint\n test_tab.getComplexConstraintsForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getFullConstraintNameListForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getConstraintObjectsForModel = MagicMock(\n return_value=[self.constraint1])\n self.widget.updateFitLine(\"test_tab\")\n # We should have 2 tabs in the model tab\n self.assertEqual(self.widget.tblTabList.rowCount(), 2)\n # One constraint in the constraint tab\n self.assertEqual(self.widget.tblConstraints.rowCount(), 1)\n # Constraint should be active\n self.assertEqual(self.widget.tblConstraints.item(0, 0).checkState(), 2)\n # Check the text\n self.assertEqual(self.widget.tblConstraints.item(0, 0).text(),\n test_tab.kernel_module.name +\n \":scale = \" +\n self.constraint1.func)\n # Add a tab with a non active constraint\n test_tab.getComplexConstraintsForModel = MagicMock(return_value=[])\n self.widget.updateFitLine(\"test_tab\")\n # There should be two constraints now\n self.assertEqual(self.widget.tblConstraints.rowCount(), 2)\n # Added constraint should not be checked since it isn't active\n self.assertEqual(self.widget.tblConstraints.item(1, 0).checkState(), 0)\n\n def testUpdateFitList(self):\n ''' see if the fit table can be updated '''\n # mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = MagicMock()\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n\n # Fit button should be disabled if no tabs are present\n ObjectLibrary.listObjects =MagicMock(return_value=False)\n self.widget.initializeFitList()\n self.assertEqual(self.widget.available_tabs, {})\n self.assertEqual(self.widget.available_constraints, {})\n self.assertEqual(self.widget.tblConstraints.rowCount(), 0)\n self.assertEqual(self.widget.tblTabList.rowCount(), 0)\n self.assertFalse(self.widget.cmdFit.isEnabled())\n\n # Add a tab\n self.widget.isTabImportable = MagicMock(return_value=True)\n ObjectLibrary.listObjects = MagicMock(return_value=[test_tab])\n self.widget.updateFitLine = MagicMock()\n self.widget.updateSignalsFromTab = MagicMock()\n self.widget.initializeFitList()\n self.widget.updateFitLine.assert_called_once()\n self.widget.updateSignalsFromTab.assert_called_once()\n self.assertTrue(self.widget.cmdFit.isEnabled())\n\n # Check if the tab list gets ordered\n self.widget.isTabImportable = MagicMock(return_value=True)\n ObjectLibrary.listObjects = MagicMock(return_value=[test_tab])\n self.widget.updateFitLine = MagicMock()\n self.widget.updateSignalsFromTab = MagicMock()\n self.widget._row_order = [test_tab]\n self.widget.orderedSublist = MagicMock()\n self.widget.initializeFitList()\n self.widget.orderedSublist.assert_called_with([test_tab], [test_tab])\n\n\n def testOnAcceptConstraint(self):\n ''' test if a constraint can be added '''\n # mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.addConstraintToRow = MagicMock()\n test_tab.getRowFromName = MagicMock(return_value=1)\n test_tab.changeCheckboxStatus = MagicMock()\n\n # mock the getObjectByName method\n self.widget.getObjectByName = MagicMock(return_value=test_tab)\n\n # add a constraint\n constraint_tuple = ('M1', self.constraint1)\n self.widget.onAcceptConstraint(constraint_tuple)\n\n # check the getObjectByName call\n self.widget.getObjectByName.assert_called_with('M1')\n\n # check the tab method calls\n test_tab.getRowFromName.assert_called_with(self.constraint1.param)\n test_tab.addConstraintToRow.assert_called_with(self.constraint1, 1)\n test_tab.changeCheckboxStatus.assert_called_with(1, True)\n\n def testFitComplete(self):\n ''' test the handling of fit results'''\n self.widget.getTabsForFit = MagicMock(return_value=[[None], [None]])\n spy = QSignalSpy(self.widget.parent.communicate.statusBarUpdateSignal)\n # test handling of fit error\n # result is None\n result = None\n self.widget.fitComplete(result)\n self.assertEqual(spy[0][0], 'Fitting failed.')\n # Result has failed\n result = MagicMock(return_value= \"foo\")\n results = [[[result]], 1.5]\n result.success = False\n result.mesg = [\"foo\", None]\n self.widget.fitComplete(results)\n self.assertEqual(spy[1][0], 'Fitting failed with the following '\n 'message: foo')\n\n # test a successful fit\n result.success = True\n test_tab = MagicMock()\n test_tab.kernel_module.name = 'M1'\n test_tab.fitComplete = MagicMock()\n result.model.name = 'M1'\n self.widget.tabs_for_fitting = {\"test_tab\": test_tab}\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n self.widget.fitComplete(results)\n self.assertEqual(test_tab.fitComplete.call_args[0][0][1], 1.5)\n self.assertEqual(test_tab.fitComplete.call_args[0][0][0],\n [[result]])\n self.assertEqual(spy[2][0], 'Fitting completed successfully in: 1.5 '\n 's.\\n')\n\n def testBatchFitComplete(self):\n ''' test the handling of batch fit results'''\n self.widget.getTabsForFit = MagicMock(return_value=[[None], [None]])\n spy = QSignalSpy(self.widget.parent.communicate.statusBarUpdateSignal)\n spy_data = QSignalSpy(\n self.widget.parent.communicate.sendDataToGridSignal)\n # test handling of fit error\n # result is None\n result = None\n self.widget.batchComplete(result)\n self.assertEqual(spy[0][0], 'Fitting failed.')\n # Result has failed\n result = MagicMock(return_value= \"foo\")\n results = [[[result]], 1.5]\n result.success = False\n result.mesg = [\"foo\", None]\n self.widget.batchComplete(results)\n self.assertEqual(spy[1][0], 'Fitting failed with the following '\n 'message: foo')\n\n # test a successful fit\n result.success = True\n self.widget.batchComplete(results)\n self.assertEqual(spy[2][0], 'Fitting completed successfully in: 1.5 '\n 's.\\n')\n self.assertEqual(spy_data[0][0], [[result], 'ConstSimulPage'])\n\n def testUncheckConstraints(self):\n '''Tests the unchecking of constraints'''\n # mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = MagicMock()\n test_tab.kernel_module.name = \"M1\"\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n\n # Add a tab with an active constraint\n test_tab.getComplexConstraintsForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getFullConstraintNameListForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getConstraintObjectsForModel = MagicMock(\n return_value=[self.constraint1])\n test_tab.getConstraintForRow = MagicMock(return_value=self.constraint1)\n self.widget.updateFitLine(\"test_tab\")\n self.widget.parent.getTabByName = MagicMock(return_value=test_tab)\n perspective = self.widget.parent.parent.perspective()\n perspective.symbol_dict = {\"M1.scale\": 1, \"M1.radius\": 1}\n self.widget.initializeFitList = MagicMock()\n\n # Constraint should be checked\n self.assertEqual(self.widget.tblConstraints.item(0, 0).checkState(), 2)\n\n self.widget.uncheckConstraint('M1:scale')\n # Should be unchecked in tblConstraint\n self.assertEqual(self.widget.tblConstraints.item(0, 0).checkState(), 0)\n # Constraint should be deactivated\n self.assertEqual(self.constraint1.active, False)\n\n def testOnConstraintChange(self):\n ''' test edition of the constraint list '''\n # mock a tab\n test_tab = MagicMock(spec=FittingWidget)\n test_tab.data_is_loaded = False\n test_tab.kernel_module = MagicMock()\n test_tab.kernel_module.name = \"M1\"\n test_tab.getRowFromName = MagicMock(return_value=0)\n ObjectLibrary.getObject = MagicMock(return_value=test_tab)\n\n # Add a constraint to the tab\n test_tab.getComplexConstraintsForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getFullConstraintNameListForModel = MagicMock(\n return_value=[('scale', self.constraint1.func)])\n test_tab.getConstraintObjectsForModel = MagicMock(\n return_value=[self.constraint1])\n\n self.widget.updateFitLine(\"test_tab\")\n\n self.widget.initializeFitList = MagicMock()\n QtWidgets.QMessageBox.critical = MagicMock()\n\n # Change the constraint to one with no equal sign\n self.widget.tblConstraints.item(0, 0).setText(\"foo\")\n\n msg = (\"Incorrect operator in constraint definition. Please use = \"\n \"sign to define constraints.\")\n (QtWidgets.QMessageBox.critical.\n assert_called_with(self.widget, \"Inconsistent constraint\", msg,\n QtWidgets.QMessageBox.Ok))\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n\n self.widget.initializeFitList.reset_mock()\n # Change the constraint to one with no colons in constrained parameter\n self.widget.tblConstraints.item(0, 0).setText(\"foo = bar\")\n msg = (\"Incorrect constrained parameter definition. Please use colons\"\n \" to separate model and parameter on the rhs of the definition, \"\n \"e.g. M1:scale\")\n (QtWidgets.QMessageBox.critical.\n assert_called_with(self.widget, \"Inconsistent constraint\", msg,\n QtWidgets.QMessageBox.Ok))\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n\n self.widget.initializeFitList.reset_mock()\n perspective = self.widget.parent.parent.perspective()\n # Change the constraint to one with an unknown symbol or with several\n # parameters on the rhs of the constraint definition\n self.widget.tblConstraints.item(0, 0).setText(\"M1:foo = bar\")\n msg = (\"Unknown parameter M1.foo used in constraint. Please use \"\n \"a single known parameter in the rhs of the constraint \"\n \"definition, e.g. M1:scale = M1.radius + 2\")\n (QtWidgets.QMessageBox.critical.\n assert_called_with(self.widget, \"Inconsistent constraint\", msg,\n QtWidgets.QMessageBox.Ok))\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n\n self.widget.initializeFitList.reset_mock()\n # Check replacement of a constraint\n perspective.symbol_dict = {\"M1.scale\": 1, \"M1.radius\": 1}\n\n self.widget.tblConstraints.item(0, 0).setText(\"M1:radius = bar\")\n constraint = Constraint(param=\"radius\", func=\"bar\",\n value_ex=\"M1.radius\")\n target = test_tab.addConstraintToRow.call_args[1]\n self.assertEqual(target[\"constraint\"].value_ex, constraint.value_ex)\n self.assertEqual(target[\"constraint\"].func, constraint.func)\n self.assertEqual(target[\"constraint\"].param, constraint.param)\n self.assertEqual(target[\"row\"], 0)\n target = test_tab.deleteConstraintOnParameter.call_args[0]\n self.assertEqual(target[0], \"scale\")\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n\n self.widget.initializeFitList.reset_mock()\n # Check the checkbox\n self.widget.tblConstraints.item(0, 0).setText(\"M1:scale = M1.sld\")\n self.assertEqual(test_tab.modifyViewOnRow.call_args[0][0], 0)\n font = QtGui.QFont()\n font.setItalic(True)\n self.assertEqual(test_tab.modifyViewOnRow.call_args[1][\"font\"],\n font)\n brush = QtGui.QBrush(QtGui.QColor('blue'))\n self.assertEqual(test_tab.modifyViewOnRow.call_args[1][\"brush\"],\n brush)\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n\n self.widget.initializeFitList.reset_mock()\n # Uncheck the checkbox\n self.widget.tblConstraints.item(0, 0).setCheckState(0)\n self.assertEqual(test_tab.modifyViewOnRow.call_args[0][0], 0)\n self.assertTrue(not test_tab.modifyViewOnRow.call_args[1])\n self.assertEqual(self.constraint1.active, False)\n # Check the reloading of the view\n self.widget.initializeFitList.assert_called_once()\n", "sub_path": "src/sas/qtgui/Perspectives/Fitting/UnitTesting/ConstraintWidgetTest.py", "file_name": "ConstraintWidgetTest.py", "file_ext": "py", "file_size_in_byte": 19967, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PyQt5.QtWidgets.QApplication.instance", "line_number": 24, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 24, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 24, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 25, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets", "line_number": 25, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 25, "usage_type": "attribute"}, {"api_name": "unittest.TestCase", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.GuiUtils.Communicate", "line_number": 33, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.GuiUtils", "line_number": 33, "usage_type": "name"}, {"api_name": "sas.qtgui.Utilities.GuiUtils.Communicate", "line_number": 34, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.GuiUtils", "line_number": 34, "usage_type": "name"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingPerspective.FittingWindow", "line_number": 59, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.ConstraintWidget.ConstraintWidget.updateSignalsFromTab", "line_number": 60, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Perspectives.Fitting.ConstraintWidget.ConstraintWidget", "line_number": 60, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 60, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.ConstraintWidget.ConstraintWidget", "line_number": 62, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.Constraint.Constraint", "line_number": 65, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.Constraint.Constraint", "line_number": 68, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QWidget", "line_number": 77, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 77, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 93, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest.mouseClick", "line_number": 97, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 97, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 97, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 97, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication.processEvents", "line_number": 98, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 98, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 98, "usage_type": "name"}, {"api_name": "PyQt5.QtTest.QTest.mouseClick", "line_number": 104, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QTest", "line_number": 104, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 104, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore", "line_number": 104, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication.processEvents", "line_number": 105, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 105, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 105, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 124, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 124, "usage_type": "name"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 127, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 127, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 127, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 141, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 141, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 143, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 144, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 144, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 144, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 159, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 159, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 161, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 163, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 163, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 163, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 172, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 174, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 176, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 191, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 201, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 201, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 203, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 204, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 204, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 204, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.listObjects", "line_number": 207, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 207, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 207, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 216, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.listObjects", "line_number": 217, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 217, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 217, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 218, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 219, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 226, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.listObjects", "line_number": 227, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 227, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 227, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 228, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 229, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 231, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 239, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 239, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 240, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 241, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 242, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 245, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 261, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QSignalSpy", "line_number": 262, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 269, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 279, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 281, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 284, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 284, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 284, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 294, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QSignalSpy", "line_number": 295, "usage_type": "call"}, {"api_name": "PyQt5.QtTest.QSignalSpy", "line_number": 296, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 304, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 322, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 322, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 324, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 326, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 326, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 326, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 329, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 331, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 333, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 335, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 337, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 340, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 354, "usage_type": "call"}, {"api_name": "sas.qtgui.Perspectives.Fitting.FittingWidget.FittingWidget", "line_number": 354, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 356, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 358, "usage_type": "call"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary.getObject", "line_number": 359, "usage_type": "attribute"}, {"api_name": "sas.qtgui.Utilities.ObjectLibrary", "line_number": 359, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 359, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 362, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 364, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 366, "usage_type": "call"}, {"api_name": "unittest.mock.MagicMock", "line_number": 371, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 372, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 372, "usage_type": "name"}, {"api_name": "unittest.mock.MagicMock", "line_number": 372, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.critical.assert_called_with", "line_number": 379, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 379, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 379, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 381, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 381, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.critical.assert_called_with", "line_number": 391, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 391, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 391, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 393, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 393, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.critical.assert_called_with", "line_number": 405, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 405, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 405, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 407, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets", "line_number": 407, "usage_type": "name"}, {"api_name": "sas.qtgui.Perspectives.Fitting.Constraint.Constraint", "line_number": 416, "usage_type": "call"}, {"api_name": "PyQt5.QtGui.QFont", "line_number": 432, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 432, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QBrush", "line_number": 436, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 436, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QColor", "line_number": 436, "usage_type": "call"}]} +{"seq_id": "372843523", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('sale', '0004_order_invoice_id'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='InvoiceNumber',\n fields=[\n ('id', models.AutoField(serialize=False, verbose_name='ID', auto_created=True, primary_key=True)),\n ('year', models.IntegerField()),\n ('last_number', models.IntegerField()),\n ('marketplace_suffix', models.CharField(max_length=1)),\n ('marketplace', models.ForeignKey(to='sale.Marketplace')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.AlterUniqueTogether(\n name='invoicenumber',\n unique_together=set([('marketplace', 'marketplace_suffix', 'year')]),\n ),\n migrations.RemoveField(\n model_name='order',\n name='invoice_id',\n ),\n ]\n", "sub_path": "m13/sale/migrations/0005_auto_20141203_0855.py", "file_name": "0005_auto_20141203_0855.py", "file_ext": "py", "file_size_in_byte": 1069, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 17, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 17, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.IntegerField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.Model", "line_number": 25, "usage_type": "attribute"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterUniqueTogether", "line_number": 27, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 27, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 31, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 31, "usage_type": "name"}]} +{"seq_id": "223944409", "text": "# -*- coding: utf-8 -*-\n# MegEngine is Licensed under the Apache License, Version 2.0 (the \"License\")\n#\n# Copyright (c) 2014-2020 Megvii Inc. All rights reserved.\n#\n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\nimport numpy as np\nimport pytest\n\nimport megengine\nfrom megengine import is_cuda_available, tensor\nfrom megengine.core._imperative_rt import CompNode\nfrom megengine.core._imperative_rt.core2 import apply\nfrom megengine.core._imperative_rt.ops import (\n delete_rng_handle,\n get_global_rng_seed,\n new_rng_handle,\n)\nfrom megengine.core.ops.builtin import GaussianRNG, UniformRNG\nfrom megengine.distributed.helper import get_device_count_by_fork\nfrom megengine.random import RNG\nfrom megengine.random.rng import _normal, _uniform\n\n\n@pytest.mark.skipif(\n get_device_count_by_fork(\"xpu\") <= 2, reason=\"xpu counts need > 2\",\n)\ndef test_gaussian_op():\n shape = (\n 8,\n 9,\n 11,\n 12,\n )\n shape = tensor(shape, dtype=\"int32\")\n op = GaussianRNG(seed=get_global_rng_seed(), mean=1.0, std=3.0)\n (output,) = apply(op, shape)\n assert np.fabs(output.numpy().mean() - 1.0) < 1e-1\n assert np.sqrt(output.numpy().var()) - 3.0 < 1e-1\n assert str(output.device) == str(CompNode(\"xpux\"))\n\n cn = CompNode(\"xpu2\")\n seed = 233333\n h = new_rng_handle(cn, seed)\n op = GaussianRNG(seed=seed, mean=3.0, std=1.0, handle=h)\n (output,) = apply(op, shape)\n delete_rng_handle(h)\n assert np.fabs(output.numpy().mean() - 3.0) < 1e-1\n assert np.sqrt(output.numpy().var()) - 1.0 < 1e-1\n assert str(output.device) == str(cn)\n\n\n@pytest.mark.skipif(\n get_device_count_by_fork(\"xpu\") <= 2, reason=\"xpu counts need > 2\",\n)\ndef test_uniform_op():\n shape = (\n 8,\n 9,\n 11,\n 12,\n )\n shape = tensor(shape, dtype=\"int32\")\n op = UniformRNG(seed=get_global_rng_seed())\n (output,) = apply(op, shape)\n assert np.fabs(output.numpy().mean() - 0.5) < 1e-1\n assert str(output.device) == str(CompNode(\"xpux\"))\n\n cn = CompNode(\"xpu2\")\n seed = 233333\n h = new_rng_handle(cn, seed)\n op = UniformRNG(seed=seed, handle=h)\n (output,) = apply(op, shape)\n delete_rng_handle(h)\n assert np.fabs(output.numpy().mean() - 0.5) < 1e-1\n assert str(output.device) == str(cn)\n\n\n@pytest.mark.skipif(\n get_device_count_by_fork(\"xpu\") <= 1, reason=\"xpu counts need > 1\",\n)\ndef test_UniformRNG():\n m1 = RNG(seed=111, device=\"xpu0\")\n m2 = RNG(seed=111, device=\"xpu1\")\n m3 = RNG(seed=222, device=\"xpu0\")\n out1 = m1.uniform(size=(100,))\n out1_ = m1.uniform(size=(100,))\n out2 = m2.uniform(size=(100,))\n out3 = m3.uniform(size=(100,))\n\n np.testing.assert_equal(out1.numpy(), out2.numpy())\n assert out1.device == \"xpu0\" and out2.device == \"xpu1\"\n assert not (out1.numpy() == out3.numpy()).all()\n assert not (out1.numpy() == out1_.numpy()).all()\n\n low = -234\n high = 123\n out = m1.uniform(low=low, high=high, size=(20, 30, 40))\n out_shp = out.shape\n if isinstance(out_shp, tuple):\n assert out_shp == (20, 30, 40)\n else:\n assert all(out.shape.numpy() == np.array([20, 30, 40]))\n assert np.abs(out.mean().numpy() - ((low + high) / 2)) / (high - low) < 0.1\n\n\n@pytest.mark.skipif(\n get_device_count_by_fork(\"xpu\") <= 1, reason=\"xpu counts need > 1\",\n)\ndef test_NormalRNG():\n m1 = RNG(seed=111, device=\"xpu0\")\n m2 = RNG(seed=111, device=\"xpu1\")\n m3 = RNG(seed=222, device=\"xpu0\")\n out1 = m1.normal(size=(100,))\n out1_ = m1.uniform(size=(100,))\n out2 = m2.normal(size=(100,))\n out3 = m3.normal(size=(100,))\n\n np.testing.assert_equal(out1.numpy(), out2.numpy())\n assert out1.device == \"xpu0\" and out2.device == \"xpu1\"\n assert not (out1.numpy() == out3.numpy()).all()\n assert not (out1.numpy() == out1_.numpy()).all()\n\n mean = -1\n std = 2\n out = m1.normal(mean=mean, std=std, size=(20, 30, 40))\n out_shp = out.shape\n if isinstance(out_shp, tuple):\n assert out_shp == (20, 30, 40)\n else:\n assert all(out.shape.numpy() == np.array([20, 30, 40]))\n assert np.abs(out.mean().numpy() - mean) / std < 0.1\n assert np.abs(np.std(out.numpy()) - std) < 0.1\n", "sub_path": "imperative/python/test/unit/random/test_rng.py", "file_name": "test_rng.py", "file_ext": "py", "file_size_in_byte": 4340, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "megengine.tensor", "line_number": 37, "usage_type": "call"}, {"api_name": "megengine.core.ops.builtin.GaussianRNG", "line_number": 38, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.get_global_rng_seed", "line_number": 38, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.core2.apply", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 41, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.CompNode", "line_number": 42, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.CompNode", "line_number": 44, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.new_rng_handle", "line_number": 46, "usage_type": "call"}, {"api_name": "megengine.core.ops.builtin.GaussianRNG", "line_number": 47, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.core2.apply", "line_number": 48, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.delete_rng_handle", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 51, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 27, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 27, "usage_type": "attribute"}, {"api_name": "megengine.distributed.helper.get_device_count_by_fork", "line_number": 28, "usage_type": "call"}, {"api_name": "megengine.tensor", "line_number": 65, "usage_type": "call"}, {"api_name": "megengine.core.ops.builtin.UniformRNG", "line_number": 66, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.get_global_rng_seed", "line_number": 66, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.core2.apply", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 68, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.CompNode", "line_number": 69, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.CompNode", "line_number": 71, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.new_rng_handle", "line_number": 73, "usage_type": "call"}, {"api_name": "megengine.core.ops.builtin.UniformRNG", "line_number": 74, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.core2.apply", "line_number": 75, "usage_type": "call"}, {"api_name": "megengine.core._imperative_rt.ops.delete_rng_handle", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.fabs", "line_number": 77, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 55, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 55, "usage_type": "attribute"}, {"api_name": "megengine.distributed.helper.get_device_count_by_fork", "line_number": 56, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 85, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 86, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 87, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 93, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 106, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 81, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 81, "usage_type": "attribute"}, {"api_name": "megengine.distributed.helper.get_device_count_by_fork", "line_number": 82, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 113, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 114, "usage_type": "call"}, {"api_name": "megengine.random.RNG", "line_number": 115, "usage_type": "call"}, {"api_name": "numpy.testing.assert_equal", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.testing", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 135, "usage_type": "call"}, {"api_name": "pytest.mark.skipif", "line_number": 109, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 109, "usage_type": "attribute"}, {"api_name": "megengine.distributed.helper.get_device_count_by_fork", "line_number": 110, "usage_type": "call"}]} +{"seq_id": "86118705", "text": "# encoding: utf-8\nfrom rest_framework import serializers\nfrom .models import Role, UserInfo, Menu, Permission,NewMenu\n\n\nclass RoleSerializer(serializers.ModelSerializer):\n create_time = serializers.DateTimeField(format='%Y-%m-%d %H:%M:%S')\n # roles = serializers.SlugRelatedField(many=True, read_only=True, slug_field='menu')\n class Meta:\n model = Role\n fields =\"__all__\"\n\n\nclass UserSerializer(serializers.ModelSerializer):\n create_time = serializers.DateTimeField(format='%Y-%m-%d %H:%M:%S')\n roles = serializers.SlugRelatedField(many=True, read_only=True,slug_field='rolename')\n\n # roles = RoleSerializer()\n class Meta:\n model = UserInfo\n fields = ('id','username','phone_number','nickname','email','status','create_time','roles')\n # fields = '__all__'\n\n\nclass PermissonSerializer(serializers.ModelSerializer):\n menu = Menu.title\n\n class Meta:\n model = Permission\n fields = ('id', 'title', 'url', 'menu')\n\nclass NewmenuSerializer(serializers.ModelSerializer):\n\n class Meta:\n model = NewMenu\n fields = ('id','name','url','perms','type','orderNum','parentid')\n\n def get_menus(self,obj):\n title_list = [self.name]\n p = self.parentid\n while p:\n title_list.insert(0, p.name)\n p = p.parentid\n title_list.append(p)\n return title_list\n\n\n\n\nclass MenuSerializer(serializers.ModelSerializer):\n parent = serializers.SlugRelatedField (many=True, read_only=True,slug_field='parent')\n class Meta:\n model = Menu\n fields = ('id', 'title', 'url', 'parent')\n", "sub_path": "apps/rbac/serializers.py", "file_name": "serializers.py", "file_ext": "py", "file_size_in_byte": 1619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 6, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 6, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 7, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 7, "usage_type": "name"}, {"api_name": "models.Role", "line_number": 10, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 14, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 14, "usage_type": "name"}, {"api_name": "rest_framework.serializers.DateTimeField", "line_number": 15, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 15, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SlugRelatedField", "line_number": 16, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 16, "usage_type": "name"}, {"api_name": "models.UserInfo", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 25, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 25, "usage_type": "name"}, {"api_name": "models.Menu.title", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.Menu", "line_number": 26, "usage_type": "name"}, {"api_name": "models.Permission", "line_number": 29, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 32, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 32, "usage_type": "name"}, {"api_name": "models.NewMenu", "line_number": 35, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ModelSerializer", "line_number": 50, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 50, "usage_type": "name"}, {"api_name": "rest_framework.serializers.SlugRelatedField", "line_number": 51, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 51, "usage_type": "name"}, {"api_name": "models.Menu", "line_number": 53, "usage_type": "name"}]} +{"seq_id": "581012537", "text": "from functools import reduce\nfrom typing import cast, MutableSequence, Sequence\n\nfrom looker_sdk import client, models\n\n\nsdk = client.setup(\"../looker.ini\")\n\n\ndef main():\n connections = get_connections()\n\n if not connections:\n print(\"No connections found.\")\n\n for connection in connections:\n if connection.name == \"looker\":\n continue\n test_results = run_connection_tests(connection)\n generate_report(connection.name, test_results)\n\n\ndef get_connections() -> Sequence[models.DBConnection]:\n \"\"\"Get list of all connections.\"\"\"\n return sdk.all_connections(fields=\"name, dialect\")\n\n\ndef run_connection_tests(\n connection: models.DBConnection\n) -> Sequence[models.DBConnectionTestResult]:\n \"\"\"Run supported tests against a given connection.\"\"\"\n assert connection.name\n assert connection.dialect and connection.dialect.connection_tests\n supported_tests: MutableSequence[str] = list(connection.dialect.connection_tests)\n test_results = sdk.test_connection(\n connection.name, models.DelimSequence(supported_tests)\n )\n return test_results\n\n\ndef generate_report(\n connection_name: str, test_results: Sequence[models.DBConnectionTestResult]\n):\n errors = list(filter(lambda test: cast(str, test.status) == \"error\", test_results))\n if errors:\n report = reduce(\n lambda failures, error: failures + f\"\\n - {error.message}\",\n errors,\n f\"{connection_name}:\",\n )\n else:\n report = f\"{connection_name}: OK\"\n print(report)\n\n\nmain()\n", "sub_path": "python/test_connections.py", "file_name": "test_connections.py", "file_ext": "py", "file_size_in_byte": 1566, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "looker_sdk.client.setup", "line_number": 7, "usage_type": "call"}, {"api_name": "looker_sdk.client", "line_number": 7, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 23, "usage_type": "name"}, {"api_name": "looker_sdk.models.DBConnection", "line_number": 23, "usage_type": "attribute"}, {"api_name": "looker_sdk.models", "line_number": 23, "usage_type": "name"}, {"api_name": "looker_sdk.models.DBConnection", "line_number": 29, "usage_type": "attribute"}, {"api_name": "looker_sdk.models", "line_number": 29, "usage_type": "name"}, {"api_name": "typing.MutableSequence", "line_number": 34, "usage_type": "name"}, {"api_name": "looker_sdk.models.DelimSequence", "line_number": 36, "usage_type": "call"}, {"api_name": "looker_sdk.models", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 30, "usage_type": "name"}, {"api_name": "looker_sdk.models.DBConnectionTestResult", "line_number": 30, "usage_type": "attribute"}, {"api_name": "looker_sdk.models", "line_number": 30, "usage_type": "name"}, {"api_name": "typing.Sequence", "line_number": 42, "usage_type": "name"}, {"api_name": "looker_sdk.models.DBConnectionTestResult", "line_number": 42, "usage_type": "attribute"}, {"api_name": "looker_sdk.models", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 44, "usage_type": "call"}, {"api_name": "functools.reduce", "line_number": 46, "usage_type": "call"}]} +{"seq_id": "308511967", "text": "from TextClassification import TextClassification, DataPreprocess\nfrom sklearn.model_selection import train_test_split\nfrom TextClassification import load_data\nimport numpy as np\n\n# load data\ndata = load_data(name='single')\nx = data['evaluation']\ny = [[i] for i in data['label']]\n\n# split train and test\nX_train, X_test, y_train, y_test = train_test_split(x, y, test_size=0.2)\n\nmodel = TextClassification()\n\n# train\nmodel.fit(x=X_train, y=y_train, method='CNN', model=None,\n x_need_preprocess=True, y_need_preprocess=True,\n epochs=1, batchsize=128, output_type='single')\n\n# predict\nlabel_set = model.label_set\ny_predict = model.predict(x=X_test, x_need_preprocess=True)\ny_predict_label = model.label2toptag(predictions=y_predict, labelset=label_set)\nprint(sum([y_predict_label[i] == y_test[i] for i in range(len(y_predict))]) / len(y_predict))\n", "sub_path": "demo/demo_net_single.py", "file_name": "demo_net_single.py", "file_ext": "py", "file_size_in_byte": 862, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "TextClassification.load_data", "line_number": 7, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 12, "usage_type": "call"}, {"api_name": "TextClassification.TextClassification", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "558826259", "text": "from pyspark import SparkConf, SparkContext\nfrom pyspark.sql.session import SparkSession\n\nfrom pyspark.sql.functions import *\nfrom pyspark.sql.types import *\nimport pyspark.sql.functions as f\nimport re,html\nfrom pyspark.sql.functions import broadcast\n\n\"\"\"\nThis pyspark script performs the joins between the questions which has the response times and the pagerank dataframe of\nthe post(question) links which is then pushed to the postgresql database\n\n\"\"\"\ndef main():\n sc = SparkContext(conf=SparkConf().setAppName(\"se\"))\n spark = SparkSession.builder.appName(\"se\").getOrCreate()\n #read in the links parquet file\n links = spark.read.load(\"s3a://xmlparq/pr_se_links.parquet\")\n #read in the posts parquet file\n posts = spark.read.load(\"s3a://xmlparq/posts.parquet\")\n #filter the questions:\n questions = posts.filter((f.col('PostTypeId')==1)).filter((f.col('AcceptedAnswerId').isNotNull()))\n questions_subset = questions.select('Id','AcceptedAnswerId','Tags','CreationDate', 'Community')\n #filter the answers in another dataframe\n answers = df3.filter((f.col('PostTypeId')==2))\n #rename the answer dataframe columns\n answers_subset = answers.select(\"Id\",\"CreationDate\",\"Community\")\n new_names = ['AnsId', 'AnsCreationDate','AnsCommunity']\n answers_subset = answers_subset.toDF(*new_names)\n #perform a join on the questions df and the answer df based on the common answer id\n qa_deets = questions_subset.join(answers_subset,(answers_subset.AnsCommunity == all_questions.Community) & (questions_subset.AcceptedAnswerId == answers_subset.AnsId))\n \n timeFmt = \"yyyy-MM-dd' 'HH:mm:ss.SSS\"\n timeDiff = (f.unix_timestamp('AnsCreationDate', format=timeFmt)\n - f.unix_timestamp('CreationDate', format=timeFmt))\n #divide duration by seconds to convert it to milli- minutes\n qa_deets = qa_deets.withColumn(\"Duration\", timeDiff/60)\n\n qa_deets_subset = qa_deets.select(\"Id\", \"Tags\",\"CreationDate\",\"Community\", \"Duration\")\n #filter off questions which have no answers\n questions_null = df3.filter((f.col('PostTypeId')==1)).filter((f.col('AcceptedAnswerId').isNull()))\n questions_null_subset = questions_null.select('Id','Tags','CreationDate','Community')\n questions_null_duration = questions_null_subset.withColumn('Duration', lit(None).cast(DoubleType()))\n #combine all questions; questions with answers and question with no answers\n all_questions = qa_deets_subset.union(questions_null_duration)\n all_questions = all_questions.withColumn('post_create_date',all_questions['CreationDate'].cast('date'))\n all_questions = all_questions.withColumnRenamed('id','COMMUNITY_QUESTION_ID')\n all_questions_subset = all_questions.select('COMMUNITY_QUESTION_ID','Community','Tags', 'post_create_date')\n\n links = links.withColumnRenamed(\"community\",\"lcommunity\")\n \"\"\" perform a join based on community and the question id to combine the pagerank score and\n response time duration in one dataframe\n \"\"\"\n cred_tags = all_questions.join(broadcast(links), (links.id == all_questions.COMMUNITY_QUESTION_ID) & (all_questions_subset.Community == links.lcommunity), \"left_outer\")\n #rename columns as per postgresql schema, round off values and write to the database:\n total_df = cred_tags.select(\"COMMUNITY_QUESTION_ID\",\"Community\",\"post_create_date\",\"Tags\",\"Duration\",\"cred_score\")\n total_df = total_df.withColumn(\"duration\",f.round(total_df[\"Duration\"],2))\n total_df = total_df.withColumn(\"pr_score\",f.round(total_df[\"cred_score\"],3))\n total_df = total_df.withColumnRenamed(\"COMMUNITY_QUESTION_ID\",\"qid\")\n total_df = total_df.withColumnRenamed(\"Community\",\"community\")\n total_df = total_df.withColumnRenamed(\"Tags\",\"tags\")\n total_df = total_df.withColumnRenamed(\"post_create_date\",\"create_date\")\n total_df_reqd = total_df.select(\"qid\",\"tags\",\"community\",\"duration\",\"create_date\",\"pr_score\")\n total_df_reqd.write.format(\"jdbc\").mode(\"append\") .option(\"url\", \"jdbc:postgresql://hostname/ls?user=xxx&password=xxx\").option(\"dbtable\", \"questions\").option(\"user\", \"postgres\").option(\"password\", \"xxx\").save()\n spark.catalog.clearCache()\n\nif __name__ == \"__main__\":\n main()\n\n\n \n \n", "sub_path": "src/spark/join_dfs.py", "file_name": "join_dfs.py", "file_ext": "py", "file_size_in_byte": 4110, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pyspark.SparkContext", "line_number": 16, "usage_type": "call"}, {"api_name": "pyspark.SparkConf", "line_number": 16, "usage_type": "call"}, {"api_name": "pyspark.sql.session.SparkSession.builder.appName", "line_number": 17, "usage_type": "call"}, {"api_name": "pyspark.sql.session.SparkSession.builder", "line_number": 17, "usage_type": "attribute"}, {"api_name": "pyspark.sql.session.SparkSession", "line_number": 17, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 23, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 23, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 26, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 26, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.unix_timestamp", "line_number": 35, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 35, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.unix_timestamp", "line_number": 36, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 36, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.col", "line_number": 42, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 42, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.broadcast", "line_number": 55, "usage_type": "call"}, {"api_name": "pyspark.sql.functions.round", "line_number": 58, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 58, "usage_type": "name"}, {"api_name": "pyspark.sql.functions.round", "line_number": 59, "usage_type": "call"}, {"api_name": "pyspark.sql.functions", "line_number": 59, "usage_type": "name"}]} +{"seq_id": "465915775", "text": "#!/usr/bin/env python3\n\n\"\"\"\nWrapper around the cromwell api\n\"\"\"\n\nimport os\nimport argparse\nfrom pathlib import Path\n\nfrom cromwell_tools import api\nfrom cromwell_tools import utilities\nfrom cromwell_tools.cromwell_auth import CromwellAuth\n\n\n# Set defaults\nDEFAULT_WEBSERVICE_PORT = 8000\nDEFAULT_WORKFLOW_OPTIONS = Path(\"/opt/cromwell/configs/options.json\")\n\n\ndef get_args():\n parser = argparse.ArgumentParser(description=\"Submit workflow to cromwell server\")\n parser.add_argument(\"--workflow-source\",\n required=True,\n help=\"Path to workflow code\")\n parser.add_argument(\"--workflow-inputs\",\n required=True,\n help=\"Path to workflow inputs configuration\")\n parser.add_argument(\"--workflow-dependencies\",\n required=False,\n help=\"Zip file containing workflow dependencies\")\n parser.add_argument(\"--webservice-port\",\n type=int,\n required=False, default=DEFAULT_WEBSERVICE_PORT,\n help=\"Port that cromwell is running on\")\n parser.add_argument(\"--workflow-options-json\",\n required=False, default=DEFAULT_WORKFLOW_OPTIONS,\n help=\"Options.json file\")\n\n return parser.parse_args()\n\n\ndef check_args(args):\n \"\"\"\n Ensure that each file is as expected\n :return:\n \"\"\"\n # Convert workflow source to path object\n workflow_source_arg = getattr(args, \"workflow_source\", None)\n workflow_source = Path(workflow_source_arg)\n if not workflow_source.is_file():\n sys.exit(1)\n setattr(args, \"workflow_source\", workflow_source)\n\n # Convert workflow inputs to path object\n workflow_inputs_arg = getattr(args, \"workflow_inputs\", None)\n workflow_inputs = Path(workflow_inputs_arg)\n if not workflow_inputs.is_file():\n sys.exit(1)\n setattr(args, \"workflow_inputs\", workflow_inputs)\n\n # Check workflow dependencies is a zip file, otherwise exit\n workflow_dependencies_arg = getattr(args, \"workflow_dependencies\", None)\n if workflow_dependencies_arg is not None:\n workflow_dependencies = Path(workflow_dependencies_arg)\n if not workflow_dependencies.is_file():\n sys.exit(1)\n setattr(args, \"workflow_dependencies\", workflow_dependencies)\n\n # Check options is a file\n workflow_options_json_arg = getattr(args, \"workflow_options_json\", None)\n if workflow_options_json_arg is not None:\n workflow_options_json = Path(workflow_options_json_arg)\n if not workflow_options_json.is_file():\n sys.exit(1)\n setattr(args, \"workflow_options_json\", workflow_options_json)\n\n return args\n\n\ndef submit_to_cromwell(args):\n # Get authentication (as no authentication)\n auth = CromwellAuth.from_no_authentication(url=\"http://localhost:{}\".format(args.webservice_port))\n\n response = api.submit(auth=auth,\n wdl_file=args.workflow_source,\n inputs_files=args.workflow_inputs,\n dependencies=args.workflow_dependencies,\n validate_labels=True)\n\n # FIXME - need to test this first a bit\n print(response)\n\n\ndef main():\n # Get args\n args = get_args()\n # Check em\n args = check_args(args)\n # Submit to cromwell\n submit_to_cromwell(args)\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "cromwell/scripts/submit_to_cromwell.py", "file_name": "submit_to_cromwell.py", "file_ext": "py", "file_size_in_byte": 3456, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pathlib.Path", "line_number": 18, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 22, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 50, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 57, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 65, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 73, "usage_type": "call"}, {"api_name": "cromwell_tools.cromwell_auth.CromwellAuth.from_no_authentication", "line_number": 83, "usage_type": "call"}, {"api_name": "cromwell_tools.cromwell_auth.CromwellAuth", "line_number": 83, "usage_type": "name"}, {"api_name": "cromwell_tools.api.submit", "line_number": 85, "usage_type": "call"}, {"api_name": "cromwell_tools.api", "line_number": 85, "usage_type": "name"}]} +{"seq_id": "530649425", "text": "import matplotlib\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Activation, Dropout\nfrom keras.optimizers import SGD\n\n\n# Generate random 50,000 images with single objects\nnum_imgs = 50000\n\nimg_size = 8\nmin_object_size = 1\nmax_object_size = 4\nnum_objects = 1\n\nbboxes = np.zeros((num_imgs, num_objects, 4))\nimgs = np.zeros((num_imgs, img_size, img_size)) # set background to 0\n\nfor i_img in range(num_imgs):\n for i_object in range(num_objects):\n w, h = np.random.randint(min_object_size, max_object_size, size=2)\n x = np.random.randint(0, img_size - w)\n y = np.random.randint(0, img_size - h)\n imgs[i_img, x:x + w, y:y + h] = 1. # set rectangle to 1\n bboxes[i_img, i_object] = [x, y, w, h]\n\nX = (imgs.reshape(num_imgs, -1) - np.mean(imgs)) / np.std(imgs)\nX.shape, np.mean(X), np.std(X)\n\ny = bboxes.reshape(num_imgs, -1) / img_size\n\n\n# Split training and testing data\ni = int(0.8 * num_imgs)\ntrain_X = X[:i]\ntest_X = X[i:]\ntrain_y = y[:i]\ntest_y = y[i:]\ntest_imgs = imgs[i:]\ntest_bboxes = bboxes[i:]\n\n# Define a simple feed forward model in keras\n# Using adadelta optimizer as it automatically chooses hyperparameters and the learning rate\nmodel = Sequential([Dense(200, input_dim=X.shape[-1]), Activation('relu'), Dropout(0.2), Dense(y.shape[-1])])\nmodel.compile('adadelta', 'mse')\n\nmodel.fit(train_X, train_y, nb_epoch=30, validation_data=(test_X, test_y), verbose=2)\n\npred_y = model.predict(test_X)\npred_bboxes = pred_y * img_size\npred_bboxes = pred_bboxes.reshape(len(pred_bboxes), num_objects, -1)\n\n\ndef IOU(bbox1, bbox2):\n '''Calculate overlap between two bounding boxes [x, y, w, h] as the area of intersection over the area of unity'''\n x1, y1, w1, h1 = bbox1[0], bbox1[1], bbox1[2], bbox1[3]\n x2, y2, w2, h2 = bbox2[0], bbox2[1], bbox2[2], bbox2[3]\n\n w_I = min(x1 + w1, x2 + w2) - max(x1, x2)\n h_I = min(y1 + h1, y2 + h2) - max(y1, y2)\n if w_I <= 0 or h_I <= 0: # no overlap\n return 0.\n I = w_I * h_I\n\n U = w1 * h1 + w2 * h2 - I\n\n return I/U\n\nplt.figure(figsize=(12, 3))\nfor i_subplot in range(1, 5):\n plt.subplot(1, 4, i_subplot)\n i = np.random.randint(len(test_imgs))\n plt.imshow(test_imgs[i].T, cmap='Greys', interpolation='none', origin='lower', extent=[0, img_size, 0, img_size])\n for pred_bbox, exp_bbox in zip(pred_bboxes[i], test_bboxes[i]):\n plt.gca().add_patch(matplotlib.patches.Rectangle((pred_bbox[0], pred_bbox[1]), pred_bbox[2], pred_bbox[3], ec='r', fc='none'))\n plt.annotate('IOU: {:.2f}'.format(IOU(pred_bbox, exp_bbox)), (pred_bbox[0], pred_bbox[1]+pred_bbox[3]+0.2), color='r')\n\nplt.show()", "sub_path": "simple_obj_dec_keras.py", "file_name": "simple_obj_dec_keras.py", "file_ext": "py", "file_size_in_byte": 2689, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.zeros", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 23, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 23, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 24, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 24, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 29, "usage_type": "call"}, {"api_name": "numpy.std", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 45, "usage_type": "call"}, {"api_name": "keras.layers.Dropout", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplot", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 72, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 73, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.patches.Rectangle", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.patches", "line_number": 76, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.annotate", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 79, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 79, "usage_type": "name"}]} +{"seq_id": "45472585", "text": "import os\nimport csv, sqlite3\nfrom tkinter import *\nfrom tkinter import ttk\n\ncd = os.path.dirname(os.path.abspath(__file__))\n\nconn = sqlite3.connect(os.path.join(cd, 'DATA\\\\Vehicles.db'))\ncur = conn.cursor()\n\nclass GUIapp():\n \n def __init__(self):\n self.root = Tk()\n self.buildControls()\n self.root.mainloop()\n\n def DBdata(self, sql, param):\n cur.execute(sql, (param,))\n dbdata = [str(row[0]) for row in cur.fetchall()]\n return(dbdata)\n\n def updateCombos(self, event): \n self.modellist = self.DBdata(\"SELECT DISTINCT [Model] FROM vehicles WHERE [Make] = ? ORDER BY [Model];\", self.makecbo.get()) \n self.modelcbo['values'] = [''] + self.modellist\n self.modelcbo.current(0)\n\n self.englist = self.DBdata(\"SELECT DISTINCT [engId] FROM vehicles WHERE [Make] = ? ORDER BY [engId];\", self.makecbo.get())\n self.engcbo['values'] = [''] + self.englist \n self.engcbo.current(0)\n\n self.translist = self.DBdata(\"SELECT DISTINCT [trany] FROM vehicles WHERE [Make] = ? ORDER BY [trany];\", self.makecbo.get())\n self.transcbo['values'] = [''] + self.translist \n self.transcbo.current(0)\n\n self.fuellist = self.DBdata(\"SELECT DISTINCT [fuelType] FROM vehicles WHERE [Make] = ? ORDER BY [fuelType];\", self.makecbo.get())\n self.fuelcbo['values'] = [''] + self.fuellist \n self.fuelcbo.current(0)\n \n\n def buildControls(self): \n self.root.wm_title(\"Report Menu\")\n self.guiframe = Frame(self.root, width=800, height=500, bd=1, relief=FLAT)\n self.guiframe.pack(padx=5, pady=5)\n\n # IMAGE\n self.photo = PhotoImage(file=\"IMG/CarIcon.png\")\n self.imglbl = Label(self.guiframe, image=self.photo)\n self.imglbl.photo = self.photo\n self.imglbl.grid(row=0, sticky=W, padx=5, pady=5)\n self.imglbl = Label(self.guiframe, text=\"Vehicle Reports Menu\", font=(\"Arial\", 14)).\\\n grid(row=0, column=1, sticky=W, padx=5, pady=5)\n\n # MAKE\n self.makelbl = Label(self.guiframe, text=\"Make\", font=(\"Arial\", 10)).grid(row=1, sticky=W, padx=5, pady=5)\n self.makevar = StringVar()\n self.makecbo = ttk.Combobox(self.guiframe, textvariable=self.makevar, font=(\"Arial\", 10), state='readonly') \n self.makelist = self.DBdata(\"SELECT DISTINCT [Make] FROM vehicles WHERE ? ORDER BY [Make];\", 1) \n self.makecbo['values'] = [''] + self.makelist \n self.makecbo.current(0)\n self.makecbo.grid(row=1, column=1, sticky=W, padx=5, pady=5)\n self.makecbo.bind(\"<>\", self.updateCombos)\n \n # MODEL\n self.modellbl = Label(self.guiframe, text=\"Model\", font=(\"Arial\", 10)).grid(row=2, sticky=W, padx=5, pady=5)\n self.modelvar = StringVar()\n self.modelcbo = ttk.Combobox(self.guiframe, textvariable=self.modelvar, font=(\"Arial\", 10),state='readonly') \n self.modelcbo.grid(row=2, column=1, sticky=W, padx=5, pady=5)\n\n # YEAR START\n self.yearStartlbl = Label(self.guiframe, text=\"Year Start\", font=(\"Arial\", 10)).grid(row=3, sticky=W, padx=5, pady=5)\n self.yearStartvar = IntVar()\n self.yearStarttxt = Entry(self.guiframe, textvariable=self.yearStartvar, relief=SOLID, font=(\"Arial\", 10), width=23).\\\n grid(row=3, column=1, sticky=W, padx=5, pady=5)\n \n # YEAR END\n self.yearEndlbl = Label(self.guiframe, text=\"Year End\", font=(\"Arial\", 10)).grid(row=4, sticky=W, padx=5, pady=5)\n self.yearEndvar = IntVar()\n self.yearEndtxt = Entry(self.guiframe, textvariable=self.yearEndvar, relief=SOLID, font=(\"Arial\", 10), width=23).\\\n grid(row=4, column=1, sticky=W, padx=5, pady=5)\n\n # ENGINE\n self.englbl = Label(self.guiframe, text=\"Engine\", font=(\"Arial\", 10)).grid(row=5, sticky=W, padx=5, pady=5)\n self.engvar = IntVar()\n self.engcbo = ttk.Combobox(self.guiframe, textvariable=self.engvar, font=(\"Arial\", 10))\n self.engcbo.grid(row=5, column=1, sticky=W, padx=5, pady=5)\n\n # TRANSMISSION\n self.translbl = Label(self.guiframe, text=\"Transmission\", font=(\"Arial\", 10)).grid(row=6, sticky=W, padx=5, pady=5)\n self.transvar = StringVar()\n self.transcbo = ttk.Combobox(self.guiframe, textvariable=self.transvar, font=(\"Arial\", 10))\n self.transcbo.grid(row=6, column=1, sticky=W, padx=5, pady=5)\n\n\n # FUEL TYPE\n self.fuellbl = Label(self.guiframe, text=\"Fuel Type\", font=(\"Arial\", 10)).grid(row=7, sticky=W, padx=5, pady=5)\n self.fuelvar = StringVar()\n self.fuelcbo = ttk.Combobox(self.guiframe, textvariable=self.fuelvar, font=(\"Arial\", 10))\n self.fuelcbo.grid(row=7, column=1, sticky=W, padx=5, pady=5)\n\n # OUTPUT REPORT BUTTON \n self.btnoutput = Button(self.guiframe, text=\"Output Report\", font=(\"Arial\", 10), width=15, command=self.outputReport).\\\n grid(row=8, column=1, sticky=W, padx=10, pady=5)\n\n def outputReport(self):\n strFilter = \"1=1\"\n params = []\n\n if self.makecbo.get() != \"\": \n strFilter = strFilter + \" AND [Make] = ?\"\n params.append(self.makecbo.get())\n \n if self.modelcbo.get() != \"\": \n strFilter = strFilter + \" AND [Model] = ?\"\n params.append(self.modelcbo.get())\n \n if self.yearEndvar.get() > 0 and self.yearEndvar.get() > 0: \n strFilter = strFilter + \" AND [YEAR] BETWEEN ? and ?\"\n params.append(self.yearStartvar.get())\n params.append(self.yearEndvar.get())\n\n if self.engcbo.get() != \"\": \n strFilter = strFilter + \" AND [engId] = ?\"\n params.append(self.engcbo.get())\n \n if self.transcbo.get() != \"\": \n strFilter = strFilter + \" AND [trany] = ?\"\n params.append(self.transcbo.get())\n \n if self.fuelcbo.get() != \"\": \n strFilter = strFilter + \" AND [fuelType] = ?\"\n params.append(self.fuelcbo.get())\n \n strSQL = \"SELECT Vehicles.Make, Vehicles.Model, Vehicles.Year,\" + \\\n \" Vehicles.EngId, Vehicles.Trany, Vehicles.FuelType,\" + \\\n \" Avg(Vehicles.barrels08) AS BarrelsAvg, Avg(Vehicles.city08) AS CityMPGAvg, \" + \\\n \" Avg(Vehicles.comb08) AS CombinedMPGAvg, Avg(Vehicles.highway08) AS HighwayMPGAvg, \" + \\\n \" Avg(Vehicles.fuelCost08) AS FuelCostAvg\" + \\\n \" FROM Vehicles\" + \\\n \" WHERE \" + strFilter + \\\n \" GROUP BY Vehicles.make, Vehicles.model, Vehicles.Year,\" + \\\n \" Vehicles.engId, Vehicles.trany, Vehicles.fuelType;\"\n \n cur.execute(strSQL, params)\n\n tempcsv = os.path.join(cd, 'DATA', 'GUIData_PY.csv')\n with open(tempcsv, 'w', newline='') as csvfile:\n writer = csv.writer(csvfile)\n writer.writerow([i[0] for i in cur.description])\n for row in cur.fetchall(): \n writer.writerow(row)\n\n messagebox.showinfo(\"SUCCESFUL OUTPUT\", \"Successfully outputted query report to csv!\") \n \n\nGUIapp()\ncur.close()\nconn.close()\n", "sub_path": "GUI_Program_PY.py", "file_name": "GUI_Program_PY.py", "file_ext": "py", "file_size_in_byte": 7415, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.dirname", "line_number": 6, "usage_type": "call"}, {"api_name": "os.path", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 6, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 57, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 57, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 67, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 67, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 85, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 85, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 91, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 91, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 98, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 98, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path", "line_number": 146, "usage_type": "attribute"}, {"api_name": "csv.writer", "line_number": 148, "usage_type": "call"}]} +{"seq_id": "340336316", "text": "from time import time\nimport pickle\nimport numpy as np\nfrom math import *\nimport sys\nimport os\nimport matplotlib.pyplot as plt\nfrom matplotlib import cm, rc\nfrom matplotlib.patches import Rectangle\nfrom matplotlib.ticker import MultipleLocator, FixedLocator, FixedFormatter\nfrom gatspy.periodic import RRLyraeTemplateModeler, LombScargleFast, LombScargle\nimport gatspy.datasets as datasets\nfrom ftperiodogram.modeler import FastTemplatePeriodogram, SlowTemplatePeriodogram\nfrom ftperiodogram.template import Template\nfrom scipy.interpolate import interp1d\nfrom scipy.stats import pearsonr\nfrom collections import namedtuple\n\nplt.style.use('./ggplot-like.mplstyle')\nrc('text', usetex=False)\nrc('font', **{'family' : 'sans-serif', 'sans-serif' : [ 'cmss10' ]})\n\n\ndefault_settings = dict(\n wspace = 0.02,\n left = 0.12,\n bottom = 0.12,\n right = 0.9,\n top = 0.9,\n locator_axtmp = 0.5,\n locator_axpdg = None,\n ratio_pdglen_tmplen = 5,\n keep_grid_sizes_equal = True,\n title_pad = 0.01,\n label_fontsize = 12,\n title_fontsize = 12,\n tick_fontsize = 10,\n annotation_fontsize = 12,\n font_color = '#555555',\n colorfunc_a = 0.8,\n colorfunc_b = 0.2,\n answer_color = 'r',\n annotate_x0 = -0.06,\n axsize = 5,\n tmp_height_frac = 0.6,\n data_params = dict(fmt='o', ecolor='0.6',\n markeredgecolor='none', markersize=3,\n markerfacecolor='k', capsize=0, linewidth=1),\n\n )\n\nrms = lambda y : sqrt(np.mean(np.power(y, 2)))\nrror = lambda ym, y : rms(y - ym) / rms(y)\n\ndef get_boundaries_for_axtmp_and_axpdg(maxfrq, settings):\n possible_gridsizes = [ 1, 2, 5, 10, 15, 20, 25, 50, 100 ]\n\n dx_all = settings['right'] - settings['left']\n ltmp, lpdg = None, None\n if settings['keep_grid_sizes_equal']:\n ratio = lambda g2 : settings['locator_axtmp'] * int(maxfrq / g2)\n i = np.argmin([ abs(ratio(g2) - settings['ratio_pdglen_tmplen'])\n for g2 in possible_gridsizes ])\n\n settings['locator_axpdg'] = possible_gridsizes[i]\n\n ltmp = (dx_all - settings['wspace']) / (1 + ratio(settings['locator_axpdg']))\n else:\n ltmp = (dx_all - settings['wspace']) / (1 + settings['ratio_pdglen_tmplen'])\n\n lpdg = settings['right'] - settings['left'] - settings['wspace'] - ltmp\n\n\n bounds_ax_tmp = ( settings['left'],\n settings['bottom'],\n ltmp,\n settings['top'] - settings['bottom']\n )\n bounds_ax_pdg = ( settings['left'] + bounds_ax_tmp[2] + settings['wspace'],\n settings['bottom'],\n settings['right'] - settings['left'] \\\n - bounds_ax_tmp[2] - 0.5 * settings['wspace'],\n settings['top'] - settings['bottom'] )\n\n return bounds_ax_tmp, bounds_ax_pdg\n\n\ndef flatten(iterable):\n out = []\n for i in iterable:\n if hasattr(i,'__iter__'):\n out.extend(flatten(i))\n else:\n out.append(i)\n return out\n\ndef generate_random_signal(n, sigma=1.0, freq=1.0, template=None):\n x = np.sort(np.random.rand(n))\n dy = sigma * np.random.normal(size=n, loc=0)\n err = np.ones_like(x) * sigma\n\n y = np.cos(freq * x)\n if not template is None:\n y = template(x * freq) + dy\n\n return x, y, err\n\n\ndef adjust_figure(f, settings):\n keywords = ['left', 'right', 'top', 'bottom', 'hspace', 'wspace']\n keywords = list(filter(lambda a : a in settings, keywords))\n\n if len(keywords) > 0:\n kwargs = { key : settings[key] for key in keywords }\n f.subplots_adjust(**kwargs)\n\ndef translate_color(ax, settings):\n for key, value in settings.iteritems():\n if isinstance(value, dict):\n settings[key] = translate_color(ax, value)\n\n if value == 'axes_background':\n settings[key] = ax.get_facecolor()\n return settings\n\ndef clean_up_figure(f, settings):\n\n # save\n if not settings['fname'] is None:\n fname = os.path.join(settings['plot_dir'], settings['fname'])\n #f.savefig(fname, dpi=settings['dpi'])\n f.savefig(fname)\n\ndef clean_up_axis(ax, settings):\n\n translate_color(ax, settings)\n\n # format x tick labels\n [ label.set_fontsize(settings['tick_fontsize']) and \\\n label.set_fontfamily('sans-serif') \\\n for label in ax.get_xticklabels()]\n\n # format y tick labels\n [ label.set_fontsize(settings['tick_fontsize']) and \\\n label.set_fontfamily('sans-serif') \\\n for label in ax.get_yticklabels()]\n\n # remove ticks\n if settings['remove_ticks']:\n ax.tick_params(axis='both', which='both', length=0)\n\n # make sure legend font\n legend = ax.get_legend()\n if not legend is None:\n for txt in legend.get_texts():\n txt.set_color(settings['font_color'])\n txt.set_family('sans-serif')\n txt.set_fontsize(settings['legend_fontsize'])\n\n legend.set_frame_on(settings['legend_frameon'])\n\n legend.get_frame().set_facecolor(settings['legend_facecolor'])\n legend.get_frame().set_edgecolor(settings['legend_edgecolor'])\n\n ax.set_xlabel(ax.get_xlabel(), color=settings['font_color'],\n fontsize=settings['label_fontsize'])\n ax.set_ylabel(ax.get_ylabel(), color=settings['font_color'],\n fontsize=settings['label_fontsize'])\n\n\ndef open_results(name, settings, mode):\n fname = os.path.join(settings['results_dir'], settings[name])\n return open(fname, mode)\n\n\n\ndef get_default_template(nharmonics=5):\n rrl_templates = datasets.rrlyrae.fetch_rrlyrae_templates()\n xt, yt = rrl_templates.get_template('100r')\n\n template = Template.from_sampled(yt, nharmonics=nharmonics)\n return template\n\n\ndef get_timing_vs_nharmonics(x, y, yerr, hvals, filename=None, overwrite=True,\n only_use_saved_data = False):\n\n # load old results\n results = {}\n if not filename is None and os.path.exists(filename):\n old_results = pickle.load(open(filename, 'rb'))\n results.update(old_results)\n\n # return if nothing to do\n if all([ h in results for h in hvals ]):\n return hvals, [ results[h] for h in hvals ]\n\n if only_use_saved_data:\n return zip(*[ (h, results[h]) for h in hvals if h in results ])\n\n for h in hvals:\n if h in results:\n continue\n\n print(\" H = \", h)\n\n template = get_default_template(nharmonics=h)\n #template.precompute()\n\n model = FastTemplatePeriodogram(template=template)\n model.fit(x, y, yerr)\n\n t0 = time()\n model.autopower()\n results[h] = time() - t0\n\n print(\" %.4f seconds\"%(results[h]))\n if not filename is None and overwrite:\n pickle.dump(results, open(filename, 'wb'))\n\n return hvals, [ results[h] for h in hvals ]\n\ndef get_timing_lombscargle(n, ofac, hfac):\n\n x, y, dy = generate_random_signal(n)\n\n Nf = int(floor(0.5 * len(x) * ofac * hfac))\n df = 1./(ofac * (max(x) - min(x)))\n f0 = df\n\n model = LombScargleFast(silence_warnings=True)\n model.fit(x, y, dy)\n t0 = time()\n model.score_frequency_grid(f0, df, Nf)\n dt = time() - t0\n\n return dt\n\ndef sort_timing_dict(tdict, col='ndata'):\n keys = tdict.keys()\n inds = np.argsort(tdict[col])\n for key in keys:\n tdict[key] = [ tdict[key][i] for i in inds ]\n return tdict\n\ndef select_from_dict(tdict, values, col='ndata'):\n tcopy = {}\n tcopy.update(tdict)\n\n inds = [ i for i in range(len(tcopy[col])) if tcopy[col][i] in values ]\n\n for key in tcopy:\n tcopy[key] = [ tcopy[key][i] for i in inds ]\n\n return tcopy\n\ndef get_timing_vs_ndata(nvals, nharmonics, filename=None, overwrite=True,\n time_gatspy=True, only_use_saved_data=False, time_lomb_scargle=False):\n #if template is None:\n template = get_default_template(nharmonics=nharmonics)\n #template.precompute()\n\n # load saved results\n results = {}\n if not filename is None and os.path.exists(filename):\n old_results = pickle.load(open(filename, 'rb'))\n results.update(old_results)\n else:\n results = { name : [] for name in [ 'nfreqs', 'ndata', 'tftp', 'tgats' ]}\n\n if only_use_saved_data:\n return select_from_dict(results, nvals)\n\n # return if nothing to do\n if all([ n in results for n in nvals ]):\n return [ results[n] for n in nvals ]\n\n for n in nvals:\n if n in results['ndata']:\n continue\n\n results['ndata'].append(n)\n\n x, y, dy = generate_random_signal(n)\n\n # time FTP\n print(\"timing: n = %d, h = %d, ftp\"%(n, nharmonics))\n\n model = FastTemplatePeriodogram(template=template)\n model.fit(x, y, dy)\n\n t0 = time()\n frq, p = model.autopower()\n results['tftp'].append( time() - t0 )\n\n print(\" done in %.4f seconds\"%(results['tftp'][-1]))\n results['nfreqs'].append(len(frq))\n\n if time_gatspy:\n # time GATSPY\n print(\"timing: n = %d, gatspy\"%(n))\n\n model = SlowTemplatePeriodogram(template=template)\n model.fit(x, y, dy)\n\n t0 = time()\n p = model.power(frq)\n results['tgats'].append(time() - t0)\n\n print(\" done in %.4f seconds\"%(results['tgats'][-1]))\n else:\n results['tgats'].append(-1)\n\n\n # SORT results\n results = sort_timing_dict(results)\n\n # save\n if not filename is None and overwrite:\n pickle.dump(results, open(filename, 'wb'))\n\n return select_from_dict(results, nvals)\n\ndef get_timing_vs_ndata_at_const_nfreq(nvals, nharmonics, max_freq, filename=None, overwrite=True, time_slow=True):\n\n #if template is None:\n template = get_default_template(nharmonics=nharmonics)\n template.precompute()\n\n # load saved results\n results = {}\n if filename is None:\n filename = './saved_results/timing_results_nh%d_maxfrq%.1f.pkl'%(nharmonics, max_freq)\n if not filename is None and os.path.exists(filename):\n old_results = pickle.load(open(filename, 'rb'))\n results.update(old_results)\n\n\n for n in nvals:\n if n in results:\n continue\n\n x, y, dy = generate_random_signal(n)\n x[0] = 0\n x[-1] = 1\n\n # time FTP\n print(\"timing: n = %d, h = %d, ftp\"%(n, nharmonics))\n\n model = FastTemplatePeriodogram(template=template)\n model.fit(x, y, dy)\n\n t0 = time()\n frq, p = model.autopower(maximum_frequency=max_freq)\n tftp = time() - t0\n\n print(\" done in %.4f seconds\"%(tftp))\n\n if time_slow:\n print(\"timing: n = %d, h = %d, slow\"%(n, nharmonics))\n model = SlowTemplatePeriodogram(template=template)\n model.fit(x, y, dy)\n\n t0 = time()\n p = model.power(frq)\n tslow = time() - t0\n\n print(\" done in %.4f seconds\"%(tslow))\n else:\n tslow = -1\n\n results[n] = (tftp, tslow)\n\n # save\n if not filename is None and overwrite:\n pickle.dump(results, open(filename, 'wb'))\n\n return zip(*[ results[n] for n in nvals ])\n\nHVAL_WITH_GATSPY_TIMING_DATA = 6\ndef plot_timing_vs_ndata(settings=default_settings):\n\n f, ax = plt.subplots(figsize=(settings['axsize'], settings['axsize']))\n\n settings = translate_color(ax, settings)\n\n fname = os.path.join(settings['results_dir'], settings['timing_filename'])\n\n nharms = settings['nharmonics']\n if not hasattr(nharms, '__iter__'):\n nharms = [ nharms ]\n\n tls = None\n if settings['plot_lomb_scargle']:\n tls = []\n for n in settings['ndata']:\n tls.append(get_timing_lombscargle(n, 10, 3))\n\n\n for i, h in enumerate(nharms):\n\n time_gatspy = (h == HVAL_WITH_GATSPY_TIMING_DATA)\n filename = fname.replace('.pkl', '_h%d.pkl'%(h))\n timing_data = get_timing_vs_ndata(settings['ndata'], h, filename=filename,\n time_gatspy=time_gatspy,\n only_use_saved_data=settings['only_use_saved_data'])\n\n ndata = np.array(timing_data['ndata'])\n nfreqs = np.array(timing_data['nfreqs'])\n tftp = np.array(timing_data['tftp'])\n tgats = np.array(timing_data['tgats'])\n\n\n\n label = None if h < max(nharms) else 'Fast template periodogram'\n color = settings['scatter_params_ftp']['c']\n lw = settings['linewidth']\n spars = {}\n spars.update(settings['scatter_params_ftp'])\n fudge = 0.7\n ls = '-' if h == max(nharms) else ':'\n #ls = '-'\n q = float(h - min(nharms)) / float(max(nharms) - min(nharms))\n\n spars['alpha'] = fudge * q + (1 - fudge)\n\n ax.scatter(ndata, tftp, label=label, **spars)\n ax.plot(ndata, tftp, color=color, lw=lw, alpha=spars['alpha'], ls=ls)\n\n # now label this\n\n xoffset = 0.08 * (settings['xlim'][1] - settings['xlim'][0])\n\n ax.text(ndata[-1] + xoffset, tftp[-1], \"$H = %d$\"%(h), ha='left', va='center',\n color=settings['font_color'], fontsize=settings['annotation_fontsize'],\n bbox=settings['bbox'])\n\n if time_gatspy:\n\n ax.scatter(ndata, tgats, label='Non-linear optimization', **settings['scatter_params_gatspy'])\n ax.plot(ndata, tgats, color=settings['scatter_params_gatspy']['c'],\n lw=settings['linewidth'])\n\n if not tls is None:\n ax.scatter(ndata, tls, **settings['scatter_params_lomb_scargle'])\n ax.plot(ndata, tls, **settings['line_params_lomb_scargle'])\n ax.plot(ndata, tls, color=ax.get_facecolor(), lw=4, zorder=8)\n\n\n ax.set_xlabel('Number of datapoints')\n ax.set_ylabel(\"Execution time [s]\")\n\n\n ax.set_xscale('log')\n ax.set_yscale('log')\n ax.set_title('Constant cadence')\n ax.set_xlim(*settings['xlim'])\n ax.set_ylim(*settings['ylim'])\n\n ax.legend(loc=settings['legend_loc'])\n\n adjust_figure(f, settings)\n clean_up_axis(ax, settings)\n clean_up_figure(f, settings)\n\ndef plot_timing_vs_ndata_const_freq(settings=default_settings):\n\n f, ax = plt.subplots(figsize=(settings['axsize'], settings['axsize']))\n\n settings = translate_color(ax, settings)\n\n #fname = os.path.join(settings['results_dir'], settings['timing_filename'])\n\n nharms = settings['nharmonics']\n if not hasattr(nharms, '__iter__'):\n nharms = [ nharms ]\n\n\n x, y, dy = generate_random_signal(10)\n x[0] = 0\n x[-1] = 1\n xoffset = 0.08 * (settings['xlim'][1] - settings['xlim'][0])\n nfrq = len(FastTemplatePeriodogram().fit(x, y, dy).autofrequency(maximum_frequency=settings['max_freq']))\n for i, h in enumerate(nharms):\n time_slow = (h==3)\n tftp, tslow = get_timing_vs_ndata_at_const_nfreq(settings['ndata'], h,\n settings['max_freq'], time_slow=time_slow)\n\n\n label = None if h < max(nharms) else 'Fast template periodogram'\n color = settings['scatter_params_ftp']['c']\n lw = settings['linewidth']\n spars = {}\n spars.update(settings['scatter_params_ftp'])\n fudge = 0.7\n ls = '-' if h == max(nharms) else ':'\n #ls = '-'\n q = 1.\n if len(nharms) > 1:\n q = float(h - min(nharms)) / float(max(nharms) - min(nharms))\n\n spars['alpha'] = fudge * q + (1 - fudge)\n\n ax.scatter(settings['ndata'], tftp, label=label, **spars)\n ax.plot(settings['ndata'], tftp, color=color, lw=lw, alpha=spars['alpha'], ls=ls)\n\n # now label this\n\n ax.text(settings['ndata'][-1] + xoffset, tftp[-1], \"$H = %d$\"%(h), ha='left', va='center',\n color=settings['font_color'], fontsize=settings['annotation_fontsize'],\n bbox=settings['bbox'])\n\n if time_slow:\n\n ax.scatter(settings['ndata'], tslow, label='Non-linear optimization', **settings['scatter_params_nonlin'])\n ax.plot(settings['ndata'], tslow, color=settings['scatter_params_nonlin']['c'],\n lw=settings['linewidth'])\n\n\n ax.text(0.05, 0.7, \"$N_f=%d$\"%(nfrq), ha='left', va='top', transform=ax.transAxes)\n ax.set_xlabel('Number of datapoints')\n ax.set_ylabel(\"Execution time [s]\")\n\n ax.set_title('Constant baseline')\n ax.set_xscale('log')\n ax.set_yscale('log')\n\n ax.set_xlim(*settings['xlim'])\n ax.set_ylim(*settings['ylim'])\n\n ax.legend(loc=settings['legend_loc'])\n\n adjust_figure(f, settings)\n clean_up_axis(ax, settings)\n clean_up_figure(f, settings)\n\n\ndef plot_timing_vs_nharmonics(settings=default_settings):\n f, ax = plt.subplots(figsize=(settings['axsize'], settings['axsize']))\n\n settings = translate_color(ax, settings)\n res_fname = lambda n : os.path.join(settings['results_dir'], 'timing_vh_n%d.pkl'%(n))\n\n hvals = settings['hvals']\n nvals = settings['nvals']\n\n if not hasattr(nvals, '__iter__'):\n nvals = [ nvals ]\n\n for n in nvals:\n\n X, Y, Yerr = generate_random_signal(n, 1.0)\n\n nh, dt = get_timing_vs_nharmonics(X, Y, Yerr, hvals, filename=res_fname(n),\n only_use_saved_data=settings['only_use_saved_data'])\n\n #print \"ndata, nharmonics, dlogt / dlogh\"\n #for i, h in enumerate(nh):\n # if i == len(nh) - 1 or i == 0: continue\n\n # #avg = lambda arr, j : (1./3.) * float(arr[j+1] + arr[j] + arr[j+1])\n # avg = lambda arr, j : (1./2.) * float(arr[j+1] + arr[j-1])\n # diff = lambda arr, j : float(arr[j + 1] - arr[j - 1])\n\n # dlogdt = diff(dt, i) / avg(dt, i)\n # dlogh = diff(hvals, i) / avg(hvals, i)\n\n # print n, h, dlogdt/dlogh\n\n #print settings['bbox']\n\n y = np.array(dt)\n nf = int(0.5 * 5 * 5 * n)\n\n if settings['divide_by'] == \"NH\":\n y /= nf * np.array(nh)\n elif settings['divide_by'] == \"N\":\n y /= nf\n\n #print settings['scatter_params']\n spars = settings['scatter_params'][\"%d\"%(n)]\n zorder = spars['zorder']\n\n ax.scatter(nh, y, label=\"$N_{\\\\rm obs} = %d$\\n$N_f=%d$\"%(n, nf), **spars)\n lpars = settings['line_params']['%d'%(n)]\n\n ax.plot(nh, y, zorder = zorder - 1, **lpars)\n #if lpars['ls'] in [ ':', '--', '-.' ]:\n ax.plot(nh, y, zorder = 7, color=ax.get_facecolor(), lw=8)\n\n # add label for normalization\n divtxt = \" / $(H \\\\times N_f)$\" if settings['divide_by'] == \"NH\"\\\n else (\" / $N_f$\" if settings['divide_by'] == \"N\" \\\n else \"\")\n\n # add axis labels\n ax.set_xlabel(\"$H$ (number of harmonics)\")\n ax.set_ylabel(\"Exec. time [s]{divtxt}\".format(divtxt=divtxt))\n\n ax.xaxis.set_major_locator(MultipleLocator(5))\n\n if settings[\"xlog\"]:\n ax.set_xscale('log')\n if settings[\"ylog\"]:\n ax.set_yscale('log')\n\n ax.set_xlim(*settings['xlim'])\n\n\n ylim = None\n if settings['keep_dlogy_equal_to_dlogx']:\n dlogx = log10(settings['xlim'][1]) - log10(settings['xlim'][0])\n ylim = (settings['ylim'][0], pow(10, log10(settings['ylim'][0]) + dlogx))\n else:\n ylim = settings['ylim']\n ax.set_ylim(*ylim)\n\n rb = settings['region_boundaries']\n region_boundaries = [ (rb[2*i], rb[2*i+1]) for i in range(len(rb)//2) ]\n region_titles = settings['region_titles']\n\n\n\n for i, (title, boundary) in enumerate(zip(region_titles, region_boundaries)):\n\n\n xmin, xmax = ax.get_xlim()\n ymin, ymax = ax.get_ylim()\n\n b0 = max([ boundary[0], xmin ])\n b1 = min([ boundary[1], xmax ])\n\n if boundary[0] > xmin and boundary[1] < xmax:\n if i + 1 == settings['highlight_region']:\n ax.fill_between([ boundary[0], boundary[1] ], [ ymin, ymin ], [ymax, ymax],\n edgecolor=settings['font_color'], hatch='////', zorder= 2,\n linewidth=1.0, linestyle=':', lw=0.5, facecolor=\"none\", alpha=1)\n\n #ax.axvspan(boundary[0], boundary[1], edgecolor=settings['font_color'], ls=':',\n # facecolor='none', fill=False, lw=1)\n else:\n ax.axvline(boundary[1], ls='-', color=ax.get_facecolor(), lw=3)\n ax.axvline(boundary[1], ls=':', color=settings['font_color'])\n\n translog = lambda a : log10(a)\n invtranslog = lambda a : pow(10, a)\n\n translin = lambda a : a\n invtranslin = lambda a : a\n\n transx, invtransx = (translog, invtranslog) if settings['xlog'] \\\n else (translin, invtranslin)\n transy, invtransy = (translog, invtranslog) if settings['ylog'] \\\n else (translin, invtranslin)\n\n dx = transx(b1) - transx(b0)\n dxfig = transx(xmax) - transx(xmin)\n\n xtext = (transx(float(b0)) - transx(xmin) + settings['xtext'] * dx ) / dxfig \\\n + settings['xtext_offset']\n\n ytext = settings['ytext']\n if dx / dxfig < 0.2:\n xcoord0 = invtransx(transx(b0) + 0.5 * (transx(b1) - transx(b0)))\n ycoord0 = invtransy(transy(b0) + 0.5 * (transy(b1) - transy(b0)))\n\n coord0 = (xcoord0, ycoord0)\n\n arrowprops = dict(ec=ax.get_facecolor(), fc=settings['font_color'],\n lw=1.5, arrowstyle='simple')\n print(settings['bbox'])\n ax.annotate(title, xy=coord0, xycoords='data', xytext=tuple(settings['xyoffset']),\n textcoords='offset points',\n horizontalalignment=settings['xtext_ha'], verticalalignment='bottom',\n arrowprops=arrowprops, color=settings['font_color'],\n fontsize=settings['annotation_fontsize'], bbox=settings['bbox'])\n else:\n ax.text(xtext, ytext, title, transform=ax.transAxes, ha=settings['xtext_ha'], va='center',\n fontsize=settings['annotation_fontsize'],\n color=settings['font_color'], bbox=settings['bbox'])\n\n\n ax.legend(loc=settings['legend_loc'], ncol=2, mode='expand')\n #ax.legend(loc=3, bbox_to_anchor=(0., 1.02, 1., .102), mode='expand')\n\n clean_up_axis(ax, settings)\n adjust_figure(f, settings)\n clean_up_figure(f, settings)\n\ndef plot_accuracy(x, y, yerr, y_temp, nharmonics, compare_with=10, settings=default_settings):\n\n template = Template.from_sampled(y_temp, nharmonics=10)\n\n # if comparing to large nharmonics, set template now\n if isinstance(compare_with, float) or isinstance(compare_with, int):\n template = Template.from_sampled(y_temp, nharmonics=int(compare_with))\n #template.nharmonics = int(compare_with)\n template.precompute()\n\n # Set the reference model\n true_model = SlowTemplatePeriodogram(template=template) \\\n if compare_with == 'slow_version' \\\n else FastTemplatePeriodogram(template=template)\n\n # fit data\n true_model.fit(x, y, yerr)\n\n results, frq, p_ans = {}, None, None\n label_formatter = lambda kind, h=None : \\\n \"$P_{\\\\rm %s}(\\\\omega%s)$\"\\\n %(kind, \"\" if h is None else \"|H=%d\"%(h))\n corrlabel = lambda R : \"$R = %.3f$\"%(R)\n\n # store results from the reference model\n # (if the reference model is FastTemplatePeriodogram)\n if isinstance(true_model, FastTemplatePeriodogram):\n frq, p_ans = true_model.autopower()\n results = { 'ans' : (frq, p_ans) }\n\n # add results from all desired harmonics\n for h in nharmonics:\n\n # set template harmonics\n #template.nharmonics = h\n template = Template.from_sampled(y_temp, nharmonics=h)\n template.precompute()\n\n # create & fit modeler\n model = FastTemplatePeriodogram(template=template)\n model.fit(x, y, yerr)\n\n # compute periodogram\n results[h] = model.autopower()\n\n # compute results for reference model\n # (if the reference model is gatspy)\n if not 'ans' in results and isinstance(true_model, SlowTemplatePeriodogram):\n frq = results[h][0]\n p_ans = true_model.power(frq)\n results['ans'] = (frq, p_ans)\n\n # Set up plot geometry\n nplots = len(nharmonics)\n if not settings['nrows'] is None:\n nrows = settings['nrows']\n else:\n nrows = max([ int(sqrt(nplots)), 1 ])\n\n ncols = 1\n while ncols * nrows < nplots:\n ncols += 1\n\n figsize = (settings['axsize'] * ncols, settings['axsize'] * nrows)\n\n # create figure\n f, axes = plt.subplots(nrows, ncols, figsize=figsize)\n\n # ensure we have a list of axes\n if not hasattr(axes, '__iter__'):\n axes = [ axes ]\n\n # some plotting definitions\n ans_label = label_formatter('slow') if compare_with == 'slow_version'\\\n else label_formatter('FTP', int(compare_with))\n\n ax_label_params = dict(fontsize=settings['label_fontsize'],\n color=settings['font_color'])\n ax_annotation_params = dict(fontsize=settings['annotation_fontsize'],\n color=settings['font_color'], bbox=settings['bbox'])\n scatter_params = settings['scatter_params']\n #print scatter_params\n\n for i, h in enumerate(nharmonics):\n # select the axes instance\n r, c = i / ncols, i % ncols\n ax = axes[c] if ncols >= 1 and nrows == 1 else axes[r][c]\n\n settings = translate_color(ax, settings)\n\n p = results[h][1]\n\n # scatterplot\n ax.plot([0, 1], [0, 1], ls=':', color=settings['font_color'], zorder=2)\n ax.scatter(p, p_ans, **scatter_params)\n\n # write the pearson R correlation\n ax.text(0.05, 0.9, corrlabel(pearsonr(p_ans, p)[0]), transform=ax.transAxes,\n ha='left', va='bottom', zorder=10, **ax_annotation_params)\n\n # many of the gatspy periodogram values are 0;\n # write pearson R using only non-zero P_gatspy values\n #if compare_with == 'slow_version':\n # nonzero_p_ans, nonzero_p = zip(*filter(lambda (Pans, P) : Pans > 0, zip(p_ans, p)))\n # Rnonzero = pearsonr(nonzero_p_ans, nonzero_p)[0]\n\n\n # ax.text(0.05, 0.9 - 0.03, \"%s; $P_{\\\\rm non-lin. opt.} > 0$\"\\\n # %(corrlabel(Rnonzero)),zorder=10,\n # transform=ax.transAxes, ha='left', va='top', **ax_annotation_params)\n\n # set plot properties\n ax.set_xlabel(label_formatter('FTP', h), **ax_label_params)\n if c == 0:\n ax.set_ylabel(ans_label, **ax_label_params)\n else:\n [ label.set_visible(False) for label in ax.get_yticklabels() ]\n\n ax.set_xlim(settings['x_and_y_min'], 1)\n ax.set_ylim(settings['x_and_y_min'], 1)\n\n name_list = [ \"0\", \"0.25\", \"0.5\", \"0.75\", \"1\" ]\n pos_list = [ 0, 0.25, 0.5, 0.75, 1 ]\n ax.xaxis.set_major_locator(FixedLocator((pos_list)))\n ax.xaxis.set_major_formatter(FixedFormatter((name_list)))\n\n ax.yaxis.set_major_locator(FixedLocator((pos_list)))\n ax.yaxis.set_major_formatter(FixedFormatter((name_list)))\n\n clean_up_axis(ax, settings)\n\n adjust_figure(f, settings)\n clean_up_figure(f, settings)\n\ndef get_multiharmonic_periodogram(x, y, err, nh, hfac=3, ofac=10):\n model = LombScargle(Nterms=nh)\n model.fit(x,y,err)\n\n pers, p = model.periodogram_auto(nyquist_factor=hfac, oversampling=ofac)\n\n return np.power(pers, -1), p\n\ndef get_bls_periodogram(x, y, err, use_pybls=False, hfac=3, ofac=10, nbin=50, qmin=0.001, qmax=0.5):\n\n df = 1./(ofac * (max(x) - min(x)))\n Nf = int(floor(0.5 * len(x) * ofac * hfac))\n f0 = 2./(max(x) - min(x))\n\n if not use_pybls:\n import pyeebls as bls\n #import bls\n u, v = np.empty(len(x)), np.empty(len(x))\n result = bls.eebls(x, y, u, v, Nf, f0, df, nbin, qmin, qmax)\n periodogram, best_period, best_power, depth, q, in1, in2 = result\n else:\n from pybls import BLS\n bls = BLS(x, y, err, fmin=f0, nf=Nf, df=df, nbin=nbin, wmin=qmin, qmax=qmax )\n periodogram = bls().p\n\n\n periodogram -= min(periodogram)\n periodogram /= max(periodogram)\n\n frq = np.linspace(f0, f0 + Nf * df, Nf)\n\n return frq, periodogram\n\ndef plot_templates_and_periodograms(x, y, err, y_temp, freq_val=None, hfac=None, ofac=None,\n nharms = None, settings=default_settings):\n\n\n p_ftps = []\n phi_data = None if freq_val is None else (x * freq_val - settings['phi0']) % 1.0\n for i, nharm in enumerate(nharms):\n\n #model.templates.values()[0].nharmonics = nharm\n #model.templates.values()[0].precompute()\n template = Template.from_sampled(y_temp, nharmonics = nharm)\n template.precompute()\n model = FastTemplatePeriodogram(template=template)\n\n # Run FTP\n model.fit(x, y, err)\n\n frq, p = model.autopower(samples_per_peak=ofac, nyquist_factor=hfac)\n\n p_ftps.append(p)\n\n # get axes boundaries\n bounds_ax_tmp, bounds_ax_pdg = \\\n get_boundaries_for_axtmp_and_axpdg(max(frq), settings)\n\n # initialize figure\n f = plt.figure(figsize=(2 * settings['axsize'], settings['axsize']))\n ax_pdg = f.add_axes(bounds_ax_pdg)\n ax_tmp = f.add_axes(bounds_ax_tmp)\n\n settings = translate_color(ax_pdg, settings)\n\n # get full template\n phi0 = np.linspace(0, 1, len(y_temp))\n y0 = y_temp\n ymin, ymax = min(-y0), max(-y0)\n\n # x position for text (H = ...); has to be in ax_tmp data coordinates\n x0 = settings['annotate_x0'] / bounds_ax_tmp[2]\n\n # functions for normalizing templates\n tmpnorm = settings['tmp_height_frac'] / (ymax - ymin)\n yoffset = 0.5 * (1 - settings['tmp_height_frac'])\n tmp_transform = lambda yt, et : ( (-yt - ymin) * tmpnorm + yoffset,\n et * tmpnorm if not et is None else None )\n\n # normalize data and template\n ydata, edata = tmp_transform(y, err)\n y0, _ = tmp_transform(y0, None)\n\n #colorfunc = lambda i : \"%.5f\"%(settings['colorfunc_a'] * (float(len(nharms) - i - 1) / float(len(nharms) - 1)) \\\n # + settings['colorfunc_b']) if i < len(nharms) - 1 \\\n # else settings['answer_color']\n colorfunc = lambda i : settings['answer_color']\n for i, (p, h) in enumerate(zip(p_ftps, nharms)):\n offset = len(p_ftps) - i - 1\n ytext = offset + 0.5\n lw = 1 if i < len(p_ftps) - 1 else 1\n\n # plot periodogram\n ax_pdg.plot(frq, p + offset, color=colorfunc(i), lw=lw, zorder=20)\n ax_pdg.plot(frq, p + offset, color=ax_pdg.get_facecolor(), lw=3, zorder=19)\n\n if (i == len(p_ftps) - 1) and settings['plot_bls']:\n frq_bls, p_bls = get_bls_periodogram(x, y, err, hfac=3, ofac=10)\n ax_pdg.plot(frq_bls, p_bls + offset, color=settings['color_bls'],\n alpha=settings['alpha_bls'], zorder=18)\n ax_pdg.plot(frq_bls, p_bls + offset, color=ax_pdg.get_facecolor(),\n zorder=17, lw=3)\n\n ax_pdg.text(0.02, 0.92, \"Box Least Squares\", color=settings['color_bls'],\n ha='left',\n va='top', bbox=settings['bbox'], fontsize=settings['annotation_fontsize'],\n transform=ax_pdg.transAxes)\n\n\n if (i > 0 and settings['plot_multiharmonic_periodogram']):\n # plot multiharmonic periodogram\n frq_mh, p_mh = get_multiharmonic_periodogram(x, y, err, h)\n\n color = settings['color_multiharmonic_periodogram']\n alpha = settings['alpha_multiharmonic_periodogram']\n ax_pdg.plot(frq_mh, p_mh + offset, color=color, alpha=alpha, zorder=16)\n ax_pdg.plot(frq_mh, p_mh + offset, color=ax_pdg.get_facecolor(), zorder=15, lw=3)\n\n if i == 1:\n df0 = (2 * freq_val - min(frq_mh)) / (max(frq_mh) - min(frq_mh))\n ind = int(df0 * len(frq_mh))\n dx = max(ax_pdg.get_xlim()) - min(ax_pdg.get_xlim())\n #ax_pdg.annotate('Multiharmonic Lomb Scargle',\n # xy = (frq_mh[ind], p_mh[ind] + offset + 0.03),\n # xycoords = 'data',\n # xytext = (frq_mh[ind] - 0.015 * dx, 1.45 + offset),\n # textcoords = 'data',\n # horizontalalignment = 'right',\n # verticalalignment = 'bottom',\n # color = settings['font_color'],\n # arrowprops = dict(ec=ax_pdg.get_facecolor(), fc=settings['font_color'],\n # lw=1.5, arrowstyle='simple'),\n # fontsize = settings['annotation_fontsize'],\n # bbox = settings['bbox'])\n ax_pdg.text(0.02, 0.98, \"Multi-harmonic Lomb Scargle\",\n fontsize=settings['annotation_fontsize'],\n bbox =settings['bbox'],\n color = color, ha='left', va='top', transform=ax_pdg.transAxes)\n\n\n\n\n # get truncated template\n #template.nharmonics = h\n template = Template.from_sampled(y_temp, nharmonics=h)\n template.precompute()\n\n phi = np.linspace(0, 1, 100)\n ytmp_trunc = template(phi)\n\n # normalize truncated template\n ytmp_trunc, _ = tmp_transform(ytmp_trunc, None)\n\n # plot truncated template\n ax_tmp.plot(phi, ytmp_trunc + offset, color=colorfunc(i), lw=lw)\n\n # plot data\n ax_tmp.errorbar(phi_data, ydata + offset, yerr=edata,\n **settings['data_params'])\n\n # write H = ...\n ax_tmp.text(x0, ytext, \"H = %d\"%(h), va='center', ha='left',\n color=settings['font_color'],\n fontsize=settings['label_fontsize'])\n\n\n # Write axis labels\n ax_tmp.set_xlabel('Phase')\n ax_pdg.set_xlabel('Frequency $[d^{-1}]$')\n\n # Draw line for correct frequency\n if freq_val is not None:\n ax_pdg.axvline(freq_val, ls=':', color='k')\n\n # Write titles\n ytitle = settings['top'] + settings['title_pad']\n xtitle_pdg = settings['left'] + bounds_ax_tmp[2] + settings['wspace'] \\\n + 0.5 * bounds_ax_pdg[2]\n xtitle_tmp = settings['left'] + 0.5 * bounds_ax_tmp[2]\n\n f.text(xtitle_pdg, ytitle,\n \"Template periodogram\", va='bottom', ha='center',\n color=settings['font_color'], fontsize=settings['title_fontsize'])\n\n f.text(xtitle_tmp, ytitle,\n \"Template fits\", va='bottom', ha='center',\n color=settings['font_color'], fontsize=settings['title_fontsize'])\n\n # Set other properties\n ax_pdg.set_xlim(0, int(max(frq) / settings['locator_axpdg'])\\\n * settings['locator_axpdg'])\n ax_pdg.set_ylim(0, len(nharms))\n\n ax_pdg.xaxis.set_major_locator(MultipleLocator(settings['locator_axpdg']))\n\n ax_tmp.set_xlim(0, 1)\n ax_tmp.set_ylim(*ax_pdg.get_ylim())\n ax_tmp.set_yticks(ax_pdg.get_yticks())\n ax_tmp.xaxis.set_major_locator(MultipleLocator(settings['locator_axtmp']))\n\n # for both axes...\n for ax in [ ax_pdg, ax_tmp ]:\n # turn of ytick labels\n [ label.set_visible(False) for label in ax.get_yticklabels()]\n ax.yaxis.set_major_locator(MultipleLocator(1))\n\n clean_up_axis(ax, settings)\n\n clean_up_figure(f, settings)\n\ndef set_defaults_in_section(settings, defaults):\n #print settings\n for dvar, dval in defaults.iteritems():\n if not dvar in settings:\n settings[dvar] = dval\n elif isinstance(dval, dict):\n settings[dvar] = set_defaults_in_section(settings[dvar], dval)\n return settings\n\ndef set_defaults(settings, defaults):\n for group, subsettings in settings.iteritems():\n settings[group] = set_defaults_in_section(subsettings, defaults)\n return settings\n\ndef dictprint(d, indent=\"\"):\n\n for key, value in d.iteritems():\n s = \"%s%s-%ds\"%(indent, \"%\", 50 - len(indent))%(key)\n #s = \"{fmt}\".format(indent=indent, fmt=fmt)\n if isinstance(value, dict):\n print(s)\n dictprint(value, indent = \"%s \"%(indent))\n else:\n print(\"{s}{value}\".format(s=s, value=value))\n\ndef inject_transit(x, y, q=0.2, freq=100, depth=1):\n phi = (x * freq) % (1.0)\n for i, ph in enumerate(phi):\n if ph > 0.5 * (1 - q) and ph < 0.5 * (1 + q):\n y[i] -= depth\n\n return y\n\ndef test_bls(n = 1000):\n freq = 200\n depth = 1\n q = 0.2\n\n x, y, err = generate_random_signal(n)\n\n y = inject_transit(x, y, q=q, freq=freq, depth=depth)\n\n frq, p = get_bls_periodogram(x, y, err, hfac=3, ofac=10)\n\n phi = (x * freq)%(1.0)\n f, (ax_lc, ax_bls) = plt.subplots(1, 2, figsize=(10, 5))\n\n ax_lc.scatter(phi, y, alpha=0.1, marker='o', s=3)\n ax_lc.set_xlim(0, 1)\n ax_lc.set_xlabel('phase')\n ax_lc.set_ylabel('mag')\n\n ax_bls.plot(frq, p, color='k')\n ax_bls.axvline(freq, ls=':', color='k')\n ax_bls.set_xlabel('freq')\n ax_bls.set_ylabel('bls')\n\n plt.show()\n\ndef plot_nobs_dt_for_surveys(settings=default_settings):\n #surveys = ConfigObj('surveys.ini', unrepr=True)\n\n\n surveys_to_use = []\n for survey, info in surveys.iteritems():\n relevant_info = [ 'nobs', 'nlc' ]\n if all([ var in info for var in relevant_info ])\\\n and not any([ info[var] is None for var in relevant_info ])\\\n and not survey in settings['skip_surveys']:\n surveys_to_use.append(survey)\n\n\n f, ax = plt.subplots()\n settings = translate_color(ax, settings)\n\n conversion = 1./(3600.)\n unit = 'CPU hours'\n #for s in surveys_to_use:\n # try:\n # surveys[s]['nobs'] * surveys[s]['nlc']\n # except:\n # print s\n # sys.exit()\n X = [ surveys[s]['nobs'] * surveys[s]['nlc'] for s in surveys_to_use ]\n Yftp = [ conversion * (float(x) / settings['ndata']) * settings['tftp'] for x in X ]\n Ygats = [ conversion * (float(surveys[s]['nobs']) / settings['ndata'])**2 \\\n * surveys[s]['nlc'] * settings['tgats'] for s in surveys_to_use ]\n\n surveys_to_use, X, Yftp, Ygats = zip(*sorted(zip(surveys_to_use, X, Yftp, Ygats), key=lambda stuff : stuff[1] ))\n ax.scatter(X, Yftp, **settings['scatter_params_ftp'])\n ax.scatter(X, Ygats, **settings['scatter_params_gats'])\n ax.set_xscale('log')\n ax.set_yscale('log')\n\n ax.set_xlabel('$N_{\\\\rm obs} \\\\times N_{\\\\rm LC}$')\n ax.set_ylabel(\"Exec time [%s]\"%(unit))\n\n if 'xlim' in settings:\n ax.set_xlim(*settings['xlim'])\n ymin, ymax = ax.get_ylim()\n xmin, xmax = ax.get_xlim()\n\n y0 = settings['y0']\n yf = settings['yf']\n for i, (survey, x) in enumerate(zip(surveys_to_use, X)):\n common_params = dict(va='center', ha='center', color=settings['font_color'],\n fontsize=settings['annotation_fontsize'], bbox=settings['bbox'])\n\n frac = (log10(x) - log10(xmin)) / (log10(xmax) - log10(xmin))\n #ytext = pow(10, ((yf - y0) * frac + y0) * (log10(ymax) - log10(ymin)) + log10(ymin))\n ytext = pow(10, 0.5 * abs(log10(Yftp[i]) - log10(Ygats[i])) \\\n + min([ log10(Yftp[i]), log10(Ygats[i]) ]))\n #u = 1\n #if i < len(surveys_to_use) - 1 and log10(X[i+1]) - log10(X[i]) < 0.75:\n #ytext = pow(10, 0.6 * (log10(ymax) - log10(ymin)) + log10(ymin))\n ax.plot([x, x], [Yftp[i], Ygats[i]], color=settings['scatter_params_ftp']['color'],\n lw=2)\n if i > 0 and log10(X[i]) - log10(X[i-1]) < settings['dlogx_min']:\n\n ax.text(x * (1 + settings['dx_text2']), ytext, survey, rotation=270, **common_params)\n #ha='left', **common_params)\n else:\n ax.text(x * (1 + settings['dx_text']), ytext, survey, rotation='vertical', **common_params)\n #ha='right', **common_params)\n ax.axvline(x, ls=':', color=settings['font_color'], zorder=1)\n ax.legend(loc='upper left')\n\n clean_up_axis(ax, settings)\n clean_up_figure(f, settings)\n\nif __name__ == '__main__':\n\n #test_bls()\n #sys.exit()\n from configobj import ConfigObj\n\n conf = ConfigObj('plotting.ini', unrepr=True)\n defs = conf['DEFAULT']\n\n conf = set_defaults(conf, defs)\n\n\n template_name = conf['template']['template_name']\n tconf = conf['template'][template_name]\n tconf = set_defaults_in_section(tconf, defs)\n dconf = conf['data']\n\n template, cn, sn = None, None, None\n x, y, err = None, None, None\n Ttemp, Ytemp = None, None\n ndata, sigma, freq = dconf['ndata'], dconf['sigma'], dconf['freq']\n ofac, hfac = conf['analysis']['ofac'], conf['analysis']['hfac']\n nharms = conf['analysis']['nharms']\n\n\n nh = [ n for n in nharms ]\n nh.append(tconf['nharm_answer'])\n\n if tconf['fourier']:\n if tconf['template_filename'] is None:\n cn, sn = tconf['cn'], tconf['sn']\n assert(not cn is None and not sn is None)\n else:\n\n cn, sn = pickle.load(open_results('template_filename', tconf, 'rb'))\n\n norm = 1./np.sqrt(sum(np.power(cn, 2) + np.power(sn, 2)))\n cn = np.multiply(cn, norm)\n sn = np.multiply(sn, norm)\n\n nha = tconf['nharm_answer']\n template = Template(c_n=cn[:nha], s_n=sn[:nha])\n\n y_phase = template(np.linspace(0, 1, 100))\n\n x, y, err = generate_random_signal(ndata, sigma, freq=freq,\n template=template)\n\n else:\n if tconf['template_filename'] is None:\n # Obtain template from RR Lyrae dataset\n rrl_templates = datasets.rrlyrae.fetch_rrlyrae_templates()\n\n Ttemp, Ytemp = rrl_templates.get_template(tconf['name'])\n\n else:\n Ttemp, Ytemp = pickle.load(open_results('template_filename', tconf, 'rb'))\n\n template = Template.from_sampled(Ytemp, nharmonics=tconf['nharm_answer'])\n\n x, y, err = generate_random_signal(ndata, sigma, freq=freq, template=template)\n\n\n template.precompute()\n\n # build model\n model = FastTemplatePeriodogram(template=template)\n y_temp = template(np.linspace(0, 1, 100))\n\n print(\"plotting timing vs ndata at constant freq\")\n plot_timing_vs_ndata_const_freq(settings=conf['timing_vs_ndata_const_freq'])\n\n print(\"plotting timing vs nharmonics\")\n plot_timing_vs_nharmonics(conf['timing_vs_nharmonics'])\n\n print(\"plotting timing vs ndata\")\n plot_timing_vs_ndata(conf['timing_vs_ndata'])\n\n #print(\"plotting nobs dt for surveys\")\n #plot_nobs_dt_for_surveys(settings=conf['nobs_dt_for_surveys'])\n\n print(\"plotting templates and periodograms\")\n plot_templates_and_periodograms(x, y, err, y_temp, nharms = nh, freq_val=freq, ofac=ofac, hfac=hfac,\n settings=conf['templates_and_periodograms'])\n\n #print(\"plotting accuracy (self)\")\n #plot_accuracy(x, y, err, y_temp, nharms, compare_with=tconf['nharm_answer'],\n # settings=conf['make_accuracy_plot'])\n\n #print(\"plotting accuracy (non-linear)\")\n #plot_accuracy(x, y, err, y_temp, [ tconf['nharm_answer'] ], compare_with='slow_version',\n # settings=conf['make_accuracy_plot_with_slow_version'])\n\n plt.show()\n", "sub_path": "scripts/generate_plots.py", "file_name": "generate_plots.py", "file_ext": "py", "file_size_in_byte": 43908, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.pyplot.style.use", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 19, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.rc", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.rc", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 98, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 98, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 99, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 99, "usage_type": "attribute"}, {"api_name": "numpy.ones_like", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 130, "usage_type": "call"}, {"api_name": "os.path", "line_number": 130, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 172, "usage_type": "call"}, {"api_name": "os.path", "line_number": 172, "usage_type": "attribute"}, {"api_name": "gatspy.datasets.rrlyrae.fetch_rrlyrae_templates", "line_number": 178, "usage_type": "call"}, {"api_name": "gatspy.datasets.rrlyrae", "line_number": 178, "usage_type": "attribute"}, {"api_name": "gatspy.datasets", "line_number": 178, "usage_type": "name"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 181, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 181, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 191, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 210, "usage_type": "call"}, {"api_name": "time.time", "line_number": 213, "usage_type": "call"}, {"api_name": "time.time", "line_number": 215, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 219, "usage_type": "call"}, {"api_name": "gatspy.periodic.LombScargleFast", "line_number": 231, "usage_type": "call"}, {"api_name": "time.time", "line_number": 233, "usage_type": "call"}, {"api_name": "time.time", "line_number": 235, "usage_type": "call"}, {"api_name": "numpy.argsort", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 265, "usage_type": "call"}, {"api_name": "os.path", "line_number": 265, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 266, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 289, "usage_type": "call"}, {"api_name": "time.time", "line_number": 292, "usage_type": "call"}, {"api_name": "time.time", "line_number": 294, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.SlowTemplatePeriodogram", "line_number": 303, "usage_type": "call"}, {"api_name": "time.time", "line_number": 306, "usage_type": "call"}, {"api_name": "time.time", "line_number": 308, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 334, "usage_type": "call"}, {"api_name": "os.path", "line_number": 334, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 335, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 350, "usage_type": "call"}, {"api_name": "time.time", "line_number": 353, "usage_type": "call"}, {"api_name": "time.time", "line_number": 355, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.SlowTemplatePeriodogram", "line_number": 361, "usage_type": "call"}, {"api_name": "time.time", "line_number": 364, "usage_type": "call"}, {"api_name": "time.time", "line_number": 366, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 376, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 383, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 383, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 387, "usage_type": "call"}, {"api_name": "os.path", "line_number": 387, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 408, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 410, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 411, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 468, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 468, "usage_type": "name"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 483, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 539, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 539, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 542, "usage_type": "call"}, {"api_name": "os.path", "line_number": 542, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 572, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 576, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 600, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 692, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 692, "usage_type": "name"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 696, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 696, "usage_type": "name"}, {"api_name": "ftperiodogram.modeler.SlowTemplatePeriodogram", "line_number": 701, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 703, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 716, "usage_type": "argument"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 725, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 725, "usage_type": "name"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 729, "usage_type": "call"}, {"api_name": "ftperiodogram.modeler.SlowTemplatePeriodogram", "line_number": 737, "usage_type": "argument"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 756, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 756, "usage_type": "name"}, {"api_name": "scipy.stats.pearsonr", "line_number": 787, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FixedLocator", "line_number": 813, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FixedFormatter", "line_number": 814, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FixedLocator", "line_number": 816, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FixedFormatter", "line_number": 817, "usage_type": "call"}, {"api_name": "gatspy.periodic.LombScargle", "line_number": 825, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 830, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 841, "usage_type": "call"}, {"api_name": "pyeebls.eebls", "line_number": 842, "usage_type": "call"}, {"api_name": "pybls.BLS", "line_number": 846, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 853, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 867, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 867, "usage_type": "name"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 869, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 883, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 883, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 890, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 968, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 968, "usage_type": "name"}, {"api_name": "numpy.linspace", "line_number": 971, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 1017, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 1022, "usage_type": "call"}, {"api_name": "matplotlib.ticker.MultipleLocator", "line_number": 1028, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 1079, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1079, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1091, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1091, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 1106, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1106, "usage_type": "name"}, {"api_name": "configobj.ConfigObj", "line_number": 1170, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 1198, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 1200, "usage_type": "call"}, {"api_name": "numpy.power", "line_number": 1200, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1201, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 1202, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 1205, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1207, "usage_type": "call"}, {"api_name": "gatspy.datasets.rrlyrae.fetch_rrlyrae_templates", "line_number": 1215, "usage_type": "call"}, {"api_name": "gatspy.datasets.rrlyrae", "line_number": 1215, "usage_type": "attribute"}, {"api_name": "gatspy.datasets", "line_number": 1215, "usage_type": "name"}, {"api_name": "pickle.load", "line_number": 1220, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template.from_sampled", "line_number": 1222, "usage_type": "call"}, {"api_name": "ftperiodogram.template.Template", "line_number": 1222, "usage_type": "name"}, {"api_name": "ftperiodogram.modeler.FastTemplatePeriodogram", "line_number": 1230, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1231, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 1257, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 1257, "usage_type": "name"}]} +{"seq_id": "135852005", "text": "import requests\nimport logging\nfrom suds import TypeNotFound\n\nlogger = logging.getLogger(__name__)\n\nAPI_VERSION = 'v201705'\nREPORTS_DEFINITIONS = {\n 'BASE_PATH': 'https://developers.google.com/adwords/api/docs/appendix/reports/',\n 'CAMPAIGN_NEGATIVE_KEYWORDS_PERFORMANCE_REPORT': 'campaign-negative-keywords-performance-report.csv',\n 'CRITERIA_PERFORMANCE_REPORT': 'criteria-performance-report.csv',\n 'AD_PERFORMANCE_REPORT': 'ad-performance-report.csv',\n 'KEYWORDS_PERFORMANCE_REPORT': 'keywords-performance-report.csv',\n 'SEARCH_QUERY_PERFORMANCE_REPORT': 'search-query-performance-report.csv',\n 'CAMPAIGN_PERFORMANCE_REPORT': 'campaign-performance-report.csv',\n 'ADGROUP_PERFORMANCE_REPORT': 'adgroup-performance-report.csv',\n 'CAMPAIGN_LOCATION_TARGET_REPORT': 'campaign-location-target-report.csv',\n 'CLICK_PERFORMANCE_REPORT': 'click-performance-report.csv',\n 'BUDGET_PERFORMANCE_REPORT': 'budget-performance-report.csv',\n 'LABEL_REPORT': 'label-report.csv',\n}\n\n\ndef get_report_csv(report_type):\n # TODO: download the csv version associated with the version of the api\n csv_url = '{}{}'.format(REPORTS_DEFINITIONS['BASE_PATH'],\n REPORTS_DEFINITIONS[report_type])\n result = requests.get(csv_url)\n if result.status_code == 200:\n return result\n csv_url = '{}{}/{}'.format(REPORTS_DEFINITIONS['BASE_PATH'],\n API_VERSION,\n REPORTS_DEFINITIONS[report_type])\n return requests.get(csv_url)\n\n\nclass SudsFactory:\n def __init__(self, suds_client):\n self.suds_client = suds_client\n\n def get_object(self, name, namespace=None):\n if namespace:\n wsdl = '{{https://adwords.google.com/api/adwords/{}/{}}}{}'.format(namespace, API_VERSION, name)\n return self.suds_client.factory.create(wsdl)\n else:\n try:\n wsdl = '{{https://adwords.google.com/api/adwords/cm/{}}}{}'.format(API_VERSION, name)\n return self.suds_client.factory.create(wsdl)\n except TypeNotFound:\n pass\n try:\n wsdl = '{{https://adwords.google.com/api/adwords/o/{}}}{}'.format(API_VERSION, name)\n return self.suds_client.factory.create(wsdl)\n except TypeNotFound:\n pass\n raise NameError(name)\n\n\nclass BaseResult:\n def __init__(self, callback, parameters):\n self.callback = callback\n self.callback_parameters = parameters\n self.result = None\n\n def __getattr__(self, item):\n try:\n return self.__getattribute__(item)\n except AttributeError:\n return self.result.__getattribute__(item)\n\n def __contains__(self, item):\n return item in self.result\n\n def __repr__(self):\n return self.result.__repr__()\n\n\nclass SimpleReturnValue(BaseResult):\n def __init__(self, callback, parameters):\n super().__init__(callback, parameters)\n self.result = callback(parameters)\n\n def __iter__(self):\n if 'value' in self:\n for entry in self.value:\n yield entry\n else:\n return iter(())\n\n def __getitem__(self, item):\n if 'value' in self:\n return self.value[item]\n else:\n raise IndexError('value not present')\n\n\nclass SimpleResult(BaseResult):\n def __init__(self, callback, parameters):\n super().__init__(callback, parameters)\n self.result = callback(parameters)\n\n\nclass PagedResult(BaseResult):\n def __iter__(self):\n start_index = self.callback_parameters.paging.startIndex\n original_start_index = start_index\n page_size = self.callback_parameters.paging.numberResults\n more_pages = True\n while more_pages:\n self.result = self.callback(self.callback_parameters)\n if 'entries' in self:\n for entry in self.entries:\n yield entry\n else:\n self.callback_parameters.paging.startIndex = original_start_index\n raise StopIteration\n start_index += page_size\n self.callback_parameters.paging.startIndex = start_index\n more_pages = start_index < self.totalNumEntries\n self.callback_parameters.paging.startIndex = original_start_index\n raise StopIteration\n\n\nclass BaseService(SudsFactory):\n def __init__(self, client, service_name):\n self.client = client\n self.service_name = service_name\n self.service = client.GetService(service_name, version=API_VERSION)\n self.suds_client = self.service.suds_client\n self.helper = None\n self.ResultProcessor = None\n\n # Default selector for get, should be overwritten if necessary.\n def prepare_get(self):\n self.helper = Selector(self.service)\n self.ResultProcessor = PagedResult\n\n def get(self, client_customer_id=None):\n \"\"\"\n\n :param client_customer_id:\n :param number_results:\n :param start_index:\n :param min_date: Default as in\n https://developers.google.com/adwords/api/docs/reference/v201601/DataService.Selector\n :param max_date: Default as in\n https://developers.google.com/adwords/api/docs/reference/v201601/DataService.Selector\n :return:\n \"\"\"\n if client_customer_id:\n self.client.SetClientCustomerId(client_customer_id)\n\n # manually setting the headers and sending the objects\n soap_header = self.get_object('SoapHeader', 'cm')\n soap_header.clientCustomerId = self.client.client_customer_id\n soap_header.developerToken = self.client.developer_token\n soap_header.userAgent = self.client.user_agent\n self.suds_client.set_options(\n soapheaders=soap_header,\n headers=self.client.oauth2_client.CreateHttpHeader())\n\n return self.ResultProcessor(self.suds_client.service.get, self.helper.selector)\n\n # alternative call that uses google's adwords api to package the\n # suds object at the cost of rebuilding the objects\n # return self.get_page(self.service.get(self.helper.selector))\n\n def mutate(self, client_customer_id=None):\n if client_customer_id:\n self.client.SetClientCustomerId(client_customer_id)\n\n # manually setting the headers and sending the objects\n soap_header = self.get_object('SoapHeader', 'cm')\n soap_header.clientCustomerId = self.client.client_customer_id\n soap_header.developerToken = self.client.developer_token\n soap_header.userAgent = self.client.user_agent\n self.suds_client.set_options(\n soapheaders=soap_header,\n headers=self.client.oauth2_client.CreateHttpHeader())\n # return self.ResultProcessor(self.suds_client.service.mutate, self.helper.operations)\n return self.ResultProcessor(self.service.mutate, self.helper.operations)\n # return self.get_result(self.service.mutate(self.helper.operations))\n\n\nclass BaseSelector(SudsFactory):\n def __init__(self, service):\n self.suds_client = service.suds_client\n self.selector = None\n\n def __getattr__(self, item):\n try:\n return self.__getattribute__(item)\n except AttributeError:\n return self.selector.__getattribute__(item)\n\n def __repr__(self):\n return self.selector.__repr__()\n\n\nclass Selector(BaseSelector):\n def __init__(self, service):\n super().__init__(service)\n self.selector = self.get_object('Selector', 'cm')\n delattr(self.selector, 'dateRange')\n self.paging.startIndex = 0\n # maximum number of results allowed by API\n # https://developers.google.com/adwords/api/docs/appendix/limits#general\n self.paging.numberResults = 10000\n\n def add_fields(self, *args):\n self.fields.extend(args)\n\n def add_predicate(self, field, operator, values):\n predicate = self.get_object('Predicate', 'cm')\n predicate.field = field\n predicate.operator = operator\n predicate.values = values\n self.predicates.append(predicate)\n", "sub_path": "adwords_client/adwordsapi/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 8185, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 5, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 34, "usage_type": "call"}, {"api_name": "suds.TypeNotFound", "line_number": 49, "usage_type": "name"}, {"api_name": "suds.TypeNotFound", "line_number": 54, "usage_type": "name"}]} +{"seq_id": "228691003", "text": "from flask import Flask, render_template, request\nimport sys\nimport requests\nimport os\nimport datetime\nimport time\n\n#testing git\n\n\napp = Flask(__name__)\nAPI_KEY = os.environ.get('WEATHER_API_KEY')\n\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n if request.method == 'POST':\n try:\n return add()\n except Exception as ex:\n print(ex)\n return render_template('index.html')\n\n\ndef add():\n city_name = request.form['city_name'].lower()\n\n r = requests.get('https://api.openweathermap.org/data/2.5/weather', params={\n 'q': city_name, 'appid': API_KEY, 'units': 'metric'}).json()\n\n city_name = r['name'].upper()\n temp = r['main']['temp']\n state = r['weather'][0]['main']\n\n city_time = datetime.datetime.utcfromtimestamp(time.time() + r['timezone']).time()\n sunrise = datetime.datetime.utcfromtimestamp(r['sys']['sunrise'] + r['timezone']).time()\n sunset = datetime.datetime.utcfromtimestamp(r['sys']['sunset'] + r['timezone']).time()\n afternoon = datetime.time(hour=12)\n\n if sunrise <= city_time < afternoon:\n day_period = 'day'\n elif afternoon <= city_time < sunset:\n day_period = 'evening-morning'\n else:\n day_period = 'night'\n\n weather_dict = {'city_name': city_name, 'day_night': day_period, 'temp': temp,\n 'state': state}\n return render_template('index.html', weather=weather_dict)\n\n\nif __name__ == '__main__':\n if len(sys.argv) > 1:\n arg_host, arg_port = sys.argv[1].split(':')\n app.run(host=arg_host, port=arg_port)\n else:\n app.run()\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1597, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "call"}, {"api_name": "os.environ.get", "line_number": 12, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 12, "usage_type": "attribute"}, {"api_name": "flask.request.method", "line_number": 17, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 17, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 22, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 28, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 35, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 35, "usage_type": "call"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 36, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 36, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcfromtimestamp", "line_number": 37, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 37, "usage_type": "attribute"}, {"api_name": "datetime.time", "line_number": 38, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 49, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 53, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 54, "usage_type": "attribute"}]} +{"seq_id": "69465207", "text": "#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport argparse\nimport glob\nimport mimetypes\nimport os\nimport subprocess\nimport tempfile\n\nimport botocore.session\nimport yattag\n\n\nclass Bucket:\n def __init__(self, name):\n self.name = name\n self.s3_service = botocore.session.get_session().get_service('s3')\n\n def region(self):\n if not hasattr(self, '_region'):\n default_endpoint = self.s3_service.get_endpoint()\n op = self.s3_service.get_operation('GetBucketLocation')\n http_response, response_data = op.call(default_endpoint,\n bucket=self.name)\n self._region = response_data['LocationConstraint']\n return self._region\n\n def endpoint(self):\n return self.s3_service.get_endpoint(self.region())\n\n def remote_url(self):\n return 's3://{}'.format(self.name)\n\n def resource_url(self, resource):\n return os.path.join(self.endpoint().host, self.name, resource)\n\n def sync(self, local_dir):\n return subprocess.check_call([\n 'aws', 's3', 'sync',\n local_dir, self.remote_url(),\n '--region', self.region()])\n\n def put(self, body, key):\n args = [\n 'aws', 's3api', 'put-object',\n '--bucket', self.name,\n '--region', self.region(),\n '--key', key\n ]\n\n content_type = mimetypes.guess_type(key)[0]\n if content_type:\n args += ['--content-type', content_type]\n\n with tempfile.NamedTemporaryFile() as f:\n f.write(body.encode('utf-8'))\n f.flush()\n args += ['--body', f.name]\n return subprocess.check_call(args)\n\n def wheels(self):\n op = self.s3_service.get_operation('ListObjects')\n http_response, response_data = op.call(self.endpoint(),\n bucket=self.name)\n\n keys = [obj['Key'] for obj in response_data['Contents']]\n keys = [key for key in keys if key.endswith('.whl')]\n\n wheels = []\n for key in keys:\n url = self.resource_url(key)\n wheels.append((key, url))\n\n return wheels\n\n def index(self):\n doc, tag, text = yattag.Doc().tagtext()\n with tag('html'):\n for name, url in self.wheels():\n with tag('a', href=url):\n text(name)\n doc.stag('br')\n\n return doc.getvalue()\n\n\ndef build_wheels(packages, index_url, requirements=None):\n packages = packages or []\n temp_dir = tempfile.mkdtemp(prefix='mkwheelhouse-')\n args = [\n 'pip', 'wheel',\n '--wheel-dir', temp_dir,\n '--find-links', index_url\n ]\n # Add a requirements.txt file if -r/--requirements specified\n if requirements is not None:\n args += ['-r', requirements]\n\n args += packages\n subprocess.check_call(args)\n return temp_dir\n\n\ndef main():\n parser = argparse.ArgumentParser(\n description='Generate and upload wheels to an Amazon S3 wheelhouse')\n parser.add_argument('-r', '--requirements')\n parser.add_argument('-e', '--exclude', action=\"append\")\n parser.add_argument('bucket')\n parser.add_argument('package', nargs='+')\n\n args = parser.parse_args()\n\n bucket = Bucket(args.bucket)\n index_url = bucket.resource_url('index.html')\n\n build_dir = build_wheels(args.package, index_url, args.requirements)\n\n # Remove exclusions from the build_dir -e/--exclude\n for exclusion in args.exclude:\n matches = glob.glob(os.path.join(build_dir, exclusion))\n map(os.remove, matches)\n\n bucket.sync(build_dir)\n bucket.put(bucket.index(), key='index.html')\n\n print('Index written to:', index_url)\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "mkwheelhouse.py", "file_name": "mkwheelhouse.py", "file_ext": "py", "file_size_in_byte": 3818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "botocore.session.session.get_session", "line_number": 19, "usage_type": "call"}, {"api_name": "botocore.session.session", "line_number": 19, "usage_type": "attribute"}, {"api_name": "botocore.session", "line_number": 19, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "subprocess.check_call", "line_number": 40, "usage_type": "call"}, {"api_name": "mimetypes.guess_type", "line_number": 53, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 57, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 61, "usage_type": "call"}, {"api_name": "yattag.Doc", "line_number": 79, "usage_type": "call"}, {"api_name": "tempfile.mkdtemp", "line_number": 91, "usage_type": "call"}, {"api_name": "subprocess.check_call", "line_number": 102, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 107, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 124, "usage_type": "attribute"}]} +{"seq_id": "491609302", "text": "#!/Library/Frameworks/Python.framework/Versions/3.6/bin/python3\n# -*- coding: utf-8 -*-\nfrom PIL import Image\n\nimage_open = Image.open('/Users/vayne/Pictures/Saved Pictures/test.jpg')\n\n# 获得尺寸\nw, h = image_open.size\nprint('Original image size: %sx%s' % (w, h))\n# 缩放50%\nimage_open.thumbnail((w//2, h//2))\nprint('Resize image to: %sx%s' %(w//2, h//2))\n# 缩放后的图片保存\nimage_open.save('test2.jpg','jpeg')", "sub_path": "com/wk/OtherModule/Pillow_test.py", "file_name": "Pillow_test.py", "file_ext": "py", "file_size_in_byte": 423, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "PIL.Image.open", "line_number": 5, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 5, "usage_type": "name"}]} +{"seq_id": "368759581", "text": "import pyglet\nimport ratcave as rc\n\n\ndef get_screen(idx):\n \"\"\"These Three lines send the data to the monitor next to the main monitor ( harder to recognize)\"\"\"\n platform = pyglet.window.get_platform()\n display = platform.get_default_display()\n screens = display.get_screens()\n return screens[idx]\n\n\ndef load_textured_mesh(reader, name, image_filename=None, image_dirname='assets/img/'):\n mesh = reader.get_mesh(name)\n if image_filename:\n mesh.textures.append(rc.Texture.from_image(image_dirname + image_filename))\n return mesh\n\n\ndef remove_image_lines_from_mtl(mtl_filename):\n lines = []\n with open(mtl_filename) as f:\n for line in f:\n if 'map_Kd' in line:\n continue\n if 'map_Bump' in line:\n continue\n lines.append(line)\n\n with open(mtl_filename, 'w') as f:\n f.writelines(lines)", "sub_path": "grass_scene/utils.py", "file_name": "utils.py", "file_ext": "py", "file_size_in_byte": 895, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pyglet.window.get_platform", "line_number": 7, "usage_type": "call"}, {"api_name": "pyglet.window", "line_number": 7, "usage_type": "attribute"}, {"api_name": "ratcave.Texture.from_image", "line_number": 16, "usage_type": "call"}, {"api_name": "ratcave.Texture", "line_number": 16, "usage_type": "attribute"}]} +{"seq_id": "413015412", "text": "#Version 1.0: final funcional, a veces se bloquea la base de datos. Interface cheff y barra sin terminar\n#\nfrom tkinter import ttk\nfrom tkinter import *\n\nfrom datetime import date\n\nfrom matplotlib.backends.backend_tkagg import (FigureCanvasTkAgg, NavigationToolbar2Tk)\nfrom matplotlib.backend_bases import key_press_handler\nfrom matplotlib.figure import Figure\nimport matplotlib.pyplot as plt\n\nimport numpy as np\n\nimport sqlite3\n\n\ndb_name='restaurante.db'\nnombre_Restaurante = 'COMIDA CORRIDA'\nbanderagerente = 0\nbanderamesero=0\n\ndef run_Query(query,parameters=()):\n\t\twith sqlite3.connect(db_name) as conn:\n\t\t\tcursor=conn.cursor()\n\t\t\tresult=cursor.execute(query,parameters)\n\t\t\tconn.commit()\n\t\treturn result\n\ndef destruir(window):\n window.destroy()\n\ndef crear_Inicio():\n window = Tk()\n window.geometry(\"300x200\")\n ventana = VentanaPrincipal(window)\n window.mainloop()\n\ndef crear_Gerente():\n nuevaVentana = Tk()\n nuevaVentana.geometry(\"400x400\")\n objeto = WindGerente(nuevaVentana,'Gerente')\n\nclass PagarCuenta:\n\tdef __init__(self,tree,saldo,propinas,mesero):\n\n\t\tself.mesero= mesero\n\t\tself.saldo = saldo\n\t\tself.propinas = propinas\n\n\t\tself.pagar_wind=Toplevel()\n\t\tself.pagar_wind.title='Pagar la cuenta'\n\t\tself.mesa=self.tree.item(self.tree.selection())['text']\n\n\t\tself.Calculos()\n\n\t\tLabel(self.pagar_wind, text='Pagar cuenta de mesa No. ' + str(self.mesa)).grid(row=0,column=1)\n\t\tLabel(self.pagar_wind, text='Total consumo: ').grid(row=1,column=1)\n\t\tEntry(self.pagar_wind,textvariable=StringVar(self.pagar_wind,value=self.total),state='readonly').grid(row=1,column=2)\n\n\t\tLabel(self.pagar_wind, text='Descuento: ').grid(row=2,column=1)\n\t\tEntry(self.pagar_wind,textvariable=StringVar(self.pagar_wind,value=self.descuento),state='readonly').grid(row=2,column=2)\n\n\t\tLabel(self.pagar_wind, text='Total a pagar: ').grid(row=3,column=1)\n\t\tEntry(self.pagar_wind,textvariable=StringVar(self.pagar_wind,value=self.nuevo_total),state='readonly').grid(row=3,column=2)\n\n\t\tLabel(self.pagar_wind, text='Total pagado: ').grid(row=4,column=1)\n\t\tself.pagado = Entry(self.pagar_wind,textvariable=StringVar(self.pagar_wind,value=self.saldo))\n\t\tself.pagado.grid(row=4,column=2)\n\n\t\tif self.propinas ==0:\n\t\t self.propinas_previas = self.total*0.1\n\t\telse:\n\t\t self.propinas_previas=self.propinas\n\n\t\tLabel(self.pagar_wind, text='Propinas: ').grid(row=5,column=1)\n\t\tself.propinas = Entry(self.pagar_wind,textvariable=StringVar(self.pagar_wind,value=self.propinas_previas))\n\t\tself.propinas.grid(row=5,column=2)\n\n\t\tButton(self.pagar_wind,text='Aceptar',command=lambda: self.matar_Cuenta()).grid(row=6,column=2,sticky=W)\n\t\tButton(self.pagar_wind,text='Cancelar',command=lambda: self.pagar_wind.destroy()).grid(row=6,column=2,sticky=E)\n\n\tdef Calculos(self):\n\t\tself.query='SELECT * FROM mesa where num = ?'\n\n\t\tself.mesa_datos=run_Query(self.query,(self.mesa,))\n\n\t\tfor row in self.mesa_datos:\n\t\t self.numero_mesa=row[0]\n\t\t self.personas=row[2]\n\t\t self.total=row[3]\n\t\t self.descuento=row[5]\n\n\t\tself.nuevo_total = self.total - self.saldo - self.descuento\n\n\tdef matar_Cuenta(self):\n\t\tself.pagado1 = int(self.pagado.get()) #+ self.saldo\n\t\tself.propinas1 = self.propinas.get()\n\t\tself.ventana = self.pagar_wind\n\t\tself.total = self.nuevo_total\n\n\n\t\tself.bandera=0\n\n\t\tif self.total>float(self.pagado1):\n\t\t\tself.ventana.destroy()\n\t\t\tself.pagar_Cuenta(float(self.pagado1),float(self.propinas1))\n\n\t\telif self.total==float(self.pagado1):\n\t\t\tself.ventana.destroy()\n\t\t\tself.cambio=Toplevel()\n\t\t\tself.cambio.title='ok'\n\t\t\tLabel(self.cambio, text='Mesa pagada con éxito ').grid(row=0,column=1)\n\t\t\tLabel(self.cambio, text='Cambio: '+'$0.00').grid(row=1,column=1)\n\t\t\tButton(self.cambio,text='Aceptar',command=lambda: destruir(self.cambio)).grid(row=2,column=1)\n\t\t\tself.bandera=1\n\n\t\telif self.total0:\n self.cantidad-=1\n #query='UPDATE mesa SET {} = ? WHERE num = ?'.format(platillo)\n #parameters=(cantidad,mesa)\n #run_Query(query,parameters)\n\n self.query='UPDATE platillos_mesas SET {} = {} WHERE platillo = ? '.format(self.mesa,self.cantidad,self.platillo)\n self.platillosss=run_Query(self.query,(self.platillo,))\n\n self.query='UPDATE mesa SET total=? WHERE num = ?'.format(self.platillo)\n self.parameters=(self.precio_total,self.mesa)\n run_Query(self.query,self.parameters)\n Entry(self.frame,textvariable=StringVar(self.frame,value=self.precio_total),state='readonly').grid(row=6,column=2,sticky=W)\n self.get_Platillos()\n\n def modificar_Mesa(self):\n self.num = self.mesa\n self.personas = self.Nuevas_personas.get()\n self.personas= int(self.personas)\n self.meseroactual = self.mesero\n self.mesero = self.Nuevo_mesero.get()\n\n if self.mesero!='' and self.personas!=0: #\n self.query= 'UPDATE mesa SET personas= ?,mesero=?,ocupada=1 WHERE num=? '\n self.parameters=(self.personas,self.mesero,self.num)\n run_Query(self.query,self.parameters)\n #print('MESERO = ALGO Y PERSONAS = DIF0 ')\n\n elif self.mesero=='' and self.personas!=0:#\n self.mesero=self.meseroactual\n self.query= 'UPDATE mesa SET personas= ?,mesero=?,ocupada=1 WHERE num=? '\n self.parameters=(self.personas,self.meseroactual,self.num)\n run_Query(self.query,self.parameters)\n #print('MESERO =NADA Y PERSONAS = DIF0 ')\n\n elif self.mesero!='' and self.personas==0:#\n self.personas=1\n self.query= 'UPDATE mesa SET personas= 1,mesero=?,ocupada=1 WHERE num=? '\n self.parameters=(self.mesero,self.num)\n run_Query(self.query,self.parameters)\n #print('MESERO = ALGO Y PERSONAS = 0 ')\n\n #elif self.mesero=='' and self.personas==0:#\n #self.query= 'UPDATE mesa SET personas= 0,mesero=0 ,ocupada=0 WHERE num=? '\n #run_Query(self.query,(self.num,))\n #print('MESERO = NADA Y PERSONAS = 0 ')\n\n self.editar_wind.destroy()\n\n self.get_Mesas()\n\n def get_Mesas(self):\n self.records = self.tree.get_children()\n for element in self.records:\n self.tree.delete(element)\n\n self.query='SELECT * FROM mesa where ocupada = 0 or mesero = ?'\n self.db_rows=run_Query(self.query,(self.meseroahora,))\n for row in self.db_rows:\n self.tree.insert('',END,text=row[0],values = row[3])\n print('Actualizada con extito')\n\nclass WindGerente(ModificarMeseros,ModificarPlatillos,WindDescuentos,WindBloquear,WindVentas,WindVentasMes):\n def __init__(self,window,id):\n self.wind = window\n self.gerente = id\n\n self.wind.title(nombre_Restaurante + ' || '+ self.gerente)\n\n self.frame=LabelFrame(self.wind, text='Acciones: ')\n self.frame.grid(row=2, column=1, columnspan=3, pady=10)\n\n ttk.Button(self.frame,text='Agregar/Quitar mesero',command=lambda: self.editar_Meseros()).grid(row=2,column=1,sticky=W+E,pady=10)\n ttk.Button(self.frame,text='Descontar cuenta',command=lambda: self.descontar_Cuentas()).grid(row=3,column=1,sticky=W+E)\n ttk.Button(self.frame,text='Modificar platillos',command=lambda: self.modificar_Platillos()).grid(row=4,column=1,sticky=W+E)\n ttk.Button(self.frame,text='Bloquear ingredientes',command=lambda: self.bloquear_Platillos()).grid(row=5,column=1,sticky=W+E)\n ttk.Button(self.frame,text='Modificar id',command=lambda: self.modificar_Usuario()).grid(row=6,column=1,sticky=W+E,pady=10)\n ttk.Button(self.frame,text='Gestionar día',command=lambda: self.Ventas()).grid(row=7,column=1,sticky=W+E)\n ttk.Button(self.frame,text='Ver mesas',command=lambda: self.VerMesas()).grid(row=8,column=1,sticky=W+E)\n ttk.Button(self.frame,text='Ver ventas',command=lambda: self.ventas_Totales()).grid(row=9,column=1,sticky=W+E,pady=10)\n ttk.Button(self.frame,text='Salir',command=lambda:[destruir(self.wind),crear_Inicio()]).grid(row=10,column=1,sticky=W+E)\n\n self.frame.pack(side='top',anchor='n')\n\n def editar_Meseros(self):\n self.wind.destroy()\n self.window=Tk()\n #application=Meseros(window)\n ModificarMeseros.__init__(self,self.window)\n\n self.window.mainloop()\n\n def modificar_Platillos(self):\n self.wind.destroy()\n self.window=Tk()\n #application=Platillosn(window)\n ModificarPlatillos.__init__(self,self.window)\n\n self.window.mainloop()\n\n def descontar_Cuentas(self):\n self.wind.destroy()\n self.window2=Tk()\n #application=wind_descuentos(window2)\n WindDescuentos.__init__(self,self.window2)\n\n self.window2.mainloop()\n\n def bloquear_Platillos(self):\n self.wind.destroy()\n self.window2=Tk()\n #ventana=wind_bloquear(self.window2)\n WindBloquear.__init__(self,self.window2)\n\n self.window2.mainloop()\n\n def modificar_Usuario(self):\n self.modificar_wind=Toplevel()\n self.modificar_wind.title='Modificar usuario'\n\n self.message=Label(self.modificar_wind,text='',fg='red')\n self.message.grid(row=0,column=0,columnspan=2,sticky=W+E)\n\n Label(self.modificar_wind,text='Nuevo usuario: ').grid(row=1,column=0)\n self.nuevo_usuario=Entry(self.modificar_wind)\n self.nuevo_usuario.grid(row=1,column=1)\n self.nuevo_usuario.focus()\n\n Label(self.modificar_wind,text='Nueva contraseña: ').grid(row=2,column=0)\n self.nueva_contra=Entry(self.modificar_wind,show='*')\n self.nueva_contra.grid(row=2,column=1)\n\n Label(self.modificar_wind,text='Confirmar contraseña: ').grid(row=3,column=0)\n self.nuevo_contra2=Entry(self.modificar_wind,show='*')\n self.nuevo_contra2.grid(row=3,column=1)\n\n Label(self.modificar_wind,text=' ').grid(row=4,column=0)\n\n Button(self.modificar_wind,text='Guardar',command=lambda: self.editar_Id()).grid(row=5,column=0,sticky=E)\n Button(self.modificar_wind,text='Salir',command=lambda: destruir(self.modificar_wind)).grid(row=5,column=1,sticky=W)\n\n def editar_Id(self):\n self.user=self.nuevo_usuario\n self.passw = self.nueva_contra\n self.passw2 = self.nuevo_contra2\n if self.passw.get() != self.passw2.get():\n self.message['text']='Las contraseñas no coinciden'\n\n else:\n self.query='UPDATE gerentes SET usuario=?,contraseña=? WHERE registro = 2'\n self.parameters=(self.user.get(),self.passw.get())\n run_Query(self.query,self.parameters)\n self.message['text']='Id y contraseña actualizados'\n self.user.delete(0,END)\n self.passw.delete(0,END)\n self.passw2.delete(0,END)\n\n def Ventas(self):\n self.ventas_wind=Toplevel()\n\n WindVentas.__init__(self,self.ventas_wind)\n\n def VerMesas(self):\n\n self.window = Toplevel()\n self.window.title(nombre_Restaurante + ' || Mesas')\n\n self.frame=LabelFrame(self.window, text='Mesas: ')\n self.frame.grid(row=0, column=1, columnspan=3, pady=10)\n\n self.tree2=ttk.Treeview(self.frame,height=10,columns=('Mesero','Total','Descuento'))\n self.tree2.grid(row=0,column=0,pady=10)\n self.tree2.heading('#0',text='Mesa',anchor=CENTER)\n self.tree2.heading('#1',text='Mesero',anchor=CENTER)\n self.tree2.heading('#2',text='Total',anchor=CENTER)\n self.tree2.heading('#3',text='Descuento',anchor=CENTER)\n\n self.query = 'SELECT num,total,mesero,descuento FROM mesa '\n self.mesasn = run_Query(self.query)\n #for row in mesasn:\n # print(row)\n self.records = self.tree2.get_children()\n\n for element in self.records:\n self.tree2.delete(element)\n\n for row in self.mesasn:\n self.tree2.insert('',END,text=row[0],values = (row[2],row[1],row[3]))\n print('Actualizada con extito')\n\n\n\n Button(self.frame,text=' Regresar ',command=lambda: destruir(self.window)).grid(row=1,column=0,sticky=E)\n\n self.window.mainloop()\n\n def ventas_Totales(self):\n self.ventas_wind=Toplevel()\n WindVentasMes.__init__(self,self.ventas_wind)\n\nclass WindCheff:\n\n\tdef __init__(self,window,id):\n\t\tself.wind = window\n\t\tself.wind.title('Lista de platillos || CHEFF')\n\t\tself.ID = id\n\t\tself.wind.geometry(\"1230x350\")\n\n\t\tself.frame=LabelFrame(self.wind, text='Platillos Nuevos: ')\n\t\tself.frame.grid(row=0, column=0, columnspan=3, pady=10)\n\n\t\tself.frame1=LabelFrame(self.wind, text='Platillos en preparación: ')\n\t\tself.frame1.grid(row=0, column=4, columnspan=3, pady=10)\n\n\t\tself.frame2=LabelFrame(self.wind, text='Platillos listos: ')\n\t\tself.frame2.grid(row=0, column=8, columnspan=3, pady=10)\n\n\t\tself.tree0=ttk.Treeview(self.frame,height=10,columns=('Cantidad'))\n\t\tself.tree0.grid(row=0,column=0,pady=10)\n\t\tself.tree0.heading('#0',text='Platillo',anchor=CENTER)\n\t\tself.tree0.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tself.tree3=ttk.Treeview(self.frame1,height=10,columns=('Cantidad'))\n\t\tself.tree3.grid(row=0,column=0,pady=10)\n\t\tself.tree3.heading('#0',text='Platillo',anchor=CENTER)\n\t\tself.tree3.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tself.tree4=ttk.Treeview(self.frame2,height=10,columns=('Cantidad'))\n\t\tself.tree4.grid(row=0,column=0,pady=10)\n\t\tself.tree4.heading('#0',text='Platillo',anchor=CENTER)\n\t\tself.tree4.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tself.cargar_Datos()\n\n\t\tttk.Button(self.wind,text='Salir',command=lambda:[destruir(self.wind),crear_Inicio()]).grid(row=1,column=0,sticky=W+E)#\n\t\tttk.Button(self.wind,text='Actualizar',command=lambda:self.actualizar_Datos()).grid(row=1,column=4,sticky=W+E)\n\t\tttk.Button(self.wind,text='Cargar datos',command=lambda: self.cargar_Datos()).grid(row=1,column=8,sticky=W+E)\n\n\t\tttk.Button(self.frame,text='Pasar a preparación',command=lambda:self.pasar_Preparacionp()).grid(row=2,column=0,sticky=W+E)\n\n\t\tttk.Button(self.frame1,text='Regresar a Fila',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=W)\n\t\tttk.Button(self.frame1,text='Marcar Listo',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=E)\n\n\t\tttk.Button(self.frame2,text='Regresar a preparación',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=W)\n\t\tttk.Button(self.frame2,text='Eliminar',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=E)\n\n\tdef cargar_Datos(self):\n\n\t\t#self.query='DELETE from controlcheffbarra WHERE bebida = 0'\n\t\t#run_Query(self.query)\n\n\t\tself.query = 'SELECT * FROM platillos_mesas WHERE mesa1!=0 OR mesa2!=0 OR mesa3!=0 OR mesa4!=0 OR mesa5!=0 OR mesa6!=0 OR mesa7!=0 '\n\t\tself.mesasn = run_Query(self.query)\n\t\tself.records = self.tree0.get_children()\n\n\t\tfor element in self.records:\n\t\t self.tree0.delete(element)\n\n\t\tfor row in self.mesasn:\n\t\t\tself.totalplatillo=row[1]+row[2]+row[3]+row[4]+row[5]+row[6]+row[7]\n\t\t\t#print('agregando ',self.totalplatillo,' ',row[0])\n\t\t\tself.query='SELECT bebida FROM platillos where nombre = ?'\n\t\t\tself.parameters = (row[0],)\n\t\t\tself.bebidasis=run_Query(self.query,self.parameters)\n\t\t\tprint('seleccionado los datos de bebidas')\n\n\t\t\tfor x in self.bebidasis:\n\t\t\t\tself.banderab = x[0]\n\n\t\t\tif self.banderab == 0:\n\t\t\t\tself.query='INSERT INTO controlcheffbarra (nombre,cantidad,estado,bebida) VALUES (?,?,?,?)'\n\t\t\t\tself.parameters =(row[0],self.totalplatillo,0,0)\n\t\t\t\t#run_Query(self.query,self.parameters)\n\t\t\t\t#print('Insertado datos en controlcheffbarra')\n\n\t\t\t\tself.tree0.insert('',END,text=row[0],values = self.totalplatillo)\n\n\t\tprint('Actualizada con extito')\n\n\tdef pasar_Preparacionp(self):\n\n\t\ttry:\n\t\t self.tree0.item(self.tree0.selection())['text'][0]\n\n\t\texcept IndexError as e:\n\t\t\tprint('No se selecciono nada')\n\t\t\treturn\n\t\telse:\n\t\t\tself.cantidad = self.tree0.item(self.tree0.selection())['values'][0]\n\t\t\tif self.cantidad > 1:\n\t\t\t\tself.platillos = self.tree0.item(self.tree0.selection())['text']\n\t\t\t\tself.tree0.insert('',END,text=self.platillos,values = self.cantidad-1)\n\t\t\t\tself.tree0.delete(self.tree0.selection())\n\t\t\telse:\n\t\t\t\tself.tree0.delete(self.tree0.selection())\n\n\n\tdef actualizar_Datos(self):\n\t\tpass\n\nclass WindBar:\n\tdef __init__(self,window,id):\n\t\tself.wind = window\n\t\tself.wind.title('Lista de bebidas || CHEFF')\n\t\tself.ID = id\n\t\tself.wind.geometry(\"1230x350\")\n\n\t\tself.frame=LabelFrame(self.wind, text='Bebidas Nuevas: ')\n\t\tself.frame.grid(row=0, column=0, columnspan=3, pady=10)\n\n\t\tself.frame1=LabelFrame(self.wind, text='Bebidas en preparación: ')\n\t\tself.frame1.grid(row=0, column=4, columnspan=3, pady=10)\n\n\t\tself.frame2=LabelFrame(self.wind, text='Bebidas listos: ')\n\t\tself.frame2.grid(row=0, column=8, columnspan=3, pady=10)\n\n\t\tself.tree0=ttk.Treeview(self.frame,height=10,columns=('Cantidad'))\n\t\tself.tree0.grid(row=0,column=0,pady=10)\n\t\tself.tree0.heading('#0',text='Bebida',anchor=CENTER)\n\t\tself.tree0.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tself.tree3=ttk.Treeview(self.frame1,height=10,columns=('Cantidad'))\n\t\tself.tree3.grid(row=0,column=0,pady=10)\n\t\tself.tree3.heading('#0',text='Bebida',anchor=CENTER)\n\t\tself.tree3.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tself.tree4=ttk.Treeview(self.frame2,height=10,columns=('Cantidad'))\n\t\tself.tree4.grid(row=0,column=0,pady=10)\n\t\tself.tree4.heading('#0',text='Bebida',anchor=CENTER)\n\t\tself.tree4.heading('#1',text='Cantidad',anchor=CENTER)\n\n\t\tttk.Button(self.wind,text='Salir',command=lambda:[destruir(self.wind),crear_Inicio()]).grid(row=1,column=0,sticky=W+E)#\n\t\tttk.Button(self.wind,text='Actualizar',command=lambda:self.actualizar_Datos()).grid(row=1,column=4,sticky=W+E)\n\t\tttk.Button(self.wind,text='Cargar datos',command=lambda: self.cargar_Datosb()).grid(row=1,column=8,sticky=W+E)\n\n\t\tttk.Button(self.frame,text='Pasar a preparación',command=lambda: self.pasar_Preparacion()).grid(row=2,column=0,sticky=W+E)\n\n\t\tttk.Button(self.frame1,text='Regresar a Fila',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=W)\n\t\tttk.Button(self.frame1,text='Marcar Listo',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=E)\n\n\t\tttk.Button(self.frame2,text='Regresar a preparación',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=W)\n\t\tttk.Button(self.frame2,text='Eliminar',command=lambda:self.actualizar_Datos()).grid(row=2,column=0,sticky=E)\n\n\tdef cargar_Datosb(self):\n\n\t\tself.query='DELETE from controlcheffbarra WHERE bebida = 1'\n\t\t#run_Query(self.query)\n\n\t\tself.query = 'SELECT * FROM platillos_mesas WHERE mesa1!=0 OR mesa2!=0 OR mesa3!=0 OR mesa4!=0 OR mesa5!=0 OR mesa6!=0 OR mesa7!=0 '\n\t\tself.mesasn = run_Query(self.query)\n\t\tself.records = self.tree0.get_children()\n\n\t\tfor element in self.records:\n\t\t self.tree0.delete(element)\n\n\t\tfor row in self.mesasn:\n\t\t\tself.totalplatillo=row[1]+row[2]+row[3]+row[4]+row[5]+row[6]+row[7]\n\t\t\t#print('agregando ',self.totalplatillo,' ',row[0])\n\t\t\tself.query='SELECT bebida FROM platillos where nombre = ?'\n\t\t\tself.parameters = (row[0],)\n\t\t\tself.bebidasis=run_Query(self.query,self.parameters)\n\t\t\t#print('seleccionado los datos de bebidas')\n\n\t\t\tfor x in self.bebidasis:\n\t\t\t\tself.banderab = x[0]\n\n\t\t\tif self.banderab == 0:\n\t\t\t\tself.query='INSERT INTO controlcheffbarra (nombre,cantidad,estado,bebida) VALUES (?,?,?,?)'\n\t\t\t\tself.parameters =(row[0],self.totalplatillo,0,0)\n\t\t\t\t#run_Query(self.query,self.parameters)\n\t\t\t\t#print('Insertado datos en controlcheffbarra')\n\n\t\t\t\tself.tree0.insert('',END,text=row[0],values = self.totalplatillo)\n\n\t\tprint('Actualizada con extito')\n\n\tdef pasar_Preparacion(self):\n\t\t#print('dentro de la funcion')\n\t\ttry:\n\t\t self.tree0.item(self.tree0.selection())['text'][0]\n\n\t\texcept IndexError as e:\n\t\t\tprint('No se selecciono nada')\n\t\t\treturn\n\n\t\tself.tree0.delete(self.tree0.selection())\n\n\tdef actualizar_Datos(self):\n\t\tpass\n\nclass VentanaPrincipal(WindWaiter,WindGerente,WindBar,WindCheff):\n def __init__(self,window):\n self.wind = window\n self.wind.title(nombre_Restaurante)\n\n self.frame = LabelFrame(self.wind, text = 'Inicio')\n self.frame.grid(row = 0,column = 0,columnspan = 3,pady = 20,sticky = W+E)\n self.frame.config(width = '200',height = '200')\n self.frame.pack(side = 'top',anchor = 'n')\n\n ttk.Button(self.frame,text = 'Iniciar sesion',command = lambda: self.meter_Id()).grid(row = 3,columnspan = 2,sticky = W+E)\n ttk.Button(self.frame,text = 'salir',command = lambda: destruir(window)).grid(row = 4,columnspan = 2,sticky = W+E)\n\n def meter_Id(self):\n\n self.introducirWind = Toplevel()\n self.introducirWind.title = 'Iniciar sesión'\n\n Label(self.introducirWind,text= 'ID').grid(row=0,column=1)\n self.ID = Entry(self.introducirWind)\n self.ID.focus()\n self.ID.grid(row=0,column=2)\n\n Button (self.introducirWind, text='Iniciar',command= lambda: self.comprobar_Id()).grid(row=1,column=2,sticky=W)\n\n def comprobar_Id(self):\n self.bandera = 0\n self.IDD = self.ID.get()\n\n self.query1 = 'SELECT usuario FROM gerentes WHERE usuario = ?'\n self.query2 = 'SELECT Usuario FROM meseros WHERE Usuario = ?'\n self.consulta = run_Query(self.query1,(self.IDD,))\n\n if self.IDD!='cheff' and self.IDD!='barra':\n\n for row in self.consulta:\n if row[0] == self.IDD:\n self.bandera = 1\n self.meter_Password()\n\n if self.bandera == 0:\n self.consulta = run_Query(self.query2,(self.IDD,))\n for row in self.consulta:\n if row[0]==self.IDD:\n self.bandera=1\n self.wind.destroy()\n self.nuevaVentana =Tk()\n self.nuevaVentana.geometry(\"400x400\")\n\n WindWaiter.__init__(self,self.nuevaVentana,self.IDD)\n \t\t\t\t\t#self.ventana = WindWaiter(self.nuevaVentana,self.ID)\n\n self.nuevaVentana.mainloop()\n\n elif self.IDD == 'cheff':\n self.bandera=1\n self.wind.destroy()\n self.nuevaVentana =Tk()\n self.nuevaVentana.geometry(\"400x400\")\n WindCheff.__init__(self,self.nuevaVentana,self.IDD)\n #self.ventana = Wind_cheff(self.nuevaVentana)\n self.nuevaVentana.mainloop()\n\n elif self.IDD == 'barra':\n self.bandera=1\n self.wind.destroy()\n self.nuevaVentana =Tk()\n self.nuevaVentana.geometry(\"400x400\")\n WindBar.__init__(self,self.nuevaVentana,self.IDD)\n #self.ventana = Wind_bar(self.nuevaVentana)\n self.nuevaVentana.mainloop()\n\n if self.bandera == 0:\n self.error=Toplevel()\n self.error.title='usuario no válido'\n Label(self.error, text='Usuario no válido ').grid(row=0,column=1)\n Button(self.error,text='regresar',command=lambda: destruir(self.error)).grid(row=1,column=1)\n print('Usuario no válido')\n\n def meter_Password(self):\n self.usuario = self.IDD\n self.contraWind = Toplevel()\n self.contraWind.title = 'Iniciar sesión de ' + self.usuario\n\n Label(self.contraWind,text= 'Contraseña:').grid(row=0,column=1)\n self.casillaPsw = Entry(self.contraWind,show=\"*\")\n self.casillaPsw.focus()\n self.casillaPsw.grid(row=0,column=2)\n\n Button (self.contraWind, text='Iniciar',command= lambda: self.verificar_Password()).grid(row=1,column=2,sticky=W)\n Button (self.contraWind, text='regresar',command= lambda: self.contraWind.destroy()).grid(row=1,column=3)\n\n def verificar_Password(self):\n \tself.bandera=0\n \tself.query='SELECT contraseña FROM gerentes WHERE usuario = ?'\n \tself.consulta=run_Query(self.query,(self.IDD,))\n\n \tfor row in self.consulta:\n \t\tif row[0] == self.casillaPsw.get():\n \t\t\tprint('AHORA ERES GERENTE')\n \t\t\tself.bandera=1\n\n \tif self.bandera == 0:\n \t\tprint('Contraseña no válida')\n \t\tself.error=Toplevel()\n \t\tself.error.title='Contraseña no válida'\n \t\tLabel(self.error, text='Contraseña no válida ').grid(row=0,column=1)\n \t\tButton(self.error,text='Regresar',command=lambda: destruir(self.error)).grid(row=1,column=1)\n\n \tif self.bandera ==1:\n self.wind.destroy()\n self.nuevaVentana = Tk()\n\n self.nuevaVentana.geometry(\"400x400\")\n WindGerente.__init__(self,self.nuevaVentana,'Gerente')\n #ventana = Wind_gerente(nuevaVentana,id)\n\n self.nuevaVentana.mainloop()\n\n#crear_Inicio()\n", "sub_path": "index.py", "file_name": "index.py", "file_ext": "py", "file_size_in_byte": 60620, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sqlite3.connect", "line_number": 24, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 188, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 188, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 193, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 193, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 198, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 198, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 199, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 199, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 200, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 200, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 205, "usage_type": "call"}, {"api_name": "tkinter.ttk.Button", "line_number": 320, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 320, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 325, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 325, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 330, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 330, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 331, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 331, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 332, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 332, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 333, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 333, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 338, "usage_type": "call"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 520, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 520, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 580, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 580, "usage_type": "name"}, {"api_name": "tkinter.ttk.Combobox", "line_number": 642, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 642, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 697, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 697, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 705, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 705, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 750, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 750, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 797, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 797, "usage_type": "name"}, {"api_name": "matplotlib.figure.Figure", "line_number": 826, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 833, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 834, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 842, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.FigureCanvasTkAgg", "line_number": 848, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_tkagg.NavigationToolbar2Tk", "line_number": 853, "usage_type": "call"}, {"api_name": "matplotlib.backend_bases.key_press_handler", "line_number": 859, "usage_type": "call"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 893, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 893, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 901, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 901, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 902, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 902, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 903, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 903, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 904, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 904, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 956, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 956, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1140, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1140, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1141, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1141, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1142, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1142, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1143, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1143, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1144, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1144, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1145, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1145, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1146, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1146, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1147, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1147, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1148, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1148, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1238, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1238, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1285, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1285, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1290, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1290, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1295, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1295, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1302, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1302, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1303, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1303, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1304, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1304, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1306, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1306, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1308, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1308, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1309, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1309, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1311, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1311, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1312, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1312, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1384, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1384, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1389, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1389, "usage_type": "name"}, {"api_name": "tkinter.ttk.Treeview", "line_number": 1394, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1394, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1399, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1399, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1400, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1400, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1401, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1401, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1403, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1403, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1405, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1405, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1406, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1406, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1408, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1408, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1409, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1409, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1468, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1468, "usage_type": "name"}, {"api_name": "tkinter.ttk.Button", "line_number": 1469, "usage_type": "call"}, {"api_name": "tkinter.ttk", "line_number": 1469, "usage_type": "name"}]} +{"seq_id": "476333385", "text": "import pytest\nfrom bs4 import BeautifulSoup\nfrom flask import url_for\n\n\ndef test_should_render_email_verification_resend_show_email_address_and_resend_verify_email(\n client,\n mocker,\n api_user_active,\n mock_get_user_by_email,\n mock_send_verify_email,\n):\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_active[\"id\"],\n \"email\": api_user_active[\"email_address\"],\n }\n response = client.get(url_for(\"main.resend_email_verification\"))\n assert response.status_code == 200\n\n page = BeautifulSoup(response.data.decode(\"utf-8\"), \"html.parser\")\n\n assert page.h1.string == \"Check your email\"\n expected = \"A new confirmation email has been sent to {}\".format(api_user_active[\"email_address\"])\n message = page.find_all(\"p\", {\"class\": \"email-confirm\"})[0].text\n assert message == expected\n mock_send_verify_email.assert_called_with(api_user_active[\"id\"], api_user_active[\"email_address\"])\n\n\ndef test_should_render_correct_resend_template_for_active_user(\n client,\n api_user_active,\n mock_get_user_by_email,\n mock_send_verify_code,\n):\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_active[\"id\"],\n \"email\": api_user_active[\"email_address\"],\n }\n response = client.get(url_for(\"main.check_and_resend_text_code\"))\n assert response.status_code == 200\n\n page = BeautifulSoup(response.data.decode(\"utf-8\"), \"html.parser\")\n assert page.h1.string == \"Re-send security code\"\n # there shouldn't be a form for updating mobile number\n assert page.find(\"form\") is None\n\n\ndef test_should_render_correct_resend_template_for_pending_user(\n client,\n mocker,\n api_user_pending,\n mock_send_verify_code,\n):\n mocker.patch(\"app.user_api_client.get_user_by_email\", return_value=api_user_pending)\n\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_pending[\"id\"],\n \"email\": api_user_pending[\"email_address\"],\n }\n response = client.get(url_for(\"main.check_and_resend_text_code\"))\n assert response.status_code == 200\n\n page = BeautifulSoup(response.data.decode(\"utf-8\"), \"html.parser\")\n assert page.h1.string == \"Check your mobile number\"\n\n expected = \"Check that your mobile phone number is correct and then re-send the security code.\"\n message = page.find_all(\"p\", {\"class\": \"phone-confirm\"})[0].text\n assert message == expected\n assert page.find(\"form\").input[\"value\"] == api_user_pending[\"mobile_number\"]\n\n\n@pytest.mark.parametrize(\n \"phone_number_to_register_with\",\n [\n \"+16502532222\",\n \"+4966921809\",\n ],\n)\ndef test_should_resend_verify_code_and_update_mobile_for_pending_user(\n client,\n mocker,\n api_user_pending,\n mock_update_user_attribute,\n mock_send_verify_code,\n phone_number_to_register_with,\n):\n mocker.patch(\"app.user_api_client.get_user_by_email\", return_value=api_user_pending)\n\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_pending[\"id\"],\n \"email\": api_user_pending[\"email_address\"],\n }\n response = client.post(\n url_for(\"main.check_and_resend_text_code\"),\n data={\"mobile_number\": phone_number_to_register_with},\n )\n assert response.status_code == 302\n assert response.location == url_for(\"main.verify\")\n\n mock_update_user_attribute.assert_called_once_with(\n api_user_pending[\"id\"],\n mobile_number=phone_number_to_register_with,\n )\n mock_send_verify_code.assert_called_once_with(\n api_user_pending[\"id\"],\n \"sms\",\n phone_number_to_register_with,\n )\n\n\ndef test_check_and_redirect_to_two_factor_if_user_active(\n client,\n api_user_active,\n mock_get_user_by_email,\n mock_send_verify_code,\n):\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_active[\"id\"],\n \"email\": api_user_active[\"email_address\"],\n }\n response = client.get(url_for(\"main.check_and_resend_verification_code\"))\n assert response.status_code == 302\n assert response.location == url_for(\"main.two_factor_sms_sent\")\n\n\ndef test_check_and_redirect_to_verify_if_user_pending(\n client,\n mocker,\n api_user_pending,\n mock_get_user_pending,\n mock_send_verify_code,\n):\n mocker.patch(\"app.user_api_client.get_user_by_email\", return_value=api_user_pending)\n\n with client.session_transaction() as session:\n session[\"user_details\"] = {\n \"id\": api_user_pending[\"id\"],\n \"email\": api_user_pending[\"email_address\"],\n }\n response = client.get(url_for(\"main.check_and_resend_verification_code\"))\n assert response.status_code == 302\n assert response.location == url_for(\"main.verify\")\n\n\n@pytest.mark.parametrize(\n \"endpoint\",\n [\n \"main.resend_email_verification\",\n \"main.check_and_resend_text_code\",\n \"main.check_and_resend_verification_code\",\n ],\n)\ndef test_redirect_to_sign_in_if_not_logged_in(\n client,\n endpoint,\n):\n response = client.get(url_for(endpoint))\n\n assert response.location == url_for(\"main.sign_in\")\n assert response.status_code == 302\n", "sub_path": "tests/app/main/views/test_code_not_received.py", "file_name": "test_code_not_received.py", "file_ext": "py", "file_size_in_byte": 5336, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.url_for", "line_number": 18, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 21, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 41, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 63, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 66, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 98, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 102, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 75, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.url_for", "line_number": 126, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 128, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 145, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 147, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 164, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 150, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 150, "usage_type": "attribute"}]} +{"seq_id": "344881434", "text": "from django.shortcuts import render\nimport json\nfrom django.http import HttpResponse\n\n\ndef index(request):\n \"\"\"\n index\n \"\"\"\n\n context = {'title': 'タイトルです',\n }\n\n return render(request, 'app/index.html', context)\n\n\ndef index2(request):\n \"\"\"\n index2\n \"\"\"\n\n context = {'title': 'タイトルです',\n }\n\n return render(request, 'app/index2.html', context)\n\n\ndef index3(request):\n \"\"\"\n index3\n \"\"\"\n\n context = {'title': 'タイトルです',\n }\n\n return render(request, 'app/index3.html', context)\n\n\ndef api_01(request):\n \"\"\"\n getでkeywordを受け取り、jsonを返すAPI\n \"\"\"\n\n dctData = {}\n\n # keywordがgetで与えられたとき(辞書のkeyが存在したら)\n if 'keyword' in request.GET:\n keyword = request.GET['keyword']\n dctData['原型'] = keyword\n dctData['丁寧語'] = keyword+'でございます'\n dctData['疑問'] = keyword+'?'\n else:\n dctData['データ'] = 'なし'\n\n return HttpResponse(\n json.dumps(dctData, ensure_ascii=False),\n content_type='application/json; charset=utf-8',\n )\n", "sub_path": "app/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1175, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.shortcuts.render", "line_number": 14, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 25, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 36, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 55, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 56, "usage_type": "call"}]} +{"seq_id": "178251788", "text": "import math\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\nx = np.linspace(1,4.99,1000)\ny = 1 / np.sqrt(5-x)\nplt.plot(x,y)\nplt.show()\n\n\n#以1/x解决积分过程中的奇异性的问题\ns = 0\nx = np.ones(500)\nfor i in range(1,501):\n x[i] = 0.01 * i\n\nfor i in range(1,500):\n si = (1/x[i+1] + 1/x[i]) * ((x[i+1]-x[i])/2)\n s = si +s\n\nprint(s)", "sub_path": ".history/奇异性_20201204110117.py", "file_name": "奇异性_20201204110117.py", "file_ext": "py", "file_size_in_byte": 354, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.linspace", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 8, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 8, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "111229502", "text": "# coding: utf-8\n\nimport six\n\nfrom huaweicloudsdkcore.utils.http_utils import sanitize_for_serialization\n\n\nclass ImportDatabaseDataReq:\n\n \"\"\"\n Attributes:\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n attribute_map (dict): The key is attribute name\n and the value is json key in definition.\n \"\"\"\n sensitive_list = []\n\n openapi_types = {\n 'files': 'list[str]',\n 'delimiter': 'str',\n 'skip_lines': 'int'\n }\n\n attribute_map = {\n 'files': 'files',\n 'delimiter': 'delimiter',\n 'skip_lines': 'skip_lines'\n }\n\n def __init__(self, files=None, delimiter=None, skip_lines=None):\n \"\"\"ImportDatabaseDataReq\n\n The model defined in huaweicloud sdk\n\n :param files: 导入文件l路径列表\n :type files: list[str]\n :param delimiter: 分隔符,常见分隔符为, ;\n :type delimiter: str\n :param skip_lines: 跳过的header行数\n :type skip_lines: int\n \"\"\"\n \n \n\n self._files = None\n self._delimiter = None\n self._skip_lines = None\n self.discriminator = None\n\n self.files = files\n self.delimiter = delimiter\n self.skip_lines = skip_lines\n\n @property\n def files(self):\n \"\"\"Gets the files of this ImportDatabaseDataReq.\n\n 导入文件l路径列表\n\n :return: The files of this ImportDatabaseDataReq.\n :rtype: list[str]\n \"\"\"\n return self._files\n\n @files.setter\n def files(self, files):\n \"\"\"Sets the files of this ImportDatabaseDataReq.\n\n 导入文件l路径列表\n\n :param files: The files of this ImportDatabaseDataReq.\n :type files: list[str]\n \"\"\"\n self._files = files\n\n @property\n def delimiter(self):\n \"\"\"Gets the delimiter of this ImportDatabaseDataReq.\n\n 分隔符,常见分隔符为, ;\n\n :return: The delimiter of this ImportDatabaseDataReq.\n :rtype: str\n \"\"\"\n return self._delimiter\n\n @delimiter.setter\n def delimiter(self, delimiter):\n \"\"\"Sets the delimiter of this ImportDatabaseDataReq.\n\n 分隔符,常见分隔符为, ;\n\n :param delimiter: The delimiter of this ImportDatabaseDataReq.\n :type delimiter: str\n \"\"\"\n self._delimiter = delimiter\n\n @property\n def skip_lines(self):\n \"\"\"Gets the skip_lines of this ImportDatabaseDataReq.\n\n 跳过的header行数\n\n :return: The skip_lines of this ImportDatabaseDataReq.\n :rtype: int\n \"\"\"\n return self._skip_lines\n\n @skip_lines.setter\n def skip_lines(self, skip_lines):\n \"\"\"Sets the skip_lines of this ImportDatabaseDataReq.\n\n 跳过的header行数\n\n :param skip_lines: The skip_lines of this ImportDatabaseDataReq.\n :type skip_lines: int\n \"\"\"\n self._skip_lines = skip_lines\n\n def to_dict(self):\n \"\"\"Returns the model properties as a dict\"\"\"\n result = {}\n\n for attr, _ in six.iteritems(self.openapi_types):\n value = getattr(self, attr)\n if isinstance(value, list):\n result[attr] = list(map(\n lambda x: x.to_dict() if hasattr(x, \"to_dict\") else x,\n value\n ))\n elif hasattr(value, \"to_dict\"):\n result[attr] = value.to_dict()\n elif isinstance(value, dict):\n result[attr] = dict(map(\n lambda item: (item[0], item[1].to_dict())\n if hasattr(item[1], \"to_dict\") else item,\n value.items()\n ))\n else:\n if attr in self.sensitive_list:\n result[attr] = \"****\"\n else:\n result[attr] = value\n\n return result\n\n def to_str(self):\n \"\"\"Returns the string representation of the model\"\"\"\n import simplejson as json\n if six.PY2:\n import sys\n reload(sys)\n sys.setdefaultencoding(\"utf-8\")\n return json.dumps(sanitize_for_serialization(self), ensure_ascii=False)\n\n def __repr__(self):\n \"\"\"For `print`\"\"\"\n return self.to_str()\n\n def __eq__(self, other):\n \"\"\"Returns true if both objects are equal\"\"\"\n if not isinstance(other, ImportDatabaseDataReq):\n return False\n\n return self.__dict__ == other.__dict__\n\n def __ne__(self, other):\n \"\"\"Returns true if both objects are not equal\"\"\"\n return not self == other\n", "sub_path": "huaweicloud-sdk-eihealth/huaweicloudsdkeihealth/v1/model/import_database_data_req.py", "file_name": "import_database_data_req.py", "file_ext": "py", "file_size_in_byte": 4664, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "six.iteritems", "line_number": 125, "usage_type": "call"}, {"api_name": "six.PY2", "line_number": 151, "usage_type": "attribute"}, {"api_name": "sys.setdefaultencoding", "line_number": 154, "usage_type": "call"}, {"api_name": "simplejson.dumps", "line_number": 155, "usage_type": "call"}, {"api_name": "huaweicloudsdkcore.utils.http_utils.sanitize_for_serialization", "line_number": 155, "usage_type": "call"}]} +{"seq_id": "180007946", "text": "import pathlib\nimport dropbox\nimport re\nimport configparser\n\n\ndef upload_to_dropbox(file, file_name):\n # target location in Dropbox\n target = \"/TTS/Magic/\" # the target folder\n targetfile = target + file_name # the target path and file name\n\n # Get the API key from the config\n config = configparser.ConfigParser()\n config.read(\"config.ini\")\n api_key = config['Dropbox']['API_V2']\n if not api_key:\n print(\"Please specify your Dropbox API key in config.ini\")\n print(\"See https://blogs.dropbox.com/developers/2014/05/generate-an-access-token-for-your-own-account/\")\n return\n\n # Create a dropbox object using an API v2 key\n d = dropbox.Dropbox(api_key)\n\n d.files_upload(file, targetfile,\n mode=dropbox.files.WriteMode(\"overwrite\"))\n\n # create a shared link\n link = d.sharing_create_shared_link(targetfile)\n\n # url which can be shared\n url = link.url\n\n # link which directly downloads by replacing ?dl=0 with ?dl=1\n dl_url = re.sub(r\"\\?dl\\=0\", \"?dl=1\", url)\n return dl_url\n", "sub_path": "Scryfall Tools/dropbox_uploader.py", "file_name": "dropbox_uploader.py", "file_ext": "py", "file_size_in_byte": 1074, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "configparser.ConfigParser", "line_number": 13, "usage_type": "call"}, {"api_name": "dropbox.Dropbox", "line_number": 22, "usage_type": "call"}, {"api_name": "dropbox.files.WriteMode", "line_number": 25, "usage_type": "call"}, {"api_name": "dropbox.files", "line_number": 25, "usage_type": "attribute"}, {"api_name": "re.sub", "line_number": 34, "usage_type": "call"}]} +{"seq_id": "81429532", "text": "#!/usr/bin/env python3\n# @Time : 2020-02-14\n# @Author : caicai\n# @File : common.py\nimport os\nimport redis\nfrom random import sample\nfrom urllib import parse\nimport sys\nfrom myscan.lib.core.data import paths, conn, cmd_line_options,logger\nimport difflib\nimport hashlib\nimport base64\n\n\ndef redis_conn():\n arg_redis = cmd_line_options.redis\n if arg_redis:\n if \"@\" in arg_redis:\n pwd, ipport = arg_redis.split(\"@\", 1)\n if not pwd:\n pwd = None\n if \":\" in ipport and ipport.count(\".\") >= 2:\n ip, port, db = ipport.split(\":\", 2)\n else:\n ip = ipport\n port = 6379\n db = 0\n logger.info(\"Redis connection args: pwd:{},ip:{},port:{},db:{}\".format(pwd, ip, port, db))\n conn.redis = redis.ConnectionPool(host=ip, password=pwd, port=int(port), db=int(db))\n red=getredis()\n\n\n else:\n # error_msg = \"Set reids connection error,please check redis-server\"\n error_msg = \"Please use --redis pass@host:port:db ,if pass is none ,like --redis @host:port:db\"\n logger.warning(error_msg)\n sys.exit()\ndef set_paths(root_path):\n \"\"\"\n Sets absolute paths for project directories and files\n \"\"\"\n paths.MYSCAN_ROOT_PATH = root_path\n paths.MYSCAN_DATA_PATH = os.path.join(paths.MYSCAN_ROOT_PATH, \"data\")\n paths.MYSCAN_PLUGINS_PATH = os.path.join(paths.MYSCAN_ROOT_PATH, \"plugins\")\n # paths.MYSCAN_MOUDLE_PATH = os.path.join(paths.MYSCAN_ROOT_PATH, \"moudle\")\n paths.MYSCAN_POCS_PATH = os.path.join(paths.MYSCAN_ROOT_PATH, \"pocs\")\n paths.MYSCAN_REPORT_PATH = os.path.join(paths.MYSCAN_ROOT_PATH, \"report\")\n paths.USER_POCS_PATH = None\n\n paths.SENSETIVE_DIR = os.path.join(paths.MYSCAN_DATA_PATH, \"sensetive-dir.txt\")\n paths.WEAK_PASS = os.path.join(paths.MYSCAN_DATA_PATH, \"password-top100.txt\")\n paths.LARGE_WEAK_PASS = os.path.join(paths.MYSCAN_DATA_PATH, \"password-top1000.txt\")\n\n # paths.MYSCAN_HOME_PATH = os.path.expanduser(\"~\")\n # _ = os.path.join(paths.MYSCAN_HOME_PATH, \".pocsuite\")\n #\n # paths.API_SHELL_HISTORY = os.path.join(_, \"api.hst\")\n # paths.OS_SHELL_HISTORY = os.path.join(_, \"os.hst\")\n # paths.SQL_SHELL_HISTORY = os.path.join(_, \"sql.hst\")\n # paths.MYSCAN_SHELL_HISTORY = os.path.join(_, \"pocsuite.hst\")\n # paths.MYSCAN_CONSOLE_HISTORY = os.path.join(_, \"console.hst\")\n #\n # paths.MYSCAN_TMP_PATH = os.path.join(_, \"tmp\")\n # paths.MYSCAN_RC_PATH = os.path.join(paths.MYSCAN_HOME_PATH, \".pocsuiterc\")\n # paths.MYSCAN_OUTPUT_PATH = paths.get(\"MYSCAN_OUTPUT_PATH\", os.path.join(_, \"output\"))\n\n\ndef get_random_str(nums):\n return ''.join(sample(\"abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ\", int(nums)))\n\n\ndef get_random_num(nums):\n return int(''.join(sample(\"123456789\", int(nums))))\n\n\ndef banner():\n return \"\\033[1;33;40m\" + '''\n# .___ ___. ____ ____ _______. ______ ___ .__ __. \n# | \\/ | \\ \\ / / / | / | / \\ | \\ | | \n# | \\ / | \\ \\/ / | (----`| ,----' / ^ \\ | \\| | \n# | |\\/| | \\_ _/ \\ \\ | | / /_\\ \\ | . ` | \n# | | | | | | .----) | | `----. / _____ \\ | |\\ | \n# |__| |__| |__| |_______/ \\______|/__/ \\__\\ |__| \\__| \n# v2.0 \n ''' + \"\\033[0m\"\n\n\ndef similar(text1, text2):\n return difflib.SequenceMatcher(None, text1, text2).quick_ratio()\n\n\ndef getredis():\n #此处windows linux的一个坑,windows不能多进程 共享socket\n\n return redis.StrictRedis(connection_pool=conn.redis)\n\n\ndef gethostportfromurl(url):\n '''\n return list [host,port]\n '''\n port = 80\n r = parse.urlparse(url)\n if \":\" not in r.netloc:\n if r.scheme == \"https\":\n port = 443\n else:\n h, p = r.netloc.split(\":\")\n return h, int(p)\n return r.netloc, port\ndef getmd5(s):\n m = hashlib.md5()\n if not isinstance(s,str):\n s=str(s)\n b = s.encode(encoding='utf-8')\n m.update(b)\n return m.hexdigest()\ndef is_base64(value: str):\n if isinstance(value,str):\n value=value.encode()\n try:\n res=base64.b64decode(value)\n return res.decode().isprintable()\n except Exception as ex:\n print(ex)\n return False\ndef verify_param(param,new,method=\"a\"):\n '''\n 处理新添加的值\n burp大哥这么说的:\n /**\n * Used to indicate a parameter within the URL query string.\n */\n static final byte PARAM_URL = 0;\n /**\n * Used to indicate a parameter within the message body.\n */\n static final byte PARAM_BODY = 1;\n /**\n * Used to indicate an HTTP cookie.\n */\n static final byte PARAM_COOKIE = 2;\n /**\n * Used to indicate an item of data within an XML structure.\n */\n static final byte PARAM_XML = 3;\n /**\n * Used to indicate the value of a tag attribute within an XML structure.\n */\n static final byte PARAM_XML_ATTR = 4;\n /**\n * Used to indicate the value of a parameter attribute within a multi-part\n * message body (such as the name of an uploaded file).\n */\n static final byte PARAM_MULTIPART_ATTR = 5;\n /**\n * Used to indicate an item of data within a JSON structure.\n */\n static final byte PARAM_JSON = 6;\n '''\n if param.get(\"type\")==1: #body,主动url编码\n if method==\"a\":\n value=parse.quote(parse.unquote(param.get(\"value\"))+new)\n else:\n value=parse.quote(new)\n return value\n if param.get(\"type\") in [0,2]: #cookie ,url ,request会自动url编码\n if method==\"a\":\n value=parse.unquote(param.get(\"value\"))+new\n else:\n value=new\n return value\n if param.get(\"type\")>2: #xml,json等不编码,按理说json要把\"等转义,后头再弄\n if method==\"a\":\n value=param.get(\"value\")+new\n else:\n value=new\n return value\n\n\n\n\n\n", "sub_path": "myscan/lib/core/common.py", "file_name": "common.py", "file_ext": "py", "file_size_in_byte": 6131, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "myscan.lib.core.data.cmd_line_options.redis", "line_number": 17, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.cmd_line_options", "line_number": 17, "usage_type": "name"}, {"api_name": "myscan.lib.core.data.logger.info", "line_number": 29, "usage_type": "call"}, {"api_name": "myscan.lib.core.data.logger", "line_number": 29, "usage_type": "name"}, {"api_name": "myscan.lib.core.data.conn.redis", "line_number": 30, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.conn", "line_number": 30, "usage_type": "name"}, {"api_name": "redis.ConnectionPool", "line_number": 30, "usage_type": "call"}, {"api_name": "myscan.lib.core.data.logger.warning", "line_number": 37, "usage_type": "call"}, {"api_name": "myscan.lib.core.data.logger", "line_number": 37, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 38, "usage_type": "call"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_ROOT_PATH", "line_number": 43, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 43, "usage_type": "name"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_DATA_PATH", "line_number": 44, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 44, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_ROOT_PATH", "line_number": 44, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_PLUGINS_PATH", "line_number": 45, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 45, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_ROOT_PATH", "line_number": 45, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_POCS_PATH", "line_number": 47, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 47, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 47, "usage_type": "call"}, {"api_name": "os.path", "line_number": 47, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_ROOT_PATH", "line_number": 47, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_REPORT_PATH", "line_number": 48, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 48, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 48, "usage_type": "call"}, {"api_name": "os.path", "line_number": 48, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_ROOT_PATH", "line_number": 48, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.USER_POCS_PATH", "line_number": 49, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 49, "usage_type": "name"}, {"api_name": "myscan.lib.core.data.paths.SENSETIVE_DIR", "line_number": 51, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 51, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 51, "usage_type": "call"}, {"api_name": "os.path", "line_number": 51, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_DATA_PATH", "line_number": 51, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.WEAK_PASS", "line_number": 52, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 52, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_DATA_PATH", "line_number": 52, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.LARGE_WEAK_PASS", "line_number": 53, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths", "line_number": 53, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.paths.MYSCAN_DATA_PATH", "line_number": 53, "usage_type": "attribute"}, {"api_name": "random.sample", "line_number": 70, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 74, "usage_type": "call"}, {"api_name": "difflib.SequenceMatcher", "line_number": 90, "usage_type": "call"}, {"api_name": "redis.StrictRedis", "line_number": 96, "usage_type": "call"}, {"api_name": "myscan.lib.core.data.conn.redis", "line_number": 96, "usage_type": "attribute"}, {"api_name": "myscan.lib.core.data.conn", "line_number": 96, "usage_type": "name"}, {"api_name": "urllib.parse.urlparse", "line_number": 104, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 104, "usage_type": "name"}, {"api_name": "hashlib.md5", "line_number": 113, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 123, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 164, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 164, "usage_type": "name"}, {"api_name": "urllib.parse.unquote", "line_number": 164, "usage_type": "call"}, {"api_name": "urllib.parse.quote", "line_number": 166, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 166, "usage_type": "name"}, {"api_name": "urllib.parse.unquote", "line_number": 170, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 170, "usage_type": "name"}]} +{"seq_id": "224218749", "text": "######## Webcam Object Detection Using Tensorflow-trained Classifier #########\n#\n# Author: Evan Juras\n# Date: 1/20/18\n# Description: \n# This program uses a TensorFlow-trained classifier to perform object detection.\n# It loads the classifier and uses it to perform object detection on a webcam feed.\n# It draws boxes, scores, and labels around the objects of interest in each frame\n# from the webcam.\n\n## Some of the code is copied from Google's example at\n## https://github.com/tensorflow/models/blob/master/research/object_detection/object_detection_tutorial.ipynb\n\n## and some is copied from Dat Tran's example at\n## https://github.com/datitran/object_detector_app/blob/master/object_detection_app.py\n\n## but I changed it to make it more understandable to me.\nimport serial\nimport face_recognition\nfrom pymongo import MongoClient, ReadPreference\nfrom sshtunnel import SSHTunnelForwarder\nimport paramiko\nimport camera\n# Import packages\nimport os\nimport cv2\nimport numpy as np\nimport tensorflow as tf\nimport sys\nfrom datetime import date\nprint(\"done import?\")\n# This is needed since the notebook is stored in the object_detection folder.\nsys.path.append(\"..\")\n\nPORT = 'COM8'\nBaudRate = 9600\nser = serial.Serial(PORT, BaudRate)\n# Import utilites\nfrom object_detection.utils import label_map_util\nfrom object_detection.utils import visualization_utils as vis_util\nprint(\"import really done?\")\n# Name of the directory containing the object detection module we're using\nMODEL_NAME = '/Users/q/Desktop/webcam/object_detection/inference_graph'\n\n# Grab path to current working directory\nCWD_PATH = os.getcwd()\nprint(\"got cwd\")\n# Path to frozen detection graph .pb file, which contains the model that is used\n# for object detection.\nPATH_TO_CKPT = os.path.join(CWD_PATH,MODEL_NAME,'frozen_inference_graph.pb')\n\n# Path to label map file\nPATH_TO_LABELS = os.path.join(CWD_PATH,MODEL_NAME,'label_map.pbtxt')\nprint(\"got all paths\")\n# Number of classes the object detector can identify\nNUM_CLASSES = 3\nprint(PATH_TO_LABELS)\n## Load the label map.\n# Label maps map indices to category names, so that when our convolution\n# network predicts `5`, we know that this corresponds to `king`.\n# Here we use internal utility functions, but anything that returns a\n# dictionary mapping integers to appropriate string labels would be fine\nlabel_map = label_map_util.load_labelmap(PATH_TO_LABELS)\nprint(\"got lb map\")\ncategories = label_map_util.convert_label_map_to_categories(label_map, max_num_classes=NUM_CLASSES, use_display_name=True)\nprint(\"parsed\")\ncategory_index = label_map_util.create_category_index(categories)\nprint(\"starting to load model\")\n# Load the Tensorflow model into memory.\ndetection_graph = tf.Graph()\nwith detection_graph.as_default():\n od_graph_def = tf.GraphDef()\n with tf.gfile.GFile(PATH_TO_CKPT, 'rb') as fid:\n serialized_graph = fid.read()\n od_graph_def.ParseFromString(serialized_graph)\n tf.import_graph_def(od_graph_def, name='')\n\n sess = tf.Session(graph=detection_graph)\n\nprint(\"model loaded\")\n# Define input and output tensors (i.e. data) for the object detection classifier\n\n# Input tensor is the image\nimage_tensor = detection_graph.get_tensor_by_name('image_tensor:0')\n\n# Output tensors are the detection boxes, scores, and classes\n# Each box represents a part of the image where a particular object was detected\ndetection_boxes = detection_graph.get_tensor_by_name('detection_boxes:0')\n\n# Each score represents level of confidence for each of the objects.\n# The score is shown on the result image, together with the class label.\ndetection_scores = detection_graph.get_tensor_by_name('detection_scores:0')\ndetection_classes = detection_graph.get_tensor_by_name('detection_classes:0')\ntoday = date.today()\ntodate = today.strftime(\"%Y-%m-%d\")\nprint(todate)\n# Number of objects detected\nnum_detections = detection_graph.get_tensor_by_name('num_detections:0')\nprint(\"ok here\")\n# Initialize webcam feed\nvideo = cv2.VideoCapture(0)\nret = video.set(3,1280)\nret = video.set(4,720)\n\nclass VideoCamera(object):\n def __init__(self, video):\n # Using OpenCV to capture from device 0. If you have trouble capturing\n # from a webcam, comment the line below out and use a video file\n # instead.\n self.video = video\n # If you decide to use video.mp4, you must have this file in the folder\n # as the main.py.\n # self.video = cv2.VideoCapture('video.mp4')\n\n def __del__(self):\n self.video.release()\n\n def get_frame(self):\n # Grab a single frame of video\n ret, frame = self.video.read()\n return frame\n\n\n\nclass FaceRecog():\n \n def __init__(self, video):\n # Using OpenCV to capture from device 0. If you have trouble capturing\n # from a webcam, comment the line below out and use a video file\n # instead.\n self.camera = VideoCamera(video)\n\n self.known_face_encodings = []\n self.known_face_names = []\n\n # Load sample pictures and learn how to recognize it.\n dirname = '/Users/q/Desktop/webcam/object_detection/knowns'\n files = os.listdir(dirname)\n for filename in files:\n name, ext = os.path.splitext(filename)\n if ext == '.jpg':\n self.known_face_names.append(name)\n pathname = os.path.join(dirname, filename)\n img = face_recognition.load_image_file(pathname)\n face_encoding = face_recognition.face_encodings(img)[0]\n self.known_face_encodings.append(face_encoding)\n\n # Initialize some variables\n self.face_locations = []\n self.face_encodings = []\n self.face_names = []\n self.process_this_frame = True\n\n def __del__(self):\n del self.camera\n\n def get_frame(self, my_face_names = []):\n # Grab a single frame of video\n frame = self.camera.get_frame()\n\n # Resize frame of video to 1/4 size for faster face recognition processing\n small_frame = cv2.resize(frame, (0, 0), fx=0.25, fy=0.25)\n\n # Convert the image from BGR color (which OpenCV uses) to RGB color (which face_recognition uses)\n rgb_small_frame = small_frame[:, :, ::-1]\n \n # Find all the faces and face encodings in the current frame of video\n self.face_locations = face_recognition.face_locations(rgb_small_frame)\n self.face_encodings = face_recognition.face_encodings(rgb_small_frame, self.face_locations)\n\n # Display the results\n for (top, right, bottom, left), name in zip(self.face_locations, my_face_names):\n # Scale back up face locations since the frame we detected in was scaled to 1/4 size\n top *= 4\n right *= 4\n bottom *= 4\n left *= 4\n\n # Draw a box around the face\n cv2.rectangle(frame, (left, top), (right, bottom), (0, 0, 255), 2)\n\n # Draw a label with a name below the face\n cv2.rectangle(frame, (left, bottom - 35), (right, bottom), (0, 0, 255), cv2.FILLED)\n font = cv2.FONT_HERSHEY_DUPLEX\n cv2.putText(frame, name, (left + 6, bottom - 6), font, 1.0, (255, 255, 255), 1)\n\n return frame\n\n def get_name(self):\n\n if self.process_this_frame:\n \n self.face_names = []\n for face_encoding in self.face_encodings:\n # See if the face is a match for the known face(s)\n distances = face_recognition.face_distance(self.known_face_encodings, face_encoding)\n min_value = min(distances)\n\n # tolerance: How much distance between faces to consider it a match. Lower is more strict.\n # 0.6 is typical best performance.\n name = \"Unknown\"\n if min_value < 0.6:\n index = np.argmin(distances)\n name = self.known_face_names[index]\n\n self.face_names.append(name)\n\n self.process_this_frame = not self.process_this_frame\n \n return self.face_names\n\n\n def get_jpg_bytes(self):\n frame = self.get_frame()\n # We are using Motion JPEG, but OpenCV defaults to capture raw images,\n # so we must encode it into JPEG in order to correctly display the\n # video stream.\n ret, jpg = cv2.imencode('.jpg', frame)\n return jpg.tobytes()\n\n\nplastic_cnt=0\nmetal_cnt=0\nglass_cnt=0\nfoundface = None\nface_recog = FaceRecog(video)\nwhile(True):\n # print(1)\n gets = face_recog.get_name()\n frameface = face_recog.get_frame(gets)\n if len(gets)>0:\n foundface = str(gets[0])\n print(foundface)\n # show the frame\n # cv2.imshow(\"Frame\", frameface)\n\n\n # SSH통해서 mongodb접속하기 \n SSH_KEY_LOCATION = 'C:/Users/q/Downloads/cs496-key.pem' \n JUMP_MACHINE_ADDRESS = '192.249.19.252'\n SSH_USER = 'root'\n REMOTE_MONGO_ADDRESS = '127.0.0.1'\n\n pkey = paramiko.RSAKey.from_private_key_file(SSH_KEY_LOCATION)\n server = SSHTunnelForwarder(\n (JUMP_MACHINE_ADDRESS, 2022),\n ssh_username=SSH_USER,\n ssh_private_key=pkey,\n remote_bind_address=(REMOTE_MONGO_ADDRESS, 27017),\n local_bind_address=('0.0.0.0', 27017)\n )\n # 접속 시작 \n server.start()\n # print(server.is_active)\n DB_NAME = 'trash'\n COLLECTION_NAME = 'logins'\n # mongodb 접속 \n client = MongoClient('mongodb://127.0.0.1:27017')\n db = client[DB_NAME]\n col = db[COLLECTION_NAME]\n # print(db, col)\n #print(col.find_one({\"userID\":\"yhs25\"}))\n # Acquire frame and expand frame dimensions to have shape: [1, None, None, 3]\n # i.e. a single-column array, where each item in the column has the pixel RGB value\n ret, frame = video.read()\n frame_expanded = np.expand_dims(frame, axis=0)\n\n # Perform the actual detection by running the model with the image as input\n (boxes, scores, classes, num) = sess.run(\n [detection_boxes, detection_scores, detection_classes, num_detections],\n feed_dict={image_tensor: frame_expanded})\n\n # Draw the results of the detection (aka 'visulaize the results')\n vis_util.visualize_boxes_and_labels_on_image_array(\n frame,\n np.squeeze(boxes),\n np.squeeze(classes).astype(np.int32),\n np.squeeze(scores),\n category_index,\n use_normalized_coordinates=True,\n line_thickness=6,\n min_score_thresh=0.95)\n # All the results have been drawn on the frame, so it's time to display it.\n cv2.imshow('Object detector', frame)\n\n # print([category_index.get(i) for i in classes[0]])\n # print(scores)\n detected_object = category_index.get(classes[0][0])['name']\n score = scores[0][0]\n if score>0.95:\n if detected_object == 'glass':\n glass_cnt+=1\n elif detected_object == 'metal':\n metal_cnt+=1\n elif detected_object =='plastic':\n plastic_cnt+=1\n\n if glass_cnt>10:\n #send signal to arduino\n glass_cnt=0\n metal_cnt=0\n plastic_cnt=0\n op = \"1\"\n ser.write(op.encode())\n if(not foundface==None):\n col.update_one({\"userID\": str(foundface)}, {\"$push\": {\"points\":{\"type\":\"glass\", \"date\":str(todate), \"point\":\"20\"}}})\n print(\"opened glass bin, \"+str(foundface)+\"+20pt!!\")\n # foundface=None\n else:\n print(\"opened glass bin\")\n elif metal_cnt>10:\n #send signal to arduino\n glass_cnt=0\n metal_cnt=0\n plastic_cnt=0\n op = \"2\"\n ser.write(op.encode())\n if(not foundface==None):\n col.update_one({\"userID\": str(foundface)}, {\"$push\": {\"points\":{\"type\":\"metal\", \"date\":str(todate), \"point\":\"30\"}}})\n print(\"opened metal bin, \"+ str(foundface)+\"+30pt!!\")\n # foundface=None\n else:\n print(\"opened metal bin\")\n elif plastic_cnt>10:\n #send signal to arduino\n glass_cnt=0\n metal_cnt=0\n plastic_cnt=0\n op = \"3\"\n ser.write(op.encode())\n if(not foundface==None):\n col.update_one({\"userID\": str(foundface)}, {\"$push\": {\"points\":{\"type\":\"plastic\", \"date\":str(todate), \"point\":\"10\"}}})\n print(\"opened plastic bin, \"+ str(foundface)+\"+10pt!!\")\n # foundface=None\n else:\n print(\"opened plastic bin\")\n # Press 'q' to quit\n if cv2.waitKey(1) == ord('q'):\n break\n\n# Clean up\nvideo.release()\ncv2.destroyAllWindows()\n", "sub_path": "Object_detection_webcam.py", "file_name": "Object_detection_webcam.py", "file_ext": "py", "file_size_in_byte": 12473, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.path.append", "line_number": 33, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "serial.Serial", "line_number": 37, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "object_detection.utils.label_map_util.load_labelmap", "line_number": 63, "usage_type": "call"}, {"api_name": "object_detection.utils.label_map_util", "line_number": 63, "usage_type": "name"}, {"api_name": "object_detection.utils.label_map_util.convert_label_map_to_categories", "line_number": 65, "usage_type": "call"}, {"api_name": "object_detection.utils.label_map_util", "line_number": 65, "usage_type": "name"}, {"api_name": "object_detection.utils.label_map_util.create_category_index", "line_number": 67, "usage_type": "call"}, {"api_name": "object_detection.utils.label_map_util", "line_number": 67, "usage_type": "name"}, {"api_name": "tensorflow.Graph", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.GraphDef", "line_number": 72, "usage_type": "call"}, {"api_name": "tensorflow.gfile.GFile", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.gfile", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.import_graph_def", "line_number": 76, "usage_type": "call"}, {"api_name": "tensorflow.Session", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.date.today", "line_number": 94, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 94, "usage_type": "name"}, {"api_name": "cv2.VideoCapture", "line_number": 101, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "face_recognition.load_image_file", "line_number": 144, "usage_type": "call"}, {"api_name": "face_recognition.face_encodings", "line_number": 145, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 162, "usage_type": "call"}, {"api_name": "face_recognition.face_locations", "line_number": 168, "usage_type": "call"}, {"api_name": "face_recognition.face_encodings", "line_number": 169, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 180, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 183, "usage_type": "call"}, {"api_name": "cv2.FILLED", "line_number": 183, "usage_type": "attribute"}, {"api_name": "cv2.FONT_HERSHEY_DUPLEX", "line_number": 184, "usage_type": "attribute"}, {"api_name": "cv2.putText", "line_number": 185, "usage_type": "call"}, {"api_name": "face_recognition.face_distance", "line_number": 196, "usage_type": "call"}, {"api_name": "numpy.argmin", "line_number": 203, "usage_type": "call"}, {"api_name": "cv2.imencode", "line_number": 218, "usage_type": "call"}, {"api_name": "paramiko.RSAKey.from_private_key_file", "line_number": 244, "usage_type": "call"}, {"api_name": "paramiko.RSAKey", "line_number": 244, "usage_type": "attribute"}, {"api_name": "sshtunnel.SSHTunnelForwarder", "line_number": 245, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 266, "usage_type": "call"}, {"api_name": "object_detection.utils.visualization_utils.visualize_boxes_and_labels_on_image_array", "line_number": 274, "usage_type": "call"}, {"api_name": "object_detection.utils.visualization_utils", "line_number": 274, "usage_type": "name"}, {"api_name": "numpy.squeeze", "line_number": 276, "usage_type": "call"}, {"api_name": "numpy.squeeze", "line_number": 277, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 277, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 278, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 284, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 338, "usage_type": "call"}, {"api_name": "cv2.destroyAllWindows", "line_number": 343, "usage_type": "call"}]} +{"seq_id": "249500146", "text": "from django.shortcuts import render\nfrom venus.apps.notes.models import Note\nfrom venus.apps.notes.forms import AddNoteForm\nfrom django.http import HttpResponse # HttpResponseRedirect\nimport json\nimport random\n\n\ndef index(request):\n \"\"\" output of all notes \"\"\"\n all_notes_list = Note.objects.order_by('id')\n return render(request, 'notes/index.html', {\n 'all_notes_list': all_notes_list\n })\n\n\ndef add_note(request):\n if request.method == 'POST':\n form = AddNoteForm(request.POST)\n response = HttpResponse()\n if form.is_valid(): # pylint: disable=E1101\n form.save() # pylint: disable=E1101\n data = request.POST\n data = data.dict()\n data['alert-type'] = 'success'\n else:\n data = form.errors # pylint: disable=E1101\n data['alert-type'] = 'error'\n response.content = json.dumps(data)\n return response\n else:\n form = AddNoteForm()\n\n return render(request, 'notes/add_note.html', {\n 'form': form,\n })\n\ndef random_note(request):\n \"\"\" output random note \"\"\"\n notes_list_ids = range(1, Note.objects.count())\n note = Note.objects.get(pk=random.choice(notes_list_ids))\n return render(request, 'notes/random_note.html', {'note': note})\n", "sub_path": "venus/venus/apps/notes/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1292, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "venus.apps.notes.models.Note.objects.order_by", "line_number": 11, "usage_type": "call"}, {"api_name": "venus.apps.notes.models.Note.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "venus.apps.notes.models.Note", "line_number": 11, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 12, "usage_type": "call"}, {"api_name": "venus.apps.notes.forms.AddNoteForm", "line_number": 19, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 20, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 29, "usage_type": "call"}, {"api_name": "venus.apps.notes.forms.AddNoteForm", "line_number": 32, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 34, "usage_type": "call"}, {"api_name": "venus.apps.notes.models.Note.objects.count", "line_number": 40, "usage_type": "call"}, {"api_name": "venus.apps.notes.models.Note.objects", "line_number": 40, "usage_type": "attribute"}, {"api_name": "venus.apps.notes.models.Note", "line_number": 40, "usage_type": "name"}, {"api_name": "venus.apps.notes.models.Note.objects.get", "line_number": 41, "usage_type": "call"}, {"api_name": "venus.apps.notes.models.Note.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "venus.apps.notes.models.Note", "line_number": 41, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}]} +{"seq_id": "23392136", "text": "import os\nimport json\nimport torch\nfrom torchvision import transforms\nimport numpy as np\nfrom PIL import Image\nimport copy\n\n\ndef imresize(im, size, interp='bilinear'):\n if interp == 'nearest':\n resample = Image.NEAREST\n elif interp == 'bilinear':\n resample = Image.BILINEAR\n elif interp == 'bicubic':\n resample = Image.BICUBIC\n else:\n raise Exception('resample method undefined!')\n\n return im.resize(size, resample)\n\n\nclass BaseDataset(torch.utils.data.Dataset):\n def __init__(self, odgt, opt, **kwargs):\n # parse options\n self.imgSizes = opt.imgSizes\n self.imgMaxSize = opt.imgMaxSize\n # max down sampling rate of network to avoid rounding during conv or pooling\n self.padding_constant = opt.padding_constant\n self.num_class = opt.num_class\n\n # parse the input list\n self.parse_input_list(odgt, **kwargs)\n\n # mean and std\n self.normalize = transforms.Normalize(\n mean=[0.485, 0.456, 0.406],\n std=[0.229, 0.224, 0.225])\n\n def parse_input_list(self, odgt, max_sample=-1, start_idx=-1, end_idx=-1):\n if isinstance(odgt, list):\n self.list_sample = odgt\n elif isinstance(odgt, str):\n self.list_sample = [json.loads(x.rstrip()) for x in open(odgt, 'r')]\n\n if max_sample > 0:\n self.list_sample = self.list_sample[0:max_sample]\n if start_idx >= 0 and end_idx >= 0: # divide file list\n self.list_sample = self.list_sample[start_idx:end_idx]\n\n self.list_sample_orig = copy.deepcopy(self.list_sample)\n\n self.num_sample = len(self.list_sample)\n assert self.num_sample > 0\n print('# samples: {}'.format(self.num_sample))\n\n def img_transform(self, img):\n # 0-255 to 0-1\n img = np.float32(np.array(img)) / 255.\n img = img.transpose((2, 0, 1))\n img = self.normalize(torch.from_numpy(img.copy()))\n return img\n\n def segm_transform(self, segm):\n # to tensor, -1 to 149\n segm = torch.from_numpy(np.array(segm)).long() - 1\n return segm\n\n def segm_one_hot(self, segm):\n segm = torch.from_numpy(np.array(segm)).long().unsqueeze(0)\n size = segm.size()\n oneHot_size = (self.num_class+1, size[1], size[2])\n segm_oneHot = torch.FloatTensor(torch.Size(oneHot_size)).zero_()\n segm_oneHot = segm_oneHot.scatter_(0, segm, 1.0)\n return segm_oneHot\n\n # Round x to the nearest multiple of p and x' >= x\n def round2nearest_multiple(self, x, p):\n return ((x - 1) // p + 1) * p\n\n\nclass TrainDataset(BaseDataset):\n def __init__(self, root_dataset, odgt, opt, ref_path, ref_start=0, ref_end=3, batch_per_gpu=1, **kwargs):\n super(TrainDataset, self).__init__(odgt, opt, **kwargs)\n self.root_dataset = root_dataset\n # down sampling rate of segm labe\n self.segm_downsampling_rate = opt.segm_downsampling_rate\n self.batch_per_gpu = batch_per_gpu\n\n # classify images into two classes: 1. h > w and 2. h <= w\n self.batch_record_list = [[], []]\n self.batch_ref_list = [[], []]\n\n # override dataset length when trainig with batch_per_gpu > 1\n self.cur_idx = 0\n self.if_shuffled = False\n\n self.ref_start = ref_start\n self.ref_end = ref_end\n\n self.random_pick = opt.random_pick\n self.no_align = opt.no_align\n self.zero_input_rgb = opt.zero_input_rgb\n self.zero_input_seg = opt.zero_input_seg\n self.random_input_seg = opt.random_input_seg\n self.RGB_mask_combine_val = opt.RGB_mask_combine_val\n\n with open(ref_path, 'r') as f:\n lines = f.readlines()\n self.ref_list = [[int(item) for item in line.strip().split()] for line in lines]\n assert len(self.ref_list) == len(self.list_sample)\n\n def _get_sub_batch(self):\n while True:\n # get a sample record\n this_sample = self.list_sample[self.cur_idx]\n this_ref_list = self.ref_list[self.cur_idx]\n if this_sample['height'] > this_sample['width']:\n self.batch_record_list[0].append(this_sample) # h > w, go to 1st class\n self.batch_ref_list[0].append(this_ref_list)\n else:\n self.batch_record_list[1].append(this_sample) # h <= w, go to 2nd class\n self.batch_ref_list[1].append(this_ref_list)\n\n # update current sample pointer\n self.cur_idx += 1\n if self.cur_idx >= self.num_sample:\n self.cur_idx = 0\n permutation = np.random.permutation(len(self.list_sample))\n self.list_sample = [self.list_sample[i] for i in permutation]\n self.ref_list = [self.ref_list[i] for i in permutation]\n\n if len(self.batch_record_list[0]) == self.batch_per_gpu:\n batch_records = self.batch_record_list[0]\n self.batch_record_list[0] = []\n ref_lists = self.batch_ref_list[0]\n self.batch_ref_list[0] = []\n break\n elif len(self.batch_record_list[1]) == self.batch_per_gpu:\n batch_records = self.batch_record_list[1]\n self.batch_record_list[1] = []\n ref_lists = self.batch_ref_list[1]\n self.batch_ref_list[1] = []\n break\n return batch_records, ref_lists\n\n def __getitem__(self, index):\n # NOTE: random shuffle for the first time. shuffle in __init__ is useless\n if not self.if_shuffled:\n np.random.seed(index)\n permutation = np.random.permutation(len(self.list_sample))\n self.list_sample = [self.list_sample[i] for i in permutation]\n self.ref_list = [self.ref_list[i] for i in permutation]\n self.if_shuffled = True\n\n # get sub-batch candidates\n batch_records, ref_lists = self._get_sub_batch()\n\n # resize all images' short edges to the chosen size\n if isinstance(self.imgSizes, list) or isinstance(self.imgSizes, tuple):\n this_short_size = np.random.choice(self.imgSizes)\n else:\n this_short_size = self.imgSizes\n\n # calculate the BATCH's height and width\n # since we concat more than one samples, the batch's h and w shall be larger than EACH sample\n batch_widths = np.zeros(self.batch_per_gpu, np.int32)\n batch_heights = np.zeros(self.batch_per_gpu, np.int32)\n for i in range(self.batch_per_gpu):\n img_height, img_width = batch_records[i]['height'], batch_records[i]['width']\n this_scale = min(\n this_short_size / min(img_height, img_width), \\\n self.imgMaxSize / max(img_height, img_width))\n batch_widths[i] = img_width * this_scale\n batch_heights[i] = img_height * this_scale\n\n # Here we must pad both input image and segmentation map to size h' and w' so that p | h' and p | w'\n batch_width = np.max(batch_widths)\n batch_height = np.max(batch_heights)\n batch_width = int(self.round2nearest_multiple(batch_width, self.padding_constant))\n batch_height = int(self.round2nearest_multiple(batch_height, self.padding_constant))\n\n assert self.padding_constant >= self.segm_downsampling_rate, \\\n 'padding constant must be equal or large than segm downsamping rate'\n batch_images = torch.zeros(\n self.batch_per_gpu, 3, batch_height, batch_width)\n batch_segms = torch.zeros(\n self.batch_per_gpu,\n batch_height // self.segm_downsampling_rate,\n batch_width // self.segm_downsampling_rate).long()\n\n batch_refs_rgb = torch.zeros(\n self.batch_per_gpu, 3, self.random_pick, batch_height, batch_width)\n\n if self.RGB_mask_combine_val:\n batch_refs_mask = torch.zeros(\n self.batch_per_gpu, 3+1+self.num_class, self.random_pick, batch_height, batch_width)\n else:\n batch_refs_mask = torch.zeros(\n self.batch_per_gpu, 1+self.num_class, self.random_pick, batch_height, batch_width)\n\n #infos = []\n\n for i in range(self.batch_per_gpu):\n #info_single = {}\n #info_single['ref_path'] = []\n\n this_record = batch_records[i]\n\n # load image and label\n image_path = os.path.join(self.root_dataset, this_record['fpath_img'])\n segm_path = os.path.join(self.root_dataset, this_record['fpath_segm'])\n\n #info_single['query_path'] = (image_path, segm_path)\n\n img = Image.open(image_path).convert('RGB')\n segm = Image.open(segm_path)\n assert(segm.mode == \"L\")\n assert(img.size[0] == segm.size[0])\n assert(img.size[1] == segm.size[1])\n\n # random_flip\n if np.random.choice([0, 1]):\n img = img.transpose(Image.FLIP_LEFT_RIGHT)\n segm = segm.transpose(Image.FLIP_LEFT_RIGHT)\n\n # note that each sample within a mini batch has different scale param\n img = imresize(img, (batch_widths[i], batch_heights[i]), interp='bilinear')\n segm = imresize(segm, (batch_widths[i], batch_heights[i]), interp='nearest')\n\n # further downsample seg label, need to avoid seg label misalignment\n segm_rounded_width = self.round2nearest_multiple(segm.size[0], self.segm_downsampling_rate)\n segm_rounded_height = self.round2nearest_multiple(segm.size[1], self.segm_downsampling_rate)\n segm_rounded = Image.new('L', (segm_rounded_width, segm_rounded_height), 0)\n segm_rounded.paste(segm, (0, 0))\n segm = imresize(\n segm_rounded,\n (segm_rounded.size[0] // self.segm_downsampling_rate, \\\n segm_rounded.size[1] // self.segm_downsampling_rate), \\\n interp='nearest')\n\n # image transform, to torch float tensor 3xHxW\n img = self.img_transform(img)\n\n # segm transform, to torch long tensor HxW\n segm = self.segm_transform(segm)\n\n # put into batch arrays\n batch_images[i][:, :img.shape[1], :img.shape[2]] = img\n batch_segms[i][:segm.shape[0], :segm.shape[1]] = segm\n\n # prepare the references\n this_ref_list = ref_lists[i]\n\n ref_perm = np.random.permutation(self.ref_end - self.ref_start)\n\n #for k in range(self.ref_end - self.ref_start):\n for idx in range(self.random_pick):\n k = ref_perm[idx]\n if self.no_align:\n ref_record1 = self.list_sample_orig[this_ref_list[k+self.ref_start]]\n ref_record2 = self.list_sample_orig[this_ref_list[k+self.ref_start+10]]\n image_path = os.path.join(self.root_dataset, ref_record1['fpath_img'])\n segm_path = os.path.join(self.root_dataset, ref_record2['fpath_segm'])\n else:\n ref_record = self.list_sample_orig[this_ref_list[k+self.ref_start]]\n image_path = os.path.join(self.root_dataset, ref_record['fpath_img'])\n segm_path = os.path.join(self.root_dataset, ref_record['fpath_segm'])\n\n #info_single['ref_path'].append((image_path, segm_path))\n\n img = Image.open(image_path).convert('RGB')\n segm = Image.open(segm_path)\n assert(segm.mode == \"L\")\n if not self.no_align:\n assert(img.size[0] == segm.size[0])\n assert(img.size[1] == segm.size[1])\n\n if np.random.choice([0, 1]):\n img = img.transpose(Image.FLIP_LEFT_RIGHT)\n segm = segm.transpose(Image.FLIP_LEFT_RIGHT)\n img = imresize(img, (batch_widths[i], batch_heights[i]), interp='bilinear')\n segm = imresize(segm, (batch_widths[i], batch_heights[i]), interp='nearest')\n\n img = self.img_transform(img)\n segm = self.segm_one_hot(segm)\n\n batch_refs_rgb[i][:, idx, :img.shape[1], :img.shape[2]] = img\n if self.RGB_mask_combine_val:\n batch_refs_mask[i][0:3, idx, :segm.shape[1], :segm.shape[2]] = img\n batch_refs_mask[i][3:, idx, :segm.shape[1], :segm.shape[2]] = segm\n else:\n batch_refs_mask[i][:, idx, :segm.shape[1], :segm.shape[2]] = segm\n\n if self.zero_input_rgb:\n batch_refs_rgb[i][:, idx, :img.shape[1], :img.shape[2]] = 0.\n\n if self.zero_input_seg:\n if self.RGB_mask_combine_val:\n batch_refs_mask[i][3:, idx, :segm.shape[1], :segm.shape[2]] = 0.\n else:\n batch_refs_mask[i][:, idx, :segm.shape[1], :segm.shape[2]] = 0.\n elif self.random_input_seg:\n if self.RGB_mask_combine_val:\n batch_refs_mask[i][3:, idx, :segm.shape[1], :segm.shape[2]] = torch.rand_like(segm)\n else:\n batch_refs_mask[i][:, idx, :segm.shape[1], :segm.shape[2]] = torch.rand_like(segm)\n\n #infos.append(info_single)\n\n output = dict()\n output['img_data'] = batch_images\n output['seg_label'] = batch_segms\n output['img_refs_rgb'] = batch_refs_rgb\n output['img_refs_mask'] = batch_refs_mask\n return output\n\n def __len__(self):\n return int(1e10) # It's a fake length due to the trick that every loader maintains its own list\n #return self.num_sampleclass\n\n\nclass ValDataset(BaseDataset):\n def __init__(self, root_dataset, odgt, opt, **kwargs):\n super(ValDataset, self).__init__(odgt, opt, **kwargs)\n self.root_dataset = root_dataset\n\n self.train_list_sample = [json.loads(x.rstrip()) for x in open(opt.list_train, 'r')]\n\n self.train_num_sample = len(self.train_list_sample)\n assert self.train_num_sample > 0\n print('# training samples: {}'.format(self.train_num_sample))\n\n self.RGB_mask_combine_val = opt.RGB_mask_combine_val\n\n self.debug_with_gt = opt.debug_with_gt\n self.debug_with_translated_gt = opt.debug_with_translated_gt\n self.debug_with_random = opt.debug_with_random\n self.debug_with_double_random = opt.debug_with_double_random\n self.debug_with_double_complete_random = opt.debug_with_double_complete_random\n self.debug_with_randomSegNoise = opt.debug_with_randomSegNoise\n\n self.ref_start = opt.ref_val_start\n self.ref_end = opt.ref_val_end\n\n with open(opt.ref_val_path, 'r') as f:\n lines = f.readlines()\n self.ref_list = [[int(item) for item in line.strip().split()] for line in lines]\n\n start_idx = kwargs['start_idx']\n end_idx = kwargs['end_idx']\n if start_idx >= 0 and end_idx >= 0: # divide file list\n self.ref_list = self.ref_list[start_idx:end_idx]\n assert len(self.ref_list) == len(self.list_sample)\n\n def __getitem__(self, index):\n this_record = self.list_sample[index]\n this_ref_list = self.ref_list[index]\n # load image and label\n image_path = os.path.join(self.root_dataset, this_record['fpath_img'])\n segm_path = os.path.join(self.root_dataset, this_record['fpath_segm'])\n img = Image.open(image_path).convert('RGB')\n segm = Image.open(segm_path)\n assert(segm.mode == \"L\")\n assert(img.size[0] == segm.size[0])\n assert(img.size[1] == segm.size[1])\n\n # segm transform, to torch long tensor HxW\n segm = self.segm_transform(segm)\n batch_segms = torch.unsqueeze(segm, 0)\n\n ori_width, ori_height = img.size\n\n img_resized_list = []\n ref_rgb_resized_list = []\n ref_mask_resized_list = []\n for this_short_size in self.imgSizes:\n # calculate target height and width\n scale = min(this_short_size / float(min(ori_height, ori_width)),\n self.imgMaxSize / float(max(ori_height, ori_width)))\n target_height, target_width = int(ori_height * scale), int(ori_width * scale)\n\n # to avoid rounding in network\n target_width = self.round2nearest_multiple(target_width, self.padding_constant)\n target_height = self.round2nearest_multiple(target_height, self.padding_constant)\n\n # resize images\n img_resized = imresize(img, (target_width, target_height), interp='bilinear')\n\n # image transform, to torch float tensor 3xHxW\n img_resized = self.img_transform(img_resized)\n img_resized = torch.unsqueeze(img_resized, 0)\n img_resized_list.append(img_resized)\n\n batch_refs_rgb = torch.zeros(\n 3, self.ref_end-self.ref_start, target_height, target_width)\n if self.RGB_mask_combine_val:\n batch_refs_mask = torch.zeros(\n 3+1+self.num_class, self.ref_end-self.ref_start, target_height, target_width)\n else:\n batch_refs_mask = torch.zeros(\n 1+self.num_class, self.ref_end-self.ref_start, target_height, target_width)\n\n for k in range(self.ref_end - self.ref_start):\n ref_record = self.train_list_sample[this_ref_list[k+self.ref_start]]\n image_ref_path = os.path.join(self.root_dataset, ref_record['fpath_img'])\n segm_ref_path = os.path.join(self.root_dataset, ref_record['fpath_segm'])\n\n img_ref = Image.open(image_ref_path).convert('RGB')\n segm_ref = Image.open(segm_ref_path)\n assert(segm_ref.mode == \"L\")\n assert(img_ref.size[0] == segm_ref.size[0])\n assert(img_ref.size[1] == segm_ref.size[1])\n\n img_ref = imresize(img_ref, (target_width, target_height), interp='bilinear')\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n\n img_ref = self.img_transform(img_ref)\n segm_ref = self.segm_one_hot(segm_ref)\n\n batch_refs_rgb[:, k, :img_ref.shape[1], :img_ref.shape[2]] = img_ref\n\n if self.RGB_mask_combine_val:\n batch_refs_mask[0:3, k, :segm_ref.shape[1], :segm_ref.shape[2]] = img_ref\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n\n if self.debug_with_gt:\n img_resized_gt = img_resized[0]\n batch_refs_rgb[:, k, :img_resized_gt.shape[1], :img_resized_gt.shape[2]] = img_resized_gt\n \n segm_ref = Image.open(os.path.join(self.root_dataset, this_record['fpath_segm']))\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n segm_ref = self.segm_one_hot(segm_ref)\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n elif self.debug_with_translated_gt:\n img_resized_gt = img_resized[0]\n translation = 20\n batch_refs_rgb[:, k, translation:img_resized_gt.shape[1], translation:img_resized_gt.shape[2]] = img_resized_gt[:,:img_resized_gt.shape[1]-translation, :img_resized_gt.shape[2]-translation]\n \n segm_ref = Image.open(os.path.join(self.root_dataset, this_record['fpath_segm']))\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n segm_ref = self.segm_one_hot(segm_ref)\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, translation:segm_ref.shape[1], translation:segm_ref.shape[2]] = segm_ref[:, :segm_ref.shape[1]-translation, :segm_ref.shape[2]-translation]\n else:\n batch_refs_mask[:, k, translation:segm_ref.shape[1], translation:segm_ref.shape[2]] = segm_ref[:, :segm_ref.shape[1]-translation, :segm_ref.shape[2]-translation]\n elif self.debug_with_random:\n img_resized_gt = img_resized[0]\n batch_refs_rgb[:, k, :img_resized_gt.shape[1], :img_resized_gt.shape[2]] = img_resized_gt\n \n segm_ref = Image.open(os.path.join(self.root_dataset, ref_record['fpath_segm']))\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n segm_ref = self.segm_one_hot(segm_ref)\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n elif self.debug_with_double_random:\n ref_record_tmp = self.train_list_sample[this_ref_list[k+100+self.ref_start]]\n segm_ref = Image.open(os.path.join(self.root_dataset, ref_record_tmp['fpath_segm']))\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n segm_ref = self.segm_one_hot(segm_ref)\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n elif self.debug_with_double_complete_random:\n ref_record_tmp = self.train_list_sample[np.random.randint(0, len(self.train_list_sample))]\n segm_ref = Image.open(os.path.join(self.root_dataset, ref_record_tmp['fpath_segm']))\n segm_ref = imresize(segm_ref, (target_width, target_height), interp='nearest')\n segm_ref = self.segm_one_hot(segm_ref)\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = segm_ref\n elif self.debug_with_randomSegNoise:\n if self.RGB_mask_combine_val:\n batch_refs_mask[3:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = torch.rand_like(segm_ref)\n else:\n batch_refs_mask[:, k, :segm_ref.shape[1], :segm_ref.shape[2]] = torch.rand_like(segm_ref)\n\n batch_refs_rgb = torch.unsqueeze(batch_refs_rgb, 0)\n batch_refs_mask = torch.unsqueeze(batch_refs_mask, 0)\n ref_rgb_resized_list.append(batch_refs_rgb)\n ref_mask_resized_list.append(batch_refs_mask)\n\n output = dict()\n output['img_ori'] = np.array(img)\n output['img_data'] = [x.contiguous() for x in img_resized_list]\n output['seg_label'] = batch_segms.contiguous()\n output['img_refs_rgb'] = [x. contiguous() for x in ref_rgb_resized_list]\n output['img_refs_mask'] = [x. contiguous() for x in ref_mask_resized_list]\n output['info'] = this_record['fpath_img']\n return output\n\n def __len__(self):\n return self.num_sample\n\n\nclass TestDataset(BaseDataset):\n def __init__(self, odgt, opt, **kwargs):\n super(TestDataset, self).__init__(odgt, opt, **kwargs)\n\n def __getitem__(self, index):\n this_record = self.list_sample[index]\n # load image\n image_path = this_record['fpath_img']\n img = Image.open(image_path).convert('RGB')\n\n ori_width, ori_height = img.size\n\n img_resized_list = []\n for this_short_size in self.imgSizes:\n # calculate target height and width\n scale = min(this_short_size / float(min(ori_height, ori_width)),\n self.imgMaxSize / float(max(ori_height, ori_width)))\n target_height, target_width = int(ori_height * scale), int(ori_width * scale)\n\n # to avoid rounding in network\n target_width = self.round2nearest_multiple(target_width, self.padding_constant)\n target_height = self.round2nearest_multiple(target_height, self.padding_constant)\n\n # resize images\n img_resized = imresize(img, (target_width, target_height), interp='bilinear')\n\n # image transform, to torch float tensor 3xHxW\n img_resized = self.img_transform(img_resized)\n img_resized = torch.unsqueeze(img_resized, 0)\n img_resized_list.append(img_resized)\n\n output = dict()\n output['img_ori'] = np.array(img)\n output['img_data'] = [x.contiguous() for x in img_resized_list]\n output['info'] = this_record['fpath_img']\n return output\n\n def __len__(self):\n return self.num_sample\n", "sub_path": "dataset_memory_separate.py", "file_name": "dataset_memory_separate.py", "file_ext": "py", "file_size_in_byte": 25391, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "PIL.Image.NEAREST", "line_number": 12, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 12, "usage_type": "name"}, {"api_name": "PIL.Image.BILINEAR", "line_number": 14, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 14, "usage_type": "name"}, {"api_name": "PIL.Image.BICUBIC", "line_number": 16, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 16, "usage_type": "name"}, {"api_name": "torch.utils", "line_number": 23, "usage_type": "attribute"}, {"api_name": "torchvision.transforms.Normalize", "line_number": 36, "usage_type": "call"}, {"api_name": "torchvision.transforms", "line_number": 36, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 44, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.Size", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 129, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 129, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 150, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 151, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 161, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 161, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 167, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 167, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 168, "usage_type": "attribute"}, {"api_name": "numpy.max", "line_number": 178, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 185, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 187, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 196, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 199, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 211, "usage_type": "call"}, {"api_name": "os.path", "line_number": 211, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 212, "usage_type": "call"}, {"api_name": "os.path", "line_number": 212, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 216, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 216, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 217, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 217, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 223, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 223, "usage_type": "attribute"}, {"api_name": "PIL.Image.FLIP_LEFT_RIGHT", "line_number": 224, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 224, "usage_type": "name"}, {"api_name": "PIL.Image.FLIP_LEFT_RIGHT", "line_number": 225, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 225, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 234, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 234, "usage_type": "name"}, {"api_name": "numpy.random.permutation", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 255, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 263, "usage_type": "call"}, {"api_name": "os.path", "line_number": 263, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 264, "usage_type": "call"}, {"api_name": "os.path", "line_number": 264, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 267, "usage_type": "call"}, {"api_name": "os.path", "line_number": 267, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 268, "usage_type": "call"}, {"api_name": "os.path", "line_number": 268, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 272, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 272, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 273, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 273, "usage_type": "name"}, {"api_name": "numpy.random.choice", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 279, "usage_type": "attribute"}, {"api_name": "PIL.Image.FLIP_LEFT_RIGHT", "line_number": 280, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 280, "usage_type": "name"}, {"api_name": "PIL.Image.FLIP_LEFT_RIGHT", "line_number": 281, "usage_type": "attribute"}, {"api_name": "PIL.Image", "line_number": 281, "usage_type": "name"}, {"api_name": "torch.rand_like", "line_number": 305, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 307, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 328, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 360, "usage_type": "call"}, {"api_name": "os.path", "line_number": 360, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 361, "usage_type": "call"}, {"api_name": "os.path", "line_number": 361, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 362, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 362, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 363, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 363, "usage_type": "name"}, {"api_name": "torch.unsqueeze", "line_number": 370, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 392, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 395, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 398, "usage_type": "call"}, {"api_name": "torch.zeros", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 406, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 407, "usage_type": "call"}, {"api_name": "os.path", "line_number": 407, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 409, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 409, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 410, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 410, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 433, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 433, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 433, "usage_type": "call"}, {"api_name": "os.path", "line_number": 433, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 445, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 445, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path", "line_number": 445, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 456, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 456, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 456, "usage_type": "call"}, {"api_name": "os.path", "line_number": 456, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 465, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 465, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 465, "usage_type": "call"}, {"api_name": "os.path", "line_number": 465, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 473, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 473, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 474, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 474, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 474, "usage_type": "call"}, {"api_name": "os.path", "line_number": 474, "usage_type": "attribute"}, {"api_name": "torch.rand_like", "line_number": 483, "usage_type": "call"}, {"api_name": "torch.rand_like", "line_number": 485, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 487, "usage_type": "call"}, {"api_name": "torch.unsqueeze", "line_number": 488, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 493, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 513, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 513, "usage_type": "name"}, {"api_name": "torch.unsqueeze", "line_number": 533, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 537, "usage_type": "call"}]} +{"seq_id": "300529281", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on 2019/11/7 10:13 AM \n\n@author: HOY\n\"\"\"\nimport pickle\nimport jieba\nimport re\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nimport pandas as pd\n\ndef predictText(X_train):\n\n data = pd.read_csv('before.csv', encoding=\"utf-8\", header=None)\n data.columns = ['n_id', 'id', 'content', 'label']\n data = data.drop(index=[0])\n\n lines = open('stopwords-master/百度停用词表.txt', 'r', encoding='utf-8')\n stop_words = [line.strip() for line in lines]\n # print(stop_words)\n word_list = []\n words_list = []\n for sent in data['content']:\n try:\n words = jieba.cut(sent)\n words = [word for word in words if word not in stop_words]\n segmented_words = ','.join(words)\n except AttributeError:\n continue\n finally:\n words_list.append(words)\n word_list.append(segmented_words.strip())\n data['tokens'] = word_list\n\n def tfidf(data):\n tfidf_vectorizer = TfidfVectorizer()\n train = tfidf_vectorizer.fit_transform(data)\n return train, tfidf_vectorizer\n # 文本特征提取\n X_train_tfidf, tfidf_vectorizer = tfidf(X_train)\n X_test_tfidf = tfidf_vectorizer.transform(data['tokens'])\n\n tfidf_path = 'tfidf.pkl'\n with open(tfidf_path, 'rb') as out_data:\n clf_tfidf = pickle.load(out_data)\n y_predicted_tfidf = clf_tfidf.predict(X_test_tfidf)\n data['label'] = y_predicted_tfidf\n\n data.to_csv('after.csv', encoding=\"utf_8_sig\",columns=[ 'id', 'content', 'label'])\n", "sub_path": "predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 1584, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pandas.read_csv", "line_number": 16, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 27, "usage_type": "call"}, {"api_name": "sklearn.feature_extraction.text.TfidfVectorizer", "line_number": 38, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "422115322", "text": "from django.conf.urls import url\nfrom . import views\n\nurlpatterns = [\n url(r'^$', views.home), \n url(r'^register$', views.register),\n url(r'^login$', views.login),\n url(r'^logout$', views.logout), \n url(r'^friends$', views.friends),\n url(r'^add$', views.add),\n url(r'^remove$', views.remove),\n url(r'^user/1$', views.user),\n url(r'^otheruser/10$', views.otheruser),\n\n]", "sub_path": "apps/first_app/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.conf.urls.url", "line_number": 5, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 10, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "589582211", "text": "from qiskit import QuantumCircuit, execute, Aer, IBMQ, QuantumRegister, ClassicalRegister\nfrom qiskit.circuit import Qubit\nfrom qiskit.aqua import AquaError\nfrom qiskit.compiler import transpile, assemble\nfrom sympy.combinatorics.graycode import GrayCode\nfrom qiskit.circuit.library.standard_gates.multi_control_rotation_gates import _apply_mcu3_graycode, mcrx, mcrz\nimport numpy as np\nimport random\nimport math\n\ndef dec_to_bin(number, n): #função pra tranformar decimal em binário\n binary = bin(number)[2:].zfill(n)\n return binary\n\ndef find_position(base_state): #função para guardar as posições onde serão aplicadas as portas Z\n position = []\n for pos in range(len(base_state)):\n if base_state[pos] == '1':\n position.append(pos)\n return position\n\ndef atualizeVector(vector, n, positions): #função que atualiza o vetor com os novos valores de amplitudes depois da aplicação do HSGS \n for base_state in range(len(vector)):\n count = 0\n pos1_base_state = find_position(dec_to_bin(base_state, n))\n for pos in positions:\n if pos in pos1_base_state:\n count += 1\n if count == len(positions):\n vector[base_state] *= -1\n\ndef makeCZ(n, circuit, ctrls, qaux, qtarget): #função que aplica porta z multi-controlada entre 3 ou mais qubits\n m=0\n circuit.ccx(ctrls[0], ctrls[1], qaux[0])\n for m in range(2, len(ctrls)):\n circuit.ccx(ctrls[m], qaux[m-2], qaux[m-1])\n \n circuit.cz(qtarget, qaux[len(ctrls)-2])\n \n for m in range(len(ctrls)-1, 1, -1):\n circuit.ccx(ctrls[m], qaux[m-2], qaux[m-1])\n circuit.ccx(ctrls[0], ctrls[1], qaux[0])\n\n return circuit\n\ndef hsgsGenerator(inputVector, circuit, q_input, n, ancila=False):\n\t#inputVector is a Python list \n\t\t#eg. inputVector=[1, -1, 1, 1]\n\t#circuit is the circuit where the HSGS will be put in\n #q_input ????\n #nSize is the input size\n #ancila esse parametro define se usamos registradores auxiliares para aplicar a porta multi-controlada Z ou não\n \n\t## this functions returns the quantum circuit that generates the quantum state whose amplitudes values are the values of inputVector using hsgsGenerator approach.\n\n if ancila == True:\n if n == 1:\n q_aux = QuantumRegister(n+1, 'q_aux')\n elif n == 2:\n q_aux = QuantumRegister(n, 'q_aux')\n else:\n q_aux = QuantumRegister(n-1, 'q_aux')\n circuit.add_register(q_aux)\n\n\n if(inputVector[0] == -1):\n inputVector = [indice * -1 for indice in inputVector]\n else:\n inputVector = [indice for indice in inputVector]\n \n outputVector = []\n for i in range(len(inputVector)):\n outputVector.append(1)\n \n for p in range(1, n+1): \n #laço pra pegar os p bits setado com 1 no estado base\n for base_state in range(2**n): \n \n #percorre todos os estados da base\n positions = find_position(dec_to_bin(base_state,n)) \n \n #encontrando as posições no estado da base onde o bit é '1'\n if len(positions) == p:\n if outputVector[base_state] != inputVector[base_state]: \n \n #checa se o vetor final já está no estado desejado\n circuit.barrier()\n if len(positions) == 1: \n \n #aplicar porta z simples para os estados da base com 1 unico qubit '1'\n circuit.z(q_input[positions[0]])\n elif len(positions) == 2: \n \n #aplicar porta cz pas os estados da base com 2 qubits setados\n circuit.cz(q_input[positions[0]], q_input[positions[1]])\n else: \n \n #aplicar uma porta control Z para 3 ou mais qubits\n # setados com ajuda de qbits auxiliares\n ctrls = [q_input[pos] for pos in positions[:len(positions)-1]]\n target = q_input[positions[len(positions)-1]]\n if ancila == True:\n makeCZ(n, circuit, q_input, q_aux, target)\n else:\n mcrz(circuit, math.pi, ctrls, target)\n \n #atualiza vetor de saída\n atualizeVector(outputVector, n, positions)\n return circuit", "sub_path": "qneuronreal-master/hsgs.py", "file_name": "hsgs.py", "file_ext": "py", "file_size_in_byte": 4512, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "qiskit.QuantumRegister", "line_number": 58, "usage_type": "call"}, {"api_name": "qiskit.QuantumRegister", "line_number": 60, "usage_type": "call"}, {"api_name": "qiskit.QuantumRegister", "line_number": 62, "usage_type": "call"}, {"api_name": "qiskit.circuit.library.standard_gates.multi_control_rotation_gates.mcrz", "line_number": 105, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 105, "usage_type": "attribute"}]} +{"seq_id": "443263211", "text": "from multiprocessing import Pool\nimport ijson.backends.yajl2_cffi as ijson\nimport simplejson as json\nfrom MyDataBase import MyColumnDatabaseWBatch\nfrom Program import skinnyProgramInBatch\nimport pickle\nimport os\n\nbackupDB = '../log/bayouSearchColDb_backup.pkl'\n\nclass parallelReadJSON():\n\n\n def __init__(self, folder, numThreads, dimension=256, batch_size=50, minJSONs = 0, maxJSONs = 70):\n self.folder = folder\n self.minJSONs = minJSONs\n self.maxJSONs = maxJSONs\n self.dimension = dimension\n self.batch_size = batch_size\n self.numThreads = numThreads\n\n\n def getSearchDatabase(self):\n\n if not os.path.exists(backupDB):\n FinalProgram_DB = self.readAllJSONs()\n #with open(backupDB , 'wb') as output:\n # pickle.dump(FinalProgram_DB, output)\n else:\n with open(backupDB, 'rb') as input:\n FinalProgram_DB = pickle.load(input)\n\n return FinalProgram_DB\n\n\n\n def readAllJSONs(self):\n\n prefix = self.folder + 'Program_output_'\n files = [ prefix + str(j) + '.json' for j in range(self.minJSONs, self.maxJSONs)]\n\n\n numThreads = self.numThreads\n print(\"Start parallel multiprocessing read JSONs\")\n fileChunks = [files[i::numThreads] for i in range(numThreads)]\n pool = Pool(processes=numThreads)\n result = pool.map(self.readMultipleJSONs, fileChunks)\n\n # pool.close()\n # pool.join()\n print(\"Done with multi-multiprocessing read JSONs\")\n #Aggregate the result\n FinalProgram_DB = []\n for item in result:\n FinalProgram_DB.extend(item)\n\n return FinalProgram_DB\n\n\n\n def readMultipleJSONs(self, files):\n\n Program_DB_all=[]\n #print (\"Starting to read \" + str(len(files)) + \" JSON files\")\n for file in files:\n Program_DB_j = self.readEachJSON(file)\n Program_DB_all.append(Program_DB_j)\n\n\n #print (\"Completed reading \" + str(len(files)) + \" JSON files\")\n return Program_DB_all\n\n\n def readEachJSON(self, fileName):\n # print(\"Starting to read \" + fileName)\n\n with open( fileName , 'r') as f:\n js = json.load(f)\n numItems = len(js['programs'])\n\n Program_DB_j = MyColumnDatabaseWBatch( numItems , self.dimension, self.batch_size)\n\n for k, jsProgram in enumerate(js['programs']):\n decodedProgram = skinnyProgramInBatch(jsProgram, k, Program_DB_j, self.batch_size)\n Program_DB_j.setValues(jsProgram, decodedProgram, k)\n\n # print(\"Read \" + fileName)\n return Program_DB_j\n", "sub_path": "src/main/python/bayou/experiments/predictMethods/SearchDB/parallelReadJSON.py", "file_name": "parallelReadJSON.py", "file_ext": "py", "file_size_in_byte": 2649, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.exists", "line_number": 25, "usage_type": "call"}, {"api_name": "os.path", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 31, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 46, "usage_type": "call"}, {"api_name": "simplejson.load", "line_number": 78, "usage_type": "call"}, {"api_name": "MyDataBase.MyColumnDatabaseWBatch", "line_number": 81, "usage_type": "call"}, {"api_name": "Program.skinnyProgramInBatch", "line_number": 84, "usage_type": "call"}]} +{"seq_id": "509628315", "text": "import pygame\nimport random\nimport math\n# import fighting_game\n\n# Player = player()\n\nclass ZombieEnemy(pygame.sprite.Sprite):\n def __init__(self, x=300, y=450):\n pygame.sprite.Sprite.__init__(self)\n self.image = pygame.image.load('images/zombie.png')\n self.rect = self.image.get_rect()\n self.rect.center = (x, y)\n self.speedy = random.randrange(1, 8)\n def move_towards_player(self, player):\n # Find direction vector (dx, dy) between enemy and player.\n dx, dy = player.rect.x - self.rect.x, player.rect.y - self.rect.y\n dist = math.hypot (dx, dy)\n dx, dy = dx / dist, dy / dist # Normalize\n # Move along this normalized vector towards the player\n self.rect.x += dx * 10\n self.rect.y += dy * 0\n\n\n# all_zombies = pygame.sprite.Group()\n\n\n# for i in range( 50 ):\n# new_x = random.randrange( 0, 10000) # random x-position\n# # new_y = random.randrange( 0, ) # random y-position\n# all_zombies.add(ZombieEnemy(new_x)) # create, and add to group", "sub_path": "Zombie.py", "file_name": "Zombie.py", "file_ext": "py", "file_size_in_byte": 1056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.sprite", "line_number": 8, "usage_type": "attribute"}, {"api_name": "pygame.sprite.Sprite.__init__", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.sprite", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 11, "usage_type": "attribute"}, {"api_name": "random.randrange", "line_number": 14, "usage_type": "call"}, {"api_name": "math.hypot", "line_number": 18, "usage_type": "call"}]} +{"seq_id": "436162462", "text": "from telegram.ext import CallbackContext, CallbackQueryHandler\nfrom telegram import Update, InlineKeyboardButton, InlineKeyboardMarkup\n\nfrom .ban import not_banned_users\nfrom strings import get_string, get_languages, _lang\n\n\n@not_banned_users\ndef set_lang(update: Update, context: CallbackContext) -> None:\n lang = update.callback_query.data.split(\"_\")[-1]\n\n if update.effective_chat.id in context.bot_data:\n context.bot_data[update.effective_chat.id][\"lang\"] = lang\n else:\n context.bot_data[update.effective_chat.id] = dict(lang=lang)\n\n update.effective_message.delete()\n\n keyboard = []\n\n for language in get_languages():\n if language != context.bot_data.get(update.effective_chat.id, {}).get(\"lang\", \"en\"):\n keyboard.append(\n [\n InlineKeyboardButton(\n get_languages()[language],\n callback_data=\"setlang_\" + language\n )\n ]\n )\n\n context.bot.send_message(\n update.effective_chat.id,\n get_string(\n \"start\",\n _lang(context, update.effective_chat.id)\n ),\n reply_markup=InlineKeyboardMarkup(keyboard)\n if keyboard != [[]] else None\n )\n\n update.callback_query.answer()\n\n\n__handlers__ = [\n [\n CallbackQueryHandler(set_lang, pattern=\"setlang_.+\")\n ]\n]\n", "sub_path": "handlers/callback.py", "file_name": "callback.py", "file_ext": "py", "file_size_in_byte": 1398, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "telegram.Update", "line_number": 9, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackContext", "line_number": 9, "usage_type": "name"}, {"api_name": "strings.get_languages", "line_number": 21, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardButton", "line_number": 25, "usage_type": "call"}, {"api_name": "strings.get_languages", "line_number": 26, "usage_type": "call"}, {"api_name": "strings.get_string", "line_number": 34, "usage_type": "call"}, {"api_name": "strings._lang", "line_number": 36, "usage_type": "call"}, {"api_name": "telegram.InlineKeyboardMarkup", "line_number": 38, "usage_type": "call"}, {"api_name": "ban.not_banned_users", "line_number": 8, "usage_type": "name"}, {"api_name": "telegram.ext.CallbackQueryHandler", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "417676125", "text": "\"\"\"\nThis module contains KMS related class and methods\ncurrently supported KMSs: Vault\n\n\"\"\"\nimport logging\nimport os\n\nimport json\nimport shlex\nimport tempfile\nimport subprocess\nimport base64\n\nfrom ocs_ci.framework import config\nfrom ocs_ci.ocs import constants, ocp, defaults\nfrom ocs_ci.ocs.exceptions import (\n VaultDeploymentError,\n VaultOperationError,\n KMSNotSupported,\n KMSConnectionDetailsError,\n KMSTokenError,\n KMSResourceCleaneupError,\n)\nfrom ocs_ci.utility import templating\nfrom ocs_ci.utility.utils import (\n load_auth_config,\n run_cmd,\n get_vault_cli,\n get_running_cluster_id,\n get_default_if_keyval_empty,\n)\n\nlogger = logging.getLogger(__name__)\n\n\nclass KMS(object):\n \"\"\"\n This is base class for any KMS integration\n\n \"\"\"\n\n def __init__(self, provider=None):\n self.kms_provider = provider\n\n def deploy(self):\n raise NotImplementedError()\n\n\nclass Vault(KMS):\n \"\"\"\n A class which handles deployment and other\n configs related to vault\n\n \"\"\"\n\n def __init__(self):\n super().__init__(\"vault\")\n self.vault_server = None\n self.port = None\n self.cluster_id = None\n # Name of kubernetes resources\n # for ca_cert, client_cert, client_key\n self.ca_cert_name = None\n self.client_cert_name = None\n self.client_key_name = None\n self.vault_root_token = None\n self.vault_namespace = None\n self.vault_deploy_mode = config.ENV_DATA.get(\"vault_deploy_mode\")\n self.vault_backend_path = None\n # Base64 encoded (with padding) token\n self.vault_path_token = None\n self.vault_policy_name = None\n\n def deploy(self):\n \"\"\"\n This function delegates the deployment of vault\n based on OCP or vault standalone external mode deployment\n\n \"\"\"\n if self.vault_deploy_mode == \"external\":\n self.deploy_vault_external()\n elif self.vault_deploy_mode == \"internal\":\n self.deploy_vault_internal()\n else:\n raise VaultDeploymentError(\"Not a supported vault deployment mode\")\n\n def deploy_vault_external(self):\n \"\"\"\n This function takes care of deployment and configuration\n for external mode vault deployment. We are assuming that\n an external vault service already exists and we will be just\n configuring the necessary OCP objects for OCS like secrets, token etc\n\n \"\"\"\n self.gather_init_vault_conf()\n # Update env vars for vault CLI usage\n self.update_vault_env_vars()\n get_vault_cli()\n self.vault_unseal()\n self.vault_create_backend_path()\n self.create_ocs_vault_resources()\n\n def gather_init_vault_conf(self):\n \"\"\"\n Gather vault configuration and init the vars\n This function currently gathers only for external mode\n Gathering for internal mode woulde be different\n\n \"\"\"\n self.vault_conf = self.gather_vault_config()\n self.vault_server = self.vault_conf[\"VAULT_ADDR\"]\n self.port = self.vault_conf[\"PORT\"]\n if not config.ENV_DATA.get(\"VAULT_SKIP_VERIFY\"):\n self.ca_cert_base64 = self.vault_conf[\"VAULT_CACERT_BASE64\"]\n self.client_cert_base64 = self.vault_conf[\"VAULT_CLIENT_CERT_BASE64\"]\n self.client_key_base64 = self.vault_conf[\"VAULT_CLIENT_KEY_BASE64\"]\n self.vault_tls_server = self.vault_conf[\"VAULT_TLS_SERVER_NAME\"]\n self.vault_root_token = self.vault_conf[\"VAULT_ROOT_TOKEN\"]\n\n def update_vault_env_vars(self):\n \"\"\"\n In order to run vault CLI we need following env vars\n VAULT_ADDR and VAULT_TOKEN\n\n \"\"\"\n os.environ[\"VAULT_ADDR\"] = f\"https://{self.vault_server}:{self.port}\"\n os.environ[\"VAULT_TOKEN\"] = self.vault_root_token\n os.environ[\"VAULT_FORMAT\"] = \"json\"\n # setup client crt so that vault cli works smoothly\n # if 'VAULT_SKIP_VERIFY' is True then no need to do\n # this call as vault would have configured for http\n if (\n not config.ENV_DATA.get(\"VAULT_SKIP_VERIFY\")\n and config.ENV_DATA.get(\"vault_deploy_mode\") == \"external\"\n ):\n self.setup_vault_client_cert()\n os.environ[\"VAULT_CACERT\"] = constants.VAULT_CLIENT_CERT_PATH\n\n def setup_vault_client_cert(self):\n \"\"\"\n For Vault cli interaction with the server we need client cert\n to talk to HTTPS on the vault server\n\n \"\"\"\n cert_str = base64.b64decode(self.client_cert_base64).decode()\n with open(constants.VAULT_CLIENT_CERT_PATH, \"w\") as cert:\n cert.write(cert_str)\n logger.info(f\"Created cert file at {constants.VAULT_CLIENT_CERT_PATH}\")\n\n def create_ocs_vault_resources(self):\n \"\"\"\n This function takes care of creating ocp resources for\n secrets like ca cert, client cert, client key and vault token\n Assumption is vault section in AUTH file contains base64 encoded\n (with padding) ca, client certs, client key and vault path token\n\n \"\"\"\n if not config.ENV_DATA.get(\"VAULT_SKIP_VERIFY\"):\n # create ca cert secret\n ca_data = templating.load_yaml(constants.EXTERNAL_VAULT_CA_CERT)\n self.ca_cert_name = get_default_if_keyval_empty(\n config.ENV_DATA, \"VAULT_CACERT\", defaults.VAULT_DEFAULT_CA_CERT\n )\n ca_data[\"metadata\"][\"name\"] = self.ca_cert_name\n ca_data[\"data\"][\"cert\"] = self.ca_cert_base64\n self.create_resource(ca_data, prefix=\"ca\")\n\n # create client cert secret\n client_cert_data = templating.load_yaml(\n constants.EXTERNAL_VAULT_CLIENT_CERT\n )\n self.client_cert_name = get_default_if_keyval_empty(\n config.ENV_DATA, \"VAULT_CLIENT_CERT\", defaults.VAULT_DEFAULT_CLIENT_CERT\n )\n client_cert_data[\"metadata\"][\"name\"] = self.client_cert_name\n client_cert_data[\"data\"][\"cert\"] = self.client_cert_base64\n self.create_resource(client_cert_data, prefix=\"clientcert\")\n\n # create client key secert\n client_key_data = templating.load_yaml(constants.EXTERNAL_VAULT_CLIENT_KEY)\n self.client_key_name = get_default_if_keyval_empty(\n config.ENV_DATA, \"VAULT_CLIENT_KEY\", defaults.VAULT_DEFAULT_CLIENT_KEY\n )\n client_key_data[\"metadata\"][\"name\"] = self.client_key_name\n client_key_data[\"data\"][\"key\"] = self.client_key_base64\n self.create_resource(client_key_data, prefix=\"clientkey\")\n\n # create oc resource secret for token\n token_data = templating.load_yaml(constants.EXTERNAL_VAULT_KMS_TOKEN)\n # token has to base64 encoded (with padding)\n token_data[\"data\"][\"token\"] = base64.b64encode(\n # encode() because b64encode expects a byte type\n self.vault_path_token.encode()\n ).decode() # decode() because b64encode returns a byte type\n self.create_resource(token_data, prefix=\"token\")\n\n # create ocs-kms-connection-details\n connection_data = templating.load_yaml(\n constants.EXTERNAL_VAULT_KMS_CONNECTION_DETAILS\n )\n connection_data[\"data\"][\"VAULT_ADDR\"] = os.environ[\"VAULT_ADDR\"]\n connection_data[\"data\"][\"VAULT_BACKEND_PATH\"] = self.vault_backend_path\n connection_data[\"data\"][\"VAULT_CACERT\"] = self.ca_cert_name\n connection_data[\"data\"][\"VAULT_CLIENT_CERT\"] = self.client_cert_name\n connection_data[\"data\"][\"VAULT_CLIENT_KEY\"] = self.client_key_name\n self.vault_namespace = config.ENV_DATA.get(\n \"VAULT_NAMESPACE\", constants.VAULT_DEFAULT_NAMESPACE\n )\n connection_data[\"data\"][\"VAULT_NAMESPACE\"] = self.vault_namespace\n connection_data[\"data\"][\"VAULT_TLS_SERVER_NAME\"] = self.vault_tls_server\n self.create_resource(connection_data, prefix=\"kmsconnection\")\n\n def create_resource(self, resource_data, prefix=None):\n \"\"\"\n Given a dictionary of resource data, this function will\n creates oc resource\n\n Args:\n resource_data (dict): yaml dictionary for resource\n prefix (str): prefix for NamedTemporaryFile\n\n \"\"\"\n resource_data_yaml = tempfile.NamedTemporaryFile(\n mode=\"w+\", prefix=prefix, delete=False\n )\n templating.dump_data_to_temp_yaml(resource_data, resource_data_yaml.name)\n run_cmd(f\"oc create -f {resource_data_yaml.name}\", timeout=300)\n\n def vault_unseal(self):\n \"\"\"\n Unseal vault if sealed\n\n Raises:\n VaultOperationError: In case unseal operation failed\n\n \"\"\"\n if self.vault_sealed():\n logger.info(\"Vault is sealed, Unsealing now..\")\n for i in range(3):\n kkey = f\"UNSEAL_KEY{i+1}\"\n self._vault_unseal(self.vault_conf[kkey])\n # Check if vault is unsealed or not\n if self.vault_sealed():\n raise VaultOperationError(\"Failed to Unseal vault\")\n else:\n logger.info(\"Vault has been successfully unsealed\")\n else:\n logger.info(\"Vault is not sealed\")\n\n def _vault_unseal(self, key):\n \"\"\"\n Execute unseal command here\n\n Args:\n key (str): unseal key\n\n \"\"\"\n unseal_cmd = f\"vault operator unseal {key}\"\n subprocess.check_output(shlex.split(unseal_cmd))\n\n def vault_sealed(self):\n \"\"\"\n Returns:\n bool: if vault is sealed then return True else False\n\n \"\"\"\n status_cmd = \"vault status --format=json\"\n output = subprocess.check_output(shlex.split(status_cmd))\n outbuf = json.loads(output)\n return outbuf[\"sealed\"]\n\n def vault_create_backend_path(self):\n \"\"\"\n create vault path to be used by OCS\n\n Raises:\n VaultOperationError exception\n \"\"\"\n if config.ENV_DATA.get(\"VAULT_BACKEND_PATH\"):\n self.vault_backend_path = config.ENV_DATA.get(\"VAULT_BACKEND_PATH\")\n else:\n # Generate backend path name using prefix \"ocs\"\n # \"ocs-\"\n self.cluster_id = get_running_cluster_id()\n self.vault_backend_path = (\n f\"{constants.VAULT_DEFAULT_PATH_PREFIX}-{self.cluster_id}\"\n )\n cmd = f\"vault secrets enable -path={self.vault_backend_path} kv\"\n out = subprocess.check_output(shlex.split(cmd))\n if \"Success\" in out.decode():\n logger.info(f\"vault path {self.vault_backend_path} created\")\n else:\n raise VaultOperationError(\n f\"Failed to create path f{self.vault_backend_path}\"\n )\n self.vault_create_policy()\n\n def vault_create_policy(self):\n \"\"\"\n Create a vault policy and generate token\n\n Raises:\n VaultOperationError exception\n\n \"\"\"\n policy = (\n f'path \"{self.vault_backend_path}/*\" {{\\n'\n f' capabilities = [\"create\", \"read\", \"update\",\"delete\"]'\n f\"\\n}}\\n\"\n f'path \"sys/mounts\" {{\\n'\n f'capabilities = [\"read\"]\\n'\n f\"}}\"\n )\n vault_hcl = tempfile.NamedTemporaryFile(mode=\"w+\", prefix=\"test\", delete=False)\n with open(vault_hcl.name, \"w\") as hcl:\n hcl.write(policy)\n\n if not config.ENV_DATA.get(\"VAULT_POLICY\"):\n self.vault_policy_name = (\n f\"{constants.VAULT_DEFAULT_POLICY_PREFIX}-\" f\"{self.cluster_id}\"\n )\n else:\n self.vault_policy_name = config.ENV_DATA.get(\"VAULT_POLICY\")\n\n cmd = f\"vault policy write {self.vault_policy_name} {vault_hcl.name}\"\n out = subprocess.check_output(shlex.split(cmd))\n if \"Success\" in out.decode():\n logger.info(f\"vault policy {self.vault_policy_name} created\")\n else:\n raise VaultOperationError(\n f\"Failed to create policy f{self.vault_policy_name}\"\n )\n self.vault_path_token = self.generate_vault_token()\n\n def generate_vault_token(self):\n \"\"\"\n Generate a token for self.vault_policy_name\n\n Returns:\n str: vault token\n\n \"\"\"\n cmd = f\"vault token create -policy={self.vault_policy_name} \" f\"--format=json\"\n out = subprocess.check_output(shlex.split(cmd))\n json_out = json.loads(out)\n return json_out[\"auth\"][\"client_token\"]\n\n def deploy_vault_internal(self):\n \"\"\"\n This function takes care of deployment and configuration for\n internal mode vault deployment on OCP\n\n \"\"\"\n pass\n\n def gather_vault_config(self):\n \"\"\"\n This function populates the vault configuration\n\n \"\"\"\n if self.vault_deploy_mode == \"external\":\n vault_conf = load_auth_config()[\"vault\"]\n return vault_conf\n\n def get_vault_backend_path(self):\n \"\"\"\n Fetch the vault backend path used for this deployment\n This can be obtained from kubernetes secret resource\n 'ocs-kms-connection-details'\n\n .. code-block:: none\n\n apiVersion: v1\n data:\n KMS_PROVIDER: vault\n KMS_SERVICE_NAME: vault\n VAULT_ADDR: https://xx.xx.xx.xx:8200\n VAULT_BACKEND_PATH: ocs\n\n \"\"\"\n if not self.vault_backend_path:\n connection_details = ocp.OCP(\n kind=\"ConfigMap\",\n resource_name=constants.VAULT_KMS_CONNECTION_DETAILS_RESOURCE,\n namespace=constants.OPENSHIFT_STORAGE_NAMESPACE,\n )\n try:\n self.vault_backend_path = connection_details.get().get(\"data\")[\n \"VAULT_BACKEND_PATH\"\n ]\n except IndexError:\n raise KMSConnectionDetailsError(\"KMS connection details not available\")\n\n def get_vault_path_token(self):\n \"\"\"\n Fetch token from kubernetes secret\n we need this to find the vault policy\n default name in case of ocs is 'ocs-kms-token'\n\n .. code-block:: none\n\n apiVersion: v1\n data:\n token: cy5DRXBKV0lVbzNFQjM1VHlGMFNURzZQWms=\n kind: Secret\n metadata:\n name: ocs-kms-token\n namespace: openshift-storage\n type: Opaque\n\n \"\"\"\n if not self.vault_path_token:\n vault_token = ocp.OCP(\n kind=\"Secret\",\n resource_name=constants.VAULT_KMS_TOKEN_RESOURCE,\n namespace=constants.OPENSHIFT_STORAGE_NAMESPACE,\n )\n try:\n token = vault_token.get().get(\"data\")[\"token\"]\n self.vault_path_token = base64.b64decode(token).decode()\n except IndexError:\n raise KMSTokenError(\"Couldn't find KMS token\")\n\n def get_vault_policy(self):\n \"\"\"\n Get the policy name based on token from vault\n\n \"\"\"\n if not self.vault_policy_name:\n cmd = f\"vault token lookup {self.vault_path_token}\"\n out = subprocess.check_output(shlex.split(cmd))\n json_out = json.loads(out)\n for policy in json_out[\"data\"][\"policies\"]:\n if self.cluster_id in policy:\n self.vault_policy_name = policy\n\n def remove_vault_backend_path(self):\n \"\"\"\n remove vault path\n\n \"\"\"\n cmd = f\"vault secrets disable {self.vault_backend_path}\"\n subprocess.check_output(shlex.split(cmd))\n # Check if path doesn't appear in the list\n cmd = \"vault secrets list --format=json\"\n out = subprocess.check_output(shlex.split(cmd))\n json_out = json.loads(out)\n for path in json_out.keys():\n if self.vault_backend_path in path:\n raise KMSResourceCleaneupError(\n f\"Path {self.vault_backend_path} not deleted\"\n )\n logger.info(f\"Vault path {self.vault_backend_path} deleted\")\n\n def remove_vault_policy(self):\n \"\"\"\n Cleanup the policy we used\n\n \"\"\"\n cmd = f\"vault policy delete {self.vault_policy_name} \"\n subprocess.check_output(shlex.split(cmd))\n # Check if policy still exists\n cmd = \"vault policy list --format=json\"\n out = subprocess.check_output(shlex.split(cmd))\n json_out = json.loads(out)\n if self.vault_policy_name in json_out:\n raise KMSResourceCleaneupError(\n f\"Policy {self.vault_policy_name} not deleted\"\n )\n logger.info(f\"Vault policy {self.vault_policy_name} deleted\")\n\n def cleanup(self):\n \"\"\"\n Cleanup the backend resources in case of external\n\n \"\"\"\n if not self.vault_server:\n self.gather_init_vault_conf()\n\n self.update_vault_env_vars()\n # get vault path\n self.get_vault_backend_path()\n # from token secret get token\n self.get_vault_path_token()\n # from token get policy\n if not self.cluster_id:\n self.cluster_id = get_running_cluster_id()\n self.get_vault_policy()\n # Delete the policy and backend path from vault\n # we need root token of vault in the env\n self.remove_vault_backend_path()\n self.remove_vault_policy()\n\n\nkms_map = {\"vault\": Vault}\n\n\ndef get_kms_deployment():\n provider = config.ENV_DATA[\"KMS_PROVIDER\"]\n try:\n return kms_map[provider]()\n except KeyError:\n raise KMSNotSupported(\"Not a supported KMS deployment\")\n", "sub_path": "ocs_ci/utility/kms.py", "file_name": "kms.py", "file_ext": "py", "file_size_in_byte": 17706, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 34, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 69, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 69, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 69, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.exceptions.VaultDeploymentError", "line_number": 86, "usage_type": "call"}, {"api_name": "ocs_ci.utility.utils.get_vault_cli", "line_number": 99, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 114, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 114, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 114, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 128, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 129, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 134, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 134, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 134, "usage_type": "name"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 135, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 135, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 135, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 138, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants.VAULT_CLIENT_CERT_PATH", "line_number": 138, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 138, "usage_type": "name"}, {"api_name": "base64.b64decode", "line_number": 146, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.constants.VAULT_CLIENT_CERT_PATH", "line_number": 147, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 147, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.VAULT_CLIENT_CERT_PATH", "line_number": 149, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 149, "usage_type": "name"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 159, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 159, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 159, "usage_type": "name"}, {"api_name": "ocs_ci.utility.templating.load_yaml", "line_number": 161, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 161, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.EXTERNAL_VAULT_CA_CERT", "line_number": 161, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 161, "usage_type": "name"}, {"api_name": "ocs_ci.utility.utils.get_default_if_keyval_empty", "line_number": 162, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 163, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 163, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.defaults.VAULT_DEFAULT_CA_CERT", "line_number": 163, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.defaults", "line_number": 163, "usage_type": "name"}, {"api_name": "ocs_ci.utility.templating.load_yaml", "line_number": 170, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 170, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.EXTERNAL_VAULT_CLIENT_CERT", "line_number": 171, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 171, "usage_type": "name"}, {"api_name": "ocs_ci.utility.utils.get_default_if_keyval_empty", "line_number": 173, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 174, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 174, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.defaults.VAULT_DEFAULT_CLIENT_CERT", "line_number": 174, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.defaults", "line_number": 174, "usage_type": "name"}, {"api_name": "ocs_ci.utility.templating.load_yaml", "line_number": 181, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 181, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.EXTERNAL_VAULT_CLIENT_KEY", "line_number": 181, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 181, "usage_type": "name"}, {"api_name": "ocs_ci.utility.utils.get_default_if_keyval_empty", "line_number": 182, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 183, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 183, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.defaults.VAULT_DEFAULT_CLIENT_KEY", "line_number": 183, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.defaults", "line_number": 183, "usage_type": "name"}, {"api_name": "ocs_ci.utility.templating.load_yaml", "line_number": 190, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 190, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.EXTERNAL_VAULT_KMS_TOKEN", "line_number": 190, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 190, "usage_type": "name"}, {"api_name": "base64.b64encode", "line_number": 192, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating.load_yaml", "line_number": 199, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 199, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.EXTERNAL_VAULT_KMS_CONNECTION_DETAILS", "line_number": 200, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 200, "usage_type": "name"}, {"api_name": "os.environ", "line_number": 202, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 207, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 207, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 207, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.VAULT_DEFAULT_NAMESPACE", "line_number": 208, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 208, "usage_type": "name"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 224, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating.dump_data_to_temp_yaml", "line_number": 227, "usage_type": "call"}, {"api_name": "ocs_ci.utility.templating", "line_number": 227, "usage_type": "name"}, {"api_name": "ocs_ci.utility.utils.run_cmd", "line_number": 228, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.VaultOperationError", "line_number": 245, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 260, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 260, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 269, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 269, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 270, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 280, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 280, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 280, "usage_type": "name"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 281, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 281, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 281, "usage_type": "name"}, {"api_name": "ocs_ci.utility.utils.get_running_cluster_id", "line_number": 285, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.constants.VAULT_DEFAULT_PATH_PREFIX", "line_number": 287, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 287, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 290, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 290, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.VaultOperationError", "line_number": 294, "usage_type": "call"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 315, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 319, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 319, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 319, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.VAULT_DEFAULT_POLICY_PREFIX", "line_number": 321, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 321, "usage_type": "name"}, {"api_name": "ocs_ci.framework.config.ENV_DATA.get", "line_number": 324, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 324, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 324, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 327, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 327, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.VaultOperationError", "line_number": 331, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 345, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 345, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 346, "usage_type": "call"}, {"api_name": "ocs_ci.utility.utils.load_auth_config", "line_number": 363, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.ocp.OCP", "line_number": 383, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.ocp", "line_number": 383, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.VAULT_KMS_CONNECTION_DETAILS_RESOURCE", "line_number": 385, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 385, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.OPENSHIFT_STORAGE_NAMESPACE", "line_number": 386, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 386, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.exceptions.KMSConnectionDetailsError", "line_number": 393, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.ocp.OCP", "line_number": 414, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.ocp", "line_number": 414, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.VAULT_KMS_TOKEN_RESOURCE", "line_number": 416, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 416, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.constants.OPENSHIFT_STORAGE_NAMESPACE", "line_number": 417, "usage_type": "attribute"}, {"api_name": "ocs_ci.ocs.constants", "line_number": 417, "usage_type": "name"}, {"api_name": "base64.b64decode", "line_number": 421, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.KMSTokenError", "line_number": 423, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 432, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 432, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 433, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 444, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 444, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 447, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 447, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 448, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.KMSResourceCleaneupError", "line_number": 451, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 462, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 462, "usage_type": "call"}, {"api_name": "subprocess.check_output", "line_number": 465, "usage_type": "call"}, {"api_name": "shlex.split", "line_number": 465, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 466, "usage_type": "call"}, {"api_name": "ocs_ci.ocs.exceptions.KMSResourceCleaneupError", "line_number": 468, "usage_type": "call"}, {"api_name": "ocs_ci.utility.utils.get_running_cluster_id", "line_number": 488, "usage_type": "call"}, {"api_name": "ocs_ci.framework.config.ENV_DATA", "line_number": 500, "usage_type": "attribute"}, {"api_name": "ocs_ci.framework.config", "line_number": 500, "usage_type": "name"}, {"api_name": "ocs_ci.ocs.exceptions.KMSNotSupported", "line_number": 504, "usage_type": "call"}]} +{"seq_id": "380732777", "text": "\"\"\" messageme.core.api.messages\n\n This module implements the core API for working with messages. See the\n README.markdown file for documentation on the various functions defined\n herein.\n\"\"\"\nimport datetime\nimport logging\nimport traceback\n\nfrom django.http import HttpResponse\n\nfrom messageme.core.models import *\nfrom messageme.core.lib import sessionHandler\nfrom messageme.core.lib import messageHandler\nfrom messageme.core.lib import rateLimiter\nfrom messageme.core.api.exceptions import *\nfrom messageme.core.gateways import pubnub_gateway\nfrom messageme.core.gateways import twilio_gateway\n\n#############################################################################\n\nlogger = logging.getLogger(__name__)\n\n#############################################################################\n\ndef send(session, topic_id, sender_name=None, message=None):\n \"\"\" Send a message.\n \"\"\"\n raise UnauthorizedException() # Disable for now.\n\n logger.debug(\"core.api.messages.send(\" +\n 'session=%s, topic_id=%s, sender_name=%s, message=%s)'\n % (repr(session), repr(topic_id), repr(sender_name),\n repr(message)))\n\n try:\n if session == None or topic_id == None or message == None:\n raise InvalidParametersException()\n\n sessionHandler.validate(session)\n\n sender = sessionHandler.get_user(session)\n\n try:\n topic = Topic.objects.get(id=topic_id)\n except Topic.DoesNotExist:\n raise NoSuchTopicException()\n\n if not topic.active:\n raise InactiveTopicException()\n\n recipient = topic.user\n phone_number = recipient.phone_number\n\n rateLimiter.ensure_phone_number_below_rate_limit(phone_number)\n\n if messageHandler.handle_special_message(sender, recipient, topic,\n message):\n msg = None\n else:\n msg = messageHandler.send_message(sender, recipient, topic,\n sender_name, message)\n rateLimiter.sms_message_sent_to(phone_number)\n\n if msg != None:\n return msg.to_dict()\n else:\n return None\n except RateLimitReachedException:\n raise\n except:\n traceback.print_exc()\n raise\n\n#############################################################################\n\ndef receive(**kwargs):\n \"\"\" Receive an incoming SMS reply from Twilio.\n \"\"\"\n raise UnauthorizedException() # Disable for now.\n\n # Extract the parameters we are interested in.\n\n From = kwargs.get(\"From\")\n To = kwargs.get(\"To\")\n Body = kwargs.get(\"Body\")\n\n # Ask the Twilio gateway to process the incoming SMS.\n\n sms_reply = twilio_gateway.receive_sms(From, To, Body)\n\n # Create the XML-formatted response to send back to Twilio.\n\n response = []\n response.append('')\n response.append('')\n if sms_reply != None:\n response.append('')\n response.append(sms_reply)\n response.append('')\n response.append('')\n\n # Finally, wrap the response in an HttpResponse so the appropriate MIME\n # type, etc, will be returned to Twilio.\n\n return HttpResponse(\"\\n\".join(response), mimetype=\"text/xml\", status=201)\n\n", "sub_path": "messageme/core/api/messages.py", "file_name": "messages.py", "file_ext": "py", "file_size_in_byte": 3386, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "messageme.core.lib.sessionHandler.validate", "line_number": 41, "usage_type": "call"}, {"api_name": "messageme.core.lib.sessionHandler", "line_number": 41, "usage_type": "name"}, {"api_name": "messageme.core.lib.sessionHandler.get_user", "line_number": 43, "usage_type": "call"}, {"api_name": "messageme.core.lib.sessionHandler", "line_number": 43, "usage_type": "name"}, {"api_name": "messageme.core.lib.rateLimiter.ensure_phone_number_below_rate_limit", "line_number": 56, "usage_type": "call"}, {"api_name": "messageme.core.lib.rateLimiter", "line_number": 56, "usage_type": "name"}, {"api_name": "messageme.core.lib.messageHandler.handle_special_message", "line_number": 58, "usage_type": "call"}, {"api_name": "messageme.core.lib.messageHandler", "line_number": 58, "usage_type": "name"}, {"api_name": "messageme.core.lib.messageHandler.send_message", "line_number": 62, "usage_type": "call"}, {"api_name": "messageme.core.lib.messageHandler", "line_number": 62, "usage_type": "name"}, {"api_name": "messageme.core.lib.rateLimiter.sms_message_sent_to", "line_number": 64, "usage_type": "call"}, {"api_name": "messageme.core.lib.rateLimiter", "line_number": 64, "usage_type": "name"}, {"api_name": "traceback.print_exc", "line_number": 73, "usage_type": "call"}, {"api_name": "messageme.core.gateways.twilio_gateway.receive_sms", "line_number": 91, "usage_type": "call"}, {"api_name": "messageme.core.gateways.twilio_gateway", "line_number": 91, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 107, "usage_type": "call"}]} +{"seq_id": "349463172", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib\n\nPI = 4*np.arctan(1.)\n\nmatplotlib.rcParams.update({'font.size': 14})\n\nlatt, vit = np.loadtxt(\"liste_points\", unpack=True, skiprows=1)\n\nrayon = 0.6957e6\n\n#vit *= (rayon/3600)*PI/180\n\nvit *= (PI/180)*(rayon/3600)*(np.cos(latt*PI/180))\nprint(np.abs(latt))\n\nplt.plot(latt, vit, \"o\")\nplt.xlabel(\"Lattitude (°)\")\nplt.ylabel(\"Vitesse de rotation (km/s)\")\n\nplt.legend()\n\nplt.savefig(\"images/vit_rot.png\")\nplt.show()\n\n\n", "sub_path": "rapport/plot_rot.py", "file_name": "plot_rot.py", "file_ext": "py", "file_size_in_byte": 480, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.arctan", "line_number": 5, "usage_type": "call"}, {"api_name": "matplotlib.rcParams.update", "line_number": 7, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 7, "usage_type": "attribute"}, {"api_name": "numpy.loadtxt", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 18, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 19, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 19, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "476060292", "text": "#!/usr/bin/env python\n\n\"\"\"\nThis script takes the four final state hadron scores and retrieve\nthe data's centroid, principal components (PC), and points along PCs.\n\"\"\"\n\nfrom __future__ import print_function\n\nfrom sklearn.decomposition import PCA\nimport numpy as np\nimport pandas as pd\n\ncol_names = [\n 'run',\n 'subrun',\n 'cycle',\n 'evt',\n 'subevt',\n 'isnumucc',\n 'trueenu',\n 'trueemu',\n 'recoemu',\n 'recotrklenact',\n 'recotrklencat',\n 'mustopz',\n 'trueehad',\n 'recoehad',\n 'calehad',\n 'recoq2',\n 'npng',\n 'remidtrkismuon',\n 'cvnelectron',\n 'cvnmuon',\n 'cvnpi0',\n 'cvnchargedpion',\n 'cvnneutron',\n 'cvnproton',\n 'weight',\n 'intmode'\n]\ndf = pd.read_csv('../make_training_and_test_datasets/grid_output/nd_p6_standard_numucc_selection_stride1_offset0.1_of_200.txt',\n delimiter='\\s+',\n names=col_names)\n\nX = df.as_matrix(['cvnpi0','cvnchargedpion','cvnneutron','cvnproton'])\npca = PCA()\npca.fit(X)\nprint(pca.explained_variance_ratio_)\nprint(pca.singular_values_)\nprint(pca.mean_)\nprint(pca.components_)\nprint(pca.components_[0])\nprint(pca.explained_variance_)\n", "sub_path": "energy_regression/nd/rhc/pca/cvn_fs_score.py", "file_name": "cvn_fs_score.py", "file_ext": "py", "file_size_in_byte": 1111, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 42, "usage_type": "call"}, {"api_name": "sklearn.decomposition.PCA", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "503160156", "text": "\"\"\"新浪微博\"\"\"\nimport re\nimport json\nimport time\nimport requests\nfrom urltomid import *\nfrom pyquery import PyQuery as pq\nfrom Capmodule.Capture_WB_Hot import Cap_wbhotnews as cw\n\nclass Wb_Comment:\n def __init__(self,url,id,ti):\n self.hot_list = []\n self.Id = id\n self.url = url\n self.taskid = ti\n self.count = 0\n self.title = ''\n self.articletime = ''\n self.agree = ''\n self.headers = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.67 Safari/537.36',\n 'Cookie': 'M_WEIBOCN_PARAMS=uicode%3D20000174;SUB=_2A25xey4sDeRhGeBL71YU9ynKyjqIHXVSh7JkrDV6PUJbkdANLUPTkW1NRxrqa38ATTDrDDSkWkA5QE9_1GCT8-ip;MLOGIN=1;SCF=AsMQYFmjzrUPq0vWH9-kdzostEzXedQiGDAyNg2vbNoeepPKo8EBDbKY3Ig2fk8YXo1W-BLYkxzrHel8Y4QimHk.;SUHB=0IIVOLvSh6Svuh;SSOLoginState=1551851132;WEIBOCN_FROM=1110006030;_T_WM=6ca1ad1a9e8e681c1bb7a6e8e62705b0;XSRF-TOKEN=e4acfc'\n }\n p1 = re.compile(r'.*/(.*?)[?]', re.S)\n cut = re.findall(p1, url)\n mid = url_to_mid(cut[0])\n self.mid = mid\n # self.pageurl = 'https://weibo.com/aj/v6/comment/big?ajwvr=6&id={}&filter=all&page=1'.format(self.mid)\n self.pageurl = 'https://weibo.com/aj/v6/comment/big?id={}&from=singleWeiBo&__rnd=1545206992544'.format(self.mid)\n def get_frist(self):\n response = requests.get(self.pageurl, headers=self.headers).text\n htmljson = json.loads(response)\n html = htmljson['data']['html'] # 获取网页的代码元素\n doc = pq(html)\n items = doc('div.list_box > div.list_ul > div').items()\n for item in items:\n ID = item.attr('comment_id') # ID\n if self.Id == ID:\n self.agree = item('div.list_con > div.WB_func.clearfix > div.WB_handle.W_fr > ul > li:nth-child(4) > span > a > span > em:nth-child(2)').text()\n if self.agree == '赞':\n self.agree = 0\n self.hot_list.append(ID)\n def get_info(self):\n resp = requests.get(self.url,headers=self.headers).text\n p1 = re.compile(r'.*', re.S)\n cuttitle = re.findall(p1, resp)\n self.title = cuttitle[0] #个人微博正文\n mat = re.search(r\"(\\d{4}-\\d{1,2}-\\d{1,2}\\s\\d{1,2}:\\d{1,2})\", resp)\n self.articletime = mat.group(0) #微博发布时间\n def main(self):\n self.get_frist()\n self.get_info()\n time.sleep(5)\n if len(self.hot_list) > 10:\n if self.Id in self.hot_list[0:10]:\n print('%s已更新至热评,正在截图' % self.taskid)\n src = cw(self.url, self.Id, self.taskid, self.hot_list, self.title, self.articletime, self.agree).capture()\n return src\n else:\n self.count += 1\n print('%s已监控%d次' % (self.taskid, self.count))\n time.sleep(60)\n self.main()\n else:\n if self.Id in self.hot_list:\n print('%s已更新至热评,正在截图' % self.taskid)\n src = cw(self.url, self.Id, self.taskid, self.hot_list, self.title, self.articletime, self.agree).capture()\n return src\n else:\n self.count += 1\n print('%s已监控%d次' % (self.taskid, self.count))\n time.sleep(60)\n self.main()", "sub_path": "daokong1.01/Capture1.01/Comtmodule/Wb_Comment.py", "file_name": "Wb_Comment.py", "file_ext": "py", "file_size_in_byte": 3476, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "re.compile", "line_number": 24, "usage_type": "call"}, {"api_name": "re.S", "line_number": 24, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 25, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 31, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 32, "usage_type": "call"}, {"api_name": "pyquery.PyQuery", "line_number": 34, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 44, "usage_type": "call"}, {"api_name": "re.compile", "line_number": 45, "usage_type": "call"}, {"api_name": "re.S", "line_number": 45, "usage_type": "attribute"}, {"api_name": "re.findall", "line_number": 46, "usage_type": "call"}, {"api_name": "re.search", "line_number": 48, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 53, "usage_type": "call"}, {"api_name": "Capmodule.Capture_WB_Hot.Cap_wbhotnews", "line_number": 57, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "Capmodule.Capture_WB_Hot.Cap_wbhotnews", "line_number": 67, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 72, "usage_type": "call"}]} +{"seq_id": "175308386", "text": "import numpy as np \r\nimport pyglet \r\nfrom points import x_points, y_points\r\n \r\n \r\n \r\ndef circle_points(center, radius, resolution = 36):\r\n\r\n angle_inc = 2 * np.pi / resolution\r\n\r\n angles = (angle_inc * i for i in range(resolution))\r\n\r\n points = tuple((center[0] + np.cos(angle) * radius, center[1] + np.sin(angle) * radius) for angle in angles)\r\n\r\n points = tuple(element for tupl in points for element in tupl)\r\n return points\r\n\r\n \r\n\r\nWIDTH = 640\r\nHEIGHT = 480\r\nX_OFFSET = 0\r\nY_OFFSET = 0\r\n\r\n\r\n\r\nclass Main(pyglet.window.Window):\r\n def __init__ (self, width, height, x_points, y_points, x_circles, y_circles, orig):\r\n super().__init__(width, height)\r\n self.x_points = x_points\r\n self.y_points = y_points\r\n self.x_circles = x_circles\r\n self.y_circles = y_circles\r\n self.show_circles = True\r\n self.ends = []\r\n self.point_colors = []\r\n self.t = 0\r\n self.circle_colors = (150,50,150)\r\n self.lever_colors = (0, 150, 0)\r\n self.batch = pyglet.graphics.Batch()\r\n self.orig = orig\r\n \r\n def update(self, dt):\r\n self.t += dt\r\n\r\n def on_draw(self):\r\n\r\n self.clear()\r\n freq_scale = .1 #Scales how fast the circles spin and draw\r\n \r\n orig = self.orig\r\n \r\n for circle in self.y_circles:\r\n orig = self.spinning_circle(orig, circle[0], circle[2] * freq_scale, circle[1] + np.pi/2, False)\r\n \r\n for circle in self.x_circles:\r\n orig = self.spinning_circle(orig, circle[0], circle[2] * freq_scale, circle[1], True)\r\n \r\n self.ends.append(orig[0])\r\n self.ends.append(orig[1])\r\n \r\n# self.point_colors.append(int(np.cos(orig[0]/200 + orig[1]/200) * 255))\r\n# self.point_colors.append(int(np.sin(orig[0]/150 - orig[1]/150) * 255))\r\n# self.point_colors.append(len(self.ends) % 255)\r\n \r\n pyglet.graphics.draw(len(self.ends) // 2, pyglet.gl.GL_LINE_STRIP, \r\n ('v2f', self.ends),\r\n ('c3B', (255,255,255) * (len(self.ends)//2)))\r\n \r\n def spinning_circle(self, orig, length, freq, starting_angle, width):\r\n if width:\r\n mid = (orig[0] + (length/2) * np.cos(np.pi * freq * self.t + starting_angle), orig[1] + (length/2) * np.sin(np.pi * freq * self.t + starting_angle))\r\n end = (mid[0] + mid[0] - orig[0], orig[1])\r\n else:\r\n mid = (orig[0] + (length/2) * np.cos(np.pi * freq * self.t + starting_angle), orig[1] + (length/2) * np.sin(np.pi * freq * self.t + starting_angle))\r\n end = (orig[0], mid[1] + mid[1] - orig[1])\r\n \r\n resolution = 32\r\n angle_inc = np.pi * 2 / resolution\r\n \r\n if self.show_circles:\r\n \r\n pyglet.graphics.draw(2, pyglet.gl.GL_LINES, ('v2f', (orig[0], orig[1], mid[0], mid[1])),\r\n ('c3B', self.lever_colors * 2))\r\n \r\n pyglet.graphics.draw(2, pyglet.gl.GL_LINES, ('v2f', (mid[0], mid[1], end[0], end[1])),\r\n ('c3B', self.lever_colors * 2))\r\n \r\n circle_points1 = [(orig[0] + (length / 2) * np.cos(angle_inc * i), orig[1] + (length / 2) * np.sin(angle_inc * i)) for i in range(resolution)]\r\n circle_points1 = tuple(element for tupl in circle_points1 for element in tupl)\r\n circle_points1 = circle_points1 + (circle_points1[0], circle_points1[1])\r\n pyglet.graphics.draw(resolution + 1, pyglet.gl.GL_LINE_STRIP, \r\n ('v2f', circle_points1), ('c3B', self.circle_colors*(resolution + 1)))\r\n \r\n circle_points2 = [(mid[0] + (length / 2) * np.cos(angle_inc * i), mid[1] + (length / 2) * np.sin(angle_inc * i)) for i in range(resolution)]\r\n circle_points2 = tuple(element for tupl in circle_points2 for element in tupl)\r\n circle_points2 = circle_points2 + (circle_points2[0], circle_points2[1])\r\n pyglet.graphics.draw(resolution + 1, pyglet.gl.GL_LINE_STRIP, \r\n ('v2f', circle_points2), ('c3B', self.circle_colors*(resolution+1)))\r\n \r\n return end\r\n \r\n \r\n \r\n def on_key_press(self, symbol, modifiers):\r\n if symbol == pyglet.window.key.S:\r\n self.show_circles = not self.show_circles\r\n\r\ndef fourier_circles_points(points, T, N, tolerance=1):\r\n f_sample = 2 * N\r\n t, dt = np.linspace(0, T, f_sample + 2, endpoint=False, retstep=True)\r\n y = np.fft.rfft(points) / t.size\r\n y *= 2\r\n \r\n offset = y[0].real\r\n radii = abs(y[1:-1])\r\n freqs = np.arange(1, len(y) - 1) #TODO: Fix this to handle actual scale of frequencies\r\n angles = np.arctan2(y[1:-1].imag, y[1:-1].real)\r\n \r\n select = np.where(radii > tolerance)\r\n \r\n return offset, radii[select], angles[select], freqs[select]\r\n\r\n\r\nT = 1\r\nx_circles = []\r\nN = len(x_points)//2 - 2\r\nx_offset, x_radii, x_angles, x_freqs = fourier_circles_points(x_points, T, N)\r\nfor xr, xa, xf in zip(x_radii, x_angles, x_freqs):\r\n x_circles.append([xr, xa, xf])\r\n\r\ny_circles = []\r\ny_offset, y_radii, y_angles, y_freqs = fourier_circles_points(y_points, T, N)\r\nfor yr, ya, yf in zip(y_radii, y_angles, y_freqs):\r\n y_circles.append([yr, ya, yf])\r\n\r\norig = (int(np.min(x_points) + np.max(x_points))//2, int(np.min(y_points) + np.max(y_points))//2)\r\n\r\nif __name__ == '__main__':\r\n game_window = Main(WIDTH, HEIGHT, x_points, y_points, x_circles, y_circles, orig)\r\n pyglet.clock.schedule_interval(game_window.update, 1/60)\r\n\r\n pyglet.app.run()\r\n", "sub_path": "Fourier Draw/fourier.py", "file_name": "fourier.py", "file_ext": "py", "file_size_in_byte": 5619, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.pi", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 13, "usage_type": "call"}, {"api_name": "pyglet.window", "line_number": 27, "usage_type": "attribute"}, {"api_name": "points.x_points", "line_number": 30, "usage_type": "name"}, {"api_name": "points.y_points", "line_number": 31, "usage_type": "name"}, {"api_name": "pyglet.graphics.Batch", "line_number": 40, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.pi", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.draw", "line_number": 66, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 66, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 72, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 75, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.draw", "line_number": 83, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyglet.graphics.draw", "line_number": 86, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 86, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 89, "usage_type": "call"}, {"api_name": "pyglet.graphics.draw", "line_number": 92, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 92, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 95, "usage_type": "call"}, {"api_name": "pyglet.graphics.draw", "line_number": 98, "usage_type": "call"}, {"api_name": "pyglet.graphics", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pyglet.gl", "line_number": 98, "usage_type": "attribute"}, {"api_name": "pyglet.window", "line_number": 106, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.fft.rfft", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 112, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.arctan2", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 120, "usage_type": "call"}, {"api_name": "points.x_points", "line_number": 127, "usage_type": "argument"}, {"api_name": "points.x_points", "line_number": 128, "usage_type": "argument"}, {"api_name": "points.y_points", "line_number": 133, "usage_type": "argument"}, {"api_name": "numpy.min", "line_number": 137, "usage_type": "call"}, {"api_name": "points.x_points", "line_number": 137, "usage_type": "argument"}, {"api_name": "numpy.max", "line_number": 137, "usage_type": "call"}, {"api_name": "points.y_points", "line_number": 137, "usage_type": "argument"}, {"api_name": "points.x_points", "line_number": 140, "usage_type": "argument"}, {"api_name": "points.y_points", "line_number": 140, "usage_type": "argument"}, {"api_name": "pyglet.clock.schedule_interval", "line_number": 141, "usage_type": "call"}, {"api_name": "pyglet.clock", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pyglet.app.run", "line_number": 143, "usage_type": "call"}, {"api_name": "pyglet.app", "line_number": 143, "usage_type": "attribute"}]} +{"seq_id": "452491593", "text": "import numpy as np\nfrom skimage.measure import points_in_poly\nimport pandas as pd\n\nfrom ._base import GeneAssignmentAlgorithm\n\n\nclass PointInPoly(GeneAssignmentAlgorithm):\n def __init__(self, verbose=False, **kwargs):\n self.verbose = verbose\n\n @classmethod\n def add_arguments(cls, parser):\n pass\n\n @staticmethod\n def _assign(cells_region, spots, use_hull=True, verbose=False):\n res = pd.DataFrame({'spot_id': range(0, spots.shape[0])})\n res['cell_id'] = None\n\n for cell_id in range(cells_region.count):\n if use_hull:\n verts = cells_region[cell_id].hull\n else:\n verts = cells_region[cell_id].coordinates\n verts = np.array(verts)\n in_poly = points_in_poly(spots, verts)\n res.loc[res.spot_id[in_poly], 'cell_id'] = cell_id\n if verbose:\n cnt = np.sum(in_poly)\n print(cell_id, cnt)\n\n return res\n\n def assign_genes(self, intensity_table, regions):\n\n x = intensity_table.coords['features'].x.values\n y = intensity_table.coords['features'].y.values\n points = pd.DataFrame(dict(x=x, y=y))\n return self._assign(regions, points, use_hull=True, verbose=self.verbose)\n", "sub_path": "starfish/pipeline/gene_assignment/point_in_poly.py", "file_name": "point_in_poly.py", "file_ext": "py", "file_size_in_byte": 1269, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "_base.GeneAssignmentAlgorithm", "line_number": 8, "usage_type": "name"}, {"api_name": "pandas.DataFrame", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 26, "usage_type": "call"}, {"api_name": "skimage.measure.points_in_poly", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 30, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 39, "usage_type": "call"}]} +{"seq_id": "370998008", "text": "#!/usr/bin/python\r\n# # -*- coding: UTF-8 -*-\r\nimport cv2\r\nimport numpy as np\r\n#import matplotlib.pyplot as plt\r\n#检测图片边缘、轮廓、分割...\r\n\r\n\r\nclass check(object):\r\n\t\"\"\"docstring for check\"\"\"\r\n\tdef __init__(self,img):\r\n\t\t#super(check, self).__init__()\r\n\t\tself.img = img\r\n\t\tself.result = ''\r\n\t\tself.ims = ''\r\n\t#轮廓检测\t\r\n\tdef rough(self,limit):\r\n\t\tret, thresh = cv2.threshold(self.img, 127, 255, 0)\r\n\r\n\t\timage, contours = cv2.findContours(thresh, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)\r\n\r\n\t\tfor c in image:\r\n\t\t\tx, y, w, h = cv2.boundingRect(c)\r\n\t\t\tif w < limit:\r\n\t\t\t\tcontinue\r\n\t\t\telse:\r\n\t\t\t\tcv2.rectangle(self.img, (x, y), (x + w, y + h), (0, 255, 5), 2)\r\n\t\t\t\tself.ims = self.img[x:x+w,y:y+h]\r\n\t\t\t\trect = cv2.minAreaRect(c)\r\n\t\t\t\tbox = cv2.boxPoints(rect) \r\n\t\t\t\tbox = np.int0(box)\r\n\t\t\t\tcv2.drawContours(self.img, [box], 0, (253, 10, 253), 2)\r\n\r\n\tdef grabCut(self):\r\n\t\tmask = np.zeros(self.img.shape[:2],np.uint8)\r\n\t\tbgdModel = np.zeros((1,65),np.float64)\r\n\t\tfgdModel = np.zeros((1,65),np.float64)\r\n\t\t#限定分割图像的范围\r\n\t\trect = (10,10,500,580) \r\n\t\tcv2.grabCut(self.img,mask,rect,bgdModel,fgdModel,5,cv2.GC_INIT_WITH_RECT)\r\n\t\tmask2 = np.where((mask==2)|(mask==0),0,1).astype('uint8')\r\n\t\tself.img = self.img*mask2[:,:,np.newaxis] \r\n\r\n\tdef watershed(self):\r\n\t\tgray = cv2.cvtColor(self.img,cv2.COLOR_BGR2GRAY)\r\n\t\tret,thresh = cv2.threshold(gray,0,255,cv2.THRESH_BINARY_INV+cv2.THRESH_OTSU)\r\n\t\tkernel = np.ones((3,3),np.uint8)\r\n\t\t#变换,腐蚀去除噪声数据\r\n\t\topening = cv2.morphologyEx(thresh,cv2.MORPH_OPEN,kernel,iterations=2)\r\n\r\n\t\t#膨胀,得到的大部分是前景区域\r\n\t\tsure_bg = cv2.dilate(opening,kernel,iterations=3)\r\n\t\tdis_transform = cv2.distanceTransform(opening,cv2.DIST_L2,5)\r\n\t\tret,sure_fg = cv2.threshold(dis_transform,0.7*dis_transform.max(),255,0)\r\n\t\tsure_fg = np.uint8(sure_fg)\r\n\t\t#前景与背景有交叉部分,通过相减处理\r\n\t\tunknown = cv2.subtract(sure_bg,sure_fg)\r\n\t\tret,markers = cv2.connectedComponents(sure_fg)\r\n\t\tmarkers = markers+1\r\n\t\tmarkers[unknown==255] = 0\r\n\t\t\r\n\t\tmarkers = cv2.watershed(self.img,markers)\r\n\r\n\t\tself.img[markers==-1] = [255,0,0]\r\n\r\n\r\n\r\n\tdef show(self):\r\n\t\tcv2.imshow('img',self.img)\r\n\t\tcv2.waitKey()\r\n\r\n", "sub_path": "checkImg.py", "file_name": "checkImg.py", "file_ext": "py", "file_size_in_byte": 2197, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "cv2.threshold", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.findContours", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.RETR_TREE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.CHAIN_APPROX_SIMPLE", "line_number": 20, "usage_type": "attribute"}, {"api_name": "cv2.boundingRect", "line_number": 23, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 27, "usage_type": "call"}, {"api_name": "cv2.minAreaRect", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.boxPoints", "line_number": 30, "usage_type": "call"}, {"api_name": "numpy.int0", "line_number": 31, "usage_type": "call"}, {"api_name": "cv2.drawContours", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 35, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 36, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 37, "usage_type": "attribute"}, {"api_name": "cv2.grabCut", "line_number": 40, "usage_type": "call"}, {"api_name": "cv2.GC_INIT_WITH_RECT", "line_number": 40, "usage_type": "attribute"}, {"api_name": "numpy.where", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.newaxis", "line_number": 42, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 45, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 45, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 46, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY_INV", "line_number": 46, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_OTSU", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 47, "usage_type": "attribute"}, {"api_name": "cv2.morphologyEx", "line_number": 49, "usage_type": "call"}, {"api_name": "cv2.MORPH_OPEN", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cv2.dilate", "line_number": 52, "usage_type": "call"}, {"api_name": "cv2.distanceTransform", "line_number": 53, "usage_type": "call"}, {"api_name": "cv2.DIST_L2", "line_number": 53, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.subtract", "line_number": 57, "usage_type": "call"}, {"api_name": "cv2.connectedComponents", "line_number": 58, "usage_type": "call"}, {"api_name": "cv2.watershed", "line_number": 62, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 69, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 70, "usage_type": "call"}]} +{"seq_id": "543534071", "text": "# Copyright 2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\").\n# You may not use this file except in compliance with the License.\n# A copy of the License is located at:\n#\n# http://aws.amazon.com/apache2.0/\n#\n# or in the \"license\" file accompanying this file. This file is\n# distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS\n# OF ANY KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations under the\n# License.\n\n\"\"\"Ion core types.\"\"\"\n\n# Python 2/3 compatibility\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nfrom collections import MutableMapping, MutableSequence\nfrom datetime import datetime, timedelta, tzinfo\nfrom math import isnan\n\nimport six\n\nfrom .util import Enum\nfrom .util import record\n\n\nclass IonType(Enum):\n \"\"\"Enumeration of the Ion data types.\"\"\"\n NULL = 0\n BOOL = 1\n INT = 2\n FLOAT = 3\n DECIMAL = 4\n TIMESTAMP = 5\n SYMBOL = 6\n STRING = 7\n CLOB = 8\n BLOB = 9\n LIST = 10\n SEXP = 11\n STRUCT = 12\n\n @property\n def is_numeric(self):\n return IonType.INT <= self <= IonType.TIMESTAMP\n\n @property\n def is_text(self):\n \"\"\"Returns whether the type is a Unicode textual type.\"\"\"\n return self is IonType.SYMBOL or self is IonType.STRING\n\n @property\n def is_lob(self):\n \"\"\"Returns whether the type is a LOB.\"\"\"\n return self is IonType.CLOB or self is IonType.BLOB\n\n @property\n def is_container(self):\n \"\"\"Returns whether the type is a container.\"\"\"\n return self >= IonType.LIST\n\n\n# TODO At some point we can add SCALAR_START/SCALAR_END for streaming large values.\nclass IonEventType(Enum):\n \"\"\"Enumeration of Ion parser or serializer events.\n\n These types do not correspond directly to the Ion type system, but they are related.\n In particular, ``null.*`` will surface as a ``SCALAR`` even though they are containers.\n\n Attributes:\n INCOMPLETE: Indicates that parsing cannot be completed due to lack of input.\n STREAM_END: Indicates that the logical stream has terminated.\n VERSION_MARKER: Indicates that the **Ion Version Marker** has been encountered.\n SCALAR: Indicates an *atomic* value has been encountered.\n CONTAINER_START: Indicates that the start of a container has been reached.\n CONTAINER_END: Indicates that the end of a container has been reached.\n \"\"\"\n INCOMPLETE = -2\n STREAM_END = -1\n VERSION_MARKER = 0\n SCALAR = 1\n CONTAINER_START = 2\n CONTAINER_END = 3\n\n @property\n def begins_value(self):\n \"\"\"Indicates if the event type is a start of a value.\"\"\"\n return self is IonEventType.SCALAR or self is IonEventType.CONTAINER_START\n\n @property\n def ends_container(self):\n \"\"\"Indicates if the event type terminates a container or stream.\"\"\"\n return self is IonEventType.STREAM_END or self is IonEventType.CONTAINER_END\n\n @property\n def is_stream_signal(self):\n \"\"\"Indicates that the event type corresponds to a stream signal.\"\"\"\n return self < 0\n\n\nclass IonEvent(record(\n 'event_type',\n ('ion_type', None),\n ('value', None),\n ('field_name', None),\n ('annotations', ()),\n ('depth', None)\n )):\n \"\"\"An parse or serialization event.\n\n Args:\n event_type (IonEventType): The type of event.\n ion_type (Optional(amazon.ion.core.IonType)): The Ion data model type\n associated with the event.\n value (Optional[any]): The data value associated with the event.\n field_name (Optional[Union[amazon.ion.symbols.SymbolToken, unicode]]): The field name\n associated with the event.\n annotations (Sequence[Union[amazon.ion.symbols.SymbolToken, unicode]]): The annotations\n associated with the event.\n depth (Optional[int]): The tree depth of the event if applicable.\n \"\"\"\n def __eq__(self, other):\n if not isinstance(other, IonEvent):\n return False\n\n if isinstance(self.value, float):\n if not isinstance(other.value, float):\n return False\n\n # Need to deal with NaN appropriately.\n if self.value != other.value and not (isnan(self.value) and isnan(other.value)):\n return False\n else:\n if self.value != other.value:\n return False\n\n # Timestamp precision has additional requirements.\n if isinstance(self.value, Timestamp) or isinstance(other.value, Timestamp):\n # Special case for timestamps to capture equivalence over precision.\n self_precision = getattr(self.value, TIMESTAMP_PRECISION_FIELD, None)\n other_precision = getattr(other.value, TIMESTAMP_PRECISION_FIELD, None)\n if self_precision != other_precision \\\n and not ((self_precision is None and other_precision is TimestampPrecision.SECOND) or\n (self_precision is TimestampPrecision.SECOND and other_precision is None)):\n # The absence of precision indicates a naive datetime, which always has SECOND precision.\n return False\n if isinstance(self.value, datetime):\n if self.value.utcoffset() != other.value.utcoffset():\n return False\n\n return (self.event_type == other.event_type\n and self.ion_type == other.ion_type\n and self.field_name == other.field_name\n and self.annotations == other.annotations\n and self.depth == other.depth\n )\n\n def derive_field_name(self, field_name):\n \"\"\"Derives a new event from this one setting the ``field_name`` attribute.\n\n Args:\n field_name (Union[amazon.ion.symbols.SymbolToken, unicode]): The field name to set.\n Returns:\n IonEvent: The newly generated event.\n \"\"\"\n cls = type(self)\n # We use ordinals to avoid thunk materialization.\n return cls(\n self[0],\n self[1],\n self[2],\n field_name,\n self[4],\n self[5]\n )\n\n def derive_annotations(self, annotations):\n \"\"\"Derives a new event from this one setting the ``annotations`` attribute.\n\n Args:\n annotations: (Sequence[Union[amazon.ion.symbols.SymbolToken, unicode]]):\n The annotations associated with the derived event.\n\n Returns:\n IonEvent: The newly generated event.\n \"\"\"\n cls = type(self)\n # We use ordinals to avoid thunk materialization.\n return cls(\n self[0],\n self[1],\n self[2],\n self[3],\n annotations,\n self[5]\n )\n\n def derive_value(self, value):\n \"\"\"Derives a new event from this one setting the ``value`` attribute.\n\n Args:\n value: (any):\n The value associated with the derived event.\n\n Returns:\n IonEvent: The newly generated non-thunk event.\n \"\"\"\n return IonEvent(\n self.event_type,\n self.ion_type,\n value,\n self.field_name,\n self.annotations,\n self.depth\n )\n\n def derive_depth(self, depth):\n \"\"\"Derives a new event from this one setting the ``depth`` attribute.\n\n Args:\n depth: (int):\n The annotations associated with the derived event.\n\n Returns:\n IonEvent: The newly generated event.\n \"\"\"\n cls = type(self)\n # We use ordinals to avoid thunk materialization.\n return cls(\n self[0],\n self[1],\n self[2],\n self[3],\n self[4],\n depth\n )\n\n\nclass MemoizingThunk(object):\n \"\"\"A :class:`callable` that invokes a ``delegate`` and caches and returns the result.\"\"\"\n def __init__(self, delegate):\n self.delegate = delegate\n\n def __call__(self):\n if hasattr(self, 'value'):\n return self.value\n self.value = self.delegate()\n return self.value\n\n def __str__(self):\n return str(self())\n\n def __repr__(self):\n return repr(self())\n\n\nclass IonThunkEvent(IonEvent):\n \"\"\"An :class:`IonEvent` whose ``value`` field is a thunk.\"\"\"\n def __new__(cls, *args, **kwargs):\n if len(args) >= 3:\n args = list(args)\n args[2] = MemoizingThunk(args[2])\n else:\n value = kwargs.get('value')\n if value is not None:\n kwargs['value'] = MemoizingThunk(kwargs['value'])\n return super(IonThunkEvent, cls).__new__(cls, *args, **kwargs)\n\n @property\n def value(self):\n # We're masking the value field, this gets around that.\n return self[2]()\n\n# Singletons for structural events\nION_STREAM_END_EVENT = IonEvent(IonEventType.STREAM_END)\nION_STREAM_INCOMPLETE_EVENT = IonEvent(IonEventType.INCOMPLETE)\nION_VERSION_MARKER_EVENT = IonEvent(\n IonEventType.VERSION_MARKER, ion_type=None, value=(1, 0), depth=0\n)\n\n\nclass DataEvent(record('type', 'data')):\n \"\"\"Event generated as a result of the writer or as input into the reader.\n\n Args:\n type (Enum): The type of event.\n data (bytes): The serialized data returned. If no data is to be serialized,\n this should be the empty byte string.\n \"\"\"\n\n\nclass Transition(record('event', 'delegate')):\n \"\"\"A pair of event and co-routine delegate.\n\n This is generally used as a result of a state-machine.\n\n Args:\n event (Union[DataEvent]): The event associated with the transition.\n delegate (Coroutine): The co-routine delegate which can be the same routine from\n whence this transition came.\n \"\"\"\n\n_MIN_OFFSET = timedelta(hours=-24)\n_MAX_OFFSET = timedelta(hours=24)\n_ZERO_DELTA = timedelta()\n\n\nclass OffsetTZInfo(tzinfo):\n \"\"\"A trivial UTC offset :class:`tzinfo`.\"\"\"\n def __init__(self, delta=_ZERO_DELTA):\n if delta <= _MIN_OFFSET or delta >= _MAX_OFFSET:\n raise ValueError('Invalid UTC offset: %s' % delta)\n self.delta = delta\n\n def dst(self, date_time):\n return timedelta()\n\n def tzname(self, date_time):\n return None\n\n def utcoffset(self, date_time):\n return self.delta\n\n def __repr__(self):\n sign = '+'\n delta = self.delta\n if delta < _ZERO_DELTA:\n sign = '-'\n delta = _ZERO_DELTA - delta\n return 'OffsetTZInfo(%s%s)' % (sign, delta)\n\n\nclass TimestampPrecision(Enum):\n \"\"\"The different levels of precision supported in an Ion timestamp.\"\"\"\n YEAR = 0\n MONTH = 1\n DAY = 2\n MINUTE = 3\n SECOND = 4\n\n @property\n def includes_month(self):\n \"\"\"Precision has at least the ``month`` field.\"\"\"\n return self >= TimestampPrecision.MONTH\n\n @property\n def includes_day(self):\n \"\"\"Precision has at least the ``day`` field.\"\"\"\n return self >= TimestampPrecision.DAY\n\n @property\n def includes_minute(self):\n \"\"\"Precision has at least the ``minute`` field.\"\"\"\n return self >= TimestampPrecision.MINUTE\n\n @property\n def includes_second(self):\n \"\"\"Precision has at least the ``second`` field.\"\"\"\n return self >= TimestampPrecision.SECOND\n\n\nTIMESTAMP_PRECISION_FIELD = 'precision'\nTIMESTAMP_FRACTION_PRECISION_FIELD = 'fractional_precision'\nMICROSECOND_PRECISION = 6\n\n\nclass Timestamp(datetime):\n \"\"\"Sub-class of :class:`datetime` that supports a precision field\n\n Notes:\n The ``precision`` field is passed as a keyword argument of the same name.\n \"\"\"\n __slots__ = [TIMESTAMP_PRECISION_FIELD, TIMESTAMP_FRACTION_PRECISION_FIELD]\n\n def __new__(cls, *args, **kwargs):\n precision = None\n fractional_precision = None\n if TIMESTAMP_PRECISION_FIELD in kwargs:\n precision = kwargs.get(TIMESTAMP_PRECISION_FIELD)\n # Make sure we mask this before we construct the datetime.\n del kwargs[TIMESTAMP_PRECISION_FIELD]\n if TIMESTAMP_FRACTION_PRECISION_FIELD in kwargs:\n fractional_precision = kwargs.get(TIMESTAMP_FRACTION_PRECISION_FIELD)\n if fractional_precision is not None and 1 > fractional_precision > MICROSECOND_PRECISION:\n raise ValueError('Cannot construct a Timestamp with fractional precision of %d digits, '\n 'which is out of the supported range of [1, %d].'\n % (fractional_precision, MICROSECOND_PRECISION,))\n # Make sure we mask this before we construct the datetime.\n del kwargs[TIMESTAMP_FRACTION_PRECISION_FIELD]\n\n instance = super(Timestamp, cls).__new__(cls, *args, **kwargs)\n setattr(instance, TIMESTAMP_PRECISION_FIELD, precision)\n setattr(instance, TIMESTAMP_FRACTION_PRECISION_FIELD, fractional_precision)\n\n return instance\n\n def __repr__(self):\n return 'Timestamp(%04d-%02d-%02dT%02d:%02d:%02d.%06d, %r, %r, %s=%s)' % \\\n (self.year, self.month, self.day,\n self.hour, self.minute, self.second, self.microsecond,\n self.tzinfo, self.precision,\n TIMESTAMP_FRACTION_PRECISION_FIELD, self.fractional_precision)\n\n @staticmethod\n def adjust_from_utc_fields(*args, **kwargs):\n \"\"\"Constructs a timestamp from UTC fields adjusted to the local offset if given.\"\"\"\n raw_ts = Timestamp(*args, **kwargs)\n offset = raw_ts.utcoffset()\n if offset is None or offset == timedelta():\n return raw_ts\n\n # XXX This returns a datetime, not a Timestamp (which has our precision if defined)\n adjusted = raw_ts + offset\n if raw_ts.precision is None:\n # No precision means we can just return a regular datetime\n return adjusted\n\n return Timestamp(\n adjusted.year,\n adjusted.month,\n adjusted.day,\n adjusted.hour,\n adjusted.minute,\n adjusted.second,\n adjusted.microsecond,\n raw_ts.tzinfo,\n precision=raw_ts.precision,\n fractional_precision=raw_ts.fractional_precision\n )\n\n\ndef timestamp(year, month=1, day=1,\n hour=0, minute=0, second=0, microsecond=None,\n off_hours=None, off_minutes=None,\n precision=None, fractional_precision=None):\n \"\"\"Shorthand for the :class:`Timestamp` constructor.\n\n Specifically, converts ``off_hours`` and ``off_minutes`` parameters to a suitable\n :class:`OffsetTZInfo` instance.\n \"\"\"\n delta = None\n if off_hours is not None:\n if off_hours < -23 or off_hours > 23:\n raise ValueError('Hour offset %d is out of required range -23..23.' % (off_hours,))\n delta = timedelta(hours=off_hours)\n if off_minutes is not None:\n if off_minutes < -59 or off_minutes > 59:\n raise ValueError('Minute offset %d is out of required range -59..59.' % (off_minutes,))\n minutes_delta = timedelta(minutes=off_minutes)\n if delta is None:\n delta = minutes_delta\n else:\n delta += minutes_delta\n\n tz = None\n if delta is not None:\n tz = OffsetTZInfo(delta)\n\n if microsecond is not None:\n if fractional_precision is None:\n fractional_precision = MICROSECOND_PRECISION\n else:\n microsecond = 0\n if fractional_precision is not None:\n raise ValueError('Cannot have fractional precision without a fractional component.')\n\n return Timestamp(\n year, month, day,\n hour, minute, second, microsecond,\n tz, precision=precision, fractional_precision=fractional_precision\n )\n\n\nclass Multimap(MutableMapping):\n \"\"\"\n Dictionary that can hold multiple values for the same key\n\n In order not to break existing customers, getting and inserting elements with ``[]`` keeps the same behaviour\n as the built-in dict. If multiple elements are already mapped to the key, ``[]` will return\n the newest one.\n\n To map multiple elements to a key, use the ``add_item`` operation.\n To retrieve all the values map to a key, use ``get_all_values``.\n \"\"\"\n\n def __init__(self, *args, **kwargs):\n super(Multimap, self).__init__()\n self.__store = {}\n if args is not None and len(args) > 0:\n for key, value in six.iteritems(args[0]):\n self.__store[key] = MultimapValue(value)\n\n def __getitem__(self, key):\n return self.__store[key][len(self.__store[key]) - 1] # Return only one in order not to break clients\n\n def __delitem__(self, key):\n del self.__store[key]\n\n def __setitem__(self, key, value):\n self.__store[key] = MultimapValue(value)\n\n def __len__(self):\n return sum([len(values) for values in six.itervalues(self.__store)])\n\n def __iter__(self):\n for key in six.iterkeys(self.__store):\n yield key\n\n def add_item(self, key, value):\n if key in self.__store:\n self.__store[key].append(value)\n else:\n self.__setitem__(key, value)\n\n def get_all_values(self, key):\n return self.__store[key]\n\n def iteritems(self):\n for key in self.__store:\n for value in self.__store[key]:\n yield (key, value)\n\n def items(self):\n output = []\n for k, v in self.iteritems():\n output.append((k, v))\n return output\n\n\nclass MultimapValue(MutableSequence):\n\n def __init__(self, *args):\n if args is not None:\n self.__store = [x for x in args]\n else:\n self.__store = []\n\n def insert(self, index, value):\n self.__setitem__(index, value)\n\n def __len__(self):\n return len(self.__store)\n\n def __getitem__(self, index):\n return self.__store[index]\n\n def __setitem__(self, index, value):\n self.__store.insert(index, value)\n\n def __delitem__(self, index):\n del self.__store[index]\n\n def __iter__(self):\n for x in self.__store:\n yield x\n", "sub_path": "amazon/ion/core.py", "file_name": "core.py", "file_ext": "py", "file_size_in_byte": 18366, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "util.Enum", "line_number": 32, "usage_type": "name"}, {"api_name": "util.Enum", "line_number": 69, "usage_type": "name"}, {"api_name": "util.record", "line_number": 106, "usage_type": "call"}, {"api_name": "math.isnan", "line_number": 136, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 152, "usage_type": "argument"}, {"api_name": "util.record", "line_number": 287, "usage_type": "call"}, {"api_name": "util.record", "line_number": 297, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 308, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 309, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 310, "usage_type": "call"}, {"api_name": "datetime.tzinfo", "line_number": 313, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 321, "usage_type": "call"}, {"api_name": "util.Enum", "line_number": 338, "usage_type": "name"}, {"api_name": "datetime.datetime", "line_number": 372, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 414, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 450, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 454, "usage_type": "call"}, {"api_name": "collections.MutableMapping", "line_number": 479, "usage_type": "name"}, {"api_name": "six.iteritems", "line_number": 495, "usage_type": "call"}, {"api_name": "six.itervalues", "line_number": 508, "usage_type": "call"}, {"api_name": "six.iterkeys", "line_number": 511, "usage_type": "call"}, {"api_name": "collections.MutableSequence", "line_number": 535, "usage_type": "name"}]} +{"seq_id": "122008806", "text": "#!/usr/bin/env python\n# coding=utf-8\n\n# Author: Junjie Wang\n# Mail: dreamboy.gns@sjtu.edu.cn\n\n# Website:http://www.dbgns.com\n# Blog:http://www.dbgns.com/blog\n\nimport torch\nimport torch.nn as nn\nimport random, time\nfrom pynn.utils.focal_loss import FocalLoss\nfrom pynn.utils.label_smooth_loss import LSRLoss\n\n\"\"\"\ntest the focal loss and label smoothing regularization cross entropy loss\n\"\"\"\n\n\ndef focal_loss_test(num_examples=1000, random_seed=0):\n \"\"\" test the focal loss\n if you want to test the alpha option,\n e.g. set alpha=0.5 and double the loss_focal, then the error will be approximately 0\n :param num_examples: number of random examples\n :param random_seed: keep the results consistent\n \"\"\"\n random.seed(random_seed)\n criterion_standard = nn.CrossEntropyLoss()\n criterion_focal = FocalLoss(gamma=0)\n max_error = 0\n since = time.time()\n for _ in range(num_examples):\n inputs = torch.rand(6400, 2) * random.randint(0, 10)\n labels = torch.rand(6400).ge(0.1).long()\n inputs = inputs.cuda()\n labels = labels.cuda()\n loss_std = criterion_standard(inputs, labels).item()\n loss_focal = criterion_focal(inputs, labels).item()\n if abs(loss_focal - loss_std) > max_error:\n max_error = abs(loss_focal - loss_std)\n time_elapsed = time.time() - since\n print(\"Test takes {0:.4f} seconds for {1} random examples, the maximum error is {2}.\".format(time_elapsed, num_examples, max_error))\n\n\ndef lsr_loss_test(num_examples=1000, random_seed=0):\n \"\"\" test the label smoothing regularization loss\n :param num_examples: number of random examples\n :param random_seed: keep the results consistent\n \"\"\"\n random.seed(random_seed)\n criterion_standard = nn.CrossEntropyLoss()\n criterion_lsr = LSRLoss(lsr_param=0)\n max_error = 0\n since = time.time()\n for _ in range(num_examples):\n inputs = torch.rand(6400, 2) * random.randint(0, 10)\n labels = torch.rand(6400).ge(0.1).long()\n inputs = inputs.cuda()\n labels = labels.cuda()\n loss_std = criterion_standard(inputs, labels).item()\n loss_lsr = criterion_lsr(inputs, labels).item()\n if abs(loss_lsr - loss_std) > max_error:\n max_error = abs(loss_lsr - loss_std)\n time_elapsed = time.time() - since\n print(\"Test takes {0:.4f} seconds for {1} random examples, the maximum error is {2}.\".format(time_elapsed, num_examples, max_error))\n\n\nif __name__ == '__main__':\n lsr_loss_test()\n\n", "sub_path": "demo/loss_test.py", "file_name": "loss_test.py", "file_ext": "py", "file_size_in_byte": 2508, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "random.seed", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "pynn.utils.focal_loss.FocalLoss", "line_number": 30, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 34, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 35, "usage_type": "call"}, {"api_name": "time.time", "line_number": 42, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 51, "usage_type": "call"}, {"api_name": "torch.nn.CrossEntropyLoss", "line_number": 52, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 52, "usage_type": "name"}, {"api_name": "pynn.utils.label_smooth_loss.LSRLoss", "line_number": 53, "usage_type": "call"}, {"api_name": "time.time", "line_number": 55, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 57, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 57, "usage_type": "call"}, {"api_name": "torch.rand", "line_number": 58, "usage_type": "call"}, {"api_name": "time.time", "line_number": 65, "usage_type": "call"}]} +{"seq_id": "651382288", "text": "\"\"\"\n.. module MultiCropAndOpenFace\n :synopsis: Script to apply cropping and OpenFace to all videos in a directory.\n\n\"\"\"\n\nimport glob\nimport json\nimport os\nimport subprocess\nfrom warnings import warn\nimport sys\nimport numpy as np\nsys.path.append(os.path.dirname(os.path.abspath(__file__)))\nimport CropAndOpenFace\nfrom timeit import default_timer as timer\n\n\ndef make_vids(input_path, output_path, emotions = False):\n \"\"\"\n Return list of vids not processed yet given a path.\n NEW: Also return only those that have emotions file\n :param path: Path to video directory\n :type path: str\n :return: list of vids to do\n \"\"\"\n folder_components = set(os.path.join(output_path, x) for x in os.listdir(output_path))\n\n #this is to find all .avi videos that are in the given input dir, even recursively\n #might not always be needed (somewhat time consuming)\n paths = []\n for root, dirs, files in os.walk(input_path):\n for file in files:\n if file.endswith(\".avi\"):\n paths.append(os.path.join(root, file))\n\n #this is for processing only those videos we have emotion annotations for\n to_process = []\n if emotions:\n for p in paths:\n x = os.path.splitext(os.path.split(p)[1])[0]\n if x + '_emotions.csv' in os.listdir('/home/emil/emotion_annotations'):\n if (os.path.join(output_path, x) + '_cropped' not in folder_components\n or 'hdfs' not in os.listdir(\n os.path.join(output_path,\n x + '_cropped'))):\n to_process.append(p)\n\n\n else:\n for p in paths:\n x = os.path.splitext(os.path.split(p)[1])[0]\n if (os.path.join(output_path, x) + '_cropped' not in folder_components\n or 'hdfs' not in os.listdir(\n os.path.join(output_path,\n x + '_cropped'))):\n to_process.append(p)\n return to_process\n\n\n\ndef make_crop_and_nose_files(path): #FOUND OUT: These are just supposed to be a collection of patient_day_vid key and ACTUAL crop file path as value, so basically a lookup dictionary.\n crop_file = os.path.join(path, 'crop_files_list.txt')\n nose_file = os.path.join(path, 'nose_files_list.txt')\n\n if not os.path.exists(crop_file):\n crop_path = sys.argv[sys.argv.index('-c') + 1]\n crop_txt_files = CropAndOpenFace.find_txt_files(crop_path)\n json.dump(crop_txt_files, open(crop_file, mode='w'))\n\n if not os.path.exists(nose_file):\n nose_path = sys.argv[sys.argv.index('-n') + 1]\n nose_txt_files = CropAndOpenFace.find_txt_files(nose_path)\n json.dump(nose_txt_files, open(nose_file, mode='w'))\n\n return json.load(open(crop_file)), json.load(open(nose_file))\n\nif __name__ == '__main__':\n input_path = sys.argv[sys.argv.index('-id') + 1]\n output_path = sys.argv[sys.argv.index('-od') + 1]\n\n vids = make_vids(input_path,output_path)\n num_GPUs = 2\n processes = []\n indices = np.linspace(0, len(vids), num=num_GPUs + 1)\n\n # TODO: make this a cmd-line arg\n CONDA_ENV = '/home/emil/miniconda3/envs/br_doc/bin/python'\n\n for index in range(len(indices) - 1):\n if '-c' not in sys.argv:\n cmd = [\"ionice -c2 -n7\",\n CONDA_ENV,\n os.path.join(\n os.path.dirname(os.path.dirname(os.path.abspath(__file__))),\n 'helpers', 'HalfCropper.py'), '-id', input_path, '-vl',\n str(int(indices[index])), '-vr',\n str(int(indices[index + 1])), '-od', output_path\n ]\n else:\n cmd = [\"ionice -c2 -n7\",\n CONDA_ENV,\n os.path.join(\n os.path.dirname(os.path.dirname(os.path.abspath(__file__))),\n 'helpers', 'HalfCropper.py'), '-id', input_path, '-od', output_path, '-vl',\n str(int(indices[index])), '-vr',\n str(int(indices[index + 1])), '-c', sys.argv[sys.argv.index('-c') + 1], '-n', sys.argv[sys.argv.index('-n') + 1]\n ]\n processes.append(\n subprocess.Popen(\n cmd, env={'CUDA_VISIBLE_DEVICES': '{0}'.format(str(index))}))\n start = timer()\n [p.wait() for p in processes]\n print(timer()-start, 'so viel zeit')\n", "sub_path": "runners/MultiCropAndOpenFace.py", "file_name": "MultiCropAndOpenFace.py", "file_ext": "py", "file_size_in_byte": 4381, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.path.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 27, "usage_type": "call"}, {"api_name": "os.walk", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 35, "usage_type": "call"}, {"api_name": "os.path", "line_number": 35, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 41, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 42, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 43, "usage_type": "call"}, {"api_name": "os.path", "line_number": 43, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 45, "usage_type": "call"}, {"api_name": "os.path", "line_number": 45, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path", "line_number": 52, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 67, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 67, "usage_type": "call"}, {"api_name": "CropAndOpenFace.find_txt_files", "line_number": 68, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 71, "usage_type": "call"}, {"api_name": "os.path", "line_number": 71, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 72, "usage_type": "call"}, {"api_name": "CropAndOpenFace.find_txt_files", "line_number": 73, "usage_type": "call"}, {"api_name": "json.dump", "line_number": 74, "usage_type": "call"}, {"api_name": "json.load", "line_number": 76, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 79, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 79, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 80, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 85, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 91, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 94, "usage_type": "call"}, {"api_name": "os.path", "line_number": 94, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path", "line_number": 104, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 104, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 107, "usage_type": "attribute"}, {"api_name": "sys.argv.index", "line_number": 107, "usage_type": "call"}, {"api_name": "subprocess.Popen", "line_number": 110, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 112, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 114, "usage_type": "call"}]} +{"seq_id": "283791803", "text": "# -*- coding: utf-8 -*-\n\nfrom mitmproxy import options\nfrom mitmproxy.proxy import ProxyConfig, ProxyServer\nfrom mitmproxy.tools.dump import DumpMaster\n\nfrom addons import Counter\n\n\nclass Server(DumpMaster):\n\n def __init__(\n self, options, with_termlog=True, with_dumper=True\n ):\n super().__init__(options, with_termlog, with_dumper)\n self.addons.add(Counter())\n\n\nif __name__ == '__main__':\n opts = options.Options()\n opts.set(\n 'listen_host=0.0.0.0',\n 'listen_port=8080'\n )\n server = Server(opts, with_dumper=False)\n config = ProxyConfig(opts)\n server.server = ProxyServer(config=config)\n server.run()\n", "sub_path": "server.py", "file_name": "server.py", "file_ext": "py", "file_size_in_byte": 665, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "mitmproxy.tools.dump.DumpMaster", "line_number": 10, "usage_type": "name"}, {"api_name": "mitmproxy.options", "line_number": 15, "usage_type": "argument"}, {"api_name": "addons.Counter", "line_number": 16, "usage_type": "call"}, {"api_name": "mitmproxy.options.Options", "line_number": 20, "usage_type": "call"}, {"api_name": "mitmproxy.options", "line_number": 20, "usage_type": "name"}, {"api_name": "mitmproxy.proxy.ProxyConfig", "line_number": 26, "usage_type": "call"}, {"api_name": "mitmproxy.proxy.ProxyServer", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "556774153", "text": "\"\"\"\r\nTest script for file_io\r\n\"\"\"\r\n\r\nfrom __future__ import absolute_import, division, print_function\r\n\r\nfrom tests.util import unittest_reporter, glob_tests\r\n\r\nimport logging\r\nlogger = logging.getLogger('file_io_test')\r\n\r\nimport os, sys, time\r\nimport shutil\r\nimport tempfile\r\nimport random\r\nimport unittest\r\nfrom io import IOBase\r\n\r\nimport iceprod.server.file_io\r\n\r\n\r\nclass file_io_test(unittest.TestCase):\r\n def setUp(self):\r\n super(file_io_test,self).setUp()\r\n self.test_dir = tempfile.mkdtemp(dir=os.getcwd())\r\n\r\n def tearDown(self):\r\n shutil.rmtree(self.test_dir)\r\n super(file_io_test,self).tearDown()\r\n\r\n @unittest_reporter\r\n def test_01_open(self):\r\n \"\"\"Test file_io open\"\"\"\r\n filename = os.path.join(self.test_dir,'test')\r\n with open(filename,'w') as f:\r\n f.write('test')\r\n fileio = iceprod.server.file_io.AsyncFileIO()\r\n\r\n fut = fileio.open(filename)\r\n ret = None\r\n try:\r\n ret = fut.result(timeout=1)\r\n if not isinstance(ret,IOBase):\r\n raise Exception('did not return a file object')\r\n if ret.mode != 'r':\r\n raise Exception('file did not open in read mode')\r\n finally:\r\n if ret is not None:\r\n try:\r\n ret.close()\r\n except Exception:\r\n pass\r\n\r\n for mode in ('r','w','a','rb','wb','ab','rb+','wb+','ab+'):\r\n fut = fileio.open(filename,mode)\r\n ret = None\r\n try:\r\n ret = fut.result(timeout=1)\r\n self.assertIsInstance(ret,IOBase)\r\n if 'r' in mode or '+' in mode:\r\n self.assertTrue(ret.readable())\r\n if 'w' in mode or 'a' in mode:\r\n self.assertTrue(ret.writable())\r\n if 'b' in mode:\r\n self.assertIn('b', ret.mode)\r\n finally:\r\n if ret is not None:\r\n try:\r\n ret.close()\r\n except Exception:\r\n pass\r\n\r\n @unittest_reporter\r\n def test_02_close(self):\r\n \"\"\"Test file_io close\"\"\"\r\n filename = os.path.join(self.test_dir,'test')\r\n with open(filename,'w') as f:\r\n f.write('test')\r\n fileio = iceprod.server.file_io.AsyncFileIO()\r\n f = open(filename)\r\n try:\r\n fut = fileio.close(f)\r\n fut.result(timeout=1)\r\n if not f.closed:\r\n raise Exception('did not close file')\r\n finally:\r\n if not f.closed:\r\n f.close\r\n\r\n @unittest_reporter\r\n def test_03_read(self):\r\n \"\"\"Test file_io read\"\"\"\r\n filename = os.path.join(self.test_dir,'test')\r\n data = 'test'\r\n with open(filename,'w') as f:\r\n f.write(data)\r\n fileio = iceprod.server.file_io.AsyncFileIO()\r\n f = open(filename)\r\n try:\r\n fut = fileio.read(f)\r\n ret = fut.result(timeout=1)\r\n if ret != data:\r\n raise Exception('did not read data')\r\n finally:\r\n if not f.closed:\r\n f.close()\r\n\r\n data = ''.join(chr(i) for i in range(256)).encode('utf-8')\r\n with open(filename,'wb') as f:\r\n f.write(data)\r\n f = open(filename,'rb')\r\n try:\r\n fut = fileio.read(f,150)\r\n ret = fut.result(timeout=1)\r\n if ret != data[:150]:\r\n raise Exception('did not read 150 chars of data')\r\n fut = fileio.read(f)\r\n ret = fut.result(timeout=1)\r\n if ret != data[150:]:\r\n raise Exception('did not read rest of data')\r\n finally:\r\n if not f.closed:\r\n f.close()\r\n\r\n @unittest_reporter\r\n def test_04_readline(self):\r\n \"\"\"Test file_io readline\"\"\"\r\n filename = os.path.join(self.test_dir,'test')\r\n data = 'test\\ndata'\r\n with open(filename,'w') as f:\r\n f.write(data)\r\n fileio = iceprod.server.file_io.AsyncFileIO()\r\n f = open(filename)\r\n try:\r\n fut = fileio.readline(f)\r\n ret = fut.result(timeout=1)\r\n if ret != data.split('\\n')[0]+'\\n':\r\n logger.info('ret = %r',ret)\r\n logger.info('first line = %r',data.split('\\n')[0]+'\\n')\r\n raise Exception('did not read first line')\r\n fut = fileio.readline(f)\r\n ret = fut.result(timeout=1)\r\n if ret != data.split('\\n')[1]:\r\n logger.info('ret = %r',ret)\r\n logger.info('2nd line = %r',data.split('\\n')[1])\r\n raise Exception('did not read second line')\r\n finally:\r\n if not f.closed:\r\n f.close()\r\n\r\n @unittest_reporter\r\n def test_05_write(self):\r\n \"\"\"Test file_io write\"\"\"\r\n filename = os.path.join(self.test_dir,'test')\r\n data = 'test\\ndata'\r\n fileio = iceprod.server.file_io.AsyncFileIO()\r\n\r\n f = open(filename,'w')\r\n try:\r\n fut = fileio.write(f,data)\r\n fut.result(timeout=1)\r\n finally:\r\n if not f.closed:\r\n f.close()\r\n if not os.path.exists(filename):\r\n raise Exception('file does not exist')\r\n ret = open(filename).read()\r\n if ret != data:\r\n logger.info('ret = %r',ret)\r\n logger.info('data = %r',data)\r\n raise Exception('did not write data')\r\n\r\n\r\ndef load_tests(loader, tests, pattern):\r\n suite = unittest.TestSuite()\r\n alltests = glob_tests(loader.getTestCaseNames(file_io_test))\r\n suite.addTests(loader.loadTestsFromNames(alltests,file_io_test))\r\n return suite\r\n", "sub_path": "tests/server/file_io_test.py", "file_name": "file_io_test.py", "file_ext": "py", "file_size_in_byte": 5811, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "unittest.TestCase", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tempfile.mkdtemp", "line_number": 25, "usage_type": "call"}, {"api_name": "os.getcwd", "line_number": 25, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io.server.file_io.AsyncFileIO", "line_number": 37, "usage_type": "call"}, {"api_name": "iceprod.server.file_io.server", "line_number": 37, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io", "line_number": 37, "usage_type": "name"}, {"api_name": "io.IOBase", "line_number": 43, "usage_type": "argument"}, {"api_name": "io.IOBase", "line_number": 59, "usage_type": "argument"}, {"api_name": "tests.util.unittest_reporter", "line_number": 31, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io.server.file_io.AsyncFileIO", "line_number": 79, "usage_type": "call"}, {"api_name": "iceprod.server.file_io.server", "line_number": 79, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io", "line_number": 79, "usage_type": "name"}, {"api_name": "tests.util.unittest_reporter", "line_number": 73, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 93, "usage_type": "call"}, {"api_name": "os.path", "line_number": 93, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io.server.file_io.AsyncFileIO", "line_number": 97, "usage_type": "call"}, {"api_name": "iceprod.server.file_io.server", "line_number": 97, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io", "line_number": 97, "usage_type": "name"}, {"api_name": "tests.util.unittest_reporter", "line_number": 90, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 128, "usage_type": "call"}, {"api_name": "os.path", "line_number": 128, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io.server.file_io.AsyncFileIO", "line_number": 132, "usage_type": "call"}, {"api_name": "iceprod.server.file_io.server", "line_number": 132, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io", "line_number": 132, "usage_type": "name"}, {"api_name": "tests.util.unittest_reporter", "line_number": 125, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 154, "usage_type": "call"}, {"api_name": "os.path", "line_number": 154, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io.server.file_io.AsyncFileIO", "line_number": 156, "usage_type": "call"}, {"api_name": "iceprod.server.file_io.server", "line_number": 156, "usage_type": "attribute"}, {"api_name": "iceprod.server.file_io", "line_number": 156, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 165, "usage_type": "call"}, {"api_name": "os.path", "line_number": 165, "usage_type": "attribute"}, {"api_name": "tests.util.unittest_reporter", "line_number": 151, "usage_type": "name"}, {"api_name": "unittest.TestSuite", "line_number": 175, "usage_type": "call"}, {"api_name": "tests.util.glob_tests", "line_number": 176, "usage_type": "call"}]} +{"seq_id": "653395350", "text": "# coding:utf-8\nfrom __future__ import print_function\nimport numpy as np\nnp.random.seed(1337)\n\nfrom keras.datasets import mnist\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout, Activation, Flatten\nfrom keras.layers import Convolution2D, MaxPooling2D\nfrom keras.utils import np_utils\nfrom keras import backend as K\nfrom keras import initializers\nfrom keras.utils.vis_utils import plot_model\n\nbatch_size = 128\nnb_classes = 10\nnb_epoch = 40\n\n# 输入数据的维度\nimg_rows, img_cols = 28, 28\n# 使用的卷积滤波器的数量\nnb_filters = 6\n# 用于 max pooling 的池化面积\npool_size = (2, 2)\n# 卷积核的尺寸\nkernel_size = (5, 5)\n\n# the data, shuffled and split between train and test sets\n(X_train, y_train), (X_test, y_test) = mnist.load_data()\n\nif K.image_dim_ordering() == 'th':\n X_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols)\n X_test = X_test.reshape(X_test.shape[0], 1, img_rows, img_cols)\n input_shape = (1, img_rows, img_cols)\nelse:\n X_train = X_train.reshape(X_train.shape[0], img_rows, img_cols, 1)\n X_test = X_test.reshape(X_test.shape[0], img_rows, img_cols, 1)\n input_shape = (img_rows, img_cols, 1)\n\n\nX_train = X_train.astype('float32')\nX_test = X_test.astype('float32')\nX_train /= 255\nX_test /= 255\nprint('X_train shape:', X_train.shape)\nprint(X_train.shape[0], 'train samples')\nprint(X_test.shape[0], 'test samples')\n\n# convert class vectors to binary class matrices\nY_train = np_utils.to_categorical(y_train, nb_classes)\nY_test = np_utils.to_categorical(y_test, nb_classes)\n\nmodel = Sequential()\n# C1 卷积层1 卷积核6个 5*5\nmodel.add(Convolution2D(nb_filters, kernel_size[0], kernel_size[1],\n border_mode='valid',\n input_shape=input_shape, kernel_initializer='random_normal'))\nmodel.add(Activation('sigmoid'))\n# S2 下采样\nmodel.add(MaxPooling2D(pool_size=pool_size))\n# C3卷积层2 16个卷积核 5*5\nmodel.add(Convolution2D(\n 16, kernel_size[0], kernel_size[1], kernel_initializer='random_normal'))\nmodel.add(Activation('sigmoid'))\n# S4 下采样\nmodel.add(MaxPooling2D(pool_size=pool_size))\n# C5 卷积层3 120个卷积核 3*3\nmodel.add(Convolution2D(120, 3, 3, kernel_initializer='random_normal'))\nmodel.add(Activation('sigmoid'))\n# 转化为一维\nmodel.add(Flatten())\n# F6 全连接层 输出层\nmodel.add(Dense(nb_classes, kernel_initializer='random_normal'))\nmodel.add(Activation('softmax'))\n# print model\nmodel.summary()\n\nplot_model(model, to_file='cnn3-3.png')\n\nmodel.compile(loss='categorical_crossentropy',\n optimizer='adadelta',\n metrics=['accuracy'])\n\nhistory = model.fit(X_train, Y_train, batch_size=batch_size, nb_epoch=nb_epoch,\n verbose=1, validation_data=(X_test, Y_test))\n\nwith open('cnn3-3.txt', 'w') as f:\n f.write(str(history.history))\n\nscore = model.evaluate(X_test, Y_test, verbose=0)\nprint('Test score:', score[0])\nprint('Test accuracy:', score[1])\n", "sub_path": "李鑫_手写数字识别算法/毕业设计源码/CNN3权值初始化/cnn3-3.py", "file_name": "cnn3-3.py", "file_ext": "py", "file_size_in_byte": 2972, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.random.seed", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 4, "usage_type": "attribute"}, {"api_name": "keras.datasets.mnist.load_data", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.datasets.mnist", "line_number": 29, "usage_type": "name"}, {"api_name": "keras.backend.image_dim_ordering", "line_number": 31, "usage_type": "call"}, {"api_name": "keras.backend", "line_number": 31, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 50, "usage_type": "name"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 51, "usage_type": "name"}, {"api_name": "keras.models.Sequential", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 64, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.layers.Convolution2D", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 74, "usage_type": "call"}, {"api_name": "keras.utils.vis_utils.plot_model", "line_number": 78, "usage_type": "call"}]} +{"seq_id": "640456995", "text": "# /us/bin/python3\n# -*- coding:utf-8 -*-\n# @Author:baixiaoling\n# @Time:2018/12/02\n# comments: 使用glove词嵌入的文本分类\n# imdb原始数据:acllmdb: 下载地址http://mng.bz/0tIo\n# glove词典下载地址:https://nlp.stanford.edu/projects/glove/ glove.6B.zip\n\nimport os\n\nimport numpy as np\nfrom keras.preprocessing.text import Tokenizer\nfrom keras.preprocessing.sequence import pad_sequences\nfrom keras.models import Sequential\nfrom keras.layers import Embedding, Flatten, Dense\nimport matplotlib.pyplot as plt\nimport win_unicode_console\nwin_unicode_console.enable()\n\n\ndef load_data():\n \"\"\"\n 数据预处理\n \"\"\"\n imdb_dir = 'aclImdb'\n train_dir = os.path.join(imdb_dir, 'train')\n labels = []\n texts = []\n for label_type in ['neg', 'pos']:\n dir_name = os.path.join(train_dir, label_type)\n for fname in os.listdir(dir_name):\n if fname.endswith(\".txt\"):\n try:\n f = open(os.path.join(dir_name, fname))\n # print(os.path.join(dir_name, fname))\n # print(f.read())\n texts.append(f.read())\n f.close()\n if label_type == \"neg\":\n labels.append(0)\n else:\n labels.append(1)\n except Exception as e:\n pass\n return texts, labels\n\n\ndef data_handle(texts, labels, maxlen, traning_samples, validation_samples, max_words):\n \"\"\"\n \"\"\"\n tokenizer = Tokenizer(num_words=max_words)\n\n tokenizer.fit_on_texts(texts)\n\n sequences = tokenizer.texts_to_sequences(texts)\n\n word_index = tokenizer.word_index\n print('Found %s unique tokens.' % len(word_index))\n data = pad_sequences(sequences, maxlen=maxlen)\n labels = np.asarray(labels)\n print(\"data shape:\", data.shape[0])\n print(\"label shape:\", labels.shape[0])\n\n indices = np.arange(data.shape[0])\n np.random.shuffle(indices)\n data = data[indices]\n labels = labels[indices]\n print('data size:', len(data))\n\n x_train = data[:traning_samples]\n y_train = labels[:traning_samples]\n x_val = data[traning_samples:traning_samples + validation_samples]\n y_val = labels[traning_samples:traning_samples + validation_samples]\n\n print(\"train size[%d], val_sizes[%d]!\" % (len(x_train), len(x_val)))\n return x_train, y_train, x_val, y_val, word_index\n\n\ndef load_glove(max_words, word_index):\n \"\"\"\n \"\"\"\n glove_dir = 'glove.6B'\n embedding_index = {}\n\n with open(os.path.join(glove_dir, \"glove.6B.100d.txt\"), 'rb') as f:\n for line in f:\n try:\n line_arr = line.split()\n word = line_arr[0]\n glove_vec = np.asarray(line_arr[1:], dtype='float32')\n embedding_index[word] = glove_vec\n except Exception as e:\n pass\n print(\"Found %s word vectors!\" % len(embedding_index))\n\n embedding_dim = 100\n\n embedding_matrix = np.zeros((max_words, embedding_dim))\n for word, i in word_index.items():\n if i < max_words:\n embedding_vector = embedding_index.get(word)\n if embedding_vector is not None:\n embedding_matrix[i] = embedding_vector\n return embedding_matrix\n\n\ndef model_func(max_words, embedding_dim, mnaxlen, embedding_matrix):\n \"\"\"\n model部分\n \"\"\"\n model = Sequential()\n model.add(Embedding(max_words, embedding_dim, input_length=mnaxlen))\n model.add(Flatten())\n model.add(Dense(32, activation='relu'))\n model.add(Dense(1, activation='sigmoid'))\n print(model.summary())\n model.layers[0].set_weights([embedding_matrix]) # 冻结嵌入层\n model.layers[0].trainable = False\n return model\n\n\ndef plot(history):\n \"\"\"\n 绘制学习曲线\n \"\"\"\n acc = history.history['acc']\n val_acc = history.history['val_acc']\n loss = history.history['loss']\n val_loss = history.history['val_loss']\n\n epochs = range(1, len(acc) + 1)\n plt.plot(epochs, acc, 'bo', label='train acc')\n plt.plot(epochs, val_acc, 'b', label='val acc')\n plt.title(' train and val accuracy')\n plt.legend()\n plt.figure()\n\n plt.plot(epochs, loss, 'bo', label='Training loss')\n plt.plot(epochs, val_loss, 'b', label='Validation loss')\n plt.title('Training and validation loss')\n plt.legend()\n plt.show()\n\n\nif __name__ == '__main__':\n \"\"\"\n \"\"\"\n # 1、加载数据\n texts, labels = load_data()\n\n maxlen = 100 # 评论超过100 截断\n traning_samples = 10000 # 训练集大小\n validation_samples = 2000 # 验证集大小\n max_words = 10000 # 只考虑前10000个常见单词\n embedding_dim = 100\n x_train, y_train, x_val, y_val, word_index = data_handle(\n texts, labels, maxlen, traning_samples, validation_samples, max_words)\n embedding_matrix = load_glove(max_words, word_index)\n\n model = model_func(max_words, embedding_dim, maxlen, embedding_matrix)\n model.compile(optimizer='rmsprop',\n loss='binary_crossentropy', metrics=['acc'])\n history = model.fit(x_train, y_train, epochs=10,\n batch_size=32, validation_data=(x_val, y_val))\n model.save_weights('pre_train_glove_mode.h5')\n\n plot(history)\n", "sub_path": "code/keras/text_dnn_imdb_glove.py", "file_name": "text_dnn_imdb_glove.py", "file_ext": "py", "file_size_in_byte": 5275, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "win_unicode_console.enable", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 26, "usage_type": "call"}, {"api_name": "os.path", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "keras.preprocessing.text.Tokenizer", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.preprocessing.sequence.pad_sequences", "line_number": 59, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 60, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random.shuffle", "line_number": 65, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 85, "usage_type": "call"}, {"api_name": "os.path", "line_number": 85, "usage_type": "attribute"}, {"api_name": "numpy.asarray", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 98, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 111, "usage_type": "call"}, {"api_name": "keras.layers.Embedding", "line_number": 112, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 113, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 114, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 115, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 138, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 138, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 139, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 139, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 140, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 140, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 141, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 141, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 142, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 142, "usage_type": "name"}]} +{"seq_id": "648185572", "text": "import numpy as np\nimport random\nimport torch\nimport torch.nn.functional as F\nfrom torch.distributions import Categorical\n\n\nclass Policy(torch.nn.Module):\n def __init__(self, action_space, input_dimension):\n super().__init__()\n self.hidden = 512\n self.fc1 = torch.nn.Linear(input_dimension*2, self.hidden)\n self.fc2 = torch.nn.Linear(self.hidden, action_space)\n # self.init_weights()\n\n def init_weights(self):\n for m in self.modules():\n if type(m) is torch.nn.Linear:\n torch.nn.init.xavier_normal_(m.weight)\n torch.nn.init.zeros_(m.bias)\n\n def forward(self, x):\n x = self.fc1(x)\n x = F.relu(x)\n x = self.fc2(x)\n return x\n\n\nclass Policy3FC(torch.nn.Module):\n def __init__(self, action_space, input_dimension):\n super().__init__()\n self.hidden1 = 512\n self.hidden2 = 64\n self.fc1 = torch.nn.Linear(input_dimension*2, self.hidden1)\n self.fc2 = torch.nn.Linear(self.hidden1, self.hidden2)\n self.fc3 = torch.nn.Linear(self.hidden2, action_space)\n # self.init_weights()\n\n def init_weights(self):\n for m in self.modules():\n if type(m) is torch.nn.Linear:\n torch.nn.init.xavier_normal_(m.weight)\n torch.nn.init.zeros_(m.bias)\n\n def forward(self, x):\n x = self.fc1(x)\n x = F.relu(x)\n x = self.fc2(x)\n x = F.relu(x)\n x = self.fc3(x)\n return x\n\n\nclass PolicyConv(torch.nn.Module):\n def __init__(self, action_space, hidden=64):\n super().__init__()\n self.action_space = action_space\n self.hidden = hidden # TODO: 64 or 128?\n self.conv1 = torch.nn.Conv2d(2, 32, 3, 2)\n self.conv2 = torch.nn.Conv2d(32, 64, 3, 2)\n self.conv3 = torch.nn.Conv2d(64, 128, 3, 2)\n self.reshaped_size = 128 * 11 * 11\n self.fc1 = torch.nn.Linear(self.reshaped_size, self.hidden)\n self.fc2 = torch.nn.Linear(self.hidden, action_space)\n # self.init_weights()\n\n def init_weights(self):\n for m in self.modules():\n if type(m) is torch.nn.Linear:\n torch.nn.init.uniform_(m.weight)\n torch.nn.init.zeros_(m.bias)\n elif type(m) is torch.nn.Conv2d:\n torch.nn.init.xavier_normal_(m.weight.data)\n if m.bias is not None:\n torch.nn.init.normal_(m.bias.data)\n\n def forward(self, x):\n x = self.conv1(x)\n x = F.relu(x)\n x = self.conv2(x)\n x = F.relu(x)\n x = self.conv3(x)\n x = F.relu(x)\n\n x = x.reshape(-1, self.reshaped_size)\n x = self.fc1(x)\n x = F.relu(x)\n x = self.fc2(x)\n\n return x\n\n\nclass Agent(object):\n def __init__(self, train_device=\"cpu\"):\n self.train_device = train_device\n self.input_dimension = 100 * 100 # downsampled by 2 -> 100x100 grid\n self.action_space = 3\n self.policy = Policy(self.action_space, self.input_dimension).to(self.train_device)\n # self.policy = PolicyConv(self.action_space, 128).to(self.train_device)\n self.optimizer = torch.optim.Adam(self.policy.parameters(), lr=1e-3)\n self.gamma = 0.99\n self.eps_clip = 0.1\n self.prev_obs = None\n self.perc_minibatch = 0.7\n self.name = \"** DaBomb3 **\"\n\n def get_action(self, obs, evaluation=True):\n stack_obs = self.preprocess(obs)\n logits = self.policy.forward(stack_obs)\n\n if evaluation:\n action = int(torch.argmax(logits[0]).detach().cpu().numpy()) # TODO: let it random also in test?\n return self.convert_action(action)\n else:\n dist = torch.distributions.Categorical(logits=logits)\n action = int(dist.sample().cpu().numpy()[0])\n action_prob = float(dist.probs[0, action].detach().cpu().numpy())\n return self.convert_action(action), action_prob, stack_obs\n\n def convert_action(self, action):\n return action + 1 if self.action_space == 2 else action\n\n def revert_action_convertion(self, action):\n return action - 1 if self.action_space == 2 else action\n\n def preprocess(self, obs):\n if \"Conv\" not in type(self.policy).__name__:\n obs = obs[::2, ::2, 0] # downsample by factor of 2\n obs[obs == 43] = 0 # erase background (background type 1)\n obs[obs != 0] = 1 # everything else (paddles, ball) just set to 1\n obs = torch.from_numpy(obs.astype(np.float32).ravel()).unsqueeze(0)\n if self.prev_obs is None:\n self.prev_obs = obs\n stack_obs = torch.cat([obs, self.prev_obs], dim=1)\n else:\n obs = obs[::2, ::2].mean(axis=-1)\n obs = np.expand_dims(obs, axis=-1)\n if self.prev_obs is None:\n self.prev_obs = obs\n stack_obs = np.concatenate((self.prev_obs, obs), axis=-1)\n stack_obs = torch.from_numpy(stack_obs).float().unsqueeze(0)\n stack_obs = stack_obs.transpose(1, 3)\n\n self.prev_obs = obs\n return stack_obs.to(self.train_device)\n\n def discount_rewards(self, reward_history):\n R = 0\n discounted_rewards = []\n for r in reward_history[::-1]:\n if r != 0:\n R = 0 # scored/lost a point in pong, so reset reward sum\n R = r + self.gamma * R\n discounted_rewards.insert(0, R)\n return torch.FloatTensor(discounted_rewards)\n\n def update_policy(self, d_obs_history, action_history, action_prob_history, reward_history):\n # Compute discounted rewards and normalize\n discounted_rewards = self.discount_rewards(reward_history)\n discounted_rewards = (discounted_rewards - discounted_rewards.mean()) / discounted_rewards.std()\n\n for _ in range(5): # TODO: check ideal number of updates\n n_batch = int(self.perc_minibatch * len(action_history)) # TODO: check ideal batch size\n idxs = random.sample(range(len(action_history)), n_batch)\n d_obs_batch = torch.cat([d_obs_history[idx] for idx in idxs], 0).to(self.train_device)\n action_batch = torch.LongTensor([action_history[idx] for idx in idxs]).to(self.train_device)\n action_prob_batch = torch.FloatTensor([action_prob_history[idx] for idx in idxs]).to(self.train_device)\n advantage_batch = torch.FloatTensor([discounted_rewards[idx] for idx in idxs]).to(self.train_device)\n # advantage_batch = (advantage_batch - advantage_batch.mean()) / advantage_batch.std()\n\n self.optimizer.zero_grad()\n vs = np.identity(self.action_space)\n ts = torch.FloatTensor(vs[action_batch.cpu().numpy()]).to(self.train_device)\n logits = self.policy.forward(d_obs_batch)\n r = torch.sum(F.softmax(logits, dim=1) * ts, dim=1) / action_prob_batch\n loss1 = r * advantage_batch\n loss2 = torch.clamp(r, 1 - self.eps_clip, 1 + self.eps_clip) * advantage_batch\n loss = -torch.min(loss1, loss2)\n loss = torch.mean(loss)\n loss.backward()\n self.optimizer.step()\n\n def reset(self):\n self.prev_obs = None\n\n def get_name(self):\n return self.name\n\n def load_model(self, name=None, evaluation=True):\n # name_file = \"{}.mdl\".format(self.name if name is None else name)\n name_file = \"model.mdl\"\n weights = torch.load(name_file, map_location=torch.device(self.train_device))\n self.policy.load_state_dict(weights, strict=False)\n if evaluation:\n self.policy.eval()\n\n def save_model(self, iteration=-1):\n hundreds_iterations = (iteration // 100) * 100\n torch.save(self.policy.state_dict(), \"{}_{}.mdl\".format(self.name, hundreds_iterations))\n # TODO: is it enough saving just the state dict? What about the optimizer?\n # https://stackoverflow.com/questions/42703500/best-way-to-save-a-trained-model-in-pytorch\n", "sub_path": "test_agents/PPOAgent_3act/agent.py", "file_name": "agent.py", "file_ext": "py", "file_size_in_byte": 8048, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "torch.nn", "line_number": 8, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 12, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 12, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 13, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 13, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 18, "usage_type": "attribute"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 19, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 19, "usage_type": "attribute"}, {"api_name": "torch.nn.init.zeros_", "line_number": 20, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 20, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.relu", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 24, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 34, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 35, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 35, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 36, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 36, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 41, "usage_type": "attribute"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 42, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 42, "usage_type": "attribute"}, {"api_name": "torch.nn.init.zeros_", "line_number": 43, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 43, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.relu", "line_number": 47, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 47, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 49, "usage_type": "name"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.nn.Conv2d", "line_number": 61, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 63, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 63, "usage_type": "attribute"}, {"api_name": "torch.nn.Linear", "line_number": 64, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 64, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 69, "usage_type": "attribute"}, {"api_name": "torch.nn.init.uniform_", "line_number": 70, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 70, "usage_type": "attribute"}, {"api_name": "torch.nn.init.zeros_", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 71, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 72, "usage_type": "attribute"}, {"api_name": "torch.nn.init.xavier_normal_", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 73, "usage_type": "attribute"}, {"api_name": "torch.nn.init.normal_", "line_number": 75, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 75, "usage_type": "attribute"}, {"api_name": "torch.nn.functional.relu", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 79, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 81, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 81, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 83, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 87, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 87, "usage_type": "name"}, {"api_name": "torch.optim.Adam", "line_number": 100, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 100, "usage_type": "attribute"}, {"api_name": "torch.argmax", "line_number": 112, "usage_type": "call"}, {"api_name": "torch.distributions.Categorical", "line_number": 115, "usage_type": "call"}, {"api_name": "torch.distributions", "line_number": 115, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 131, "usage_type": "attribute"}, {"api_name": "torch.cat", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 155, "usage_type": "call"}, {"api_name": "random.sample", "line_number": 164, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 165, "usage_type": "call"}, {"api_name": "torch.LongTensor", "line_number": 166, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 167, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.identity", "line_number": 172, "usage_type": "call"}, {"api_name": "torch.FloatTensor", "line_number": 173, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.nn.functional.softmax", "line_number": 175, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 175, "usage_type": "name"}, {"api_name": "torch.clamp", "line_number": 177, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 178, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 179, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.device", "line_number": 192, "usage_type": "call"}, {"api_name": "torch.save", "line_number": 199, "usage_type": "call"}]} +{"seq_id": "571599184", "text": "from flask import Blueprint\nfrom flask import render_template\n\nfrom models import Post, Tag\n\nposts = Blueprint('posts', __name__, template_folder = 'templates')\n\n@posts.route('/')\ndef index():\n posts = Post.query.all()\n return render_template('posts/index.html', posts = posts)\n\n# сделаем из title ссылки для перехода на тело posts\n@posts.route('/')\ndef post_detail(slug):\n post = Post.query.filter(Post.slug == slug).first()\n tags = post.tags\n return render_template('posts/post_detail.html', post = post, tags = tags)\n\n@posts.route('/tag/')\ndef tag_detail(slug):\n tag = Tag.query.filter(Tag.slug == slug).first()\n posts = tag.posts.all()\n return render_template('posts/tag_detail.html', tag = tag, posts = posts)\n", "sub_path": "posts/blueprint.py", "file_name": "blueprint.py", "file_ext": "py", "file_size_in_byte": 783, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Blueprint", "line_number": 6, "usage_type": "call"}, {"api_name": "models.Post.query.all", "line_number": 10, "usage_type": "call"}, {"api_name": "models.Post.query", "line_number": 10, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 10, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 11, "usage_type": "call"}, {"api_name": "models.Post.query.filter", "line_number": 16, "usage_type": "call"}, {"api_name": "models.Post.query", "line_number": 16, "usage_type": "attribute"}, {"api_name": "models.Post", "line_number": 16, "usage_type": "name"}, {"api_name": "models.Post.slug", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 18, "usage_type": "call"}, {"api_name": "models.Tag.query.filter", "line_number": 22, "usage_type": "call"}, {"api_name": "models.Tag.query", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.Tag", "line_number": 22, "usage_type": "name"}, {"api_name": "models.Tag.slug", "line_number": 22, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "97873797", "text": "# -*- coding: utf-8 -*-\n\nimport os\nimport pickle\nfrom PIL import Image, ImageDraw, ImageFont\nimport random\nimport numpy\nfrom keras.models import Model, load_model\nfrom keras.layers import Input, Dense, Dropout, Conv2D, MaxPooling2D, AveragePooling2D, GlobalAveragePooling2D, PReLU, BatchNormalization, concatenate\nfrom custom_layers import GlobalStandardPooling2D\nfrom keras.utils import np_utils\n\n\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\n\n\nfonts_dir = 'fonts/'\nraw_img_dir = 'raw_img/'\ncharacters = []\nfonts_name = []\n\nBATCH = 256\n\n\n# read characters\ncharacters = open('characters.txt', 'r',\n encoding='utf-8').read()\ncharacters = list(characters)\nrandom.shuffle(characters)\nchar_num = len(characters)\nepoch_num = char_num // BATCH + 1\nprint('succeeded: reading characters')\nprint(epoch_num)\n\n\n# read fonts name\nfor i, font_name in enumerate([name for name in os.listdir(fonts_dir) if name[0] != '.']):\n # output number of fonts\n fonts_num = i + 1\n print(str(i), font_name.replace('.ttf', ''))\n fonts_name.append(font_name)\nwith open('fonts_name.dat', 'rb+') as f:\n pickle.dump(fonts_name, f)\nprint('succeeded: reading fonts name')\n\n\n# the model\ninput = Input(shape=(128, 128, 1), name='input')\nx = Conv2D(input_shape=(128, 128, 1),\n filters=16,\n kernel_size=3,\n padding='same')(input)\nx = PReLU()(x)\nx = MaxPooling2D(pool_size=2)(x)\nx = BatchNormalization()(x)\nx = Conv2D(filters=32,\n kernel_size=3,\n padding='same')(x)\nx = PReLU()(x)\nx = MaxPooling2D(pool_size=2)(x)\nx = BatchNormalization()(x)\nx = Conv2D(filters=64,\n kernel_size=3,\n padding='same')(x)\nx = PReLU()(x)\nx = MaxPooling2D(pool_size=2)(x)\nx = BatchNormalization()(x)\nx = AveragePooling2D(pool_size=2)(x)\nx = BatchNormalization()(x)\navgpool = GlobalAveragePooling2D()(x)\nstdpool = GlobalStandardPooling2D()(x)\nx = concatenate([avgpool, stdpool])\nx = Dense(units=128,\n kernel_initializer='random_normal')(x)\nx = PReLU()(x)\nx = BatchNormalization()(x)\noutput = Dense(units=fonts_num,\n kernel_initializer='random_normal',\n activation='softmax',\n name='output')(x)\nmodel = Model(inputs=input, outputs=output)\n\n# print model\nprint(model.summary())\n\n# compile model\nmodel.compile(\n loss='categorical_crossentropy',\n optimizer='adam',\n metrics=['accuracy']\n)\n\n\nfor epoch in range(epoch_num):\n if os.path.exists('model_data/style_discriminator.h5'):\n model = load_model('model_data/style_discriminator.h5')\n #\n raw_imgs = []\n labels = []\n if epoch == epoch_num:\n characters_using = characters\n else:\n characters_using = characters[:BATCH - 1]\n characters = characters[BATCH:]\n\n #\n for i, font_name in enumerate(fonts_name):\n # read font by using truetype\n font = ImageFont.truetype(fonts_dir + font_name, 96)\n for character in characters_using:\n # create an img\n img = Image.new(mode='L', size=(128, 128), color=255)\n draw = ImageDraw.Draw(img)\n # make the font drawn on center\n text_size = draw.textsize(character, font)\n text_w = text_size[0]\n text_h = text_size[1]\n draw.text((64 - text_w / 2, 64 - text_h / 2),\n character, font=font, fill=0)\n\n raw_imgs.append(list(img.getdata()))\n labels.append(i)\n\n # randomize dataset\n dataset = list(zip(raw_imgs, labels))\n random.shuffle(dataset)\n raw_imgs[:], labels[:] = zip(*dataset)\n print('succeeded: randomizing data set')\n\n # tranfer the dataset into a numpy array\n raw_imgs = numpy.array(raw_imgs)\n raw_imgs = raw_imgs.reshape(\n raw_imgs.shape[0], 128, 128, 1).astype('float32') / 255\n labels = numpy.array(labels)\n labels = np_utils.to_categorical(labels)\n\n # train model\n history = model.fit(x=raw_imgs,\n y=labels,\n validation_split=0.2,\n initial_epoch=epoch * 100,\n epochs=100,\n batch_size=128,\n verbose=2)\n\n # save model\n model.save('model_data/style_discriminator.h5')\n", "sub_path": "style_discrimination.py", "file_name": "style_discrimination.py", "file_ext": "py", "file_size_in_byte": 4235, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.environ", "line_number": 14, "usage_type": "attribute"}, {"api_name": "random.shuffle", "line_number": 29, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 37, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 43, "usage_type": "call"}, {"api_name": "keras.layers.Input", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.PReLU", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 55, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers.PReLU", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 60, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 61, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 62, "usage_type": "call"}, {"api_name": "keras.layers.PReLU", "line_number": 65, "usage_type": "call"}, {"api_name": "keras.layers.MaxPooling2D", "line_number": 66, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 67, "usage_type": "call"}, {"api_name": "keras.layers.AveragePooling2D", "line_number": 68, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 69, "usage_type": "call"}, {"api_name": "keras.layers.GlobalAveragePooling2D", "line_number": 70, "usage_type": "call"}, {"api_name": "custom_layers.GlobalStandardPooling2D", "line_number": 71, "usage_type": "call"}, {"api_name": "keras.layers.concatenate", "line_number": 72, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 73, "usage_type": "call"}, {"api_name": "keras.layers.PReLU", "line_number": 75, "usage_type": "call"}, {"api_name": "keras.layers.BatchNormalization", "line_number": 76, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 77, "usage_type": "call"}, {"api_name": "keras.models.Model", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 95, "usage_type": "call"}, {"api_name": "os.path", "line_number": 95, "usage_type": "attribute"}, {"api_name": "keras.models.load_model", "line_number": 96, "usage_type": "call"}, {"api_name": "PIL.ImageFont.truetype", "line_number": 109, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 109, "usage_type": "name"}, {"api_name": "PIL.Image.new", "line_number": 112, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 112, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 113, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 113, "usage_type": "name"}, {"api_name": "random.shuffle", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 134, "usage_type": "call"}, {"api_name": "keras.utils.np_utils.to_categorical", "line_number": 135, "usage_type": "call"}, {"api_name": "keras.utils.np_utils", "line_number": 135, "usage_type": "name"}]} +{"seq_id": "566373409", "text": "#! /usr/bin/env python\nfrom flask import Flask, render_template,g, request, redirect, url_for\nfrom redis import Redis\nfrom flask_sqlalchemy import SQLAlchemy # 変更\nfrom hamlish_jinja import HamlishExtension\nfrom werkzeug import ImmutableDict\nimport os\n\n\n\nclass FlaskWithHamlish(Flask):\n jinja_options = ImmutableDict(\n extensions=[HamlishExtension]\n )\napp = FlaskWithHamlish(__name__)\ndb_uri = os.environ.get('DATABASE_URL') or \"postgresql://python:triple4649@pythonpostgres/python\"\napp.config['SQLALCHEMY_DATABASE_URI'] = db_uri # 追加\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = True\napp.config['testing'] = True\ndb = SQLAlchemy(app) # 追加\nredis = Redis(host='redis', port=6379)\n\n@app.route('/')\ndef hello():\n redis.incr('hits')\n return 'Hello World! I have been seen %s times.' % redis.get('hits')\n@app.route('/test')\ndef hello_world():\n return render_template('view/index.html') # 変更\n\n@app.route('/regist_image')\ndef show_regist_image():\n return render_template('view/bas64.html') # 変更\n \n \n@app.route('/imageentry', methods=['POST'])\ndef add_entry():\n from logic.mywebapp import add_Image_toDB\n add_Image_toDB(request.form)\n return redirect(url_for('show_regist_image'))\n\n\n@app.route('/wasource')\ndef get_wasresource():\n import urllib\n from logic.wasdata import add_Word_toDB\n querry=urllib.parse.urlencode({\"path\":\"/pdf/xml/2017h29h_sc_am2_qs.pdf.xml\"})\n url ='http://mywebapp_web_1:9080/MyWebApp/sample/Area?{0}'.format(querry)\n import urllib.request\n with urllib.request.urlopen(url) as response:\n body = response.read().decode('utf-8')\n return render_template('view/was.haml',resouce=body)\n \nif __name__ == '__main__':\n #init()\n app.run(host='0.0.0.0',debug=True)\n\n", "sub_path": "app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.Flask", "line_number": 11, "usage_type": "name"}, {"api_name": "werkzeug.ImmutableDict", "line_number": 12, "usage_type": "call"}, {"api_name": "hamlish_jinja.HamlishExtension", "line_number": 13, "usage_type": "name"}, {"api_name": "os.environ.get", "line_number": 16, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 16, "usage_type": "attribute"}, {"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 20, "usage_type": "call"}, {"api_name": "redis.Redis", "line_number": 21, "usage_type": "call"}, {"api_name": "redis.incr", "line_number": 25, "usage_type": "call"}, {"api_name": "redis.get", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 33, "usage_type": "call"}, {"api_name": "logic.mywebapp.add_Image_toDB", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.request.form", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "flask.redirect", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 40, "usage_type": "call"}, {"api_name": "urllib.parse.urlencode", "line_number": 47, "usage_type": "call"}, {"api_name": "urllib.parse", "line_number": 47, "usage_type": "attribute"}, {"api_name": "urllib.request.urlopen", "line_number": 50, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 50, "usage_type": "attribute"}, {"api_name": "flask.render_template", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "505041170", "text": "# Try to add seasonal change for Minecraft resource packs.\n# Requires pillow(PIL)\n\nimport os\nimport json\nfrom PIL import Image\n\nlang = 'en'\navailable_list = ['acacia_leaves', 'birch_leaves',\n 'dark_oak_leaves', 'jungle_leaves', 'oak_leaves', 'spruce_leaves']\ndefault_color = {'acacia_leaves': (174, 164, 42, 255), 'birch_leaves': (26, 191, 0, 255), 'dark_oak_leaves': (\n 26, 191, 0, 255), 'jungle_leaves': (26, 191, 0, 255), 'oak_leaves': (26, 191, 0, 255), 'spruce_leaves': (96, 161, 123, 255)}\nt_r, t_g, t_b, t_a = target_color = (255, 192, 0, 255)\ntranslation = {'zh': {\"Fuck! \": \"你🐴死了\"}}\n\n# Seasonal leaves frames\nframes = 24\n\n# example root_dir: 'G:\\sjfhsjfh\\hmcl'\ndir = ''\ndir_list = []\n\n# Function definition.\n\n# language choosing.\n\n\ndef trans(text: str) -> str:\n \"\"\"Translation. \"\"\"\n global lang, translation\n if lang != 'en':\n try:\n dictionary = translation[lang]\n except:\n print(\"* Translation Error\")\n print(\"* Language \", lang, \" not found, using en instead. \")\n lang = 'en'\n return text\n try:\n text = dictionary[text]\n except:\n pass\n return text\n\n# Img processing & .mcmeta writing\n\n\ndef convert(img_dir: str, tint: bool = False) -> None:\n \"\"\"Processing the .png image only. \"\"\"\n global frames, default_color, t_r, t_g, t_b, t_a\n\n # Img part\n if os.path.isfile(img_dir):\n pname, fname = os.path.split(img_dir)\n name, ename = os.path.splitext(fname)\n try:\n img = Image.open(img_dir)\n except:\n print(trans(\"* Unable to open image:\"), img_dir)\n exit()\n # Save the raw image\n img.save(os.path.join(pname, 'raaaaaw_' + name + ename))\n w, h = img.size\n ans_size = (w, h * 2 * (frames - 1))\n ans = Image.new('RGBA', ans_size, '#00000000')\n raw_pix = img.load()\n\n # Detect if black-and-white\n for x in range(0, w):\n for y in range(0, h):\n r_r, r_g, r_b, r_a = raw_pix[x, y]\n if abs(r_r - r_g) > 1 or abs(r_r - r_b) > 1 or abs(r_g - r_b) > 1:\n tint = True\n break\n\n # Forced coloring (only green now)\n if tint == False:\n for x in range(0, w):\n for y in range(0, h):\n r_r, r_g, r_b, r_a = raw_pix[x, y]\n if r_a > 0:\n r_r, r_g, r_b, r_a = default_color[name]\n raw_pix[x, y] = (r_r, r_g, r_b, r_a)\n\n # Frames\n for i in range(0, frames):\n frame = img.copy()\n frame_pix = frame.load()\n\n # COLOR!!!!!!!!!!\n for x in range(0, w):\n for y in range(0, h):\n r, g, b, a = frame_pix[x, y]\n r_r, r_g, r_b, r_a = raw_pix[x, y]\n if r_a == 255:\n r = round(r_r + (i / (frames - 1)) * (t_r - r_r))\n g = round(r_g + (i / (frames - 1)) * (t_g - r_g))\n b = round(r_b + (i / (frames - 1)) * (t_b - r_b))\n frame_pix[x, y] = (r, g, b, a)\n\n # Paste the frame on the ans img\n ans.paste(frame, (0, i * h))\n if i != 1:\n ans.paste(frame, (0, (2 * frames - 1 - i) * h))\n\n # Save the output\n ans.save(img_dir)\n\n # If no, use default texture(generated)\n else:\n pname, fname = os.path.split(img_dir)\n name, ename = os.path.splitext(fname)\n pydir = os.path.split(__file__)[0]\n print(trans(\"* No texture found for\"), name,\n trans(\", trying to use the generated texture. \"))\n try:\n ans = Image.open(os.path.join(\n pydir, 'default_texture', 'generated', fname))\n except:\n print(trans(\"* No generated texture found for\"), name,\n trans(\". Please make sure the tool is downloaded completely. \"))\n exit()\n ans.save(img_dir)\n\n # mcmeta part\n try:\n with open(os.path.join(pname, fname + '.mcmeta'), 'r') as mcmeta_file:\n raw_mcmeta = mcmeta_file.read()\n # Later to support animated textures\n #mcmeta = json.loads(raw_mcmeta)\n # mcmeta[\"animation\"]\n with open(os.path.join(pname, 'raaaaaw_' + fname + '.mcmeta'), 'w') as backup:\n backup.write(raw_mcmeta)\n except:\n with open(os.path.join(pname, fname + '.mcmeta'), 'w') as mcmeta_file:\n mcmeta = {'animation': {'frametime': round(4383000 / frames)}}\n mcmeta = json.dumps(mcmeta, ensure_ascii=False, indent=4)\n mcmeta_file.write(mcmeta)\n\n\n# Main\n# Get and enter the resourcepacks folder in .minecraft folder\nok = False\nwhile ok == False:\n try:\n dir_get = input(\n trans(\"* Please enter your .minecraft folder directory: \\n\"))\n dir_list = os.listdir(dir_get)\n ok = True\n except:\n print(trans(\"* The path is not available. \\n\"))\n if '.minecraft' in dir_list:\n dir = os.path.join(dir_get, '.minecraft', 'resourcepacks')\n elif 'resourcepacks' in dir_list:\n dir = os.path.join(dir_get, 'resourcepacks')\n if os.path.isdir(dir):\n ok = True\n else:\n ok = False\n print(trans(\"* No 'resourcepacks' folder found in '.minecraft' folder. \"))\ndir_list = os.listdir(dir)\n\n# Try to find resource packs\nresourcepacks = []\nfor f in dir_list:\n if os.path.isdir(os.path.join(dir, f)):\n print(trans(\"* Resource pack < \") + f + trans(\" > detected. \"))\n resourcepacks.append(f)\n\n# If no resource packs found\nif len(resourcepacks) == 0:\n print(trans(\"* No resource packs found in 'resourcepacks' folder. \\n\"))\n quit()\n\n# Choose a resource pack\nok = False\nwhile ok == False:\n target_resourcepack = input(\n trans(\"* Please choose a resource pack: (Name is case sensitive)\\n\"))\n if target_resourcepack in resourcepacks:\n ok = True\n else:\n print(trans(\"* Resource pack < \") +\n target_resourcepack + trans(\" > not found. \"))\n\n# Enter the resource pack\ndir = os.path.join(dir, target_resourcepack, 'assets', 'minecraft')\n\n# Edit the model json file\ntry:\n with open(os.path.join(dir, 'models', 'leaves.json'), 'r') as leaves:\n setting = leaves.read()\n setting = json.loads(setting)\n for face in ['down', 'up', 'north', 'south', 'west', 'east']:\n for i in range(len(setting['elements'])):\n try:\n del setting['elements'][i]['faces'][face]['tintindex']\n except:\n pass\n setting = json.dumps(setting)\n\n# Add one according to the default texture\nexcept:\n with open(os.path.join(os.path.split(__file__)[0], 'default_texture', 'leaves.json'), 'r') as leaves:\n setting = leaves.read()\n\n# Write the model json file\nwith open(os.path.join(dir, 'models', 'leaves.json'), 'w') as leaves:\n leaves.write(setting)\n\n# Process the leaves\nfor leaf in available_list:\n convert(os.path.join(dir, 'textures', 'block', leaf + '.png'))\n print(trans(\"* Processed texture < \") + leaf +\n '.png' + trans(\" > successfully. \"))\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 7249, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.isfile", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path", "line_number": 53, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 55, "usage_type": "call"}, {"api_name": "os.path", "line_number": 55, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 57, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 57, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "PIL.Image.new", "line_number": 65, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 65, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 111, "usage_type": "call"}, {"api_name": "os.path", "line_number": 111, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 112, "usage_type": "call"}, {"api_name": "os.path", "line_number": 112, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 113, "usage_type": "call"}, {"api_name": "os.path", "line_number": 113, "usage_type": "attribute"}, {"api_name": "PIL.Image.open", "line_number": 117, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 117, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 117, "usage_type": "call"}, {"api_name": "os.path", "line_number": 117, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path", "line_number": 132, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 135, "usage_type": "call"}, {"api_name": "os.path", "line_number": 135, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 137, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 156, "usage_type": "call"}, {"api_name": "os.path", "line_number": 156, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 161, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path", "line_number": 166, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 166, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 187, "usage_type": "call"}, {"api_name": "os.path", "line_number": 187, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 191, "usage_type": "call"}, {"api_name": "os.path", "line_number": 191, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 193, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 200, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 208, "usage_type": "call"}, {"api_name": "os.path", "line_number": 208, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 213, "usage_type": "call"}, {"api_name": "os.path", "line_number": 213, "usage_type": "attribute"}]} +{"seq_id": "233007377", "text": "# import the necessary packages\nimport numpy as np\nimport cv2\n\ndef getWidthOfTrunk(imagePath):\n\timage = cv2.imread(imagePath)\n\timage=cv2.bilateralFilter(image,9,75,75)\n\n\tlower = np.array([15,0,33])\n\tupper = np.array([30,255,255])\n\tmask = cv2.inRange(image, lower, upper)\n\tres = cv2.bitwise_and(image,image,mask=mask)\n\n\tdescript=cv2.MSER_create(10, 50, 14400, 0.001, .2, 200, 1.01, 0.003, 3)\n\tregions = descript.detectRegions(res, None)\n\n\thulls = [cv2.convexHull(p.reshape(-1, 1, 2)) for p in regions]\n\tcv2.polylines(res, hulls, 1, (0, 0, 255))\n\n\tboxes = [cv2.boundingRect(p) for p in regions]\n\tbestBox = boxes[0]\n\tfor box in boxes:\n\t\tratio = box[3]/box[2]\n\t\tif ratio>(bestBox[3]/bestBox[2]): bestBox=box\n\tx,y,w,h=bestBox\n\timg = cv2.rectangle(res,(x,y),(x+w,y+h),(0,0,255),2)\n\n\tcv2.imshow(\"original\", image)\n\tcv2.imshow(\"processed\", img)\n\tcv2.waitKey(0)\n\n\treturn w\n", "sub_path": "TrunkRecognition.py", "file_name": "TrunkRecognition.py", "file_ext": "py", "file_size_in_byte": 864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cv2.imread", "line_number": 6, "usage_type": "call"}, {"api_name": "cv2.bilateralFilter", "line_number": 7, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 10, "usage_type": "call"}, {"api_name": "cv2.inRange", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.bitwise_and", "line_number": 12, "usage_type": "call"}, {"api_name": "cv2.MSER_create", "line_number": 14, "usage_type": "call"}, {"api_name": "cv2.convexHull", "line_number": 17, "usage_type": "call"}, {"api_name": "cv2.polylines", "line_number": 18, "usage_type": "call"}, {"api_name": "cv2.boundingRect", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.rectangle", "line_number": 26, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 28, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 29, "usage_type": "call"}, {"api_name": "cv2.waitKey", "line_number": 30, "usage_type": "call"}]} +{"seq_id": "171083653", "text": "from __future__ import annotations\n\nfrom binascii import hexlify\nfrom collections import namedtuple\nfrom hashlib import sha256\nimport logging\nimport time\nfrom typing import Any, List\n\nfrom ipv8.keyvault.crypto import default_eccrypto\nfrom ipv8.messaging.serialization import default_serializer, PackError\n\nfrom bami.backbone.datastore.database import BaseDB\nfrom bami.backbone.utils import (\n BytesLinks,\n decode_links,\n Dot,\n EMPTY_PK,\n EMPTY_SIG,\n encode_links,\n GENESIS_DOT,\n GENESIS_LINK,\n GENESIS_SEQ,\n Links,\n shorten,\n UNKNOWN_SEQ,\n)\nfrom bami.backbone.payload import BlockPayload\n\nSKIP_ATTRIBUTES = {\n \"key\",\n \"serializer\",\n \"crypto\",\n \"_logger\",\n \"_previous\",\n \"_links\",\n}\n\n\nclass BamiBlock(object):\n \"\"\"\n Container for Plexus block information\n \"\"\"\n\n Data = namedtuple(\n \"Data\",\n [\n \"type\",\n \"transaction\",\n \"public_key\",\n \"sequence_number\",\n \"previous\",\n \"links\",\n \"com_prefix\",\n \"com_id\",\n \"com_seq_num\",\n \"timestamp\",\n \"insert_time\",\n \"signature\",\n ],\n )\n\n def __init__(self, data: List = None, serializer=default_serializer) -> None:\n \"\"\"\n Create a new PlexusBlock or load from an existing database entry.\n\n :param data: Optional data to initialize this block with.\n :type data: Block.Data or list\n :param serializer: An optional custom serializer to use for this block.\n :type serializer: Serializer\n \"\"\"\n super(BamiBlock, self).__init__()\n self.serializer = serializer\n if data is None:\n # data\n self.type = b\"unknown\"\n self.transaction = b\"\"\n # block identity\n self.public_key = EMPTY_PK\n self.sequence_number = GENESIS_SEQ\n\n # previous hash in the personal chain\n self.previous = GENESIS_LINK\n self._previous = encode_links(self.previous)\n\n # Linked blocks => links to the block in other chains\n self.links = GENESIS_LINK\n self._links = encode_links(self.links)\n\n # Metadata for community identifiers\n self.com_prefix = b\"\"\n self.com_id = EMPTY_PK\n self.com_seq_num: int = UNKNOWN_SEQ\n\n # Creation timestamp\n self.timestamp = int(time.time() * 1000)\n # Signature for the block\n self.signature = EMPTY_SIG\n # debug stuff\n self.insert_time = None\n else:\n self.transaction = data[1] if isinstance(data[1], bytes) else bytes(data[1])\n self._previous = (\n BytesLinks(data[4]) if isinstance(data[4], bytes) else bytes(data[4])\n )\n self._links = (\n BytesLinks(data[5]) if isinstance(data[5], bytes) else bytes(data[5])\n )\n\n self.previous = decode_links(self._previous)\n self.links = decode_links(self._links)\n\n self.type, self.public_key, self.sequence_number = data[0], data[2], data[3]\n self.com_prefix, self.com_id, self.com_seq_num = (\n data[6],\n data[7],\n int(data[8]),\n )\n self.signature, self.timestamp, self.insert_time = (\n data[9],\n data[10],\n data[11],\n )\n\n self.type = (\n self.type\n if isinstance(self.type, bytes)\n else str(self.type).encode(\"utf-8\")\n )\n self.public_key = (\n self.public_key\n if isinstance(self.public_key, bytes)\n else bytes(self.public_key)\n )\n self.signature = (\n self.signature\n if isinstance(self.signature, bytes)\n else bytes(self.signature)\n )\n\n self.hash = self.calculate_hash()\n self.crypto = default_eccrypto\n self._logger = logging.getLogger(self.__class__.__name__)\n\n def __str__(self):\n # This makes debugging and logging easier\n return \"Block {0} from ...{1}:{2} links {3} for {4} type {5} cseq {6} cid {7}.{8}\".format(\n self.short_hash,\n shorten(self.public_key),\n self.sequence_number,\n self.links,\n self.transaction,\n self.type,\n self.com_seq_num,\n self.com_prefix,\n self.com_id,\n )\n\n @property\n def short_hash(self):\n return shorten(self.hash)\n\n def __hash__(self):\n return self.hash_number\n\n @property\n def pers_dot(self) -> Dot:\n return Dot((self.sequence_number, self.short_hash))\n\n @property\n def com_dot(self) -> Dot:\n return Dot((self.com_seq_num, self.short_hash))\n\n @property\n def hash_number(self):\n \"\"\"\n Return the hash of this block as a number (used as crawl ID).\n \"\"\"\n return int(hexlify(self.hash), 16) % 100000000\n\n def calculate_hash(self) -> bytes:\n return sha256(self.pack()).digest()\n\n def __eq__(self, other: object) -> bool:\n if not isinstance(other, BamiBlock):\n return False\n return self.pack() == other.pack()\n\n @property\n def is_peer_genesis(self) -> bool:\n return self.sequence_number == GENESIS_SEQ and self.previous == GENESIS_LINK\n\n def block_args(self, signature: bool = True) -> List[Any]:\n args = [\n self.type,\n self.transaction,\n self.public_key,\n self.sequence_number,\n self._previous,\n self._links,\n self.com_prefix,\n self.com_id,\n self.com_seq_num,\n self.signature if signature else EMPTY_SIG,\n self.timestamp,\n ]\n return args\n\n def to_block_payload(self, signature: bool = True) -> BlockPayload:\n return BlockPayload(*self.block_args(signature))\n\n def pack(self, signature: bool = True) -> bytes:\n \"\"\"\n Encode the block\n Args:\n signature: False to pack EMPTY_SIG in the signature location, true to pack the signature field\n Returns:\n Block bytes\n \"\"\"\n return self.serializer.pack_multiple(\n self.to_block_payload(signature).to_pack_list()\n )[0]\n\n @classmethod\n def unpack(\n cls, block_blob: bytes, serializer: Any = default_serializer\n ) -> BamiBlock:\n payload = serializer.ez_unpack_serializables([BlockPayload], block_blob)\n return BamiBlock.from_payload(payload[0], serializer)\n\n @classmethod\n def from_payload(\n cls, payload: BlockPayload, serializer=default_serializer\n ) -> BamiBlock:\n \"\"\"\n Create a block according to a given payload and serializer.\n This method can be used when receiving a block from the network.\n \"\"\"\n return cls(\n [\n payload.type,\n payload.transaction,\n payload.public_key,\n payload.sequence_number,\n payload.previous,\n payload.links,\n payload.com_prefix,\n payload.com_id,\n payload.com_seq_num,\n payload.signature,\n payload.timestamp,\n time.time(),\n ],\n serializer,\n )\n\n def sign(self, key):\n \"\"\"\n Signs this block with the given key\n :param key: the key to sign this block with\n \"\"\"\n self.signature = self.crypto.create_signature(key, self.pack(signature=False))\n self.hash = self.calculate_hash()\n\n @classmethod\n def create(\n cls,\n block_type: bytes,\n transaction: bytes,\n database: BaseDB,\n public_key: bytes,\n com_prefix: bytes = b\"\",\n com_id: bytes = None,\n com_links: Links = None,\n pers_links: Links = None,\n use_consistent_links: bool = True,\n ):\n \"\"\"\n Create PlexusBlock wrt local database knowledge.\n\n Args:\n block_type: type of the block in bytes\n transaction: transaction blob bytes\n database: local database with chains\n public_key: public key of the block creator\n com_prefix: prefix for the chain identification [optional]\n com_id: id of the community which block is part of [optional]\n com_links: Explicitly link with these blocks [optional]\n pers_links: Create a block at a certain [optional]\n use_consistent_links: Build on top of blocks that are known. By default: True\n\n Returns:\n PlexusBlock\n\n \"\"\"\n\n if public_key == com_id:\n full_pers_chain_id = com_prefix + public_key\n else:\n full_pers_chain_id = public_key\n personal_chain = database.get_chain(full_pers_chain_id)\n # Decide to link blocks in the personal chain:\n if not personal_chain:\n # There are no blocks in the personal chain yet\n last_link = Links((GENESIS_DOT,))\n else:\n last_link = (\n personal_chain.consistent_terminal\n if use_consistent_links\n else personal_chain.terminal\n )\n\n # Fork personal chain at the\n if pers_links:\n # There is an explicit link for the previous link\n last_link = pers_links\n\n per_seq_num = max(last_link)[0] + 1\n\n # TODO: Add link filtering and choose links\n ret = cls()\n ret.type = block_type\n ret.transaction = transaction\n ret.sequence_number = per_seq_num\n ret.previous = last_link\n\n # --- Community related logic ---\n if com_id:\n ret.com_id = com_id\n # There is community specified => will base block on the latest known information + filters\n if com_links:\n last_com_links = com_links\n com_seq_num = max(last_com_links)[0]\n else:\n com_chain = database.get_chain(com_prefix + com_id)\n if not com_chain:\n last_com_links = Links((GENESIS_DOT,))\n else:\n last_com_links = (\n com_chain.consistent_terminal\n if use_consistent_links\n else com_chain.terminal\n )\n # TODO: add link filtering here\n com_seq_num = max(last_com_links)[0] + 1\n\n ret.links = last_com_links\n ret.com_seq_num = com_seq_num\n ret.com_id = com_id\n ret.com_prefix = com_prefix\n\n ret._links = encode_links(ret.links)\n ret._previous = encode_links(ret.previous)\n\n ret.public_key = public_key\n ret.signature = EMPTY_SIG\n ret.hash = ret.calculate_hash()\n return ret\n\n def block_invariants_valid(self) -> bool:\n \"\"\"Verify that block is valid wrt block invariants\"\"\"\n # 1. Sequence number should not be prior to genesis\n if self.sequence_number < GENESIS_SEQ and self.com_seq_num < GENESIS_SEQ:\n self._logger.error(\"Sequence number wrong\", self.sequence_number)\n return False\n # 2. Timestamp should be non negative\n if self.timestamp < 0:\n self._logger.error(\"Timestamp negative\")\n return False\n # 3. Public key and signature should be valid\n if not self.crypto.is_valid_public_bin(self.public_key):\n self._logger.error(\"Public key is not valid\")\n return False\n else:\n try:\n pck = self.pack(signature=False)\n except PackError:\n pck = None\n if pck is None or not self.crypto.is_valid_signature(\n self.crypto.key_from_public_bin(self.public_key), pck, self.signature\n ):\n self._logger.error(\"Cannot pack the block, or signature is not valid\")\n return False\n return True\n", "sub_path": "src/bami/backbone/block.py", "file_name": "block.py", "file_ext": "py", "file_size_in_byte": 12211, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "collections.namedtuple", "line_number": 45, "usage_type": "call"}, {"api_name": "typing.List", "line_number": 63, "usage_type": "name"}, {"api_name": "ipv8.messaging.serialization.default_serializer", "line_number": 63, "usage_type": "name"}, {"api_name": "bami.backbone.utils.EMPTY_PK", "line_number": 79, "usage_type": "name"}, {"api_name": "bami.backbone.utils.GENESIS_SEQ", "line_number": 80, "usage_type": "name"}, {"api_name": "bami.backbone.utils.GENESIS_LINK", "line_number": 83, "usage_type": "name"}, {"api_name": "bami.backbone.utils.encode_links", "line_number": 84, "usage_type": "call"}, {"api_name": "bami.backbone.utils.GENESIS_LINK", "line_number": 87, "usage_type": "name"}, {"api_name": "bami.backbone.utils.encode_links", "line_number": 88, "usage_type": "call"}, {"api_name": "bami.backbone.utils.EMPTY_PK", "line_number": 92, "usage_type": "name"}, {"api_name": "bami.backbone.utils.UNKNOWN_SEQ", "line_number": 93, "usage_type": "name"}, {"api_name": "time.time", "line_number": 96, "usage_type": "call"}, {"api_name": "bami.backbone.utils.EMPTY_SIG", "line_number": 98, "usage_type": "name"}, {"api_name": "bami.backbone.utils.BytesLinks", "line_number": 104, "usage_type": "call"}, {"api_name": "bami.backbone.utils.BytesLinks", "line_number": 107, "usage_type": "call"}, {"api_name": "bami.backbone.utils.decode_links", "line_number": 110, "usage_type": "call"}, {"api_name": "bami.backbone.utils.decode_links", "line_number": 111, "usage_type": "call"}, {"api_name": "ipv8.keyvault.crypto.default_eccrypto", "line_number": 142, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 143, "usage_type": "call"}, {"api_name": "bami.backbone.utils.shorten", "line_number": 149, "usage_type": "call"}, {"api_name": "bami.backbone.utils.shorten", "line_number": 161, "usage_type": "call"}, {"api_name": "bami.backbone.utils.Dot", "line_number": 168, "usage_type": "call"}, {"api_name": "bami.backbone.utils.Dot", "line_number": 167, "usage_type": "name"}, {"api_name": "bami.backbone.utils.Dot", "line_number": 172, "usage_type": "call"}, {"api_name": "bami.backbone.utils.Dot", "line_number": 171, "usage_type": "name"}, {"api_name": "binascii.hexlify", "line_number": 179, "usage_type": "call"}, {"api_name": "hashlib.sha256", "line_number": 182, "usage_type": "call"}, {"api_name": "bami.backbone.utils.GENESIS_SEQ", "line_number": 191, "usage_type": "name"}, {"api_name": "bami.backbone.utils.GENESIS_LINK", "line_number": 191, "usage_type": "name"}, {"api_name": "bami.backbone.utils.EMPTY_SIG", "line_number": 204, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 193, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 193, "usage_type": "name"}, {"api_name": "bami.backbone.payload.BlockPayload", "line_number": 210, "usage_type": "call"}, {"api_name": "bami.backbone.payload.BlockPayload", "line_number": 209, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 226, "usage_type": "name"}, {"api_name": "ipv8.messaging.serialization.default_serializer", "line_number": 226, "usage_type": "name"}, {"api_name": "bami.backbone.payload.BlockPayload", "line_number": 228, "usage_type": "name"}, {"api_name": "bami.backbone.payload.BlockPayload", "line_number": 233, "usage_type": "name"}, {"api_name": "ipv8.messaging.serialization.default_serializer", "line_number": 233, "usage_type": "name"}, {"api_name": "time.time", "line_number": 252, "usage_type": "call"}, {"api_name": "bami.backbone.datastore.database.BaseDB", "line_number": 270, "usage_type": "name"}, {"api_name": "bami.backbone.utils.Links", "line_number": 274, "usage_type": "name"}, {"api_name": "bami.backbone.utils.Links", "line_number": 275, "usage_type": "name"}, {"api_name": "bami.backbone.utils.Links", "line_number": 305, "usage_type": "call"}, {"api_name": "bami.backbone.utils.GENESIS_DOT", "line_number": 305, "usage_type": "name"}, {"api_name": "bami.backbone.utils.Links", "line_number": 337, "usage_type": "call"}, {"api_name": "bami.backbone.utils.GENESIS_DOT", "line_number": 337, "usage_type": "name"}, {"api_name": "bami.backbone.utils.encode_links", "line_number": 352, "usage_type": "call"}, {"api_name": "bami.backbone.utils.encode_links", "line_number": 353, "usage_type": "call"}, {"api_name": "bami.backbone.utils.EMPTY_SIG", "line_number": 356, "usage_type": "name"}, {"api_name": "bami.backbone.utils.GENESIS_SEQ", "line_number": 363, "usage_type": "name"}, {"api_name": "ipv8.messaging.serialization.PackError", "line_number": 377, "usage_type": "name"}]} +{"seq_id": "296225870", "text": "#imports\nfrom flask import Flask, render_template\nfrom flask_restful import Api\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_bcrypt import Bcrypt\nfrom flask_cors import CORS\nfrom flask_jwt_extended import JWTManager\n\n#defined in global access and configured when app is created\ndb = SQLAlchemy()\nauth_api = Api()\nbcrypt = Bcrypt()\ndb_api = Api()\n\n# app factory function\ndef create_app(config_filename=None):\n #create app and link to react build folder\n app = Flask(\n __name__, \n instance_relative_config=True, \n static_folder='../react_frontend/build/static', \n template_folder='../react_frontend/build')\n \n #configure app with /instance/*.cfg files\n app.config.from_pyfile(config_filename)\n \n #initialise addition extensions\n initialise_extensions(app)\n\n #all routes direct to react application index.html \n @app.route('/', defaults={'path': ''})\n @app.route('/')\n def react(path):\n return render_template('index.html')\n \n return app\n\n# helper function\ndef initialise_extensions(app):\n# initialise extensions with created app\n \n #import and configure api routes\n from flask_backend import auth\n from flask_backend import database\n auth_api.init_app(app)\n db_api.init_app(app)\n\n #configure CORS for development\n CORS(app, resources={\n r'/auth/*': {'origins': '*'},\n r'/db/*':{'origins:': '*'}})\n\n #configure JWTs\n jwt = JWTManager(app)\n\n #handle blacklisting of JWTs\n @jwt.token_in_blacklist_loader\n def check_if_token_in_blacklist(decrypted_token):\n jti = decrypted_token['jti']\n return models.Revoked_Token.is_jti_blacklisted(jti)\n \n #configure password hashing\n bcrypt.init_app(app)\n #import models and configure databse\n import flask_backend.models\n db.init_app(app) \n \n #migrate or seed db\n from flask_backend.seed import seed_db\n with app.app_context():\n #seed - should only be run for development\n # seed_db()\n #migrate\n db.create_all()\n db.create_all(bind=['auth']) ", "sub_path": "flask_backend/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2101, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 10, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 11, "usage_type": "call"}, {"api_name": "flask_bcrypt.Bcrypt", "line_number": 12, "usage_type": "call"}, {"api_name": "flask_restful.Api", "line_number": 13, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 18, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 34, "usage_type": "call"}, {"api_name": "flask_cors.CORS", "line_number": 49, "usage_type": "call"}, {"api_name": "flask_jwt_extended.JWTManager", "line_number": 54, "usage_type": "call"}]} +{"seq_id": "308482559", "text": "#!usr/bin/python3\n\nimport matplotlib\n\nmatplotlib.use(\"agg\")\n\nimport numpy as np\n\nimport matplotlib.pyplot as plt\n\nfil1=open('/home/richard/outhome2/outhome/Fano33sh2.txt',\"r\")\nx1,y1=[],[]\nfor k in fil1:\n row=k.split()\n x1.append(float(row[0]))\n y1.append(float(row[1]))\nx1s,y1s = zip(*sorted(zip(x1,y1)))\n\ncol=[]\t\nfor y in range(1,31):\n\tfile=open('/home/richard/outhome/mechf%d.txt' % (y),\"r\")\n\tfor k in file:\n\t\trow=k.split()\n\t\tcol.append(float(row[1]))\ncola=np.array(col)\nx1sa=np.arange(0.56,0.86,0.01)\n\nfil2=open('/home/richard/outhome2/outhome/Fano43sh2.txt',\"r\")\nx2,y2=[],[]\nfor k in fil2:\n row=k.split()\n x2.append(float(row[0]))\n y2.append(float(row[1]))\nx2s,y2s = zip(*sorted(zip(x2,y2)))\n\nfil3=open('/home/richard/outhome2/outhome/Fano56sh.txt',\"r\")\nx3,y3=[],[]\nfor k in fil3:\n row=k.split()\n x3.append(float(row[0]))\n y3.append(float(row[1]))\nx3s,y3s = zip(*sorted(zip(x3,y3)))\n\nfil4=open('/home/richard/outhome2/outhome/Fano72sh.txt',\"r\")\nx4,y4=[],[]\nfor k in fil4:\n row=k.split()\n x4.append(float(row[0]))\n y4.append(float(row[1]))\nx4s,y4s = zip(*sorted(zip(x4,y4)))\n\nfil5=open('/home/richard/outhome2/outhome/Fano94sh.txt',\"r\")\nx5,y5=[],[]\nfor k in fil5:\n row=k.split()\n x5.append(float(row[0]))\n y5.append(float(row[1]))\nx5s,y5s = zip(*sorted(zip(x5,y5)))\n\n#xsa=np.array(xs)\n#ysa=np.array(ys)\nplt.xlabel('bias Force F')\nplt.ylabel('Fano factor')\nplt.yscale('log')\nplt.plot(x1s,y1s,label='kT=0.033')\nplt.plot(x2s,y2s,label='kT=0.043')\nplt.plot(x3s,y3s,label='kT=0.056')\nplt.plot(x4s,y4s,label='kT=0.072')\nplt.plot(x5s,y5s,label='kT=0.094')\nplt.legend()\nplt.savefig('mechfsh.pdf')\n", "sub_path": "pythonplots/rangeplots/runmechfsh.py", "file_name": "runmechfsh.py", "file_ext": "py", "file_size_in_byte": 1637, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.use", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yscale", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 67, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 67, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 70, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 70, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 71, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 71, "usage_type": "name"}]} +{"seq_id": "622572325", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nfrom math import *\n\n#n = int(input(\"n = \"))\nn = 10000000\n\nt = np.linspace(1, 40, n)\ndt = t[1]-t[0]\n\nprint(dt)\n\ng = 9.8\nrho = 1.293\n\nSx = 0.5\nSy = 0.5\n\nalpha_st = 5\n\nk = 0.1\n\nCd = 0.025\n\nm = 0.6\n\nT = (g*Sx*Cd)/(k*alpha_st*Sy)\nprint(T)\n\ndef alpha(t):\n S = 70*np.exp(-0.3*t)*np.sin(0.1*t) + 5\n return(S)\n\nalphaf = alpha(t)\n\ndef fligh(Sx, Sy, m):\n X = [0, 0]\n Y = [0, 0]\n x = 0\n y = 0\n for i in range(1, len(t)-1):\n x = ((T - (0.5/m)*rho*Cd*Sx*(((X[i]-X[i-1])/dt)*((X[i]-X[i-1])/dt)))*(dt*dt)) + 2*X[i] - X[i-1]\n y = ((((0.5/m)*rho*k*alphaf[i]*Sy*(((X[i]-X[i-1])/dt)*((X[i]-X[i-1])/dt))) - g)*(dt*dt)) + 2*Y[i] - Y[i-1]\n X.append(x)\n Y.append(y)\n return(X, Y)\n\nX, Y = fligh(Sx, Sy, m)\n\n\nplt.plot(t, alphaf, color='red')\nplt.show()\n\nplt.plot(t, X, color='green')\nplt.show()\nplt.plot(t, Y, color='blue')\nplt.show()\n\nplt.plot(X, Y, color='blue')\nplt.show()\n", "sub_path": "avioninator.py", "file_name": "avioninator.py", "file_ext": "py", "file_size_in_byte": 953, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.linspace", "line_number": 8, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}]} +{"seq_id": "389460865", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n'多进程'\n\n__author__ = 'garyhu'\n\nimport os;\nfrom multiprocessing import Process;\n\nprint(\"start process %s ....\" % os.getpid());\n\n# pid = os.fork();\n\n# if pid == 0:\n# print('child process: %s, parent process: %s' % (os.getpid(),os.getppid()))\n# else :\n# print('the parent process(%s) create child process(%s)' % (os.getpid(),pid)) \n \n\ndef run_proc(name):\n print(\"Running child process %s (%s) ...\" % (name,os.getpid()));\n \nif __name__ == '__main__':\n print(\"Parent process is %s.\" % os.getpid());\n p = Process(target=run_proc,args=('my_test',));\n print('child process will start...')\n p.start();\n p.join();\n print(\"child process is end\") \n \n ", "sub_path": "process/python_multiprocessing.py", "file_name": "python_multiprocessing.py", "file_ext": "py", "file_size_in_byte": 744, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.getpid", "line_number": 11, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 22, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 25, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "563578426", "text": "from includes.sumo import SimEnv\n\nimport time\nimport traci\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport statistics as stat\n\n\nfrom includes.DDQAgent import DDQAgent\nfrom includes.utils import plot_learning_curve\n\nfrom traffic_gen import GenerateTraffic\nimport random\n\ndef save_plot_neg(list1, episode):\n print(\"Saving plot\")\n x = [i for i in range(len(list1))]\n\n plt.plot(x, list1)\n plt.grid()\n plt.savefig(\"plots21_march/neg_reward_episodes_\"+str(episode)+\".png\") # save as png\n plt.close()\n\ndef save_plot_pos(list1, episode):\n print(\"Saving plot\")\n x = [i for i in range(len(list1))]\n\n plt.plot(x, list1)\n plt.grid()\n plt.savefig(\"plots21_march/pos_reward_episodes_\"+str(episode)+\".png\") # save as png\n plt.close()\n\ndef save_plot_reward(list1, episode):\n print(\"Saving plot\")\n x = [i for i in range(len(list1))]\n\n plt.plot(x, list1)\n plt.grid()\n plt.savefig(\"plots21_march/reward_episodes_\"+str(episode)+\".png\") # save as png\n plt.close()\n\ndef save_plot(list1, episode):\n print(\"Saving plot\")\n x = [i for i in range(len(list1))]\n\n plt.plot(x, list1)\n plt.grid()\n plt.savefig(\"plots21_march/\"+str(episode)+\".png\") # save as png\n plt.close()\n\ndef single_step(agent, n_steps, episode, observation, env):\n score= 0\n action = agent.choose_action(observation)\n observation_, reward, info = agent.step(action, n_steps, env)\n score += reward\n \"\"\"\n if not load_checkpoint:# and n_steps>100:\n agent.store_transition(observation, action, reward, observation_)\n agent.learn()\n \"\"\" \n observation = observation_\n\n #if n_steps > 100:\n scores.append(score)\n steps_array.append(n_steps)\n \n avg_score = np.mean(scores[-100:])\n #print('-- episode: ', episode, 'steps', n_steps)\n\n return observation, reward, action\n\n\nif __name__ == '__main__':\n env = SimEnv()\n #env.start_sumo()\n \n delayTime = 1\n\n Central = \"TL0\"\n n_games = 5200\n n_steps = 0\n\n scores = []\n eps_history = []\n steps_array = []\n\n best_score = -np.inf\n load_checkpoint = False\n\n episodes = 250\n\n EPS = (1-0.01)/episodes\n\n agent = DDQAgent(gamma=0.99, epsilon=1, lr=0.0001,\n input_dims=8,\n n_actions=4, mem_size=50000, eps_min=0.01,\n batch_size=64, replace=3500, eps_dec=EPS,\n chkpt_dir='models/', algo='DQNAgent',\n env_name='SUMO_tlc', TLC_name = Central)\n\n if load_checkpoint:\n agent.load_models()\n\n episodic_reward_neg =[]\n episodic_reward_pos =[]\n episodic_reward_sum =[] \n episodic_reward_avg =[] \n\n #env.start_sumo()\n \n eps_history = []\n \n for episode in range(1, episodes):\n neg_reward_current_episode = []\n pos_reward_current_episode = []\n reward_current_episode = []\n n_steps = 0\n\n # Reset Traffic\n traffic = GenerateTraffic(n_games+600)\n traffic_flow = 0#random.randint(0,3)\n print(\"Generating traffic for episode: \", episode, \" and flow: \", traffic_flow)\n traffic.set_traffic_flow(episode, traffic_flow)\n\n #Start Sumo\n \n env.start_sumo()\n\n obs, reward, info = agent.step(0, n_steps, env) #taking random action\n action = 0\n reward_ = agent.get_reward()\n start_time = time.time()\n\n green_time = 30\n \n busy_count = green_time + n_steps\n\n for i in range(n_games):\n if n_steps > busy_count:\n\n observation_ = agent.get_state()\n reward_ = agent.get_reward()\n print(\"\\n \")\n\n print('-- episode: ', episode, 'steps', n_steps, \"Reward: \", reward_)\n print(\"Avg Wait time: \", agent.get_avg_wait_time())\n print(\"Queue length array: \", agent.acc_wait_time)\n\n if not load_checkpoint:# and n_steps>100:\n agent.store_transition(obs, action, reward_, observation_)\n agent.learn()\n\n print(\"Performing action\")\n obs, reward, action = single_step(agent, n_steps, episode, observation_, env)\n #observation = obs\n\n busy_count = green_time + n_steps\n #print(\"Busy count: \", busy_count)\n\n print(\"obs_ac: \", obs)\n\n if reward_ < 0:# and n_steps > 100:\n neg_reward_current_episode.append(reward_)\n if reward_ >0:# and n_steps > 100:\n pos_reward_current_episode.append(reward_)\n\n if True:#n_steps > 100: \n reward_current_episode.append(reward_)\n\n time.sleep(delayTime)\n #env.simulationStep()\n n_steps += 1\n else:\n #print(\"Agent is busy: \", n_steps-busy_count)\n n_steps += 1\n env.simulationStep()\n# print(\"Obs: \", observation_)\n\n # Plots for each episode\n save_plot(neg_reward_current_episode, str(episode)+\"_neg_subreward\")\n save_plot(pos_reward_current_episode, str(episode)+\"_pos_subreward\")\n save_plot(reward_current_episode, str(episode)+\"_overall_subreward\")\n\n\n print(\"######################################################\")\n print('episode: ', episode,'neg_reward: ', sum(neg_reward_current_episode),\n 'pos_reward: ', sum(pos_reward_current_episode),\n 'cumm reward: ', sum(reward_current_episode)\n )\n print(\"Time taken (minutes): \", ((time.time() - start_time)/60))\n print(\"Generating new traffic scenario\")\n print(\"######################################################\")\n\n # Plots for overall episodes\n episodic_reward_neg.append(sum(neg_reward_current_episode))\n episodic_reward_pos.append(sum(pos_reward_current_episode))\n episodic_reward_sum.append(sum(reward_current_episode))\n episodic_reward_avg.append(stat.mean(reward_current_episode))\n \n #env.close_sumo() \n #if episode %10 == 0:\n save_plot(episodic_reward_neg, str(episode)+\"_neg\")\n save_plot(episodic_reward_pos, str(episode)+\"_pos\")\n save_plot(episodic_reward_sum, str(episode)+\"_sum\")\n save_plot(episodic_reward_avg, str(episode)+\"_avg\")\n \n agent.save_models()\n\n eps_history.append(agent.epsilon)\n save_plot_neg(eps_history, str(episode)+\"_eps\")\n\n agent.decrement_epsilon()\n\n env.close_sumo() \n \n\n\n #print(\"Episodic reward list: \", episodic_reward_neg)\n# save_plot(episodic_reward_neg, \"Final\")\n save_plot_neg(episodic_reward_neg, \"Final\")\n #save_plot_pos(episodic_reward_pos, \"Final\")\n #save_plot_reward(episodic_reward_sum, \"Final\")", "sub_path": "main_github.py", "file_name": "main_github.py", "file_ext": "py", "file_size_in_byte": 6804, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.pyplot.plot", "line_number": 20, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 20, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 38, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 38, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 39, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 39, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 40, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 40, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "numpy.mean", "line_number": 68, "usage_type": "call"}, {"api_name": "includes.sumo.SimEnv", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 88, "usage_type": "attribute"}, {"api_name": "includes.DDQAgent.DDQAgent", "line_number": 95, "usage_type": "call"}, {"api_name": "traffic_gen.GenerateTraffic", "line_number": 121, "usage_type": "call"}, {"api_name": "time.time", "line_number": 133, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 171, "usage_type": "call"}, {"api_name": "time.time", "line_number": 191, "usage_type": "call"}, {"api_name": "statistics.mean", "line_number": 199, "usage_type": "call"}]} +{"seq_id": "228536173", "text": "import collections\nclass Solution:\n def firstUniqChar(self, s: str) -> int:\n #Time - O(n)\n #Space - O(n)\n if len(s) == 0:\n return -1\n hashmap = collections.Counter(s) #a dictionary with count of all characters\n reqset = set() #stores all unique characters\n for key in hashmap:\n if hashmap[key] == 1:\n reqset.add(key)\n if len(reqset) == 0:\n return -1\n for i in range(len(s)): #While looping, we will return the first ever character which is in reqset\n if s[i] in reqset:\n return i\n return -1", "sub_path": "Week2/FirstUniqueCharacter.py", "file_name": "FirstUniqueCharacter.py", "file_ext": "py", "file_size_in_byte": 630, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "collections.Counter", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "560516219", "text": "from utils.adapter.adapter import RabbitMqOutAdapter, NirvanaOutAdapter, RabbitMqInAdapter, ClassDependencyInAdapter, \\\n StatsClassOutAdapter, FooOut, FooIn\n\n__out_adapters = {\n 'rabbitmq': RabbitMqOutAdapter,\n 'nirvana': NirvanaOutAdapter,\n 'statsclass': StatsClassOutAdapter,\n None: FooOut\n}\n\n__in_adapters = {\n 'rabbitmq': RabbitMqInAdapter,\n 'classdependency': ClassDependencyInAdapter,\n None: FooIn\n}\n\n\ndef get_out_adapter(identifier, **kwargs):\n params = {}\n adapter_class = __out_adapters[identifier]\n\n if identifier == 'rabbitmq':\n params['exchange_name'] = kwargs.get('exchange_name')\n params['mq_connection_string'] = kwargs.get('mq_connection_string')\n\n return adapter_class(**params)\n\n\ndef get_in_adapter(identifier, **kwargs):\n params = {}\n adapter_class = __in_adapters[identifier]\n\n if identifier == 'rabbitmq':\n params['queue_name'] = kwargs.get('queue_name')\n params['mq_connection_string'] = kwargs.get('mq_connection_string')\n\n return adapter_class(**params)\n", "sub_path": "utils/adapter/adapterfactory.py", "file_name": "adapterfactory.py", "file_ext": "py", "file_size_in_byte": 1054, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "utils.adapter.adapter.RabbitMqOutAdapter", "line_number": 5, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.NirvanaOutAdapter", "line_number": 6, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.StatsClassOutAdapter", "line_number": 7, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.FooOut", "line_number": 8, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.RabbitMqInAdapter", "line_number": 12, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.ClassDependencyInAdapter", "line_number": 13, "usage_type": "name"}, {"api_name": "utils.adapter.adapter.FooIn", "line_number": 14, "usage_type": "name"}]} +{"seq_id": "350143011", "text": "\"\"\"\nCopyright 2018 Conijn.io. or its affiliates. All Rights Reserved.\n\"\"\"\nimport pytest\nfrom aws_lambda_event_handler import SNSMessage\nfrom aws_lambda_event_handler import LambdaHandlerException\nfrom ..fixtures import sns_event\n\n@pytest.mark.parametrize('event, exception', [\n (None, LambdaHandlerException),\n (sns_event(message='{\"Foo\": \"Bar\"}', attributes={'myKey': {'Value': 'myValue'}}), None)\n])\ndef test_sns_message(event, exception):\n \"\"\"\n LambdaHandlerException\n \"\"\"\n if exception is None:\n message = SNSMessage(event)\n assert message.get_message() == event['Sns']['Message']\n assert message.get_attribute('myKey') == 'myValue'\n assert message.get_attribute('otherKey', 'otherValue') == 'otherValue'\n assert message.get_message_json() == '{\"Foo\": \"Bar\"}'\n assert message.get_message_object() == {\"Foo\": \"Bar\"}\n else:\n with pytest.raises(exception):\n message = SNSMessage(event)\n", "sub_path": "tests/models/test_sns_message.py", "file_name": "test_sns_message.py", "file_ext": "py", "file_size_in_byte": 968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "aws_lambda_event_handler.SNSMessage", "line_number": 18, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 25, "usage_type": "call"}, {"api_name": "aws_lambda_event_handler.SNSMessage", "line_number": 26, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 9, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 9, "usage_type": "attribute"}, {"api_name": "aws_lambda_event_handler.LambdaHandlerException", "line_number": 10, "usage_type": "name"}, {"api_name": "fixtures.sns_event", "line_number": 11, "usage_type": "call"}]} +{"seq_id": "114749592", "text": "import base64\nimport requests\nimport datetime\nimport os\n\nCLIENT_ID = os.environ[\"SPOTIPY_CLIENT_ID\"]\nCLIENT_SECRET = os.environ[\"SPOTIPY_CLIENT_SECRET\"]\n\nclient_creds = f\"{CLIENT_ID}:{CLIENT_SECRET}\"\nclient_creds_64 = base64.b64encode(client_creds.encode())\n\ntoken_url = 'https://accounts.spotify.com/api/token'\ntoken_data = {\n \"grant_type\" : \"client_credentials\"\n}\ntoken_headers = {\n \"Authorization\" : f\"Basic {client_creds_64.decode()}\"\n}\n\nr = requests.post(token_url, data = token_data, headers = token_headers)\n\nif r.status_code in range(200, 299) :\n token_response_data = r.json()\n now = datetime.datetime.now()\n access_token = token_response_data[\"access_token\"]\n expires_in = token_response_data[\"expires_in\"]\n remaining_token_time = now + datetime.timedelta(seconds = expires_in)\n did_expire = remaining_token_time < now\n\n #print(\"token: %s\" % access_token)\n #print(\"expired\") if did_expire else print(f\"expires in: {expires_in}\")\nelse :\n print(\"invalid request\")\n", "sub_path": "spotifyAuth.py", "file_name": "spotifyAuth.py", "file_ext": "py", "file_size_in_byte": 1005, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.environ", "line_number": 6, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 7, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 10, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 20, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 24, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 24, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 27, "usage_type": "call"}]} +{"seq_id": "607060766", "text": "\"\"\"Tests for letsencrypt.achallenges.\"\"\"\nimport unittest\n\nimport OpenSSL\n\nfrom acme import challenges\nfrom acme import crypto_util as acme_crypto_util\nfrom acme import jose\n\nfrom letsencrypt.tests import acme_util\nfrom letsencrypt.tests import test_util\n\n\nclass DVSNITest(unittest.TestCase):\n \"\"\"Tests for letsencrypt.achallenges.DVSNI.\"\"\"\n\n def setUp(self):\n self.chall = acme_util.chall_to_challb(\n challenges.DVSNI(r=\"r_value\", nonce=\"12345ABCDE\"), \"pending\")\n self.response = challenges.DVSNIResponse()\n key = jose.JWKRSA.load(test_util.load_vector(\"rsa512_key.pem\"))\n\n from letsencrypt.achallenges import DVSNI\n self.achall = DVSNI(challb=self.chall, domain=\"example.com\", key=key)\n\n def test_proxy(self):\n self.assertEqual(self.chall.r, self.achall.r)\n self.assertEqual(self.chall.nonce, self.achall.nonce)\n\n def test_gen_cert_and_response(self):\n cert_pem, _ = self.achall.gen_cert_and_response(s=self.response.s)\n\n cert = OpenSSL.crypto.load_certificate(\n OpenSSL.crypto.FILETYPE_PEM, cert_pem)\n self.assertEqual(cert.get_subject().CN, \"example.com\")\n # pylint: disable=protected-access\n self.assertEqual(acme_crypto_util._pyopenssl_cert_or_req_san(cert), [\n \"example.com\", self.chall.nonce_domain,\n self.response.z_domain(self.chall)])\n\n\nif __name__ == \"__main__\":\n unittest.main() # pragma: no cover\n", "sub_path": "letsencrypt/tests/achallenges_test.py", "file_name": "achallenges_test.py", "file_ext": "py", "file_size_in_byte": 1454, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 14, "usage_type": "attribute"}, {"api_name": "letsencrypt.tests.acme_util.chall_to_challb", "line_number": 18, "usage_type": "call"}, {"api_name": "letsencrypt.tests.acme_util", "line_number": 18, "usage_type": "name"}, {"api_name": "acme.challenges.DVSNI", "line_number": 19, "usage_type": "call"}, {"api_name": "acme.challenges", "line_number": 19, "usage_type": "name"}, {"api_name": "acme.challenges.DVSNIResponse", "line_number": 20, "usage_type": "call"}, {"api_name": "acme.challenges", "line_number": 20, "usage_type": "name"}, {"api_name": "acme.jose.JWKRSA.load", "line_number": 21, "usage_type": "call"}, {"api_name": "acme.jose.JWKRSA", "line_number": 21, "usage_type": "attribute"}, {"api_name": "acme.jose", "line_number": 21, "usage_type": "name"}, {"api_name": "letsencrypt.tests.test_util.load_vector", "line_number": 21, "usage_type": "call"}, {"api_name": "letsencrypt.tests.test_util", "line_number": 21, "usage_type": "name"}, {"api_name": "letsencrypt.achallenges.DVSNI", "line_number": 24, "usage_type": "call"}, {"api_name": "OpenSSL.crypto.load_certificate", "line_number": 33, "usage_type": "call"}, {"api_name": "OpenSSL.crypto", "line_number": 33, "usage_type": "attribute"}, {"api_name": "OpenSSL.crypto", "line_number": 34, "usage_type": "attribute"}, {"api_name": "acme.crypto_util._pyopenssl_cert_or_req_san", "line_number": 37, "usage_type": "call"}, {"api_name": "acme.crypto_util", "line_number": 37, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 43, "usage_type": "call"}]} +{"seq_id": "388213492", "text": "\"\"\"\nScatterplot visualizer used to plot the error of F2 estimate amplifiers. Hard-coded to read input from\noutput file produced by Visualizer class. \n\n\"\"\"\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport re\nfrom sys import stdin\nfrom scipy.stats import gaussian_kde\n\n# Computes sliding average over all trials\ndef sliding_mean(data_array, xaxis, minElement=1, window=10): \n data_array = np.array(data_array)\n newdict = dict()\n\n for i in range(len(data_array)):\n \tif xaxis[i] not in newdict:\n \t\tnewdict[xaxis[i]] = data_array[i]\n \telse:\n \t\tnewdict[xaxis[i]] = (newdict[xaxis[i]] + data_array[i])/2\n\n new_list = [] \n for i in range(len(newdict)): \n\n indices = range(max(i - window + 1, minElement), \n min(i + window + 1, len(newdict))) \n avg = 0 \n for j in indices: \n avg += newdict[j] \n avg /= float(len(indices)) \n \tnew_list.append(avg) \n\n return np.array(new_list) \n\nf = open('tempfile7624561794', 'r')\n\n# First three lines are label info\ntitle = f.readline()\nyaxis = f.readline()\nxaxis = f.readline()\n\nxs = []\nys = []\n# mymap = plt.get_cmap(\"rainbow\")\n\n\nfor line in f.readlines():\n\tx1, y1 = (x for x in line.split(','))\n\txs.append(float(x1))\n\tys.append(float(y1.strip(\" \")))\n\nxs = np.array(xs)\nys = np.array(ys)\n\n\nN = len(xs)\nmaxX = max(xs)\nminX = min(xs)\nmaxY = max(ys)\nminY = min(ys)\n\n# Color by point density\nxy = np.vstack([xs,ys])\nz = gaussian_kde(xy)(xy)\n\n# # Place dense points on top\nidx = z.argsort()\n# print xs\nxs, ys, z = xs[idx], ys[idx], z[idx]\n# print xs\n\n# Compute average over all trials\nfirstX = []\nfor i in range(int(minX), int(maxX)+1):\n\tfirstX.append(i)\n\n# mean_PlyCount = sliding_mean(ys, xs, minElement=int(minX), window=10)\n\n# Remove distracting formatting\nfig, ax = plt.subplots()\nax.spines[\"top\"].set_visible(False) \nax.spines[\"right\"].set_visible(False)\nax.get_xaxis().tick_bottom() \nax.get_yaxis().tick_left() \n\n# Limit whitespace around points of interest\n# plt.xlim(int(minX - maxX*.05), int(maxX + maxX*.05))\nplt.ylim(minY, maxY)\nplt.xlim(minX, maxX)\n\n# plt.scatter(xs, ys, s=area, c=colors, alpha=0.5)\nax.scatter(xs, ys, alpha=0.5, c = z)\n# ax.scatter(xs, ys, alpha=0.5)\n\n# ax.plot(firstX, mean_PlyCount, color=\"red\", lw=3) \n\n# Add axis laels \nplt.title(title)\nplt.xlabel(xaxis, fontsize = 16) \nplt.ylabel(yaxis, fontsize = 16) \nplt.show()\n\n", "sub_path": "dump/scatter.py", "file_name": "scatter.py", "file_ext": "py", "file_size_in_byte": 2396, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.array", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 65, "usage_type": "call"}, {"api_name": "scipy.stats.gaussian_kde", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 82, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 82, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 102, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 102, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 103, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 103, "usage_type": "name"}]} +{"seq_id": "563123088", "text": "# importing\nimport pymongo\nimport json\nimport pandas as pd\nimport os\n\n# csv path\ndirname = os.path.dirname(__file__)\nfile = os.path.join(dirname, 'co-emissions.csv')\n\n# reading csv and assigning to 'data'\ndata = pd.read_csv(file)\n\n################################ clean up ###############################\n# dropping all columns before 1980 (1980 - 2017 remains)\ndata.drop(data[data.Year < 2000].index, inplace=True)\n\n# dropping rows with all null values in rows\ndata.dropna(how=\"all\", inplace=True)\n\n# dropping rows with all null values in columns\ndata.dropna(axis=\"columns\", how=\"all\", inplace=True)\n\n# filling NA values\ndata[\"Entity\"].fillna(\"No Country\", inplace=True)\ndata[\"Code\"].fillna(\"No Code\", inplace=True)\ndata[\"Year\"].fillna(\"No Year\", inplace=True)\ndata[\"Per capita CO2 emissions (tonnes per capita)\"].fillna(0, inplace=True)\n\n# Sort by Year && Country\ndata.sort_values([\"Year\", \"Entity\"], inplace=True)\n\n# renaming columns\ndata.rename(columns={\"Entity\": \"Country\",\n \"Per capita CO2 emissions (tonnes per capita)\": \"CO2 emissions (metric tons)\"},\n inplace=True)\n\n# changing 'year' column datatype to 'str'\ndata = data.astype({\"Year\": str})\n####################################################################################\n################################ Storing to Database ###############################\n# database connection setup\nclient = ('mongodb://localhost:27017/')\nmongo_client = pymongo.MongoClient(client)\ndatabase = mongo_client['emission-data']\nemission_collection = database['emission']\n\n# converting 'data' to JSON\nrecords = json.loads(data.T.to_json()).values()\n\n# inserting into mongodb\ntry:\n emission_collection.insert_many(records)\n print('Database created successfully')\nexcept pymongo.errors.ServerSelectionTimeoutError as err:\n print(err)\n\nmongo_client.close()\n", "sub_path": "Task_3-4/scripts/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1840, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.dirname", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 9, "usage_type": "call"}, {"api_name": "os.path", "line_number": 9, "usage_type": "attribute"}, {"api_name": "pandas.read_csv", "line_number": 12, "usage_type": "call"}, {"api_name": "pymongo.MongoClient", "line_number": 44, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "pymongo.errors", "line_number": 55, "usage_type": "attribute"}]} +{"seq_id": "509449018", "text": "from django.shortcuts import render, redirect\nfrom django.http import HttpResponse, HttpResponseRedirect, JsonResponse, HttpRequest\nfrom .forms import registration_form, NewWorkoutForm\nfrom .models import NewWorkout, User_Profile_Model, WorkoutWeek\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib.auth import authenticate, login\nfrom django.views.decorators.csrf import csrf_exempt\nfrom .utils import create_workout_week\n\n\ndef new_workout(request):\n form = NewWorkoutForm()\n context = {\n \"form\":form\n }\n return render(request, \"new_workout.html\", context)\n\ndef index(request):\n if request.user.is_authenticated:\n return redirect(\"/home/\")\n else:\n return render(request, \"index.html\")\n\ndef update(request, week, workout):\n cur_user = request.user\n new_workout = NewWorkout.objects.get(user_id=cur_user.id)\n workout_weeks = WorkoutWeek.objects.filter(associated_workout=new_workout.id)\n workout_week = workout_weeks.get(name=week)\n if workout == 'bench': \n WorkoutWeek.objects.filter(pk=workout_week.id).update(bench_done=True)\n if workout == 'squat': \n WorkoutWeek.objects.filter(pk=workout_week.id).update(squat_done=True)\n if workout == 'deadlift': \n WorkoutWeek.objects.filter(pk=workout_week.id).update(deadlift_done=True) \n if workout == 'overhead': \n WorkoutWeek.objects.filter(pk=workout_week.id).update(overhead_done=True) \n # check if they have any more unfinished workouts\n weeksdone = 0\n for single_week in workout_weeks:\n if single_week.bench_done and single_week.squat_done and single_week.deadlift_done and single_week.overhead_done:\n weeksdone += 1\n if weeksdone == 4:\n new_max_squat = new_workout.max_squat + 10\n #NewWorkout.objects.filter(user_id=cur_user.id).update(max_squat=new_max_squat)\n new_max_bench = new_workout.max_bench + 5\n #NewWorkout.objects.filter(user_id=cur_user.id).update(max_bench=new_max_bench)\n new_max_deadlift = new_workout.max_deadlift + 10\n #NewWorkout.objects.filter(user_id=cur_user.id).update(max_deadlift=new_max_deadlift)\n new_max_overhead = new_workout.max_overhead + 5\n #NewWorkout.objects.filter(user_id=cur_user.id).update(max_overhead=new_max_overhead)\n form = NewWorkoutForm({\n 'max_squat' : new_max_squat,\n 'max_bench' : new_max_bench,\n 'max_deadlift' : new_max_deadlift,\n 'max_overhead' : new_max_overhead\n })\n congrats_message = \"\"\" Congratulations {cur_user}! You have completed your\n first 4 week workout cycle. It is suggested that you increase\n your 1 Rep Max's used to calculate your workout by 10 lbs for legs\n and 5 lbs for upper body. I have taken the liberty of doing that based\n on the 1 Rep Maxes you originally entered. I would reccomend keeping these\n suggested increases unless you failed to complete workouts durring this 4\n week cycle then I would reduce legs by 10 and upper body by 5 and repeat the \n last cycle again. Good luck and keep up the hard work!\n \"\"\"\n context = {\n \"form\":form,\n \"message\": congrats_message,\n \"finished_workout\": True\n } \n\n\n\n # new_request = HttpRequest()\n # new_request.mode = 'POST'\n # new_request.META = request.META\n return render(request, \"new_workout.html\", context)\n #return redirect('/new_workout/')\n #make a new workout\n return redirect(\"/home/\")\n\n@login_required(login_url=\"/\")\ndef home(request):\n if request.method == \"GET\":\n cur_user = request.user\n has_workout = NewWorkout.objects.filter(user_id=cur_user.id).count()\n if has_workout == 0:\n context = {\n \"user\": request.user.username,\n \"hasWorkout\": False,\n \"message\": \"The gym is empty and so is your training plan! Not much to do here if you dont have a workout, you should go create a new workout.\"\n }\n return render(request, \"home.html\", context)\n\n workout_dict = {}\n workout_dict['workout_weeks'] = []\n user_workout = NewWorkout.objects.get(user_id=cur_user.id)\n user_workout_weeks = WorkoutWeek.objects.filter(associated_workout=user_workout.id)\n for week in user_workout_weeks:\n workout_dict['workout_weeks'] += [{\n \"name\": week.name,\n \"pretty_name\": week.prettyName,\n \"reps\": week.reps,\n \"bench\": week.bench,\n \"bench_done\": week.bench_done,\n \"squat\": week.squat,\n \"squat_done\": week.squat_done,\n \"deadlift\": week.deadlift,\n \"deadlift_done\": week.deadlift_done,\n \"overhead\": week.overhead,\n \"overhead_done\": week.overhead_done,\n }]\n context = {\n \"user\": request.user.username,\n \"hasWorkout\": True,\n \"workoutWeeks\": workout_dict['workout_weeks'],\n }\n print(context)\n return render(request, \"home.html\", context)\n ## This is when you create a new_workout and new workout_schedule\n if request.method == 'POST':\n # get current user from request\n cur_user = request.user\n\n # Get the current user profile from db\n cur_user_profile = User_Profile_Model.objects.get(user_id=cur_user.id)\n\n #Check if user has a workout already if so clean it up.\n has_workout = NewWorkout.objects.filter(user_id=cur_user.id).count()\n\n form = NewWorkoutForm(request.POST)\n if form.is_valid():\n if has_workout > 0:\n NewWorkout.objects.filter(user_id=cur_user.id).delete()\n cur_user_profile.has_workout = 0\n cur_user_profile.save()\n\n new_workout = NewWorkout(\n max_squat=form.cleaned_data['max_squat'],\n max_bench=form.cleaned_data['max_bench'],\n max_deadlift=form.cleaned_data['max_deadlift'],\n max_overhead=form.cleaned_data['max_overhead'], \n user_id=cur_user\n )\n new_workout.save()\n cur_user_profile.has_workout = 1\n cur_user_profile.save()\n workout_multiplier = .65\n for i in range(1, 4):\n create_workout_week(request, workout_multiplier, '_week_' + str(i))\n workout_multiplier += .05\n workout_multiplier = .4\n create_workout_week(request, workout_multiplier, '_week_4')\n return redirect('/home/')\n #else:\n # mod_workout = NewWorkout.objects.filter(user_id=cur_user.id) \n # if has_workout > 0:\n # new_workout = NewWorkout(\n # max_squat = mod_workout.max_squat,\n # max_bench = mod_workout.max_bench,\n # max_deadlift = mod_workout.max_deadlift,\n # max_overhead = mod_workout.max_overhead,\n # user_id=cur_user\n # )\n # NewWorkout.objects.filter(user_id=cur_user.id).delete()\n # cur_user_profile.has_workout = 0\n # cur_user_profile.save()\n # new_workout.save()\n # cur_user_profile.has_workout = 1\n # cur_user_profile.save() \n # workout_multiplier = .65\n # for i in range(1, 4):\n # create_workout_week(request, workout_multiplier, '_week_' + str(i))\n # workout_multiplier += .05\n # workout_multiplier = .4\n # create_workout_week(request, workout_multiplier, '_week_4')\n # return redirect('/home/') \n return render(request, \"home.html\")\n\n\n# auto login adapted from https://stackoverflow.com/questions/3222549/how-to-automatically-login-a-user-after-registration-in-django\ndef register(request):\n if request.method == 'POST':\n form = registration_form(request.POST)\n if form.is_valid():\n form.save(commit=True)\n newUser = authenticate(\n username=form.cleaned_data['username'],\n password=form.cleaned_data['password1'],\n )\n login(request, newUser)\n return redirect(\"/home/\")\n else:\n form = registration_form()\n context = {\n \"form\":form\n }\n return render(request,\"registration/register.html\",context)", "sub_path": "code/strength_trainer/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 8835, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "forms.NewWorkoutForm", "line_number": 12, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 20, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 22, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects.get", "line_number": 26, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects", "line_number": 26, "usage_type": "attribute"}, {"api_name": "models.NewWorkout", "line_number": 26, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 27, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 27, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 30, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 30, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 30, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 32, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 32, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 32, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 34, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 34, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 36, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 36, "usage_type": "name"}, {"api_name": "forms.NewWorkoutForm", "line_number": 51, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 77, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 80, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects.filter", "line_number": 86, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects", "line_number": 86, "usage_type": "attribute"}, {"api_name": "models.NewWorkout", "line_number": 86, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 93, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects.get", "line_number": 97, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects", "line_number": 97, "usage_type": "attribute"}, {"api_name": "models.NewWorkout", "line_number": 97, "usage_type": "name"}, {"api_name": "models.WorkoutWeek.objects.filter", "line_number": 98, "usage_type": "call"}, {"api_name": "models.WorkoutWeek.objects", "line_number": 98, "usage_type": "attribute"}, {"api_name": "models.WorkoutWeek", "line_number": 98, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 119, "usage_type": "call"}, {"api_name": "models.User_Profile_Model.objects.get", "line_number": 126, "usage_type": "call"}, {"api_name": "models.User_Profile_Model.objects", "line_number": 126, "usage_type": "attribute"}, {"api_name": "models.User_Profile_Model", "line_number": 126, "usage_type": "name"}, {"api_name": "models.NewWorkout.objects.filter", "line_number": 129, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects", "line_number": 129, "usage_type": "attribute"}, {"api_name": "models.NewWorkout", "line_number": 129, "usage_type": "name"}, {"api_name": "forms.NewWorkoutForm", "line_number": 131, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects.filter", "line_number": 134, "usage_type": "call"}, {"api_name": "models.NewWorkout.objects", "line_number": 134, "usage_type": "attribute"}, {"api_name": "models.NewWorkout", "line_number": 134, "usage_type": "name"}, {"api_name": "models.NewWorkout", "line_number": 138, "usage_type": "call"}, {"api_name": "utils.create_workout_week", "line_number": 150, "usage_type": "call"}, {"api_name": "utils.create_workout_week", "line_number": 153, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 154, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 178, "usage_type": "call"}, {"api_name": "django.contrib.auth.decorators.login_required", "line_number": 82, "usage_type": "call"}, {"api_name": "forms.registration_form", "line_number": 184, "usage_type": "call"}, {"api_name": "django.contrib.auth.authenticate", "line_number": 187, "usage_type": "call"}, {"api_name": "django.contrib.auth.login", "line_number": 191, "usage_type": "call"}, {"api_name": "django.shortcuts.redirect", "line_number": 192, "usage_type": "call"}, {"api_name": "forms.registration_form", "line_number": 194, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 198, "usage_type": "call"}]} +{"seq_id": "61898219", "text": "import numpy as np\nfrom ase.cell import Cell\nfrom ase.lattice import bravais_lattices, UnsupportedLattice\nfrom ase.build import bulk, fcc111\nfrom ase.test.testsuite import must_raise\n\nbravais = {}\nfor name in bravais_lattices:\n bravais[name.lower()] = bravais_lattices[name]\n\n\ndef check_single(name, cell, pbc=None):\n c = Cell(cell)\n\n try:\n print('TEST', c, pbc)\n if pbc[:2].all() or sum(pbc) == 1:\n lattice = c.get_bravais_lattice(pbc=pbc)\n else:\n with must_raise(UnsupportedLattice):\n lattice = c.get_bravais_lattice(pbc=pbc)\n return\n except RuntimeError:\n print('error checking {}'.format(name))\n raise\n name1 = lattice.name.lower()\n latname = name.split('@')[0]\n ok = latname == name1\n print(name, '-->', name1, 'OK' if ok else 'ERR', c.cellpar())\n assert ok, 'Expected {} but found {}'.format(latname, name1)\n\n\ndef check(name, cell, pbc=None):\n if pbc is None:\n pbc = cell.any(1)\n pbc = np.asarray(pbc)\n cell = Cell(cell)\n\n # Check all three positive permutations:\n check_single(name + '@012', cell[[0, 1, 2]], pbc=pbc[[0, 1, 2]])\n # 2D lattice determination only supports pbc=(1,1,0) and hence we\n # check the permutations only for 3D lattices:\n if cell.rank == 3 and pbc.sum() != 1:\n check_single(name + '@201', cell[[2, 0, 1]], pbc=pbc[[2, 0, 1]])\n check_single(name + '@120', cell[[1, 2, 0]], pbc=pbc[[1, 2, 0]])\n\n\ncheck('cub', bravais['cub'](3.3).tocell())\ncheck('fcc', bravais['fcc'](3.4).tocell())\ncheck('fcc', bulk('Au').cell)\ncheck('bcc', bravais['bcc'](3.5).tocell())\ncheck('bcc', bulk('Fe').cell)\ncheck('tet', bravais['tet'](4., 5.).tocell())\ncheck('tet', np.diag([4., 5., 5.]))\ncheck('tet', np.diag([5., 4., 5.]))\ncheck('tet', np.diag([5., 5., 4.]))\ncheck('bct', bravais['bct'](3., 4.).tocell())\ncheck('orc', bravais['orc'](3., 4., 5.).tocell())\ncheck('orcf', bravais['orcf'](4., 5., 7.).tocell())\ncheck('orci', bravais['orci'](2., 5., 6.).tocell())\ncheck('orcc', bravais['orcc'](3., 4., 5.).tocell())\ncheck('hex', fcc111('Au', size=(1, 1, 3), periodic=True).cell)\ncheck('hex', bravais['hex'](5., 6.).tocell())\ncheck('rhl', bravais['rhl'](4., 54.).tocell())\ncheck('mcl', bravais['mcl'](2., 3., 4., 62.).tocell())\ncheck('mclc', bravais['mclc'](3., 4., 5., 75.).tocell())\ncheck('tri', bravais['tri'](7., 6., 5., 65., 70., 80.).tocell())\n\n# For 2D materials we have to check both the tocell() method\n# but also for realistic cell nonzero nonperiodic axis.\ncheck('sqr', bravais['sqr'](3.).tocell())\ncheck('sqr', Cell(np.diag([3., 3., 10.])),\n pbc=np.array([True, True, False]))\n\ncheck('crect', bravais['crect'](3., 40).tocell())\n\nalpha = 40 / 360 * 2 * np.pi\na = 3\nx = np.cos(alpha)\ny = np.sin(alpha)\n\ncrectcell = np.array([[a, 0, 0],\n [a * x, a * y, 0],\n [0, 0, 10]])\ncheck('crect', Cell(crectcell), pbc=[1, 1, 0])\n\ncheck('rect', bravais['rect'](3., 4.).tocell())\ncheck('rect', Cell.new([3, 4, 10]), pbc=[1, 1, 0])\n\ncheck('hex2d', bravais['hex2d'](3.).tocell())\nx = 0.5 * np.sqrt(3)\nhexcell = np.array([[a, 0, 0],\n [-0.5 * a, x * a, 0],\n [0., 0., 0.]])\ncheck('hex2d', Cell(hexcell))\n\ncheck('obl', bravais['obl'](3., 4., 40).tocell())\n\nb = 4\nx = np.cos(alpha)\ny = np.sin(alpha)\noblcell = np.array([[a, 0, 0],\n [b * x, b * y, 0],\n [0, 0, 10]])\ncheck('obl', Cell(oblcell), pbc=np.array([True, True, False]))\n\n# 1-d:\ncheck('line', Cell(np.diag([a, 0, 0.0])))\ncheck('line', Cell(np.diag([a, 1, 1.0])), pbc=np.array([1, 0, 0]))\ncheck('line', Cell(np.diag([0.0, 0, a])))\ncheck('line', Cell(np.diag([1.0, 1, a])), pbc=np.array([0, 0, 1]))\n", "sub_path": "test/bravais_check.py", "file_name": "bravais_check.py", "file_ext": "py", "file_size_in_byte": 3723, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "ase.lattice.bravais_lattices", "line_number": 8, "usage_type": "name"}, {"api_name": "ase.lattice.bravais_lattices", "line_number": 9, "usage_type": "name"}, {"api_name": "ase.cell.Cell", "line_number": 13, "usage_type": "call"}, {"api_name": "ase.test.testsuite.must_raise", "line_number": 20, "usage_type": "call"}, {"api_name": "ase.lattice.UnsupportedLattice", "line_number": 20, "usage_type": "argument"}, {"api_name": "numpy.asarray", "line_number": 36, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 37, "usage_type": "call"}, {"api_name": "ase.build.bulk", "line_number": 50, "usage_type": "call"}, {"api_name": "ase.build.bulk", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 56, "usage_type": "call"}, {"api_name": "ase.build.fcc111", "line_number": 62, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 73, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 77, "usage_type": "attribute"}, {"api_name": "numpy.cos", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 82, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 85, "usage_type": "call"}, {"api_name": "ase.cell.Cell.new", "line_number": 88, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 88, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 92, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 101, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 102, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 105, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 108, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 108, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 109, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 110, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 110, "usage_type": "call"}, {"api_name": "ase.cell.Cell", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 111, "usage_type": "call"}]} +{"seq_id": "653027017", "text": "import pygame\nimport math\nimport random\nimport sys\n\n## Press R to make a repeller, space to turn off gravity, arrow keys for wind\n\n## Colours\nRED = [255, 0, 0]\nGREEN = [0, 255, 0]\nBLUE = [0, 0, 255]\nORANGE = [255, 128, 0]\nYELLOW = [242, 224, 63]\nBLACK = [0, 0, 0, 255]\nWHITE = [255, 255, 255]\nLIGHT_BLUE = [0, 255, 255]\nPINK = [255, 153, 255]\nLIME = [150, 255, 0]\n\n## Constants\n\nbounce = False\ngravity = True\nparticleFade = True\nwindLeft = False\nwindRight = False\n\nwind_right_force = (2.5, 0)\nwind_left_force = (-2.5, 0)\n\ng = 0.1 ## acceleration due to gravity\nradius_constant = 3\nframerate = 120 ## frames per second\n\nv_lim = 0.5 ## max velocity of particles\nf_limit = 5\n\nparticle_num = 1 ## Particles created per system\noffset = 3 ## Maximum offset particle is moved by\nparticle_size = 15 ## size of particles\ndecay_speed = 1.7 ## speed of decay\nlifespan = 255\n\nrepel_mass = 50\nrepeller_strength = 40\n\n## Setting window size, defining screen\nwin_size = (900, 900)\nscreen = pygame.display.set_mode(win_size)\n\n## Defining centre of screen\ncentre = (int(win_size[0]/2), int(win_size[1]/2))\n\n## Initialise window\npygame.init()\n\n## Defining clock object\nclock = pygame.time.Clock()\n\n## Window runtime loop\nwin_loop = True\n\n## Return magnitude, direction of a given vector\ndef magDir(x, y):\n return math.sqrt(x**2 + y**2), round(math.atan2(y, x), 5)\n\n## Return x, y components of given vector\ndef resolve(mag, t):\n return mag*math.cos(t), mag*math.sin(t)\n\n## Return direction of given vector\ndef direction(x, y):\n return math.atan2(y, x)\n\n## Return magnitude of given vector\ndef mag(x, y):\n return math.sqrt(x**2 + y**2)\n\n## Scale given vector to given magnitude\ndef setMag(x, y, m):\n \n magnitude = mag(x, y)\n \n if magnitude == 0:\n return 0, 0\n else:\n \n x = x/magnitude * m\n y = y/magnitude * m\n \n return x, y\n\n## Scale given vector to unit length\ndef unitVector(x, y):\n x, y = setMag(x, y, 1)\n \n return x, y\n\n## Limit given vector to given max length\ndef limit(x, y, a):\n if mag(x, y) >= a:\n x, y = setMag(x, y, a)\n \n return x, y\n\ndef distance(body1, body2):\n ## Difference in x-pos and y-pos\n delta_x = body2.px - body1.px\n delta_y = body2.py - body1.py\n \n ## Uses Pythagorean Theorem\n return math.sqrt((delta_x)**2 + (delta_y)**2)\n\n## Single Particle object\n## (colour, mass, pos, vel, acc, force, lifespan)\nclass Particle(object):\n def __init__(self, col, mass, p, v, a, f, lifespan):\n self.col = col\n self.mass = mass\n \n ## Radius of circle is proportional to sqrt of mass (S.A = kr^2)\n self.radius = int(math.sqrt(mass) * radius_constant)\n \n ## Components of pos, vel, acc, force\n self.px = p[0]\n self.py = p[1]\n \n self.vx = v[0]\n self.vy = v[1]\n \n self.ax = a[0]\n self.ay = a[1]\n \n self.fx = f[0]\n self.fy = f[1]\n \n ## Lifespan\n self.lifespan = lifespan\n \n ## Weight is vector with y-component f=ma=mg\n self.weight = [0, self.mass * g]\n \n self.prevPos = (0, 0)\n \n ## Adds velocity to position\n def changePos(self):\n \n self.prevPos = (self.px, self.py)\n \n self.px = int(self.px + self.vx)\n self.py = int(self.py + self.vy)\n \n ## Adds acceleration to velocity\n def changeVel(self):\n self.vx += self.ax\n self.vy += self.ay\n \n ## f=ma --> new a = f/mass\n def changeAcc(self):\n \n self.ax += self.fx/self.mass\n self.ay += self.fy/self.mass\n \n \n ## Add given force to total (resultant) force\n def applyForce(self, force):\n self.fx += force[0]\n self.fy += force[1]\n #print (self.fy)\n \n ## Force arrow that doesn't work properly\n #drawArrow(screen, (100, 255, 0), (self.px, self.py),(self.px +force[0] , self.py+force[1]*0.1 ))\n \n ## Accelerate body to mouse location\n def moveToMouse(self, mouse):\n ## Figure out direction\n self.ax = mouse[0] - self.px\n self.ay = mouse[1] - self.py\n \n ## Set magnitude to value of acceleration (constant at the top)\n self.ax, self.ay = setMag(self.ax, self.ay, acc_val)\n \n ## Change the velocity, limit to constant\n self.changeVel()\n self.vx, self.vy = limit(self.vx, self.vy, vel_limit)\n \n ## Change position\n self.changePos()\n \n ## Bounce off walls of screen\n def bounce(self):\n if self.px + self.vx >= win_size[0] or self.px + self.vx < 0:\n self.vx = -self.vx\n \n if self.py + self.vy >= win_size[1] or self.py + self.vy < 0:\n self.vy = -self.vy\n \n ## Check if still alive (using lifespan value)\n def isAlive(self):\n if self.lifespan <= 0:\n return False\n else:\n return True\n \n \n ## Full movement algorithm\n def move(self): \n ## Accelerate to mouse location\n #self.moveToMouse(mouse)\n \n ## Bounce of screen edge\n if bounce == True:\n self.bounce()\n \n ## Change acceleration\n self.changeAcc()\n #i.ax, i.ay = limit(i.ax, i.ay, acc_val)\n \n ## Change velocity\n self.changeVel()\n #i.vx, i.vy = limit(i.vx, i.vy, vel_limit)\n \n ## Change position\n self.changePos()\n \n ## Changes colour of particle - decrements RGB values\n def fade(self):\n for r in range(len(self.col)):\n #print (r, particle.col[r])\n if self.col[r] > decay_speed:\n self.col[r] = self.col[r] - decay_speed\n else:\n self.col[r] = decay_speed\n \n def replacePrevPos(self):\n pygame.draw.circle(screen, BLACK, self.prevPos, self.radius)\n \nclass Repeller(Particle):\n def __init__(self, G, p):\n self.G = G\n self.px = p[0]\n self.py = p[1]\n \n super().__init__(WHITE, repel_mass, p, (0, 0), (0, 0), (0, 0), 0)\n \n def repel(self, body):\n r = distance(self, body)\n dir_x, dir_y = unitVector((self.px - body.px), (self.py-body.py))\n \n if r != 0:\n strength = self.G * (self.mass * body.mass) / r**2\n else:\n strength = 75\n \n fx, fy = setMag(dir_x, dir_y, strength)\n \n fx, fy = limit(fx, fy, f_limit)\n \n body.fx += fx\n body.fy += fy\n\n## Particle system class\nclass ParticleSystem(object):\n ## Takes origin of particles, colour\n def __init__(self, origin, c):\n self.origin = origin\n ## List of all Particle objects in system\n self.particles = []\n self.c = c\n self.positions = []\n \n ## Method to generate Particle objects\n def makeEmitter(self):\n \n ## Makes particle_num Particle objects\n for j in range(particle_num):\n \n ## Generate new object with appropriate attributes, random initial velocity\n p = Particle([self.c[0], self.c[1], self.c[2]],\n particle_size, self.origin,\n (random.uniform(-v_lim, 3*v_lim), random.uniform(-v_lim, 3*v_lim)),\n (0, 0),\n (0, 0),\n lifespan)\n ## Adds new particle to list of all particles in system\n self.particles.append(p)\n \n ## Method to move and update the system\n def moveParticles(self):\n\n ## Iterating over all Particle objects in system\n for particle in self.particles:\n \n ## Set current acceleration, force to zero\n particle.ax, particle.ay = 0, 0\n \n \n ## Apply weight force, random offset force\n if gravity == True:\n particle.applyForce(particle.weight)\n \n \n particle.applyForce((random.uniform(-offset, offset), (random.uniform(-offset, offset))))\n \n ## Deincrement lifespan\n particle.lifespan -= decay_speed\n \n ## Run movement algorithm\n particle.move()\n \n ## Draw particle to screen\n pygame.draw.circle(screen, particle.col, (int(particle.px), int(particle.py)), particle.radius)\n \n ## Check if particle is still alive\n if particle.isAlive() == False:\n ## If particle is dead, remove from particle list\n self.particles.remove(particle)\n \n self.positions.append((particle.px, particle.py))\n \n ## Changes colour (fading affect)\n if particleFade == True:\n particle.fade()\n \n particle.fx, particle.fy = 0, 0\n\n def finished(self):\n if len(self.particles) == 0:\n return True\n else:\n return False\n \n def applyForce(self, force):\n for i in self.particles:\n i.applyForce(force)\n \n def repel(self):\n for repeller in repellers:\n for p in self.particles:\n repeller.repel(p)\n\n## Get mouse position\nmouse = pygame.mouse.get_pos()\n\n## Initialise empty systems array to hold all particle systems\nsystems = []\nrepellers = []\n\n## Fill screen with black\nscreen.fill(BLACK)\n\n## Main runtime loop\nwhile win_loop:\n \n ## Refills background\n screen.fill(BLACK)\n \n ## Getting coordinates of mouse position\n mouse = pygame.mouse.get_pos()\n \n ## Iterates over all events\n for i in pygame.event.get():\n \n ## Checks for QUIT event to break loop\n if i.type == pygame.QUIT:\n win_loop = False\n \n ## Checks for mouse click to make new particle system\n if i.type == pygame.MOUSEBUTTONDOWN:\n \n ## Create random colour\n p = [random.randint(0,255), random.randint(0,255), random.randint(0,255)]\n \n ## Create new ParticleSystem object, origin=mouse position and col = random colour\n system = ParticleSystem(mouse, p)\n ## Generate particle objects at origin with makeEmitter method\n system.makeEmitter()\n \n ## Add new Particle system to list of all systems\n systems.append(system)\n \n \n if i.type == pygame.KEYDOWN:\n if i.key == pygame.K_RIGHT:\n windRight = True \n if i.key == pygame.K_LEFT:\n windLeft = True\n if i.key == pygame.K_SPACE:\n gravity = False\n \n if i.key == pygame.K_r:\n r = Repeller(repeller_strength, mouse)\n repellers.append(r)\n \n if i.type == pygame.KEYUP:\n if i.key == pygame.K_RIGHT:\n windRight = False\n if i.key == pygame.K_LEFT:\n windLeft = False\n \n else:\n gravity = True\n \n \n \n \n ## Iterate over all systems\n for system in systems:\n ## Generate particle objects at origin with makeEmitter method\n system.makeEmitter()\n \n if windRight == True:\n system.applyForce(wind_right_force)\n if windLeft == True:\n system.applyForce(wind_left_force)\n \n system.repel()\n \n ## Move and update all particle objects in system\n system.moveParticles()\n \n for repeller in repellers:\n pygame.draw.circle(screen, repeller.col, (int(repeller.px), int(repeller.py)), repeller.radius)\n\n\n ## Update screen\n pygame.display.flip()\n clock.tick(framerate)\n \n## Close window\npygame.quit()\nsys.exit()", "sub_path": "My Weird Physics Toys/Particle System Things/Particle Emitters.py", "file_name": "Particle Emitters.py", "file_ext": "py", "file_size_in_byte": 11879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.display.set_mode", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 55, "usage_type": "call"}, {"api_name": "pygame.time.Clock", "line_number": 58, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 58, "usage_type": "attribute"}, {"api_name": "math.sqrt", "line_number": 65, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 65, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 69, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 69, "usage_type": "call"}, {"api_name": "math.atan2", "line_number": 73, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 77, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 112, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 122, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 236, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 236, "usage_type": "attribute"}, {"api_name": "random.uniform", "line_number": 281, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 303, "usage_type": "call"}, {"api_name": "pygame.draw.circle", "line_number": 312, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 312, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 343, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 343, "usage_type": "attribute"}, {"api_name": "pygame.mouse.get_pos", "line_number": 359, "usage_type": "call"}, {"api_name": "pygame.mouse", "line_number": 359, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 362, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 362, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 365, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 369, "usage_type": "attribute"}, {"api_name": "random.randint", "line_number": 372, "usage_type": "call"}, {"api_name": "pygame.KEYDOWN", "line_number": 383, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 384, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 386, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 388, "usage_type": "attribute"}, {"api_name": "pygame.K_r", "line_number": 391, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 395, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 396, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 398, "usage_type": "attribute"}, {"api_name": "pygame.draw.circle", "line_number": 423, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 423, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 427, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 427, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 431, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 432, "usage_type": "call"}]} +{"seq_id": "384546042", "text": "# -*- coding: utf-8 -*-\n\nfrom . import Graph\nfrom pygsp.utils import build_logger\n\nfrom collections import Counter\nfrom copy import deepcopy\nimport numpy as np\nfrom scipy import sparse, spatial\n\n\nclass Community(Graph):\n r\"\"\"\n Create a community graph.\n\n Parameters\n ----------\n N : int\n Number of nodes (default = 256)\n Nc : int (optional)\n Number of communities (default = :math:`round(\\sqrt{N}/2)`)\n min_comm : int (optional)\n Minimum size of the communities (default = round(N/Nc/3))\n min_deg : int (optional)\n Minimum degree of each node (default = 0, NOT IMPLEMENTED YET)\n comm_sizes : int (optional)\n Size of the communities (default = random)\n size_ratio : float (optional)\n Ratio between the radius of world and the radius of communities (default = 1)\n world_density : float (optional)\n Probability of a random edge between two different communities (default = 1/N)\n comm_density : float (optional)\n Probability of a random edge inside any community (default = None, not used if None)\n k_neigh : int (optional)\n Number of intra-community connections (default = None, not used if None or comm_density is defined)\n epsilon : float (optional)\n Max distance at which two nodes sharing a community are connected\n (default = :math:`sqrt(2\\sqrt{N})/2`, not used if k_neigh or comm_density is defined)\n\n Examples\n --------\n >>> from pygsp import graphs\n >>> G = graphs.Community()\n\n \"\"\"\n def __init__(self, N=256, **kwargs):\n\n # Parameter initialisation #\n N = int(N)\n Nc = int(kwargs.pop('Nc', int(round(np.sqrt(N)/2.))))\n min_comm = int(kwargs.pop('min_comm', int(round(N / (3. * Nc)))))\n min_deg = int(kwargs.pop('min_deg', 0))\n comm_sizes = kwargs.pop('comm_sizes', np.array([]))\n size_ratio = float(kwargs.pop('size_ratio', 1.))\n world_density = float(kwargs.pop('world_density', 1. / N))\n world_density = world_density if 0 <= world_density <= 1 else 1. / N\n comm_density = kwargs.pop('comm_density', None)\n k_neigh = kwargs.pop('k_neigh', None)\n epsilon = float(kwargs.pop('epsilon', np.sqrt(2 * np.sqrt(N)) / 2))\n\n self.logger = build_logger(__name__, **kwargs)\n w_data = [[], [[], []]]\n\n try:\n if len(comm_sizes) > 0:\n if np.sum(comm_sizes) != N:\n raise ValueError('GSP_COMMUNITY: The sum of the community sizes has to be equal to N.')\n if len(comm_sizes) != Nc:\n raise ValueError('GSP_COMMUNITY: The length of the community sizes has to be equal to Nc.')\n\n except TypeError:\n raise TypeError(\"GSP_COMMUNITY: comm_sizes expected to be a list or array, got {}\".format(type(comm_sizes)))\n\n if min_comm * Nc > N:\n raise ValueError('GSP_COMMUNITY: The constraint on minimum size for communities is unsolvable.')\n\n info = {'node_com': None, 'comm_sizes': None, 'world_rad': None,\n 'world_density': world_density, 'min_comm': min_comm}\n\n # Communities construction #\n if comm_sizes.shape[0] == 0:\n mandatory_labels = np.tile(np.arange(Nc), (min_comm,)) # min_comm labels for each of the Nc communities\n remaining_labels = np.random.choice(Nc, N - min_comm * Nc) # random choice for the remaining labels\n info['node_com'] = np.sort(np.concatenate((mandatory_labels, remaining_labels)))\n else:\n # create labels based on the constraint given for the community sizes. No random assignation here.\n info['node_com'] = np.concatenate([[val] * cnt for (val, cnt) in enumerate(comm_sizes)])\n\n counts = Counter(info['node_com'])\n info['comm_sizes'] = np.array([cnt[1] for cnt in sorted(counts.items())])\n info['world_rad'] = size_ratio * np.sqrt(N)\n\n # Intra-community edges construction #\n if comm_density:\n # random picking edges following the community density (same for all communities)\n comm_density = float(comm_density)\n comm_density = comm_density if 0. <= comm_density <= 1. else 0.1\n info['comm_density'] = comm_density\n self.logger.info(\"GSP_COMMUNITY: Constructed using community density = {}\".format(comm_density))\n elif k_neigh:\n # k-NN among the nodes in the same community (same k for all communities)\n k_neigh = int(k_neigh)\n k_neigh = k_neigh if k_neigh > 0 else 10\n info['k_neigh'] = k_neigh\n self.logger.info(\"GSP_COMMUNITY: Constructed using K-NN with k = {}\".format(k_neigh))\n else:\n # epsilon-NN among the nodes in the same community (same eps for all communities)\n info['epsilon'] = epsilon\n self.logger.info(\"GSP_COMMUNITY: Constructed using eps-NN with eps = {}\".format(epsilon))\n\n # Coordinates #\n info['com_coords'] = info['world_rad'] * np.array(list(zip(\n np.cos(2 * np.pi * np.arange(1, Nc + 1) / Nc),\n np.sin(2 * np.pi * np.arange(1, Nc + 1) / Nc))))\n\n coords = np.random.rand(N, 2) # nodes' coordinates inside the community\n coords = np.array([[elem[0] * np.cos(2 * np.pi * elem[1]),\n elem[0] * np.sin(2 * np.pi * elem[1])] for elem in coords])\n\n for i in range(N):\n # set coordinates as an offset from the center of the community it belongs to\n comm_idx = info['node_com'][i]\n comm_rad = np.sqrt(info['comm_sizes'][comm_idx])\n coords[i] = info['com_coords'][comm_idx] + comm_rad * coords[i]\n\n first_node = 0\n for i in range(Nc):\n com_siz = info['comm_sizes'][i]\n M = com_siz * (com_siz - 1) / 2\n\n if comm_density:\n nb_edges = int(comm_density * M)\n tril_ind = np.tril_indices(com_siz, -1)\n indices = np.random.permutation(M)[:nb_edges]\n\n w_data[0] += [1] * nb_edges\n w_data[1][0] += [first_node + tril_ind[1][elem] for elem in indices]\n w_data[1][1] += [first_node + tril_ind[0][elem] for elem in indices]\n\n elif k_neigh:\n comm_coords = coords[first_node:first_node + com_siz]\n kdtree = spatial.KDTree(comm_coords)\n __, indices = kdtree.query(comm_coords, k=k_neigh + 1)\n\n pairs_set = set()\n map(lambda row: map(lambda elm: pairs_set.add((min(row[0], elm), max(row[0], elm))), row[1:]), indices)\n\n w_data[0] += [1] * len(pairs_set)\n w_data[1][0] += [first_node + pair[0] for pair in pairs_set]\n w_data[1][1] += [first_node + pair[1] for pair in pairs_set]\n\n else:\n comm_coords = coords[first_node:first_node + com_siz]\n kdtree = spatial.KDTree(comm_coords)\n pairs_set = kdtree.query_pairs(epsilon)\n\n w_data[0] += [1] * len(pairs_set)\n w_data[1][0] += [first_node + elem[0] for elem in pairs_set]\n w_data[1][1] += [first_node + elem[1] for elem in pairs_set]\n\n first_node += com_siz\n\n # Inter-community edges construction #\n M = (N**2 - np.sum([com_siz**2 for com_siz in info['comm_sizes']])) / 2\n nb_edges = int(world_density * M)\n\n if world_density < 0.35:\n # use regression sampling\n inter_edges = set()\n while len(inter_edges) < nb_edges:\n new_point = np.random.randint(0, N, 2)\n if info['node_com'][min(new_point)] != info['node_com'][max(new_point)]:\n inter_edges.add((min(new_point), max(new_point)))\n else:\n # use random permutation\n indices = np.random.permutation(M)[:nb_edges]\n all_points, first_col = [], 0\n for i in range(Nc - 1):\n nb_col = info['comm_sizes'][i]\n first_row = np.sum(info['comm_sizes'][:i+1])\n\n for j in range(i+1, Nc):\n nb_row = info['comm_sizes'][j]\n all_points += [(first_row + r, first_col + c) for r in range(nb_row) for c in range(nb_col)]\n\n first_row += nb_row\n first_col += nb_col\n\n inter_edges = np.array(all_points)[indices]\n\n w_data[0] += [1] * nb_edges\n w_data[1][0] += [elem[0] for elem in inter_edges]\n w_data[1][1] += [elem[1] for elem in inter_edges]\n\n w_data[0] += w_data[0]\n tmp_w_data = deepcopy(w_data[1][0])\n w_data[1][0] += w_data[1][1]\n w_data[1][1] += tmp_w_data\n w_data[1] = tuple(w_data[1])\n\n W = sparse.coo_matrix(tuple(w_data), shape=(N, N))\n\n for key, value in {'Nc': Nc, 'info': info}.items():\n setattr(self, key, value)\n\n super(Community, self).__init__(W=W, gtype='Community', coords=coords, **kwargs)\n", "sub_path": "pygsp/graphs/community.py", "file_name": "community.py", "file_ext": "py", "file_size_in_byte": 9025, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.sqrt", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 59, "usage_type": "call"}, {"api_name": "pygsp.utils.build_logger", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 66, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 83, "usage_type": "attribute"}, {"api_name": "numpy.sort", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 87, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 89, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 90, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 113, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 114, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 114, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 116, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 117, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 117, "usage_type": "attribute"}, {"api_name": "numpy.sin", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 118, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 123, "usage_type": "call"}, {"api_name": "numpy.tril_indices", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 134, "usage_type": "attribute"}, {"api_name": "scipy.spatial.KDTree", "line_number": 142, "usage_type": "call"}, {"api_name": "scipy.spatial", "line_number": 142, "usage_type": "name"}, {"api_name": "scipy.spatial.KDTree", "line_number": 154, "usage_type": "call"}, {"api_name": "scipy.spatial", "line_number": 154, "usage_type": "name"}, {"api_name": "numpy.sum", "line_number": 164, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 171, "usage_type": "attribute"}, {"api_name": "numpy.random.permutation", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 176, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 189, "usage_type": "call"}, {"api_name": "copy.deepcopy", "line_number": 196, "usage_type": "call"}, {"api_name": "scipy.sparse.coo_matrix", "line_number": 201, "usage_type": "call"}, {"api_name": "scipy.sparse", "line_number": 201, "usage_type": "name"}]} +{"seq_id": "132361406", "text": "import numpy as np\nimport xarray as xr\nfrom os import listdir, path, makedirs\nfrom datetime import datetime\n\nMERRA2 = '/n/mickley/users/ktoshima/MERRA2/'\nsave_path = path.join(MERRA2, 'mo')\nmakedirs(save_path, exist_ok=True)\n\nfor var in ['RH']:\n year = 1980\n dss = []\n for year_file in sorted(listdir(path.join(MERRA2, var))):\n file_path = path.join(MERRA2, var, year_file)\n ds = xr.open_dataset(file_path)\n for m in range(1, 13):\n start = np.datetime64('-'.join([str(year), str(m).zfill(2)]))\n end = start + np.timedelta64(1, 'M') - np.timedelta64(1, 'D')\n ds_mo = ds.sel(time=slice(start, end))\n dr_max = ds_mo[var + '_max'].max(axis=0, keepdims=True)\n dr_min = ds_mo[var + '_min'].min(axis=0, keepdims=True)\n dr_avg = ds_mo[var + '_avg'].mean(axis=0, keepdims=True)\n ds_T2M_mo = xr.merge([dr_max, dr_min, dr_avg])\n date = start\n ds_T2M_mo['time'] = ('time', [date])\n dss.append(ds_T2M_mo)\n year += 1\n comb_ds = xr.concat(dss, dim='time')\n comb_ds.to_netcdf(path.join(save_path, 'canada_{var}_mo.nc'.format(var=var)))\n", "sub_path": "merra_monthly.py", "file_name": "merra_monthly.py", "file_ext": "py", "file_size_in_byte": 1173, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 8, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 13, "usage_type": "call"}, {"api_name": "os.path", "line_number": 13, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "name"}, {"api_name": "xarray.open_dataset", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.datetime64", "line_number": 17, "usage_type": "call"}, {"api_name": "numpy.timedelta64", "line_number": 18, "usage_type": "call"}, {"api_name": "xarray.merge", "line_number": 23, "usage_type": "call"}, {"api_name": "xarray.concat", "line_number": 28, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "name"}]} +{"seq_id": "482189913", "text": "# -*- coding: utf-8 -*-\n\nfrom __future__ import unicode_literals\n\nfrom mock import Mock\nimport pytest\n\nfrom h.views import panels\n\n@pytest.mark.usefixtures('routes')\ndef test_navbar_sets_null_username_when_logged_out(pyramid_request):\n pyramid_request.authenticated_user = None\n result = panels.navbar({}, pyramid_request)\n assert result['username'] == None\n\n\n@pytest.mark.usefixtures('routes')\ndef test_navbar_sets_username_when_logged_in(pyramid_request, authenticated_user):\n pyramid_request.authenticated_user = authenticated_user\n result = panels.navbar({}, pyramid_request)\n\n assert result['username'] == 'vannevar'\n\n\n@pytest.mark.usefixtures('routes')\ndef test_navbar_lists_groups_when_logged_in(pyramid_request, authenticated_user):\n pyramid_request.authenticated_user = authenticated_user\n result = panels.navbar({}, pyramid_request)\n\n titles = [group.name for group in authenticated_user.groups]\n\n assert result['groups_menu_items'] == [\n {'title': titles[0], 'link': 'http://example.com/groups/id1/first'},\n {'title': titles[1], 'link': 'http://example.com/groups/id2/second'},\n ]\n\n\n@pytest.mark.usefixtures('routes')\ndef test_navbar_username_link_when_logged_in(pyramid_request, authenticated_user):\n pyramid_request.authenticated_user = authenticated_user\n result = panels.navbar({}, pyramid_request)\n\n assert result['username_link'] == 'http://example.com/search?q=user:vannevar'\n\n\n@pytest.fixture\ndef routes(pyramid_config):\n pyramid_config.add_route('account', '/account')\n pyramid_config.add_route('account_profile', '/account/profile')\n pyramid_config.add_route('account_notifications', '/account/notifications')\n pyramid_config.add_route('account_developer', '/account/developer')\n pyramid_config.add_route('activity.search', '/search')\n pyramid_config.add_route('group_create', '/groups/new')\n pyramid_config.add_route('group_read', '/groups/:pubid/:slug')\n pyramid_config.add_route('logout', '/logout')\n\n\n@pytest.fixture\ndef authenticated_user():\n groups = [\n Mock(pubid='id1', slug='first'),\n Mock(pubid='id2', slug='second'),\n ]\n authenticated_user = Mock(username='vannevar', groups=groups)\n return authenticated_user\n", "sub_path": "tests/h/views/panels_test.py", "file_name": "panels_test.py", "file_ext": "py", "file_size_in_byte": 2248, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "h.views.panels.navbar", "line_number": 13, "usage_type": "call"}, {"api_name": "h.views.panels", "line_number": 13, "usage_type": "name"}, {"api_name": "pytest.mark.usefixtures", "line_number": 10, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 10, "usage_type": "attribute"}, {"api_name": "h.views.panels.navbar", "line_number": 20, "usage_type": "call"}, {"api_name": "h.views.panels", "line_number": 20, "usage_type": "name"}, {"api_name": "pytest.mark.usefixtures", "line_number": 17, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 17, "usage_type": "attribute"}, {"api_name": "h.views.panels.navbar", "line_number": 28, "usage_type": "call"}, {"api_name": "h.views.panels", "line_number": 28, "usage_type": "name"}, {"api_name": "pytest.mark.usefixtures", "line_number": 25, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 25, "usage_type": "attribute"}, {"api_name": "h.views.panels.navbar", "line_number": 41, "usage_type": "call"}, {"api_name": "h.views.panels", "line_number": 41, "usage_type": "name"}, {"api_name": "pytest.mark.usefixtures", "line_number": 38, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 38, "usage_type": "attribute"}, {"api_name": "pytest.fixture", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mock.Mock", "line_number": 61, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 62, "usage_type": "call"}, {"api_name": "mock.Mock", "line_number": 64, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 58, "usage_type": "attribute"}]} +{"seq_id": "94212050", "text": "'''\n新增防具:防具提供,防護點數,受到攻擊先扣防護點數,防護點數不夠才扣血量\n商店賣武器:可加一般攻擊\n新增武器:武器可以增加一般攻擊力\n新增防具:防具提供,防護點數,受到攻擊先扣防護點數,防護點數不夠才扣血量\n'''\nfrom PIL import Image\nimport time\nimport random\n\ndef close_game(sts):\n path = 'output.txt'\n f = open(path, 'w')\n for i in range(len(sts)):\n f.write(str(sts[i]) + \"\\n\")\n f.close()\ndef read_file():\n path = 'output.txt'\n f = open(path, 'r')\n text = []\n for line in f:\n text.append(int(line))\n print(text)\n f.close()\n return text\ndef update_life(life):\n get_life = random.randint(1, 3)\n new_life = life + get_life\n print(\"Your recovry Life = %d Your life %d\" % (get_life, new_life))\n return new_life\n\ndef update_money(money):\n get_money = random.randint(30, 50)\n new_money = money + get_money\n print(\"Your Get Money = %d Your Money =%d\" % (get_money, new_money))\n return new_money\n\ndef fighting(life, magic, money, katana):\n status= [0, 0, 0, 0, 0]\n up_life = life\n up_magic = magic\n up_money = money\n monster_life = random.randint(2, 5)\n print(\"Monster Life = %d\" % monster_life)\n\n while True:\n act = input('Use magic attack , regular attack or katana? m/r/k:')\n if( act == 'm' and up_magic > 1):\n attack = random.randint(4, 10)\n up_magic -= 1\n elif( act == 'k' and katana == 1):\n filename = 'price.jpg'\n img = Image.open(filename)\n img.show()\n attack = random.randint(50, 100)\n katana -= 1\n else:\n print('使出一般攻擊、魔力點數未滿 或 錢不夠使用武士刀(k)!')\n attack = random.randint(1, 3)\n print(\"You make damage %d\" % attack)\n monster_life -= attack\n time.sleep(1)\n print(\"Monster Life %d\" % monster_life)\n if (monster_life < 1):\n print(\"You beat monster\")\n up_money += + random.randint(10, 20)\n status[0] = 1\n status[1] = up_life\n status[2] = up_magic\n status[3] = up_money\n return status\n print(\"Monster Attack\")\n time.sleep(1)\n up_life -= 3\n print(\"You get hurt, Life=%d\" % up_life)\n if ( up_life <= 0):\n print(\"You dead \\n\")\n status[0] = 0\n status[1] = up_life\n status[2] = up_magic\n status[3] = up_money\n return status\n\ndef store (life, magic, money, katana):\n status = [0, 0, 0, 0, 0]\n while True:\n act = input('what do you want to buy\\n'\n \"'l'/life:\\n\"\n \"'m'/magic:\\n\"\n \"'k'/katana:\\n\"\n \"'q'/quit:\\n\")\n if (act == 'q'):\n print('88')\n break\n elif(act == 'k' and money >= 10000):\n katana += 1\n money -= 10000\n print(' katana= %d Money = %d' % (katana, money))\n elif(money < 100) or (act == 'k' and money <= 10000):\n print('Money is not enough , quit store')\n break\n elif(act == 'l'):\n life += random.randint(1, 3)\n money -= 100\n print('Life = %d Money = %d' % (life, money))\n elif(act == 'm'):\n magic += random.randint(2, 6)\n money -= 100\n print('Magic = %d Money = %d' % (magic, money))\n else:\n continue\n status[0] = 1\n status[1] = life\n status[2] = magic\n status[3] = money\n status[4] = katana\n return status\n\n\n#main loop\n#sts=[1, 10, 0, 10000, 0] #是否生存/生命/錢/武士刀\nsts = read_file()\nwhile True:\n rev=input(\"Do you want 'c' continue 's'go to the store 'q' quit the game\")\n if ( rev == \"c\" ):\n gen_event = random.randint(1, 3)\n if ( gen_event == 1 ):\n sts[1] = update_life(sts[1])\n if ( gen_event == 2 ):\n sts[3] = update_money(sts[3])\n if ( gen_event == 3 ):\n sts = fighting(sts[1], sts[2], sts[3], sts[4])\n if( sts[0] == 0 ):\n print(\"Game Over\")\n break\n print(\"sts = %s\" %sts)\n elif ( rev == \"q\" ):\n print(\"88\")\n close_game(sts)\n break\n elif ( rev == \"s\" ):\n sts=store(sts[1], sts[2], sts[3], sts[4])\n print(\"%s\" % sts)\n else:\n continue", "sub_path": "class/10/adventure3.py", "file_name": "adventure3.py", "file_ext": "py", "file_size_in_byte": 4475, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "random.randint", "line_number": 27, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 33, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 43, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 49, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 53, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 53, "usage_type": "name"}, {"api_name": "random.randint", "line_number": 55, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 59, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 62, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 66, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 73, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 103, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 107, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 126, "usage_type": "call"}]} +{"seq_id": "10131306", "text": "import jieba #分词包\nimport matplotlib.pyplot as plt #matplotlib作为常用的可视化工具\nimport matplotlib\nfrom collections import Counter\nfrom wordcloud import WordCloud #词云包\nmatplotlib.rcParams['figure.figsize'] = (10.0, 5.0)\njieba.add_word('古北水镇')\n#停用词\nstop_word=[\":\",\"\\n\",\"的\",\"那\",\"如\",\"。\",\",\",\",\",\" \",\"!\",\"~\",\"了\",\"也\",\"是\",\"、\",\"“\",\"”\"]\ndata=[]\nwith open('英文.txt', 'r') as f:\n for line in f.readlines(): # 按行读文件\n #jieba分词\n line=jieba.cut(line)\n for i in line:\n #去掉停用词\n if i not in stop_word:\n data.append(i)\n\n#print(Counter(data))\n#统计词频并排序\ntop_1000 = Counter(data).most_common(1000)\n# data=[]\n# for i in top_1000:\n# #print(i)\n# data.append(i[0])\nmatplotlib.rcParams['figure.figsize'] = (12.0, 12.0) # 设定图像尺寸\nwordcloud=WordCloud(font_path=\"/System/Library/Fonts/STHeiti Light.ttc\",background_color=\"white\",max_font_size=80) # 设定词云的字体路径,防止中文出错\nword_frequence = {x[0]:x[1] for x in top_1000} # 取前1000个词频最高的词语\nprint(word_frequence)\nwordcloud=wordcloud.fit_words(word_frequence)\nplt.imshow(wordcloud)\nplt.show()\n\n", "sub_path": "词云/英文词云.py", "file_name": "英文词云.py", "file_ext": "py", "file_size_in_byte": 1261, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "matplotlib.rcParams", "line_number": 6, "usage_type": "attribute"}, {"api_name": "jieba.add_word", "line_number": 7, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 14, "usage_type": "call"}, {"api_name": "collections.Counter", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.rcParams", "line_number": 27, "usage_type": "attribute"}, {"api_name": "wordcloud.WordCloud", "line_number": 28, "usage_type": "call"}, {"api_name": "wordcloud.fit_words", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}]} +{"seq_id": "379535192", "text": "from django.urls import path, include\nfrom myuser import views\nfrom rest_framework.urlpatterns import format_suffix_patterns\nfrom rest_framework import renderers\nfrom rest_framework_jwt.views import obtain_jwt_token, refresh_jwt_token\n\n\nusers = views.MyUserViewSet.as_view({\n 'get': 'list',\n 'post': 'create',\n})\n\nuser_detail = views.MyUserViewSet.as_view({\n 'get': 'retrieve',\n})\n\nurlpatterns = format_suffix_patterns([\n path('', users, name='myusers'),\n path('/', user_detail, name='myuser-detail'),\n path('api-auth/',include('rest_framework.urls')),\n path('login/', obtain_jwt_token, name='login'),\n path('refresh/', refresh_jwt_token, name='token_refresh'),\n path('captcha/', views.get_captcha, name='get_captcha'),\n])\n\n", "sub_path": "tutorial/myuser/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 759, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "myuser.views.MyUserViewSet.as_view", "line_number": 8, "usage_type": "call"}, {"api_name": "myuser.views.MyUserViewSet", "line_number": 8, "usage_type": "attribute"}, {"api_name": "myuser.views", "line_number": 8, "usage_type": "name"}, {"api_name": "myuser.views.MyUserViewSet.as_view", "line_number": 13, "usage_type": "call"}, {"api_name": "myuser.views.MyUserViewSet", "line_number": 13, "usage_type": "attribute"}, {"api_name": "myuser.views", "line_number": 13, "usage_type": "name"}, {"api_name": "rest_framework.urlpatterns.format_suffix_patterns", "line_number": 17, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 18, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 19, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.include", "line_number": 20, "usage_type": "call"}, {"api_name": "django.urls.path", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework_jwt.views.obtain_jwt_token", "line_number": 21, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 22, "usage_type": "call"}, {"api_name": "rest_framework_jwt.views.refresh_jwt_token", "line_number": 22, "usage_type": "argument"}, {"api_name": "django.urls.path", "line_number": 23, "usage_type": "call"}, {"api_name": "myuser.views.get_captcha", "line_number": 23, "usage_type": "attribute"}, {"api_name": "myuser.views", "line_number": 23, "usage_type": "name"}]} +{"seq_id": "451599989", "text": "from __future__ import print_function\n\nimport argparse\nimport re\nimport sys\n\n\nIS_PY3 = sys.version_info[0] == 3\n\n\ndef main():\n parser = argparse.ArgumentParser(\n prog='bump',\n description=\"Bumps package version numbers\")\n parser.add_argument('file', help='File to update')\n parser.add_argument('-M', dest='major', action='store_true',\n help=\"Bump major version\")\n parser.add_argument('-m', dest='minor', action='store_true',\n help=\"Bump minor version\")\n parser.add_argument('-b', dest='build', action='store_true',\n help=\"Bump build version\")\n args = parser.parse_args()\n\n with open(args.file, 'rb') as f:\n m = re.search('\\s*version\\s*=\\s*(\\'|\")?([^\\'\",]+)(\\'|\")?',\n f.read().decode('utf-8'), re.I)\n\n if m:\n version_string = m.group(2)\n try:\n version = list(map(int, version_string.split('.')))\n except ValueError:\n print(\"Invalid version string:\", version_string)\n while len(version) < 3:\n version += [0]\n if args.major:\n version = version[0] + 1, 0, 0\n elif args.minor and len(version) > 1:\n version = version[0], version[1] + 1, 0\n elif len(version) > 2:\n version = version[0], version[1], version[2] + 1\n else:\n print(\"Invalid version string:\", version_string)\n\n new_version_string = '.'.join(map(str, version))\n print(version_string, '=>', new_version_string)\n\n __input = input if IS_PY3 else raw_input\n\n if __input('Is this ok? y/n ').lower() == 'y':\n with open(args.file, 'wb') as f:\n content = bytes(m.string, 'utf-8') if IS_PY3 else m.string\n if IS_PY3:\n new_version_string = bytes(new_version_string, 'utf-8')\n f.write(content[:m.start(2)] + new_version_string +\n content[m.end(2):])\n print('Updated', args.file)\n else:\n print('Canceled')\n\n else:\n print(\"No version definition found\")\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "bump.py", "file_name": "bump.py", "file_ext": "py", "file_size_in_byte": 2172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.version_info", "line_number": 8, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 12, "usage_type": "call"}, {"api_name": "re.search", "line_number": 25, "usage_type": "call"}, {"api_name": "re.I", "line_number": 26, "usage_type": "attribute"}]} +{"seq_id": "301309590", "text": "import os\n#os.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'\nimport tensorflow as tf\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn import metrics\nimport seaborn as sns\nfrom sklearn.model_selection import train_test_split\nfrom rnn_v2 import data_preprocessing\nimport pickle\n\nn_classes = 16\nnum_units = 64\ntime_steps = 50\nn_input = 6\nlearning_rate=0.01\n\nfile_name=[#'dataset_v2/HJH_2018_10_03_3_log.txt', 'dataset_v2/HJH_2018_10_04_3_log.txt',\n #'dataset_v2/HJH_2018_10_05_2_log.txt','dataset_v2/HJH_2018_10_06_3_log.txt',\n #'dataset_v2/HJH_2018_10_12_3_log.txt','dataset_v2/HJH_2018_10_13_1_log.txt',\n #'dataset_v2/HJH_2018_10_15_3_log.txt','dataset_v2/HJH_2018_10_16_1_log.txt',\n #'dataset_v2/HJH_2018_10_17_3_log.txt','dataset_v2/HJH_2018_10_22_3_log.txt',\n 'dataset_v2/HJH_2018_10_24_3_log.txt']\n\n\nget_df_data = data_preprocessing.get_data(file_name)\n\nreshaped_segments, reshaped_labels=data_preprocessing.data_shape(get_df_data)\n\n\n#데이터를 랜덤으로 섞어 훈련데이터와 테스트 데이터 setting함\nx_train, x_test, y_train, y_test = train_test_split(reshaped_segments, reshaped_labels, test_size=0.2, random_state=42)\n\n\n#weights and biases of appropriate shape to accomplish above task\nout_weights=tf.Variable(tf.random_normal([num_units,n_classes]),name=\"weights\")\nout_bias=tf.Variable(tf.random_normal([n_classes]),name=\"bias\")\n\n#defining placeholders\n#input image placeholder\nx=tf.placeholder(\"float\",[None,time_steps,n_input])\n#input label placeholder\ny=tf.placeholder(\"float\",[None,n_classes])\n\n#processing the input tensor from [batch_size,n_steps,n_input] to \"time_steps\" number of [batch_size,n_input] tensors\ninput=tf.unstack(x ,time_steps,1,name=\"input_tensor\")\n\n#defining the network\n#lstm_layer=rnn.BasicLSTMCell(num_units,forget_bias=1)\n#lstm_layer=rnn.MultiRNNCell([rnn.BasicLSTMCell(num_units) for _ in range(3)])\n#lstm_layer=rnn.LSTMBlockCell(num_units,forget_bias=1)\nlstm_layer=tf.nn.rnn_cell.BasicLSTMCell(num_units)\n#lstm_layer=tf.nn.rnn_cell.GRUCell(num_units)\n#lstm_layer=tf.nn.rnn_cell.LSTMCell(num_units,forget_bias=1)\noutputs,_= tf.contrib.rnn.static_rnn(lstm_layer,input,dtype=\"float32\")\n\n#converting last output of dimension [batch_size,num_units] to [batch_size,n_classes] by out_weight multiplication\nprediction=tf.add(tf.matmul(outputs[-1],out_weights), out_bias, name=\"output\")\n\n#loss_function\nloss=tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=prediction,labels=y))\n#optimization\nopt=tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(loss)\n\n#model evaluation\ncorrect_prediction=tf.equal(tf.argmax(prediction,1),tf.argmax(y,1))\naccuracy=tf.reduce_mean(tf.cast(correct_prediction,tf.float32))\n\n#initialize variables\ninit=tf.global_variables_initializer()\noutput_dir=\"./test\"\nbatch_size=100\nsaver = tf.train.Saver()\n\nwith tf.Session() as sess:\n sess.run(init)\n # added\n tf.train.write_graph(sess.graph_def, '.', output_dir + '/model.pbtxt')\n total_batch = int(x_train.shape[0] / batch_size)\n iter=1\n while iter<100:\n for i in range(total_batch):\n #한번에 batchc_size만큼 학습\n batch_x = x_train[i*batch_size:(i+1)*batch_size]\n batch_y = y_train[i*batch_size:(i+1)*batch_size]\n\n batch_x=batch_x.reshape((batch_size,time_steps,n_input))\n\n sess.run(opt, feed_dict={x: batch_x, y: batch_y})\n\n if iter %10==0:\n acc=sess.run(accuracy,feed_dict={x:batch_x,y:batch_y})\n los=sess.run(loss,feed_dict={x:batch_x,y:batch_y})\n print(\"For iter \",iter)\n print(\"Accuracy \",acc)\n print(\"Loss \",los)\n print(\"__________________\")\n\n\n\n filename = saver.save(sess, output_dir + '/model.ckpt')\n\n iter=iter+1\n", "sub_path": "rnn_v2/test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 3793, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "rnn_v2.data_preprocessing.get_data", "line_number": 26, "usage_type": "call"}, {"api_name": "rnn_v2.data_preprocessing", "line_number": 26, "usage_type": "name"}, {"api_name": "rnn_v2.data_preprocessing.data_shape", "line_number": 28, "usage_type": "call"}, {"api_name": "rnn_v2.data_preprocessing", "line_number": 28, "usage_type": "name"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 32, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 36, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.random_normal", "line_number": 37, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 41, "usage_type": "call"}, {"api_name": "tensorflow.placeholder", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.unstack", "line_number": 46, "usage_type": "call"}, {"api_name": "tensorflow.nn.rnn_cell.BasicLSTMCell", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tensorflow.contrib.rnn.static_rnn", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tensorflow.add", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.nn.softmax_cross_entropy_with_logits", "line_number": 61, "usage_type": "call"}, {"api_name": "tensorflow.nn", "line_number": 61, "usage_type": "attribute"}, {"api_name": "tensorflow.train.AdamOptimizer", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 63, "usage_type": "attribute"}, {"api_name": "tensorflow.equal", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.argmax", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 67, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 67, "usage_type": "attribute"}, {"api_name": "tensorflow.global_variables_initializer", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.train.Saver", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 73, "usage_type": "attribute"}, {"api_name": "tensorflow.Session", "line_number": 75, "usage_type": "call"}, {"api_name": "tensorflow.train.write_graph", "line_number": 78, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 78, "usage_type": "attribute"}]} +{"seq_id": "438839092", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Jun 3 17:03:39 2018\n@author: witek\n\n\n%matplotlib auto\n%matplotlib inline\n\n\na = CSTairfoil([0.05833333, 0.22333333, 0.255 , 0.11666667, 0.13833333], .5,1,\n [0.05166667, 0.1525 , 0.17 , 0.09166667, 0.045] , .5 ,1 , 100 )\n \n\"\"\"\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom auxiliaryFunctions import myArrays, readReport\nfrom xfoilTools import xFoilBB\n\nclass CSTcurve:\n def __init__(self,P,N1,N2, points):\n self.P = P\n self.N1 = N1\n self.N2 = N2 \n self.x = []\n self.y=[] \n self.CST_curve(P, points)\n \n def Bernstein(i,p,phi):\n # returns value of bernstein for single point (phi), inputs are parameters\n return np.math.factorial(p)/(np.math.factorial(i)*np.math.factorial(p-i))*(phi**i)*((1-phi)**(p-i))\n \n def C(phi, N1, N2):\n return phi**N1 * (1-phi)**N2\n \n def S(phi,P):\n sum=0\n l=len(P)\n for i in range(l):\n sum += P[i]*CSTcurve.Bernstein(i,l,phi)\n return sum\n \n def zeta(phi, P, N1, N2 ): \n return CSTcurve.C(phi, N1, N2) * CSTcurve.S(phi, P) + phi*.0005\n \n def CST_curve(self , P, points):\n #defines coordiantes of points which define CST curve\n \n self.x = myArrays(points).sine\n\n \n for i in range(points):\n #self.x.append( i/100. )\n self.y.append( CSTcurve.zeta(i/(points-1), self.P, self.N1, self.N2) ) \n \n def plot_CST(self):\n plt.figure()\n plt.plot(self.x, self.y)\n plt.grid(True) \n plt.title('CST curve formed using given vector P')\n plt.show()\n \n \n \"\"\"=========================================================\n ============================================================\n =========================================================\"\"\"\nclass CSTairfoil:\n \"\"\"\n initializes airfoil with specified parameters\n using CST curves for top and bottom surfaces\n \"\"\"\n \n def __init__(self,P, N1, N2, Q, M1, M2 ,points, num = 1):\n \"\"\"\n P and Q are vectors of variables for top and bottom airfoil curves\n Ns and Ms are parameters for defining smoothness of LE and TE\n points is number of points for each curve\n num is used for NX to set 3D wing section number\n \n \"\"\"\n \n \n \n self.top = CSTcurve(P, N1, N2, points)\n self.bottom = CSTcurve(Q, M1, M2, points)\n self.maxthickness = None\n self.area = None\n self.LD = None\n \n self.HFlift = None\n self.HFdrag = None\n self.LFlift = None\n self.LFdrag = None\n \n self.path = \"X:/OptimizationCode/\"\n airfoilpath = self.path+ \"airfoils/\"\n \n \n self.xfoilfile = airfoilpath+\"xfoilairfoil.dat\" \n self.icemfile = airfoilpath+\"icemairfoil.txt\"\n self.NXfile = airfoilpath+\"NX/nxairfoil{}.txt\".format(num)\n self.icemjournalpath = self.path + 'icem replays/replay.rpl'\n self.fluentjournalpath = self.path + 'fluent/journal/journal.jou'\n \n \n Pcase = []\n Pcase2 = ''\n Qcase = []\n Qcase2 = ''\n for i in range(len(P)):\n Pcase.append('{:.3f},'.format(P[i]))\n Pcase2 += Pcase[i]\n Qcase.append('{:.3f},'.format(Q[i]))\n Qcase2 += Qcase[i]\n \n self.casename = 'CST,'+Pcase2+Qcase2[:-1]\n \n for i in range(len(self.bottom.y)):\n self.bottom.y[i] = -self.bottom.y[i]\n \n \n def plotAirfoil(self):\n fig = plt.figure( figsize=(4, 1.5), dpi=300, facecolor='w', edgecolor='k' )\n plt.plot(self.top.x, self.top.y,self.bottom.x, self.bottom.y,linewidth = 1)\n #plt.grid(True)\n plt.axis('equal')\n plt.xlabel('x/c', fontsize=8)\n plt.ylabel('y/c', fontsize=8)\n plt.xticks(fontsize=7, rotation=0)\n plt.yticks(fontsize=7, rotation=0)\n plt.tightlayout()\n plt.show()\n \n def saveXFOIL(self):\n \"\"\" saves airfoil coords to .dat file with specified path \"\"\"\n # close trailing edge\n xfoilbottom = self.bottom.y\n xfoilbottom[-1] = 0\n xfoiltop = self.top.y\n xfoiltop[-1] = 0\n \n # save coords to file\n \n with open(self.xfoilfile, \"w\") as text_file:\n print(\"airfoilCST\", file = text_file)\n for i in range(len(self.bottom.x)-1,0,-1):\n print(\"{} {}\".format(self.bottom.x[i], xfoilbottom[i]), file=text_file)\n for i in range(len(self.top.x)):\n print(\"{} {}\".format(self.top.x[i], xfoiltop[i]), file=text_file)\n \n \n def saveICEM(self):\n \"\"\" saves airfoil coords in ICEM format with selected path \"\"\"\n \n xfoilbottom = self.bottom.y\n xfoilbottom[-1] = 0\n xfoiltop = self.top.y\n xfoiltop[-1] = 0\n \n # save coords to file\n \n with open(self.icemfile, \"w\") as text_file:\n print(\"{} 1\".format(len(self.bottom.x)+len(self.top.x)-3), file = text_file)\n for i in range(len(self.bottom.x)-2,0,-1):\n print(\"{} {} 0\".format(self.bottom.x[i], xfoilbottom[i]), file=text_file)\n for i in range( len(self.top.x)-1):\n print(\"{} {} 0\".format(self.top.x[i], xfoiltop[i]), file=text_file)\n \n def saveNX(self):\n \"\"\" saves airfoil coords in ICEM format with selected path \"\"\"\n \n xfoilbottom = self.bottom.y\n xfoilbottom[-1] = 0\n xfoiltop = self.top.y\n xfoiltop[-1] = 0\n z=0\n # save coords to file\n \n with open(self.NXfile, \"w\") as text_file:\n for i in range(len(self.bottom.x)-2,0,-1):\n print(\"{} {} {}\".format(self.bottom.x[i], xfoilbottom[i], z), file=text_file)\n for i in range( len(self.top.x)-1):\n print(\"{} {} {}\".format(self.top.x[i], xfoiltop[i], z), file=text_file) \n \n \n \n \n \n \n def evaluate(self):\n thickness = np.zeros(len(self.bottom.x))\n area=0\n for i in range(len(self.bottom.x)):\n thickness[i] = self.top.y[i]-self.bottom.y[i]\n try:\n area += thickness[i]*np.abs(self.top.x[i]-self.top.x[i+1])\n except IndexError: area +=0\n self.maxthickness = thickness.max()\n self.area = area\n \n \n \n \n def xFoilRun(self, Clreq, re = 500000, iters = 200, a1 = -1, a2 = 5, a3 = .1):\n \"\"\" runs xfoil blackbox of current airfoil with specified cl and other flow settings \"\"\"\n self.saveXFOIL()\n cd, alfa = xFoilBB(self.xfoilfile, Clreq)\n self.evaluate()\n if self.maxthickness<0.1:\n cd = cd* (1+(.1-self.maxthickness)*10)\n \n try:\n self.LD = Clreq/cd\n except ZeroDivisionError: self.LD=0\n self.alfa = alfa\n return cd\n \n \n \n \n \n def fluentRun(self, v, alfa):\n \"\"\" runs fluent analysis of current airfoil at specified angle of attack \"\"\"\n from scripts2 import journalFluent, journalICEM\n from xfoilTools import subprocess_cmd\n \n \n self.casename += ',{:.2f}'.format(alfa)\n \n self.saveICEM()\n # create journals\n journalICEM(self.icemjournalpath, self.icemfile)\n journalFluent(self.casename, self.fluentjournalpath, v , alfa, iters = 1000)\n \n # run commands, ICEM and Fluent\n ICEMrun ='\"C:\\\\Program Files\\\\ANSYS Inc\\\\v180\\\\icemcfd\\\\win64_amd\\\\bin\\\\icemcfd\" -script'\n ICEMscr = '\"X:\\OptimizationCode\\\\icem replays\\\\replay.rpl\"'\n ICEM = ICEMrun + ' ' + ICEMscr \n FLUENTrun = '\"C:\\\\Program Files\\\\ANSYS Inc\\\\v180\\\\fluent\\\\ntbin\\\\win64\\\\fluent.exe\" 2d -t6 -wait -i'\n FLUENTscr = '\"'+self.fluentjournalpath+'\"'\n FLUENT = FLUENTrun + ' '+ FLUENTscr\n \n subprocess_cmd(ICEM)\n subprocess_cmd(FLUENT)\n subprocess_cmd('del \"X:\\\\OptimizationCode\\\\fluent\\\\reports\\\\lift-rfile.out\"') \n subprocess_cmd('del \"X:\\\\OptimizationCode\\\\fluent\\\\reports\\\\drag-rfile.out\"') \n report = self.path+\"fluent/reports/report-\"+self.casename+\".out\"\n self.HFlift, self.HFdrag = readReport(report)\n \n", "sub_path": "CSTobjective.py", "file_name": "CSTobjective.py", "file_ext": "py", "file_size_in_byte": 8493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.math.factorial", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.math", "line_number": 31, "usage_type": "attribute"}, {"api_name": "auxiliaryFunctions.myArrays", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 123, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 123, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 124, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 124, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 126, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 126, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 128, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 128, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xticks", "line_number": 129, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 129, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.yticks", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 130, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tightlayout", "line_number": 131, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 131, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 196, "usage_type": "call"}, {"api_name": "xfoilTools.xFoilBB", "line_number": 207, "usage_type": "call"}, {"api_name": "scripts2.journalICEM", "line_number": 232, "usage_type": "call"}, {"api_name": "scripts2.journalFluent", "line_number": 233, "usage_type": "call"}, {"api_name": "xfoilTools.subprocess_cmd", "line_number": 243, "usage_type": "call"}, {"api_name": "xfoilTools.subprocess_cmd", "line_number": 244, "usage_type": "call"}, {"api_name": "xfoilTools.subprocess_cmd", "line_number": 245, "usage_type": "call"}, {"api_name": "xfoilTools.subprocess_cmd", "line_number": 246, "usage_type": "call"}, {"api_name": "auxiliaryFunctions.readReport", "line_number": 248, "usage_type": "call"}]} +{"seq_id": "302632419", "text": "import requests\nimport os\nfrom xml.dom import minidom\nfrom models import DomainInfo\nimport xmltodict\n\n__author__ = 'goran.vrbaski'\n\n\nclass NameSilo:\n def __init__(self, token: str, sandbox: bool=True):\n \"\"\"\n\n :param token: access token from namesilo.com\n :param sandbox: true or false\n \"\"\"\n self.__token = token\n if sandbox:\n self.__base_url = \"http://sandbox.namesilo.com/api/\"\n else:\n self.__base_url = \"https://www.namesilo.com/api/\"\n\n def __process_request(self, content):\n return minidom.parseString(content)\n\n def __get_content(self, url):\n return self.__process_request(requests.get(os.path.join(self.__base_url, url)).content.decode())\n\n def __get_content_xml(self, url):\n return xmltodict.parse(requests.get(os.path.join(self.__base_url, url)).content.decode())\n\n def check_domains(self, domain_names):\n \"\"\"\n\n :param domain_names:\n :return:\n \"\"\"\n available_domains = []\n url_extend = \"checkRegisterAvailability?version=1&type=xml&key=%s&domains=%s\" % (self.__token, domain_names)\n data = self.__get_content(url_extend)\n get_availabe_domains = data.getElementsByTagName(\"available\")\n for domain in get_availabe_domains[0].childNodes:\n available_domains.append(domain.childNodes[0].data)\n\n return available_domains\n\n def get_domain_info(self, domain_name):\n url_extend = \"getDomainInfo?version=1&type=xml&key=%s&domain=%s\" % (self.__token, domain_name)\n content = self.__get_content_xml(url_extend)\n return DomainInfo(content)\n\n def list_domains(self):\n \"\"\"\n :return: list of registered domains\n \"\"\"\n domain_list = []\n url_extend = \"listDomains?version=1&type=xml&key=%s\" % self.__token\n parsed_content = self.__get_content(url_extend)\n for domain in parsed_content.getElementsByTagName(\"domain\"):\n domain_list.append(domain.childNodes[0].data)\n\n return domain_list\n\n def register_domain(self, domain_name, years=1, auto_renew=0, private=0):\n \"\"\"\n\n :param domain_name: name of domain\n :param years:\n :param auto_renew:\n :param private:\n :return:\n \"\"\"\n url_extend = \"registerDomain?version=1&type=xml&key=%s&domain=%s&years=%s&private=%s&auto_renew=%s\" % \\\n (self.__token, domain_name, years, private, auto_renew)\n parsed_content = self.__get_content(url_extend)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2537, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "xml.dom.minidom.parseString", "line_number": 24, "usage_type": "call"}, {"api_name": "xml.dom.minidom", "line_number": 24, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 27, "usage_type": "call"}, {"api_name": "os.path", "line_number": 27, "usage_type": "attribute"}, {"api_name": "xmltodict.parse", "line_number": 30, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 30, "usage_type": "call"}, {"api_name": "os.path", "line_number": 30, "usage_type": "attribute"}, {"api_name": "models.DomainInfo", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "522897401", "text": "import unittest\nimport random\nfrom datetime import date\nfrom modules import analysis, scraper\nfrom inference.ranknode import RankNode\nimport preprocess\n\n# Test cases for Analysis API \nclass TestAnalysis(unittest.TestCase):\n\n # Method to test fantasy recommendations \n def test_fantasy_rec(self):\n score_list = analysis.fantasy_recommendations()\n self.assertTrue(len(score_list) > 100, \"Missing Players\")\n self.assertTrue(score_list[0][1] > score_list[-1][1])\n \n # Method to test dataframe creation \n def test_create_df(self):\n df, p_map = analysis.create_player_dataframe()\n self.assertTrue(True)\n \n # Method to test clustering algorithm \n def test_cluster(self):\n c = 100\n res = analysis.build_stat_clusters(c)\n self.assertTrue(len(res), c)\n \n def test_query_filter(self):\n flag_1 = analysis.isNBA(\"Will this team make it to the finals?\")\n flag_0 = analysis.isNBA(\"Michael Jordan is good!\")\n flag_n = analysis.isNBA(\"This is a random query...\")\n self.assertEqual(flag_1, 1)\n self.assertEqual(flag_0, 0)\n self.assertEqual(flag_n, -1)\n\n# Test cases for web scraper \nclass TestScraper(unittest.TestCase):\n\n # Method to test web scraper for Player Efficiency Rating\n def test_get_per(self):\n year = int(date.today().year)\n per_list = scraper.get_per(year)\n for t in per_list:\n self.assertTrue(isinstance(t[0], str))\n self.assertTrue(isinstance(t[1], float))\n \n # Method to test scraper for player names\n def test_get_names(self):\n year = int(date.today().year)\n names = scraper.get_player_names(year)\n for n in names:\n self.assertTrue(isinstance(n, str))\n \n # Method to test scraper for playoff bracket\n def test_get_playoff_bracket(self):\n bracket = scraper.get_playoff_bracket()\n levels = {\n \"Western Conference First Round\",\n \"Eastern Conference First Round\",\n \"Western Conference Semifinals\",\n \"Eastern Conference Semifinals\",\n \"Western Conference Finals\",\n \"Eastern Conference Finals\",\n \"Finals\"\n }\n for l in bracket:\n self.assertTrue(l in levels)\n \n # Method to test get target name \n def test_get_target_name(self):\n lebron_jamess = [\"LeBron James\", \"Lebron James\", \"Lebron\"]\n for i in lebron_jamess:\n self.assertEqual(scraper.get_target_name(i), \"LeBron James\")\n \n shaquille_oneals = [\"Shaquille O'Neal\", \"Shaq O'Niel\", \"Shakiel O'Neal\"]\n for i in shaquille_oneals:\n self.assertEqual(scraper.get_target_name(i), \"Shaquille O'Neal\")\n \n klay_thompsons = [\"Klay Thompson\", \"Clay Thompson\", \"Clay Thomson\", \"Klay Thomson\"]\n for i in klay_thompsons:\n self.assertEqual(scraper.get_target_name(i), \"Klay Thompson\")\n \n kobe_bryants = [\"Kobe Bryant\", \"kobe\", \"Kobe\"] \n for i in kobe_bryants:\n self.assertEqual(scraper.get_target_name(i), \"Kobe Bryant\")\n \n wilt_chamberlains = [\"Wilt Chamberlain\", \"Wilt\", \"wilt\"] \n for i in wilt_chamberlains:\n self.assertEqual(scraper.get_target_name(i), \"Wilt Chamberlain\")\n\n dennis_rodmans = [\"Dennis Rodman\", \"Rodman\", \"rodman\"] \n for i in dennis_rodmans:\n self.assertEqual(scraper.get_target_name(i), \"Dennis Rodman\")\n\n # Method to test get player url \n def test_get_player_url(self):\n self.assertEqual(scraper.get_player_url(\"Kobe Bryant\"), \"https://www.basketball-reference.com/players/b/bryanko01.html\")\n self.assertEqual(scraper.get_player_url(\"LeBron James\"), \"https://www.basketball-reference.com/players/j/jamesle01.html\")\n self.assertEqual(scraper.get_player_url(\"Dennis Rodman\"), \"https://www.basketball-reference.com/players/r/rodmade01.html\")\n\n # Method to test advanced stat scraper \n def test_get_adv_stats(self):\n names = [\"Kobe Bryant\", \"Lebron James\", \"Klay Thompson\"]\n stats = [\"true shooting percentage\", \"total rebound percentage\", \"defensive plus/minus\"]\n for i in range(5):\n random_name = random.choice(names)\n random_stat = random.choice(stats)\n stat = scraper.get_adv_stat(random_name, random_stat)\n self.assertTrue(isinstance(stat, float))\n\n# Test cases for rank node\nclass TestRankNode(unittest.TestCase):\n\n # Method to test rank node response \n def test_node_response(self):\n query = \"query\"\n node = RankNode(query)\n resp = node.response()\n stat = [int(word) for word in resp.split() if word.replace('.','').isdigit()]\n self.assertTrue(isinstance(resp, str))\n \n # Method to test rank node metric conversion\n def test_metric2stat(self):\n node = RankNode(\"Query\")\n test_map = {\n \"true shooting percentage\" : \"shooting\",\n \"defensive plus/minus\" : \"defending\",\n \"player efficiency rating\" : \"player\",\n }\n\n for stat in test_map:\n metric = test_map[stat]\n predicted_stat = node.metric2stat(metric)\n self.assertEqual(predicted_stat, stat)\n\n metric = \"This is nothing\"\n predicted_stat = node.metric2stat(metric)\n self.assertIsNone(predicted_stat)\n \n # Method to test metric extraction\n def test_extract_metric(self):\n node = RankNode(\"Who is a better shooter Kobe or Lebron?\")\n metric = node.extract_metric()\n self.assertEqual(metric, \"shooter\")\n \n # Method to test name extraction\n def test_extract_names(self):\n node = RankNode(\"Who is a better shooter Kobe Bryant or Lebron James?\")\n name1, name2 = node.extract_names()\n names = set([name1, name2])\n self.assertTrue(\"Kobe Bryant\" in names)\n self.assertTrue(\"Lebron James\" in names)\n \n # Method to test stat getter\n def test_get_stat(self):\n node = RankNode(\"Query\")\n names = [\"Kobe Bryant\", \"Lebron James\", \"Klay Thompson\"]\n stats = [\"true shooting percentage\", \"total rebound percentage\", \"defensive plus/minus\"]\n for i in range(5):\n random_name = random.choice(names)\n random_stat = random.choice(stats)\n stat = node.get_stat(random_name, random_stat)\n self.assertTrue(isinstance(stat, float))\n\n\nif __name__ == '__main__':\n unittest.main()", "sub_path": "test.py", "file_name": "test.py", "file_ext": "py", "file_size_in_byte": 6520, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 9, "usage_type": "attribute"}, {"api_name": "modules.analysis.fantasy_recommendations", "line_number": 13, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 13, "usage_type": "name"}, {"api_name": "modules.analysis.create_player_dataframe", "line_number": 19, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 19, "usage_type": "name"}, {"api_name": "modules.analysis.build_stat_clusters", "line_number": 25, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 25, "usage_type": "name"}, {"api_name": "modules.analysis.isNBA", "line_number": 29, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 29, "usage_type": "name"}, {"api_name": "modules.analysis.isNBA", "line_number": 30, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 30, "usage_type": "name"}, {"api_name": "modules.analysis.isNBA", "line_number": 31, "usage_type": "call"}, {"api_name": "modules.analysis", "line_number": 31, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 37, "usage_type": "attribute"}, {"api_name": "datetime.date.today", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 41, "usage_type": "name"}, {"api_name": "modules.scraper.get_per", "line_number": 42, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 42, "usage_type": "name"}, {"api_name": "datetime.date.today", "line_number": 49, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 49, "usage_type": "name"}, {"api_name": "modules.scraper.get_player_names", "line_number": 50, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 50, "usage_type": "name"}, {"api_name": "modules.scraper.get_playoff_bracket", "line_number": 56, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 56, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 73, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 73, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 77, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 77, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 81, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 81, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 85, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 85, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 89, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 89, "usage_type": "name"}, {"api_name": "modules.scraper.get_target_name", "line_number": 93, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 93, "usage_type": "name"}, {"api_name": "modules.scraper.get_player_url", "line_number": 97, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 97, "usage_type": "name"}, {"api_name": "modules.scraper.get_player_url", "line_number": 98, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 98, "usage_type": "name"}, {"api_name": "modules.scraper.get_player_url", "line_number": 99, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 99, "usage_type": "name"}, {"api_name": "random.choice", "line_number": 106, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 107, "usage_type": "call"}, {"api_name": "modules.scraper.get_adv_stat", "line_number": 108, "usage_type": "call"}, {"api_name": "modules.scraper", "line_number": 108, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 112, "usage_type": "attribute"}, {"api_name": "inference.ranknode.RankNode", "line_number": 117, "usage_type": "call"}, {"api_name": "inference.ranknode.RankNode", "line_number": 124, "usage_type": "call"}, {"api_name": "inference.ranknode.RankNode", "line_number": 142, "usage_type": "call"}, {"api_name": "inference.ranknode.RankNode", "line_number": 148, "usage_type": "call"}, {"api_name": "inference.ranknode.RankNode", "line_number": 156, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 160, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 161, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 167, "usage_type": "call"}]} +{"seq_id": "610928245", "text": "import json\n\nfrom src.model.Sample import Sample\nfrom src.sbs_pipeline_helper import *\nfrom src.service import analysis_service, crop_service, request_service, \\\n tx_method_service\n\n\ndef parse(file_path):\n if os.path.exists(file_path):\n data = read_tab_separated_file(file_path)\n output_dict = {}\n if data:\n for row in data:\n key = row[0]\n val = row[1]\n output_dict[key] = val\n request_id = output_dict.get('Genomic_Request_ID', None)\n sample_id = output_dict.get('Sample_ID', None)\n php_id = output_dict.get('PHP_ID', None)\n prep_method = output_dict.get('Prep_Method', None)\n sbs_version = output_dict.get('SBS_VERSION', None)\n sbs_status = output_dict.get('SBS_STATUS', None)\n organism = output_dict.get('Organism', None)\n geno_type = output_dict.get('Genotype', None)\n ref_genome = output_dict.get('REFERENCE_GENOME', None)\n event_id = output_dict.get('Event_ID', None)\n secondary_php_ids = output_dict.get('SECONDARY_PHP_IDS', None)\n backbone_call = output_dict.get('Backbone_Call', None)\n on_target_rate = output_dict.get('ON_TARGET_RATE', None)\n\n sample = Sample(request_id, sbs_version, prep_method,\n sample_id, sbs_status, geno_type, ref_genome,\n event_id, php_id, backbone_call, on_target_rate,\n organism, secondary_php_ids)\n return sample\n else:\n raise Exception('File path [{}] does not exists'\n .format(file_path))\n\n\ndef get_sample_metadata(sample_id):\n \"\"\"\n Gets sample metadata for a sample from PSR\n \"\"\"\n requests.packages.urllib3.disable_warnings()\n psr_url = \"{}/samples/{}\".format(psr_host, sample_id)\n try:\n sample_mdata = requests.get(psr_url, verify=False).json()\n logger.info(\"psr url '%s' and return sample meta data '%s' \", psr_url, sample_mdata)\n return sample_mdata\n except Exception as e:\n logger.error(\"Error while retrieving PSR data for sample \\\n '%s': %s\", sample_id, e)\n raise \n\n\ndef save(analysis_id, sample):\n # save crop details\n crop_dict = {\"organism\": sample.organism}\n crop_res = crop_service.save(crop_dict)\n\n sample_metadata = get_sample_metadata(sample.sample_id)\n\n # save transformation_method details\n tx_method_dict = {\"tx_method\": sample_metadata['transformation_method'],\n \"organism\": sample.organism}\n tx_method_res = tx_method_service.save(tx_method_dict)\n\n # save request details\n request_dict = {}\n if 'method_name' in tx_method_res['data'] \\\n and tx_method_res['status']['message'] == 'Success':\n request_dict[\"tx_method\"] = tx_method_res['data']['method_name']\n\n if 'id' in crop_res.get('data') and crop_res.get('status').get('message') == 'Success':\n request_dict[\"request_id\"] = sample.request_id\n request_dict[\"organism_id\"] = crop_res.get('data').get('id')\n request_dict[\"sbs_status\"] = sample.sbs_status\n res = request_service.save(request_dict)\n\n # update sample details\n sample_params = {\"geno_type\": sample.geno_type,\n \"organism\": sample.organism,\n \"request_id\": sample.request_id,\n \"event_id\": sample.event_id,\n \"construct_name\": sample.php_id,\n \"sample_name\": sample_metadata['sample_name'],\n \"eu_id\": sample_metadata['euid'],\n \"si_id\": sample_metadata['siid']\n }\n\n logger.info(\"Updating sample...\")\n res = requests.put(\"{}/samples/{}\".format(api_url, sample.sample_id),\n json.dumps(sample_params),\n auth=client_auth\n ).json()\n\n logger.info(\"Response: {}\".format(json.dumps(res, indent=4)))\n\n # update analysis details\n analysis_dict = {\"sample_id\": sample.sample_id,\n \"reference\": sample.ref_genome,\n \"analysis_id\": analysis_id,\n \"construct_name\": sample.php_id,\n \"sbs_version\": sample.sbs_version,\n \"backbone_call\": sample.backbone_call,\n \"geno_type\": sample.geno_type,\n \"organism\": sample.organism,\n \"event_id\": sample.event_id,\n \"target_rate\": sample.on_target_rate,\n \"single_read_count\": sample_metadata['single_read_count'],\n \"paired_read_count\": sample_metadata['paired_read_count'],\n \"prep_method\": sample.prep_method,\n \"job_status\": \"Awaiting QC\"\n }\n\n logger.info(\"Adding Pipeline configuration to database from file {}\".format(sample.prep_method))\n pipeline_config_res = requests.post(\"{}/pipelines/configs/s3/{}.txt\".format(api_url, sample.prep_method),\n auth=client_auth)\n raise_for_status(pipeline_config_res)\n logger.info(\"Pipeline configuration added successfully : {}\".format(pipeline_config_res.json()))\n if pipeline_config_res.json() and pipeline_config_res.json()['data']:\n analysis_dict['configuration_id'] = pipeline_config_res.json()[\"data\"][\"id\"]\n\n analysis_service.update(analysis_dict)\n \n # save primary and secondary map records\n constructs = []\n if sample.php_id:\n constructs.append(sample.php_id)\n if sample.secondary_php_ids:\n constructs.extend(sample.secondary_php_ids.split(\";\"))\n constructs = [map_name for map_name in constructs if map_name]\n try:\n logger.info(\"Available maps in driver-file {}:\".format(constructs))\n if constructs:\n for construct_id in constructs:\n map_res = requests.post(\"{}/maps\".format(api_url),\n json.dumps({\"construct_id\": construct_id}),\n auth=client_auth\n )\n raise_for_status(map_res)\n logger.info(\"Saved construct id: {}\".format(construct_id))\n\n map = map_res.json().get('data')\n map_analysis_res = requests.post(\"{}/analyses/{}/maps/{}\"\n .format(api_url, analysis_id,\n map['id']),\n json.dumps({\"read_count\": None}),\n auth=client_auth\n )\n raise_for_status(map_analysis_res)\n logger.info('Saved map-analysis record for construct: {} with response: {}' \\\n .format(construct_id, map_analysis_res.json()))\n except Exception as e:\n logger.error(\"Failed to add maps info with error as {}\".format(str(e)))\n raise\n", "sub_path": "src/service/sample_service.py", "file_name": "sample_service.py", "file_ext": "py", "file_size_in_byte": 7061, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "src.model.Sample.Sample", "line_number": 32, "usage_type": "call"}, {"api_name": "src.service.crop_service.save", "line_number": 61, "usage_type": "call"}, {"api_name": "src.service.crop_service", "line_number": 61, "usage_type": "name"}, {"api_name": "src.service.tx_method_service.save", "line_number": 68, "usage_type": "call"}, {"api_name": "src.service.tx_method_service", "line_number": 68, "usage_type": "name"}, {"api_name": "src.service.request_service.save", "line_number": 80, "usage_type": "call"}, {"api_name": "src.service.request_service", "line_number": 80, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 95, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 99, "usage_type": "call"}, {"api_name": "src.service.analysis_service.update", "line_number": 126, "usage_type": "call"}, {"api_name": "src.service.analysis_service", "line_number": 126, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 140, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 150, "usage_type": "call"}]} +{"seq_id": "123946396", "text": "# coding: utf-8\nimport os\nimport setuptools\n\nDESCRIPTION = \"Translation Gummy is a magical gadget which enables user to be able to speak and understand other languages.\"\n\nhere = os.path.abspath(os.path.dirname(__file__))\nwith open(os.path.join(here, \"README.md\"), encoding=\"utf-8\") as f:\n LONG_DESCRIPTION = f.read()\nwith open(\"requirements.txt\", mode=\"r\") as f:\n INSTALL_REQUIRES = [line.rstrip(\"\\n\") for line in f.readlines() if line[0]!=(\"#\")]\n\ndef setup_package():\n metadata = dict(\n name=\"Translation-Gummy\",\n version=\"0.1.0\",\n description=DESCRIPTION,\n long_description=LONG_DESCRIPTION,\n long_description_content_type=\"text/markdown\",\n author=\"Shuto Iwasaki\",\n author_email=\"cabernet.rock@gmail.com\",\n license=\"MIT\",\n project_urls={\n \"Documentation\" : \"https://iwasakishuto.github.io/Translation-Gummy/index.html\",\n \"Bug Reports\" : \"https://github.com/iwasakishuto/Translation-Gummy/issues\",\n \"Source Code\" : \"https://github.com/iwasakishuto/Translation-Gummy\",\n \"Say Thanks!\" : \"https://twitter.com/cabernet_rock\",\n },\n packages=setuptools.find_packages(),\n package_data={\"gummy\": [\"templates/*\"]},\n python_requires=\">=3.6\",\n install_requires=INSTALL_REQUIRES,\n extras_require={\n \"tests\": [\"pytest\"],\n },\n classifiers=[\n \"Development Status :: 5 - Production/Stable\",\n \"Environment :: Console\",\n \"Intended Audience :: Other Audience\",\n \"License :: OSI Approved :: MIT License\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Topic :: Software Development :: Libraries\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n ],\n entry_points = {\n \"console_scripts\": [\n \"gummy-journal=gummy.cli:translate_journal\",\n \"gummy-translate=gummy.cli:translate_text\",\n ],\n },\n )\n setuptools.setup(**metadata)\n\nif __name__ == \"__main__\":\n setup_package()\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 2256, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "setuptools.find_packages", "line_number": 29, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 55, "usage_type": "call"}]} +{"seq_id": "297750976", "text": "#!/usr/bin/env python3\nimport dataclasses\nfrom typing import Generator, List\n\nfrom colorama import Fore\n\nfrom pwncat.enumerate import FactData\nfrom pwncat import util\nimport pwncat\n\nname = \"pwncat.enumerate.system\"\nprovides = \"system.init\"\nper_user = False\n\n\n@dataclasses.dataclass\nclass InitSystemData(FactData):\n\n init: util.Init\n version: str\n\n def __str__(self):\n return f\"Running [blue]{self.init}[/blue]\"\n\n @property\n def description(self):\n return self.version\n\n\ndef enumerate() -> Generator[FactData, None, None]:\n \"\"\"\n Enumerate system init service\n :return:\n \"\"\"\n\n init = util.Init.UNKNOWN\n version = None\n\n # Try to get the command name of the running init process\n try:\n with pwncat.victim.open(\"/proc/1/comm\", \"r\") as filp:\n comm = filp.read().strip()\n if comm is not None:\n if \"systemd\" in comm.lower():\n init = util.Init.SYSTEMD\n elif \"sysv\" in comm.lower():\n init = util.Init.SYSV\n elif \"upstart\" in comm.lower():\n init = util.Init.UPSTART\n except (PermissionError, FileNotFoundError):\n comm = None\n\n # Try to get the command name of the running init process\n try:\n with pwncat.victim.open(\"/proc/1/cmdline\", \"r\") as filp:\n comm = filp.read().strip().split(\"\\x00\")[0]\n except (PermissionError, FileNotFoundError):\n comm = None\n\n if comm is not None:\n if \"systemd\" in comm.lower():\n init = util.Init.SYSTEMD\n elif \"sysv\" in comm.lower():\n init = util.Init.SYSV\n elif \"upstart\" in comm.lower():\n init = util.Init.UPSTART\n\n with pwncat.victim.subprocess(f\"{comm} --version\", \"r\") as filp:\n version = filp.read().decode(\"utf-8\").strip()\n if \"systemd\" in version.lower():\n init = util.Init.SYSTEMD\n elif \"sysv\" in version.lower():\n init = util.Init.SYSV\n elif \"upstart\" in version.lower():\n init = util.Init.UPSTART\n\n # No need to provide an empty version string. They apparently don't support \"--version\"\n if version == \"\":\n version = None\n\n yield InitSystemData(init, version)\n", "sub_path": "pwncat/enumerate/system/init.py", "file_name": "init.py", "file_ext": "py", "file_size_in_byte": 2203, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pwncat.enumerate.FactData", "line_number": 17, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 19, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 16, "usage_type": "attribute"}, {"api_name": "pwncat.util.Init", "line_number": 36, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 36, "usage_type": "name"}, {"api_name": "pwncat.victim.open", "line_number": 41, "usage_type": "call"}, {"api_name": "pwncat.victim", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pwncat.util.Init", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 45, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 47, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 49, "usage_type": "name"}, {"api_name": "pwncat.victim.open", "line_number": 55, "usage_type": "call"}, {"api_name": "pwncat.victim", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pwncat.util.Init", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 62, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 64, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 66, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 66, "usage_type": "name"}, {"api_name": "pwncat.victim.subprocess", "line_number": 68, "usage_type": "call"}, {"api_name": "pwncat.victim", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pwncat.util.Init", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 71, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 73, "usage_type": "name"}, {"api_name": "pwncat.util.Init", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pwncat.util", "line_number": 75, "usage_type": "name"}, {"api_name": "typing.Generator", "line_number": 30, "usage_type": "name"}, {"api_name": "pwncat.enumerate.FactData", "line_number": 30, "usage_type": "name"}]} +{"seq_id": "549590934", "text": "from collections import deque\n\nn, m = map(int, input().split())\nshape = []\nfor _ in range(n):\n shape.append(list(input()))\n\ndx = [-1, 1, 0, 0]\ndy = [0, 0, -1, 1]\n \ngraph = {} # 2차원 배열을 그래프로 변경\nfor i in range(n):\n for j in range(m):\n graph[(i,j)] = []\n if shape[i][j] == '-': # '-' 모양인 경우\n for k in range(2,4):\n if 0 <= i+dx[k] <= n-1 and 0 <= j+dy[k] <= m-1: # 같은 행에 인접한 인덱스가 범위 안에 있는 경우\n if shape[i][j] == shape[i+dx[k]][j+dy[k]]: # 같은 모양인지 확인\n graph[(i,j)].append((i+dx[k], j+dy[k]))\n else: # '|' 모양인 경우\n for k in range(0,2):\n if 0 <= i+dx[k] <= n-1 and 0 <= j+dy[k] <= m-1: # 같은 열에 인접한 인덱스가 범위 안에 있는 경우\n if shape[i][j] == shape[i+dx[k]][j+dy[k]]: # 같은 모양인지 확인\n graph[(i,j)].append((i+dx[k], j+dy[k]))\n\ndef bfs(graph, start, visited):\n if visited[start[0]][start[1]] != True: # 방문한 적이 없는 경우\n queue = deque([start])\n visited[start[0]][start[1]] = True # 방문 처리\n while queue:\n v = queue.popleft()\n for i in graph[v]:\n if visited[i[0]][i[1]] != True:\n queue.append(i)\n visited[i[0]][i[1]] = True # 같은 나무 판자 방문 처리\n return True\n return False # 이미 방문한 경우\n\nresult = 0\nfor i in range(n):\n for j in range(m):\n if bfs(graph, (i,j), shape) == True: # 방문한 것이 없는 경우 같은 나무 판자 모두 방문 처리\n result += 1\n \nprint(result)", "sub_path": "BOJ/1388.py", "file_name": "1388.py", "file_ext": "py", "file_size_in_byte": 1783, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "collections.deque", "line_number": 28, "usage_type": "call"}]} +{"seq_id": "14824923", "text": "#!/bin/env python\n\nimport sys,os,glob,re\nfrom timeit import default_timer as timer\n\nimport tensorflow as tf\nfrom tensorflow.keras import layers\nfrom tensorflow.keras import backend as K\n\nimport numpy as np\nfrom numpy import genfromtxt\n\n#define arguments required for this script\nimport argparse\nparser = argparse.ArgumentParser()\n\nparser.add_argument('-J', '--MMR', default=2, required=True, help='Expected number of DNA repair deficiency signatures.')\nparser.add_argument('--noweights', action='store_true', default=False, required=False, help='No weights for genomic regions, i.e. set W and R to 1. Do not optimalize W and R.')\nparser.add_argument('-t', '--tensorboard', required=False, help='path to tensorboard folder. tensorboard logging will be disabled if not specified')\nparser.add_argument('-s', '--logsuffix', required=False, help='optional suffix to attach to the log counter in tensorboard')\nparser.add_argument('-o', '--out', required=False, help='path to the output file into which the trained tensors will be written')\nparser.add_argument('-i', '--input', required=True, help='path to the input file containing the given tensors')\n\nargs = parser.parse_args()\n\n# creates a configuration\ndef getConfig():\n config = {}\n\n # get the tensorboard logging folder if specified\n if args.tensorboard != None:\n\n logsuffix = \"\" if args.logsuffix == None else \"_\"+args.logsuffix\n\n maxnum = 0\n for the_file in os.listdir(args.tensorboard+\"/\"):\n if os.path.isdir(args.tensorboard+\"/\"+the_file):\n try:\n counter_prefix = re.search(r'\\d+', the_file).group()\n maxnum = max(maxnum, int(counter_prefix))\n except:\n pass\n\n config['tensorboard'] = os.path.abspath(args.tensorboard+\"/\"+str(maxnum+1)+logsuffix)\n\n else:\n config['tensorboard'] = None\n\n if args.out != None:\n config['output'] = os.path.abspath(args.out)\n else:\n config['output'] = None\n\n config['input'] = os.path.abspath(args.input)\n\n # Add data dimensions\n config['G'] = None # Number of patient, samples, genomes\n config['L'] = None # Number of genomic regions, e.g. early/late replication region, inter/intragenic\n config['K'] = None # Number of mutations categories, e.g. A[C>T]G, C[T>A]G\n config['N'] = None # Number of PRIMARY mutational signatures\n config['J'] = int(args.MMR) # Number of DNA REPAIR mutational signatures. Given as input.\n\n # Are regionsweighted?\n config['noweights'] = args.noweights\n\n # Add ADAM specific properties\n config['optimizer_iterations'] = [5000,5000,20000,20000,5000,5000,5000,5000,10000,10000,10000]\n config['optimizer_stepsize'] = [ 50, 10, 1, 0.1, 50, 25, 10, 10, 0.1, 0.01,0.005]\n config['optimizer_user_update_steps'] = 500 #Number of updates to print during optimization\n\n return config\n\n# plots the loaded model into tensorboard, has no effect if no tensorboard\n# folder was specified\ndef graphModel( ):\n if config['tensorboard'] != None:\n\n #because we are executing egerly, no graph is being\n #created. Hence, we need to manually initiate a trace\n #before we can visualize the computation\n tf.summary.trace_on(graph=True)\n foo_g = tf.function(compute_loss)\n foo_g()\n\n #now we can write the graph to tensorboard\n writer = config['writer']\n with writer.as_default():\n tf.summary.trace_export(\n name='tf2_graph',\n step=0\n )\n\n# Writes all trainable tensors to file.\n# If config['ouput'] == None, this function has no effect.\n# One line per tensor, tab separated. [ID, Shape, Serialized Data]\n# Serialized Data and Shape are comma separated internally, eg\n# Tensor1 2,2 1,2,3,4\n# Tensor2 2,3 1,2,3,4,5,6\ndef exportData(variables):\n if config['output'] != None:\n with open(config['output'], 'w') as file:\n for name in variables:\n tensor_py = eval(name).numpy()\n\n #write ID:\n file.write(name)\n file.write(\"\\t\")\n\n #write shape\n file.write(','.join(map(lambda x: str(x),tensor_py.shape)) )\n file.write(\"\\t\")\n\n #write serialized tensor\n file.write( ','.join(map(lambda x: str(x), tensor_py.reshape([-1]))) )\n file.write(\"\\n\")\n\n\n# See https://stackoverflow.com/questions/59309114/tensor-n-mode-product-in-tensorflow\ndef n_mode_product(x, u, n, name=None):\n n = int(n)\n # We need one letter per dimension\n # (maybe you could find a workaround for this limitation)\n if n > 26:\n raise ValueError('n is too large.')\n ind = ''.join(chr(ord('a') + i) for i in range(n))\n exp = f\"{ind}K...,JK->{ind}J...\"\n return tf.einsum(exp, x, u, name=name)\n\n\n\ndef readInputData(verbose=False):\n #get data from file\n data = {}\n with open(config['input']) as file:\n for line in file:\n line = line.split()\n mat = np.array(list(map(lambda x: float(x), line[2].split(',')))) #matrix or tensor\n shape = list(map(lambda x: int(x), line[1].split(','))) #matrix/tensor shape\n mat = mat.reshape(shape)\n name = line[0]\n #save the tensor in data dict\n data[name] = mat\n\n if verbose:\n print(f\"Read {len(data)} matrices: {','.join(data.keys())}\")\n\n return data\n\n\n# Function generating input tensor P.\ndef getTensor(name, verbose=False):\n\n T = tf.Variable(\n initial_value=tf.convert_to_tensor(value=data[name]),\n trainable=False,\n name=name+\"_input\"\n )\n\n if verbose:\n print(f\"Got {name} of shape {T.shape}, min {tf.reduce_min(T)}, max {tf.reduce_max(T)}\")\n\n return T\n\n\ndef getRandomTensor(shape, min, max, trainable=True, clip_min=np.NINF, clip_max=np.infty, sum_one=False, sum_one_axis=None, name=None, verbose=False):\n \"\"\"Function generating a random tensor of shape containing values.\n\n Parameters\n ----------\n shape : array or tuple\n The shape of the tensor to generate\n min : float\n smallest element in tensor (initial value)\n max : float\n largest element in tensor (initial value)\n trainable : boolean\n Whether this tensor should be optimized\n clip_min: float\n If optimized, any values below clip_min will be clipped\n clip_max: float\n If optimized, any values above clip_max will be clipped\n sum_one : boolean\n Whether to constain this tensor to have all elements sum to one\n sum_one_axis : array or integer\n If sum_one is true, it reduces this tensor along the dimensions given in axis\n name : string\n Description of the tensor\n verbose: bool\n If true, print message regarding shape of the tensor\n\n Returns\n -------\n tf.Tensor\n Tensor with the properties as described by the parameters\n\n \"\"\"\n\n tensor = np.random.uniform(size=shape) * (max - min) + min\n\n #constraints\n if sum_one: #probabilities along sum_one_axis\n constraint = lambda x: tf.clip_by_value(x,clip_min,clip_max) / tf.reduce_sum(tf.clip_by_value(x,clip_min,clip_max), axis=sum_one_axis, keepdims=True)\n else:\n constraint = lambda x: tf.clip_by_value(x, clip_min,clip_max)\n\n tensor = tf.Variable(\n initial_value=tf.convert_to_tensor(value=tensor),\n trainable=trainable,\n constraint=constraint,\n name=name)\n\n if verbose:\n print(f\"Generated initial {name} of shape {tensor.shape}, min {tf.reduce_min(tensor)}, max {tf.reduce_max(tensor)}\")\n\n return tensor\n\n\ndef optimizeModel(loss, iters, stepsize):\n\n #print directives\n update_steps = tf.constant(config['optimizer_user_update_steps'])\n epoch = tf.Variable(0)\n num_it = tf.constant(iters)\n cont = tf.Variable(True)\n #current best loss\n best_loss = tf.Variable(tf.cast(loss(), tf.float32))\n #best parameters so far\n W_best = tf.Variable(tf.identity(W), shape=W.shape)\n A_best = tf.Variable(tf.identity(A), shape=A.shape)\n Q_best = tf.Variable(tf.identity(Q), shape=Q.shape)\n R_best = tf.Variable(tf.identity(R), shape=R.shape)\n D_best = tf.Variable(tf.identity(D), shape=D.shape)\n\n # Define a training operation for tensforflow, this can be exchanged with other optimizers if desired\n train_op = tf.optimizers.Adam(stepsize)\n #train_op = tf.optimizers.Nadam(stepsize)\n #train_op = tf.optimizers.Adadelta(stepsize)\n\n @tf.function\n def doOpt():\n while epoch < num_it and cont:\n\n # evaluate graph. this will compute gradients, and update trainable\n # variables under the hood\n train = train_op.minimize(loss, trainable_variables)\n loss_val = tf.cast(loss(), tf.float32)\n\n # store the best weights so far, these are the ones we will return\n tf.cond(\n tf.less(loss_val,best_loss),\n lambda: [W_best.assign(W), A_best.assign(A), Q_best.assign(Q), R_best.assign(R), D_best.assign(D), best_loss.assign(loss_val)],\n lambda: [W_best, A_best, Q_best, R_best, D_best, loss_val] ##dummy for the graph\n )\n\n epoch.assign(epoch + 1)\n\n # print update?\n tf.cond(tf.equal(epoch % update_steps, 0),\n lambda: [ #True\n tf.print(\"Epoch\", epoch, \"loss\", loss_val, output_stream=sys.stdout) #print update message\n ],\n lambda: [ #False\n tf.no_op() #just a placeholder\n ])\n\n\n # Perform optimization\n tf.print(\"Epoch\", 0, \"loss\", tf.cast(loss(), tf.float32), output_stream=sys.stdout)\n doOpt()\n\n # Save best results at the end of the current iteration\n W.assign(W_best)\n A.assign(A_best)\n Q.assign(Q_best)\n R.assign(R_best)\n D.assign(D_best)\n\n\n# Define a loss. We use the Frobenuis loss.\n# This function will be wrapped into a GradientTape during eager\n# execution\ndef compute_loss():\n # Mhat = term1 * term2 + term3\n\n term1 = n_mode_product(tf.multiply(P,W), A, 0)\n term2 = 1 + n_mode_product(tf.multiply(Q,R), D, 0)\n\n Mhat = term1 * term2\n\n loss_value = tf.norm(\n tensor=M-Mhat,\n ord='euclidean',\n name='frobenius_norm'\n )\n\n return loss_value\n\n###############################################################################\n#MAIN\nverbose=True\nminval = 1e-10\nmaxval = 1000\n\nconfig = getConfig()\n\n\n# Create logging folder if required\nif config['tensorboard'] != None:\n os.mkdir(config['tensorboard'])\n config['writer'] = tf.summary.create_file_writer(config['tensorboard'])\n print(\"Enabled tensorboard logging to folder %s\" % config['tensorboard'])\nelse:\n config['writer'] = None\n print(\"No tensorboard directory specified, disabling logging.\")\n\n#read input matrices\ndata = readInputData(verbose=verbose)\n\n# Initialise tensors\n# Tensor ranks are:\n# M -> G x K x L (input) - mutation counts of category k in region l of genome g\n# P -> N x K x 1 (input) - PRIMARY mutational signature matrix (N signatures, K mutation categories)\n# P -> N x K x 1 (input) - PRIMARY mutational signature matrix (N signatures, K mutation categories)\n# W -> N x 1 x L (trainable or constant) - regional activity matrix for PRIMARY signatures (regional weights)\n# A -> G x N (trainable) - genome activity matrix for PRIMARY signatures\n# Q -> J x K x 1 (trainable) - MMR mutational signature matrix\n# R -> J x 1 x L (trainable or constant) - regional activity matrix for MMR signatures (regional weights)\n# D -> G x J (trainable) - genome activity matrix for MMR signatures\n\n# Adding background signature\n# PB -> 1 x K x 1 (input)\n# P <- tf.concat(P,PB)\n# N <- N + 1\n\n# input mutation count matrix\nM = getTensor('M', verbose=verbose)\n\n# PRIMARY mutational signature matrix + BACKGROUND signature\nPprimary = getTensor('P', verbose=verbose)\nPbackground = getTensor('PB', verbose=verbose)\nP = tf.concat([Pprimary,Pbackground],0)\nP = tf.reshape(P, P.shape+[1])\n\n# set config dimentions\nconfig['G'], config['K'], config['L'] = M.shape\nconfig['N'] = P.shape[0] #number of primary signatures + 1 BG signature\n\n# regional activity matrix for PRIMARY signatures (regional weights)\nW = getRandomTensor(shape=[config['N'],1,config['L']],\n min=minval if not config['noweights'] else 1.0,\n max=1.0,\n trainable=not config['noweights'],\n clip_min=minval,\n sum_one=True,\n sum_one_axis=2, #reduce sum over L (3rd dim)\n name=\"W\",\n verbose=verbose)\n# genome activity matrix for PRIMARY signatures\nA = getRandomTensor(shape=[config['G'],config['N']],\n min=0,\n max=maxval,\n trainable=True,\n clip_min=0,\n name=\"A\",\n verbose=verbose)\n# MMR mutational signature matrix\nQ = getRandomTensor(shape=[config['J'],config['K'],1],\n min=minval,\n max=1.0,\n trainable=True,\n clip_min=minval,\n sum_one=True,\n sum_one_axis=1, #reduce sum over K (2nd dim)\n name=\"Q\",\n verbose=verbose)\n# regional activity matrix for MMR signatures (regional weights)\nR = getRandomTensor(shape=[config['J'],1,config['L']],\n min=minval if not config['noweights'] else 1.0,\n max=1.0,\n trainable=not config['noweights'],\n clip_min=minval,\n sum_one=True,\n sum_one_axis=2, #reduce sum over L (3rd dim)\n name=\"R\",\n verbose=verbose)\n# genome activity matrix for MMR signatures\nD = getRandomTensor(shape=[config['G'],config['J']],\n min=0,\n max=maxval,\n trainable=True,\n clip_min=0,\n name=\"D\",\n verbose=verbose)\n\n\n#Define which tensors should be optimized\ntrainable_variables = [W,A,Q,R,D] if not config['noweights'] else [A,Q,D]\n\nprint(f\"> Config: {config}\")\n\nprint(\"> Model definition complete, commencing optimization\")\nt0= timer()\n# Perform optimization\nfor iters,stepsize in zip(config['optimizer_iterations'],config['optimizer_stepsize']):\n tf.print(\"### New optimization: iterations\", iters, \"step size\", stepsize, output_stream=sys.stdout)\n optimizeModel(compute_loss, iters, stepsize)\n tf.print(\"Best loss\", tf.cast(compute_loss(), tf.float32), output_stream=sys.stdout)\nt1 = timer()\n\nprint(f\"> Optimization complete. Time elapsed: {t1 - t0} seconds\\n\") # CPU seconds elapsed (floating point)\n\n#post processing\ngraphModel( )\nexportData(['W','A','Q','R','D'])\n", "sub_path": "repairsig.py", "file_name": "repairsig.py", "file_ext": "py", "file_size_in_byte": 15162, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 15, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 36, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 37, "usage_type": "call"}, {"api_name": "os.path", "line_number": 37, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 54, "usage_type": "call"}, {"api_name": "os.path", "line_number": 54, "usage_type": "attribute"}, {"api_name": "tensorflow.summary.trace_on", "line_number": 81, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 81, "usage_type": "attribute"}, {"api_name": "tensorflow.function", "line_number": 82, "usage_type": "call"}, {"api_name": "tensorflow.summary.trace_export", "line_number": 88, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 88, "usage_type": "attribute"}, {"api_name": "tensorflow.einsum", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 137, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 153, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 154, "usage_type": "call"}, {"api_name": "tensorflow.reduce_min", "line_number": 160, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.NINF", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.infty", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 198, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 198, "usage_type": "attribute"}, {"api_name": "tensorflow.clip_by_value", "line_number": 202, "usage_type": "call"}, {"api_name": "tensorflow.reduce_sum", "line_number": 202, "usage_type": "call"}, {"api_name": "tensorflow.clip_by_value", "line_number": 204, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 206, "usage_type": "call"}, {"api_name": "tensorflow.convert_to_tensor", "line_number": 207, "usage_type": "call"}, {"api_name": "tensorflow.reduce_min", "line_number": 213, "usage_type": "call"}, {"api_name": "tensorflow.reduce_max", "line_number": 213, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 221, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 222, "usage_type": "call"}, {"api_name": "tensorflow.constant", "line_number": 223, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 224, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 226, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 226, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 226, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.identity", "line_number": 228, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.identity", "line_number": 229, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.identity", "line_number": 230, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.identity", "line_number": 231, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 232, "usage_type": "call"}, {"api_name": "tensorflow.identity", "line_number": 232, "usage_type": "call"}, {"api_name": "tensorflow.optimizers.Adam", "line_number": 235, "usage_type": "call"}, {"api_name": "tensorflow.optimizers", "line_number": 235, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 246, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 246, "usage_type": "attribute"}, {"api_name": "tensorflow.cond", "line_number": 249, "usage_type": "call"}, {"api_name": "tensorflow.less", "line_number": 250, "usage_type": "call"}, {"api_name": "tensorflow.cond", "line_number": 258, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 258, "usage_type": "call"}, {"api_name": "tensorflow.print", "line_number": 260, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 260, "usage_type": "attribute"}, {"api_name": "tensorflow.no_op", "line_number": 263, "usage_type": "call"}, {"api_name": "tensorflow.function", "line_number": 239, "usage_type": "attribute"}, {"api_name": "tensorflow.print", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 268, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 268, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 268, "usage_type": "attribute"}, {"api_name": "tensorflow.multiply", "line_number": 285, "usage_type": "call"}, {"api_name": "tensorflow.multiply", "line_number": 286, "usage_type": "call"}, {"api_name": "tensorflow.norm", "line_number": 290, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 309, "usage_type": "call"}, {"api_name": "tensorflow.summary.create_file_writer", "line_number": 310, "usage_type": "call"}, {"api_name": "tensorflow.summary", "line_number": 310, "usage_type": "attribute"}, {"api_name": "tensorflow.concat", "line_number": 341, "usage_type": "call"}, {"api_name": "tensorflow.reshape", "line_number": 342, "usage_type": "call"}, {"api_name": "timeit.default_timer", "line_number": 402, "usage_type": "call"}, {"api_name": "tensorflow.print", "line_number": 405, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 405, "usage_type": "attribute"}, {"api_name": "tensorflow.print", "line_number": 407, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 407, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 407, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 407, "usage_type": "attribute"}, {"api_name": "timeit.default_timer", "line_number": 408, "usage_type": "call"}]} +{"seq_id": "174777550", "text": "import numpy as np\nfrom scipy.spatial.distance import euclidean\nfrom scipy.special import expit\nfrom sklearn.base import BaseEstimator\nfrom numpy import logaddexp\nimport time\n\nclass LogReg(BaseEstimator):\n def __init__(self, lambda_1=0.0, lambda_2=1.0, gd_type='full',\n tolerance=1e-6, max_iter=1000, w0=None, alpha=0.001, history=False):\n \"\"\"\n lambda_1: L1 regularization param\n lambda_2: L2 regularization param\n gd_type: 'full' or 'stochastic'\n tolerance: for stopping gradient descent\n max_iter: maximum number of steps in gradient descent\n w0: np.array of shape (d) - init weights\n alpha: learning rate\n \"\"\"\n self.lambda_1 = lambda_1\n self.lambda_2 = lambda_2\n self.gd_type = gd_type\n self.tolerance = tolerance\n self.max_iter = max_iter\n self.w0 = w0\n self.alpha = alpha\n self.w = None\n self.loss_history = []\n self.iteration_time = []\n self.history = history\n\n def fit(self, X, y):\n # нормируем матрицу объектов признаков\n X = np.array(X)\n y = np.array(y)\n\n # инициализируем массив весов\n if self.w0 is not None:\n self.w = self.w0\n else:\n self.w = np.zeros(X[0].size)\n\n for i in range(self.max_iter):\n from_ = time.time()\n w_new = self.w - self.alpha * self.calc_gradient(X, y)\n \n if self.history:\n time_wasted = time.time() - from_\n self.loss_history.append(self.calc_loss(X, y))\n self.iteration_time.append(time_wasted)\n \n if euclidean(w_new, self.w) < self.tolerance:\n self.w = w_new\n break\n \n self.w = w_new\n\n\n return self\n\n def predict_proba(self, X):\n \"\"\"\n X: np.array of shape (l, d)\n ---\n output: np.array of shape (l, 2) where\n first column has probabilities of -1\n second column has probabilities of +1\n \"\"\"\n if self.w is None:\n raise Exception('Not trained yet')\n\n X = np.array(X)\n\n predictions = np.empty((2, len(X)))\n t = expit(np.dot(X, self.w))\n predictions[0] = np.ones(len(X)) - t\n predictions[1] = t\n predictions = predictions.T\n\n return predictions\n\n def predict(self, X):\n \"\"\"\n X: np.array of shape (l, d)\n ---\n output: np.array of shape (l, 2) where\n first column has probabilities of -1\n second column has probabilities of +1\n \"\"\"\n\n X = np.array(X)\n if self.w is None:\n raise Exception('Not trained yet')\n\n predictions = np.zeros(len(X), dtype=int)\n for i in range(len(X)):\n t = expit(np.dot(self.w, X[i]))\n\n if t > 0.5:\n predictions[i] = 1\n else:\n predictions[i] = 0\n\n return predictions\n\n def calc_gradient(self, X, y):\n \"\"\"\n X: np.array of shape (l, d) (l can be equal to 1 if stochastic)\n y: np.array of shape (l)\n ---\n output: np.array of shape (d)\n \"\"\"\n grad = np.zeros(len(X[0]))\n \n if self.gd_type == 'full':\n grad = np.dot(y * expit(-y * np.dot(X, self.w)), -X)\n grad /= len(X)\n else:\n rand_idx = np.random.randint(0, len(X))\n tmp = y[rand_idx] * X[rand_idx]\n tmp /= -(1 + np.exp(y[rand_idx] * np.dot(self.w, X[rand_idx])))\n grad += tmp\n \n \n grad += self.lambda_2 * self.w\n\n return grad\n\n def calc_loss(self, X, y):\n \"\"\"\n X: np.array of shape (l, d)\n y: np.array of shape (l)\n ---\n output: float\n \"\"\"\n total_loss = (np.log(1 / expit(y * (np.dot(X, self.w))))).sum() / len(X)\n\n total_loss += (self.lambda_2 / 2) * np.linalg.norm(self.w)**2\n\n return total_loss\n\n", "sub_path": "homework-MO/2/logreg.py", "file_name": "logreg.py", "file_ext": "py", "file_size_in_byte": 4084, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sklearn.base.BaseEstimator", "line_number": 8, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 44, "usage_type": "call"}, {"api_name": "time.time", "line_number": 48, "usage_type": "call"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 72, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 74, "usage_type": "call"}, {"api_name": "scipy.special.expit", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 76, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 95, "usage_type": "call"}, {"api_name": "scipy.special.expit", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 116, "usage_type": "call"}, {"api_name": "scipy.special.expit", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 119, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 119, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 136, "usage_type": "call"}, {"api_name": "scipy.special.expit", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 138, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 138, "usage_type": "attribute"}]} +{"seq_id": "61382298", "text": "'''\nExercise 3a\n'''\n\nimport random\nimport numpy as np\nfrom matplotlib import pyplot as plt\nimport simulator_class2 as sim\n\nrandom.seed(42)\nnp.random.seed(42)\n\nif __name__ == '__main__':\n \n Tq = 1500 # max average queueing time\n SERVICE_TIMES = [*range(1, 1000, 5)]\n \n num_sim = 1\n tot_queueing_delay = np.zeros(len(SERVICE_TIMES))\n for seed in range(num_sim):\n random.seed(seed)\n np.random.seed(seed)\n queueing_delay = []\n for SERVICE in SERVICE_TIMES:\n \n # MDC\n ARRIVAL = 500\n LOAD = SERVICE/ARRIVAL\n BUFFER_SIZE = 10\n FOG_NODES = 1\n \n # CDC\n f = 0.8\n SERVICE_CLOUD = 100\n CLOUD_BUFFER_SIZE = 20\n CLOUD_SERVERS = 1\n \n # SIMULATION PARAMS\n SIM_TIME = 300000\n \n # data storage object\n data = sim.Measure(0,0,0,0,0,0,0,0,0,0,[],[],[],\n [],[],[],[],[],[],[],[],[])\n data_cloud = sim.Measure(0,0,0,0,0,0,0,0,0,0,[],[],[],\n [],[],[],[],[],[],[],[],[])\n \n # simulator\n s = sim.Simulator(data, data_cloud, LOAD, SERVICE, ARRIVAL, BUFFER_SIZE, \n FOG_NODES, SIM_TIME, f, CLOUD_SERVERS, CLOUD_BUFFER_SIZE, \n SERVICE_CLOUD)\n print_everything = False\n data, data_cloud, time, _, _ = s.simulate(print_everything)\n \n # cumulate statistics\n queueing_delay.append((data.delay + data_cloud.delay) / (data.dep + data_cloud.dep))\n tot_queueing_delay += np.array(queueing_delay)\n # average multiple simulations \n tot_queueing_delay /= num_sim\n \n # Average queueing delay with progressively faster MDC service\n plt.plot(SERVICE_TIMES, tot_queueing_delay, label='Queueing delay')\n plt.plot(Tq * np.ones(1000) , '--', label='Tq')\n plt.grid()\n plt.legend()\n plt.xlabel(\"Average service MDC time [ms]\")\n plt.ylabel(\"Average queueing delay [ms]\")\n #plt.xlim([0,200])\n plt.title('Average queueing delay with progressively slower MDC service')\n plt.show()\n\n", "sub_path": "Lab2/ex3a.py", "file_name": "ex3a.py", "file_ext": "py", "file_size_in_byte": 2237, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "random.seed", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 19, "usage_type": "call"}, {"api_name": "random.seed", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 22, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 22, "usage_type": "attribute"}, {"api_name": "simulator_class2.Measure", "line_number": 42, "usage_type": "call"}, {"api_name": "simulator_class2.Measure", "line_number": 44, "usage_type": "call"}, {"api_name": "simulator_class2.Simulator", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.grid", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 65, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 65, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 66, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 66, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 68, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 68, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 69, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 69, "usage_type": "name"}]} +{"seq_id": "109214809", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('main', '0035_auto_20151018_1528'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='ViewHistory',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('create_date', models.DateTimeField(auto_now=True)),\n ('competition', models.ForeignKey(to='main.Competition')),\n ('from_user', models.ForeignKey(related_name='viewed_photos', to=settings.AUTH_USER_MODEL)),\n ('photo', models.ForeignKey(related_name='view_history', to='main.Photo')),\n ],\n ),\n ]\n", "sub_path": "main/migrations/0036_viewhistory.py", "file_name": "0036_viewhistory.py", "file_ext": "py", "file_size_in_byte": 851, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.CreateModel", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.AutoField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.models.DateTimeField", "line_number": 19, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 19, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 21, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 21, "usage_type": "name"}, {"api_name": "django.conf.settings.AUTH_USER_MODEL", "line_number": 21, "usage_type": "attribute"}, {"api_name": "django.conf.settings", "line_number": 21, "usage_type": "name"}, {"api_name": "django.db.models.ForeignKey", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 22, "usage_type": "name"}]} +{"seq_id": "282786830", "text": "import numpy as np\nfrom pathlib import Path\nfrom torch.utils.data import DataLoader\n\nfrom data.dataset import Pix2PixDataset\nfrom utils import config\n\ndef get_dataloader(type, batch_size=2):\n\n DIR = Path(config.PATH_TO_DATA)\n files = np.random.permutation(list(DIR.rglob('*.jpg')))\n\n if type == 'train':\n files = files[:-9]\n elif type == 'val':\n files = files[-9:-3]\n batch_size = 6\n elif type == 'test':\n files = files[-3:]\n batch_size = 3\n\n data = Pix2PixDataset(files)\n data_loader = DataLoader(data, batch_size=batch_size, drop_last=True, shuffle=True)\n return data_loader\n\n", "sub_path": "data/dataloaders.py", "file_name": "dataloaders.py", "file_ext": "py", "file_size_in_byte": 637, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pathlib.Path", "line_number": 10, "usage_type": "call"}, {"api_name": "utils.config.PATH_TO_DATA", "line_number": 10, "usage_type": "attribute"}, {"api_name": "utils.config", "line_number": 10, "usage_type": "name"}, {"api_name": "numpy.random.permutation", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "data.dataset", "line_number": 22, "usage_type": "name"}, {"api_name": "data.dataset.Pix2PixDataset", "line_number": 22, "usage_type": "call"}, {"api_name": "torch.utils.data.DataLoader", "line_number": 23, "usage_type": "call"}, {"api_name": "data.dataset", "line_number": 23, "usage_type": "argument"}]} +{"seq_id": "154613166", "text": "#\n# Copyright 2014 Red Hat, Inc.\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA\n#\n# Refer to the README and COPYING files for full details of the license\n#\n\nimport nose.tools as nt\nimport os\nimport tempfile\n\nfrom ovirtsdk.xml import params\n\nfrom ovirtlago import testlib\n\n\n# AAA\nAAA_LDAP_USER = 'user1'\nAAA_LDAP_AUTHZ_PROVIDER = 'lago.local-authz'\nHOSTNAME_389DS = testlib.get_prefixed_name('engine')\n\n\n@testlib.with_ovirt_prefix\ndef add_ldap_provider(prefix):\n engine = prefix.virt_env.engine_vm()\n machine_389ds = prefix.virt_env.get_vm(HOSTNAME_389DS)\n\n answer_file_src = os.path.join(\n os.environ.get('SUITE'),\n 'aaa-ldap-answer-file.conf'\n )\n\n with open(answer_file_src, 'r') as f:\n content = f.read()\n content = content.replace('@389DS_IP@', machine_389ds.ip())\n\n with tempfile.NamedTemporaryFile(delete=False) as temp:\n temp.write(content)\n engine.copy_to(temp.name, '/root/aaa-ldap-answer-file.conf')\n os.unlink(temp.name)\n\n result = machine_389ds.ssh(\n [\n 'systemctl',\n 'start',\n 'dirsrv@lago',\n ],\n )\n nt.eq_(\n result.code, 0, 'Failed to start LDAP server. Exit code %s' % result.code\n )\n\n result = engine.ssh(\n [\n 'ovirt-engine-extension-aaa-ldap-setup',\n '--config-append=/root/aaa-ldap-answer-file.conf',\n '--log=/var/log/ovirt-engine-extension-aaa-ldap-setup.log',\n ],\n )\n nt.eq_(\n result.code, 0, 'aaa-ldap-setup failed. Exit code is %s' % result.code\n )\n\n engine.service('ovirt-engine')._request_stop()\n testlib.assert_true_within_long(\n lambda: not engine.service('ovirt-engine').alive()\n )\n engine.service('ovirt-engine')._request_start()\n testlib.assert_true_within_long(\n lambda: engine.service('ovirt-engine').alive()\n )\n\n\n@testlib.with_ovirt_api\ndef add_ldap_user(api):\n p = params.User(\n user_name=AAA_LDAP_USER,\n domain=params.Domain(\n name=AAA_LDAP_AUTHZ_PROVIDER\n ),\n )\n nt.assert_true(api.users.add(p))\n\n\n_TEST_LIST = [\n #add_ldap_provider,\n #add_ldap_user,\n]\n\n\ndef test_gen():\n for t in testlib.test_sequence_gen(_TEST_LIST):\n test_gen.__name__ = t.description\n yield t\n", "sub_path": "basic-suite-master/test-scenarios/099_aaa-ldap.py", "file_name": "099_aaa-ldap.py", "file_ext": "py", "file_size_in_byte": 2944, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "ovirtlago.testlib.get_prefixed_name", "line_number": 33, "usage_type": "call"}, {"api_name": "ovirtlago.testlib", "line_number": 33, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 41, "usage_type": "call"}, {"api_name": "os.path", "line_number": 41, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 42, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tempfile.NamedTemporaryFile", "line_number": 50, "usage_type": "call"}, {"api_name": "os.unlink", "line_number": 53, "usage_type": "call"}, {"api_name": "nose.tools.eq_", "line_number": 62, "usage_type": "call"}, {"api_name": "nose.tools", "line_number": 62, "usage_type": "name"}, {"api_name": "nose.tools.eq_", "line_number": 73, "usage_type": "call"}, {"api_name": "nose.tools", "line_number": 73, "usage_type": "name"}, {"api_name": "ovirtlago.testlib.assert_true_within_long", "line_number": 78, "usage_type": "call"}, {"api_name": "ovirtlago.testlib", "line_number": 78, "usage_type": "name"}, {"api_name": "ovirtlago.testlib.assert_true_within_long", "line_number": 82, "usage_type": "call"}, {"api_name": "ovirtlago.testlib", "line_number": 82, "usage_type": "name"}, {"api_name": "ovirtlago.testlib.with_ovirt_prefix", "line_number": 36, "usage_type": "attribute"}, {"api_name": "ovirtlago.testlib", "line_number": 36, "usage_type": "name"}, {"api_name": "ovirtsdk.xml.params.User", "line_number": 89, "usage_type": "call"}, {"api_name": "ovirtsdk.xml.params", "line_number": 89, "usage_type": "name"}, {"api_name": "ovirtsdk.xml.params.Domain", "line_number": 91, "usage_type": "call"}, {"api_name": "ovirtsdk.xml.params", "line_number": 91, "usage_type": "name"}, {"api_name": "nose.tools.assert_true", "line_number": 95, "usage_type": "call"}, {"api_name": "nose.tools", "line_number": 95, "usage_type": "name"}, {"api_name": "ovirtlago.testlib.with_ovirt_api", "line_number": 87, "usage_type": "attribute"}, {"api_name": "ovirtlago.testlib", "line_number": 87, "usage_type": "name"}, {"api_name": "ovirtlago.testlib.test_sequence_gen", "line_number": 105, "usage_type": "call"}, {"api_name": "ovirtlago.testlib", "line_number": 105, "usage_type": "name"}]} +{"seq_id": "280857456", "text": "import sys\nimport requests\nimport time\nfrom mapper import Map\nfrom miner import *\nfrom cpu import *\n\nimport random\n\n\nclass Scripter:\n #Main Script Class\n\n def __init__(self, key, command='go to the nearest shrine'):\n #this will insert the api key into requests\n self.apiKey = key\n\n self.url = \"https://lambda-treasure-hunt.herokuapp.com/api/adv/\" #add additional url information at the end. \n self.headers = {'Content-Type': 'application/json',\n 'Authorization': 'Token ' + self.apiKey} \n\n #map json will be used to pass map graph to other users. \n self.map = Map()\n\n\n\n #whch command/script is currently running\n self.command = command\n\n #status of player: \n self.player_name = \"\",\n self.player_cooldown = 1,\n self.player_encumbrance = 0, # How much are you carrying?\n self.player_strength = 0, # How much can you carry?\n self.player_speed = 0, # How fast do you travel?\n self.player_gold = 0,\n self.player_inventory = [],\n self.player_status = [],\n self.player_errors = [],\n self.player_messages = []\n self.player_location = ''\n self.player_mine = ''\n\n def getInit(self):\n response = requests.get(self.url + 'init', headers=self.headers) \n # extracting data in json format \n data = response.json() \n\n #set the player location & cooldown\n self.player_location = data['room_id']\n self.player_cooldown = data['cooldown']\n\n print(data)\n #this add to map function sends our current information to our Map Class and makes sure that it is mappd. \n return data\n\n def getStatus(self):\n response = requests.post(self.url + 'status', headers=self.headers) \n # extracting data in json format \n data = response.json() \n\n #set the player information\n self.player_name = data['name']\n self.player_encumbrance = data['encumbrance']\n self.player_strength = data['strength']\n self.player_gold = data['gold']\n self.player_inventory = data['inventory']\n self.player_status = data['status']\n self.player_errors = data['errors']\n self.player_messages = data['messages']\n\n print(data)\n #this add to map function sends our current information to our Map Class and makes sure that it is mappd. \n return data\n\n def findPath(self, destination):\n time.sleep(self.player_cooldown)\n path = self.map.findPath(self.player_location, destination)\n \n print(path)\n path.pop(0)\n for roomId in path:\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n\n print(f'room {destination} reached')\n\n def getCoin(self, amount):\n \n coinsMined = 0\n new_proof = ''\n\n while coinsMined < amount:\n # if self.player_mine == '':\n newLocation = self.wishingWell()\n # else:\n #newLocation = 242\n path = self.map.findPath(self.player_location, int(newLocation))\n\n print(path)\n\n path.pop(0)\n for roomId in path:\n\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n #print(currentRoom)\n\n #get last proof\n while new_proof == '':\n response = requests.get(\"https://lambda-treasure-hunt.herokuapp.com/api/bc/last_proof\", headers=self.headers)\n data = response.json() \n print(data.get('difficulty'))\n new_proof = proof_of_work(data.get('proof'), data.get('difficulty'))\n\n if new_proof != '':\n post_data = {\"proof\": new_proof}\n\n r = requests.post(url=\"https://lambda-treasure-hunt.herokuapp.com/api/bc/mine/\", headers=self.headers, json=post_data)\n\n data = r.json()\n\n for errors in data['errors']:\n if errors == 'Proof already submitted: +10s CD':\n new_proof = ''\n \n self.player_cooldown = data['cooldown']\n time.sleep(self.player_cooldown)\n \n print(data)\n self.player_cooldown = data['cooldown']\n coinsMined += 1\n self.player_mine = ''\n new_proof = ''\n time.sleep(self.player_cooldown)\n\n def getGold(self, amount):\n print(self.player_gold)\n while self.player_gold < amount:\n if self.player_encumbrance < self.player_strength - 2:\n #go to a random room path\n path = self.map.findPath(self.player_location, random.randint(2,499))\n #check for an item. \n path.pop(0)\n print(path)\n for roomId in path:\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n print(currentRoom)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n for item in currentRoom['items']:\n time.sleep(self.player_cooldown)\n json = {\"name\":item}\n response = requests.post(self.url + 'take', headers=self.headers, json=json)\n data = response.json() \n self.player_cooldown = data['cooldown']\n time.sleep(self.player_cooldown)\n self.getStatus()\n if self.player_encumbrance > self.player_strength - 2:\n break\n else: \n #go to shop\n path = self.map.findPath(self.player_location, 1)\n path.pop(0)\n print(path)\n for roomId in path:\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n #sell all\n for item in self.player_inventory:\n time.sleep(self.player_cooldown)\n json = {\"name\":item}\n response = requests.post(self.url + 'sell', headers=self.headers, json=json)\n data = response.json() \n print(data)\n self.player_cooldown = data['cooldown']\n\n json[\"confirm\"] = \"yes\"\n time.sleep(self.player_cooldown)\n response = requests.post(self.url + 'sell', headers=self.headers, json=json)\n data = response.json() \n self.player_cooldown = data['cooldown']\n time.sleep(self.player_cooldown)\n self.getStatus()\n print(self.player_gold)\n\n def wishingWell(self):\n path = self.map.findPath(self.player_location, 55)\n path.pop(0)\n print(path)\n for roomId in path:\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n #print(currentRoom)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n json = {\"name\":\"wishing well\"}\n time.sleep(self.player_cooldown)\n response = requests.post(self.url + 'examine', headers=self.headers, json=json)\n data = response.json() \n\n #write the wishing well message to wishingwell.txt\n with open('wishingwell.txt', 'w') as outfile:\n outfile.write(data['description'][39:])\n outfile.close()\n #CPU translator\n cpu = CPU()\n cpu.load('wishingwell.txt')\n \n\n self.player_cooldown = data['cooldown']\n time.sleep(self.player_cooldown)\n self.player_mine = ''.join(cpu.run()[23:])\n return self.player_mine\n\n def changeName(self, newName):\n #467\n path = self.map.findPath(self.player_location, 467)\n #check for an item. \n path.pop(0)\n print(path)\n for roomId in path:\n direction = self.map.findDirection(self.player_location, roomId)\n time.sleep(self.player_cooldown)\n currentRoom = self.travel(direction)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n json = {\"name\":newName}\n time.sleep(self.player_cooldown)\n response = requests.post(self.url + 'change_name', headers=self.headers, json=json)\n data = response.json() \n print(data)\n self.player_cooldown = data['cooldown']\n time.sleep(self.player_cooldown)\n json[\"confirm\"] = \"aye\"\n response = requests.post(self.url + 'change_name', headers=self.headers, json=json)\n data = response.json() \n self.player_name = data['name']\n self.player_cooldown = data['cooldown']\n print('Your name has been changed')\n time.sleep(self.player_cooldown)\n self.getStatus()\n\n def mapper(self):\n #travel backwards to new exits\n reverse_path = [] \n directions = {'n': 's', 's': 'n', 'w': 'e', 'e': 'w'}\n\n currentRoom = self.getInit()\n\n previousRoom = ''\n direction = ''\n\n while self.map.length < 499: #make sure we dont have all 500 rooms mapped already\n \n\n if self.map.checkIfRoomMapped(currentRoom['room_id']) is not True:\n print('has not been mapped')\n #because current room is not mapped, add to map:\n self.map.addToMap(currentRoom)\n #check if we came from a previous direction\n\n if previousRoom:\n #update ids if we came from a previous room. \n self.map.updateRoomExits(previousRoom['room_id'], currentRoom['room_id'], direction)\n\n #if all exits have been explored\n while self.map.unexploredExits(currentRoom['room_id']):\n with open(\"reverse.txt\", \"r+\") as f:\n data = f.read()\n print(\"we will now move in reverse\")\n reverse_direction = data[-1]\n\n time.sleep(self.player_cooldown) #this waits the cooldown timer amount\n currentRoom = self.travel(str(reverse_direction))\n\n print(currentRoom)\n with open('reverse.txt', \"w\") as f:\n f.write(data[:-1])\n f.close()\n self.player_location = currentRoom['room_id']\n self.player_cooldown = currentRoom['cooldown']\n\n #go to first available exit in this current room. \n direction = self.map.getOneExit(currentRoom['room_id'])\n\n #write our direction to this file so that we can have a reverse path\n with open(\"reverse.txt\", \"a\") as f:\n f.write(f'{directions[direction]}')\n previousRoom = currentRoom\n time.sleep(self.player_cooldown) #this waits the cooldown timer amount\n currentRoom = self.travel(direction)\n self.player_cooldown = currentRoom['cooldown']\n self.player_location = currentRoom['room_id']\n\n def travel(self, direction, withDash=False):\n print(f'we are moving {direction}')\n\n if direction == 'n':\n json = {\"direction\":\"n\"}\n if self.map.knowId(self.player_location, direction):\n json[\"next_room_id\"] = self.map.knowId(self.player_location, direction)\n response = requests.post(self.url + 'fly', headers=self.headers, json=json)\n data = response.json() \n return data\n\n if direction == 'e':\n json = {\"direction\":\"e\"}\n if self.map.knowId(self.player_location, direction):\n json[\"next_room_id\"] = self.map.knowId(self.player_location, direction)\n response = requests.post(self.url + 'fly', headers=self.headers, json=json) \n data = response.json() \n return data\n\n if direction == 's':\n json = {\"direction\":\"s\"}\n if self.map.knowId(self.player_location, direction):\n json[\"next_room_id\"] = self.map.knowId(self.player_location, direction)\n response = requests.post(self.url + 'fly', headers=self.headers, json=json) \n data = response.json() \n return data\n\n if direction == 'w':\n json = {\"direction\":\"w\"}\n if self.map.knowId(self.player_location, direction):\n json[\"next_room_id\"] = self.map.knowId(self.player_location, direction)\n response = requests.post(self.url + 'fly', headers=self.headers, json=json) \n data = response.json() \n return data\n\n", "sub_path": "scripter.py", "file_name": "scripter.py", "file_ext": "py", "file_size_in_byte": 13662, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "mapper.Map", "line_number": 23, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 45, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 58, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 77, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 84, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 109, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 117, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 125, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 134, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 141, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 148, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 154, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 160, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 162, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 165, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 176, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 182, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 184, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 190, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 191, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 194, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 204, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 210, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 211, "usage_type": "call"}, {"api_name": "cpu.load", "line_number": 220, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 224, "usage_type": "call"}, {"api_name": "cpu.run", "line_number": 225, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 236, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 241, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 242, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 246, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 248, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 253, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 286, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 303, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 315, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 323, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 331, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 339, "usage_type": "call"}]} +{"seq_id": "203911663", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Dec 10 16:52:42 2019\n\n@author: nrchilku\n\"\"\"\nimport nengo\nimport os\nimport sys\nlmu_path = os.path.abspath(\"../lmu\")\nsys.path.append(lmu_path)\nfrom lmu import LMUCell\n\nfrom keras.utils import multi_gpu_model\nfrom keras.initializers import Constant\nfrom keras.models import Sequential, load_model\nfrom keras.layers import Dense, Flatten, Reshape, Activation, Conv2D, Conv2DTranspose, MaxPooling2D, RNN, LSTM, GRU, TimeDistributed\nfrom keras.callbacks import EarlyStopping\nfrom keras.initializers import Constant\nfrom keras.optimizers import SGD\nimport numpy as np\n\n# Data\nX = np.load('../data/X_5_40_16_4096.npy')\nY = np.load('../data/Y_5_40_16_4096.npy')\nY = np.reshape(Y, (4096, 16, 40, 40, 1))\nY = np.where(Y>1, 1, 0)\nX_test = np.load('../data/X_5_40_16_100.npy')\n\n# LMU layer\ndef lmu_layer(return_sequences=False,**kwargs):\n return RNN(LMUCell(units=800,\n order=1200,\n theta=16,\n input_encoders_initializer=Constant(1),\n hidden_encoders_initializer=Constant(0),\n memory_encoders_initializer=Constant(0),\n input_kernel_initializer=Constant(0),\n hidden_kernel_initializer=Constant(0),\n memory_kernel_initializer='glorot_normal',\n ),\n return_sequences=return_sequences,\n **kwargs)\n# Model\nmodel = Sequential()\nmodel.add(TimeDistributed(Conv2D(64, (3, 3)), input_shape=(16, 40, 40, 3)))\nmodel.add(TimeDistributed(Conv2D(128, (3, 3))))\nmodel.add(TimeDistributed(Conv2D(256, (3, 3))))\nmodel.add(TimeDistributed(Conv2D(256, (3, 3))))\nmodel.add(TimeDistributed(Conv2D(128, (3, 3))))\nmodel.add(TimeDistributed(Conv2D(1, (3, 3))))\nmodel.add(TimeDistributed(Flatten()))\nmodel.add(LSTM(800, return_sequences=True))\nmodel.add(lmu_layer(return_sequences=True))\nmodel.add(TimeDistributed(Dense(400)))\nmodel.add(TimeDistributed(Dense(400)))\nmodel.add(TimeDistributed(Reshape((20, 20, 1))))\nmodel.add(TimeDistributed(Conv2DTranspose(1, (2, 2), strides=2)))\nmodel.add(Activation('sigmoid'))\n\nmodel.summary()\n#model = multi_gpu_model(model, gpus=4)\n#model.compile(loss='binary_crossentropy', optimizer='adam', metrics=['accuracy'])\n#model.fit(X, Y, batch_size=32, epochs=100, validation_split=0.05, callbacks=[EarlyStopping(restore_best_weights=True, patience=40)])\n\n", "sub_path": "models_GPU/CNN_LMU.py", "file_name": "CNN_LMU.py", "file_ext": "py", "file_size_in_byte": 2443, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.abspath", "line_number": 11, "usage_type": "call"}, {"api_name": "os.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.load", "line_number": 25, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 26, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 29, "usage_type": "call"}, {"api_name": "keras.layers.RNN", "line_number": 33, "usage_type": "call"}, {"api_name": "lmu.LMUCell", "line_number": 33, "usage_type": "call"}, {"api_name": "keras.initializers.Constant", "line_number": 36, "usage_type": "call"}, {"api_name": "keras.initializers.Constant", "line_number": 37, "usage_type": "call"}, {"api_name": "keras.initializers.Constant", "line_number": 38, "usage_type": "call"}, {"api_name": "keras.initializers.Constant", "line_number": 39, "usage_type": "call"}, {"api_name": "keras.initializers.Constant", "line_number": 40, "usage_type": "call"}, {"api_name": "keras.models.Sequential", "line_number": 46, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 47, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 48, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 49, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 50, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 51, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.Conv2D", "line_number": 52, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.Flatten", "line_number": 53, "usage_type": "call"}, {"api_name": "keras.layers.LSTM", "line_number": 54, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 56, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.Dense", "line_number": 57, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.Reshape", "line_number": 58, "usage_type": "call"}, {"api_name": "keras.layers.TimeDistributed", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Conv2DTranspose", "line_number": 59, "usage_type": "call"}, {"api_name": "keras.layers.Activation", "line_number": 60, "usage_type": "call"}]} +{"seq_id": "305164758", "text": "\"\"\"setuptools based setup module\"\"\"\n\nfrom setuptools import setup\n\n# Convert the markdown readme to ReST using Pandoc\ntry:\n import pypandoc\n long_description = pypandoc.convert('README.md', 'rst')\nexcept ImportError:\n long_description = open('README.md').read()\n\n\nsetup(\n name='sayminimal',\n version=\"3.0.0\",\n description='A minimalist write-only Twitter/Mastodon client.',\n long_description=long_description,\n url='https://github.com/mduo13/sayminimal',\n author='mDuo13',\n author_email='mduo13@gmail.com',\n license='GPLv3',\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Environment :: X11 Applications :: GTK',\n 'Intended Audience :: End Users/Desktop',\n 'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',\n 'Operating System :: POSIX',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Topic :: Communications',\n ],\n keywords='twitter mastodon social microblogging',\n packages=[\n 'sayminimal',\n ],\n entry_points={\n 'console_scripts': [\n 'sayminimal = sayminimal.tweet:main',\n ],\n },\n install_requires=[\n 'PyYAML',\n 'tweepy',\n 'Mastodon.py',\n ],\n package_data={\n '': [\"fusion.glade\"],\n }\n)\n", "sub_path": "pypi_install_script/sayminimal-3.0.0.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1477, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pypandoc.convert", "line_number": 8, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "355737750", "text": "from django.db.models import F\nfrom django.http import HttpResponse\nfrom django.shortcuts import render, redirect\nimport jsonpickle\nfrom cart.models import CartItem\nfrom utils.alipay import AliPay\nimport uuid\nfrom order.models import *\nfrom datetime import datetime\n\n\ndef index(request):\n user = request.session.get('user', '')\n cartitems = request.GET.get('cartitems', '')\n if user:\n return redirect('/order/orderlist?cartitems='+cartitems)\n # return render(request, 'order.html')\n return render(request, 'login.html', {'cartitems': cartitems, 'redirect':'order'})\n\n\ndef orderlist(request):\n cartitems = request.GET.get('cartitems', '')\n\n result = jsonpickle.loads('[' + cartitems+ ']')\n\n try:\n goods = [CartItem.objects.get(goods_id=i['goodsid'], color_id=i['colorid'], size_id=i['sizeid'], user=request.session['user']) for i in result]\n\n except Exception as e:\n pass\n return render(request, 'order.html', {'goodss':goods})\n\n\nalipay = AliPay(appid='2016092100565827',\n app_notify_url='http://127.0.0.1:8000/order/checkpay/',\n app_private_key_path='order/keys/self_privite_key.txt',\n alipay_public_key_path='order/keys/alipay_public_key.txt',\n return_url='http://127.0.0.1:8000/order/checkpay/',\n debug=True)\n\ndef toorder(request):\n payway = request.GET.get('payway')\n address_id = request.GET.get('address')\n cartitems = request.GET.get('cartitems')\n totle_rmb = 0\n\n if payway == 'alipay':\n cartitems = jsonpickle.loads(cartitems)\n cartitem = [CartItem.objects.get(color_id=i['colorid'], goods_id=i['goodsid'], size_id=i['sizeid'], count=int(i['count']), user=request.session.get('user', ''), isdelete=False) for i in cartitems if i]\n for i in cartitem:\n totle_rmb += i.goods.price * i.count\n\n out_trade_num = uuid.uuid4().hex\n\n params = alipay.direct_pay(subject='电商支付', out_trade_no=out_trade_num, total_amount=str(totle_rmb))\n\n url = alipay.gateway + \"?\" + params\n\n user = request.session.get('user', '')\n\n order = Order.objects.create(order_num=datetime.now().strftime('%Y%m%d%H%M%S'), out_trade_num=out_trade_num, status='未支付', pay_way=payway, address=Address.objects.get(id=address_id,user=user), user=user)\n [OrderItem.objects.create(color_id=i['colorid'], goods_id=i['goodsid'], size_id=i['sizeid'], count=int(i['count']), order=order) for i in cartitems if i]\n\n return redirect(url)\n\n\ndef checkpay(request):\n params = request.GET.dict()\n sign = params.pop('sign')\n\n #进行校验是否支付成功\n if alipay.verify(params,sign):\n out_trade_num = params.get('out_trade_num')\n trade_no = params.get('trade_no')\n order = Order.objects.get(out_trade_num=out_trade_num)\n order.status = '待收货'\n order.trade_no = trade_no\n order.save()\n [ Inventory.objects.filter(goods=i.goods, size=i.size, color=i.color).update(count=F('count')-int(i.count)) for i in order.orderitem_set.all() if i]\n [ CartItem.objects.filter(goods=i.goods, size=i.size, color=i.color, count=i.count, user=request.session.get('user')).update(isdelete=True) for i in order.orderitem_set.all() if i]\n return HttpResponse('支付成功!')\n\n return HttpResponse('支付失败!')", "sub_path": "order/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3377, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.shortcuts.redirect", "line_number": 16, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 18, "usage_type": "call"}, {"api_name": "jsonpickle.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "cart.models.CartItem.objects.get", "line_number": 27, "usage_type": "call"}, {"api_name": "cart.models.CartItem.objects", "line_number": 27, "usage_type": "attribute"}, {"api_name": "cart.models.CartItem", "line_number": 27, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 31, "usage_type": "call"}, {"api_name": "utils.alipay.AliPay", "line_number": 34, "usage_type": "call"}, {"api_name": "jsonpickle.loads", "line_number": 48, "usage_type": "call"}, {"api_name": "cart.models.CartItem.objects.get", "line_number": 49, "usage_type": "call"}, {"api_name": "cart.models.CartItem.objects", "line_number": 49, "usage_type": "attribute"}, {"api_name": "cart.models.CartItem", "line_number": 49, "usage_type": "name"}, {"api_name": "uuid.uuid4", "line_number": 53, "usage_type": "call"}, {"api_name": "order.models", "line_number": 61, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 61, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 61, "usage_type": "name"}, {"api_name": "order.models", "line_number": 62, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 64, "usage_type": "call"}, {"api_name": "order.models", "line_number": 75, "usage_type": "name"}, {"api_name": "order.models.status", "line_number": 76, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 76, "usage_type": "name"}, {"api_name": "order.models.trade_no", "line_number": 77, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 77, "usage_type": "name"}, {"api_name": "order.models.save", "line_number": 78, "usage_type": "call"}, {"api_name": "order.models", "line_number": 78, "usage_type": "name"}, {"api_name": "django.db.models.F", "line_number": 79, "usage_type": "call"}, {"api_name": "order.models.orderitem_set.all", "line_number": 79, "usage_type": "call"}, {"api_name": "order.models.orderitem_set", "line_number": 79, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 79, "usage_type": "name"}, {"api_name": "cart.models.CartItem.objects.filter", "line_number": 80, "usage_type": "call"}, {"api_name": "cart.models.CartItem.objects", "line_number": 80, "usage_type": "attribute"}, {"api_name": "cart.models.CartItem", "line_number": 80, "usage_type": "name"}, {"api_name": "order.models.orderitem_set.all", "line_number": 80, "usage_type": "call"}, {"api_name": "order.models.orderitem_set", "line_number": 80, "usage_type": "attribute"}, {"api_name": "order.models", "line_number": 80, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 81, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 83, "usage_type": "call"}]} +{"seq_id": "646096253", "text": "# ref: https://towardsdatascience.com/deploying-a-machine-learning-model-as-a-rest-api-4a03b865c166\n# ref2: https://towardsdatascience.com/a-flask-api-for-serving-scikit-learn-models-c8bcdaa41daa\n# ref3: https://github.com/amirziai/sklearnflask\n# ref4: https://www.lynda.com/Flask-tutorials/Web-API-Development-Flask/\nimport sys\nimport os\nimport shutil\nimport time\nimport traceback\n\nfrom flask import Flask, request, jsonify\nfrom sklearn.externals import joblib\nimport pandas as pd\nfrom pandas.io.json import json_normalize\n\napp = Flask(__name__)\n\n# load the model\nmodel = joblib.load('../modeling/rf_clf.pkl')\n\n# load the training data\ntrain_df_template = pd.read_csv('../../data/train_complete.csv', index_col= 0)\n\n# get dummy columns from data template\ndummy_cols = [\"workclass\", \"education\",\n \"marital_stat\", \"occupation\",\n \"relationship\", \"race\",\n \"sex\", \"native_country\"]\n\ntrain_df_template_with_dummies = pd.get_dummies(train_df_template, columns= dummy_cols)\n\ntrain_df_template_with_dummies_no_label = train_df_template_with_dummies.drop(['label'], axis=1)\n\n# save the column names\nmodel_columns = list(train_df_template_with_dummies_no_label.columns)\n\n@app.route('/predict', methods=['POST'])\ndef predict():\n jsonData = request.json\n client_df = pd.get_dummies(pd.DataFrame.from_dict(json_normalize(jsonData), orient='columns'))\n # ref: https://github.com/amirziai/sklearnflask/issues/3\n query = client_df.reindex(columns=model_columns, fill_value=0)\n \n # ref: https://stackoverflow.com/questions/26646362/numpy-array-is-not-json-serializable\n prediction = model.predict(query).tolist()\n\n return jsonify({\n \"prediction\": prediction\n })\n \n\nif __name__ == \"__main__\":\n app.run(debug=True)", "sub_path": "src/flask/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 16, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib.load", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.externals.joblib", "line_number": 19, "usage_type": "name"}, {"api_name": "pandas.read_csv", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.get_dummies", "line_number": 30, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "pandas.get_dummies", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 40, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pandas.io.json.json_normalize", "line_number": 40, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "324889893", "text": "#! /usr/bin/env python\n\"\"\"SearchSource: Unknown source (default for all other sources)\"\"\"\nfrom __future__ import annotations\n\nimport re\nimport typing\nfrom dataclasses import dataclass\nfrom enum import Enum\nfrom pathlib import Path\n\nimport dacite\nimport zope.interface\nfrom dacite import from_dict\nfrom dataclasses_jsonschema import JsonSchemaMixin\nfrom thefuzz import fuzz\n\nimport colrev.env.language_service\nimport colrev.env.package_manager\nimport colrev.exceptions as colrev_exceptions\nimport colrev.ops.search\nimport colrev.record\n\n\n# pylint: disable=unused-argument\n# pylint: disable=duplicate-code\n\n\n@zope.interface.implementer(\n colrev.env.package_manager.SearchSourcePackageEndpointInterface\n)\n@dataclass\nclass UnknownSearchSource(JsonSchemaMixin):\n \"\"\"SearchSource for unknown search results\"\"\"\n\n settings_class = colrev.env.package_manager.DefaultSourceSettings\n\n source_identifier = \"colrev_built_in.unknown_source\"\n search_type = colrev.settings.SearchType.DB\n api_search_supported = False\n ci_supported: bool = False\n heuristic_status = colrev.env.package_manager.SearchSourceHeuristicStatus.na\n short_name = \"Unknown Source\"\n link = (\n \"https://github.com/CoLRev-Ecosystem/colrev/blob/main/\"\n + \"colrev/ops/built_in/search_sources/unknown_source.py\"\n )\n\n HTML_CLEANER = re.compile(\"<.*?>\")\n\n def __init__(\n self, *, source_operation: colrev.operation.CheckOperation, settings: dict\n ) -> None:\n converters = {Path: Path, Enum: Enum}\n self.search_source = from_dict(\n data_class=self.settings_class,\n data=settings,\n config=dacite.Config(type_hooks=converters, cast=[Enum]), # type: ignore\n )\n self.review_manager = source_operation.review_manager\n self.language_service = colrev.env.language_service.LanguageService()\n\n @classmethod\n def heuristic(cls, filename: Path, data: str) -> dict:\n \"\"\"Source heuristic for unknown sources\"\"\"\n\n result = {\"confidence\": 0.0}\n\n return result\n\n @classmethod\n def add_endpoint(\n cls, search_operation: colrev.ops.search.Search, query: str\n ) -> typing.Optional[colrev.settings.SearchSource]:\n \"\"\"Add SearchSource as an endpoint (based on query provided to colrev search -a )\"\"\"\n return None\n\n def validate_source(\n self,\n search_operation: colrev.ops.search.Search,\n source: colrev.settings.SearchSource,\n ) -> None:\n \"\"\"Validate the SearchSource (parameters etc.)\"\"\"\n\n search_operation.review_manager.logger.debug(\n f\"Validate SearchSource {source.filename}\"\n )\n\n if \"query_file\" in source.search_parameters:\n if not Path(source.search_parameters[\"query_file\"]).is_file():\n raise colrev_exceptions.InvalidQueryException(\n f\"File does not exist: query_file {source.search_parameters['query_file']} \"\n f\"for ({source.filename})\"\n )\n\n search_operation.review_manager.logger.debug(\n f\"SearchSource {source.filename} validated\"\n )\n\n def run_search(\n self, search_operation: colrev.ops.search.Search, rerun: bool\n ) -> None:\n \"\"\"Run a search of Crossref\"\"\"\n\n def get_masterdata(\n self,\n prep_operation: colrev.ops.prep.Prep,\n record: colrev.record.Record,\n save_feed: bool = True,\n timeout: int = 10,\n ) -> colrev.record.Record:\n \"\"\"Not implemented\"\"\"\n return record\n\n def load_fixes(\n self,\n load_operation: colrev.ops.load.Load,\n source: colrev.settings.SearchSource,\n records: typing.Dict,\n ) -> dict:\n \"\"\"Load fixes for unknown sources\"\"\"\n\n return records\n\n def prepare(\n self, record: colrev.record.PrepRecord, source: colrev.settings.SearchSource\n ) -> colrev.record.Record:\n \"\"\"Source-specific preparation for unknown sources\"\"\"\n\n # pylint: disable=too-many-branches\n # pylint: disable=too-many-statements\n\n if not record.has_inconsistent_fields() or record.masterdata_is_curated():\n return record\n\n if (\n \"colrev_built_in.md_to_bib\"\n == source.load_conversion_package_endpoint[\"endpoint\"]\n ):\n if \"misc\" == record.data[\"ENTRYTYPE\"] and \"publisher\" in record.data:\n record.update_field(\n key=\"ENTRYTYPE\", value=\"book\", source=\"unkown_source_prep\"\n )\n if record.data.get(\"year\", \"year\") == record.data.get(\"date\", \"date\"):\n record.remove_field(key=\"date\")\n if (\n \"inbook\" == record.data[\"ENTRYTYPE\"]\n and \"chapter\" not in record.data\n and \"title\" in record.data\n ):\n record.rename_field(key=\"title\", new_key=\"chapter\")\n\n if \"UNKNOWN\" != record.data.get(\"author\", \"UNKNOWN\"):\n # fix name format\n if (1 == len(record.data[\"author\"].split(\" \")[0])) or (\n \", \" not in record.data[\"author\"]\n ):\n record.update_field(\n key=\"author\",\n value=colrev.record.PrepRecord.format_author_field(\n input_string=record.data[\"author\"]\n ),\n source=\"unkown_source_prep\",\n keep_source_if_equal=True,\n )\n\n if \"UNKNOWN\" != record.data.get(\"title\", \"UNKNOWN\"):\n record.format_if_mostly_upper(key=\"title\")\n\n if \"date\" in record.data and \"year\" not in record.data:\n year = re.search(r\"\\d{4}\", record.data[\"date\"])\n if year:\n record.update_field(\n key=\"year\",\n value=year.group(0),\n source=\"unkown_source_prep\",\n keep_source_if_equal=True,\n )\n\n if \"UNKNOWN\" != record.data.get(\"journal\", \"UNKNOWN\"):\n if len(record.data[\"journal\"]) > 10 and \"UNKNOWN\" != record.data[\"journal\"]:\n record.format_if_mostly_upper(key=\"journal\", case=\"title\")\n\n # Prepare the record by heuristically correcting erroneous ENTRYTYPEs\n padding = 40\n\n if (\n \"dissertation\" in record.data.get(\"fulltext\", \"NA\").lower()\n and record.data[\"ENTRYTYPE\"] != \"phdthesis\"\n ):\n prior_e_type = record.data[\"ENTRYTYPE\"]\n record.update_field(\n key=\"ENTRYTYPE\", value=\"phdthesis\", source=\"unkown_source_prep\"\n )\n self.review_manager.report_logger.info(\n f' {record.data[\"ID\"]}'.ljust(padding, \" \")\n + f\"Set from {prior_e_type} to phdthesis \"\n '(\"dissertation\" in fulltext link)'\n )\n\n if (\n \"thesis\" in record.data.get(\"fulltext\", \"NA\").lower()\n and record.data[\"ENTRYTYPE\"] != \"phdthesis\"\n ):\n prior_e_type = record.data[\"ENTRYTYPE\"]\n record.update_field(\n key=\"ENTRYTYPE\", value=\"phdthesis\", source=\"unkown_source_prep\"\n )\n self.review_manager.report_logger.info(\n f' {record.data[\"ID\"]}'.ljust(padding, \" \")\n + f\"Set from {prior_e_type} to phdthesis \"\n '(\"thesis\" in fulltext link)'\n )\n\n if (\n \"This thesis\" in record.data.get(\"abstract\", \"NA\").lower()\n and record.data[\"ENTRYTYPE\"] != \"phdthesis\"\n ):\n prior_e_type = record.data[\"ENTRYTYPE\"]\n record.update_field(\n key=\"ENTRYTYPE\", value=\"phdthesis\", source=\"unkown_source_prep\"\n )\n self.review_manager.report_logger.info(\n f' {record.data[\"ID\"]}'.ljust(padding, \" \")\n + f\"Set from {prior_e_type} to phdthesis \"\n '(\"thesis\" in abstract)'\n )\n\n # Journal articles should not have booktitles/series set.\n if \"article\" == record.data[\"ENTRYTYPE\"]:\n if \"booktitle\" in record.data:\n if \"journal\" not in record.data:\n record.update_field(\n key=\"journal\",\n value=record.data[\"booktitle\"],\n source=\"unkown_source_prep\",\n )\n record.remove_field(key=\"booktitle\")\n if \"series\" in record.data:\n if \"journal\" not in record.data:\n record.update_field(\n key=\"journal\",\n value=record.data[\"series\"],\n source=\"unkown_source_prep\",\n )\n record.remove_field(key=\"series\")\n\n if \"article\" == record.data[\"ENTRYTYPE\"]:\n if \"journal\" not in record.data:\n if \"series\" in record.data:\n journal_string = record.data[\"series\"]\n record.update_field(\n key=\"journal\", value=journal_string, source=\"unkown_source_prep\"\n )\n record.remove_field(key=\"series\")\n\n if \"UNKNOWN\" != record.data.get(\"booktitle\", \"UNKNOWN\"):\n if (\n \"UNKNOWN\" != record.data[\"booktitle\"]\n and \"inbook\" != record.data[\"ENTRYTYPE\"]\n ):\n record.format_if_mostly_upper(key=\"booktitle\", case=\"title\")\n\n stripped_btitle = re.sub(r\"\\d{4}\", \"\", record.data[\"booktitle\"])\n stripped_btitle = re.sub(r\"\\d{1,2}th\", \"\", stripped_btitle)\n stripped_btitle = re.sub(r\"\\d{1,2}nd\", \"\", stripped_btitle)\n stripped_btitle = re.sub(r\"\\d{1,2}rd\", \"\", stripped_btitle)\n stripped_btitle = re.sub(r\"\\d{1,2}st\", \"\", stripped_btitle)\n stripped_btitle = re.sub(r\"\\([A-Z]{3,6}\\)\", \"\", stripped_btitle)\n stripped_btitle = stripped_btitle.replace(\n \"Proceedings of the\", \"\"\n ).replace(\"Proceedings\", \"\")\n stripped_btitle = stripped_btitle.lstrip().rstrip()\n record.update_field(\n key=\"booktitle\",\n value=stripped_btitle,\n source=\"unkown_source_prep\",\n keep_source_if_equal=True,\n )\n\n record.unify_pages_field()\n if \"pages\" in record.data:\n if (\n not re.match(r\"^\\d*$\", record.data[\"pages\"])\n and not re.match(r\"^\\d*--\\d*$\", record.data[\"pages\"])\n and not re.match(r\"^[xivXIV]*--[xivXIV]*$\", record.data[\"pages\"])\n ):\n self.review_manager.report_logger.info(\n f' {record.data[\"ID\"]}:'.ljust(padding, \" \")\n + f'Unusual pages: {record.data[\"pages\"]}'\n )\n\n if \"UNKNOWN\" != record.data.get(\"volume\", \"UNKNOWN\"):\n record.update_field(\n key=\"volume\",\n value=record.data[\"volume\"].replace(\"Volume \", \"\"),\n source=\"unkown_source_prep\",\n keep_source_if_equal=True,\n )\n\n if \"url\" in record.data and \"fulltext\" in record.data:\n if record.data[\"url\"] == record.data[\"fulltext\"]:\n record.remove_field(key=\"fulltext\")\n\n if \"language\" in record.data:\n try:\n self.language_service.unify_to_iso_639_3_language_codes(record=record)\n record.update_field(\n key=\"language\",\n value=record.data[\"language\"],\n source=\"unkown_source_prep\",\n keep_source_if_equal=True,\n )\n except colrev_exceptions.InvalidLanguageCodeException:\n del record.data[\"language\"]\n\n for field in list(record.data.keys()):\n # Note : some dois (and their provenance) contain html entities\n if field in [\n \"colrev_masterdata_provenance\",\n \"colrev_data_provenance\",\n \"doi\",\n ]:\n continue\n if field in [\"author\", \"title\", \"journal\"]:\n record.data[field] = re.sub(r\"\\s+\", \" \", record.data[field])\n record.data[field] = re.sub(self.HTML_CLEANER, \"\", record.data[field])\n\n if \"article\" == record.data[\"ENTRYTYPE\"]:\n if \"journal\" in record.data and \"booktitle\" in record.data:\n if (\n fuzz.partial_ratio(\n record.data[\"journal\"].lower(), record.data[\"booktitle\"].lower()\n )\n / 100\n > 0.9\n ):\n record.remove_field(key=\"booktitle\")\n if \"inproceedings\" == record.data[\"ENTRYTYPE\"]:\n if \"journal\" in record.data and \"booktitle\" in record.data:\n if (\n fuzz.partial_ratio(\n record.data[\"journal\"].lower(), record.data[\"booktitle\"].lower()\n )\n / 100\n > 0.9\n ):\n record.remove_field(key=\"journal\")\n\n if record.data.get(\"publisher\", \"\") in [\"researchgate.net\"]:\n record.remove_field(key=\"publisher\")\n\n # Replace nicknames in parentheses\n if \"author\" in record.data:\n record.data[\"author\"] = re.sub(r\"\\([^)]*\\)\", \"\", record.data[\"author\"])\n record.data[\"author\"] = record.data[\"author\"].replace(\" \", \" \").rstrip()\n\n return record\n\n\nif __name__ == \"__main__\":\n pass\n", "sub_path": "colrev/ops/built_in/search_sources/unknown_source.py", "file_name": "unknown_source.py", "file_ext": "py", "file_size_in_byte": 13691, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "dataclasses_jsonschema.JsonSchemaMixin", "line_number": 32, "usage_type": "name"}, {"api_name": "colrev.env.language_service.env", "line_number": 35, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 35, "usage_type": "name"}, {"api_name": "colrev.env.language_service.settings", "line_number": 38, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 38, "usage_type": "name"}, {"api_name": "colrev.env.language_service.env", "line_number": 41, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 41, "usage_type": "name"}, {"api_name": "re.compile", "line_number": 48, "usage_type": "call"}, {"api_name": "colrev.env.language_service.operation", "line_number": 51, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 51, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 53, "usage_type": "name"}, {"api_name": "enum.Enum", "line_number": 53, "usage_type": "name"}, {"api_name": "dacite.from_dict", "line_number": 54, "usage_type": "call"}, {"api_name": "dacite.Config", "line_number": 57, "usage_type": "call"}, {"api_name": "enum.Enum", "line_number": 57, "usage_type": "name"}, {"api_name": "colrev.env.language_service.env.language_service.LanguageService", "line_number": 60, "usage_type": "call"}, {"api_name": "colrev.env.language_service.env", "line_number": 60, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 60, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 63, "usage_type": "name"}, {"api_name": "colrev.env.language_service.ops", "line_number": 72, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 72, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 73, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service.settings", "line_number": 73, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 73, "usage_type": "name"}, {"api_name": "colrev.env.language_service.ops", "line_number": 79, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 79, "usage_type": "name"}, {"api_name": "colrev.env.language_service.settings", "line_number": 80, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 80, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 89, "usage_type": "call"}, {"api_name": "colrev.exceptions.InvalidQueryException", "line_number": 90, "usage_type": "call"}, {"api_name": "colrev.exceptions", "line_number": 90, "usage_type": "name"}, {"api_name": "colrev.env.language_service.ops", "line_number": 100, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 100, "usage_type": "name"}, {"api_name": "colrev.env.language_service.ops", "line_number": 106, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 106, "usage_type": "name"}, {"api_name": "colrev.env.language_service.record", "line_number": 107, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 107, "usage_type": "name"}, {"api_name": "colrev.env.language_service.record", "line_number": 110, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 110, "usage_type": "name"}, {"api_name": "colrev.env.language_service.ops", "line_number": 116, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 116, "usage_type": "name"}, {"api_name": "colrev.env.language_service.settings", "line_number": 117, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 117, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 118, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service.record", "line_number": 125, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 125, "usage_type": "name"}, {"api_name": "colrev.env.language_service.settings", "line_number": 125, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service.record.PrepRecord.format_author_field", "line_number": 159, "usage_type": "call"}, {"api_name": "colrev.env.language_service.record", "line_number": 159, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 159, "usage_type": "name"}, {"api_name": "re.search", "line_number": 170, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 263, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 264, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 265, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 266, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 267, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 268, "usage_type": "call"}, {"api_name": "re.match", "line_number": 283, "usage_type": "call"}, {"api_name": "re.match", "line_number": 284, "usage_type": "call"}, {"api_name": "re.match", "line_number": 285, "usage_type": "call"}, {"api_name": "colrev.exceptions.InvalidLanguageCodeException", "line_number": 313, "usage_type": "attribute"}, {"api_name": "colrev.exceptions", "line_number": 313, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 325, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 326, "usage_type": "call"}, {"api_name": "thefuzz.fuzz.partial_ratio", "line_number": 331, "usage_type": "call"}, {"api_name": "thefuzz.fuzz", "line_number": 331, "usage_type": "name"}, {"api_name": "thefuzz.fuzz.partial_ratio", "line_number": 341, "usage_type": "call"}, {"api_name": "thefuzz.fuzz", "line_number": 341, "usage_type": "name"}, {"api_name": "re.sub", "line_number": 354, "usage_type": "call"}, {"api_name": "colrev.env.language_service.record", "line_number": 126, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 126, "usage_type": "name"}, {"api_name": "zope.interface.interface.implementer", "line_number": 28, "usage_type": "call"}, {"api_name": "zope.interface.interface", "line_number": 28, "usage_type": "attribute"}, {"api_name": "zope.interface", "line_number": 28, "usage_type": "name"}, {"api_name": "colrev.env.language_service.env", "line_number": 29, "usage_type": "attribute"}, {"api_name": "colrev.env.language_service", "line_number": 29, "usage_type": "name"}, {"api_name": "dataclasses.dataclass", "line_number": 31, "usage_type": "name"}]} +{"seq_id": "306109653", "text": "# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\nfrom __future__ import (absolute_import, division, print_function,\n unicode_literals)\n\nimport copy\nimport time\nimport sys\n\nfrom ...utils.compat import argparse\n\nfrom ... import log, __version__\n\nfrom .constants import SSL_SUPPORT\nfrom .errors import SAMPHubError\nfrom .hub import SAMPHubServer\n\nif SSL_SUPPORT:\n import ssl\n\n__all__ = ['main']\n\n\ndef hub_script(timeout=0):\n \"\"\"\n This main function is executed by the ``samp_hub`` command line tool.\n \"\"\"\n\n parser = argparse.ArgumentParser(prog=\"samp_hub \" + __version__)\n\n parser.add_argument(\"-k\", \"--secret\", dest=\"secret\", metavar=\"CODE\",\n help=\"custom secret code.\")\n\n parser.add_argument(\"-d\", \"--addr\", dest=\"addr\", metavar=\"ADDR\",\n help=\"listening address (or IP).\")\n\n parser.add_argument(\"-p\", \"--port\", dest=\"port\", metavar=\"PORT\", type=int,\n help=\"listening port number.\")\n\n parser.add_argument(\"-f\", \"--lockfile\", dest=\"lockfile\", metavar=\"FILE\",\n help=\"custom lockfile.\")\n\n parser.add_argument(\"-w\", \"--no-web-profile\", dest=\"web_profile\", action=\"store_false\",\n help=\"run the Hub disabling the Web Profile.\", default=True)\n\n parser.add_argument(\"-P\", \"--pool-size\", dest=\"pool_size\", metavar=\"SIZE\", type=int,\n help=\"the socket connections pool size.\", default=20)\n\n timeout_group = parser.add_argument_group(\"Timeout group\",\n \"Special options to setup hub and client timeouts.\"\n \"It contains a set of special options that allows to set up the Hub and \"\n \"clients inactivity timeouts, that is the Hub or client inactivity time \"\n \"interval after which the Hub shuts down or unregisters the client. \"\n \"Notification of samp.hub.disconnect MType is sent to the clients \"\n \"forcibly unregistered for timeout expiration.\")\n\n timeout_group.add_argument(\"-t\", \"--timeout\", dest=\"timeout\", metavar=\"SECONDS\",\n help=\"set the Hub inactivity timeout in SECONDS. By default it \"\n \"is set to 0, that is the Hub never expires.\", type=int, default=0)\n\n timeout_group.add_argument(\"-c\", \"--client-timeout\", dest=\"client_timeout\", metavar=\"SECONDS\",\n help=\"set the client inactivity timeout in SECONDS. By default it \"\n \"is set to 0, that is the client never expires.\", type=int, default=0)\n\n parser.add_argument_group(timeout_group)\n\n log_group = parser.add_argument_group(\"Logging options\",\n \"Additional options which allow to customize the logging output. By \"\n \"default the SAMP Hub uses the standard output and standard error \"\n \"devices to print out INFO level logging messages. Using the options \"\n \"here below it is possible to modify the logging level and also \"\n \"specify the output files where redirect the logging messages.\")\n\n log_group.add_argument(\"-L\", \"--log-level\", dest=\"loglevel\", metavar=\"LEVEL\",\n help=\"set the Hub instance log level (OFF, ERROR, WARNING, INFO, DEBUG).\",\n type=str, choices=[\"OFF\", \"ERROR\", \"WARNING\", \"INFO\", \"DEBUG\"], default='INFO')\n\n log_group.add_argument(\"-O\", \"--log-output\", dest=\"logout\", metavar=\"FILE\",\n help=\"set the output file for the log messages.\", default=\"\")\n\n parser.add_argument_group(log_group)\n\n adv_group = parser.add_argument_group(\"Advanced group\",\n \"Advanced options addressed to facilitate administrative tasks and \"\n \"allow new non-standard Hub behaviors. In particular the --label \"\n \"options is used to assign a value to hub.label token and is used to \"\n \"assign a name to the Hub instance. \"\n \"The very special --multi option allows to start a Hub in multi-instance mode. \"\n \"Multi-instance mode is a non-standard Hub behavior that enables \"\n \"multiple contemporaneous running Hubs. Multi-instance hubs place \"\n \"their non-standard lock-files within the /.samp-1 \"\n \"directory naming them making use of the format: \"\n \"samp-hub--, where PID is the Hub process ID while ID is an \"\n \"internal ID (integer).\")\n\n adv_group.add_argument(\"-l\", \"--label\", dest=\"label\", metavar=\"LABEL\",\n help=\"assign a LABEL to the Hub.\", default=\"\")\n\n adv_group.add_argument(\"-m\", \"--multi\", dest=\"mode\",\n help=\"run the Hub in multi-instance mode generating a custom \"\n \"lockfile with a random name.\",\n action=\"store_const\", const='multiple', default='single')\n\n parser.add_argument_group(adv_group)\n\n if SSL_SUPPORT:\n\n ssl_group = parser.add_argument_group(\"SSL group\", \"Additional options to launch \"\n \"the Hub instance using the Secure Sockets Layer (HTTPS). \"\n \"The --key-file and --cert-file parameters specify optional \"\n \"files which contain a certificate to be used to identify \"\n \"the local side of the connection. \"\n \"Often the private key is stored in the same file as the \"\n \"certificate; in this case, only the --cert-file parameter need \"\n \"be passed. If the private key is stored in a separate file, \"\n \"both parameters must be used. If the private key is stored \"\n \"in the certificate file, it should come before the first certificate \"\n \"in the certificate chain.\")\n\n ssl_group.add_argument(\"-s\", \"--https\", dest=\"https\", action=\"store_true\",\n help=\"run the Hub using the Secure Sockets Layer.\", default=False)\n\n ssl_group.add_argument(\"-C\", \"--cert-file\", dest=\"cert_file\", metavar=\"FILE\",\n help=\"set the certificate file.\", default=None)\n\n ssl_group.add_argument(\"-K\", \"--key-file\", dest=\"key_file\", metavar=\"FILE\",\n help=\"set the key file. By default this option is ignored, \"\n \"assuming that the private key is stored in the certificate file.\", default=None)\n\n ssl_group.add_argument(\"--cert-reqs\", dest=\"cert_reqs\", metavar=\"STRING\",\n help=\"this option specifies whether a certificate \"\n \"is required from the client side of the connection, and whether \"\n \"it will be validated if provided. It must be one of the three \"\n \"values NONE (certificates ignored, default), OPTIONAL (not \"\n \"required, but validated if provided), or REQUIRED (required \"\n \"and validated). If the value of this option is not NONE, \"\n \"then the --ca-certs option must point to a file of CA certificates.\",\n type=str, choices=[\"NONE\", \"OPTIONAL\", \"REQUIRED\"], default=\"NONE\")\n\n ssl_group.add_argument(\"--ca-certs\", dest=\"ca_certs\", metavar=\"FILE\",\n help=\"the --ca-certs file contains a set of concatenated \"\n \"\\\"certification authority\\\" certificates, which are used to \"\n \"validate certificates passed from the client end of the \"\n \"connection.\", default=None)\n\n ssl_group.add_argument(\"--ssl-version\", dest=\"ssl_version\", metavar=\"STRING\",\n help=\"the --ssl-version option specifies which version of the \"\n \"SSL protocol to use. Typically, the server chooses a particular \"\n \"protocol version, and the client must adapt to the server's choice. \"\n \"Most of the versions are not interoperable with the other versions. \"\n \"If not specified the default SSL version is taken from the default in \"\n \"the Python standard `ssl` library for the version of Python that is \"\n \"installed. Other SSL protocol versions are: SSLv2, SSLv3, SSLv23, \"\n \"TLSv1, TLSv1_1, TLSv1_2 but not all of them may be available on all \"\n \"versions of Python.\",\n type=str,\n choices=[\"SSLv23\", \"SSLv2\", \"SSLv3\", \"TLSv1\", \"TLSv1_1\", \"TLSv1_2\"],\n default=None)\n\n parser.add_argument_group(ssl_group)\n\n options = parser.parse_args()\n\n try:\n\n if SSL_SUPPORT:\n\n # Set ssl options properly\n\n if options.cert_reqs == \"OPTIONAL\":\n options.cert_reqs = ssl.CERT_OPTIONAL\n elif options.cert_reqs == \"REQUIRED\":\n options.cert_reqs = ssl.CERT_REQUIRED\n else:\n options.cert_reqs = ssl.CERT_NONE\n\n if options.ssl_version is not None:\n if hasattr(ssl, 'PROTOCOL_' + options.ssl_version):\n options.ssl_version = getattr(\n ssl, 'PROTOCOL_' + options.ssl_version)\n else:\n raise ValueError(\n \"SSL protocol '{0}' not supported on this version of \"\n \"Python\".format(options.ssl_version))\n\n if options.loglevel in (\"OFF\", \"ERROR\", \"WARNING\", \"DEBUG\", \"INFO\"):\n log.setLevel(options.loglevel)\n\n if options.logout != \"\":\n context = log.log_to_file(options.logout)\n else:\n class dummy_context(object):\n\n def __enter__(self):\n pass\n\n def __exit__(self, exc_type, exc_value, traceback):\n pass\n context = dummy_context()\n\n with context:\n\n args = copy.deepcopy(options.__dict__)\n del(args[\"loglevel\"])\n del(args[\"logout\"])\n\n hub = SAMPHubServer(**args)\n hub.start(False)\n\n if not timeout:\n while hub.is_running:\n time.sleep(0.01)\n else:\n time.sleep(timeout)\n hub.stop()\n\n except KeyboardInterrupt:\n try:\n hub.stop()\n except NameError:\n pass\n except IOError as e:\n print(\"[SAMP] Error: I/O error({0}): {1}\".format(e.errno, e.strerror))\n sys.exit(1)\n except SystemExit:\n pass\n", "sub_path": "pkgs/astropy-1.1.2-np110py27_0/lib/python2.7/site-packages/astropy/vo/samp/hub_script.py", "file_name": "hub_script.py", "file_ext": "py", "file_size_in_byte": 11796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "constants.SSL_SUPPORT", "line_number": 18, "usage_type": "name"}, {"api_name": "utils.compat.argparse.ArgumentParser", "line_number": 29, "usage_type": "call"}, {"api_name": "utils.compat.argparse", "line_number": 29, "usage_type": "name"}, {"api_name": "constants.SSL_SUPPORT", "line_number": 106, "usage_type": "name"}, {"api_name": "constants.SSL_SUPPORT", "line_number": 166, "usage_type": "name"}, {"api_name": "ssl.CERT_OPTIONAL", "line_number": 171, "usage_type": "attribute"}, {"api_name": "ssl.CERT_REQUIRED", "line_number": 173, "usage_type": "attribute"}, {"api_name": "ssl.CERT_NONE", "line_number": 175, "usage_type": "attribute"}, {"api_name": "copy.deepcopy", "line_number": 203, "usage_type": "call"}, {"api_name": "hub.SAMPHubServer", "line_number": 207, "usage_type": "call"}, {"api_name": "hub.start", "line_number": 208, "usage_type": "call"}, {"api_name": "hub.is_running", "line_number": 211, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 212, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 214, "usage_type": "call"}, {"api_name": "hub.stop", "line_number": 215, "usage_type": "call"}, {"api_name": "hub.stop", "line_number": 219, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 224, "usage_type": "call"}]} +{"seq_id": "55084471", "text": "from django.shortcuts import render,redirect\nfrom django.utils import timezone\nfrom .forms import BookForms,ModelBookForms,SearchForm\n#render the class names\nfrom newhome.models import Book\nfrom django.contrib import messages\n#from .forms1 import BookForms\n\n\n#from newhome.models import Book\n# Create your views here.\n\ndef form_view(request):\n msg=''\n if request.method =='POST':\n form=BookForms(request.POST)\n if form.is_valid():\n book=Book.objects.create(\n name=form.cleaned_data.get('name'),\n book_author=form.cleaned_data.get('author')\n )\n book.save()\n msg='Book Added Successfully!!!'\n else:\n msg=form.errors\n else:\n form=BookForms()\n return render(request,'forms.html',{\"msg\":msg, \"forms\":form})\n\n\ndef model_view(request):\n msg=''\n if request.method =='POST':\n form=ModelBookForms(request.POST)\n if form.is_valid():\n form.save()\n msg ='Book Added Successfully!!!'\n else:\n msg=form.errors\n else:\n form=ModelBookForms()\n return render(request,'forms.html',{\"msg\":msg,\"form\":form})\n\ndef html_form(request):\n value=''\n if request.method=='POST':\n value=request.POST.get('name')\n return render(request,'values.html',{'value':value})\n else:\n return render(request,'design.html')\n\ndef booksearch(request):\n if request.method== 'POST':\n form= SearchForm(request.POST)\n if form.is_valid():\n q=form.cleaned_data.get('q')\n #book = Book.objects.filter(name__contains=q, purchase_date__lte=timezone.now)\n book = Book.objects.filter(name__contains=q)\n form =None\n return render(request,'showtables.html',{'book':book,'form':SearchForm()})\n else:\n form=SearchForm\n book=Book.objects.all()\n return render(request,'showtables.html',{'book':book,'form':form})\n\n\ndef deletebook(request,id):\n print('id',id)\n book.delete()\n messages.success(request,'Deleted'+str(id)+'Successfully!!')\n return redirect(\"/\")\n\n\ndef editbook(request):\n book = Book.objects.get(id = id)\n if request.method==\"POST\":\n form= ModelBookForms(requst.POST)\n if form.is_valid():\n form.save()\n messages.success(request,'Book updated Successfully!!!')\n return redirect(\"/\")\n else:\n form= ModelBookForms(instance = book)\n return render(request,\"editbook.html\",{'form':form})\n \n \n \n #form=CustomForms()\n #book=Book.Objects.all()\n #book=Book.Objects.filter(name='',purchase_date='')\n \n\n\n\"\"\"def newform_view(request):\n form=BookForms()\n #book=Book.Objects.all()\n #book=Book.Objects.filter(name='',purchase_date='')\n context={\n \"head\":\"Book form created here using python\",\n \"forms\":form,\n \n #\"books\":book\n }\n return render(request,'newforms.html',context)\n\n\ndef form1_view(request):\n form1=BookForms()\n #book=Book.Objects.all()\n #book=Book.Objects.filter(name='',purchase_date='')\n context={\n \"head\":\"Book form created here using python\",\n \"forms\":form1,\n \n #\"books\":book\n }\n return render(request,'forms1.html',context)\"\"\"\n\n", "sub_path": "newhome/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 3303, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "forms.BookForms", "line_number": 16, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects.create", "line_number": 18, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects", "line_number": 18, "usage_type": "attribute"}, {"api_name": "newhome.models.Book", "line_number": 18, "usage_type": "name"}, {"api_name": "forms.BookForms", "line_number": 27, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 28, "usage_type": "call"}, {"api_name": "forms.ModelBookForms", "line_number": 34, "usage_type": "call"}, {"api_name": "forms.ModelBookForms", "line_number": 41, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 42, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 48, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 50, "usage_type": "call"}, {"api_name": "forms.SearchForm", "line_number": 54, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects.filter", "line_number": 58, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects", "line_number": 58, "usage_type": "attribute"}, {"api_name": "newhome.models.Book", "line_number": 58, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 60, "usage_type": "call"}, {"api_name": "forms.SearchForm", "line_number": 60, "usage_type": "call"}, {"api_name": "forms.SearchForm", "line_number": 62, "usage_type": "name"}, {"api_name": "newhome.models.Book.objects.all", "line_number": 63, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects", "line_number": 63, "usage_type": "attribute"}, {"api_name": "newhome.models.Book", "line_number": 63, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 64, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 70, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 70, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 71, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects.get", "line_number": 75, "usage_type": "call"}, {"api_name": "newhome.models.Book.objects", "line_number": 75, "usage_type": "attribute"}, {"api_name": "newhome.models.Book", "line_number": 75, "usage_type": "name"}, {"api_name": "forms.ModelBookForms", "line_number": 77, "usage_type": "call"}, {"api_name": "django.contrib.messages.success", "line_number": 80, "usage_type": "call"}, {"api_name": "django.contrib.messages", "line_number": 80, "usage_type": "name"}, {"api_name": "django.shortcuts.redirect", "line_number": 81, "usage_type": "call"}, {"api_name": "forms.ModelBookForms", "line_number": 83, "usage_type": "call"}, {"api_name": "django.shortcuts.render", "line_number": 84, "usage_type": "call"}]} +{"seq_id": "632681529", "text": "import numpy as np\n\nimport theano\nfrom lasagne import layers, nonlinearities\nfrom lasagne.updates import adam\nfrom lasagne.objectives import aggregate, categorical_crossentropy\nfrom nolearn.lasagne import NeuralNet, BatchIterator, TrainSplit\n\n\ndef log_softmax(x):\n xdev = x - x.max(1, keepdims=True)\n return xdev - theano.tensor.log(theano.tensor.sum(theano.tensor.exp(xdev), axis=1, keepdims=True))\n\n\ndef categorical_crossentropy_logdomain(log_predictions, targets):\n return -theano.tensor.sum(targets * log_predictions, axis=1)\n\n\ndef loss(prediction, target):\n return aggregate(categorical_crossentropy_logdomain(prediction,target))\n\n\ndef create_architecture(description, input_shape):\n \n if np.sum(description)==0:\n description[0] += 1\n\n architecture = [\n #Input\n (layers.InputLayer, {'shape': (None, input_shape[1], input_shape[2])})]\n #(layers.DropoutLayer, {'p' : 0.2})]\n\n for num_units in description:\n if num_units != 0:\n architecture.append((layers.DenseLayer, {'num_units' : num_units, 'nonlinearity' : nonlinearities.rectify}))\n #architecture.append((layers.DropoutLayer, {'p': 0.5}))\n\n architecture.append((layers.DenseLayer, {'num_units' : 3, 'nonlinearity' : log_softmax}))\n\n return architecture\n\n\ndef generate_net(description, loss_function, input_shape, train_batch_size=32, epochs=1, verbose=0):\n nnet = NeuralNet(\n #y_tensor_type = theano.tensor.matrix,\n layers = create_architecture(description, input_shape),\n batch_iterator_train = BatchIterator(batch_size = train_batch_size),\n #batch_iterator_test = BatchIterator(batch_size = val_shape),\n\n update = adam,\n\n regression = True,\n train_split = TrainSplit(eval_size=0.0),\n max_epochs = epochs,\n objective_loss_function = loss_function,\n verbose = verbose,\n )\n return nnet\n\n\ndef score_fn_binarized(net, X, y):\n predictions = net.predict(X)\n return np.sum(np.argmax(predictions, axis=1)==np.argmax(y, axis=1))/np.float32(len(np.argmax(y, axis=1)))\n\n\ndef score_fn(net, X, y):\n predictions = net.predict(X)\n return np.sum(predictions==y)/np.float32(len(y))", "sub_path": "code/utils/neural_net.py", "file_name": "neural_net.py", "file_ext": "py", "file_size_in_byte": 2160, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "theano.tensor.log", "line_number": 12, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 12, "usage_type": "attribute"}, {"api_name": "theano.tensor.sum", "line_number": 12, "usage_type": "call"}, {"api_name": "theano.tensor.exp", "line_number": 12, "usage_type": "call"}, {"api_name": "theano.tensor.sum", "line_number": 16, "usage_type": "call"}, {"api_name": "theano.tensor", "line_number": 16, "usage_type": "attribute"}, {"api_name": "lasagne.objectives.aggregate", "line_number": 20, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 25, "usage_type": "call"}, {"api_name": "lasagne.layers.InputLayer", "line_number": 30, "usage_type": "attribute"}, {"api_name": "lasagne.layers", "line_number": 30, "usage_type": "name"}, {"api_name": "lasagne.layers.DenseLayer", "line_number": 35, "usage_type": "attribute"}, {"api_name": "lasagne.layers", "line_number": 35, "usage_type": "name"}, {"api_name": "lasagne.nonlinearities.rectify", "line_number": 35, "usage_type": "attribute"}, {"api_name": "lasagne.nonlinearities", "line_number": 35, "usage_type": "name"}, {"api_name": "lasagne.layers.DenseLayer", "line_number": 38, "usage_type": "attribute"}, {"api_name": "lasagne.layers", "line_number": 38, "usage_type": "name"}, {"api_name": "nolearn.lasagne.NeuralNet", "line_number": 44, "usage_type": "call"}, {"api_name": "nolearn.lasagne.BatchIterator", "line_number": 47, "usage_type": "call"}, {"api_name": "lasagne.updates.adam", "line_number": 50, "usage_type": "name"}, {"api_name": "nolearn.lasagne.TrainSplit", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 63, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 68, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 68, "usage_type": "call"}]} +{"seq_id": "75263185", "text": "import os\nfrom setuptools import setup, find_packages\n\ndef __path(filename):\n return os.path.join(os.path.dirname(__file__), filename)\n\nVERSION = '0.9.11'\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetup( \n name='datahub_core',\n version=VERSION,\n url='https://github.com/finos/datahub', \n packages=find_packages(exclude=['contrib', 'docs', 'tests']),\n package_data={'datahub_core': [\n './data/codes.txt',\n './data/company_names.txt',\n './data/country-codes.txt',\n './data/sic-codes.txt',\n './data/sic-conventions.json',\n './data/sic-ranges.txt',\n './data/addresses/US.json',\n './data/funds/asset_class.csv',\n './data/funds/classes.csv',\n './data/funds/asset_class.csv',\n './data/funds/dividend_treatment.csv',\n './data/funds/regions.csv',\n './data/**'\n ]},\n author='grovesy',\n author_email=\"paul.groves@citi.com\",\n description='Synthetic data generation tooling',\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n include_package_data=True,\n zip_safe=True,\n install_requires=[\n 'Faker',\n 'numpy',\n 'pandas',\n 'scipy', \n ]\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 1270, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.join", "line_number": 5, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 5, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 12, "usage_type": "call"}, {"api_name": "setuptools.find_packages", "line_number": 16, "usage_type": "call"}]} +{"seq_id": "341269771", "text": "from flask_restful import Resource, Api\nfrom flask import jsonify\nimport sqlite3\nfrom ubc_grades_api.api import app\n\nfrom helpers import yearsessions, basic_element_factory, subjects\n\nclass Sections(Resource):\n def get(self, yearsession, subject, course):\n query = \"SELECT DISTINCT section FROM grades WHERE yearsession = ? AND subject = ? AND course = ?;\"\n\n conn = sqlite3.connect(app.config['DATABASE_NAME'])\n conn.row_factory = basic_element_factory\n cur = conn.cursor()\n results = cur.execute(query, [yearsession, subject, course]).fetchall()\n\n return jsonify(results)\n\nclass Courses(Resource):\n def get(self, yearsession, subject):\n query = \"SELECT DISTINCT course FROM grades WHERE yearsession = ? AND subject = ?;\"\n\n conn = sqlite3.connect(app.config['DATABASE_NAME'])\n conn.row_factory = basic_element_factory\n cur = conn.cursor()\n results = cur.execute(query, [yearsession, subject]).fetchall()\n\n return jsonify(results)\n\nclass CoursesNoYearsession(Resource):\n def get(self, subject):\n query = \"SELECT DISTINCT course FROM grades WHERE subject = ?;\"\n\n conn = sqlite3.connect(app.config['DATABASE_NAME'])\n conn.row_factory = basic_element_factory\n cur = conn.cursor()\n results = cur.execute(query, [subject]).fetchall()\n\n return jsonify(results)\n\nclass Subjects(Resource):\n def get(self, yearsession=None):\n if yearsession == None:\n return jsonify(subjects)\n\n query = \"SELECT DISTINCT subject FROM grades WHERE yearsession = ?;\"\n\n conn = sqlite3.connect(app.config['DATABASE_NAME'])\n conn.row_factory = basic_element_factory\n cur = conn.cursor()\n results = cur.execute(query, [yearsession]).fetchall()\n\n return jsonify(results)\n\nclass YearSessions(Resource):\n def get(self):\n return jsonify(yearsessions)", "sub_path": "ubc_grades_api/filters.py", "file_name": "filters.py", "file_ext": "py", "file_size_in_byte": 1920, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask_restful.Resource", "line_number": 8, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 12, "usage_type": "call"}, {"api_name": "ubc_grades_api.api.app.config", "line_number": 12, "usage_type": "attribute"}, {"api_name": "ubc_grades_api.api.app", "line_number": 12, "usage_type": "name"}, {"api_name": "helpers.basic_element_factory", "line_number": 13, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 17, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 19, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 23, "usage_type": "call"}, {"api_name": "ubc_grades_api.api.app.config", "line_number": 23, "usage_type": "attribute"}, {"api_name": "ubc_grades_api.api.app", "line_number": 23, "usage_type": "name"}, {"api_name": "helpers.basic_element_factory", "line_number": 24, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 28, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 30, "usage_type": "name"}, {"api_name": "sqlite3.connect", "line_number": 34, "usage_type": "call"}, {"api_name": "ubc_grades_api.api.app.config", "line_number": 34, "usage_type": "attribute"}, {"api_name": "ubc_grades_api.api.app", "line_number": 34, "usage_type": "name"}, {"api_name": "helpers.basic_element_factory", "line_number": 35, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 39, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 41, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 44, "usage_type": "call"}, {"api_name": "helpers.subjects", "line_number": 44, "usage_type": "argument"}, {"api_name": "sqlite3.connect", "line_number": 48, "usage_type": "call"}, {"api_name": "ubc_grades_api.api.app.config", "line_number": 48, "usage_type": "attribute"}, {"api_name": "ubc_grades_api.api.app", "line_number": 48, "usage_type": "name"}, {"api_name": "helpers.basic_element_factory", "line_number": 49, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 53, "usage_type": "call"}, {"api_name": "flask_restful.Resource", "line_number": 55, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 57, "usage_type": "call"}, {"api_name": "helpers.yearsessions", "line_number": 57, "usage_type": "argument"}]} +{"seq_id": "26353700", "text": "\"\"\"\n 作者: Yang\n 版本:2.0\n 功能:判断日期是这一年的第几天.\n 2.0新增功能:用列表替换元组.\n 日期:22/08/2018\n\"\"\"\nimport datetime\n\n\ndef is_leap_year(year):\n \"\"\"\n 判断year是否为闰年\n 是:返回Ture\n 否:返回False\n \"\"\"\n # 默认不是闰年,假如是则用True赋值.\n is_leap = False\n\n if (year % 400 == 0) or (year % 4 == 0 and year % 100 != 0):\n is_leap = True\n return is_leap\n\n\ndef main():\n \"\"\"\n 主函数\n \"\"\"\n input_date_str = input('请输入日期(yyyy/mm/dd):')\n input_date = datetime.datetime.strptime(input_date_str, '%Y/%m/%d')\n year = input_date.year\n month = input_date.month\n day = input_date.day\n # 元组-----------元组不能改变元素的值.\n # ping_year = (31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31)\n # 列表-----------列表可以改变元素的值.\n # 假如是闰年,则将2月改为29天.\n ping_year = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31]\n if is_leap_year(year):\n ping_year[1] = 29\n days = sum(ping_year[:month - 1]) + day\n print(input_date)\n print('这一天是{}年的第{}天'.format(year, days))\n\n\nif __name__ == '__main__':\n main()\n\n", "sub_path": "which_day/which_day_v2.0.py", "file_name": "which_day_v2.0.py", "file_ext": "py", "file_size_in_byte": 1259, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 30, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 30, "usage_type": "attribute"}]} +{"seq_id": "639897192", "text": "'''5/09/2018 Use JRA-55 and GPCP monthly data to plot 1979-2016 wind and precip difference between MJJAS and NDJFM'''\n\nimport numpy as np\nimport xarray as xr\nimport matplotlib.pyplot as plt\nimport sh\nfrom pylab import rcParams\n\nplot_dir = '/scratch/rg419/plots/monsoon_review_figs/'\nmkdir = sh.mkdir.bake('-p')\nmkdir(plot_dir)\n \n# Load in data\ndata_u = xr.open_dataset('/disca/share/reanalysis_links/jra_55/1958_2016/ucomp_monthly/atmos_monthly_together.nc')\ndata_v = xr.open_dataset('/disca/share/reanalysis_links/jra_55/1958_2016/vcomp_monthly/atmos_monthly_together.nc')\ndata_p = xr.open_dataset('/disca/share/rg419/CMAP_precip.mon.mean.nc')\n#name_temp = '/scratch/rg419/obs_and_reanalysis/GPCP_monthly/gpcp_cdr_v23rB1_y%04d_m%02d.nc'\n#names = [name_temp % (y, m) for y in range(1979,2017) for m in range(1,13)]\n#data_p = xr.open_mfdataset(names)\nland_mask='/scratch/rg419/python_scripts/land_era/ERA-I_Invariant_0125.nc'\nland = xr.open_dataset(land_mask)\n\n#print(data_u.lon)\n#print(data_p.longitude)\n#print(land.longitude)\n\n#print(data_u.sel(time=slice('1979','2016')))\ndata_u = data_u.sel(time=slice('1979','2016'))\ndata_v = data_v.sel(time=slice('1979','2016'))\ndata_p = data_p.sel(time=slice('1979','2016'))\n\n# Make climatologies\nu_clim = data_u.groupby('time.month').mean('time')\nv_clim = data_v.groupby('time.month').mean('time')\np_clim = data_p.groupby('time.month').mean('time')\nprint('means taken')\n\nu_MJJAS = u_clim.sel(month=[5,6,7,8,9]).mean('month') \nv_MJJAS = v_clim.sel(month=[5,6,7,8,9]).mean('month') \np_MJJAS = p_clim.sel(month=[5,6,7,8,9]).mean('month') \n\nu_NDJFM = u_clim.sel(month=[11,12,1,2,3]).mean('month')\nv_NDJFM = v_clim.sel(month=[11,12,1,2,3]).mean('month')\np_NDJFM = p_clim.sel(month=[11,12,1,2,3]).mean('month')\n\n# Calculate MJJAS and NDJFM difference\nu_diff = u_clim.sel(month=[5,6,7,8,9]).mean('month') - u_clim.sel(month=[11,12,1,2,3]).mean('month')\nv_diff = v_clim.sel(month=[5,6,7,8,9]).mean('month') - v_clim.sel(month=[11,12,1,2,3]).mean('month')\np_diff = p_clim.sel(month=[5,6,7,8,9]).mean('month') - p_clim.sel(month=[11,12,1,2,3]).mean('month')\n\n# Define monsoon region using precip\nmonsoon_nh = (p_diff.precip > 2.) & (p_clim.precip.sel(month=[5,6,7,8,9]).sum('month')/p_clim.precip.sum('month') > 0.55) \nmonsoon_sh = (p_diff.precip < -2.) & (p_clim.precip.sel(month=[11,12,1,2,3]).sum('month')/p_clim.precip.sum('month') > 0.55) \nlats_nh = [p_diff.lat[i] for i in range(len(p_diff.lat)) if p_diff.lat[i] > 0]\nlats_sh = [p_diff.lat[i] for i in range(len(p_diff.lat)) if p_diff.lat[i] < 0]\n\n#lats_tropics = [p_diff.lat[i] for i in range(len(p_diff.lat)) if p_diff.lat[i] > -30. and p_diff.lat[i] < 30.]\n\n#itcz_aug= np.zeros(len(p_clim.nlon),)\n#itcz_feb= np.zeros(len(p_clim.nlon),)\n\n#for i in range(len(p_clim.nlon)):\n #print(p_clim.latitude[p_clim.precip.sel(month=8)[:,i] == p_clim.precip.sel(month=8)[:,i].max('nlat')])\n# itcz_aug[i] = p_clim.latitude[p_clim.precip.sel(month=8)[:,i] == p_clim.precip.sel(month=8, nlat=lats_tropics)[:,i].max('nlat')].values\n# itcz_feb[i] = p_clim.latitude[p_clim.precip.sel(month=2)[:,i] == p_clim.precip.sel(month=2, nlat=lats_tropics)[:,i].max('nlat')].values\n \n#print(itcz_aug)\n\n# Make plots\n\n# Set figure parameters\nrcParams['figure.figsize'] = 9, 15\nrcParams['font.size'] = 18\n\nref_arrow=5\narrowdir='uv'\n\nfig, (ax1, ax2, ax3) = plt.subplots(3, 1, sharex='col')\n\n# Contour plot of precipitation difference\nf1 = p_diff.precip.plot.contourf(ax=ax1, x='lon', y='lat', cmap='RdBu', levels=np.arange(-10.,10.1,1.), add_colorbar=False, extend='both', add_labels=False)\nmonsoon_nh.sel(lat=lats_nh).plot.contour(ax=ax1, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nmonsoon_sh.sel(lat=lats_sh).plot.contour(ax=ax1, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nland.lsm[0,:,:].plot.contour(ax=ax1, x='longitude', y='latitude', levels=np.arange(-1.,2.,1.), add_labels=False, colors='k', linewidths=2.)\n#plt.plot(p_diff.longitude, itcz_aug, 'r')\n#plt.plot(p_diff.longitude, itcz_feb, 'b')\nb = ax1.quiver(u_diff.lon[::5], u_diff.lat[::2], u_diff.var33.sel(lev=85000.)[::2,::5], v_diff.var34.sel(lev=85000.)[::2,::5], angles=arrowdir, scale=200.)\nax1.set_title('MJJAS - NDJFM')\nax1.quiverkey(b, 5.,65., ref_arrow, str(ref_arrow) + ' m/s', coordinates='data', fontproperties={'weight': 'bold', 'size': 10}, color='k', labelcolor='k', labelsep=0.03, zorder=10)\n\n# Contour plot of MJJAS precipitation\nf1 = p_MJJAS.precip.plot.contourf(ax=ax2, x='lon', y='lat', cmap='RdBu', levels=np.arange(-10.,10.1,1.), add_colorbar=False, extend='both', add_labels=False)\nmonsoon_nh.sel(lat=lats_nh).plot.contour(ax=ax2, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nmonsoon_sh.sel(lat=lats_sh).plot.contour(ax=ax2, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nland.lsm[0,:,:].plot.contour(ax=ax2, x='longitude', y='latitude', levels=np.arange(-1.,2.,1.), add_labels=False, colors='k', linewidths=2.)\n#plt.plot(p_diff.longitude, itcz_aug, 'r')\n#plt.plot(p_diff.longitude, itcz_feb, 'b')\nb = ax2.quiver(u_MJJAS.lon[::5], u_MJJAS.lat[::2], u_MJJAS.var33.sel(lev=85000.)[::2,::5], v_MJJAS.var34.sel(lev=85000.)[::2,::5], angles=arrowdir, scale=200.)\nax2.set_title('MJJAS')\nax2.quiverkey(b, 5.,65., ref_arrow, str(ref_arrow) + ' m/s', coordinates='data', fontproperties={'weight': 'bold', 'size': 10}, color='k', labelcolor='k', labelsep=0.03, zorder=10)\n\n# Contour plot of NDJFM precipitation\nf1 = p_NDJFM.precip.plot.contourf(ax=ax3, x='lon', y='lat', cmap='RdBu', levels=np.arange(-10.,10.1,1.), add_colorbar=False, extend='both', add_labels=False)\nmonsoon_nh.sel(lat=lats_nh).plot.contour(ax=ax3, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nmonsoon_sh.sel(lat=lats_sh).plot.contour(ax=ax3, x='lon', y='lat', levels=np.arange(-1.,2.,1.), colors='m', add_colorbar=False, linewidths=2, add_labels=False)\nland.lsm[0,:,:].plot.contour(ax=ax3, x='longitude', y='latitude', levels=np.arange(-1.,2.,1.), add_labels=False, colors='k', linewidths=2.)\n#plt.plot(p_diff.longitude, itcz_aug, 'r')\n#plt.plot(p_diff.longitude, itcz_feb, 'b')\nb = ax3.quiver(u_NDJFM.lon[::5], u_NDJFM.lat[::2], u_NDJFM.var33.sel(lev=85000.)[::2,::5], v_NDJFM.var34.sel(lev=85000.)[::2,::5], angles=arrowdir, scale=200.)\nax3.set_title('NDJFM')\nax3.quiverkey(b, 5.,65., ref_arrow, str(ref_arrow) + ' m/s', coordinates='data', fontproperties={'weight': 'bold', 'size': 10}, color='k', labelcolor='k', labelsep=0.03, zorder=10)\n#ax3.quiverkey(b, 45.,-90., ref_arrow, str(ref_arrow) + ' m/s', coordinates='data', fontproperties={'weight': 'bold', 'size': 10}, color='k', labelcolor='k', labelsep=0.03, zorder=10)\n\nfor ax in [ax1,ax2,ax3]:\n ax.grid(True,linestyle=':')\n ax.set_ylabel('Latitude')\n ax.set_ylim(-60.,60.)\n ax.set_xticks(np.arange(0.,360.,90.))\n ax.set_yticks(np.arange(-60.,61.,30.))\n\nax1.text(-50, 60, 'a)')\nax2.text(-50, 60, 'b)')\nax3.text(-50, 60, 'c)')\n\nax3.set_xlabel('Longitude')\n\nplt.subplots_adjust(left=0.12, right=0.97, top=0.97, bottom=0.02, hspace=0.2, wspace=0.1)\n\ncb1=fig.colorbar(f1, ax=[ax1,ax2,ax3], use_gridspec=True, orientation = 'horizontal',fraction=0.05, pad=0.07, aspect=30, shrink=0.5)\ncb1.set_label('Precipitation, mm/day')\n\nplt.savefig(plot_dir + 'wind_precip_reversal_winter_summer.pdf', format='pdf')\nplt.close()\n\ndata_u.close()\ndata_v.close()\ndata_p.close()", "sub_path": "monsoon_review_figs/wind_precip_reversal_winter_summer.py", "file_name": "wind_precip_reversal_winter_summer.py", "file_ext": "py", "file_size_in_byte": 7577, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sh.mkdir.bake", "line_number": 10, "usage_type": "call"}, {"api_name": "sh.mkdir", "line_number": 10, "usage_type": "attribute"}, {"api_name": "xarray.open_dataset", "line_number": 14, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 15, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 16, "usage_type": "call"}, {"api_name": "xarray.open_dataset", "line_number": 21, "usage_type": "call"}, {"api_name": "pylab.rcParams", "line_number": 72, "usage_type": "name"}, {"api_name": "pylab.rcParams", "line_number": 73, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 81, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 82, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 84, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 94, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 103, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 104, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 105, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 106, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 119, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots_adjust", "line_number": 127, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 127, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}]} +{"seq_id": "167076122", "text": "\"\"\"\nSIR disease model\nS' = -beta*S*I\nI' = beta*S*I - nu*I\nR' = nu*I\n\"\"\"\n\nimport numpy as np\nfrom ODESOLVER import ForwardEuler\nfrom matplotlib import pyplot as plt\n\nclass SIR:\n def __init__(self, nu, beta, S0, I0, R0):\n #nu and Beta are paramters in ODE , meaning that the S0 I0 and R0 are the inital values\n\n if isinstance(nu, (float, int)):\n # Is number?\n self.nu = lambda t: nu\n\t\t\t#for any t u put in you get the same nu out \n elif callable(nu):\n self.nu = nu\n\n if isinstance(beta, (float, int)):\n self.beta = lambda t: beta \n\t\t\t#for any t u put in you get the same beta out\n elif callable(beta):\n self.beta = beta\n\n self.initial_conditions = [S0, I0, R0]\n\n def __call__(self, u, t):\n\n S, I, R = u \n\n return np.asarray([\n -self.beta(t)*S*I, # Susceptibles\n self.beta(t)*S*I - self.nu(t)*I, # Infected\n self.nu(t)*I # Recovered\n ])\n\nif __name__ == \"__main__\":\n\n\t#parameters that help model the virus. beta is the probabilty that a person gets infected, nu is the probabilty of an infected recovering, the third one is the size of the control group (population), the fourth one is the number of infected people being at the start of the simluation (DAY 0) and the last one is teh number of removed (dead) or recovered.\n #The parameters given are random for testing\n SIR = SIR(0.140, 0.285, 1500, 1, 0)\n solver = ForwardEuler(SIR)\n solver.set_initial_conditions(SIR.initial_conditions)\n\n time_steps = np.linspace(0, 100, 100000001)\n u, t = solver.solve(time_steps)\n\n plt.plot(t, u[:, 0], label=\"Susceptible\")\n plt.plot(t, u[:, 1], label=\"Infected\")\n plt.plot(t, u[:, 2], label=\"Recovered\")\n plt.legend()\n plt.show()", "sub_path": "SIR.py", "file_name": "SIR.py", "file_ext": "py", "file_size_in_byte": 1806, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.asarray", "line_number": 35, "usage_type": "call"}, {"api_name": "ODESOLVER.ForwardEuler", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}]} +{"seq_id": "419313978", "text": "from flask import Flask, render_template, wrappers\nfrom webapp.weather import weather_by_city\nfrom webapp.python_org_news import get_python_news\nCITY = {'moscow':('55.89','37.47'),'murmansk':('68.97', '33.07')}\n\ndef create_app():\n app = Flask(__name__)\n app.config.from_pyfile('config.py')\n\n @app.route('//')\n def index(city):\n print(city)\n title = 'Новости Python'\n get_city = CITY.get(city, CITY['moscow'])\n weather = weather_by_city(*get_city)\n get_news_list = get_python_news()\n return render_template('index.html', page_title = title, weather_text = weather, news_list = get_news_list)\n\n return app\n\n ", "sub_path": "webapp/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 679, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask.Flask", "line_number": 7, "usage_type": "call"}, {"api_name": "webapp.weather.weather_by_city", "line_number": 15, "usage_type": "call"}, {"api_name": "webapp.python_org_news.get_python_news", "line_number": 16, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 17, "usage_type": "call"}]} +{"seq_id": "334829645", "text": "import gym\r\nimport sys\r\nimport numpy as np\r\nimport random\r\nimport time\r\nfrom IPython.display import clear_output\r\nfrom PIL import Image\r\nimport cv2\r\nimport matplotlib.pyplot as plt\r\nimport pickle\r\nfrom matplotlib import style\r\nfrom gym.envs.registration import register\r\nfrom os.path import isfile, join\r\nimport yaml\r\nimport os\r\nimport inspect\r\ncurrentdir = os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))\r\nparentdir = os.path.dirname(currentdir)\r\nsys.path.insert(0, parentdir) \r\n# from max_flow import FindMaximumMatching\r\n# from max_flow import make_circuit_video\r\n\r\n# import wandb\r\n\r\nrandom.seed(0)\r\n\r\nfloor_plan_real = [[0,0,0,0,0,0,0,0,0,0],\r\n [0,1,1,1,1,1,1,1,0,0],\r\n [0,0,0,0,0,0,0,0,0,0],\r\n [0,0,0,0,0,0,0,0,0,0],\r\n [0,1,1,1,1,1,1,1,0,0],\r\n [0,0,0,0,0,0,0,0,0,0],\r\n [0,0,0,0,0,0,0,0,0,0],\r\n [0,1,1,1,1,1,1,1,0,0],\r\n [0,0,0,0,0,0,0,0,0,0]]\r\n\r\nagent_skills_2task = [[1,0,1],\r\n [0,0,1],\r\n [1,1,0]\r\n ]\r\n\r\nsingle_agent = [[0,1,0]]\r\n\r\n# wandb.init(project=\"MaxFlow Reward Tracking\")\r\nprint(\"Creating Environment...\")\r\nenv = gym.make('gym_oaas:oaas-v0', floor_plan=floor_plan_real, agent_skills=agent_skills_2task)\r\nprint(\"Environment Created!\")\r\n\r\nprint(env.task_dict_map)\r\nprint(env.task_dict)\r\nprint(\"Possible Schedules = \",env.possible_schedules)\r\nprint(\"Possible Actions = \",env.action_sets)\r\n\r\nsys.exit(0)\r\n\r\n#random_scheduler = np.zeros((env.n, len(env.task_dict)))\r\n#random_scheduler = [[0] * len(env.task_dict)] * env.n\r\nrandom_scheduler = [[0] * len(env.task_dict) for i in range(env.n)]\r\nprint(random_scheduler)\r\n#sys.exit(0)\r\n\r\nfor action_set in range(len(env.action_sets)):\r\n\r\n if not env.action_sets[action_set]:\r\n continue\r\n scheudled_task = random.choice(env.action_sets[action_set])\r\n print(\"Schedule picked: \", scheudled_task)\r\n print(\"action_set: \", action_set)\r\n if env.task_dict[env.task_dict_map[scheudled_task]].max_agents == 0:\r\n for rest in range(action_set, len(env.action_sets)):\r\n env.possible_schedules[rest][scheudled_task] = None\r\n if scheudled_task in env.action_sets[rest]:\r\n env.action_sets[rest].remove(scheudled_task)\r\n\r\n continue\r\n\r\n random_scheduler[action_set][scheudled_task] = 1\r\n print(random_scheduler)\r\n print(\"--------------\")\r\n env.task_dict[env.task_dict_map[scheudled_task]].max_agents -= 1\r\n \r\n if env.task_dict[env.task_dict_map[scheudled_task]].max_agents == 0:\r\n for rest in range(action_set, len(env.action_sets)):\r\n env.possible_schedules[rest][scheudled_task] = None\r\n if scheudled_task in env.action_sets[rest]:\r\n env.action_sets[rest].remove(scheudled_task)\r\n\r\nprint('Random Scheduler = ',random_scheduler)\r\n\r\n\r\n# RANDOM SCHEDULER #\r\n\r\nsys.exit(0)\r\n\r\nprint(\"Num Employees = \",len(env.employee_dict))\r\nprint(\"Num Tasks = \", len(env.task_dict))\r\ntemp_reward = []\r\n#sys.exit(0)\r\n\r\n# -------- Max Flow Scheduler - Baseline ------- #\r\nfor iter in range(5):\r\n\r\n env.fetch_new_tasks()\r\n\r\n # print(\"Reached this and printing data\")\r\n # print(env.employee_dict)\r\n # print(env.task_dict)\r\n\r\n num_agents = len(env.agent_skills)\r\n num_possible_tasks = len(env.agent_skills[0])\r\n\r\n max_flow_schedule = np.zeros((num_agents,num_possible_tasks))\r\n vertices = []\r\n edges = []\r\n\r\n for key in env.employee_dict:\r\n vertices.append(key)\r\n for key in env.task_dict:\r\n vertices.append(key)\r\n\r\n for agent in env.employee_dict:\r\n for skill in range(len(env.employee_dict[agent].skill)):\r\n #print(\"Agent = \", agent)\r\n #print(\"Skill = \", skill)\r\n #print(env.employee_dict[agent].skill[skill])\r\n if env.employee_dict[agent].skill[skill] == 1 and env.task_dict_map[skill] in env.task_dict:\r\n edges.append({agent,env.employee_dict[agent].task_dict_map[skill]})\r\n\r\n print(\"Vertices: \",vertices)\r\n print(\"Edges: \",edges)\r\n\r\n f = FindMaximumMatching(edges, vertices)\r\n f.find_maximum_matching()\r\n # print(\"Schedule complete: \",f.matching)\r\n\r\n empl_idx = []\r\n task_idx = []\r\n\r\n for matching in f.matching:\r\n for node in matching:\r\n if 'employee' in node:\r\n idx = node.replace('employee', '')\r\n empl_idx.append(int(idx))\r\n else:\r\n task_idx.append(env.task_dict_inverse_map[node])\r\n\r\n max_flow_schedule.astype(int)\r\n max_flow_schedule = max_flow_schedule.tolist()\r\n\r\n #print(\"Employee indices = \", empl_idx)\r\n #print(\"Task indices = \", task_idx)\r\n #print(\"max_flow_sched_size = (\", len(max_flow_schedule),\",\",len(max_flow_schedule[0]),\")\")\r\n #print(len(empl_idx))\r\n for i in range(len(empl_idx)):\r\n #print(\"At iteration \", i, \"setting maxflow to 1: (\",empl_idx[i],\",\",task_idx[i],\")\")\r\n #print(i)\r\n max_flow_schedule[empl_idx[i]][task_idx[i]] = 1\r\n\r\n \r\n print(max_flow_schedule)\r\n\r\n task_action = [[1,0,0],\r\n [0,0,1],\r\n [0,1,0]]\r\n\r\n #print(type(task_action))\r\n #print(type(max_flow_schedule))\r\n\r\n\r\n \r\n #sys.exit(0)\r\n #make_circuit_video('animation.gif', fps=1)\r\n\r\n #sys.exit(0)\r\n\r\n # -------- High Level Scheduler ------- #\r\n\r\n # env.render()\r\n\r\n # sys.exit(0)\r\n # task_action = [[1,0,0],\r\n # [0,0,1],\r\n # [0,1,0],\r\n # [0,0,1],\r\n # [0,1,0]]\r\n\r\n for i in range(15):\r\n observation, reward, done, info = env.schedule_step(max_flow_schedule)\r\n temp_reward.append(reward)\r\n #env.render()\r\n #env.create_tasks\r\n\r\n#env.close()\r\nplt.plot(temp_reward)\r\nplt.show()\r\n#make_circuit_video('animation.gif', fps=1)\r\nsys.exit(0)\r\n# -------- Reward Function Demonstration ------- #\r\n\r\n# for i in range(500):\r\n# #print(\"i = \",i)\r\n# #print(\"i % 2 \",i %2)\r\n# if i<25:\r\n# if i % 2 == 0:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 1\r\n# action.append(action_i)\r\n\r\n# elif i % 2 == 1:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 2\r\n# action.append(action_i)\r\n\r\n# elif i>=25 and i < 60:\r\n# if i % 2 == 0:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 3\r\n# action.append(action_i)\r\n\r\n# elif i % 2 == 1:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 4\r\n# action.append(action_i)\r\n\r\n# elif i>=60:\r\n# if i % 2 == 0:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 1\r\n# action.append(action_i)\r\n\r\n# elif i % 2 == 1:\r\n# action = []\r\n# for i in range(len(env.action_space)):\r\n# action_i = 1\r\n# action.append(action_i)\r\n\r\n# env.step(action)\r\n# env.render()\r\n\r\n# sys.exit(0)\r\n\r\n# -------- Random Action Behavior ------- #\r\n\r\nfor i in range(15):\r\n action = []\r\n for i in range(len(env.action_space)):\r\n action_i = env.action_space[i].sample()\r\n action.append(action_i)\r\n env.step(action)\r\n env.render()\r\n\r\nenv.close()\r\nsys.exit(0)\r\n\r\n\r\n# -------- MADDPG information ------- #\r\n\r\nprint('number of agents', env.n)\r\nprint('observation space', env.observation_space)\r\nprint('action space', env.action_space)\r\nprint('n actions', env.action_space[0].n)\r\n\r\nobservation = env.reset()\r\nprint(observation)\r\n\r\n#no_op = 0\r\nno_op = np.array([1,0,0,0,0])\r\n\r\naction = [no_op]\r\nprint(\"Action = \",action)\r\n\r\nobs_, reward, done, info = env.step(action)\r\n\r\n\r\nprint(reward)\r\nprint(done)\r\nenv.render()\r\n\r\nsys.exit(0)\r\n\r\n# -------- Random Action Behavior ------- #\r\n\r\nfor i in range(5):\r\n action = []\r\n for i in range(len(env.action_space)):\r\n action_i = env.action_space[i].sample()\r\n action.append(action_i)\r\n env.step(action)\r\n env.render()\r\n\r\n#env.close()\r\nsys.exit(0)\r\n\r\n# -------- Archive ------- #\r\n\r\nprint(\"Action = \",action)\r\n\r\nobs_, reward, done, info = env.step(action)\r\nprint(reward)\r\nprint(done)\r\nsys.exit(0)\r\nprint(env.reset())\r\n\r\n\r\nsys.exit(0)\r\n\r\nfor i in range(10):\r\n action = []\r\n for i in range(len(env.action_space)):\r\n action_i = env.action_space[i].sample()\r\n action.append(action_i)\r\n env.step(action)\r\n env.render()\r\n\r\nsys.exit(0)\r\n\r\nenv = DummyVecEnv([lambda: env])\r\nmodel = PPO('MlpPolicy', env, verbose = 1)\r\nmodel.learn(total_timesteps=100000)\r\nPPO_path = os.path.join('Training', 'Saved Models', 'PPO_model')\r\nmodel.save(PPO_path)\r\n\r\nsys.exit(0)\r\nlog_path = os.path.join('Training', 'Logs')\r\nmodel = PPO('MlpPolicy', env, verbose=1, tensorboard_log=log_path)\r\nmodel.learn(total_timesteps=100000)\r\nPPO_path = os.path.join('Training', 'Saved Models', 'PPO_model')\r\nmodel.save(PPO_path)\r\n\r\n\r\nsys.exit(0)\r\n\r\n\r\n#env = make_vec_env('gym_oaas:oaas-v0', floor_plan=floor_plan1, agent_skills=agent_skills1, n_envs=4)\r\nenv = DummyVecEnv([lambda: env])\r\nmodel = PPO(\"MlpPolicy\", env, verbose=1)\r\nmodel.learn(total_timesteps=25000)\r\n\r\n#model.learn(total_timesteps=20000)\r\nsys.exit(0)\r\n\r\n\r\nlog_path = os.path.join('Training', 'Logs')\r\nmodel = PPO('MlpPolicy', env, verbose=1, tensorboard_log=log_path)\r\nmodel.learn(total_timesteps=4000)\r\n\r\n\r\n\r\nsys.exit(0)\r\n\r\naction_space_size = env.action_space\r\nstate_space_size = env.observation_space\r\n#SIZE = env.\r\n\r\nif start_q_table1 is None:\r\n # initialize the q-table#\r\n q_table1 = {}\r\n for i in range(-SIZE+1, SIZE):\r\n for ii in range(-SIZE+1, SIZE):\r\n for iii in range(-SIZE+1, SIZE):\r\n for iiii in range(-SIZE+1, SIZE):\r\n q_table1[((i, ii), (iii, iiii))] = [np.random.uniform(-5, 0) for i in range(4)]\r\n #print(\"Size of q_table is \", len(q_table1))\r\n\r\nelse:\r\n with open(start_q_table1, \"rb\") as f:\r\n q_table1 = pickle.load(f)\r\n #print(\"Size of q_table is \", len(q_table1))\r\n\r\nnum_episodes = 10000\r\nmax_steps_per_episode = 100\r\n\r\nlearning_rate = 0.1\r\ndiscount_rate = 0.99\r\n\r\nexploration_rate = 1\r\nmax_exploration_rate = 1\r\nmin_exploration_rate = 0.01\r\nexploration_decay_rate = 0.01\r\n\r\nrewards_all_episodes = []\r\n\r\n# Q-learning algorithm\r\nfor episode in range(num_episodes):\r\n # initialize new episode params\r\n state = env.reset()\r\n done = False\r\n rewards_current_episode = 0\r\n\r\n for step in range(max_steps_per_episode): \r\n\r\n # Exploration-exploitation trade-off\r\n exploration_rate_threshold = random.uniform(0, 1) \r\n if exploration_rate_threshold > exploration_rate:\r\n action = np.argmax(q_table[state,:]) \r\n else:\r\n action = env.action_space.sample()\r\n \r\n # Take new action\r\n new_state, reward, done, info = env.step(action)\r\n\r\n # Update Q-table for Q(s,a)\r\n q_table[state, action] = q_table[state, action] * (1 - learning_rate) + \\\r\n learning_rate * (reward + discount_rate * np.max(q_table[new_state, :]))\r\n\r\n # Set new state\r\n state = new_state\r\n\r\n # Add new reward \r\n rewards_current_episode += reward \r\n\r\n # Exploration rate decay\r\n exploration_rate = min_exploration_rate + \\\r\n (max_exploration_rate - min_exploration_rate) * np.exp(-exploration_decay_rate*episode)\r\n \r\n # Add current episode reward to total rewards list\r\n rewards_all_episodes.append(rewards_current_episode)\r\n\r\n\r\n# Calculate and print the average reward per thousand episodes\r\nrewards_per_thousand_episodes = np.split(np.array(rewards_all_episodes),num_episodes/1000)\r\ncount = 1000\r\n\r\nprint(\"********Average reward per thousand episodes********\\n\")\r\nfor r in rewards_per_thousand_episodes:\r\n print(count, \": \", str(sum(r/1000)))\r\n count += 1000", "sub_path": "gym-oaas/solvers/random_scheduler.py", "file_name": "random_scheduler.py", "file_ext": "py", "file_size_in_byte": 12093, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.path.dirname", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path", "line_number": 17, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 17, "usage_type": "call"}, {"api_name": "inspect.getfile", "line_number": 17, "usage_type": "call"}, {"api_name": "inspect.currentframe", "line_number": 17, "usage_type": "call"}, {"api_name": "os.path.dirname", "line_number": 18, "usage_type": "call"}, {"api_name": "os.path", "line_number": 18, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 19, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "random.seed", "line_number": 25, "usage_type": "call"}, {"api_name": "gym.make", "line_number": 46, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 54, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 66, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 93, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 194, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 194, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 195, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 195, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 197, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 258, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 272, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 284, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 297, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 306, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 310, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 320, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 325, "usage_type": "call"}, {"api_name": "os.path", "line_number": 325, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 328, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 329, "usage_type": "call"}, {"api_name": "os.path", "line_number": 329, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 332, "usage_type": "call"}, {"api_name": "os.path", "line_number": 332, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 336, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 345, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path", "line_number": 348, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 354, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 367, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 367, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 372, "usage_type": "call"}, {"api_name": "random.uniform", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 400, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 409, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 419, "usage_type": "call"}, {"api_name": "numpy.split", "line_number": 426, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 426, "usage_type": "call"}]} +{"seq_id": "497223603", "text": "# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport argparse\nimport numpy as np\nfrom scipy.special import softmax\n\nimport paddle\nfrom paddle import inference\nfrom paddlenlp.data import Stack, Tuple, Pad\nfrom paddlenlp.transformers import BertTokenizer\n\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument(\n \"--model_path\",\n default=None,\n type=str,\n required=True,\n help=\"The path prefix of inference model to be used.\", )\n parser.add_argument(\n \"--device\",\n default=\"gpu\",\n type=str,\n choices=[\"cpu\", \"gpu\", \"xpu\"],\n help=\"The device to select to train the model, is must be cpu/gpu/xpu.\")\n parser.add_argument(\n \"--max_seq_length\",\n default=128,\n type=int,\n help=\"The maximum total input sequence length after tokenization. Sequences longer \"\n \"than this will be truncated, sequences shorter will be padded.\", )\n args = parser.parse_args()\n return args\n\n\ndef convert_example(example, tokenizer, label_list, max_seq_length=128):\n text = example\n encoded_inputs = tokenizer(text=text, max_seq_len=max_seq_length)\n input_ids = encoded_inputs[\"input_ids\"]\n segment_ids = encoded_inputs[\"token_type_ids\"]\n\n return input_ids, segment_ids\n\n\nclass Predictor(object):\n def __init__(self, predictor, input_handles, output_handle, tokenizer,\n max_seq_length):\n self.predictor = predictor\n self.input_handles = input_handles\n self.output_handle = output_handle\n self.tokenizer = tokenizer\n self.max_seq_length = max_seq_length\n\n @classmethod\n def create_predictor(cls, args):\n max_seq_length = args.max_seq_length\n config = paddle.inference.Config(args.model_path + \".pdmodel\",\n args.model_path + \".pdiparams\")\n if args.device == \"gpu\":\n # Set GPU configs accordingly\n config.enable_use_gpu(100, 0)\n elif args.device == \"cpu\":\n # Set CPU configs accordingly,\n # such as enable_mkldnn, set_cpu_math_library_num_threads\n config.disable_gpu()\n elif args.device == \"xpu\":\n # Set XPU configs accordingly\n config.enable_xpu(100)\n config.switch_use_feed_fetch_ops(False)\n predictor = paddle.inference.create_predictor(config)\n input_handles = [\n predictor.get_input_handle(name)\n for name in predictor.get_input_names()\n ]\n output_handle = predictor.get_output_handle(predictor.get_output_names()\n [0])\n tokenizer = BertTokenizer.from_pretrained(\n os.path.dirname(args.model_path))\n\n return cls(predictor, input_handles, output_handle, tokenizer,\n max_seq_length)\n\n def predict(self, data, label_map, batch_size=1):\n examples = []\n for text in data:\n input_ids, segment_ids = convert_example(\n text,\n self.tokenizer,\n label_list=label_map.values(),\n max_seq_length=self.max_seq_length)\n examples.append((input_ids, segment_ids))\n\n batchify_fn = lambda samples, fn=Tuple(\n Pad(axis=0, pad_val=self.tokenizer.pad_token_id, dtype=\"int64\"), # input\n Pad(axis=0, pad_val=self.tokenizer.pad_token_id, dtype=\"int64\"), # segment\n ): fn(samples)\n\n # Seperates data into some batches.\n batches = [\n examples[idx:idx + batch_size]\n for idx in range(0, len(examples), batch_size)\n ]\n\n outputs = []\n results = []\n for batch in batches:\n input_ids, segment_ids = batchify_fn(batch)\n self.input_handles[0].copy_from_cpu(input_ids)\n self.input_handles[1].copy_from_cpu(segment_ids)\n self.predictor.run()\n logits = self.output_handle.copy_to_cpu()\n probs = softmax(logits, axis=1)\n idx = np.argmax(probs, axis=1)\n idx = idx.tolist()\n labels = [label_map[i] for i in idx]\n outputs.extend(probs)\n results.extend(labels)\n return outputs, results\n\n\ndef main():\n args = parse_args()\n predictor = Predictor.create_predictor(args)\n\n data = [\n 'against shimmering cinematography that lends the setting the ethereal beauty of an asian landscape painting',\n 'the situation in a well-balanced fashion',\n 'at achieving the modest , crowd-pleasing goals it sets for itself',\n 'so pat it makes your teeth hurt',\n 'this new jangle of noise , mayhem and stupidity must be a serious contender for the title .'\n ]\n label_map = {0: 'negative', 1: 'positive'}\n\n outputs, results = predictor.predict(data, label_map)\n for idx, text in enumerate(data):\n print(\n 'Data: {} \\n Label: {} \\n Negative prob: {} \\n Positive prob: {} \\n '.\n format(text, results[idx], outputs[idx][0], outputs[idx][1]))\n\n\nif __name__ == \"__main__\":\n main()\n", "sub_path": "examples/language_model/bert/predict.py", "file_name": "predict.py", "file_ext": "py", "file_size_in_byte": 5695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 27, "usage_type": "call"}, {"api_name": "paddle.inference.Config", "line_number": 71, "usage_type": "call"}, {"api_name": "paddle.inference", "line_number": 71, "usage_type": "attribute"}, {"api_name": "paddle.inference.create_predictor", "line_number": 84, "usage_type": "call"}, {"api_name": "paddle.inference", "line_number": 84, "usage_type": "attribute"}, {"api_name": "paddlenlp.transformers.BertTokenizer.from_pretrained", "line_number": 91, "usage_type": "call"}, {"api_name": "paddlenlp.transformers.BertTokenizer", "line_number": 91, "usage_type": "name"}, {"api_name": "os.path.dirname", "line_number": 92, "usage_type": "call"}, {"api_name": "os.path", "line_number": 92, "usage_type": "attribute"}, {"api_name": "paddlenlp.data.Tuple", "line_number": 107, "usage_type": "call"}, {"api_name": "paddlenlp.data.Pad", "line_number": 108, "usage_type": "call"}, {"api_name": "paddlenlp.data.Pad", "line_number": 109, "usage_type": "call"}, {"api_name": "scipy.special.softmax", "line_number": 126, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 127, "usage_type": "call"}]} +{"seq_id": "536991304", "text": "import math, sys\nfrom lux.game import Game\nfrom lux.game_map import Cell, RESOURCE_TYPES\nfrom lux.constants import Constants\nfrom lux.game_constants import GAME_CONSTANTS\nfrom lux import annotate\nimport logging\n\nlogging.basicConfig(filename='agent.log', level=logging.INFO)\n\nDIRECTIONS = Constants.DIRECTIONS\ngame_state = None\nTARGET_LOCS = {}\nUNIT_LOCATIONS = {}\n\n\ndef get_resource_cells(m):\n width, height = m.width, m.height\n resource_tiles: list[Cell] = []\n for y in range(height):\n for x in range(width):\n cell = m.get_cell(x, y)\n if cell.has_resource():\n resource_tiles.append(cell)\n return resource_tiles\n\ndef get_adjacent_cells(cell, m):\n adj_cells: list[Cell] = []\n x, y = cell.pos.x, cell.pos.y\n if x > 0:\n adj_cells.append(m.get_cell(x-1,y))\n if x < (m.width-1):\n adj_cells.append(m.get_cell(x+1,y))\n if y > 0:\n adj_cells.append(m.get_cell(x,y-1))\n if y < (m.width-1):\n adj_cells.append(m.get_cell(x,y+1))\n return adj_cells\n\ndef get_cell_value(cell, p):\n if not cell.has_resource():\n return 0\n if cell.resource.type == Constants.RESOURCE_TYPES.COAL:\n if not p.researched_coal():\n return 0\n else:\n return 50\n if cell.resource.type == Constants.RESOURCE_TYPES.URANIUM:\n if not p.researched_uranium():\n return 0\n else:\n return 80\n return 20\n\ndef get_energy(unit):\n return unit.cargo.wood + unit.cargo.coal * 10 + unit.cargo.uranium * 40\n\ndef get_map_values(m, p):\n width, height = m.width, m.height\n d = {}\n for y in range(height):\n for x in range(width):\n adj = get_adjacent_cells(m.get_cell(x,y), m)\n adj.append(m.get_cell(x,y))\n d[(x,y)] = sum([get_cell_value(cell, p) for cell in adj])\n return d\n\ndef cities_powered(p, day_cycle):\n status = True\n if day_cycle > 15:\n for k, city in p.cities.items():\n if (3*len(city.citytiles)) > p.city_tile_count:\n if city.fuel < math.ceil(city.get_light_upkeep() + 180):\n status = False \n break\n return status\n\ndef is_empty(c):\n if c.has_resource():\n return False\n if c.citytile is None:\n return True\n return False\n\ndef get_coords(c):\n return (c.pos.x, c.pos.y)\n\ndef get_expansion_sites(city, m):\n borders_dup: list[Cell] = []\n for ct in city.citytiles:\n borders_dup.append(get_adjacent_cells(ct, m))\n if type(borders_dup[0]) is list:\n borders_dup = [c for sublist in borders_dup for c in sublist]\n borders = [c for c in set(borders_dup) if is_empty(c)]\n return borders\n\ndef take_step(u, target, m, allow_city, opp_locs, my_cities):\n global UNIT_LOCATIONS\n target = m.get_cell(target[0], target[1])\n \n if u.pos == target.pos:\n return None\n occ_loc = [UNIT_LOCATIONS[id] for id in UNIT_LOCATIONS.keys()]\n occ_loc = [coord for coord in occ_loc if m.get_cell(coord[0], coord[1]).citytile is None]\n occ_loc.append(opp_locs)\n if not allow_city:\n occ_loc.append(my_cities)\n #logging.info(f\"{u.id} position: {(u.pos.x, u.pos.y)} target: {(target.pos.x, target.pos.y)}\")\n #logging.info(f\"cannot step to: {occ_loc}\")\n if u.pos.y > target.pos.y:\n if (u.pos.x, u.pos.y - 1) not in occ_loc:\n UNIT_LOCATIONS[u.id] = (u.pos.x, u.pos.y - 1)\n return u.move(DIRECTIONS.NORTH)\n elif u.pos.y < target.pos.y:\n if (u.pos.x, u.pos.y + 1) not in occ_loc:\n UNIT_LOCATIONS[u.id] = (u.pos.x, u.pos.y + 1)\n return u.move(DIRECTIONS.SOUTH)\n if u.pos.x > target.pos.x:\n if (u.pos.x - 1, u.pos.y) not in occ_loc:\n UNIT_LOCATIONS[u.id] = (u.pos.x - 1, u.pos.y)\n return u.move(DIRECTIONS.WEST)\n \n if u.pos.x < target.pos.x:\n if (u.pos.x + 1, u.pos.y) not in occ_loc:\n UNIT_LOCATIONS[u.id] = (u.pos.x + 1, u.pos.y)\n return u.move(DIRECTIONS.EAST)\n return None\n\ndef get_gather_target(u, p, m, resource_tiles, allow_city, values):\n best_val = 0.0\n best_tile = None\n taken_targets = [TARGET_LOCS[id] for id in TARGET_LOCS.keys() if id != u.id]\n for tile in values.keys():\n if tile in taken_targets:\n continue\n if not allow_city:\n if m.get_cell(tile[0], tile[1]).citytile is not None:\n continue\n dist = abs(u.pos.x - tile[0]) + abs(u.pos.y - tile[1])\n if dist < 15:\n val = 1.0 * values[tile] / math.log(dist + 2)\n if val > best_val:\n best_val = val\n best_tile = m.get_cell(tile[0], tile[1])\n if best_tile is not None:\n return get_coords(best_tile)\n return None\n\ndef find_home(u, p, m):\n if m.get_cell_by_pos(u.pos).citytile is not None:\n return None\n closest_dist = math.inf\n closest_city_tile = None\n for k, city in p.cities.items():\n for city_tile in city.citytiles:\n dist = city_tile.pos.distance_to(u.pos)\n if dist < closest_dist:\n closest_dist = dist\n closest_city_tile = city_tile\n return get_coords(closest_city_tile)\n\ndef get_build_loc(u, p, m):\n target_loc = None\n target_dist = math.inf\n if p.city_tile_count == 0:\n if is_empty(m.get_cell_by_pos(u.pos)):\n return u.build_city()\n else:\n adj = get_adjacent_cells(m.get_cell_by_pos(u.pos), m)\n adj = [x for x in adj if is_empty(x)]\n if len(adj) > 0:\n return u.move(u.pos.direction_to(adj[0].pos))\n else:\n return u.move(DIRECTIONS.SOUTH)\n for k, city in p.cities.items():\n expansions = get_expansion_sites(city, m)\n taken_targets = [TARGET_LOCS[id] for id in TARGET_LOCS.keys() if id != u.id]\n expansions = [x for x in expansions if get_coords(x) not in taken_targets]\n if len(expansions) > 0:\n if u.pos in [c.pos for c in expansions]:\n return u.build_city() \n else:\n for site in expansions:\n dist = site.pos.distance_to(u.pos)\n if dist < target_dist:\n target_dist = dist\n target_loc = site\n if target_dist > 5 and u.can_build(m):\n return u.build_city() \n if target_loc is not None: \n return get_coords(target_loc)\n return None\n\ndef agent(observation, configuration):\n global TARGET_LOCS\n global UNIT_LOCATIONS\n global game_state\n\n ### Do not edit ###\n if observation[\"step\"] == 0:\n game_state = Game()\n game_state._initialize(observation[\"updates\"])\n game_state._update(observation[\"updates\"][2:])\n game_state.id = observation.player\n else:\n game_state._update(observation[\"updates\"])\n\n starting_locs = {}\n \n actions = []\n\n ### AI Code goes down here! ### \n player = game_state.players[observation.player]\n opponent = game_state.players[(observation.player + 1) % 2]\n width, height = game_state.map.width, game_state.map.height\n\n resource_tiles = get_resource_cells(game_state.map)\n unit_count = len(player.units)\n map_values = get_map_values(game_state.map, player)\n day_cycle = game_state.turn % 40\n allow_cities = {}\n ids_to_skip = []\n to_build = []\n\n # we iterate over all our units and do something with them\n for unit in player.units:\n UNIT_LOCATIONS[unit.id] = (unit.pos.x, unit.pos.y)\n starting_locs[unit.id] = (unit.pos.x, unit.pos.y)\n allow_cities[unit.id] = True\n threshold = 0\n target = None\n allow_city = True\n if unit.is_worker():\n if int(unit.id[2:]) % 3 == 0:\n if day_cycle > 27:\n threshold = 4 * min(10, 40 - day_cycle)\n if unit.cargo.wood < threshold:\n target = find_home(unit, player, game_state.map)\n elif not unit.can_act():\n continue\n elif unit.get_cargo_space_left() > 0 and (unit.get_cargo_space_left() <= 60 or day_cycle < 30):\n if unit_count > 2:\n target = get_gather_target(unit, player, game_state.map, resource_tiles, True, map_values)\n else:\n target = get_gather_target(unit, player, game_state.map, resource_tiles, False, map_values)\n elif unit.get_cargo_space_left() == 0 and (cities_powered(player, day_cycle) or player.city_tile_count == 0):\n target = get_build_loc(unit, player, game_state.map) \n to_build.append(unit.id)\n else:\n target = find_home(unit, player, game_state.map)\n if type(target) != str and target is not None:\n TARGET_LOCS[unit.id] = target\n else:\n actions.append(target)\n ids_to_skip.append(unit.id)\n \n\n opp_cities = []\n for k, city in opponent.cities.items():\n for ct in city.citytiles:\n opp_cities.append(get_coords(ct))\n my_cities = []\n for k, city in player.cities.items():\n for ct in city.citytiles:\n my_cities.append(get_coords(ct))\n moves_happened = True\n while moves_happened:\n moves_happened = False\n to_iter = [x for x in player.units if x.id not in ids_to_skip]\n for unit in to_iter:\n if unit.can_act():\n if starting_locs[unit.id][0] == UNIT_LOCATIONS[unit.id][0] and starting_locs[unit.id][1] == UNIT_LOCATIONS[unit.id][1]:\n if unit.id in TARGET_LOCS.keys():\n if unit.id in to_build:\n move_cmd = take_step(unit, TARGET_LOCS[unit.id], game_state.map, False, opp_cities, my_cities)\n else:\n move_cmd = take_step(unit, TARGET_LOCS[unit.id], game_state.map, True, opp_cities, my_cities)\n if move_cmd is not None:\n actions.append(move_cmd)\n moves_happened = True\n\n can_build = player.city_tile_count - unit_count\n for k, city in player.cities.items():\n for ct in city.citytiles:\n if ct.can_act():\n if can_build > 0:\n actions.append(ct.build_worker())\n can_build = can_build - 1\n else:\n actions.append(ct.research())\n \n\n # you can add debug annotations using the functions in the annotate object\n # actions.append(annotate.circle(0, 0))\n actions = [x for x in actions if x is not None]\n #logging.info(f\"\\n\\n\")\n #logging.info(f\"turn {game_state.turn}: locations {[(id, UNIT_LOCATIONS[id]) for id in UNIT_LOCATIONS.keys()]}\")\n return actions\n", "sub_path": "aggro_bot/agent.py", "file_name": "agent.py", "file_ext": "py", "file_size_in_byte": 10864, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "lux.constants.Constants.DIRECTIONS", "line_number": 11, "usage_type": "attribute"}, {"api_name": "lux.constants.Constants", "line_number": 11, "usage_type": "name"}, {"api_name": "lux.game_map.Cell", "line_number": 19, "usage_type": "name"}, {"api_name": "lux.game_map.Cell", "line_number": 28, "usage_type": "name"}, {"api_name": "lux.constants.Constants.RESOURCE_TYPES", "line_number": 43, "usage_type": "attribute"}, {"api_name": "lux.constants.Constants", "line_number": 43, "usage_type": "name"}, {"api_name": "lux.constants.Constants.RESOURCE_TYPES", "line_number": 48, "usage_type": "attribute"}, {"api_name": "lux.constants.Constants", "line_number": 48, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 73, "usage_type": "call"}, {"api_name": "lux.game_map.Cell", "line_number": 89, "usage_type": "name"}, {"api_name": "math.log", "line_number": 141, "usage_type": "call"}, {"api_name": "math.inf", "line_number": 152, "usage_type": "attribute"}, {"api_name": "math.inf", "line_number": 164, "usage_type": "attribute"}, {"api_name": "lux.game.Game", "line_number": 201, "usage_type": "call"}]} +{"seq_id": "285737648", "text": "import numpy as np\nfrom numpy import ndarray\nimport scipy.stats\n\nimport data_gen_funcs\nimport data_gen_funcs_bernoulli\nimport data_gen_funcs_multinomial\nfrom data_feature_range import FeatureRange\nfrom common import get_normal_dist_entropy\nfrom dataset import Dataset\nfrom support_sim_settings import SupportSimSettingsContinuous\n\nclass DataGenerator:\n \"\"\"\n Simulation engine\n \"\"\"\n def __init__(self,\n sim_func_form: str,\n sim_func_name: str,\n num_p: int,\n num_classes: int =1,\n noise_sd: float =0,\n std_dev_x: float =1,\n max_x: float =1):\n self.num_p = num_p\n self.std_dev_x = std_dev_x\n self.num_classes = num_classes\n self.max_x = max_x\n self.min_x = -max_x\n self.noise_sd = noise_sd\n self.sim_func_form = sim_func_form\n if sim_func_form == \"gaussian\":\n self.mu_func = getattr(data_gen_funcs, sim_func_name + \"_mu\")\n self.raw_sigma_func = getattr(data_gen_funcs, sim_func_name + \"_sigma\")\n elif sim_func_form == \"bernoulli\":\n self.mu_func = getattr(data_gen_funcs_bernoulli, sim_func_name + \"_mu\")\n elif sim_func_form == \"multinomial\":\n self.mu_func = getattr(data_gen_funcs_multinomial, sim_func_name + \"_mu\")\n else:\n print(sim_func_form)\n raise ValueError(\"huh?\")\n\n def sigma_func(self, xs: ndarray):\n \"\"\"\n @return sigma when Y|X is gaussian\n \"\"\"\n if self.sim_func_form == \"gaussian\":\n return self.noise_sd * self.raw_sigma_func(xs)\n elif self.sim_func_form == \"bernoulli\":\n raise ValueError(\"sure?\")\n\n def entropy_func(self, xs: ndarray):\n \"\"\"\n @return sigma when Y|X is gaussian\n \"\"\"\n if self.sim_func_form == \"gaussian\":\n sigma = self.noise_sd * self.raw_sigma_func(xs)\n return get_normal_dist_entropy(sigma)\n elif self.sim_func_form == \"bernoulli\":\n p = self.mu_func(xs)\n return -p * np.log(p) - (1 - p) * np.log(1 - p)\n else:\n p = self.mu_func(xs)\n return np.sum(-p * np.log(p), axis=1)\n\n def create_data(self, num_obs: int, seed:int = None):\n \"\"\"\n @param num_obs: number of observations\n @param seed: if given, set the seed before generating data\n\n @param tuple with Dataset, SupportSimSettingsContinuous\n \"\"\"\n if seed is not None:\n np.random.seed(seed)\n dataset = self._create_data(num_obs)\n support_sim_settings = self._create_support_sim_settings()\n return dataset, support_sim_settings\n\n def create_data_given_x(self, xs: ndarray):\n \"\"\"\n For the given Xs, generate responses and dataset\n regression-type data only\n @return Dataset\n \"\"\"\n size_n = xs.shape[0]\n mu_true = self.mu_func(xs)\n if len(mu_true.shape) == 1:\n mu_true = np.reshape(mu_true, (size_n, 1))\n if self.sim_func_form == \"gaussian\":\n sigma_true = np.reshape(self.sigma_func(xs), (size_n, 1))\n\n true_distribution = scipy.stats.norm(mu_true, sigma_true)\n y = true_distribution.rvs(size=mu_true.shape)\n true_prob = true_distribution.pdf(y)\n elif self.sim_func_form == \"bernoulli\":\n true_distribution = scipy.stats.binom(n=1, p=mu_true)\n y = true_distribution.rvs(size=mu_true.shape)\n true_prob = true_distribution.pmf(y)\n elif self.sim_func_form == \"multinomial\":\n # We have to do per row because multinomial is not nice and doesn't take in\n # 2D probability matrices\n all_ys = []\n all_probs = []\n for i in range(mu_true.shape[0]):\n mu_row = mu_true[i,:]\n true_distribution = scipy.stats.multinomial(n=1, p=mu_row)\n y = true_distribution.rvs(size=1)\n all_ys.append(y)\n true_prob = true_distribution.pmf(y)\n all_probs.append(true_prob)\n y = np.vstack(all_ys)\n true_prob = np.vstack(all_probs)\n\n # Print entropy of Y|X for fun\n entropies = self.entropy_func(xs)\n print(\"ENTROPY\", np.mean(entropies), np.sqrt(np.var(entropies)), np.max(entropies))\n print((entropies > 0.8).mean())\n print((entropies > 1.2).mean())\n print((entropies > 1.6).mean())\n print((entropies > 2.0).mean())\n\n return Dataset(xs, y, true_pdf=true_prob, num_classes=self.num_classes)\n\n def generate_x(self, size_n: int, buffer_factor: int = 10):\n \"\"\"\n Generates x from a gaussian distribution\n \"\"\"\n xs = np.random.randn(size_n * buffer_factor, self.num_p) * self.std_dev_x\n # Only keep the ones within the support\n in_support = np.sum(\n (xs >= self.min_x) * (xs <= self.max_x),\n axis=1) == self.num_p\n xs_in_support = xs[in_support,:]\n assert np.sum(in_support) >= size_n\n return xs_in_support[:size_n, :]\n\n def get_x_pdf(self, xs: ndarray):\n \"\"\"\n @return pdf of X\n \"\"\"\n all_pdfs = 1\n for i in range(xs.shape[1]):\n all_pdfs *= scipy.stats.norm.pdf(xs[:,i], loc=0, scale=self.std_dev_x)\n return all_pdfs\n\n def get_prediction_interval(self, xs: ndarray, alpha: float):\n mus = self.mu_func(xs).reshape((-1,1))\n sigmas = self.sigma_func(xs).reshape((-1,1))\n z_factor = scipy.stats.norm.ppf(1 - alpha/2)\n lower = mus - z_factor * sigmas\n upper = mus + z_factor * sigmas\n return np.concatenate([lower, upper], axis=1)\n\n def _create_data(self, size_n: int):\n \"\"\"\n regression-type data only\n @return Dataset\n \"\"\"\n # Generate some x's from a gaussian distribution\n data_gen_xs = self.generate_x(size_n)\n return self.create_data_given_x(data_gen_xs)\n\n def _create_support_sim_settings(self):\n return SupportSimSettingsContinuous(self.num_p, self.min_x, self.max_x)\n", "sub_path": "data_generator.py", "file_name": "data_generator.py", "file_ext": "py", "file_size_in_byte": 6128, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "numpy.ndarray", "line_number": 43, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 52, "usage_type": "name"}, {"api_name": "common.get_normal_dist_entropy", "line_number": 58, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 61, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 64, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 74, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 74, "usage_type": "attribute"}, {"api_name": "numpy.ndarray", "line_number": 79, "usage_type": "name"}, {"api_name": "numpy.reshape", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 90, "usage_type": "call"}, {"api_name": "scipy.stats.stats.norm", "line_number": 92, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 92, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 92, "usage_type": "name"}, {"api_name": "scipy.stats.stats.binom", "line_number": 96, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 96, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 96, "usage_type": "name"}, {"api_name": "scipy.stats.stats.multinomial", "line_number": 106, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 106, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 106, "usage_type": "name"}, {"api_name": "numpy.vstack", "line_number": 111, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 112, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.var", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 116, "usage_type": "call"}, {"api_name": "dataset.Dataset", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 128, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 130, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 134, "usage_type": "call"}, {"api_name": "numpy.ndarray", "line_number": 137, "usage_type": "name"}, {"api_name": "scipy.stats.stats.norm.pdf", "line_number": 143, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 143, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 143, "usage_type": "name"}, {"api_name": "numpy.ndarray", "line_number": 146, "usage_type": "name"}, {"api_name": "scipy.stats.stats.norm.ppf", "line_number": 149, "usage_type": "call"}, {"api_name": "scipy.stats.stats", "line_number": 149, "usage_type": "attribute"}, {"api_name": "scipy.stats", "line_number": 149, "usage_type": "name"}, {"api_name": "numpy.concatenate", "line_number": 152, "usage_type": "call"}, {"api_name": "support_sim_settings.SupportSimSettingsContinuous", "line_number": 164, "usage_type": "call"}]} +{"seq_id": "510003454", "text": "import jieba\nimport xlrd\nimport xlwt\n\ndata = xlrd.open_workbook('PolicyData.xls')\ntable = data.sheet_by_name(u'sheet1')\nncols = table.nrows\nfor i in range(ncols-1):\n txt = table.cell(i+1,0).value\n words = jieba.cut(txt,cut_all=True)\n print(','.join(words))\n \n", "sub_path": "slip_data.py", "file_name": "slip_data.py", "file_ext": "py", "file_size_in_byte": 275, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "xlrd.open_workbook", "line_number": 5, "usage_type": "call"}, {"api_name": "jieba.cut", "line_number": 10, "usage_type": "call"}]} +{"seq_id": "126084602", "text": "import numpy as np\nimport cv2 as cv\nimport glob\nimport os\npath = r'G:\\data\\train_results\\um_road_000000.png'\ndir_path = r'G:\\data\\train_results\\*.png'\nsave_dir_path = r'G:\\data\\train_label_results'\nif not os.path.exists(save_dir_path):\n os.makedirs(save_dir_path)\n\nfiles = np.array(sorted(glob.glob(dir_path)))\n\nprint(files)\nfor image in files[:56]:\n image_name = image.split('\\\\')[-1]\n image = cv.imread(path, cv.IMREAD_GRAYSCALE)\n #cv.imshow('原图', image)\n THRESH = 100\n ret, binary = cv.threshold(image, THRESH, 1, cv.THRESH_BINARY|cv.THRESH_TRIANGLE)\n #cv.imshow('binary', binary)\n cv.imwrite(save_dir_path + '/' + image_name, binary)\n #cv.waitKey(0)\n\n", "sub_path": "v5/convert_gt_2_label.py", "file_name": "convert_gt_2_label.py", "file_ext": "py", "file_size_in_byte": 686, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.exists", "line_number": 8, "usage_type": "call"}, {"api_name": "os.path", "line_number": 8, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 11, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 11, "usage_type": "call"}, {"api_name": "cv2.imread", "line_number": 16, "usage_type": "call"}, {"api_name": "cv2.IMREAD_GRAYSCALE", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.threshold", "line_number": 19, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.THRESH_TRIANGLE", "line_number": 19, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 21, "usage_type": "call"}]} +{"seq_id": "143714761", "text": "# Copyright 2002 by Tarjei Mikkelsen. All rights reserved.\n# Revisions copyright 2018 by Maximilian Greil. All rights reserved.\n#\n# This file is part of the Biopython distribution and governed by your\n# choice of the \"Biopython License Agreement\" or the \"BSD 3-Clause License\".\n# Please see the LICENSE file that should have been included as part of this\n# package.\n\n\"\"\"get/set abstraction for graph representation.\"\"\"\n\nfrom functools import reduce\n\n\nclass Graph:\n \"\"\"A directed graph abstraction with labeled edges.\"\"\"\n\n def __init__(self, nodes=()):\n \"\"\"Initialize a new Graph object.\"\"\"\n self._adjacency_list = {} # maps parent -> set of child objects\n for n in nodes:\n self._adjacency_list[n] = set()\n self._label_map = {} # maps label -> set of (parent, child) pairs\n self._edge_map = {} # maps (parent, child) pair -> label\n\n def __eq__(self, g):\n \"\"\"Return true if g is equal to this graph.\"\"\"\n return (\n isinstance(g, Graph)\n and self._adjacency_list == g._adjacency_list\n and self._label_map == g._label_map\n and self._edge_map == g._edge_map\n )\n\n def __repr__(self):\n \"\"\"Return a unique string representation of this graph.\"\"\"\n s = \"\"\n\n def __str__(self):\n \"\"\"Return a concise string description of this graph.\"\"\"\n nodenum = len(self._adjacency_list)\n edgenum = reduce(\n lambda x, y: x + y, [len(v) for v in self._adjacency_list.values()]\n )\n labelnum = len(self._label_map)\n return (\n \"\"\n )\n\n def add_node(self, node):\n \"\"\"Add a node to this graph.\"\"\"\n if node not in self._adjacency_list:\n self._adjacency_list[node] = set()\n\n def add_edge(self, source, to, label=None):\n \"\"\"Add an edge to this graph.\"\"\"\n if source not in self._adjacency_list:\n raise ValueError(\"Unknown node: \" + str(source))\n if to not in self._adjacency_list:\n raise ValueError(\"Unknown node: \" + str(to))\n if (source, to) in self._edge_map:\n raise ValueError(str(source) + \" -> \" + str(to) + \" exists\")\n self._adjacency_list[source].add(to)\n if label not in self._label_map:\n self._label_map[label] = set()\n self._label_map[label].add((source, to))\n self._edge_map[(source, to)] = label\n\n def child_edges(self, parent):\n \"\"\"Return a list of (child, label) pairs for parent.\"\"\"\n if parent not in self._adjacency_list:\n raise ValueError(\"Unknown node: \" + str(parent))\n return [\n (x, self._edge_map[(parent, x)])\n for x in sorted(self._adjacency_list[parent])\n ]\n\n def children(self, parent):\n \"\"\"Return a list of unique children for parent.\"\"\"\n return sorted(self._adjacency_list[parent])\n\n def edges(self, label):\n \"\"\"Return a list of all the edges with this label.\"\"\"\n if label not in self._label_map:\n raise ValueError(\"Unknown label: \" + str(label))\n return sorted(self._label_map[label])\n\n def labels(self):\n \"\"\"Return a list of all the edge labels in this graph.\"\"\"\n return sorted(self._label_map.keys())\n\n def nodes(self):\n \"\"\"Return a list of the nodes in this graph.\"\"\"\n return list(self._adjacency_list.keys())\n\n def parent_edges(self, child):\n \"\"\"Return a list of (parent, label) pairs for child.\"\"\"\n if child not in self._adjacency_list:\n raise ValueError(\"Unknown node: \" + str(child))\n parents = []\n for parent, children in self._adjacency_list.items():\n for x in children:\n if x == child:\n parents.append((parent, self._edge_map[(parent, child)]))\n return sorted(parents)\n\n def parents(self, child):\n \"\"\"Return a list of unique parents for child.\"\"\"\n return sorted({x[0] for x in self.parent_edges(child)})\n\n def remove_node(self, node):\n \"\"\"Remove node and all edges connected to it.\"\"\"\n if node not in self._adjacency_list:\n raise ValueError(\"Unknown node: \" + str(node))\n # remove node (and all out-edges) from adjacency list\n del self._adjacency_list[node]\n # remove all in-edges from adjacency list\n for n in self._adjacency_list.keys():\n self._adjacency_list[n] = {x for x in self._adjacency_list[n] if x != node}\n # remove all referring pairs in label map\n for label in list(self._label_map.keys()): # we're editing this!\n lm = {\n x for x in self._label_map[label] if (x[0] != node) and (x[1] != node)\n }\n # remove the entry completely if the label is now unused\n if lm:\n self._label_map[label] = lm\n else:\n del self._label_map[label]\n # remove all referring entries in edge map\n for edge in list(self._edge_map.keys()): # we're editing this!\n if edge[0] == node or edge[1] == node:\n del self._edge_map[edge]\n\n def remove_edge(self, parent, child, label):\n \"\"\"Remove edge (NOT IMPLEMENTED).\"\"\"\n # hm , this is a multigraph - how should this be implemented?\n raise NotImplementedError(\"remove_edge is not yet implemented\")\n", "sub_path": "env/Lib/site-packages/Bio/Pathway/Rep/Graph.py", "file_name": "Graph.py", "file_ext": "py", "file_size_in_byte": 5869, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "functools.reduce", "line_number": 47, "usage_type": "call"}]} +{"seq_id": "239726830", "text": "# -*- coding: utf-8 -*--\nfrom sqlalchemy.sql.elements import and_\nfrom pyramid.view import view_config\nimport pyramid.httpexceptions as exc\nfrom pyramid.response import Response\n\n\nfrom infolica.exceptions.custom_error import CustomError\nfrom infolica.models.constant import Constant\nfrom infolica.models.models import Facture, Numero, TableauEmoluments, VAffaire, VNumerosAffaires\nfrom infolica.models.models import EmolumentAffaire, Emolument, EmolumentAffaireRepartition\nfrom infolica.scripts.utils import Utils\nfrom infolica.scripts.authentication import check_connected\nimport json\nimport datetime\nimport requests\n\n###########################################################\n# EMOLUMENTS\n###########################################################\n\n\n@view_config(route_name='tableau_emoluments', request_method='GET', renderer='json')\ndef tableau_emoluments_view(request):\n \"\"\"\n Return table of emoluments \n \"\"\"\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n query = request.dbsession.query(TableauEmoluments).order_by(TableauEmoluments.id).all()\n return Utils.serialize_many(query)\n\n\n@view_config(route_name='emolument_affaire', request_method='GET', renderer='json')\ndef emolument_affaire_view(request):\n \"\"\"\n Return emolument_affaire\n \"\"\"\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n affaire_id = request.params['affaire_id'] if 'affaire_id' in request.params else None\n\n query = request.dbsession.query(EmolumentAffaire)\n \n if not affaire_id is None:\n query = query.filter(EmolumentAffaire.affaire_id == affaire_id)\n \n emolument_affaire = query.all()\n \n \n # Récupérer les données des bâtiments\n result = []\n for emolument_affaire_i in emolument_affaire:\n query_bat = request.dbsession.query(\n Emolument.batiment,\n Emolument.batiment_f\n ).filter(\n Emolument.emolument_affaire_id == emolument_affaire_i.id\n ).filter(\n Emolument.batiment > 0\n ).group_by(\n Emolument.batiment,\n Emolument.batiment_f\n ).order_by(\n Emolument.batiment.asc() # Really important with respect to implementation of loading form_detail_batiment in front !! \n ).all()\n \n batiment_f = [y for _, y in query_bat]\n numeros = []\n numeros_id = []\n if emolument_affaire_i.numeros_id and len(emolument_affaire_i.numeros_id) > 0:\n numeros_id = emolument_affaire_i.numeros_id\n numeros = Utils.serialize_many(request.dbsession.query(VNumerosAffaires).filter(\n and_(\n VNumerosAffaires.numero_id.in_(tuple(emolument_affaire_i.numeros_id)),\n VNumerosAffaires.affaire_id == emolument_affaire_i.affaire_id\n )\n ).all())\n \n # Récupérer les liens sur la facture (répartition des montants)\n facture_repartition = Utils.serialize_many(\n request.dbsession.query(EmolumentAffaireRepartition).filter(\n EmolumentAffaireRepartition.emolument_affaire_id == emolument_affaire_i.id\n ).all()\n )\n\n result.append(\n Utils._params(\n id = emolument_affaire_i.id,\n affaire_id = emolument_affaire_i.affaire_id,\n pente_pc = emolument_affaire_i.pente_pc,\n diff_visibilite_pc = emolument_affaire_i.diff_visibilite_pc,\n trafic_pc = emolument_affaire_i.trafic_pc,\n zi = emolument_affaire_i.zi ,\n indice_application = emolument_affaire_i.indice_application,\n tva_pc = emolument_affaire_i.tva_pc,\n remarque = emolument_affaire_i.remarque,\n nb_batiments = len(batiment_f),\n batiment_f = batiment_f,\n numeros_id = numeros_id,\n numeros = numeros,\n facture_type_id = emolument_affaire_i.facture_type_id,\n utilise = emolument_affaire_i.utilise,\n facture_repartition = facture_repartition\n )\n )\n \n return result\n\n\n@view_config(route_name='emolument', request_method='GET', renderer='json')\ndef emolument_view(request):\n \"\"\"\n Return emoluments of emoluments_affaire \n \"\"\"\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n # get TableauEmoluments\n tableauEmoluments = request.dbsession.query(\n TableauEmoluments\n ).filter(\n TableauEmoluments.date_sortie == None\n ).all()\n\n tableauEmoluments = Utils.serialize_many(tableauEmoluments)\n\n # get emolumentsAffaire saved with emolument_affaire_id\n emolument_affaire_id = request.params['emolument_affaire_id'] if 'emolument_affaire_id' in request.params else None\n\n emolumentsAffaire = request.dbsession.query(\n Emolument\n ).filter(\n Emolument.emolument_affaire_id == emolument_affaire_id\n ).all()\n\n emolumentsAffaire = Utils.serialize_many(emolumentsAffaire)\n\n # check if emolumentAffaire is used\n emolumentAffaireUsed = request.dbsession.query(\n EmolumentAffaire.utilise\n ).filter(\n EmolumentAffaire.id == emolument_affaire_id\n ).scalar()\n\n for ea in emolumentsAffaire:\n for te in tableauEmoluments:\n if ea['tableau_emolument_id'] == te['id'] and not emolumentAffaireUsed and not te['montant'] == 0:\n ea['prix_unitaire'] = te['montant']\n ea['montant'] = ea['nombre'] * ea['prix_unitaire'] * ea['batiment_f']\n\n return emolumentsAffaire\n\n\n@view_config(route_name='emolument_affaire', request_method='POST', renderer='json')\ndef emolument_affaire_new_view(request):\n \"\"\"\n Add new emolument_affaire\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n params = request.params\n data = json.loads(params[\"data\"])\n\n record = EmolumentAffaire()\n record = Utils.set_model_record(record, data)\n\n request.dbsession.add(record)\n request.dbsession.flush()\n\n return {\"emolument_affaire_id\": record.id}\n\n\n@view_config(route_name='emolument', request_method='POST', renderer='json')\ndef emolument_new_view(request):\n \"\"\"\n Add new emolument\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n params = request.params\n data = json.loads(params['data'])\n emolument_affaire_id = params['emolument_affaire_id']\n\n for batiment_i in data:\n for emolument_i in batiment_i:\n if float(batiment_i[emolument_i]['montant']) > 0 and float(batiment_i[emolument_i]['nombre']) > 0:\n params = Utils._params(\n emolument_affaire_id=int(emolument_affaire_id),\n tableau_emolument_id=int(batiment_i[emolument_i]['tableau_emolument_id']),\n position=batiment_i[emolument_i]['nom'],\n prix_unitaire=float(batiment_i[emolument_i]['prix_unitaire']),\n nombre=float(batiment_i[emolument_i]['nombre']),\n batiment=int(batiment_i[emolument_i]['batiment']),\n batiment_f=float(batiment_i[emolument_i]['batiment_f']),\n montant=float(batiment_i[emolument_i]['montant'])\n )\n\n record = Emolument()\n record = Utils.set_model_record(record, params)\n\n request.dbsession.add(record)\n\n return Utils.get_data_save_response(Constant.SUCCESS_SAVE.format(Emolument.__tablename__))\n\n\n@view_config(route_name='emolument_affaire', request_method='PUT', renderer='json')\ndef update_emolument_affaire_view(request):\n \"\"\"\n Update emolument_affaire\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n params = request.params\n data = json.loads(params[\"data\"])\n\n record_id = request.params['emolument_affaire_id'] if 'emolument_affaire_id' in request.params else None \n\n record = request.dbsession.query(EmolumentAffaire).filter(\n EmolumentAffaire.id == record_id\n ).first()\n\n record = Utils.set_model_record(record, data)\n\n return Utils.get_data_save_response(Constant.SUCCESS_SAVE.format(EmolumentAffaire.__tablename__))\n\n\n@view_config(route_name='emolument_affaire_freeze', request_method='PUT', renderer='json')\ndef update_emolument_affaire_freeze_view(request):\n \"\"\"\n Freeze emolument_affaire\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n params = request.params\n \n record_id = params['emolument_affaire_id'] if 'emolument_affaire_id' in params else None\n\n record = request.dbsession.query(EmolumentAffaire).filter(\n EmolumentAffaire.id == record_id\n ).first()\n\n record = Utils.set_model_record(record, params)\n\n return Utils.get_data_save_response(Constant.SUCCESS_SAVE.format(EmolumentAffaire.__tablename__))\n\n\n@view_config(route_name='emolument', request_method='PUT', renderer='json')\ndef update_emolument_view(request):\n \"\"\"\n Update emolument\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n params = request.params\n data = json.loads(params['data'])\n emolument_affaire_id = params['emolument_affaire_id']\n\n # Query existing data\n query = request.dbsession.query(Emolument).filter(\n Emolument.emolument_affaire_id == emolument_affaire_id\n )\n\n emoluments = query.all()\n\n for batiment_i in data:\n for emolument_i in batiment_i:\n record = None\n for index, item in enumerate(emoluments):\n if (item.batiment == batiment_i[emolument_i]['batiment'] and \n item.tableau_emolument_id == batiment_i[emolument_i]['tableau_emolument_id']):\n record = emoluments.pop(index)\n break\n \n\n if not record is None:\n # comparer les valeurs enregistrées\n if (not float(record.montant) == float(batiment_i[emolument_i]['montant']) \\\n or not record.position == batiment_i[emolument_i]['nom'] \\\n or not float(record.prix_unitaire) == float(batiment_i[emolument_i]['prix_unitaire']) \\\n or not float(record.nombre) == float(batiment_i[emolument_i]['nombre']) \\\n or not float(record.batiment_f) == float(batiment_i[emolument_i]['batiment_f'])):\n\n # Mettre à jour les données si le nouveau montant n'est pas nul\n if float(batiment_i[emolument_i]['montant']) > 0:\n params = Utils._params(\n position=batiment_i[emolument_i]['nom'],\n prix_unitaire=float(batiment_i[emolument_i]['prix_unitaire']),\n nombre=float(batiment_i[emolument_i]['nombre']),\n batiment_f=float(batiment_i[emolument_i]['batiment_f']),\n montant=float(batiment_i[emolument_i]['montant'])\n )\n\n record = Utils.set_model_record(record, params)\n else:\n # supprimer l'émolument\n request.dbsession.delete(record)\n \n else:\n if float(batiment_i[emolument_i]['montant']) > 0 and float(batiment_i[emolument_i]['nombre']) > 0:\n params = Utils._params(\n emolument_affaire_id=int(emolument_affaire_id),\n tableau_emolument_id=int(batiment_i[emolument_i]['tableau_emolument_id']),\n position=batiment_i[emolument_i]['nom'],\n prix_unitaire=float(batiment_i[emolument_i]['prix_unitaire']),\n nombre=float(batiment_i[emolument_i]['nombre']),\n batiment=int(batiment_i[emolument_i]['batiment']),\n batiment_f=float(batiment_i[emolument_i]['batiment_f']),\n montant=float(batiment_i[emolument_i]['montant'])\n )\n\n record = Emolument()\n record = Utils.set_model_record(record, params)\n\n request.dbsession.add(record)\n\n # delete all remaining emoluments\n for item in emoluments:\n request.dbsession.delete(item)\n\n return Utils.get_data_save_response(Constant.SUCCESS_SAVE.format(Emolument.__tablename__))\n\n\n@view_config(route_name='emolument_affaire', request_method='DELETE', renderer='json')\ndef emolument_affaire_delete_view(request):\n \"\"\"\n Delete emolument_affaire\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n emolument_affaire_id = request.params['emolument_affaire_id'] if \"emolument_affaire_id\" in request.params else None\n affaire_id = request.params['affaire_id'] if \"affaire_id\" in request.params else None\n\n\n # Remove from Emolument\n records = request.dbsession.query(Emolument).filter(\n Emolument.emolument_affaire_id == emolument_affaire_id\n ).all()\n\n for record in records:\n request.dbsession.delete(record)\n\n\n # Remove from EmolumentAffaire\n record = request.dbsession.query(EmolumentAffaire).filter(\n EmolumentAffaire.id == emolument_affaire_id\n ).filter(\n EmolumentAffaire.affaire_id == affaire_id\n ).first()\n\n if not record:\n raise CustomError(\n CustomError.RECORD_WITH_ID_NOT_FOUND.format(EmolumentAffaire.__tablename__, emolument_affaire_id))\n\n request.dbsession.delete(record)\n\n return Utils.get_data_save_response(Constant.SUCCESS_DELETE.format(Emolument.__tablename__))\n\n\n#######################################\n### EMOLUMENT AFFAIRE REPARTITION ###\n#######################################\n\n@view_config(route_name='emolument_affaire_repartiton', request_method='GET', renderer='json')\ndef emolument_affaire_repartiton_view(request):\n \"\"\"\n get emolument_affaire_repartiton\n \"\"\"\n # Check connected\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n records = request.dbsession.query(EmolumentAffaireRepartition)\n \n if \"emolument_affaire_id\" in request.params:\n records = records.filter(EmolumentAffaireRepartition.emolument_affaire_id == request.params[\"emolument_affaire_id\"])\n \n if \"facture_id\" in request.params:\n records = records.filter(EmolumentAffaireRepartition.facture_id == request.params[\"facture_id\"])\n \n records = records.all()\n\n if len(records)>0:\n return Utils.serialize_many(records)\n else:\n return []\n\n\n@view_config(route_name='emolument_affaire_repartiton', request_method='POST', renderer='json')\ndef emolument_affaire_repartiton_new_view(request):\n \"\"\"\n Add new emolument_affaire_repartiton\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n emolument_affaire_id = request.params[\"emolument_affaire_id\"] if \"emolument_affaire_id\" in request.params else None\n emolument_facture_repartition = json.loads(request.params[\"emolument_facture_repartition\"]) if \"emolument_facture_repartition\" in request.params else None\n\n # get records of current emolument_affaire_id\n emolumentAffaireRepartition = request.dbsession.query(EmolumentAffaireRepartition).filter(\n EmolumentAffaireRepartition.emolument_affaire_id == emolument_affaire_id\n ).all()\n \n\n # iterate through emolument_facture_repartition\n for efr_i in emolument_facture_repartition:\n # test if efr_i exists already in db\n record = None\n for idx, eaf_i in enumerate(emolumentAffaireRepartition):\n if eaf_i.facture_id == efr_i['id']:\n # La relation existe déjà, la modifier\n record = emolumentAffaireRepartition.pop(idx)\n break\n \n if not record is None:\n if float(record.repartition) != float(efr_i['emolument_repartition']):\n if float(efr_i['emolument_repartition']) == 0:\n # supprimer l'entrée car la répartition est nulle\n request.dbsession.delete(record)\n else:\n # enregistrer la nouvelle répartition\n params = Utils._params(\n facture_id = efr_i['id'],\n emolument_affaire_id = emolument_affaire_id,\n repartition = efr_i['emolument_repartition']\n )\n record = Utils.set_model_record(record, params)\n \n else:\n # Créer l'entrée inexistante et si la répartition est non nulle\n if float(efr_i['emolument_repartition']) > 0:\n record = EmolumentAffaireRepartition()\n params = Utils._params(\n facture_id = efr_i['id'],\n emolument_affaire_id = emolument_affaire_id,\n repartition = efr_i['emolument_repartition']\n )\n record = Utils.set_model_record(record, params)\n\n request.dbsession.add(record)\n\n # remove items in db not posted\n for item in emolumentAffaireRepartition:\n request.dbsession.delete(item) \n\n return Utils.get_data_save_response(Constant.SUCCESS_SAVE.format(EmolumentAffaireRepartition.__tablename__))\n\n\n@view_config(route_name='emolument_affaire_repartiton', request_method='DELETE', renderer='json')\ndef emolument_affaire_repartiton_delete_view(request):\n \"\"\"\n Delete emolument_affaire_repartiton by emolument_affaire_id\n \"\"\"\n # Check authorization\n if not Utils.has_permission(request, request.registry.settings['affaire_facture_edition']):\n raise exc.HTTPForbidden()\n\n emolument_affaire_id = request.params[\"emolument_affaire_id\"] if \"emolument_affaire_id\" in request.params else None\n\n records = request.dbsession.query(EmolumentAffaireRepartition).filter(\n EmolumentAffaireRepartition.emolument_affaire_id == emolument_affaire_id\n ).all()\n\n for record in records:\n request.dbsession.delete(record)\n\n\n@view_config(route_name='export_emoluments_pdf', request_method='POST')\ndef export_emoluments_pdf_view(request):\n \"\"\"\n Create PDF of emoluments\n \"\"\"\n # Check authorization\n if not check_connected(request):\n raise exc.HTTPForbidden()\n\n now = datetime.datetime.now()\n\n # get request params\n tableau_emoluments_id = request.params['tableau_emoluments_id'] if 'tableau_emoluments_id' in request.params else None\n affaire_id = request.params['affaire_id'] if 'affaire_id' in request.params else None\n tableau_emoluments_html = request.params['tableau_emoluments_html'] if 'tableau_emoluments_html' in request.params else None\n tableau_recapitulatif_html = request.params['tableau_recapitulatif_html'] if 'tableau_recapitulatif_html' in request.params else None\n\n # get facture_id\n factures = request.dbsession.query(\n Facture\n ).join(\n EmolumentAffaireRepartition\n ).filter(\n EmolumentAffaireRepartition.emolument_affaire_id == tableau_emoluments_id\n ).all()\n\n #get affaire\n affaire = request.dbsession.query(VAffaire).filter(VAffaire.id == affaire_id).first()\n\n # get bf_emolument\n numeros_id = request.dbsession.query(EmolumentAffaire).filter(EmolumentAffaire.id == tableau_emoluments_id).first().numeros_id\n emolument_bf_html = None\n if numeros_id is not None:\n emolument_bf_html = []\n for num_id in numeros_id:\n emolument_bf_html.append(\n str(request.dbsession.query(Numero).filter(Numero.id == num_id).first().numero)\n )\n\n emolument_bf_html = \", \".join(emolument_bf_html)\n\n\n d = {\"now\": now.strftime(\"%d.%m.%Y, %H:%M:%S\")}\n\n header_str = \"\"\n header_str += \"\"\n header_str += '

    DÉPARTEMENT DU DÉVELOPPEMENT
    \\\n TERRITORIAL ET DE L\\'ENVIRONNEMENT

    \\\n SERVICE DE LA GÉOMATIQUE ET
    \\\n DU REGISTRE FONCIER

    '\n\n header_str += \"

    Tableau des émoluments de la mensuration officielle

    \"\n header_str += \"

    Affaire n° \" + str(affaire_id) + \" sur le cadastre: \" + affaire.cadastre + \"

    \"\n \n # numéros de BF s'ils sont rattachés\n if emolument_bf_html is not None and emolument_bf_html != \"\":\n header_str += \"

    Bien(s)-fonds n° \" + emolument_bf_html + \"

    \"\n header_str += \"

    Emolument n° \" + str(tableau_emoluments_id) + \"

    \"\n \n # edit facture_html\n if len(factures) == 0:\n factures_html = \"Aucune facture rattachée à l'émolument\"\n else:\n factures_html = []\n for facture in factures:\n if facture.sap is not None and facture.date is not None:\n factures_html.append(\"n°\" + str(facture.sap) + \" du \" + datetime.datetime.strftime(facture.date, \"%d.%m.%Y\"))\n factures_html = \" / \".join(factures_html)\n if factures_html == \"\" or factures_html == []:\n factures_html = \"La facture n'a pas encore été envoyée\"\n\n header_str += '

    Facture(s): ' + factures_html + \"

    \"\n\n\n tableau_emoluments_html = header_str + tableau_emoluments_html\n tableau_emoluments_html += '

    Récapitulatif

    '\n tableau_emoluments_html += tableau_recapitulatif_html + \"\"\n\n filename = \"Tableau_émoluments_\" + str(tableau_emoluments_id) + \"_Affaire_\" + str(affaire_id) + \".pdf\"\n\n result = requests.post(request.registry.settings['weasyprint_baseurl'] + filename, data=tableau_emoluments_html)\n\n response = Response(result.content)\n params = response.content_type_params\n params['filename'] = filename\n response.content_type = 'application/pdf'\n response.content_type_params = params\n return response\n", "sub_path": "back/infolica/views/facture_emolument.py", "file_name": "facture_emolument.py", "file_ext": "py", "file_size_in_byte": 26221, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "infolica.scripts.authentication.check_connected", "line_number": 29, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 30, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 30, "usage_type": "name"}, {"api_name": "infolica.models.models.TableauEmoluments", "line_number": 32, "usage_type": "argument"}, {"api_name": "infolica.models.models.TableauEmoluments.id", "line_number": 32, "usage_type": "attribute"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 33, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 33, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 23, "usage_type": "call"}, {"api_name": "infolica.scripts.authentication.check_connected", "line_number": 42, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 43, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 43, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 47, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaire.affaire_id", "line_number": 50, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 50, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment", "line_number": 59, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 59, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment_f", "line_number": 60, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 60, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.emolument_affaire_id", "line_number": 62, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 62, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment", "line_number": 64, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 64, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment", "line_number": 66, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 66, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment_f", "line_number": 67, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 67, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.batiment.asc", "line_number": 69, "usage_type": "call"}, {"api_name": "infolica.models.models.Emolument.batiment", "line_number": 69, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 69, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 77, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 77, "usage_type": "name"}, {"api_name": "infolica.models.models.VNumerosAffaires", "line_number": 77, "usage_type": "argument"}, {"api_name": "sqlalchemy.sql.elements.and_", "line_number": 78, "usage_type": "call"}, {"api_name": "infolica.models.models.VNumerosAffaires.numero_id.in_", "line_number": 79, "usage_type": "call"}, {"api_name": "infolica.models.models.VNumerosAffaires.numero_id", "line_number": 79, "usage_type": "attribute"}, {"api_name": "infolica.models.models.VNumerosAffaires", "line_number": 79, "usage_type": "name"}, {"api_name": "infolica.models.models.VNumerosAffaires.affaire_id", "line_number": 80, "usage_type": "attribute"}, {"api_name": "infolica.models.models.VNumerosAffaires", "line_number": 80, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 85, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 85, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 86, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.emolument_affaire_id", "line_number": 87, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 87, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 92, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 92, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 36, "usage_type": "call"}, {"api_name": "infolica.scripts.authentication.check_connected", "line_number": 121, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 122, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 122, "usage_type": "name"}, {"api_name": "infolica.models.models.TableauEmoluments", "line_number": 126, "usage_type": "argument"}, {"api_name": "infolica.models.models.TableauEmoluments.date_sortie", "line_number": 128, "usage_type": "attribute"}, {"api_name": "infolica.models.models.TableauEmoluments", "line_number": 128, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 131, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 131, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument", "line_number": 137, "usage_type": "argument"}, {"api_name": "infolica.models.models.Emolument.emolument_affaire_id", "line_number": 139, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 139, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 142, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 142, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.utilise", "line_number": 146, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 146, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.id", "line_number": 148, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 148, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 115, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 166, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 166, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 167, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 167, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 170, "usage_type": "call"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 172, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 173, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 173, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 160, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 187, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 187, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 188, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 188, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 191, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 197, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 197, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument", "line_number": 208, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 209, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 209, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 213, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 213, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE.format", "line_number": 213, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE", "line_number": 213, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 213, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.__tablename__", "line_number": 213, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 213, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 181, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 222, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 222, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 223, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 223, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 226, "usage_type": "call"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 230, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaire.id", "line_number": 231, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 231, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 234, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 234, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 236, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 236, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE.format", "line_number": 236, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE", "line_number": 236, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 236, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.__tablename__", "line_number": 236, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 236, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 216, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 245, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 245, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 246, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 246, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 252, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaire.id", "line_number": 253, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 253, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 256, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 256, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 258, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 258, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE.format", "line_number": 258, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE", "line_number": 258, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 258, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.__tablename__", "line_number": 258, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 258, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 239, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 267, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 267, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 268, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 268, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 271, "usage_type": "call"}, {"api_name": "infolica.models.models.Emolument", "line_number": 275, "usage_type": "argument"}, {"api_name": "infolica.models.models.Emolument.emolument_affaire_id", "line_number": 276, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 276, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 301, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 301, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 309, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 309, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 316, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 316, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument", "line_number": 327, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 328, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 328, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 336, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 336, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE.format", "line_number": 336, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE", "line_number": 336, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 336, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.__tablename__", "line_number": 336, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 336, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 261, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 345, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 345, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 346, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 346, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument", "line_number": 353, "usage_type": "argument"}, {"api_name": "infolica.models.models.Emolument.emolument_affaire_id", "line_number": 354, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 354, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 362, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaire.id", "line_number": 363, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 363, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.affaire_id", "line_number": 365, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 365, "usage_type": "name"}, {"api_name": "infolica.exceptions.custom_error.CustomError", "line_number": 369, "usage_type": "call"}, {"api_name": "infolica.exceptions.custom_error.CustomError.RECORD_WITH_ID_NOT_FOUND.format", "line_number": 370, "usage_type": "call"}, {"api_name": "infolica.exceptions.custom_error.CustomError.RECORD_WITH_ID_NOT_FOUND", "line_number": 370, "usage_type": "attribute"}, {"api_name": "infolica.exceptions.custom_error.CustomError", "line_number": 370, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaire.__tablename__", "line_number": 370, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 370, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 374, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 374, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_DELETE.format", "line_number": 374, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_DELETE", "line_number": 374, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 374, "usage_type": "name"}, {"api_name": "infolica.models.models.Emolument.__tablename__", "line_number": 374, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Emolument", "line_number": 374, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 339, "usage_type": "call"}, {"api_name": "infolica.scripts.authentication.check_connected", "line_number": 387, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 388, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 388, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 390, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.emolument_affaire_id", "line_number": 393, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 393, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.facture_id", "line_number": 396, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 396, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.serialize_many", "line_number": 401, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 401, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 381, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 412, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 412, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 413, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 413, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 416, "usage_type": "call"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 419, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.emolument_affaire_id", "line_number": 420, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 420, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 441, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 441, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 446, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 446, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 451, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils._params", "line_number": 452, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 452, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.set_model_record", "line_number": 457, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 457, "usage_type": "name"}, {"api_name": "infolica.scripts.utils.Utils.get_data_save_response", "line_number": 465, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 465, "usage_type": "name"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE.format", "line_number": 465, "usage_type": "call"}, {"api_name": "infolica.models.constant.Constant.SUCCESS_SAVE", "line_number": 465, "usage_type": "attribute"}, {"api_name": "infolica.models.constant.Constant", "line_number": 465, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.__tablename__", "line_number": 465, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 465, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 406, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils.has_permission", "line_number": 474, "usage_type": "call"}, {"api_name": "infolica.scripts.utils.Utils", "line_number": 474, "usage_type": "name"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 475, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 475, "usage_type": "name"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 479, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.emolument_affaire_id", "line_number": 480, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 480, "usage_type": "name"}, {"api_name": "pyramid.view.view_config", "line_number": 468, "usage_type": "call"}, {"api_name": "infolica.scripts.authentication.check_connected", "line_number": 493, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions.HTTPForbidden", "line_number": 494, "usage_type": "call"}, {"api_name": "pyramid.httpexceptions", "line_number": 494, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 496, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 496, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 508, "usage_type": "argument"}, {"api_name": "infolica.models.models.Facture", "line_number": 506, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition.emolument_affaire_id", "line_number": 510, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaireRepartition", "line_number": 510, "usage_type": "name"}, {"api_name": "infolica.models.models.VAffaire", "line_number": 514, "usage_type": "argument"}, {"api_name": "infolica.models.models.VAffaire.id", "line_number": 514, "usage_type": "attribute"}, {"api_name": "infolica.models.models.EmolumentAffaire", "line_number": 517, "usage_type": "argument"}, {"api_name": "infolica.models.models.EmolumentAffaire.id", "line_number": 517, "usage_type": "attribute"}, {"api_name": "infolica.models.models.Numero", "line_number": 523, "usage_type": "argument"}, {"api_name": "infolica.models.models.Numero.id", "line_number": 523, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strftime", "line_number": 616, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 616, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 630, "usage_type": "call"}, {"api_name": "pyramid.response.Response", "line_number": 632, "usage_type": "call"}, {"api_name": "pyramid.view.view_config", "line_number": 487, "usage_type": "call"}]} +{"seq_id": "610059072", "text": "from app import create_app , db\n\nfrom app.auth.models import EmpDetails , EmpIdscheck\n\nfrom sqlalchemy import exc\n\nfrom app.catalog.models import EmpIds\n\nflask_app = create_app('dev')\nwith flask_app.app_context():\n db.create_all()\n try:\n if not EmpDetails.query.filter_by(emp_id=10786).first():\n EmpDetails.create_user(empid=10786,\n name=\"saikrishna\",\n email=\"krishnakrrish8@gmail.com\",\n gender=\"Male\",\n login=\"1:00 PM\",\n logout=\"10:00 PM\",\n password='secret',\n confirm='secret',\n hno='1-7-208,Maruthi nagar',\n address=\"Santosh nagar,Hyderabad\",\n pincode=500060)\n except exc.IntegrityError:\n flask_app.run()\n", "sub_path": "run.py", "file_name": "run.py", "file_ext": "py", "file_size_in_byte": 952, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "app.create_app", "line_number": 9, "usage_type": "call"}, {"api_name": "app.db.create_all", "line_number": 11, "usage_type": "call"}, {"api_name": "app.db", "line_number": 11, "usage_type": "name"}, {"api_name": "app.auth.models.EmpDetails.query.filter_by", "line_number": 13, "usage_type": "call"}, {"api_name": "app.auth.models.EmpDetails.query", "line_number": 13, "usage_type": "attribute"}, {"api_name": "app.auth.models.EmpDetails", "line_number": 13, "usage_type": "name"}, {"api_name": "app.auth.models.EmpDetails.create_user", "line_number": 14, "usage_type": "call"}, {"api_name": "app.auth.models.EmpDetails", "line_number": 14, "usage_type": "name"}, {"api_name": "sqlalchemy.exc.IntegrityError", "line_number": 25, "usage_type": "attribute"}, {"api_name": "sqlalchemy.exc", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "128928591", "text": "import os\nimport warnings\nfrom datetime import datetime\n\nfrom bs4 import BeautifulSoup\nfrom core.db.rdb_handler import RDBHandler\nfrom scrapy.crawler import CrawlerProcess\nfrom scrapy.utils.project import get_project_settings\n\nimport scrapy\nfrom core import util\nfrom core.aws.aws_handler import AwsHandler\nfrom core.db.mongodb_handler import MongodbHandler\n\n\nclass TestSpider(scrapy.Spider):\n name = \"test_spider\"\n\n def __init__(self, offset=0, chunksize=0, job_id='', task_id=''):\n # config 로딩 및 데이터 베이스 연결\n self.offset = offset\n self.chunksize = chunksize\n self.job_id = job_id\n self.task_id = task_id\n self.starttime = datetime.now()\n\n self.env = util.get_env()\n self.config = util.load_config(self.env)\n\n self.aws_handler = AwsHandler(self.config)\n self.rdb_handler = RDBHandler(self.config)\n self.mongo_handler = MongodbHandler(self.config)\n\n # 크롤링 진행 시에 필요한 기본적인 url들\n self.start_urls = ['https://neezmoa.com']\n self.list_url = \"https://neezmoa.com/index.php?mid=youtuber&page={0}\"\n self.max_page = 0\n\n # warning 로그를 무시하기 위한 설정, 에러 로그 수집을 위한 sentry 설정\n warnings.simplefilter('ignore')\n\n # task 객체의 상태를 WAIT에서 running으로 변경\n self.mongo_handler.update_doc_set(collection='tasks', doc_id=self.task_id, key='status', value='RUNNING')\n # job 객체에 현재 대기 중인 테스크 개수는 -1, 동작 중인 테스크 개수는 +1로 업데이트\n self.mongo_handler.update_job_task_running(job_id=self.job_id)\n\n # 여기서 해당 크롤링 프로세스의 정보를 담은 json을 생성한 다음 몽고 디비에 저장해준다.\n def closed(self, reason):\n stats = self.crawler.stats.get_stats()\n logfile_path = os.path.join(util.project_home(), 'tmp/{0}.txt'.format(self.task_id))\n _, logfile_url = self.aws_handler.upload_file_to_s3(filepath=logfile_path,\n savepath='{0}/{1}.txt'.format(self.job_id, self.task_id))\n endtime = datetime.now()\n timecost = (endtime - self.starttime).seconds\n\n self.mongo_handler.update_job_task_done(job_id=self.job_id)\n self.mongo_handler.update_task_done(task_id=self.task_id, timecost=timecost, logfile_url=logfile_url,\n stats=stats, reason=reason)\n\n\n # 한 페이지에 40명의 유투버가 포함되어 있는 리스트 페이지를 요청\n def parse(self, response):\n yield scrapy.FormRequest(url='https://neezmoa.com/___verify',\n method='POST',\n callback=self.parse_firstapge)\n\n def parse_firstapge(self, response):\n soup = BeautifulSoup(response.text, 'lxml')\n page_list = soup.find_all('a', class_='pagination-num')\n self.max_page = int(page_list[-1].contents[0])\n\n for i in range(self.offset, self.offset + self.chunksize):\n url = self.list_url.format(i+1)\n yield scrapy.Request(url=url, callback=self.parse_list)\n\n # 리스트 페이지를 파싱한 다음, 각 유투버 별 상세 페이지를 요청\n def parse_list(self, response):\n soup = BeautifulSoup(response.text, 'lxml')\n creator_titles = soup.find_all('div', class_='yook-tb-channeltitle')\n creator_urls = soup.find_all('a', class_='ab-link')\n creator_titleurls = []\n\n def start(self, job_id, task_id, offset, chunksize):\n settings = get_project_settings()\n settings['USER_AGENT'] = 'Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)'\n settings['HTTPERROR_ALLOWED_CODES'] = [403]\n settings['LOG_LEVEL'] = 'INFO'\n settings['LOG_FILE'] = os.path.join(util.project_home(), 'tmp/{0}.txt'.format(task_id))\n\n process = CrawlerProcess(settings)\n process.crawl(TestSpider, job_id=job_id, task_id=task_id, offset=offset, chunksize=chunksize)\n process.start()\n\n", "sub_path": "crawler/spiders/test_spider.py", "file_name": "test_spider.py", "file_ext": "py", "file_size_in_byte": 4110, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "scrapy.Spider", "line_number": 16, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 25, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 25, "usage_type": "name"}, {"api_name": "core.util.get_env", "line_number": 27, "usage_type": "call"}, {"api_name": "core.util", "line_number": 27, "usage_type": "name"}, {"api_name": "core.util.load_config", "line_number": 28, "usage_type": "call"}, {"api_name": "core.util", "line_number": 28, "usage_type": "name"}, {"api_name": "core.aws.aws_handler.AwsHandler", "line_number": 30, "usage_type": "call"}, {"api_name": "core.db.rdb_handler.RDBHandler", "line_number": 31, "usage_type": "call"}, {"api_name": "core.db.mongodb_handler.MongodbHandler", "line_number": 32, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 40, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 50, "usage_type": "call"}, {"api_name": "os.path", "line_number": 50, "usage_type": "attribute"}, {"api_name": "core.util.project_home", "line_number": 50, "usage_type": "call"}, {"api_name": "core.util", "line_number": 50, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 53, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 53, "usage_type": "name"}, {"api_name": "scrapy.FormRequest", "line_number": 63, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 68, "usage_type": "call"}, {"api_name": "scrapy.Request", "line_number": 74, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 78, "usage_type": "call"}, {"api_name": "scrapy.utils.project.get_project_settings", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 88, "usage_type": "call"}, {"api_name": "os.path", "line_number": 88, "usage_type": "attribute"}, {"api_name": "core.util.project_home", "line_number": 88, "usage_type": "call"}, {"api_name": "core.util", "line_number": 88, "usage_type": "name"}, {"api_name": "scrapy.crawler.CrawlerProcess", "line_number": 90, "usage_type": "call"}]} +{"seq_id": "599782055", "text": "#!/usr/bin/env python\n# -*- coding: iso-8859-1 -*-\n\nimport sys, re, argparse\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nimport matplotlib\nfrom math import *\nfrom collections import defaultdict\n\n\n# du fait d'erreurs de calcul, on se retrouve parfois avec des distances négatives\n# on prend ici une valeur minimale de distance, positive (pour pouvoir prendre la racine) et non nulle (pour pouvoir prendre l'inverse)\nMINDIST = 1e-18\n\n\nclass Example:\n\t\"\"\"\n\tUn exemple : \n\tvector = représentation vectorielle (Ovector) d'un objet\n\tgold_class = la classe gold pour cet objet\n\t\"\"\"\n\tdef __init__(self, example_number, gold_class):\n\t\tself.gold_class = gold_class\n\t\tself.example_number = example_number\n\t\tself.vector = Ovector()\n\n\tdef add_feat(self, featname, val):\n\t\tself.vector.add_feat(featname, val)\n\n\nclass Ovector:\n\t\"\"\"\n\tUn vecteur représentant un objet\n\n\tmembres\n\t- f= simple dictionnaire nom_de_trait => valeur\n\t\t Les traits non stockés correspondent à une valeur nulle\n\t- norm_square : la norme au carré\n\t\"\"\"\n\tdef __init__(self):\n\t\tself.f = {}\n\t\tself.norm_square = 0\n\n\tdef add_feat(self, featname, val = 0.0):\n\t\tself.f[featname] = val\n\t\tself.norm_square += val*val\n\n\n\tdef prettyprint(self):\n\t\tfor feat in sorted(self.f, lambda x,y: cmp( self.f[y], self.f[x] ) or cmp(x,y)):\n\t\t\tprint (feat + \"\t\" + str(self.f[feat]))\n\n\tdef distance_to_vector(self, other_vector):\n\t\t\"\"\" distance euclidienne entre self et other_vector, en ayant precalculé les normes au carre de chacun \"\"\"\n\t\t# NB: passer par la formulation sigma [ (ai - bi)^2 ] = sigma (ai^2) + sigma (bi^2) -2 sigma (ai*bi) \n\t\t#\t\t\t\t\t\t\t\t\t\t\t\t\t= norm_square(A) + norm_square(B) - 2 A.B\n\n\t\treturn sqrt(self.norm_square + other_vector.norm_square - 2* self.dot_product(other_vector))\n\n\tdef dot_product(self, other_vector):\n\t\t\"\"\" rend le produit scalaire de self et other_vector \"\"\"\n\t\tdot = 0\n\t\tfor feat in self.f :\n\t\t\tif feat in other_vector.f :\n\t\t\t\tdot += self.f[feat]*other_vector.f[feat]\n\t\treturn dot\n\n\tdef cosinus(self, other_vector):\n\t\t\"\"\" rend le cosinus de self et other_vector \"\"\"\n\t\treturn self.dot_product(other_vector) / sqrt(self.norm_square * other_vector.norm_square)\n\n\nclass KNN:\n\t\"\"\"\n\tK-NN pour la classification de documents (multiclasse)\n\n\tmembres = \n\n\tk = l'hyperparametre K : le nombre de voisins a considerer\n\n\texamples = liste d'instances de Example\n\n\tclasses = liste des classes (telles que recensées dans les exemples)\n\n\t\"\"\"\n\tdef __init__(self, examples, K = 1, weight_neighbors = None, use_cosinus = False, trace = False):\n\t\t\"\"\" \n\t\tsimple positionnement des membres et recensement des classes connues\n\t\t\"\"\"\n\t\t# les exemples : liste d'instances de Example\n\t\tself.examples = examples\n\t\t# le nb de voisins\n\t\tself.K = K\n\t\t# booleen : on pondere les voisins (par inverse de la distance) ou pas\n\t\tself.weight_neighbors = weight_neighbors\n\n\t\t# booleen : pour utiliser plutot la similarité cosinus\n\t\tself.use_cosinus = use_cosinus\n\n\t\tself.trace = trace\n\t\t\n\n\tdef weigth(self, x) :\n\t\tif not self.weight_neighbors :\n\t\t\treturn 1\n\t\telif self.use_cosinus :\n\t\t\treturn x\n\t\telse :\n\t\t\treturn 1/x\n\n\tdef classify(self, ovector):\n\t\t\"\"\"\n\t\tRéalise la prédiction du classifieur K-NN pour le ovector\n\t\tpour les valeurs de k allant de 1 à self.K\n\n\t\tA partir d'un vecteur de traits représentant un objet\n\t\tretourne un vecteur des classes assignées de longueur K : \n\t\tla classe à la i-eme position est la classe assignée par l'algo K-NN, avec K = i\n\t\t\"\"\"\n\t\t#Select the classifying function\n\t\tif self.use_cosinus :\n\t\t\tprox = Ovector.cosinus\n\t\telse :\n\t\t\tprox = Ovector.distance_to_vector\n\t\t\n\t\t#Classify\n\t\tneighbors = []\n\t\tfor example in self.examples :\n\t\t\tneighbors.append((example.gold_class, prox(example.vector, ovector)))\n\n\t\t#Sort neighbors\n\t\tbest = sorted(neighbors, key = lambda x : x[1], reverse = self.use_cosinus)\n\n\t\t#Extracts nearests neighbors\n\t\tresult = []\n\t\ti_class = defaultdict(int)\n\t\tfor i in range(self.K):\n\t\t\ti_class[best[i][0]] += self.weigth(best[i][1])\n\t\t\ttop_ones = 0 # top_ones is for preventing list out of range while sorting alphabeticals.\n\t\t\tclasses = sorted(i_class.items(), key = lambda x : x[1], reverse = True)\n\t\t\talpha = set()\n\t\t\twhile top_ones < len(classes) and classes[top_ones][1] == classes[0][1] :\n\t\t\t\talpha.add(classes[top_ones][0])\n\t\t\t\ttop_ones += 1\n\t\t\tresult.append(sorted(alpha)[0])\n\t\treturn result\n\n\n\tdef evaluate_on_test_set(self, test_examples):\n\t\t\"\"\" Application du classifieur sur une liste d'exemples de test, et evaluation (accuracy) \n\t\tpour les valeurs de k allant de 1 à self.K\n\t\tRetourne une liste d'accuracy (pour les valeurs de k à self.K)\n\t\t\"\"\"\n\t\tresults = [0 for i in range(self.K)]\n\t\tfor example in test_examples :\n\t\t\tclass_example = self.classify(example.vector)\n\t\t\tresult_example = [1 if example.gold_class == class_i else 0 for class_i in class_example]\n\t\t\tresults = [x + y for x, y in zip(results, result_example)]\n\t\treturn [acc / len(test_examples) for acc in results]\n\t\t\n\t\t\n\ndef read_examples(infile):\n\t\"\"\" Lit un fichier d'exemples \n\tet retourne une liste d'instances de Example\n\t\"\"\"\n\tstream = open(infile)\n\texamples = []\n\texample = None\n\twhile True:\n\t\tline = stream.readline()\n\t\tif not line:\n\t\t\tbreak\n\t\tline = line[0:-1]\n\t\tif line.startswith(\"EXAMPLE_NB\"):\n\t\t\tif example is not None:\n\t\t\t\texamples.append(example)\n\t\t\tcols = line.split('\t')\n\t\t\tgold_class = cols[3]\n\t\t\texample_number = cols[1]\n\t\t\texample = Example(example_number, gold_class)\n\t\telif line and example is not None:\n\t\t\t(featname, val) = line.split('\t')\n\t\t\texample.add_feat(featname, float(val))\n\n\tif example is not None:\n\t\texamples.append(example)\n\treturn examples\n\n\n\nusage = \"\"\" CLASSIFIEUR de DOCUMENTS, de type K-NN\n\n \"\"\"+sys.argv[0]+\"\"\" [options] EXAMPLES_FILE TEST_FILE\n\n EXAMPLES_FILE et TEST_FILE sont au format *.examples\n\n\"\"\"\n\nparser = argparse.ArgumentParser(usage = usage)\nparser.add_argument('examples_file', default = None,\n\t\t\t\t\thelp = 'Exemples utilisés comme voisins pour la prédiction KNN (au format .examples)')\n\nparser.add_argument('test_file', default = None,\n\t\t\t\t\thelp = 'Exemples de test (au format .examples)')\n\nparser.add_argument('-k', \"--k\", default = 1, type = int,\n\t\t\t\t\thelp = 'Hyperparametre K : le nombre max de voisins a considerer pour la classification (toutes les valeurs de 1 a k seront testees). Default = 1')\n\nparser.add_argument('-v', \"--trace\", action = \"store_true\",\n\t\t\t\t\thelp = \"A utiliser pour declencher un mode verbeux. Default = False\")\n\nparser.add_argument('-w', \"--weight_neighbors\", action = \"store_true\",\n\t\t\t\t\thelp = \"Ponderation des voisins : si cosinus: ponderation par le cosinus, si distance, ponderation par l'inverse de la distance. Defaut = None\")\n\nparser.add_argument('-c', \"--use_cosinus\", action = \"store_true\",\n\t\t\t\t\thelp = \"A utiliser pour passer a une mesure de similarite cosinus, au lieu d'une distance euclidienne. Default = False\")\n\nparser.add_argument('-t', \"--tune\", action = \"store_true\", default = \"\",\n\t\t\t\t\thelp = 'A utiliser pour declencher le tuning des hyperparametres: cos ou dist, avec ou sans ponderation, et figure des performances en fonction de k. Default = False')\n\nparser.add_argument('-f', \"--figure_file\", default = \"graphique.pdf\",\n\t\t\t\t\thelp = 'Pour le mode \"tune\": Base de nom de fichier pour graphique precision en fonction de K. Default = \"graphique.pdf\".')\n\nargs = parser.parse_args()\n\n#------------------------------------------------------------\n# Chargement des exemples d'apprentissage du classifieur KNN\ntraining_examples = read_examples(args.examples_file)\n# Chargement des exemples de test\ntest_examples = read_examples(args.test_file)\n\nif args.tune :\n\ttuner = defaultdict(object)\n\tfor cos in [True, False] : \n\t\tfor weight in [True, False] :\n\t\t\tclassifier = KNN( examples = training_examples,\n\t\t\t\t\t\t\tK = args.k,\n\t\t\t\t\t\t\tweight_neighbors = weight,\n\t\t\t\t\t\t\tuse_cosinus = cos,\n\t\t\t\t\t\t\ttrace = args.trace)\n\t\t\ttuner[(\"Cosinus\" if cos else \"Euclide\") + \" \" + (\"Pondere\" if weight else \"Standard\")] = classifier.evaluate_on_test_set(test_examples)\n\tdf = pd.DataFrame.from_dict(tuner)\n\tdf.index = range(1,args.k+1)\n\tprint(df)\n\tax = sns.lineplot(data = df)\n\tplt.show()\nelse :\n\n\tmyclassifier = KNN( examples = training_examples,\n\t\t\t\t\t\tK = args.k,\n\t\t\t\t\t\tweight_neighbors = args.weight_neighbors,\n\t\t\t\t\t\tuse_cosinus = args.use_cosinus,\n\t\t\t\t\t\ttrace = args.trace)\n\n\t# classification et evaluation sur les exemples de test\n\taccuracies = myclassifier.evaluate_on_test_set(test_examples)\n\tfor i in range(len(accuracies)):\n\t\tprint(\"ACCURACY FOR K =\", i+1, \" : \",\"{:.2%}\".format(accuracies[i]),\n\t\t\t\t\"(weight =\", args.weight_neighbors, \"dist_or_cos =\", \"cos)\" if args.use_cosinus else \"dist)\" )\n\n\n\n", "sub_path": "AA/M1_AA_TD_knn/gadioux_knn.py", "file_name": "gadioux_knn.py", "file_ext": "py", "file_size_in_byte": 8587, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "collections.defaultdict", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 196, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 202, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 236, "usage_type": "call"}, {"api_name": "pandas.DataFrame.from_dict", "line_number": 245, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 245, "usage_type": "attribute"}, {"api_name": "seaborn.lineplot", "line_number": 248, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 249, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 249, "usage_type": "name"}]} +{"seq_id": "623925055", "text": "#from playsound import playsound\nimport logging\nimport wave\nimport sys\n\nclass Sound_builder():\n\n\tdef Build_Output_Sound(self, directory, input_file, by_who = 'Oleg'):\n\t\tlogging.basicConfig(filename=\"Sound_builder.log\", level=logging.INFO)\n\t\tlogging.debug(\"initializing...\")\n\n\t\tinput_text = open(input_file,'r').read()\n\n\t\tlogging.info(\"Trying to write input text from {}: {}\".format(input_file, input_text))\n\n\t\toutfile = \"Output_PooP.wav\"\n\t\tinfiles = []\n\t\tdata= []\n\n\t\tsplited_input_text = input_text.split()\n\n\t\tfor word in splited_input_text:\n\t\t\tfor ltr in word:\n\t\t\t\ttry:\n\t\t\t\t\t#playsound('Oleg/{}.wav'.format(ltr.upper()),True)\n\t\t\t\t\topen(directory + \"/Sound/{}/{}.wav\".format(by_who, ltr.upper()), \"r\")\n\t\t\t\t\tinfiles.append(directory + '/Sound/{}/{}.wav'.format(by_who, ltr.upper()))\n\n\t\t\t\texcept Exception as ex:\n\t\t\t\t\tlogging.error(\"Error at symbol {}: \".format(ltr, ex))\n\t\t\t\t\tpass\n\t\t\t\n\t\t\tinfiles.append(directory + '/Sound/{}/_.wav'.format(by_who))\n\n\t\tfor infile in infiles:\n\t\t\tw = wave.open(infile, 'rb')\n\t\t\tdata.append( [w.getparams(), w.readframes(w.getnframes())] )\n\t\t\tw.close()\n\n\t\tlogging.info(\"Writing output sound at file: \" + outfile)\n\n\t\toutput = wave.open(outfile, 'wb')\n\t\toutput.setparams(data[0][0])\n\n\t\tfor i in range(len(data)):\n\t\t\toutput.writeframes(data[i][1])\n\t\toutput.close()\n\n\t\tlogging.info(\"Successful writed output sound at file: \" + outfile)\n\t\t\n\t\treturn outfile", "sub_path": "NeuroPlusPoop/Sound/Pooper.py", "file_name": "Pooper.py", "file_ext": "py", "file_size_in_byte": 1380, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.basicConfig", "line_number": 9, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 9, "usage_type": "attribute"}, {"api_name": "logging.debug", "line_number": 10, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.error", "line_number": 30, "usage_type": "call"}, {"api_name": "wave.open", "line_number": 36, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 40, "usage_type": "call"}, {"api_name": "wave.open", "line_number": 42, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 49, "usage_type": "call"}]} +{"seq_id": "112911333", "text": "import os\nimport sys\nimport json\nimport sqlite3\nimport pandas\n\ndef loadData(tableName,conn,csvfile):\n print(\"loading\",tableName)\n df = pandas.read_csv(csvfile)\n df.to_sql(tableName, conn, if_exists='append', index=False)\n print(\"done loading\",tableName)\n\ndef run(dbFilename):\n conn = sqlite3.connect(dbFilename)\n cursor = conn.cursor()\n\n cursor.execute(\"DROP TABLE IF EXISTS movies\");\n cursor.execute(\"\"\"\n CREATE TABLE if not exists movies (\n IMDB_Rating DOUBLE,\n Production_Budget DOUBLE,\n Release_Date DOUBLE,\n Rotten_Tomatoes_Rating DOUBLE,\n Running_Time_min DOUBLE,\n US_DVD_Sales DOUBLE,\n US_Gross DOUBLE,\n Worldwide_Gross DOUBLE)\n \"\"\")\n loadData(\"movies\",conn,os.path.join(os.path.dirname(__file__),'..','..','dataset_movies_100M_fixed.csv'))\n\n cursor.execute(\"DROP TABLE IF EXISTS weather\");\n cursor.execute(\"\"\"\n CREATE TABLE if not exists weather(\n ELEVATION DOUBLE,\n LATITUDE DOUBLE,\n LONGITUDE DOUBLE,\n PRECIPITATION DOUBLE,\n RECORD_DATE DOUBLE,\n SNOW DOUBLE,\n TEMP_MAX DOUBLE,\n TEMP_MIN DOUBLE,\n WIND DOUBLE)\n \"\"\")\n loadData(\"weather\",conn,os.path.join(os.path.dirname(__file__),'..','..',\"dataset_weather_100M_fixed.csv\"))\n\n cursor.execute(\"DROP TABLE IF EXISTS flights\");\n cursor.execute(\"\"\"\n CREATE TABLE if not exists flights(\n AIR_TIME DOUBLE,\n ARR_DELAY DOUBLE,\n ARR_TIME DOUBLE,\n DEP_DELAY DOUBLE,\n DEP_TIME DOUBLE,\n DISTANCE DOUBLE,\n FL_DATE TEXT)\n \"\"\")\n loadData(\"flights\",conn,os.path.join(os.path.dirname(__file__),'..','..',\"dataset_flights_100M.csv\"))\n conn.commit()\n conn.close()\n\nif __name__ == \"__main__\":\n try:\n sqliteConfig = json.load(open(os.path.join(os.path.dirname(__file__),'..','..','sqlite.config.json')))\n dbFilename = sqliteConfig['dbFilename']\n run(dbFilename)\n except Exception as e:\n print(e)\n print(\"usage: python\",sys.argv[0],\"[db filename]\")\n sys.exit(0)\n", "sub_path": "setup/sqlite/load_100M.py", "file_name": "load_100M.py", "file_ext": "py", "file_size_in_byte": 2106, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pandas.read_csv", "line_number": 9, "usage_type": "call"}, {"api_name": "sqlite3.connect", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path", "line_number": 29, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 29, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path", "line_number": 44, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 44, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 57, "usage_type": "call"}, {"api_name": "os.path", "line_number": 57, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 57, "usage_type": "call"}, {"api_name": "json.load", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 63, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 68, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 69, "usage_type": "call"}]} +{"seq_id": "258028902", "text": "#!/usr/bin/env python\n\n# MIT License\n#\n# Copyright (c) 2020 Roberto Chamorro / project contributors\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n#\n# Special thanks to\n# https://github.com/bear/python-twitter/\n# and\n# https://github.com/halcy/Mastodon.py\n\n# TODO: Write tests\nimport os\nimport sys\nimport string\nimport random\nimport time\nfrom json.decoder import JSONDecodeError\n\nimport requests\nimport shutil\nimport re\nimport mimetypes\nimport json\nimport yaml\n\n# Try to import libmagic\n# if it fails just use mimetypes\ntry:\n import magic\nexcept ImportError:\n magic = None\n\nfrom datetime import datetime, timedelta\n\n# Parameter to choose whether to update bio, avatar and banner or not - save some bandwidth\ntry:\n arg = sys.argv[1]\nexcept IndexError:\n arg = \"\"\n print(f\"Usage: {sys.argv[0]} [noProfile]\")\n\n\nclass User(object):\n def __init__(self, user_cfg: dict, cfg: dict):\n self.twitter_token = cfg['twitter_token']\n self.signature = \"\"\n self.media_upload = False\n self.support_account = None\n # iterate attrs defined in config\n for attribute in user_cfg:\n self.__setattr__(attribute, user_cfg[attribute])\n self.twitter_url = \"http://twitter.com/\" + self.twitter_username\n try:\n if not hasattr(self, \"max_tweets\"):\n self.max_tweets = cfg['max_tweets']\n except (KeyError, AttributeError):\n # Limit to 50 last tweets - just to make a bit easier and faster to process given how often it is pulled\n self.max_tweets = 50\n pass\n try:\n if not hasattr(self, \"visibility\"):\n self.visibility = cfg['visibility']\n except (KeyError, AttributeError):\n self.visibility = \"unlisted\"\n pass\n if self.visibility not in (\"public\", \"unlisted\", \"private\", \"direct\"):\n raise KeyError(\"Visibility not supported! Values allowed are: public, unlisted, private and direct\")\n try:\n if not hasattr(self, \"sensitive\"):\n self.sensitive = cfg['sensitive']\n except (KeyError, AttributeError):\n self.sensitive = \"true\"\n pass\n if hasattr(self, \"rich_text\"):\n if self.rich_text:\n self.content_type = \"text/markdown\"\n try:\n if not hasattr(self, \"pleroma_base_url\"):\n self.pleroma_base_url = cfg['pleroma_base_url']\n except KeyError:\n raise KeyError(\"No Pleroma URL defined in config! [pleroma_base_url]\")\n try:\n if not hasattr(self, \"twitter_base_url\"):\n self.twitter_base_url = cfg['twitter_base_url']\n except KeyError:\n self.twitter_base_url = \"https://api.twitter.com/1.1\"\n pass\n try:\n if not hasattr(self, \"twitter_base_url_v2\"):\n self.twitter_base_url = cfg['twitter_base_url_v2']\n except KeyError:\n self.twitter_base_url_v2 = \"https://api.twitter.com/2\"\n pass\n if not hasattr(self, \"nitter\"):\n try:\n if cfg['nitter']:\n self.twitter_url = \"http://nitter.net/\" + self.twitter_username\n except KeyError:\n pass\n else:\n if self.nitter:\n self.twitter_url = \"http://nitter.net/\" + self.twitter_username\n self.profile_image_url = None\n self.profile_banner_url = None\n self.display_name = None\n try:\n self.fields = self.replace_vars_in_str(str(user_cfg['fields']))\n self.fields = eval(self.fields)\n except KeyError:\n self.fields = []\n self.bio_text = self.replace_vars_in_str(str(user_cfg['bio_text']))\n # Auth\n self.header_pleroma = {\"Authorization\": \"Bearer \" + self.pleroma_token}\n self.header_twitter = {\"Authorization\": \"Bearer \" + self.twitter_token}\n self.tweets = self._get_tweets('v2')\n self.pinned_tweet_id = self._get_pinned_tweet_id()\n self.last_post_pleroma = None\n # Filesystem\n script_path = os.path.dirname(sys.argv[0])\n self.base_path = os.path.abspath(script_path)\n self.users_path = os.path.join(self.base_path, 'users')\n self.user_path = os.path.join(self.users_path, self.twitter_username)\n self.tweets_temp_path = os.path.join(self.user_path, 'tweets')\n self.avatar_path = os.path.join(self.user_path, 'profile.jpg')\n self.header_path = os.path.join(self.user_path, 'banner.jpg')\n if not os.path.isdir(self.users_path):\n os.mkdir(self.users_path)\n if not os.path.isdir(self.user_path):\n os.mkdir(self.user_path)\n if not os.path.isdir(self.tweets_temp_path):\n os.mkdir(self.tweets_temp_path)\n # Get Twitter info on instance creation\n self._get_twitter_info()\n return\n\n def _get_twitter_info(self):\n \"\"\"Updates User object attributes with current Twitter info\n\n This includes:\n\n * Bio text\n * Profile image url\n * Banner image url\n * Screen name\n\n :return: None\n \"\"\"\n twitter_user_url = self.twitter_base_url + '/users/show.json?screen_name=' + self.twitter_username\n response = requests.get(twitter_user_url, headers=self.header_twitter)\n if not response.ok:\n response.raise_for_status()\n user_twitter = json.loads(response.text)\n self.bio_text = self.bio_text + user_twitter['description']\n # Check if user has profile image\n if 'profile_image_url' in user_twitter.keys():\n self.profile_image_url = user_twitter['profile_image_url']\n # Check if user has banner image\n if 'profile_banner_url' in user_twitter.keys():\n self.profile_banner_url = user_twitter['profile_banner_url']\n self.display_name = user_twitter['name']\n return\n\n def _get_tweets(self, version: str, tweet_id=None):\n \"\"\"Gathers last 'max_tweets' tweets from the user and returns them as an dict\n :param version: Twitter API version to use to retrieve the tweets\n :type version: string\n :param tweet_id: Tweet ID to retrieve\n :type tweet_id: int\n\n :returns: last 'max_tweets' tweets\n :rtype: dict\n \"\"\"\n if version == 'v1.1':\n if tweet_id:\n twitter_status_url = f\"{self.twitter_base_url}/statuses/show.json?id={str(tweet_id)}\"\n response = requests.get(twitter_status_url, headers=self.header_twitter)\n if not response.ok:\n response.raise_for_status()\n tweet = json.loads(response.text)\n return tweet\n else:\n twitter_status_url = self.twitter_base_url + '/statuses/user_timeline.json?screen_name=' + \\\n self.twitter_username + '&count=' + str(self.max_tweets) + '&include_rts=true'\n response = requests.get(twitter_status_url, headers=self.header_twitter)\n if not response.ok:\n response.raise_for_status()\n tweets = json.loads(response.text)\n return tweets\n elif version == 'v2':\n params = {}\n if tweet_id:\n url = f\"{self.twitter_base_url_v2}/tweets/{tweet_id}\"\n else:\n url = f\"{self.twitter_base_url_v2}/tweets/search/recent\" # this only gets tweets from last week\n params.update({\"max_results\": self.max_tweets,\n \"query\": \"from:\" + self.twitter_username})\n # Tweet number must be between 10 and 100\n if not (100 >= self.max_tweets > 10):\n raise ValueError(f\"max_tweets must be between 10 and 100. max_tweets: {self.max_tweets}\")\n\n params.update({\"poll.fields\": \"duration_minutes,end_datetime,id,options,voting_status\",\n \"media.fields\": \"duration_ms,height,media_key,\"\n \"preview_image_url,\"\n \"type,\"\n \"url,\"\n \"width,\"\n \"public_metrics\",\n \"expansions\": \"attachments.poll_ids,\"\n \"attachments.media_keys,\"\n \"author_id,\"\n \"entities.mentions.username,\"\n \"geo.place_id,\"\n \"in_reply_to_user_id,\"\n \"referenced_tweets.id,\"\n \"referenced_tweets.id.author_id\",\n \"tweet.fields\": \"attachments,\"\n \"author_id,\"\n \"context_annotations,\"\n \"conversation_id,\"\n \"created_at,\"\n \"entities,\"\n \"geo,id,\"\n \"in_reply_to_user_id,\"\n \"lang,\"\n \"public_metrics,\"\n \"possibly_sensitive,\"\n \"referenced_tweets,\"\n \"source,\"\n \"text,\"\n \"withheld\"\n })\n response = requests.get(url, headers=self.header_twitter, params=params)\n if not response.ok:\n response.raise_for_status()\n tweets_v2 = json.loads(response.text)\n return tweets_v2\n else:\n raise ValueError(f\"API version not supported: {version}\")\n\n def get_tweets(self):\n return self.tweets\n\n def get_date_last_pleroma_post(self):\n \"\"\"Gathers last post from the user in Pleroma and returns the date of creation.\n \n :returns: Date of last Pleroma post in '%Y-%m-%d %H:%M:%S' format\n \"\"\"\n pleroma_posts_url = self.pleroma_base_url + '/api/v1/accounts/' + self.pleroma_username + '/statuses'\n response = requests.get(pleroma_posts_url, headers=self.header_pleroma)\n if not response.ok:\n response.raise_for_status()\n posts = json.loads(response.text)\n if posts:\n date_pleroma = datetime.strftime(datetime.strptime(posts[0]['created_at'], '%Y-%m-%dT%H:%M:%S.000Z'),\n '%Y-%m-%d %H:%M:%S')\n else:\n date_pleroma = datetime.strftime(datetime.now() - timedelta(days=2), '%Y-%m-%d %H:%M:%S')\n\n return date_pleroma\n\n def _get_pinned_tweet_id(self):\n \"\"\"Retrieves the pinned tweet by the user\n\n :returns: ID of currently pinned tweet\n \"\"\"\n url = self.twitter_base_url_v2 + '/users/by/username/' + self.twitter_username\n params = {\"user.fields\": \"pinned_tweet_id\", \"expansions\": \"pinned_tweet_id\", \"tweet.fields\": \"entities\"}\n response = requests.get(url, headers=self.header_twitter, params=params)\n if not response.ok:\n response.raise_for_status()\n try:\n data = json.loads(response.text)\n pinned_tweet = data['includes']['tweets'][0]\n pinned_tweet_id = pinned_tweet['id']\n except (JSONDecodeError, KeyError):\n pinned_tweet_id = None\n pass\n return pinned_tweet_id\n\n def get_pinned_tweet_(self):\n return self.pinned_tweet_id\n\n def process_tweets(self, tweets_to_post):\n \"\"\"Transforms tweets for posting them to Pleroma\n Expands shortened URLs\n Downloads tweet related media and prepares them for upload\n\n :param tweets_to_post: Dict of tweet objects to be processed\n :type tweets_to_post: dict\n :returns: Tweets ready to be published\n :rtype: list\n \"\"\"\n for tweet in tweets_to_post['data']:\n media = []\n # Replace shortened links\n try:\n for url_entity in tweet['entities']['urls']:\n matching_pattern = url_entity['url']\n matches = re.findall(matching_pattern, tweet['text'])\n for match in matches:\n tweet['text'] = re.sub(match, url_entity['expanded_url'], tweet['text'])\n except KeyError:\n # URI regex\n matching_pattern = r'(?i)\\b((?:https?://|www\\d{0,3}[.]|[a-z0-9.\\-]+[.][a-z]{2,4}/)(?:[^\\s()<>]+|\\(([' \\\n r'^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\))+(?:\\(([^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\)|[^\\s`!()\\[' \\\n r'\\]{};:\\'\".,<>?«»“”‘’]))'\n matches = re.finditer(matching_pattern, tweet['text'])\n for matchNum, match in enumerate(matches, start=1):\n # don't be brave trying to unwound an URL when it gets cut off\n if not match.group().__contains__(\"…\"):\n session = requests.Session() # so connections are recycled\n response = session.head(match.group(), allow_redirects=True)\n if not response.ok:\n response.raise_for_status()\n expanded_url = response.url\n tweet['text'] = re.sub(match.group(), expanded_url, tweet['text'])\n if hasattr(self, \"rich_text\"):\n if self.rich_text:\n matches = re.findall(r'\\B\\@\\w+', tweet['text'])\n for match in matches:\n mention_link = \"[\" + match + \"](https://twitter.com/\" + match[1:] + \")\"\n tweet['text'] = re.sub(match, mention_link, tweet['text'])\n try:\n if self.nitter:\n matching_pattern = \"https://twitter.com\"\n matches = re.findall(matching_pattern, tweet['text'])\n for match in matches:\n tweet['text'] = re.sub(match, \"https://nitter.net\", tweet['text'])\n except AttributeError:\n pass\n try:\n for item in tweet['attachments']['media_keys']:\n for media_include in tweets_to_post['includes']['media']:\n if item == media_include['media_key']:\n # Video download not implemented in v2 yet\n # fallback to v1.1\n if media_include['type'] == 'video' or media_include['type'] == 'animated_gif':\n tweet_video = self._get_tweets('v1.1', tweet['id'])\n for extended_media in tweet_video['extended_entities']['media']:\n media.append(extended_media)\n else:\n media.append(media_include)\n except KeyError:\n pass\n # Create folder to store attachments related to the tweet ID\n tweet_path = os.path.join(self.tweets_temp_path, tweet['id'])\n if not os.path.isdir(tweet_path):\n os.mkdir(tweet_path)\n # Download media only if we plan to upload it later\n if self.media_upload:\n for idx, item in enumerate(media):\n if item['type'] != 'video' and item['type'] != 'animated_gif':\n media_url = item['url']\n else:\n bitrate = 0\n for variant in item['video_info']['variants']:\n try:\n if variant['bitrate'] >= bitrate:\n media_url = variant['url']\n except KeyError:\n pass\n response = requests.get(media_url, stream=True)\n if not response.ok:\n response.raise_for_status()\n response.raw.decode_content = True\n filename = str(idx) + mimetypes.guess_extension(response.headers['Content-Type'])\n with open(os.path.join(self.tweets_temp_path, tweet['id'], filename), 'wb') as outfile:\n shutil.copyfileobj(response.raw, outfile)\n\n # Process poll if exists and no media is used\n try:\n if tweet['attachments']['poll_ids'] and not media:\n\n # tweet_poll = tweet['includes']['polls']\n poll_url = self.twitter_base_url_v2 + '/tweets'\n\n params = {\n 'ids': tweet['id'],\n 'expansions': 'attachments.poll_ids',\n 'poll.fields': 'duration_minutes,'\n 'options'\n }\n\n response = requests.get(\n poll_url, headers=self.header_twitter, params=params)\n if not response.ok:\n response.raise_for_status()\n tweet_poll = json.loads(response.content)[\n 'includes']['polls'][0]\n\n pleroma_poll = {'options': [option['label']\n for option in tweet_poll['options']],\n 'expires_in': tweet_poll['duration_minutes'] * 60}\n\n # Add poll to tweet\n tweet['polls'] = pleroma_poll\n\n else:\n tweet['polls'] = None\n except KeyError:\n tweet['polls'] = None\n pass\n return tweets_to_post\n\n def post_pleroma(self, tweet_id: str, tweet_text: str, poll: dict) -> str:\n \"\"\"Post the given text to the Pleroma instance associated with the User object\n \n :param tweet_id: It will be used to link to the Twitter status if 'signature' is True and to find related media\n :type tweet_id: str\n :param tweet_text: Literal text to use when creating the post.\n :type tweet_text: str\n :returns: id of post\n :rtype: str\n \"\"\"\n # TODO: transform twitter links to nitter links, if self.nitter 'true' in resolved shortened urls\n pleroma_post_url = self.pleroma_base_url + '/api/v1/statuses'\n pleroma_media_url = self.pleroma_base_url + '/api/v1/media'\n\n tweet_folder = os.path.join(self.tweets_temp_path, tweet_id)\n media_files = os.listdir(tweet_folder)\n media_ids = []\n if self.media_upload:\n for file in media_files:\n media_file = open(os.path.join(tweet_folder, file), 'rb')\n media_size = os.stat(os.path.join(tweet_folder, file)).st_size\n mime_type = guess_type(os.path.join(tweet_folder, file))\n timestamp = str(datetime.now().timestamp())\n file_name = \"pleromapyupload_\" + timestamp + \"_\" + random_string(10) + \\\n mimetypes.guess_extension(mime_type)\n file_description = (file_name, media_file, mime_type)\n files = {\"file\": file_description}\n response = requests.post(pleroma_media_url, headers=self.header_pleroma, files=files)\n if not response.ok:\n response.raise_for_status()\n try:\n media_ids.append(json.loads(response.text)['id'])\n except (KeyError, JSONDecodeError):\n print(\"Error uploading media:\\t\" + str(response.text))\n pass\n\n if self.signature:\n signature = '\\n\\n 🐦🔗: ' + self.twitter_url + '/status/' + tweet_id\n tweet_text = tweet_text + signature\n\n data = {\"status\": tweet_text,\n \"sensitive\": str(self.sensitive),\n \"visibility\": self.visibility,\n \"media_ids[]\": media_ids}\n\n if poll:\n data.update({\"poll[options][]\": poll['options'],\n \"poll[expires_in]\": poll['expires_in']})\n\n if hasattr(self, \"rich_text\"):\n if self.rich_text:\n data.update({\"content_type\": self.content_type})\n response = requests.post(pleroma_post_url, data, headers=self.header_pleroma)\n if not response.ok:\n response.raise_for_status()\n print(\"Post in Pleroma:\\t\" + str(response))\n post_id = json.loads(response.text)['id']\n return post_id\n\n def pin_pleroma(self, id_post):\n \"\"\"Tries to unpin previous pinned post if a file containing the ID of the previous post exists, then\n proceeds to pin the post with ID 'id_post'\n\n :param id_post: ID of post to pin\n :returns: ID of post pinned\n :rtype: str\n \"\"\"\n if os.path.isfile(os.path.join(self.user_path, 'pinned_id_pleroma.txt')):\n with open(os.path.join(self.user_path, 'pinned_id_pleroma.txt'), 'r') as file:\n previous_pinned_post_id = file.readline().rstrip()\n unpin_url = self.pleroma_base_url + '/api/v1/statuses/' + previous_pinned_post_id + '/unpin'\n response = requests.post(unpin_url, headers=self.header_pleroma)\n if not response.ok:\n response.raise_for_status()\n print(\"Unpinning previous:\\t\" + response.text)\n pin_url = self.pleroma_base_url + '/api/v1/statuses/' + id_post + '/pin'\n response = requests.post(pin_url, headers=self.header_pleroma)\n print(\"Pinning post:\\t\" + str(response.text))\n try:\n pin_id = json.loads(response.text)['id']\n except KeyError:\n pin_id = None\n pass\n return pin_id\n\n def update_pleroma(self):\n \"\"\"Update the Pleroma user info with the one retrieved from Twitter when the User object was instantiated.\n This includes:\n \n * Profile image (if exists)\n * Banner image (if exists)\n * Bio text\n * Screen name\n * Additional metadata fields\n \n :returns: None \n \"\"\"\n # Get the biggest resolution for the profile picture (400x400) instead of 'normal'\n if self.profile_image_url:\n profile_img_big = re.sub(r\"normal\", \"400x400\", self.profile_image_url)\n response = requests.get(profile_img_big, stream=True)\n if not response.ok:\n response.raise_for_status()\n response.raw.decode_content = True\n with open(self.avatar_path, 'wb') as outfile:\n shutil.copyfileobj(response.raw, outfile)\n\n if self.profile_banner_url:\n response = requests.get(self.profile_banner_url, stream=True)\n if not response.ok:\n response.raise_for_status()\n response.raw.decode_content = True\n with open(self.header_path, 'wb') as outfile:\n shutil.copyfileobj(response.raw, outfile)\n\n # Set it on Pleroma\n cred_url = self.pleroma_base_url + '/api/v1/accounts/update_credentials'\n\n # Construct fields\n fields = []\n for field_item in self.fields:\n field = (field_item['name'], field_item['value'])\n fields.append(field)\n data = {\"note\": self.bio_text,\n \"display_name\": self.display_name}\n\n if self.profile_image_url:\n data.update({\"avatar\": self.avatar_path})\n\n if self.profile_banner_url:\n data.update({\"header\": self.header_path})\n\n if len(fields) > 4:\n raise Exception(\"Maximum number of metadata fields is 4. Exiting...\")\n for idx, (field_name, field_value) in enumerate(fields):\n data['fields_attributes[' + str(idx) + '][name]'] = field_name\n data['fields_attributes[' + str(idx) + '][value]'] = field_value\n\n if self.profile_image_url:\n avatar = open(self.avatar_path, 'rb')\n avatar_mime_type = guess_type(self.avatar_path)\n timestamp = str(datetime.now().timestamp())\n avatar_file_name = \"pleromapyupload_\" + timestamp + \"_\" + random_string(10) + mimetypes.guess_extension(\n avatar_mime_type)\n\n if self.profile_banner_url:\n header = open(self.header_path, 'rb')\n header_mime_type = guess_type(self.header_path)\n header_file_name = \"pleromapyupload_\" + timestamp + \"_\" + random_string(10) + mimetypes.guess_extension(\n header_mime_type)\n\n files = {}\n\n if self.profile_image_url:\n files.update(\n {\"avatar\": (avatar_file_name, avatar, avatar_mime_type)})\n if self.profile_banner_url:\n files.update(\n {\"header\": (header_file_name, header, header_mime_type)})\n response = requests.patch(cred_url, data, headers=self.header_pleroma, files=files)\n if not response.ok:\n response.raise_for_status()\n print(\"Updating profile:\\t\" + str(response)) # for debugging\n return\n\n def replace_vars_in_str(self, text: str, var_name: str = None) -> str:\n \"\"\"Returns a string with \"{{ var_name }}\" replaced with var_name's value\n If no 'var_name' is provided, locals() (or self if it's an object method)\n will be used and all variables found in locals() (or object attributes)\n will be replaced with their values.\n\n :param text: String to be parsed, replacing \"{{ var_name }}\" with var_name's value. Multiple occurrences are\n supported.\n :type text: str\n :param var_name: Name of the variable to be replace. Multiple occurrences are supported. If not provided,\n locals() will be used and all variables will be replaced with their values.\n :type var_name: str\n\n :returns: A string with {{ var_name }} replaced with var_name's value.\n :rtype: str\n \"\"\"\n # Jinja-style string replacement i.e. vars encapsulated in {{ and }}\n if var_name is not None:\n matching_pattern = r'(?<=\\{{)( ' + var_name + r' )(?=\\}})'\n matches = re.findall(matching_pattern, text)\n else:\n matches = re.findall(r'(?<={{)(.*?)(?=}})', text)\n for match in matches:\n pattern = r'{{ ' + match.strip() + ' }}'\n # Get attribute value if it's a method\n # go for locals() if not, fallback to globals()\n try:\n value = getattr(self, match.strip())\n except NameError:\n try:\n value = locals()[match.strip()]\n except NameError:\n value = globals()[match.strip()]\n text = re.sub(pattern, value, text)\n return text\n\n\ndef guess_type(media_file: str) -> str:\n \"\"\"Try to guess what MIME type the given file is.\n\n :param media_file: The file to perform the guessing on\n :returns: the MIME type result of guessing\n :rtype: str\n \"\"\"\n mime_type = None\n try:\n mime_type = magic.from_file(media_file, mime=True)\n except AttributeError:\n mime_type = mimetypes.guess_type(media_file)[0]\n return mime_type\n\n\ndef random_string(length: int) -> str:\n \"\"\"Returns a string of random characters of length 'length'\n :param length: How long the string to return must be\n :type length: int\n :returns: an alpha-numerical string of specified length with random characters\n :rtype: str\n \"\"\"\n return ''.join(random.choice(string.ascii_lowercase + string.digits) for _ in range(length))\n\n\ndef main():\n script_path = os.path.dirname(sys.argv[0])\n base_path = os.path.abspath(script_path)\n with open(os.path.join(base_path, 'config.yml'), 'r') as stream:\n config = yaml.safe_load(stream)\n user_dict = config['users']\n\n for user_item in user_dict:\n user = User(user_item, config)\n date_pleroma = user.get_date_last_pleroma_post()\n tweets = user.get_tweets()\n # Put oldest first to iterate them and post them in order\n tweets['data'].reverse()\n tweets_to_post = {'data': [], 'includes': tweets['includes']}\n # Get rid of old tweets\n for tweet in tweets['data']:\n created_at = tweet['created_at']\n date_twitter = datetime.strftime(datetime.strptime(created_at, '%Y-%m-%dT%H:%M:%S.000Z'),\n '%Y-%m-%d %H:%M:%S')\n if date_twitter > date_pleroma:\n tweets_to_post['data'].append(tweet)\n\n tweets_to_post = user.process_tweets(tweets_to_post)\n print('tweets:', tweets_to_post['data'])\n for tweet in tweets_to_post['data']:\n user.post_pleroma(tweet['id'], tweet['text'], tweet['polls'])\n time.sleep(2)\n # Pinned tweet\n print(\"Current pinned:\\t\" + str(user.pinned_tweet_id))\n if os.path.isfile(os.path.join(user.user_path, 'pinned_id.txt')):\n with open(os.path.join(user.user_path, 'pinned_id.txt'), 'r') as file:\n previous_pinned_tweet_id = file.readline().rstrip()\n else:\n previous_pinned_tweet_id = None\n print(\"Previous pinned:\\t\" + str(previous_pinned_tweet_id))\n if (user.pinned_tweet_id != previous_pinned_tweet_id) or \\\n ((user.pinned_tweet_id is not None) and (previous_pinned_tweet_id is None)):\n pinned_tweet = user._get_tweets(\"v2\", user.pinned_tweet_id)\n tweets_to_post = {'data': [pinned_tweet['data']], 'includes': tweets['includes']}\n tweets_to_post = user.process_tweets(tweets_to_post)\n id_post_to_pin = user.post_pleroma(user.pinned_tweet_id, tweets_to_post['data'][0]['text'], None)\n pleroma_pinned_post = user.pin_pleroma(id_post_to_pin)\n with open(os.path.join(user.user_path, 'pinned_id.txt'), 'w') as file:\n file.write(user.pinned_tweet_id + '\\n')\n if pleroma_pinned_post is not None:\n with open(os.path.join(user.user_path, 'pinned_id_pleroma.txt'), 'w') as file:\n file.write(pleroma_pinned_post + '\\n')\n\n if not arg == \"noProfile\":\n user.update_pleroma()\n # Clean-up\n shutil.rmtree(user.tweets_temp_path)\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "updateInfoPleroma.py", "file_name": "updateInfoPleroma.py", "file_ext": "py", "file_size_in_byte": 31754, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.argv", "line_number": 56, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 59, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 138, "usage_type": "call"}, {"api_name": "os.path", "line_number": 138, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 138, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 139, "usage_type": "call"}, {"api_name": "os.path", "line_number": 139, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 140, "usage_type": "call"}, {"api_name": "os.path", "line_number": 140, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 142, "usage_type": "call"}, {"api_name": "os.path", "line_number": 142, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 143, "usage_type": "call"}, {"api_name": "os.path", "line_number": 143, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 144, "usage_type": "call"}, {"api_name": "os.path", "line_number": 144, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path", "line_number": 145, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 146, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 147, "usage_type": "call"}, {"api_name": "os.path", "line_number": 147, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 148, "usage_type": "call"}, {"api_name": "os.path.isdir", "line_number": 149, "usage_type": "call"}, {"api_name": "os.path", "line_number": 149, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 150, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 168, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 171, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 195, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 198, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 203, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 206, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 251, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 254, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 268, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 271, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 273, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 273, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 273, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 276, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 276, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 276, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 276, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 287, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 291, "usage_type": "call"}, {"api_name": "json.decoder.JSONDecodeError", "line_number": 294, "usage_type": "name"}, {"api_name": "re.findall", "line_number": 318, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 320, "usage_type": "call"}, {"api_name": "re.finditer", "line_number": 326, "usage_type": "call"}, {"api_name": "requests.Session", "line_number": 330, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 335, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 338, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 341, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 345, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 347, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 365, "usage_type": "call"}, {"api_name": "os.path", "line_number": 365, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 366, "usage_type": "call"}, {"api_name": "os.path", "line_number": 366, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 367, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 381, "usage_type": "call"}, {"api_name": "mimetypes.guess_extension", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "shutil.copyfileobj", "line_number": 387, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 403, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 407, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 438, "usage_type": "call"}, {"api_name": "os.path", "line_number": 438, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 443, "usage_type": "call"}, {"api_name": "os.path", "line_number": 443, "usage_type": "attribute"}, {"api_name": "os.stat", "line_number": 444, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 444, "usage_type": "call"}, {"api_name": "os.path", "line_number": 444, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 445, "usage_type": "call"}, {"api_name": "os.path", "line_number": 445, "usage_type": "attribute"}, {"api_name": "datetime.datetime.now", "line_number": 446, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 446, "usage_type": "name"}, {"api_name": "mimetypes.guess_extension", "line_number": 448, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 451, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 455, "usage_type": "call"}, {"api_name": "json.decoder.JSONDecodeError", "line_number": 456, "usage_type": "name"}, {"api_name": "requests.post", "line_number": 476, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 480, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path", "line_number": 491, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 492, "usage_type": "call"}, {"api_name": "os.path", "line_number": 492, "usage_type": "attribute"}, {"api_name": "requests.post", "line_number": 495, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 500, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 503, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 523, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 524, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 529, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 532, "usage_type": "call"}, {"api_name": "shutil.copyfileobj", "line_number": 537, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 565, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 565, "usage_type": "name"}, {"api_name": "mimetypes.guess_extension", "line_number": 566, "usage_type": "call"}, {"api_name": "mimetypes.guess_extension", "line_number": 572, "usage_type": "call"}, {"api_name": "requests.patch", "line_number": 583, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 608, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 610, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 622, "usage_type": "call"}, {"api_name": "magic.from_file", "line_number": 635, "usage_type": "call"}, {"api_name": "mimetypes.guess_type", "line_number": 637, "usage_type": "call"}, {"api_name": "random.choice", "line_number": 648, "usage_type": "call"}, {"api_name": "string.ascii_lowercase", "line_number": 648, "usage_type": "attribute"}, {"api_name": "string.digits", "line_number": 648, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 652, "usage_type": "call"}, {"api_name": "os.path", "line_number": 652, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 652, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 653, "usage_type": "call"}, {"api_name": "os.path", "line_number": 653, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 654, "usage_type": "call"}, {"api_name": "os.path", "line_number": 654, "usage_type": "attribute"}, {"api_name": "yaml.safe_load", "line_number": 655, "usage_type": "call"}, {"api_name": "datetime.datetime.strftime", "line_number": 668, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 668, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 668, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 677, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 680, "usage_type": "call"}, {"api_name": "os.path", "line_number": 680, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 680, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 681, "usage_type": "call"}, {"api_name": "os.path", "line_number": 681, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 693, "usage_type": "call"}, {"api_name": "os.path", "line_number": 693, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 696, "usage_type": "call"}, {"api_name": "os.path", "line_number": 696, "usage_type": "attribute"}, {"api_name": "shutil.rmtree", "line_number": 702, "usage_type": "call"}]} +{"seq_id": "338213791", "text": "#!/usr/bin/env python\nimport csv\nimport pymysql\nconnection = pymysql.connect(host='localhost', port=8889, user='root', passwd='root', db='ad_trends', unix_socket=\"/Applications/MAMP/tmp/mysql/mysql.sock\")\ncursor = connection.cursor()\n\n\ndef insert_one(table, column, value):\n sql = \"insert ignore into {table}({column}) values({value});\".format(\n table=table, column=column, value=value.replace(\"'\", \"\")\n )\n try:\n cursor.execute(sql)\n except Exception as e:\n print(sql)\n print(e)\n\n\ndef insert_main(row):\n row = {k: v.replace(\"'\", \"\") for k, v in row.items()}\n sql = \"\"\"\n insert into raw_data\n (day, ad_groups_id, campaign_id, clicks, avg_position, cost_conv, ctr, impressions, avg_cpc)\n values\n (\n '{day}',\n (select id from ad_groups where name = '{ad_group}'),\n (select id from campaigns where name = '{campaign}'),\n {clicks},\n '{avg_position}',\n '{cost_conv}',\n '{ctr}',\n {impressions},\n '{avg_cpc}'\n );\n \"\"\".format(**row)\n return sql\n\n\nfilename = \"../data/avg.-pos-better-than-4-cleaned.csv\"\nwith open(filename, \"r\") as f:\n reader = csv.DictReader(f, delimiter=\",\", quotechar='\"')\n for row in reader:\n del(row['account'])\n\n # campaign\n insert_one(\"campaigns\", \"name\", \"'{}'\".format(row[\"campaign\"].strip()))\n\n # ad_groups\n insert_one(\"ad_groups\", \"name\", \"'{}'\".format(row[\"ad_group\"].strip()))\n\n # main data table\n row['ctr'] = row['ctr'].replace('%', '')\n sql = insert_main(row)\n try:\n cursor.execute(sql)\n connection.commit()\n except Exception as e:\n connection.rollback()\n print(sql)\n print(e)\n\n\n\n\n\nconnection.close()\n", "sub_path": "develop/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 1749, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pymysql.connect", "line_number": 4, "usage_type": "call"}, {"api_name": "csv.DictReader", "line_number": 42, "usage_type": "call"}]} +{"seq_id": "310650172", "text": "import matplotlib.pyplot as plt\n\n# plot the average idp of R/D's in each congress \ndef plotidp():\n\txsr = []; xsd = []; ysr = []; ysd = []\n\tfin = open('./idp_year.txt')\n\tlines = fin.readlines()\n\tfor line in lines:\n\t\tls = line.split('\\t')\n\t\tyear = 1790 + int(ls[0]) * 2\n\t\taver_r = float(ls[1])\n\t\taver_d = float(ls[2])\n\t\tif aver_r != 0:\n\t\t\txsr.append(year)\n\t\t\tysr.append(aver_r)\n\t\tif aver_d != 0:\n\t\t\txsd.append(year)\n\t\t\tysd.append(aver_d)\n\tfin.close()\n\n\tfig = plt.figure()\n\tline_r = plt.scatter(xsr, ysr, c = ['r' for u in ysr], label = 'Line 1')\n\tline_d = plt.scatter(xsd, ysd, c = ['b' for u in ysd], label = 'Line 2')\n\tplt.xlabel('Year')\n\tplt.ylabel('Average Ideal Point')\n\tplt.xlim([1789,2020])\n\tplt.legend([line_r, line_d], ['Republican', 'Democrat'], loc = 2)\n\tplt.show()\n\tfig.savefig('idp_each_year_no_constraint.png')\n\n\n# plot the number of R/D's in each congress\ndef plotnum():\n\txs = []; ysr = []; ysd = []; ystotal = []\n\tfin = open('./idp_year.txt')\n\tlines = fin.readlines()\n\tfor line in lines:\n\t\tls = line.split('\\t')\n\t\tyear = 1790 + int(ls[0]) * 2\n\t\taver_r = float(ls[3])\n\t\taver_d = float(ls[4])\n\t\txs.append(year)\n\t\tysr.append(aver_r)\n\t\tysd.append(aver_d)\n\t\tystotal.append(aver_r + aver_d)\n\tfin.close()\n\n\tfig = plt.figure()\n\tline_r = plt.scatter(xs, ysr, c = ['r' for u in ysr], alpha = 1)\n\tline_d = plt.scatter(xs, ysd, c = ['b' for u in ysd], alpha = 1)\n\tline_g = plt.scatter(xs, ystotal, c = ['g' for u in ystotal], alpha = 0.3)\n\tplt.xlim([1789,2020])\n\tplt.ylim([-10,600])\n\tplt.xlabel('Year')\n\tplt.ylabel('Number of Lawmakers')\n\tplt.legend([line_r, line_d, line_g], ['Republican', 'Democrat', 'Sum'], loc = 2)\n#\tplt.show()\n\tfig.savefig('idp_num_each_year_no_constraint.png')\n\n\nif __name__ == '__main__':\n\tplotidp()\n\tplotnum()\n\n", "sub_path": "tmp/plot_ave.py", "file_name": "plot_ave.py", "file_ext": "py", "file_size_in_byte": 1739, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 21, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 22, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 22, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 23, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 23, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 24, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 24, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlim", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylim", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}]} +{"seq_id": "418217860", "text": "# This script uses the Duffy node management api to get fresh machines to run\n# your CI tests on. Once allocated you will be able to ssh into that machine\n# as the root user and setup the environ\n#\n# XXX: You need to add your own api key below, and also set the right cmd= line\n# needed to run the tests\n#\n# Please note, this is a basic script, there is no error handling and there are\n# no real tests for any exceptions. Patches welcome!\n\nimport json\nimport os\nimport urllib\nimport sys\n\nfrom ci.lib import _print, run_cmd, provision\n\nurl_base = os.environ.get('URL_BASE')\napi = os.environ.get('API')\nver = \"7\"\narch = \"x86_64\"\ncount = 4\n\nrepo_url = os.environ.get('ghprbAuthorRepoGitUrl') or \\\n os.environ.get('GIT_URL')\nrepo_branch = os.environ.get('ghprbSourceBranch') or \\\n os.environ.get('ghprbTargetBranch') or 'master'\n\n\ndef get_nodes(ver=\"7\", arch=\"x86_64\", count=4):\n get_nodes_url = \"%s/Node/get?key=%s&ver=%s&arch=%s&count=%s\" % (\n url_base, api, ver, arch, count)\n\n resp = urllib.urlopen(get_nodes_url).read()\n data = json.loads(resp)\n with open('env.properties', 'a') as f:\n f.write('DUFFY_SSID=%s' % data['ssid'])\n f.close()\n _print(resp)\n return data['hosts']\n\n\ndef fail_nodes():\n with open('env.properties') as f:\n s = f.read()\n\n ssid = None\n for line in s.splitlines():\n key, value = line.split('=')\n if key == 'DUFFY_SSID':\n ssid = value\n break\n\n fail_nodes_url = \"{url_base}/Node/fail?key={key}&ssid={ssid}\".format(\n url_base=url_base, key=api, ssid=ssid)\n resp = urllib.urlopen(fail_nodes_url).read()\n _print(resp)\n\n\ndef print_nodes():\n with open('env.properties') as f:\n s = f.read()\n\n _print('\\n'.join(s.splitlines()[3:]))\n\n\ndef generate_ansible_inventory(jenkins_master_host, jenkins_slave_host,\n openshift_host, scanner_host):\n\n ansible_inventory = (\"\"\"\n[all:children]\njenkins_master\njenkins_slaves\nopenshift\nscanner_worker\n\n[jenkins_master]\n{jenkins_master_host}\n\n[jenkins_slaves]\n{jenkins_slave_host}\n\n[openshift]\n{openshift_host}\n\n[scanner_worker]\n{scanner_host}\n\n[all:vars]\npublic_registry= {jenkins_slave_host}\ncopy_ssl_certs=true\nopenshift_startup_delay=150\nbeanstalk_server={openshift_host}\ntest=true\ncccp_source_repo={repo_url}\ncccp_source_branch={repo_branch}\njenkins_public_key_file = jenkins.key.pub\n\n[jenkins_master:vars]\njenkins_private_key_file = jenkins.key\ncccp_index_repo=https://github.com/rtnpro/container-index.git\noc_slave={jenkins_slave_host}\"\"\").format(\n jenkins_master_host=jenkins_master_host,\n jenkins_slave_host=jenkins_slave_host,\n openshift_host=openshift_host,\n repo_url=repo_url,\n repo_branch=repo_branch,\n scanner_host=scanner_host)\n\n with open('hosts', 'w') as f:\n f.write(ansible_inventory)\n\n\ndef setup_controller(controller):\n # provision controller\n run_cmd(\n \"scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no \"\n \"~/.ssh/id_rsa root@%s:/root/.ssh/id_rsa\" % controller\n )\n\n run_cmd(\n \"yum install -y git epel-release && \"\n \"yum install -y python-pip && \"\n \"yum install -y gcc libffi-devel python-devel openssl-devel && \"\n \"yum install -y python2-jenkins-job-builder && \"\n \"pip install ansible==2.1.1\",\n host=controller)\n\n run_cmd(\n \"scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -r \"\n \"./ root@%s:/root/container-pipeline-service\" % controller)\n\n\ndef run():\n os.environ.pop('CCCP_CI_PROVISIONED', None)\n os.environ.pop('CCCP_CI_HOSTS', None)\n\n nodes = get_nodes(count=5)\n\n jenkins_master_host = nodes[0]\n jenkins_slave_host = nodes[1]\n openshift_host = nodes[2]\n scanner_host = nodes[3]\n controller = nodes.pop()\n\n nodes_env = (\n \"\\nJENKINS_MASTER_HOST=%s\\n\"\n \"JENKINS_SLAVE_HOST=%s\\n\"\n \"OPENSHIFT_HOST=%s\\n\"\n \"CONTROLLER=%s\\n\"\n \"SCANNER_HOST=%s\\n\"\n ) % (jenkins_master_host, jenkins_slave_host,\n openshift_host, controller, scanner_host)\n\n with open('env.properties', 'a') as f:\n f.write(nodes_env)\n\n hosts_data = {\n 'openshift': {\n 'host': openshift_host,\n 'remote_user': 'root'\n },\n 'jenkins_master': {\n 'host': jenkins_master_host,\n 'remote_user': 'root'\n },\n 'jenkins_slave': {\n 'host': jenkins_slave_host,\n 'remote_user': 'root'\n },\n 'controller': {\n 'host': controller,\n 'user': 'root',\n 'workdir': '/root/container-pipeline-service',\n # relative to this workdir\n 'inventory_path': 'hosts'\n }\n }\n\n _print(hosts_data)\n\n generate_ansible_inventory(jenkins_master_host,\n jenkins_slave_host,\n openshift_host,\n scanner_host)\n\n run_cmd('iptables -F', host=openshift_host)\n run_cmd('iptables -F', host=jenkins_slave_host)\n\n setup_controller(controller)\n\n provision(hosts_data['controller'])\n\n os.environ['CCCP_CI_PROVISIONED'] = \"true\"\n\n os.environ['CCCP_CI_HOSTS'] = json.dumps(hosts_data)\n\n run_cmd('~/venv/bin/nosetests ci/tests', stream=True)\n\n os.environ.pop('CCCP_CI_PROVISIONED', None)\n os.environ.pop('CCCP_CI_HOSTS', None)\n\n\nif __name__ == '__main__':\n try:\n run()\n except Exception as e:\n _print('Build failed: %s' % e)\n _print('Reserving nodes for debugging...')\n fail_nodes()\n _print('=' * 10 + 'Node Info' + '=' * 10)\n print_nodes()\n sys.exit(1)\n", "sub_path": "ci/cccp_ci.py", "file_name": "cccp_ci.py", "file_ext": "py", "file_size_in_byte": 5695, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.environ.get", "line_number": 18, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 18, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 19, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 24, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 24, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 25, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 25, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 26, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.environ.get", "line_number": 27, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 27, "usage_type": "attribute"}, {"api_name": "urllib.urlopen", "line_number": 34, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 35, "usage_type": "call"}, {"api_name": "ci.lib._print", "line_number": 39, "usage_type": "call"}, {"api_name": "urllib.urlopen", "line_number": 56, "usage_type": "call"}, {"api_name": "ci.lib._print", "line_number": 57, "usage_type": "call"}, {"api_name": "ci.lib._print", "line_number": 64, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 116, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 121, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 129, "usage_type": "call"}, {"api_name": "os.environ.pop", "line_number": 135, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 135, "usage_type": "attribute"}, {"api_name": "os.environ.pop", "line_number": 136, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 136, "usage_type": "attribute"}, {"api_name": "ci.lib._print", "line_number": 180, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 187, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 188, "usage_type": "call"}, {"api_name": "ci.lib.provision", "line_number": 192, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 194, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 196, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 196, "usage_type": "call"}, {"api_name": "ci.lib.run_cmd", "line_number": 198, "usage_type": "call"}, {"api_name": "os.environ.pop", "line_number": 200, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 200, "usage_type": "attribute"}, {"api_name": "os.environ.pop", "line_number": 201, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 201, "usage_type": "attribute"}, {"api_name": "ci.lib._print", "line_number": 208, "usage_type": "call"}, {"api_name": "ci.lib._print", "line_number": 209, "usage_type": "call"}, {"api_name": "ci.lib._print", "line_number": 211, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 213, "usage_type": "call"}]} +{"seq_id": "128296106", "text": "import json\n\ninput_file = 'BLXXXgrd02/useful_data.txt'\ntarget_file_prefix = 'BLXXXgrd02/useful_part'\n\n# Read in python list of utterances and their words\nwith open(input_file, 'r') as f:\n\tutterances = json.loads(f.read())\n\nprint(\"Collected useful data\")\n\n# Convert json output from unicode to string\nutterances = [[str(file[0]), [str(word).lower() for word in file[1]]] for file in utterances]\n\n# Initialise separate lists for each part\npart_utts_list = [[], [], [], [], []]\n\n\n# Sort utterances into each correct part list\n# Remove , 'sp' and tokens\n# map %hesitation% to 'um'\n\nfor item in utterances:\n\tfileName = item[0]\n\tsentence = ''\n\t\n\t# Determine part\n\tpart_letter = fileName[23]\n\tpart_num = 100\n\t\n\tif part_letter == 'A':\n\t\tpart_num = 0\n\telif part_letter == 'B':\n\t\tpart_num = 1\n\telif part_letter == 'C':\n\t\tpart_num = 2\n\telif part_letter == 'D':\n\t\tpart_num = 3\n\telif part_letter == 'E':\n\t\tpart_num = 4\n\telse:\n\t\tprint(\"part not in range\")\n\t\tcontinue\n\n\tfor word in item[1]:\n\t\tif word == '' or word == 'sil' or word == 'sp':\n\t\t\tcontinue\n\t\telif word == '%hesitation%':\n\t\t\tnew_word = 'um'\n\t\telif '%partial%' in word:\n\t\t\tnew_word = word[:-10]\n\t\telse:\n\t\t\tnew_word = word\n\t\tsentence = sentence + ' ' + new_word\n\n\tpart_utts_list[part_num].append([fileName, sentence])\n\n\n# Output the part lists to separate files\nprint(\"Writing to files\")\n\nfor i in range(5):\n\ttarget_file = target_file_prefix + str(i+1) + '.txt'\n\twith open(target_file, 'w') as f:\n\t\tf.truncate(0)\n\t\tprint(len(part_utts_list[i]))\n\t\tf.write(json.dumps(part_utts_list[i]))\n\n\n\t\n\t\t\t\n", "sub_path": "data/make_part_data.py", "file_name": "make_part_data.py", "file_ext": "py", "file_size_in_byte": 1551, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "json.loads", "line_number": 8, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 67, "usage_type": "call"}]} +{"seq_id": "428491392", "text": "# Import packages\r\nfrom scipy.spatial import distance as dist\r\nimport numpy as np\r\nimport dlib\r\nimport cv2\r\nfrom cv2 import cv2\r\n\r\nimage = cv2.imread('Images/coffee_break.png') # Load input (coffee break) image\r\nimage = cv2.resize(image, (100,100)) # Resize image to 100x100\r\n\r\n'''\r\nEye Aspect Ratio (E.A.R.)\r\nFunction to calculate eye aspect ratio as in paper :\r\n\"Real-Time Eye Blink Detection using Facial Landmarks [Soukupova, Cech]\" \r\nLandmarks | 0 1 2 3 4 5\r\n Left Eye : [36,37,38,39,40,41]\r\nRight Eye : [42,43,44,45,46,47]\r\n'''\r\ndef eye_aspect_ratio(eye):\r\n # Vertical distances\r\n dist1 = dist.euclidean(eye[1], eye[5]) # P2-P6\r\n dist2 = dist.euclidean(eye[2], eye[4]) # P3-P5\r\n # Horiontal distance\r\n dist3 = dist.euclidean(eye[0], eye[3]) # P1-P4\r\n\r\n # Eye Aspect Ratio (E.A.R.)\r\n ear = (dist1 + dist2) / (2.0 * dist3)\r\n\r\n return ear\r\n\r\n\r\n'''\r\nLips Aspect Ratio (L.A.R.)\r\nFunction to calculate lips aspect ratio in the same way as in E.A.R.\r\nLandmarks | 0 1 2 3 4 5 6 7\r\n Lips : [60,61,62,63,64,65,66,67]\r\n'''\r\ndef lips_aspect_ratio(lips):\r\n # Vertical distance\r\n dist1 = dist.euclidean(lips[2], lips[6]) # L3-L7\r\n # Horiontal distance\r\n dist2 = dist.euclidean(lips[0], lips[4]) # L1-L5\r\n\r\n # Lips Aspect Ratio (L.A.R.)\r\n lar = float(dist1/dist2)\r\n\r\n return lar\r\n\r\n\r\n'''\r\nFacial Landmarks for any face part\r\nFunction to calculate facial landmark point coordinates (x,y),\r\ndraw them on frame and return a numpy array with the corresponding points\r\n'''\r\ndef draw_landmarks(face_part, landmarks):\r\n landmarks_list = []\r\n for point in face_part:\r\n x, y = landmarks.part(point).x, landmarks.part(point).y\r\n landmarks_list.append([x,y])\r\n cv2.circle(frame, (x,y), 2, (0,0,255), -1)\r\n \r\n return np.array(landmarks_list)\r\n\r\n\r\n# DLIB - Face Detector\r\ndetector = dlib.get_frontal_face_detector()\r\n# DLIB - Predictor\r\npredictor = dlib.shape_predictor('Models/shape_predictor_68_face_landmarks.dat')\r\n\r\n# Video Capture\r\ncap = cv2.VideoCapture(1)\r\n\r\n# Text settings\r\nfont = cv2.FONT_HERSHEY_SIMPLEX\r\nfont_scale = 0.7\r\n\r\n# Initializations\r\nframes = 0\r\n\r\n# ear & lar, threshold values\r\near_thresh = 0.3\r\nlar_thresh = 0.5\r\n\r\n# Blink initializations\r\nblink_counter, total_blinks = 0, 0\r\n# Yawn initializations\r\nyawn_counter, total_yawns = 0, 0\r\n\r\nwhile True:\r\n _, frame = cap.read()\r\n frame = cv2.flip(frame, 1) # May not be necessary\r\n h, w = frame.shape[: 2] # Height and Width of frame\r\n \r\n frames += 1\r\n\r\n # Grayscale\r\n gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\r\n\r\n # Detect faces in the gray frame\r\n faces = detector(gray, 0)\r\n\r\n # Loop through each face\r\n for face in faces:\r\n # Determine facial landmarks\r\n facial_landmarks = predictor(gray, face)\r\n\r\n # Landmark indexes for eyes and lips\r\n left_eye = [36,37,38,39,40,41]\r\n right_eye = [42,43,44,45,46,47]\r\n \r\n lips = [60,61,62,63,64,65,66,67]\r\n\r\n # Convert to numpy array the above lists and\r\n # draw the corresponding facial landmark points on frame\r\n left_eye_points = draw_landmarks(left_eye, facial_landmarks)\r\n right_eye_points = draw_landmarks(right_eye, facial_landmarks)\r\n\r\n lips_points = draw_landmarks(lips, facial_landmarks)\r\n\r\n # Find and draw the convex hulls of left and right eye, and lips\r\n left_eye_hull = cv2.convexHull(left_eye_points) \r\n cv2.drawContours(frame, [left_eye_hull], -1, (0, 255, 0), 1)\r\n \r\n right_eye_hull = cv2.convexHull(right_eye_points)\r\n cv2.drawContours(frame, [right_eye_hull], -1, (0, 255, 0), 1)\r\n\r\n lips_hull = cv2.convexHull(lips_points)\r\n cv2.drawContours(frame, [lips_hull], -1, (0, 255, 0), 1)\r\n\r\n # Calculate E.A.R. and L.A.R.\r\n left_ear = eye_aspect_ratio(left_eye_points) # Left eye aspect ratio\r\n right_ear = eye_aspect_ratio(right_eye_points) # Right eye aspect ratio\r\n ear = (left_ear + right_ear) / 2.0 # Average eye aspect ratio\r\n cv2.putText(frame, \"E.A.R. : {:.2f}\".format(ear), (10,30), font, font_scale, (0,0,255), 2)\r\n\r\n lar = lips_aspect_ratio(lips_points) # Lips aspect ratio\r\n cv2.putText(frame, \"L.A.R. : {:.2f}\".format(lar), (10,90), font, font_scale, (0,0,255), 2)\r\n\r\n # Check for blinks or yawns\r\n # BLINK\r\n if ear < ear_thresh:\r\n blink_counter += 1\r\n else:\r\n if blink_counter > 3:\r\n total_blinks += 1\r\n blink_counter = 0\r\n cv2.putText(frame, \"Blinks: {}\".format(total_blinks), (10, 50), font, font_scale, (0, 0, 255), 2)\r\n\r\n # YAWN\r\n if lar > lar_thresh:\r\n yawn_counter += 1\r\n else:\r\n if yawn_counter > 1:\r\n total_yawns += 1\r\n yawn_counter = 0\r\n cv2.putText(frame, \"Yawns: {}\".format(total_yawns), (10, 110), font, font_scale, (0, 0, 255), 2)\r\n \r\n # Drowsiness Detection\r\n if total_yawns > 2 or total_blinks > 3:\r\n frame[20:120, w-120:w-20] = image # Show coffee break image\r\n cv2.putText(frame, \"ALERT\", (w-120, 160), font, 1.2, (0, 0, 255), 4)\r\n\r\n cv2.imshow('Frame', frame)\r\n key = cv2.waitKey(1)\r\n if key == 27:\r\n break\r\n if key==ord('r') or key==ord('R'):\r\n total_blinks, total_yawns = 0, 0 # Reset calculations by pressing 'r' or 'R'\r\n\r\ncap.release()\r\ncv2.destroyAllWindows()", "sub_path": "yawn detection/random.py", "file_name": "random.py", "file_ext": "py", "file_size_in_byte": 5656, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cv2.cv2.imread", "line_number": 8, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 8, "usage_type": "name"}, {"api_name": "cv2.cv2.resize", "line_number": 9, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 9, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 21, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 21, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 22, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 22, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 24, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 24, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 40, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 40, "usage_type": "name"}, {"api_name": "scipy.spatial.distance.euclidean", "line_number": 42, "usage_type": "call"}, {"api_name": "scipy.spatial.distance", "line_number": 42, "usage_type": "name"}, {"api_name": "cv2.cv2.circle", "line_number": 60, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 60, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "dlib.get_frontal_face_detector", "line_number": 66, "usage_type": "call"}, {"api_name": "dlib.shape_predictor", "line_number": 68, "usage_type": "call"}, {"api_name": "cv2.cv2.VideoCapture", "line_number": 71, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 71, "usage_type": "name"}, {"api_name": "cv2.cv2.FONT_HERSHEY_SIMPLEX", "line_number": 74, "usage_type": "attribute"}, {"api_name": "cv2.cv2", "line_number": 74, "usage_type": "name"}, {"api_name": "cv2.cv2.flip", "line_number": 91, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 91, "usage_type": "name"}, {"api_name": "cv2.cv2.cvtColor", "line_number": 97, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 97, "usage_type": "name"}, {"api_name": "cv2.cv2.COLOR_BGR2GRAY", "line_number": 97, "usage_type": "attribute"}, {"api_name": "cv2.cv2.convexHull", "line_number": 121, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 121, "usage_type": "name"}, {"api_name": "cv2.cv2.drawContours", "line_number": 122, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 122, "usage_type": "name"}, {"api_name": "cv2.cv2.convexHull", "line_number": 124, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 124, "usage_type": "name"}, {"api_name": "cv2.cv2.drawContours", "line_number": 125, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 125, "usage_type": "name"}, {"api_name": "cv2.cv2.convexHull", "line_number": 127, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 127, "usage_type": "name"}, {"api_name": "cv2.cv2.drawContours", "line_number": 128, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 128, "usage_type": "name"}, {"api_name": "cv2.cv2.putText", "line_number": 134, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 134, "usage_type": "name"}, {"api_name": "cv2.cv2.putText", "line_number": 137, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 137, "usage_type": "name"}, {"api_name": "cv2.cv2.putText", "line_number": 147, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 147, "usage_type": "name"}, {"api_name": "cv2.cv2.putText", "line_number": 156, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 156, "usage_type": "name"}, {"api_name": "cv2.cv2.putText", "line_number": 161, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 161, "usage_type": "name"}, {"api_name": "cv2.cv2.imshow", "line_number": 163, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 163, "usage_type": "name"}, {"api_name": "cv2.cv2.waitKey", "line_number": 164, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 164, "usage_type": "name"}, {"api_name": "cv2.cv2.destroyAllWindows", "line_number": 171, "usage_type": "call"}, {"api_name": "cv2.cv2", "line_number": 171, "usage_type": "name"}]} +{"seq_id": "86074383", "text": "from mpl_toolkits.mplot3d import Axes3D\nfrom matplotlib import cm\nfrom matplotlib.ticker import LinearLocator, FormatStrFormatter\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport math\n\nfig = plt.figure()\nax = fig.gca(projection='3d')\nH = 5\nX = np.arange(-5, 6, 1)\nY = np.arange(-5, 6, 1)\nX, Y = np.meshgrid(X, Y)\nR = np.sqrt(X**2 + Y**2)\nZ = H-R\n#surf = ax.plot_surface(X, Y, Z, rstride=1, cstride=1, cmap=cm.coolwarm, linewidth=0, antialiased=False)\n#ax.set_zlim(-1.01, 1.01)\nax.set_zlim(0,6)\n\nax.zaxis.set_major_locator(LinearLocator(10))\nax.zaxis.set_major_formatter(FormatStrFormatter('%.02f'))\n\n\ndef put2(x,y,depth):\n\tresX,resY,resZ=[],[],[]\n\ttheta = 50*math.pi/180\n\tRot = np.matrix([[math.cos(theta), -math.sin(theta)],[math.sin(theta), math.cos(theta)]])\n\tK = 0.25\n\ta = 0.2\n\tif depth < 1: return\n\tarea = depth**3*10\n\tfor d in range(0,360*2,60):\n\t\tt = d * math.pi/180\n\t\tr = K * math.exp(t*a)\n\t\tm = np.matrix((r*math.cos(t), r*math.sin(t))).T\n\t\tfor k in range(0,360,50):\n\t\t\tu = x+m[0]\n\t\t\tv = y+m[1]\n\t\t\tresX.append(float(u))\n\t\t\tresY.append(float(v))\n\t\t\tresZ.append(H-math.sqrt(u**2+v**2)+0.3)\n\t\t\t#c = ax.scatter(u,v, H-math.sqrt(u**2+v**2)+0.3,c=2, s=area, cmap=plt.get_cmap(\"Greens\"), norm=mpl.colors.Normalize(vmin=0,vmax=3))\n\t\t\tm = Rot * m\n\t\t\t# x,y = u,v and go back\n\n\treturn resX,resY,resZ\nx,y,z=put2(0,0,2)\n\nimport rodrigues\n \nax.scatter(x,y,z, s=100, cmap=plt.get_cmap(\"Greens\"), norm=mpl.colors.Normalize(vmin=0,vmax=3))\nM = rodrigues.rrot(np.matrix([x,y,z]), np.matrix('1;1;1') / math.sqrt(3), math.pi)\nM = np.squeeze(np.asarray(M))\n#ax.scatter(x,y,z, s=100, cmap=plt.get_cmap(\"Greens\"), norm=mpl.colors.Normalize(vmin=0,vmax=3))\nax.scatter(M[0],M[1],M[2], s=100, cmap=plt.get_cmap(\"Greens\"), norm=mpl.colors.Normalize(vmin=0,vmax=3))\n\nplt.show()\n", "sub_path": "draw3dsurface.py", "file_name": "draw3dsurface.py", "file_ext": "py", "file_size_in_byte": 1796, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.pyplot.figure", "line_number": 9, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 9, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 14, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 15, "usage_type": "call"}, {"api_name": "matplotlib.ticker.LinearLocator", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.ticker.FormatStrFormatter", "line_number": 22, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 27, "usage_type": "attribute"}, {"api_name": "numpy.matrix", "line_number": 28, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 28, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 28, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 34, "usage_type": "attribute"}, {"api_name": "math.exp", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 36, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 36, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 36, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 52, "usage_type": "attribute"}, {"api_name": "rodrigues.rrot", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.matrix", "line_number": 53, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 53, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 53, "usage_type": "attribute"}, {"api_name": "numpy.squeeze", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.asarray", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.get_cmap", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 56, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}]} +{"seq_id": "140378953", "text": "# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Jan 13 10:34:59 2021\r\n\r\n@author: felipecas\r\n\"\"\"\r\n\r\nfrom __future__ import print_function\r\nfrom ortools.linear_solver import pywraplp\r\n\r\ndef WyndorGlassCo():\r\n \"\"\"Wyndor Glass Co\"\"\"\r\n #Initiate a Glop solver, naming it WyndorGlassCo\r\n solver = pywraplp.Solver('WyndorGlassCo',\r\n pywraplp.Solver.GLOP_LINEAR_PROGRAMMING)\r\n \r\n #Create the two decision variables and let them take on non negative real values\r\n x1 = solver.NumVar(0, solver.infinity(), \"x1\")\r\n x2 = solver.NumVar(0, solver.infinity(), \"x2\")\r\n \r\n #Objective function\r\n objective = solver.Objective()\r\n objective.SetCoefficient(x1, 3)\r\n objective.SetCoefficient(x2, 5)\r\n objective.SetMaximization()\r\n\r\n #Add the constraints to the model\r\n \r\n #Constraint 1: Plant 1 constraint\r\n constraint1 = solver.Constraint(-solver.infinity(), 4)\r\n constraint1.SetCoefficient(x1, 1)\r\n constraint1.SetCoefficient(x2, 0)\r\n \r\n #Constraint 2: Plant 2 constraint\r\n constraint2 = solver.Constraint(-solver.infinity(), 12)\r\n constraint2.SetCoefficient(x1, 0)\r\n constraint2.SetCoefficient(x2, 2)\r\n \r\n #Constraint 3: Plant 3 constraint\r\n constraint3 = solver.Constraint(-solver.infinity(), 18)\r\n constraint3.SetCoefficient(x1, 3)\r\n constraint3.SetCoefficient(x2, 2)\r\n \r\n #Solve the problem\r\n solver.Solve()\r\n opt_solution = 3*x1.solution_value() + 5*x2.solution_value()\r\n print('Number of variables =', solver.NumVariables())\r\n print('Number of constraints =', solver.NumConstraints())\r\n \r\n #Print the value of each decision variable in the solution\r\n print('Solution: ')\r\n print('x1 =', x1.solution_value())\r\n print('x2 =', x2.solution_value())\r\n \r\n #Print the value of the objective function in the solution\r\n print('Optimal objective value =', opt_solution)\r\n \r\nWyndorGlassCo()", "sub_path": "Hillier_Example_3_1_OR_Tools.py", "file_name": "Hillier_Example_3_1_OR_Tools.py", "file_ext": "py", "file_size_in_byte": 1922, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "ortools.linear_solver.pywraplp.Solver", "line_number": 14, "usage_type": "call"}, {"api_name": "ortools.linear_solver.pywraplp", "line_number": 14, "usage_type": "name"}, {"api_name": "ortools.linear_solver.pywraplp.Solver", "line_number": 15, "usage_type": "attribute"}, {"api_name": "ortools.linear_solver.pywraplp", "line_number": 15, "usage_type": "name"}]} +{"seq_id": "649083088", "text": "from json_parser import JsonParser\nfrom data import DataType, DataStore\nfrom multiprocessing import Process\nfrom multiprocessing.dummy import Pool\nimport csv\nimport requests\nfrom pprint import pprint\n\nclass Market(object):\n \"\"\"\n This class handles gathering general market data from the IEX Trading API v1.0\n \"\"\"\n #constants\n PATH_SYMBOLS = \"symbols/\"\n FILENAME_NASDAQ = \"nasdaq.csv\"\n FILENAME_NYSE = \"nyse.csv\"\n FILENAME_AMEX = \"amex.csv\"\n\n BATCH_LIMIT_IEX = 100\n API_URL_IEX = \"https://api.iextrading.com/1.0\"\n THREAD_COUNT = 30\n\n def __init__(self):\n self.symbols = self._read_symbols()\n\n #initialize datastores\n self.earnings_data = DataStore(\"earnings\", JsonParser.parse_hier1)\n self.quote_data = DataStore(\"quote\", JsonParser.parse_flat )\n self.financials_data = DataStore(\"financials\", JsonParser.parse_hier1)\n self.stats_data = DataStore(\"stats\", JsonParser.parse_flat )\n self.company_data = DataStore(\"company\", JsonParser.parse_flat )\n self.datastores = [\n self.earnings_data,\n self.quote_data,\n self.financials_data,\n self.stats_data,\n self.company_data,\n ]\n\n #populate data\n self.data_refresh()\n\n def data_refresh(self):\n symbol_batches = list(self._splits(self.symbols, self.BATCH_LIMIT_IEX))\n print(\"Market data refresh...\")\n\n #parallelize API requests with a thread pool. Python threads are a reasonable\n #choice here because the bottleneck is network latency, not processing power.\n pool = Pool(self.THREAD_COUNT)\n json_list = pool.map(self._api_request, symbol_batches)\n pool.close()\n pool.join()\n\n #combine json responses\n json_data = {}\n for json_response in json_list:\n json_data.update(json_response)\n\n #parallelize parsing with a process for each data type.\n #Use processes instead of threads because python's threads often don't take\n #advantage of multiple cores.\n processes = []\n for datastore in self.datastores:\n process = Process(target=datastore.parse_and_save,\n args=(json_data,),\n name=datastore.get_name())\n process.start()\n processes.append(process)\n print(\"Begin parsing '\" + process.name + \"'\")\n\n #wait for all processes to complete before proceeding\n for process in processes:\n process.join()\n print(\"Finished parsing \" + process.name)\n\n def get_most_active(self):\n request_base = \"/stock/market/list/mostactive\"\n #this response will make a fine addition to my collection\n response_json = requests.get(url=self.API_URL_IEX + request_base).json()\n if len(response_json) == 0:\n return None\n df = JsonParser.parse_list(response_json)\n print(df)\n return df\n\n def get_gainers(self):\n request_base = \"/stock/market/list/gainers\"\n #this response will make a fine addition to my collection\n response_json = requests.get(url=self.API_URL_IEX + request_base).json()\n if len(response_json) == 0:\n return None\n df = JsonParser.parse_list(response_json)\n print(df)\n return df\n\n def get_losers(self):\n request_base = \"/stock/market/list/losers\"\n #this response will make a fine addition to my collection\n response_json = requests.get(url=self.API_URL_IEX + request_base).json()\n if len(response_json) == 0:\n return None\n df = JsonParser.parse_list(response_json)\n print(df)\n return df\n\n def get_profile(self, symbol):\n pass\n\n def _api_request(self, batch):\n request_base = \"/stock/market/batch?\"\n datatypes = \",\".join([d.get_name() for d in self.datastores])\n #set up the parameters for API request\n params = dict(\n symbols = \",\".join(batch),\n types = datatypes\n )\n #this response will make a fine addition to my collection\n response_json = requests.get(url=self.API_URL_IEX + request_base, params=params).json()\n return response_json\n\n def _read_symbols(self):\n \"\"\"\n Read a list of symbols (tickers) into memory\n \"\"\"\n symbols = set()\n with open(self.PATH_SYMBOLS + self.FILENAME_NASDAQ, \"r\") as f:\n reader = csv.reader(f, delimiter=',')\n for row in reader:\n symbols.add(row[0])\n with open(self.PATH_SYMBOLS + self.FILENAME_NYSE, \"r\") as f:\n reader = csv.reader(f, delimiter=',')\n for row in reader:\n symbols.add(row[0])\n with open(self.PATH_SYMBOLS + self.FILENAME_AMEX, \"r\") as f:\n reader = csv.reader(f, delimiter=',')\n for row in reader:\n symbols.add(row[0])\n return sorted(list(symbols))\n\n def _splits(self, l, n):\n #yield successive n-sized splits from list l\n for i in range(0, len(l), n):\n yield l[i:i + n]\n", "sub_path": "market.py", "file_name": "market.py", "file_ext": "py", "file_size_in_byte": 5191, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "data.DataStore", "line_number": 27, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_hier1", "line_number": 27, "usage_type": "attribute"}, {"api_name": "json_parser.JsonParser", "line_number": 27, "usage_type": "name"}, {"api_name": "data.DataStore", "line_number": 28, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_flat", "line_number": 28, "usage_type": "attribute"}, {"api_name": "json_parser.JsonParser", "line_number": 28, "usage_type": "name"}, {"api_name": "data.DataStore", "line_number": 29, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_hier1", "line_number": 29, "usage_type": "attribute"}, {"api_name": "json_parser.JsonParser", "line_number": 29, "usage_type": "name"}, {"api_name": "data.DataStore", "line_number": 30, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_flat", "line_number": 30, "usage_type": "attribute"}, {"api_name": "json_parser.JsonParser", "line_number": 30, "usage_type": "name"}, {"api_name": "data.DataStore", "line_number": 31, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_flat", "line_number": 31, "usage_type": "attribute"}, {"api_name": "json_parser.JsonParser", "line_number": 31, "usage_type": "name"}, {"api_name": "multiprocessing.dummy.Pool", "line_number": 49, "usage_type": "call"}, {"api_name": "multiprocessing.Process", "line_number": 64, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 79, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_list", "line_number": 82, "usage_type": "call"}, {"api_name": "json_parser.JsonParser", "line_number": 82, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 89, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_list", "line_number": 92, "usage_type": "call"}, {"api_name": "json_parser.JsonParser", "line_number": 92, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 99, "usage_type": "call"}, {"api_name": "json_parser.JsonParser.parse_list", "line_number": 102, "usage_type": "call"}, {"api_name": "json_parser.JsonParser", "line_number": 102, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 118, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 127, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 131, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 135, "usage_type": "call"}]} +{"seq_id": "533782210", "text": "'''\nThis example demonstrates the usage of vtuIO.PVDIO at the results of the OGS-benchmark\nElliptic problem with Dirichlet-type boundary conditions,\ni.e. it requires the presence of:\n\nsquare_1e2_pcs_0.pvd\nsquare_1e2_pcs_0_ts_0_t_0.000000.vtu\nsquare_1e2_pcs_0_ts_1_t_1.000000.vtu\n\nThe pressure at a point is read and plotted over time (two time points)\n'''\n\nimport matplotlib.pyplot as plt \t# for fancy plots\n\nimport vtuIO\t# to read and process (point interpolation) vtu- and pvd-files \n# class methods for information\n#\tPVDIO\n# __init__(self, folder, filename, dim=3):\n# readPVD(self,filename):\n# readTimeSeries(self,fieldname, pts = {'pt0': (0.0,0.0,0.0)}):\n# readTimeStep(self, timestep, fieldname):\n\n\n# read pvd-file specified by path and filename\n# dim refers to the actual dimension:\n# 2D data in 2D are OK (dim=2), 3D data in 3D are OK (dim=3).\n# Note that for 2D data in 3D, e.g. x,y!=0 and z=0, \n# dim must be set to 2, otherwise the interpolator fails.\n# Currently PVDIO assumes 2D data in 3D at x,y and ignores z.\npvdfile=vtuIO.PVDIO(\"square_1e2_pcs_0.pvd\", dim=2)\n\n# get time vector from pvd-data (list)\ntime=pvdfile.timesteps\n\n# define points for interpolation (dictionary)\nselected_points={'pt0': (0.25, 0.5, 0.0), 'pt1': (0.75, 0.5, 0.0)}\n\n# read and interpolate from vtu-files listed in pvd\npressure_interpolation=pvdfile.read_time_series('pressure', selected_points)\n\n# read pressure at pt0 from interpolations (dictionary)\npressure_pt0=pressure_interpolation['pt0']\n\n# plot some result\nplt.plot(time, pressure_pt0)\ntitlestring=\"At point \"+str(selected_points['pt0'])\nplt.title(titlestring)\nplt.xlabel('t')\nplt.ylabel('p')\nplt.show()\n\n# do something with pt1 or whatever you like\n# ...\n\n", "sub_path": "examples/pvd_read_example.py", "file_name": "pvd_read_example.py", "file_ext": "py", "file_size_in_byte": 1715, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "vtuIO.PVDIO", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 47, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 48, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}]} +{"seq_id": "85165808", "text": "#!/usr/bin/env python3\n#\n# MIT\n# ciko@afra-berlin.de\nimport pydle, threading, subprocess, time, asyncio\n\ndef mac_tester():\n\tglobal current_users\n\twhile True:\n\t\t# Load the macs. in the loop for auto reload\n\t\tmacs = {}\n\t\twith open(\"registered_macs\", \"r\") as f:\n\t\t\tfor line in f.readlines():\n\t\t\t\tif len(line.strip()) > 0:\n\t\t\t\t\tmacs[line.split()[0].upper()] = line.split()[1]\n\n\t\t# Scan for all macs in the current network\n\t\tscan_result = subprocess.check_output([\"nmap\", \"-sPn\", \"172.23.42.1-254\"], universal_newlines=True)\n\t\tcurrent_users = []\n\t\tfor line in scan_result.split(\"\\n\"):\n\t\t\twords = line.split()\n\t\t\tif len(words) >= 2:\n\t\t\t\tif words[0] == \"MAC\":\n\t\t\t\t\tmac_address = words[2].upper()\n\t\t\t\t\tif mac_address in macs.keys():\n\t\t\t\t\t\tcurrent_users.append(macs[mac_address])\n\n\t\tcurrent_users = list(set(current_users)) # Dont duplicate users\n\t\ttime.sleep(60)\n\n\nclass MyOwnBot(pydle.Client):\n\t@asyncio.coroutine\n\tdef on_connect(self):\n\t\t yield from self.join('#afra')\n\n\t@asyncio.coroutine\n\tdef on_message(self, target, source, message):\n\t\tglobal current_users\n\t\t# don't respond to our own messages, as this leads to a positive feedback loop\n\t\tif source != self.nickname and (\".presence\" in message or \".present\" in message):\n\t\t\tif len(current_users) == 0:\n\t\t\t\tm = \"Nobody wants to be surveilled.\"\n\t\t\telse:\n\t\t\t\tm = \"Now at AfRA: \" + \", \".join(current_users)\n\t\t\tyield from self.message(target, m)\n\n\ncurrent_users = []\nthreading.Thread(target=mac_tester).start()\n\nclient = MyOwnBot(\"pr3s3nce\", realname=\"AfRA attendance bot\")\nclient.run('chat.freenode.net', tls=True, tls_verify=False)\n\n", "sub_path": "mac_kicker.py", "file_name": "mac_kicker.py", "file_ext": "py", "file_size_in_byte": 1579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "subprocess.check_output", "line_number": 18, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 29, "usage_type": "call"}, {"api_name": "pydle.Client", "line_number": 32, "usage_type": "attribute"}, {"api_name": "asyncio.coroutine", "line_number": 33, "usage_type": "attribute"}, {"api_name": "asyncio.coroutine", "line_number": 37, "usage_type": "attribute"}, {"api_name": "threading.Thread", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "587101937", "text": "import json\nimport logging\nimport requests\nimport os\n\nfrom text_analytics.abstract_nlp_service import NLPService\nfrom text_analytics.enhance import *\nfrom text_analytics.quickUMLS.semtype_lookup import lookup\nfrom text_analytics.quickUMLS.semtype_lookup import get_semantic_type_list\n\nlogger = logging.getLogger()\n\nclass QuickUMLSService(NLPService):\n types_can_handle = {'AllergyIntolerance': enhance_allergy_intolerance_payload_to_fhir,\n 'Immunization': enhance_immunization_payload_to_fhir,\n 'DiagnosticReport': enhance_diagnostic_report_payload_to_fhir,\n 'DocumentReference': enhance_document_reference_payload_to_fhir\n }\n\n PROCESS_TYPE_UNSTRUCTURED = \"QuickUMLS Unstructured\"\n PROCESS_TYPE_STRUCTURED = \"QuickUMLS Structured\"\n\n def __init__(self, json_string):\n config_dict = json.loads(json_string)\n self.quickUMLS_url = config_dict[\"config\"][\"endpoint\"]\n self.jsonString = json_string\n self.config_name = config_dict[\"name\"]\n\n\n def process(self, text):\n if type(text) is bytes:\n request_body = {\"text\": text.decode('utf-8')}\n else:\n request_body = {\"text\": text}\n logger.info(\"Calling QUICKUMLS-\" + self.config_name)\n resp = requests.post(self.quickUMLS_url, json=request_body)\n concepts = json.loads(resp.text)\n conceptsList = []\n if concepts is not None:\n for concept in concepts:\n conceptsList.append(self.concept_to_dict(concept))\n return {\"concepts\": conceptsList}\n\n @staticmethod\n def concept_to_dict(concept):\n output = {\"Structure\": \"Concept\"}\n output[\"generatingService\"] = \"quickUMLS\"\n output[\"coveredText\"] = concept[\"ngram\"] if \"ngram\" in concept else None\n output[\"cui\"] = concept[\"cui\"] if \"cui\" in concept else None\n output[\"begin\"] = concept[\"start\"] if \"start\" in concept else None\n output[\"end\"] = concept[\"end\"] if \"end\" in concept else None\n output[\"preferredName\"] = concept[\"term\"] if \"term\" in concept else None\n output[\"type\"] = get_semantic_type_list(concept[\"semtypes\"]) if \"semtypes\" in concept and len(concept[\"semtypes\"]) > 0 else None\n output[\"negated\"] = False\n return output\n", "sub_path": "services/nlp-insights/text_analytics/quickUMLS/quickUMLS_service.py", "file_name": "quickUMLS_service.py", "file_ext": "py", "file_size_in_byte": 2329, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "text_analytics.abstract_nlp_service.NLPService", "line_number": 13, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 24, "usage_type": "call"}, {"api_name": "requests.post", "line_number": 36, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 37, "usage_type": "call"}, {"api_name": "text_analytics.quickUMLS.semtype_lookup.get_semantic_type_list", "line_number": 53, "usage_type": "call"}]} +{"seq_id": "545900918", "text": "# Copyright 2021 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Tests for driblet.cloud_env_setup.\"\"\"\n\nimport argparse\nimport unittest\nimport mock\nimport parameterized\nfrom gps_building_blocks.cloud.utils import cloud_api\nfrom gps_building_blocks.cloud.utils import cloud_composer\nfrom gps_building_blocks.cloud.utils import cloud_storage\nfrom driblet import cloud_env_setup\n\n\nclass CloudEnvSetupTest(unittest.TestCase):\n\n def setUp(self):\n super(CloudEnvSetupTest, self).setUp()\n self.addCleanup(mock.patch.stopall)\n self.mock_parse_args = mock.patch.object(\n cloud_env_setup, 'parse_arguments', autospec=True).start()\n self.service_account_name = 'my-service@my-project.iam.gserviceaccount.com'\n self.composer_env_name = 'test-env'\n self.project_id = 'project_id'\n self.pypi_packages = {'test-package': '2.0'}\n self.modeling_platform = 'TensorFlow'\n self.mock_parse_args.return_value = argparse.Namespace(\n project_id=self.project_id,\n service_account_name=self.service_account_name,\n modeling_platform=self.modeling_platform,\n composer_env_name=cloud_env_setup._COMPOSER_ENV_NAME,\n dags_folder=cloud_env_setup._DAGS_FOLDER,\n saved_model_folder=cloud_env_setup._SAVED_MODEL_FOLDER,\n gcs_path_prefix=cloud_env_setup._GCS_PATH_PREFIX,\n config_file=cloud_env_setup._FEATURES_CONFIG_FILE)\n\n # Setup cloud_api mocks.\n self.mock_cloud_api_utils = mock.patch.object(\n cloud_api, 'CloudApiUtils', autospec=True).start()\n self.mock_cloud_composer_utils = mock.patch.object(\n cloud_composer, 'CloudComposerUtils', autospec=True).start()\n self.mock_cloud_storage_utils = mock.patch.object(\n cloud_storage, 'CloudStorageUtils', autospec=True).start()\n\n def test_enable_apis_enables_cloud_apis(self):\n mock_enable_apis = self.mock_cloud_api_utils.return_value.enable_apis\n\n cloud_env_setup.enable_apis(self.project_id, self.service_account_name)\n\n self.mock_cloud_api_utils.assert_called_once_with(\n project_id=self.project_id,\n service_account_name=self.service_account_name)\n mock_enable_apis.assert_called_once_with(\n cloud_env_setup._APIS_TO_BE_ENABLED)\n\n def test_install_pypi_packages_raises_value_error(self):\n modeling_platform = 'CustomPlatform'\n\n with self.assertRaises(ValueError):\n cloud_env_setup.install_pypi_packages(self.mock_cloud_composer_utils,\n self.composer_env_name,\n modeling_platform,\n self.pypi_packages)\n\n @parameterized.parameterized.expand([['TensorFlow'], ['AutoMl']])\n def test_install_pypi_packages_installs_python_packages(\n self, modeling_platform):\n mock_install_packages = (\n self.mock_cloud_composer_utils.install_python_packages)\n expected_pypi_packages = self.pypi_packages.copy()\n\n if modeling_platform == 'TensorFlow':\n expected_pypi_packages.update(cloud_env_setup._TENSORFLOW_MODULES)\n\n cloud_env_setup.install_pypi_packages(self.mock_cloud_composer_utils,\n self.composer_env_name,\n modeling_platform, self.pypi_packages)\n\n mock_install_packages.assert_called_once_with(self.composer_env_name,\n expected_pypi_packages)\n\n def test_copy_dags_to_gcs_copies_files_and_dirs_to_gcs(self):\n mock_upload_directory = (\n self.mock_cloud_storage_utils.upload_directory_to_url)\n dags_folder = 'local_dags_folder'\n dags_folder_url = 'gcs_dags_folder/subfolder'\n self.mock_cloud_composer_utils.get_dags_folder.return_value = (\n dags_folder_url)\n\n cloud_env_setup.copy_dags_to_gcs(self.mock_cloud_composer_utils,\n self.mock_cloud_storage_utils,\n self.composer_env_name, dags_folder)\n\n mock_upload_directory.assert_called_once_with(dags_folder,\n 'gcs_dags_folder/dags')\n\n def test_copy_model_to_gcs_copies_saved_model_and_config_file_to_gcs(self):\n mock_upload_directory = (\n self.mock_cloud_storage_utils.upload_directory_to_url)\n mock_upload_file_to_url = self.mock_cloud_storage_utils.upload_file_to_url\n\n cloud_env_setup.copy_model_to_gcs(self.mock_cloud_storage_utils,\n 'test_path_prefix', 'src/saved_model',\n 'config.cfg')\n\n self.assertEqual(mock_upload_directory.call_count, 1)\n\n self.assertEqual(mock_upload_file_to_url.call_count, 1)\n\n\nif __name__ == '__main__':\n unittest.main()\n", "sub_path": "tests/cloud_env_setup_test.py", "file_name": "cloud_env_setup_test.py", "file_ext": "py", "file_size_in_byte": 5235, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "unittest.TestCase", "line_number": 27, "usage_type": "attribute"}, {"api_name": "mock.patch", "line_number": 31, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 32, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 33, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 32, "usage_type": "attribute"}, {"api_name": "argparse.Namespace", "line_number": 39, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup._COMPOSER_ENV_NAME", "line_number": 43, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 43, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._DAGS_FOLDER", "line_number": 44, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 44, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._SAVED_MODEL_FOLDER", "line_number": 45, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 45, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._GCS_PATH_PREFIX", "line_number": 46, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 46, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._FEATURES_CONFIG_FILE", "line_number": 47, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 47, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 50, "usage_type": "call"}, {"api_name": "gps_building_blocks.cloud.utils.cloud_api", "line_number": 51, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 50, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 52, "usage_type": "call"}, {"api_name": "gps_building_blocks.cloud.utils.cloud_composer", "line_number": 53, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 52, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 54, "usage_type": "call"}, {"api_name": "gps_building_blocks.cloud.utils.cloud_storage", "line_number": 55, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 54, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup.enable_apis", "line_number": 60, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 60, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._APIS_TO_BE_ENABLED", "line_number": 66, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 66, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup.install_pypi_packages", "line_number": 72, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 72, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup._TENSORFLOW_MODULES", "line_number": 85, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup", "line_number": 85, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup.install_pypi_packages", "line_number": 87, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 87, "usage_type": "name"}, {"api_name": "parameterized.parameterized.expand", "line_number": 77, "usage_type": "call"}, {"api_name": "parameterized.parameterized", "line_number": 77, "usage_type": "attribute"}, {"api_name": "driblet.cloud_env_setup.copy_dags_to_gcs", "line_number": 102, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 102, "usage_type": "name"}, {"api_name": "driblet.cloud_env_setup.copy_model_to_gcs", "line_number": 114, "usage_type": "call"}, {"api_name": "driblet.cloud_env_setup", "line_number": 114, "usage_type": "name"}, {"api_name": "unittest.main", "line_number": 124, "usage_type": "call"}]} +{"seq_id": "180782717", "text": "\"\"\"The module provides auxiliary functions for the estimation process\"\"\"\nfrom statsmodels.tools.sm_exceptions import PerfectSeparationError\nfrom scipy.stats import norm\nimport statsmodels.api as sm\nimport pandas as pd\nimport numpy as np\n\nfrom grmpy.simulate.simulate_auxiliary import construct_covariance_matrix\nfrom grmpy.simulate.simulate_auxiliary import simulate_unobservables\nfrom grmpy.simulate.simulate_auxiliary import simulate_covariates\nfrom grmpy.simulate.simulate_auxiliary import simulate_outcomes\n\n\ndef log_likelihood(init_dict, data_frame, rslt, dict_=None):\n \"\"\"The function provides the log-likelihood function for the minimization process.\"\"\"\n beta1, beta0, gamma, sd1, sd0, sdv, rho1v, rho0v, choice = \\\n _prepare_arguments(init_dict, rslt)\n likl = []\n for i in [0.0, 1.0]:\n if i == 1.00:\n beta, gamma, rho, sd, sdv = beta1, gamma, rho1v, sd1, sdv\n else:\n beta, gamma, rho, sd, sdv = beta0, gamma, rho0v, sd0, sdv\n data = data_frame[data_frame['D'] == i]\n X = data.filter(regex=r'^X\\_')\n Z = data.filter(regex=r'^Z\\_')\n g = pd.concat((X, Z), axis=1)\n choice_ = pd.DataFrame.sum(choice * g, axis=1)\n part1 = (data['Y'] - pd.DataFrame.sum(beta * X, axis=1)) / sd\n part2 = (choice_ - rho * sdv * part1) / (np.sqrt((1 - rho ** 2) * sdv ** 2))\n dist_1, dist_2 = norm.pdf(part1), norm.cdf(part2)\n if i == 1.00:\n contrib = (1.0 / sd) * dist_1 * dist_2\n else:\n contrib = (1.0 / sd) * dist_1 * (1.0 - dist_2)\n likl.append(contrib)\n likl = np.append(likl[0], likl[1])\n likl = - np.mean(np.log(np.clip(likl, 1e-20, np.inf)))\n if dict_ is None:\n pass\n else:\n dict_['crit'][str(len(dict_['crit']))] = likl\n return likl\n\n\ndef _prepare_arguments(init_dict, rslt):\n \"\"\"The function prepares the coefficients for the log-liklihood function.\"\"\"\n beta1 = np.array(rslt['TREATED']['all'])\n beta0 = np.array(rslt['UNTREATED']['all'])\n gamma = np.array(rslt['COST']['all'])\n sd1 = rslt['DIST']['all'][1]\n sd0 = rslt['DIST']['all'][0]\n sdv = init_dict['DIST']['all'][5]\n rho1, rho0 = rslt['DIST']['all'][3], rslt['DIST']['all'][2]\n choice = np.concatenate(((np.subtract(beta1, beta0)), -gamma))\n\n return beta1, beta0, gamma, sd1, sd0, sdv, rho1, rho0, choice\n\n\ndef start_values(init_dict, data_frame, option):\n \"\"\"The function selects the start values for the minimization process.\"\"\"\n\n if not isinstance(init_dict, dict):\n raise AssertionError()\n numbers = [init_dict['AUX']['num_covars_out'], init_dict['AUX']['num_covars_cost']]\n\n if option == 'init':\n # Set coefficients equal the true init file values\n x0 = init_dict['AUX']['init_values'][:2 * numbers[0] + numbers[1]]\n sd_ = None\n elif option == 'auto':\n\n try:\n\n # Estimate beta1 and beta0:\n beta = []\n sd_ = []\n for i in [0.0, 1.0]:\n Y, X = data_frame.Y[data_frame.D == i], data_frame.filter(regex=r'^X\\_')[\n data_frame.D == i]\n ols_results = sm.OLS(Y, X).fit()\n beta += [ols_results.params]\n sd_ += [np.sqrt(ols_results.scale)]\n\n # Estimate gamma via probit\n X = data_frame.filter(regex=r'^X\\_')\n Z = (data_frame.filter(regex=r'^Z\\_')).drop('Z_0', axis=1)\n XZ = np.concatenate((X, Z), axis=1)\n probitRslt = sm.Probit(data_frame.D, XZ).fit(disp=0)\n gamma = probitRslt.params\n gamma_const = np.subtract(np.subtract(beta[1][0], beta[0][0]), gamma[0])\n if len(init_dict['COST']['all']) == 1:\n gamma = [gamma_const]\n else:\n gamma = np.concatenate(([gamma_const], gamma[-(numbers[1] - 1):]))\n # Arange starting values\n x0 = np.concatenate((beta[1], beta[0]))\n x0 = np.concatenate((x0, gamma))\n\n except (PerfectSeparationError, ValueError):\n msg = 'The estimation process wasn`t able to provide automatic start values due to ' \\\n 'perfect seperation. \\n ' \\\n ' The intialization specifications are used as start ' \\\n 'values during the further process.'\n # Set coefficients equal the true init file values\n x0 = init_dict['AUX']['init_values'][:2 * numbers[0] + numbers[1]]\n sd_ = None\n init_dict['ESTIMATION']['warning'] = msg\n option = 'init'\n\n x0, start = provide_cholesky_decom(init_dict, x0, option, sd_)\n init_dict['AUX']['starting_values'] = x0[:]\n init_dict['AUX']['start_values'] = start\n x0 = np.array(x0)\n\n return x0\n\n\ndef distribute_parameters(init_dict, start_values, dict_=None):\n \"\"\"The function generates a dictionary for the representation of the optimization output.\"\"\"\n if dict_ is None:\n pass\n else:\n dict_['parameter'][str(len(dict_['parameter']))] = start_values\n\n num_covars_out = init_dict['AUX']['num_covars_out']\n rslt = dict()\n\n rslt['TREATED'] = dict()\n rslt['UNTREATED'] = dict()\n rslt['COST'] = dict()\n rslt['DIST'] = dict()\n\n # Distribute parameters\n rslt['TREATED']['all'] = start_values[:num_covars_out]\n rslt['UNTREATED']['all'] = start_values[num_covars_out:(2 * num_covars_out)]\n rslt['COST']['all'] = start_values[(2 * num_covars_out):(-6)]\n\n rslt['DIST']['all'] = backward_cholesky_transformation(start_values, True)\n\n # Update auxiliary versions\n rslt['AUX'] = dict()\n rslt['AUX']['x_internal'] = start_values[:]\n rslt['AUX']['init_values'] = init_dict['AUX']['init_values']\n return rslt\n\n\ndef minimizing_interface(start_values, init_dict, data_frame, dict_):\n \"\"\"The function provides the minimization interface for the estimation process.\"\"\"\n # Collect arguments\n rslt = distribute_parameters(init_dict, start_values, dict_)\n\n # Calculate liklihood for pre specified arguments\n likl = log_likelihood(init_dict, data_frame, rslt, dict_)\n\n return likl\n\n\ndef calculate_criteria(init_dict, data_frame, start_values):\n \"\"\"The function calculates the criteria function value.\"\"\"\n rslt = distribute_parameters(init_dict, start_values)\n criteria = log_likelihood(init_dict, data_frame, rslt)\n return criteria\n\n\ndef print_logfile(init_dict, rslt):\n \"\"\"The function writes the log file for the estimation process.\"\"\"\n # Adjust output\n init_dict, rslt = adjust_print_output(init_dict, rslt)\n with open('est.grmpy.info', 'w') as file_:\n\n for label in ['Optimization Information', 'Criterion Function', 'Economic Parameters']:\n header = '\\n \\n {:<10}\\n\\n'.format(label)\n file_.write(header)\n if label == 'Optimization Information':\n for section in ['Optimizer', 'Start values', 'Success', 'Status',\n 'Number of Evaluations',\n 'Criterion', 'Message', 'Warning']:\n fmt = ' {:<10}' + ' {:<20}'\n if section == 'Number of Evaluations':\n if len(str(rslt['nfev'])) == 4:\n fmt += ' {:>21}\\n'\n else:\n fmt += ' {:>20}\\n'\n file_.write(fmt.format('', section + ':', rslt['nfev']))\n elif section == 'Start values':\n fmt += ' {:>23}\\n'\n file_.write(fmt.format('', section + ':',\n init_dict['ESTIMATION']['start']))\n elif section == 'Optimizer':\n if init_dict['ESTIMATION']['optimizer'] == 'SCIPY-POWELL':\n fmt += ' {:>31}\\n'\n else:\n fmt += ' {:>29}\\n'\n file_.write(fmt.format('', section + ':',\n init_dict['ESTIMATION']['optimizer']))\n elif section == 'Criterion':\n fmt += ' {:>20.4f}\\n'\n file_.write(fmt.format('', section + ':', rslt['crit']))\n elif section in ['Message', 'Warning']:\n fmt += ' {:>20}\\n'\n file_.write(fmt.format('', section + ':', rslt[section.lower()]) + '\\n')\n if section == 'Warning':\n if 'warning' in init_dict['ESTIMATION'].keys():\n file_.write(fmt.format('', '', init_dict['ESTIMATION']['warning']))\n else:\n fmt += ' {:>20}\\n'\n file_.write(fmt.format('', section + ':', rslt[section.lower()]))\n elif label == 'Criterion Function':\n fmt = ' {:<10}' * 2 + ' {:>20}' * 2 + '\\n\\n'\n file_.write(fmt.format('', '', 'Start', 'Current'))\n file_.write('\\n' + fmt.format('', '', init_dict['AUX']['criteria'], rslt['crit']))\n\n else:\n file_.write(fmt.format(*['', 'Identifier', 'Start', 'Current']) + '\\n\\n')\n fmt = ' {:>10}' * 2 + ' {:>20.4f}' * 2\n for i in range(len(rslt['AUX']['x_internal'])):\n file_.write('{0}\\n'.format(\n fmt.format('', str(i), init_dict['AUX']['starting_values'][i],\n rslt['AUX']['x_internal'][i])))\n\n\ndef optimizer_options(init_dict_):\n \"\"\"The function provides the optimizer options given the initialization dictionary.\"\"\"\n method = init_dict_['ESTIMATION']['optimizer'].split('-')[1]\n opt_dict = init_dict_['SCIPY-' + method]\n\n return opt_dict, method\n\n\ndef simulate_estimation(init_dict, rslt, data_frame, start=False):\n \"\"\"The function simulates a new sample based on the estimated coefficients.\"\"\"\n\n # Distribute information\n seed = init_dict['SIMULATION']['seed']\n\n # Determine parametrization and read in /simulate observables\n if start is True:\n start_dict, rslt_dict = process_results(init_dict, rslt, start)\n dicts = [start_dict, rslt_dict]\n X = data_frame.filter(regex=r'^X\\_')\n Z = data_frame.filter(regex=r'^Z\\_')\n else:\n rslt_dict = process_results(init_dict, rslt, start)\n dicts = [rslt_dict]\n X = simulate_covariates(rslt_dict, 'TREATED')\n Z = simulate_covariates(rslt_dict, 'COST')\n\n data_frames = []\n for dict_ in dicts:\n # Set seed value\n np.random.seed(seed)\n # Simulate unobservables\n U, _ = simulate_unobservables(dict_)\n\n # Simulate endogeneous variables\n Y, D, Y_1, Y_0 = simulate_outcomes(dict_, X, Z, U)\n\n df = write_output_estimation(Y, D, X, Z, Y_1, Y_0)\n data_frames += [df]\n\n if start is True:\n return data_frames[0], data_frames[1]\n else:\n return data_frames[0]\n\n\ndef process_results(init_dict, rslt, start=False):\n \"\"\"The function processes the results dictionary for the following simulation.\"\"\"\n rslt_dict = {}\n start_dict = {}\n dicts = [rslt_dict, start_dict]\n for dict_ in dicts:\n dict_['SIMULATION'] = {}\n dict_['SIMULATION']['agents'] = init_dict['ESTIMATION']['agents']\n dict_['SIMULATION']['seed'] = init_dict['SIMULATION']['seed']\n if dict_ == rslt_dict:\n for key_ in ['TREATED', 'UNTREATED', 'COST']:\n dict_[key_] = {}\n dict_[key_]['types'] = init_dict[key_]['types']\n dict_[key_]['all'] = rslt[key_]['all']\n dict_ = transform_rslt_DIST(rslt['AUX']['x_internal'], dict_)\n else:\n if start is True:\n num_treated = len(init_dict['TREATED']['all'])\n for key_ in ['TREATED', 'UNTREATED', 'COST']:\n dict_[key_] = {}\n dict_[key_]['types'] = init_dict[key_]['types']\n dict_['TREATED']['all'] = init_dict['AUX']['starting_values'][:num_treated]\n dict_['UNTREATED']['all'] = init_dict['AUX']['starting_values'][\n num_treated:2 * num_treated]\n dict_['COST']['all'] = init_dict['AUX']['starting_values'][2 * num_treated:-6]\n dict_ = transform_rslt_DIST(init_dict['AUX']['starting_values'][-6:], dict_)\n return start_dict, rslt_dict\n else:\n return rslt_dict\n\n\ndef write_descriptives(init_dict, df1, rslt):\n \"\"\"The function writes the info file including the descriptives of the original and the\n estimated sample.\n \"\"\"\n df3, df2 = simulate_estimation(init_dict, rslt, df1, True)\n with open('descriptives.grmpy.txt', 'w') as file_:\n # First we note some basic information ab out the dataset.\n header = '\\n\\n Number of Observations \\n\\n'\n file_.write(header)\n info_ = []\n for i, label in enumerate([df1, df2, df3]):\n info_ += [[label.shape[0], (label['D'] == 1).sum(), (label['D'] == 0).sum()]]\n\n fmt = ' {:<25}' + ' {:>20}' * 3 + '\\n\\n\\n'\n file_.write(fmt.format(*['Sample', 'Observed', 'Simulated (finish)',\n 'Simulated (start)']))\n\n for i, label in enumerate(['All', 'Treated', 'Untreated']):\n str_ = ' {:<25}' + ' {:>20}' * 3 + '\\n'\n file_.write(str_.format(label, info_[0][i], info_[1][i], info_[2][i]))\n\n header = '\\n\\n Distribution of Outcomes\\n\\n'\n file_.write(header)\n for group in ['All', 'Treated', 'Untreated']:\n header = '\\n\\n ' ' {:<10}'.format(group) + '\\n\\n'\n file_.write(header)\n fmt = ' {:<25}' + ' {:>20}' * 5 + '\\n\\n'\n args = ['', 'Mean', 'Std-Dev.', '25%', '50%', '75%']\n file_.write(fmt.format(*args))\n\n for sample in ['Observed Sample', 'Simulated Sample (finish)',\n 'Simulated Sample (start)']:\n\n if sample == 'Observed Sample':\n data_frame = df1\n elif sample == 'Simulated Sample (finish)':\n data_frame = df2\n else:\n data_frame = df3\n\n data = data_frame['Y']\n\n if group == 'Treated':\n data = data[data_frame['D'] == 1]\n elif group == 'Untreated':\n data = data[data_frame['D'] == 0]\n else:\n pass\n fmt = ' {:<25}' + ' {:>20.4f}' * 5 + '\\n'\n info = list(data.describe().tolist()[i] for i in [1, 2, 4, 5, 6])\n if pd.isnull(info).all():\n fmt = ' {:<10}' + ' {:>20}' * 5 + '\\n'\n info = ['---'] * 5\n elif pd.isnull(info[1]):\n info[1] = '---'\n fmt = ' {:<25}' ' {:>20.4f}' ' {:>20}' + ' {:>20.4f}' * 3 + '\\n'\n\n file_.write(fmt.format(*[sample] + info))\n\n\ndef write_output_estimation(Y, D, X, Z, Y_1, Y_0):\n \"\"\"The function converts the simulated variables to a panda data frame.\"\"\"\n\n # Stack arrays\n data = np.column_stack((Y, D, X, Z, Y_1, Y_0))\n\n # Construct list of column labels\n column = ['Y', 'D']\n for i in range(X.shape[1]):\n str_ = 'X_' + str(i)\n column.append(str_)\n for i in range(Z.shape[1]):\n str_ = 'Z_' + str(i)\n column.append(str_)\n column += ['Y1', 'Y0']\n\n # Generate data frame\n df = pd.DataFrame(data=data, columns=column)\n df['D'] = df['D'].apply(np.int64)\n return df\n\n\ndef process_rslt(init_dict, dict_, rslt):\n \"\"\"The function checks if the criteria function value is smaller for the optimization output as\n for the start values.\n \"\"\"\n\n x = min(dict_['crit'], key=dict_['crit'].get)\n if dict_['crit'][str(x)] <= rslt['crit']:\n warning = 'The optimization algorithm has failed to provide the parametrization that ' \\\n 'leads to the minimal criterion function value. \\n ' \\\n ' The estimation output is automatically adjusted.'\n\n rslt['warning'] = warning\n\n if dict_['crit'][str(x)] < init_dict['AUX']['criteria']:\n rslt['AUX']['x_internal'] = dict_['parameter'][str(x)].tolist()\n rslt['crit'] = dict_['crit'][str(x)]\n else:\n rslt['AUX']['x_internal'] = init_dict['AUX']['starting_values']\n rslt['crit'] = init_dict['AUX']['criteria']\n else:\n rslt['warning'] = '---'\n\n\ndef bfgs_dict():\n \"\"\"The function provides a dictionary for tracking the criteria function values and the\n associated parametrization.\n \"\"\"\n rslt_dict = {'parameter': {}, 'crit': {}}\n return rslt_dict\n\n\ndef adjust_output(opt_rslt, init_dict, start_values, dict_=None):\n \"\"\"The function adds different information of the minimization process to the estimation\n output.\"\"\"\n rslt = distribute_parameters(init_dict, start_values)\n rslt['success'], rslt['status'] = opt_rslt['success'], opt_rslt['status']\n rslt['message'], rslt['nfev'], rslt['crit'] = opt_rslt['message'], opt_rslt['nfev'], \\\n opt_rslt['fun']\n\n process_rslt(init_dict, dict_, rslt)\n\n return rslt\n\n\ndef adjust_output_maxiter_zero(init_dict, start_values):\n \"\"\"The function returns a result dictionary if the maximum number of evaluations is zero.\"\"\"\n num_covars_out = init_dict['AUX']['num_covars_out']\n rslt = dict()\n rslt['TREATED'] = dict()\n rslt['UNTREATED'] = dict()\n rslt['COST'] = dict()\n rslt['DIST'] = dict()\n\n # Distribute parameters\n rslt['TREATED']['all'] = start_values[:num_covars_out]\n rslt['UNTREATED']['all'] = start_values[num_covars_out:(2 * num_covars_out)]\n rslt['COST']['all'] = start_values[(2 * num_covars_out):(-6)]\n\n rslt['DIST']['all'] = start_values[-6:]\n\n # Update auxiliary versions\n rslt['AUX'] = dict()\n rslt['AUX']['x_internal'] = start_values[:]\n\n rslt['AUX']['init_values'] = init_dict['AUX']['init_values'][:]\n rslt['success'], rslt['status'] = False, 2\n rslt['message'], rslt['nfev'], rslt['crit'] = '---', 0, init_dict['AUX']['criteria']\n rslt['warning'] = '---'\n\n return rslt\n\n\ndef adjust_print_output(init_dict, rslt):\n \"\"\"The function arranges the distributional parameters.\"\"\"\n\n for dict_ in [init_dict, rslt]:\n if dict_ == init_dict:\n key_ = 'starting_values'\n else:\n key_ = 'x_internal'\n if not isinstance(dict_['AUX'][key_], list):\n dict_['AUX'][key_] = dict_['AUX'][key_].tolist()\n dict_['AUX'][key_] = backward_cholesky_transformation(dict_['AUX'][key_])\n dict_['AUX'][key_] = np.array(dict_['AUX'][key_])\n\n return init_dict, rslt\n\n\ndef transform_rslt_DIST(rslt, dict_):\n \"\"\"The function converts the correlation parameters from the estimation outcome to\n covariances for the simulation of the estimation sample.\n \"\"\"\n dict_['DIST'] = {}\n place_holder = rslt[-6:]\n cov01 = place_holder[1] * place_holder[0] * place_holder[3]\n cov0V = place_holder[2] * place_holder[0] * place_holder[5]\n cov1V = place_holder[4] * place_holder[3] * place_holder[5]\n\n dict_['DIST']['all'] = [place_holder[0], cov01, cov0V, place_holder[3], cov1V, place_holder[5]]\n\n for i, element in enumerate(dict_['DIST']['all']):\n dict_['DIST']['all'][i] = round(element, 4)\n\n return dict_\n\n\ndef provide_cholesky_decom(init_dict, x0, option, sd_=None):\n \"\"\"The function transforms the start covariance matrix into its cholesky decomposition.\"\"\"\n if option == 'init':\n cov = construct_covariance_matrix(init_dict)\n L = np.linalg.cholesky(cov)\n L = L[np.tril_indices(3)]\n distribution_characteristics = init_dict['AUX']['init_values'][-6:]\n x0 = np.concatenate((x0, L))\n\n elif option == 'auto':\n distribution_characteristics = [sd_[0], init_dict['DIST']['all'][1], 0, sd_[1], 0,\n init_dict['DIST']['all'][5]]\n cov = np.zeros((3, 3))\n cov[np.triu_indices(3)] = [distribution_characteristics]\n cov[np.tril_indices(3, k=-1)] = cov[np.triu_indices(3, k=1)]\n cov[np.diag_indices(3)] **= 2\n L = np.linalg.cholesky(cov)\n L = L[np.tril_indices(3)]\n x0 = np.concatenate((x0, L))\n init_dict['AUX']['cholesky_decomposition'] = L.tolist()\n start = [i for i in x0] + distribution_characteristics\n\n return x0, start\n\n\ndef backward_cholesky_transformation(x0, dist=False, test=False):\n \"\"\"The function creates a positive semi definite covariance matrix from the given cholesky\n decomposition elements.\n \"\"\"\n start_cholesky = x0[-6:]\n\n cholesky = np.zeros((3, 3))\n cholesky[np.tril_indices(3)] = start_cholesky\n cov = np.dot(cholesky, cholesky.T)\n sdv = cov[2, 2] ** 0.5\n # What do we want to use here ? the sdv from the cholesky decomposition or the sdv from the init dict??\n if dist is True:\n sd0 = cov[0, 0] ** 0.5\n sd1 = cov[1, 1] ** 0.5\n rho0 = cov[0, 2] / (sd0 * sdv)\n rho1 = cov[1, 2] / (sd1 * sdv)\n\n dist_parameter = [sd0, sd1, rho0, rho1]\n return dist_parameter\n else:\n dist_para = cov[np.triu_indices(3)]\n sd0, sd1, sdv = dist_para[0] ** 0.5, dist_para[3] ** 0.5, dist_para[5] ** 0.5\n rho0, rho1 = dist_para[2] / (sd0 * sdv), dist_para[4] / (sd1 * sdv)\n rho01 = dist_para[1] / (sd0 * sd1)\n if test is False:\n output = x0[:-6] + [sd0, rho01, rho0, sd1, rho1, sdv]\n else:\n output = [sd0, rho01, rho0, sd1, rho1, sdv]\n return output\n", "sub_path": "grmpy/estimate/estimate_auxiliary.py", "file_name": "estimate_auxiliary.py", "file_ext": "py", "file_size_in_byte": 21913, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.concat", "line_number": 27, "usage_type": "call"}, {"api_name": "pandas.DataFrame.sum", "line_number": 28, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame.sum", "line_number": 29, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 29, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 30, "usage_type": "call"}, {"api_name": "scipy.stats.norm.pdf", "line_number": 31, "usage_type": "call"}, {"api_name": "scipy.stats.norm", "line_number": 31, "usage_type": "name"}, {"api_name": "scipy.stats.norm.cdf", "line_number": 31, "usage_type": "call"}, {"api_name": "numpy.append", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 38, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 55, "usage_type": "call"}, {"api_name": "numpy.subtract", "line_number": 55, "usage_type": "call"}, {"api_name": "statsmodels.api.OLS", "line_number": 81, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 88, "usage_type": "call"}, {"api_name": "statsmodels.api.Probit", "line_number": 89, "usage_type": "call"}, {"api_name": "statsmodels.api", "line_number": 89, "usage_type": "name"}, {"api_name": "numpy.subtract", "line_number": 91, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 97, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 98, "usage_type": "call"}, {"api_name": "statsmodels.tools.sm_exceptions.PerfectSeparationError", "line_number": 100, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 114, "usage_type": "call"}, {"api_name": "grmpy.simulate.simulate_auxiliary.simulate_covariates", "line_number": 246, "usage_type": "call"}, {"api_name": "grmpy.simulate.simulate_auxiliary.simulate_covariates", "line_number": 247, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 252, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 252, "usage_type": "attribute"}, {"api_name": "grmpy.simulate.simulate_auxiliary.simulate_unobservables", "line_number": 254, "usage_type": "call"}, {"api_name": "grmpy.simulate.simulate_auxiliary.simulate_outcomes", "line_number": 257, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 349, "usage_type": "call"}, {"api_name": "pandas.isnull", "line_number": 352, "usage_type": "call"}, {"api_name": "numpy.column_stack", "line_number": 363, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.int64", "line_number": 377, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 464, "usage_type": "call"}, {"api_name": "grmpy.simulate.simulate_auxiliary.construct_covariance_matrix", "line_number": 490, "usage_type": "call"}, {"api_name": "numpy.linalg.cholesky", "line_number": 491, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 491, "usage_type": "attribute"}, {"api_name": "numpy.tril_indices", "line_number": 492, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 494, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 499, "usage_type": "call"}, {"api_name": "numpy.triu_indices", "line_number": 500, "usage_type": "call"}, {"api_name": "numpy.tril_indices", "line_number": 501, "usage_type": "call"}, {"api_name": "numpy.triu_indices", "line_number": 501, "usage_type": "call"}, {"api_name": "numpy.diag_indices", "line_number": 502, "usage_type": "call"}, {"api_name": "numpy.linalg.cholesky", "line_number": 503, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 503, "usage_type": "attribute"}, {"api_name": "numpy.tril_indices", "line_number": 504, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 505, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 518, "usage_type": "call"}, {"api_name": "numpy.tril_indices", "line_number": 519, "usage_type": "call"}, {"api_name": "numpy.dot", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.triu_indices", "line_number": 532, "usage_type": "call"}]} +{"seq_id": "571378768", "text": "from pathlib import Path # noqa\nimport os\nimport sys\n\nsys.path.append(os.getcwd())\n\n# pylint: disable=wrong-import-position\nfrom core.lib.mac_utils import MacAddressUtils # noqa\nfrom core.lib.common import print_json # noqa\nfrom core.pandas_utils.split_csv_data import extract_data_as_separate_csv # noqa\n\n\nif __name__ == '__main__':\n scripts_dir_path = Path(os.path.abspath(__file__)).parent\n project_dir_path = scripts_dir_path.parent\n input_file_path = project_dir_path / 'fixtures/results/results.csv'\n result_dir_path = project_dir_path / 'fixtures/results/splitted/'\n summary = extract_data_as_separate_csv(\n file_path=input_file_path,\n result_folder=result_dir_path,\n filter_columns=['src_mac', 'dst_mac'],\n sort_by_columns=['timestamp'],\n transform_function=MacAddressUtils().int_to_mac\n )\n print_json(summary)\n", "sub_path": "tools/split_data_file_to_separate_csv.py", "file_name": "split_data_file_to_separate_csv.py", "file_ext": "py", "file_size_in_byte": 879, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "sys.path.append", "line_number": 5, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 5, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 5, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "core.pandas_utils.split_csv_data.extract_data_as_separate_csv", "line_number": 18, "usage_type": "call"}, {"api_name": "core.lib.mac_utils.MacAddressUtils", "line_number": 23, "usage_type": "call"}, {"api_name": "core.lib.common.print_json", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "562824694", "text": "# Licensed under a 3-clause BSD style license - see LICENSE.rst\nfrom __future__ import absolute_import, division, print_function, unicode_literals\nimport logging\nimport click\nfrom ..irf._utils_old import read_json, write_all\n\nlog = logging.getLogger(__name__)\n\n\n@click.command('fit')\n@click.option('--counts', default='counts.fits',\n help='Counts FITS file name')\n@click.option('--exposure', default='exposure.fits',\n help='Exposure FITS file name')\n@click.option('--background', default='background.fits',\n help='Background FITS file name')\n@click.option('--psf', type=str, default=None,\n help='PSF JSON file name')\n@click.option('--sources', default='sources.json',\n help='Sources JSON file name (contains start '\n 'values for fit of Gaussians)')\n@click.option('--roi', type=str, default=None,\n help='Region of interest (ROI) file name (ds9 reg format)')\n@click.option('--outfile', default='fit_results.json',\n help='Output JSON file with fit results')\ndef cli_image_fit(counts, exposure, background, psf,\n sources, roi, outfile):\n \"\"\"Fit morphology model to image using Sherpa.\n\n Uses initial parameters from a JSON file (for now only Gaussians).\n \"\"\"\n import sherpa.astro.ui\n from ..irf import SherpaMultiGaussPSF\n\n # ---------------------------------------------------------\n # Load images, PSF and sources\n # ---------------------------------------------------------\n log.info('Clearing the sherpa session')\n sherpa.astro.ui.clean()\n\n log.info('Reading counts: {}'.format(counts))\n sherpa.astro.ui.load_image(counts)\n\n log.info('Reading exposure: {}'.format(exposure))\n sherpa.astro.ui.load_table_model('exposure', exposure)\n\n log.info('Reading background: {}'.format(background))\n sherpa.astro.ui.load_table_model('background', background)\n\n if psf:\n log.info('Reading PSF: {}'.format(psf))\n SherpaMultiGaussPSF(psf).set()\n else:\n log.warning(\"No PSF convolution.\")\n\n if roi:\n log.info('Reading ROI: {}'.format(roi))\n sherpa.astro.ui.notice2d(roi)\n else:\n log.info('No ROI selected.')\n\n log.info('Reading sources: {}'.format(sources))\n read_json(sources, sherpa.astro.ui.set_source)\n\n # ---------------------------------------------------------\n # Set up the full model and freeze PSF, exposure, background\n # ---------------------------------------------------------\n # Scale exposure by 1e-10 to get ampl or order unity and avoid some fitting problems\n name = sherpa.astro.ui.get_source().name\n if psf:\n full_model = 'background + 1e-12 * exposure * psf ({})'.format(name)\n sherpa.astro.ui.set_full_model(full_model)\n sherpa.astro.ui.freeze('background', 'exposure', 'psf')\n else:\n full_model = 'background + 1e-12 * exposure * {}'.format(name)\n sherpa.astro.ui.set_full_model(full_model)\n sherpa.astro.ui.freeze('background', 'exposure')\n\n # ---------------------------------------------------------\n # Set up the fit\n # ---------------------------------------------------------\n sherpa.astro.ui.set_coord('image')\n sherpa.astro.ui.set_stat('cash')\n sherpa.astro.ui.set_method('levmar') # levmar, neldermead, moncar\n sherpa.astro.ui.set_method_opt('maxfev', int(1e3))\n sherpa.astro.ui.set_method_opt('verbose', 10)\n\n # ---------------------------------------------------------\n # Fit and save information we care about\n # ---------------------------------------------------------\n # sherpa.astro.ui.show_all() # Prints info about data and model\n sherpa.astro.ui.fit() # Does the fit\n # sherpa.astro.ui.covar() # Computes symmetric errors (fast)\n # conf() # Computes asymmetric errors (slow)\n # image_fit() # Shows data, model, residuals in ds9\n log.info('Writing {}'.format(outfile))\n write_all(outfile)\n\n # Save model image\n sherpa.astro.ui.set_par('background.ampl', 0)\n sherpa.astro.ui.notice2d()\n log.info('Writing model.fits')\n sherpa.astro.ui.save_model('model.fits', clobber=True)\n sherpa.astro.ui.clean()\n", "sub_path": "gammapy/scripts/image_fit.py", "file_name": "image_fit.py", "file_ext": "py", "file_size_in_byte": 4195, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro.ui.clean", "line_number": 39, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 39, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 39, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.load_image", "line_number": 42, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 42, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 42, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.load_table_model", "line_number": 45, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 45, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 45, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.load_table_model", "line_number": 48, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 48, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 48, "usage_type": "name"}, {"api_name": "irf.SherpaMultiGaussPSF", "line_number": 52, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro.ui.notice2d", "line_number": 58, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 58, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 58, "usage_type": "name"}, {"api_name": "irf._utils_old.read_json", "line_number": 63, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 63, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 63, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.get_source", "line_number": 69, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 69, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 69, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_full_model", "line_number": 72, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 72, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 72, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.freeze", "line_number": 73, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 73, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 73, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_full_model", "line_number": 76, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 76, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 76, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.freeze", "line_number": 77, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 77, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 77, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_coord", "line_number": 82, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 82, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 82, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_stat", "line_number": 83, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 83, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_method", "line_number": 84, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 84, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 84, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_method_opt", "line_number": 85, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 85, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 85, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.set_method_opt", "line_number": 86, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 86, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 86, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.fit", "line_number": 92, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 92, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 92, "usage_type": "name"}, {"api_name": "irf._utils_old.write_all", "line_number": 97, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro.ui.set_par", "line_number": 100, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 100, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 100, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.notice2d", "line_number": 101, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 101, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 101, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.save_model", "line_number": 103, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 103, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 103, "usage_type": "name"}, {"api_name": "sherpa.astro.ui.astro.ui.clean", "line_number": 104, "usage_type": "call"}, {"api_name": "sherpa.astro.ui.astro", "line_number": 104, "usage_type": "attribute"}, {"api_name": "sherpa.astro.ui", "line_number": 104, "usage_type": "name"}, {"api_name": "click.command", "line_number": 10, "usage_type": "call"}, {"api_name": "click.option", "line_number": 11, "usage_type": "call"}, {"api_name": "click.option", "line_number": 13, "usage_type": "call"}, {"api_name": "click.option", "line_number": 15, "usage_type": "call"}, {"api_name": "click.option", "line_number": 17, "usage_type": "call"}, {"api_name": "click.option", "line_number": 19, "usage_type": "call"}, {"api_name": "click.option", "line_number": 22, "usage_type": "call"}, {"api_name": "click.option", "line_number": 24, "usage_type": "call"}]} +{"seq_id": "211727058", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 2.6 (62161)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-i686/egg/coils/logic/contact/cache_service.py\n# Compiled at: 2012-10-12 07:02:39\nimport os, time\nfrom sqlalchemy import and_\nfrom coils.foundation import *\nfrom coils.core import *\nfrom utility import *\n\nclass ContactCacheService(Service):\n __service__ = 'coils.contacts.cache'\n __auto_dispatch__ = True\n __is_worker__ = True\n\n def __init__(self):\n self._cursor = None\n self._iter = 0\n self._next_run = 0.0\n Service.__init__(self)\n return\n\n def prepare(self):\n Service.prepare(self)\n self._ticktock = time.time()\n self._ctx = AdministrativeContext()\n\n @property\n def ticktock(self):\n if time.time() - selk._ticktock > 59:\n self._ticktock = time.time()\n return True\n return False\n\n def _read_data(self):\n self.log.info('Retrieving Contact list for vCard cache fill.')\n query = self._ctx.db_session().query(Contact.object_id, Contact.version).filter(and_(Contact.status != 'archived', Contact.first_name is not None, Contact.last_name is not None)).distinct()\n self._cursor = query.all()\n self._iter = 0\n return\n\n def work(self):\n if time.time() > self._next_run:\n if self._cursor is None:\n self._read_data()\n self._count = len(self._cursor)\n else:\n if self._iter >= self._count:\n self.log.info('Refill of vCard cache complete.')\n self._cursor = None\n self._next_run = time.time() + 43200.0\n return\n self.log.debug(('Walking contact cache; items {0}...{1}').format(self._iter, self._iter + 150))\n for i in range(150):\n if self._iter < self._count:\n if not is_vcard_cached(self._cursor[self._iter].object_id, self._cursor[self._iter].version):\n self._ctx.run_command('contact::get-as-vcard', id=self._cursor[self._iter].object_id, access_check=False)\n self._iter = self._iter + 1\n else:\n break\n\n return", "sub_path": "pycfiles/OpenGroupware-0.1.48-py2.6/cache_service.py", "file_name": "cache_service.py", "file_ext": "py", "file_size_in_byte": 2351, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "time.time", "line_number": 27, "usage_type": "call"}, {"api_name": "time.time", "line_number": 32, "usage_type": "call"}, {"api_name": "time.time", "line_number": 33, "usage_type": "call"}, {"api_name": "sqlalchemy.and_", "line_number": 39, "usage_type": "call"}, {"api_name": "time.time", "line_number": 45, "usage_type": "call"}, {"api_name": "time.time", "line_number": 53, "usage_type": "call"}]} +{"seq_id": "470811988", "text": "import pygame as pg\nimport sys\nfrom random import randint\n\ndef quit():\n for event in pg.event.get():\n if event.type == pg.QUIT:\n sys.exit()\n\ndef check_speed(speed):\n\n move = randint(0,1)\n new_speed = [-1,1]\n\n if speed[0] == 0:\n speed[0] = new_speed[move]\n if speed[1] == 0:\n speed[1] = new_speed[move]\n\ndef input_box():\n\n color_inactive = pg.Color('lightskyblue3')\n color_active = pg.Color('dodgerblue2')\n color = color_inactive\n input_box = pg.Rect(100, 100, 140, 32)\n\n ment = 'YOU ARE THE BEST SCORE!!!'\n ment_surface = ment_font.render(ment,False,(0,0,0))\n\n active = False\n text = ''\n done = False\n\n while not done:\n for event in pg.event.get():\n quit()\n if event.type == pg.MOUSEBUTTONDOWN:\n if input_box.collidepoint(event.pos):\n active = not active\n else:\n active = False\n color = color_active if active else color_inactive\n if event.type == pg.KEYDOWN:\n if active:\n if event.key == pg.K_RETURN:\n print(text)\n return text\n elif event.key == pg.K_BACKSPACE:\n text = text[:-1]\n else:\n text += event.unicode\n\n\n txt_surface = font.render(text, True, color)\n # Resize the box if the text is too long.\n width = max(200, txt_surface.get_width() + 10)\n input_box.w = width\n # Blit the text.\n screen.blit(txt_surface, (input_box.x + 5, input_box.y))\n screen.blit(ment_surface, (5,300))\n # Blit the input_box rect.\n pg.draw.rect(screen, color, input_box, 2)\n\n pg.display.flip()\n\nwhite = 255,255,255\nblue = 0, 128, 255\n\npg.init()\npg.font.init()\n\nscreen = pg.display.set_mode((600,600))\nfont = pg.font.SysFont('Arial',32)\nment_font = pg.font.SysFont('Arial', 50)", "sub_path": "obstacle_avoid/source.py", "file_name": "source.py", "file_ext": "py", "file_size_in_byte": 1990, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pygame.event.get", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 7, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 8, "usage_type": "call"}, {"api_name": "random.randint", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 22, "usage_type": "call"}, {"api_name": "pygame.Color", "line_number": 23, "usage_type": "call"}, {"api_name": "pygame.Rect", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 35, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 35, "usage_type": "attribute"}, {"api_name": "pygame.MOUSEBUTTONDOWN", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 43, "usage_type": "attribute"}, {"api_name": "pygame.K_RETURN", "line_number": 45, "usage_type": "attribute"}, {"api_name": "pygame.K_BACKSPACE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "pygame.draw.rect", "line_number": 62, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 62, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 64, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygame.init", "line_number": 69, "usage_type": "call"}, {"api_name": "pygame.font.init", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 70, "usage_type": "attribute"}, {"api_name": "pygame.display.set_mode", "line_number": 72, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 73, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pygame.font.SysFont", "line_number": 74, "usage_type": "call"}, {"api_name": "pygame.font", "line_number": 74, "usage_type": "attribute"}]} +{"seq_id": "265072117", "text": "from django.shortcuts import render\nfrom django.http import JsonResponse\nfrom blog.models import Post\n\n# Create your views here.\n\ndef get_posts(request):\n '''\n :param request: HttpRequest object, which should be in json format\n :returns: Json in format like beyond.\n\n Example return\n ==============\n\n { posts:\n [\n {\n title: example title,\n content: example content\n date : 26-02-2017,\n language: pl,\n author: rafix\n },\n {\n title: example title2,\n content: example content2,\n date : 26-03-2017,\n language: en,\n author: ator\n }\n ]\n }\n\n Request example\n ====================\n\n GET /posts?language=pl&count=3&category=space\n\n If no count param specified, method will return all posts matching other params.\n You can skip other params to acquire simillar effect.\n\n If language param is not specified, default is 'pl'\n\n Invalid request\n ===============\n\n If request is not made by ajax, server is returning error 500\n\n Errors\n ======\n Error messages (In Code 500 etc.) are returned in 'text' key.\n '''\n\n if (not request.is_ajax()):\n return JsonResponse({'text': 'Resquest seems not to be ajax'}, status=500)\n # Commented for debugging\n\n response = {\n 'posts':\n [\n ]\n }\n\n language = request.GET['language'] if 'language' in request.GET else 'pl'\n count = int(request.GET['count']) if 'count' in request.GET else 0\n\n if count == 0:\n db_result = Post.objects.get(category=request.GET['category']) if 'category' in request.GET else Post.objects.all()\n else:\n db_result = Post.objects.get(category=request.GET['category'])[:count] if 'category' in request.GET else Post.objects.all()[:count]\n\n for post in db_result:\n\n post = {\n 'title': post.content.english_title if language == 'en' else post.content.polish_title,\n 'content': post.content.english_content if language == 'en' else post.content.polish_content,\n 'date': post.date,\n 'language': language,\n 'author': post.author\n }\n response['posts'].append(post)\n\n return JsonResponse(response)\n\n", "sub_path": "microchip/blog/views_bc.py", "file_name": "views_bc.py", "file_ext": "py", "file_size_in_byte": 2399, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.http.JsonResponse", "line_number": 55, "usage_type": "call"}, {"api_name": "blog.models.Post.objects.get", "line_number": 68, "usage_type": "call"}, {"api_name": "blog.models.Post.objects", "line_number": 68, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 68, "usage_type": "name"}, {"api_name": "blog.models.Post.objects.all", "line_number": 68, "usage_type": "call"}, {"api_name": "blog.models.Post.objects.get", "line_number": 70, "usage_type": "call"}, {"api_name": "blog.models.Post.objects", "line_number": 70, "usage_type": "attribute"}, {"api_name": "blog.models.Post", "line_number": 70, "usage_type": "name"}, {"api_name": "blog.models.Post.objects.all", "line_number": 70, "usage_type": "call"}, {"api_name": "django.http.JsonResponse", "line_number": 83, "usage_type": "call"}]} +{"seq_id": "132377284", "text": "#!/usr/bin/env python\n# -*-coding:utf-8-*-\n# @Time : 2017/11/1 ~ 2019/9/1\n# @Author : Allen Woo\nfrom flask import request\nfrom flask_babel import gettext\n\nfrom apps.app import mdbs\nfrom apps.core.flask.reqparse import arg_verify\nfrom apps.modules.post.process.post_process import get_posts_pr\nfrom apps.utils.format.obj_format import str_to_num\nfrom apps.utils.paging.paging import datas_paging\nfrom apps.utils.upload.get_filepath import get_avatar_url\n\n\ndef search_process():\n \"\"\"\n 搜索(暂不支持全文搜索)\n 只能搜索文章, 用户\n :return:\n \"\"\"\n\n keyword = request.argget.all('keyword')\n target = request.argget.all('target')\n page = str_to_num(request.argget.all('page', 1))\n pre = str_to_num(request.argget.all('pre', 10))\n\n s, r = arg_verify(reqargs=[(gettext(\"keyword\"), keyword)], required=True)\n if not s:\n return r\n\n data = {\"posts\": {}, \"users\": {}}\n # post\n if not target or target == \"post\":\n data[\"posts\"] = {}\n data[\"posts\"][\"items\"] = get_posts_pr(\n field={\n \"title\": 1,\n \"issue_time\": 1,\n \"brief_content\": 1},\n page=page,\n pre=pre,\n status=\"is_issued\",\n sort=None,\n time_range=None,\n matching_rec=None,\n keyword=keyword,\n other_filter=None,\n is_admin=False,\n get_userinfo=False)[\"posts\"]\n data[\"posts\"][\"kw\"] = keyword\n\n if not target or target == \"user\":\n # user\n data[\"users\"] = {\"kw\": keyword, \"items\": []}\n query_conditions = {\n \"is_delete\": {\n \"$in\": [\n False, 0]}, \"active\": {\n \"$in\": [\n True, 1]}}\n keyword = {\"$regex\": keyword, \"$options\": \"$i\"}\n query_conditions[\"$or\"] = [{\"username\": keyword},\n {\"email\": keyword},\n {\"custom_domain\": keyword}\n ]\n us = mdbs[\"user\"].db.user.find(\n query_conditions, {\n \"_id\": 1, \"username\": 1, \"avatar_url\": 1, \"custom_domain\": 1, \"gender\": 1, })\n\n data_cnt = us.count(True)\n users = list(us.skip(pre * (page - 1)).limit(pre))\n for user in users:\n user['_id'] = str(user['_id'])\n user[\"avatar_url\"][\"url\"] = get_avatar_url(user[\"avatar_url\"])\n\n data[\"users\"][\"items\"] = datas_paging(\n pre=pre, page_num=page, data_cnt=data_cnt, datas=users)\n\n return data\n", "sub_path": "apps/modules/search/process/search.py", "file_name": "search.py", "file_ext": "py", "file_size_in_byte": 2578, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.request.argget.all", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.request.argget", "line_number": 23, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 23, "usage_type": "name"}, {"api_name": "flask.request.argget.all", "line_number": 24, "usage_type": "call"}, {"api_name": "flask.request.argget", "line_number": 24, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 24, "usage_type": "name"}, {"api_name": "apps.utils.format.obj_format.str_to_num", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request.argget.all", "line_number": 25, "usage_type": "call"}, {"api_name": "flask.request.argget", "line_number": 25, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 25, "usage_type": "name"}, {"api_name": "apps.utils.format.obj_format.str_to_num", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.request.argget.all", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.request.argget", "line_number": 26, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 26, "usage_type": "name"}, {"api_name": "apps.core.flask.reqparse.arg_verify", "line_number": 28, "usage_type": "call"}, {"api_name": "flask_babel.gettext", "line_number": 28, "usage_type": "call"}, {"api_name": "apps.modules.post.process.post_process.get_posts_pr", "line_number": 36, "usage_type": "call"}, {"api_name": "apps.app.mdbs", "line_number": 67, "usage_type": "name"}, {"api_name": "apps.utils.upload.get_filepath.get_avatar_url", "line_number": 75, "usage_type": "call"}, {"api_name": "apps.utils.paging.paging.datas_paging", "line_number": 77, "usage_type": "call"}]} +{"seq_id": "553863736", "text": "# Copyright 2016 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\nimport logging\nimport os\nimport queue\nimport threading\nimport time\nfrom typing import List, Optional, Set\n\nfrom pants.base.exiter import PANTS_SUCCEEDED_EXIT_CODE\nfrom pants.base.specs import Specs\nfrom pants.engine.fs import PathGlobs, Snapshot\nfrom pants.engine.unions import UnionMembership\nfrom pants.goal.run_tracker import RunTracker\nfrom pants.init.engine_initializer import LegacyGraphScheduler, LegacyGraphSession\nfrom pants.option.options import Options\nfrom pants.option.options_bootstrapper import OptionsBootstrapper\nfrom pants.pantsd.service.fs_event_service import FSEventService\nfrom pants.pantsd.service.pants_service import PantsService\n\n\nclass SchedulerService(PantsService):\n \"\"\"The pantsd scheduler service.\n\n This service holds an online Scheduler instance that is primed via watchman filesystem events.\n \"\"\"\n\n QUEUE_SIZE = 64\n INVALIDATION_WATCHER_LIVENESS_CHECK_INTERVAL = 1\n\n def __init__(\n self,\n *,\n fs_event_service: Optional[FSEventService],\n legacy_graph_scheduler: LegacyGraphScheduler,\n build_root: str,\n invalidation_globs: List[str],\n pantsd_pidfile: Optional[str],\n union_membership: UnionMembership,\n ) -> None:\n \"\"\"\n :param fs_event_service: An unstarted FSEventService instance for setting up filesystem event handlers.\n :param legacy_graph_scheduler: The LegacyGraphScheduler instance for graph construction.\n :param build_root: The current build root.\n :param invalidation_globs: A list of `globs` that when encountered in filesystem event\n subscriptions will tear down the daemon.\n :param pantsd_pidfile: The path to the pantsd pidfile for fs event monitoring.\n \"\"\"\n super().__init__()\n self._fs_event_service = fs_event_service\n self._graph_helper = legacy_graph_scheduler\n self._invalidation_globs = invalidation_globs\n self._build_root = build_root\n self._pantsd_pidfile = pantsd_pidfile\n self._union_membership = union_membership\n\n self._scheduler = legacy_graph_scheduler.scheduler\n # This session is only used for checking whether any invalidation globs have been invalidated.\n # It is not involved with a build itself; just with deciding when we should restart pantsd.\n self._scheduler_session = self._scheduler.new_session(\n zipkin_trace_v2=False, build_id=\"scheduler_service_session\",\n )\n self._logger = logging.getLogger(__name__)\n self._event_queue: queue.Queue = queue.Queue(maxsize=self.QUEUE_SIZE)\n self._watchman_is_running = threading.Event()\n self._invalidating_snapshot = None\n self._invalidating_files: Set[str] = set()\n\n self._loop_condition = LoopCondition()\n\n def _get_snapshot(self):\n \"\"\"Returns a Snapshot of the input globs.\"\"\"\n return self._scheduler_session.product_request(\n Snapshot, subjects=[PathGlobs(self._invalidation_globs)]\n )[0]\n\n def setup(self, services):\n \"\"\"Service setup.\"\"\"\n super().setup(services)\n # Register filesystem event handlers on an FSEventService instance.\n if self._fs_event_service is not None:\n self._fs_event_service.register_all_files_handler(\n self._enqueue_fs_event, self._fs_event_service.PANTS_ALL_FILES_SUBSCRIPTION_NAME\n )\n\n # N.B. We compute the invalidating fileset eagerly at launch with an assumption that files\n # that exist at startup are the only ones that can affect the running daemon.\n if self._fs_event_service is not None:\n if self._invalidation_globs:\n self._invalidating_snapshot = self._get_snapshot()\n self._invalidating_files = self._invalidating_snapshot.files\n self._logger.info(\n \"watching invalidating files: {}\".format(self._invalidating_files)\n )\n\n if self._pantsd_pidfile:\n self._fs_event_service.register_pidfile_handler(\n self._pantsd_pidfile, self._enqueue_fs_event\n )\n\n def _enqueue_fs_event(self, event):\n \"\"\"Watchman filesystem event handler for BUILD/requirements.txt updates.\n\n Called via a thread.\n \"\"\"\n self._logger.info(\n \"enqueuing {} changes for subscription {}\".format(\n len(event[\"files\"]), event[\"subscription\"]\n )\n )\n self._event_queue.put(event)\n\n def _maybe_invalidate_scheduler_batch(self):\n new_snapshot = self._get_snapshot()\n if (\n self._invalidating_snapshot\n and new_snapshot.digest != self._invalidating_snapshot.digest\n ):\n self._logger.critical(\n \"saw file events covered by invalidation globs [{}], terminating the daemon.\".format(\n self._invalidating_files\n )\n )\n self.terminate()\n\n def _maybe_invalidate_scheduler_pidfile(self):\n new_pid = self._check_pid_changed()\n if new_pid is not False:\n self._logger.critical(\n \"{} says pantsd PID is {} but my PID is: {}: terminating\".format(\n self._pantsd_pidfile, new_pid, os.getpid(),\n )\n )\n self.terminate()\n\n def _check_pid_changed(self):\n \"\"\"Reads pidfile and returns False if its PID is ours, else a printable (maybe falsey)\n value.\"\"\"\n try:\n with open(os.path.join(self._build_root, self._pantsd_pidfile), \"r\") as f:\n pid_from_file = f.read()\n except IOError:\n return \"[no file could be read]\"\n if int(pid_from_file) != os.getpid():\n return pid_from_file\n else:\n return False\n\n def _handle_batch_event(self, files):\n self._logger.debug(\"handling change event for: %s\", files)\n\n invalidated = self._scheduler.invalidate_files(files)\n if invalidated:\n self._loop_condition.notify_all()\n\n self._maybe_invalidate_scheduler_batch()\n\n def _process_event_queue(self):\n \"\"\"File event notification queue processor.\"\"\"\n try:\n event = self._event_queue.get(timeout=0.05)\n except queue.Empty:\n return\n\n try:\n subscription, is_initial_event, files = (\n event[\"subscription\"],\n event[\"is_fresh_instance\"],\n event[\"files\"],\n )\n except (KeyError, UnicodeDecodeError) as e:\n self._logger.warning(\"%r raised by invalid watchman event: %s\", e, event)\n return\n\n self._logger.debug(\n \"processing {} files for subscription {} (first_event={})\".format(\n len(files), subscription, is_initial_event\n )\n )\n\n # The first watchman event for all_files is a listing of all files - ignore it.\n if (\n not is_initial_event\n and self._fs_event_service is not None\n and subscription == self._fs_event_service.PANTS_ALL_FILES_SUBSCRIPTION_NAME\n ):\n self._handle_batch_event(files)\n\n # However, we do want to check for the initial event in the pid file creation.\n if subscription == self._fs_event_service.PANTS_PID_SUBSCRIPTION_NAME:\n self._maybe_invalidate_scheduler_pidfile()\n\n if not self._watchman_is_running.is_set():\n self._watchman_is_running.set()\n\n self._event_queue.task_done()\n\n def _check_invalidation_watcher_liveness(self):\n time.sleep(self.INVALIDATION_WATCHER_LIVENESS_CHECK_INTERVAL)\n if not self._scheduler.check_invalidation_watcher_liveness():\n # Watcher failed for some reason\n self._logger.critical(\n \"The graph invalidation watcher failed, so we are shutting down. Check the pantsd.log for details\"\n )\n self.terminate()\n\n def prepare_graph(self, options: Options) -> LegacyGraphSession:\n # If any nodes exist in the product graph, wait for the initial watchman event to avoid\n # racing watchman startup vs invalidation events.\n if self._fs_event_service is not None and self._scheduler.graph_len() > 0:\n self._logger.debug(\n f\"fs event service is running and graph_len > 0: waiting for initial watchman event\"\n )\n self._watchman_is_running.wait()\n\n global_options = options.for_global_scope()\n build_id = RunTracker.global_instance().run_id\n v2_ui = global_options.get(\"v2_ui\", False)\n zipkin_trace_v2 = options.for_scope(\"reporting\").zipkin_trace_v2\n return self._graph_helper.new_session(zipkin_trace_v2, build_id, v2_ui)\n\n def graph_run_v2(\n self,\n session: LegacyGraphSession,\n specs: Specs,\n options: Options,\n options_bootstrapper: OptionsBootstrapper,\n ) -> int:\n \"\"\"Perform an entire v2 run.\n\n The exit_code in the return indicates whether any issue was encountered.\n \"\"\"\n\n global_options = options.for_global_scope()\n perform_loop = global_options.get(\"loop\", False)\n v2 = global_options.v2\n\n if not perform_loop:\n return self._body(session, options, options_bootstrapper, specs, v2)\n\n # TODO: See https://github.com/pantsbuild/pants/issues/6288 regarding Ctrl+C handling.\n iterations = global_options.loop_max\n exit_code = PANTS_SUCCEEDED_EXIT_CODE\n\n while iterations and not self._state.is_terminating:\n try:\n exit_code = self._body(session, options, options_bootstrapper, specs, v2)\n except session.scheduler_session.execution_error_type as e:\n self._logger.warning(e)\n\n iterations -= 1\n while (\n iterations\n and not self._state.is_terminating\n and not self._loop_condition.wait(timeout=1)\n ):\n continue\n\n return exit_code\n\n def _body(\n self,\n session: LegacyGraphSession,\n options: Options,\n options_bootstrapper: OptionsBootstrapper,\n specs: Specs,\n v2: bool,\n ) -> int:\n exit_code = PANTS_SUCCEEDED_EXIT_CODE\n\n _, ambiguous_goals, v2_goals = options.goals_by_version\n\n if v2_goals or (ambiguous_goals and v2):\n goals = v2_goals + (ambiguous_goals if v2 else tuple())\n\n # N.B. @goal_rules run pre-fork in order to cache the products they request during execution.\n exit_code = session.run_goal_rules(\n options_bootstrapper=options_bootstrapper,\n union_membership=self._union_membership,\n options=options,\n goals=goals,\n specs=specs,\n )\n\n return exit_code\n\n def run(self):\n \"\"\"Main service entrypoint.\"\"\"\n while not self._state.is_terminating:\n if self._fs_event_service is not None:\n self._process_event_queue()\n else:\n self._check_invalidation_watcher_liveness()\n self._state.maybe_pause()\n\n\nclass LoopCondition:\n \"\"\"A wrapped condition variable to handle deciding when loop consumers should re-run.\n\n Any number of threads may wait and/or notify the condition.\n \"\"\"\n\n def __init__(self):\n super().__init__()\n self._condition = threading.Condition(threading.Lock())\n self._iteration = 0\n\n def notify_all(self):\n \"\"\"Notifies all threads waiting for the condition.\"\"\"\n with self._condition:\n self._iteration += 1\n self._condition.notify_all()\n\n def wait(self, timeout):\n \"\"\"Waits for the condition for at most the given timeout and returns True if the condition\n triggered.\n\n Generally called in a loop until the condition triggers.\n \"\"\"\n\n with self._condition:\n previous_iteration = self._iteration\n self._condition.wait(timeout)\n return previous_iteration != self._iteration\n", "sub_path": "src/python/pants/pantsd/service/scheduler_service.py", "file_name": "scheduler_service.py", "file_ext": "py", "file_size_in_byte": 12341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pants.pantsd.service.pants_service.PantsService", "line_number": 23, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 35, "usage_type": "name"}, {"api_name": "pants.pantsd.service.fs_event_service.FSEventService", "line_number": 35, "usage_type": "name"}, {"api_name": "pants.init.engine_initializer.LegacyGraphScheduler", "line_number": 36, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 38, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 39, "usage_type": "name"}, {"api_name": "pants.engine.unions.UnionMembership", "line_number": 40, "usage_type": "name"}, {"api_name": "logging.getLogger", "line_number": 64, "usage_type": "call"}, {"api_name": "queue.Queue", "line_number": 65, "usage_type": "attribute"}, {"api_name": "threading.Event", "line_number": 66, "usage_type": "call"}, {"api_name": "typing.Set", "line_number": 68, "usage_type": "name"}, {"api_name": "pants.engine.fs.Snapshot", "line_number": 75, "usage_type": "argument"}, {"api_name": "pants.engine.fs.PathGlobs", "line_number": 75, "usage_type": "call"}, {"api_name": "os.getpid", "line_number": 132, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 141, "usage_type": "call"}, {"api_name": "os.path", "line_number": 141, "usage_type": "attribute"}, {"api_name": "os.getpid", "line_number": 145, "usage_type": "call"}, {"api_name": "queue.Empty", "line_number": 163, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 200, "usage_type": "call"}, {"api_name": "pants.option.options.Options", "line_number": 208, "usage_type": "name"}, {"api_name": "pants.goal.run_tracker.RunTracker.global_instance", "line_number": 218, "usage_type": "call"}, {"api_name": "pants.goal.run_tracker.RunTracker", "line_number": 218, "usage_type": "name"}, {"api_name": "pants.init.engine_initializer.LegacyGraphSession", "line_number": 208, "usage_type": "name"}, {"api_name": "pants.init.engine_initializer.LegacyGraphSession", "line_number": 225, "usage_type": "name"}, {"api_name": "pants.base.specs.Specs", "line_number": 226, "usage_type": "name"}, {"api_name": "pants.option.options.Options", "line_number": 227, "usage_type": "name"}, {"api_name": "pants.option.options_bootstrapper.OptionsBootstrapper", "line_number": 228, "usage_type": "name"}, {"api_name": "pants.base.exiter.PANTS_SUCCEEDED_EXIT_CODE", "line_number": 244, "usage_type": "name"}, {"api_name": "pants.init.engine_initializer.LegacyGraphSession", "line_number": 264, "usage_type": "name"}, {"api_name": "pants.option.options.Options", "line_number": 265, "usage_type": "name"}, {"api_name": "pants.option.options_bootstrapper.OptionsBootstrapper", "line_number": 266, "usage_type": "name"}, {"api_name": "pants.base.specs.Specs", "line_number": 267, "usage_type": "name"}, {"api_name": "pants.base.exiter.PANTS_SUCCEEDED_EXIT_CODE", "line_number": 270, "usage_type": "name"}, {"api_name": "threading.Condition", "line_number": 306, "usage_type": "call"}, {"api_name": "threading.Lock", "line_number": 306, "usage_type": "call"}]} +{"seq_id": "61737889", "text": "from typing import Dict\nfrom typing import List\n\nimport requests\nfrom app.api_helpers import get_providers\nfrom app.api_helpers import unique_id\nfrom app.api_helpers import valid_original_title\nfrom app.api_helpers import valid_release_date\nfrom app.api_helpers import valid_title\nfrom app.config import get_settings\nfrom app.db import models\nfrom app.schemas import Media\nfrom app.schemas import Movie\nfrom app.schemas import Person\nfrom app.schemas import TV\n\n\ntmdb_url = get_settings().tmdb_url\ntmdb_key = get_settings().tmdb_key\n\n\ndef fetch_trending_movies(page: int) -> Dict:\n url = f'{tmdb_url}trending/movie/week?api_key={tmdb_key}&page={page}'\n return requests.get(url).json()[\"results\"]\n\n\ndef fetch_trending_tv(page: int) -> Dict:\n url = f'{tmdb_url}trending/tv/week?api_key={tmdb_key}&page={page}'\n return requests.get(url).json()[\"results\"]\n\n\ndef media_converter(mixed_list: List[Dict]) -> List[Media]:\n \"\"\"Takes a list movie/tv json [\"results\"] and converts it to Media\"\"\"\n movie_url = 'https://api.themoviedb.org/3/genre/movie/list?' \\\n f'api_key={tmdb_key}'\n tv_url = 'https://api.themoviedb.org/3/genre/tv/list?' \\\n f'api_key={tmdb_key}'\n\n movie_genres = requests.get(movie_url).json()\n tv_genres = requests.get(tv_url).json()\n\n movie_genre_dict = {\n genre['id']: genre['name'] for genre in movie_genres['genres']\n }\n tv_genre_dict = {\n genre['id']: genre['name'] for genre in tv_genres['genres']\n }\n\n # Only keeps the unique keys\n genre_dict = {**movie_genre_dict, **tv_genre_dict}\n\n return [\n # pydantic Media schema\n Media(\n id=unique_id(media),\n title=valid_title(media),\n original_title=valid_original_title(media),\n overview=media.get('overview'),\n release_date=valid_release_date(media),\n genres=[\n genre_dict.get(genre_id) for genre_id in media.get('genre_ids')\n ] if media.get('genre_ids') else ['Unknown'],\n poster_path=media.get('poster_path'),\n popularity=media.get('popularity')\n ).dict()\n for media in mixed_list\n ]\n\n\nasync def get_person_from_id(person_id: int):\n \"\"\" Gets data of a person from an id\n \"\"\"\n # Here we make 3 api calls into 1 using the append_to_response header\n url = f'{tmdb_url}person/{person_id}?api_key={tmdb_key}' \\\n '&append_to_response=movie_credits,tv_credits'\n\n person = requests.get(url).json()\n\n # pydantic schema for a person\n return Person(\n id=person.get('id'),\n name=person.get('name'),\n birthdate=person.get('birthdate'),\n deathday=person.get('deathday'),\n biography=person.get('biography'),\n place_of_birth=person.get('place_of_birth'),\n also_known_as=person.get('also_known_as'),\n profile_path=person.get('profile_path'),\n gender=person.get('gender'),\n movie_credits=person.get('movie_credits').get('cast'),\n tv_credits=person.get('tv_credits').get('cast')\n )\n\n\ndef get_movie_from_id(movie_id: int, country_code: str = 'DK') -> Movie:\n \"\"\" Gets data of a movie from an id\n \"\"\"\n # Here we make 3 api calls into 1 using the append_to_response header\n url = f'{tmdb_url}movie/{movie_id}?api_key={tmdb_key}' \\\n '&append_to_response=watch/providers,recommendations,credits'\n\n movies = requests.get(url).json()\n\n # pydantic schema for a movie\n return Movie(\n id=movies.get('id'),\n title=movies.get('title'),\n release_date=movies.get('release_date'),\n overview=movies.get('overview'),\n genres=[\n genre.get('name') for genre in movies.get('genres')\n ],\n imdb_id=movies.get('imdb_id'),\n runtime=movies.get('runtime'),\n providers=get_providers(movies.get('watch/providers'), country_code),\n recommendations=get_recommendations(movies.get('recommendations')),\n poster_path=movies.get('poster_path'),\n cast=movies.get('credits').get('cast'),\n popularity=movies.get('popularity'),\n backdrop_path=movies.get('backdrop_path')\n )\n\n\ndef get_tv_from_id(tv_id: int, country_code: str = 'DK') -> TV:\n \"\"\" Gets data of a tv series from an id\n \"\"\"\n # Here we make 3 api calls into 1 using the append_to_response header\n url = f'{tmdb_url}tv/{tv_id}?api_key={tmdb_key}' \\\n '&append_to_response=watch/providers,recommendations,credits'\n\n tv = requests.get(url).json()\n\n # pydantic schema for a tv series\n return TV(\n id=tv.get('id'),\n name=tv.get('name'),\n first_air_date=tv.get('first_air_date'),\n overview=tv.get('overview'),\n genres=[\n genre.get('name') for genre in tv.get('genres')\n ],\n episode_run_time=tv.get('episode_run_time'),\n providers=get_providers(tv.get('watch/providers'), country_code),\n recommendations=get_recommendations(tv.get('recommendations')),\n poster_path=tv.get('poster_path'),\n popularity=tv.get('popularity'),\n number_of_seasons=tv.get('number_of_seasons'),\n seasons=tv.get('seasons'),\n backdrop_path=tv.get('backdrop_path'),\n cast=tv.get('credits').get('cast')\n )\n\n\ndef get_genres() -> Dict:\n \"\"\"Gets genres from movies and tv-series to translate genre_ids\n \"\"\"\n movie_url = 'https://api.themoviedb.org/3/genre/movie/list?' \\\n f'api_key={tmdb_key}'\n tv_url = 'https://api.themoviedb.org/3/genre/tv/list?' \\\n f'api_key={tmdb_key}'\n\n movie_genres = requests.get(movie_url).json()\n tv_genres = requests.get(tv_url).json()\n\n movie_genre_dict = {\n genre['id']: genre['name'] for genre in movie_genres['genres']\n }\n tv_genre_dict = {\n genre['id']: genre['name'] for genre in tv_genres['genres']\n }\n\n # Only keeps the unique keys\n return {**movie_genre_dict, **tv_genre_dict}\n\n\ndef request_providers(media: models.Media):\n tmdb_url = get_settings().tmdb_url\n tmdb_key = get_settings().tmdb_key\n\n try:\n if media.id[0] == 'm':\n url = f'{tmdb_url}movie/{media.id[1:]}?api_key={tmdb_key}' \\\n '&append_to_response=watch/providers'\n\n elif media.id[0] == 't':\n url = f'{tmdb_url}tv/{media.id[1:]}?api_key={tmdb_key}' \\\n '&append_to_response=watch/providers'\n\n media_provider_append = requests.get(url).json()\n return {\n 'media_id': media.id,\n 'data': get_providers(media_provider_append.get('watch/providers'))\n }\n\n except Exception as e:\n print(e)\n\n\ndef get_recommendations(recommendations: Dict) -> List[Dict]:\n \"\"\" Gets list of recommended movies for a movie\n \"\"\"\n\n return [result for result in recommendations['results']]\n", "sub_path": "backend/app/api.py", "file_name": "api.py", "file_ext": "py", "file_size_in_byte": 6839, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "app.config.get_settings", "line_number": 18, "usage_type": "call"}, {"api_name": "app.config.get_settings", "line_number": 19, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 24, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 22, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 29, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 27, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 32, "usage_type": "name"}, {"api_name": "typing.Dict", "line_number": 32, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 39, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 40, "usage_type": "call"}, {"api_name": "app.schemas.Media", "line_number": 54, "usage_type": "call"}, {"api_name": "app.api_helpers.unique_id", "line_number": 55, "usage_type": "call"}, {"api_name": "app.api_helpers.valid_title", "line_number": 56, "usage_type": "call"}, {"api_name": "app.api_helpers.valid_original_title", "line_number": 57, "usage_type": "call"}, {"api_name": "app.api_helpers.valid_release_date", "line_number": 59, "usage_type": "call"}, {"api_name": "app.schemas.Media", "line_number": 32, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 77, "usage_type": "call"}, {"api_name": "app.schemas.Person", "line_number": 80, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 102, "usage_type": "call"}, {"api_name": "app.schemas.Movie", "line_number": 105, "usage_type": "call"}, {"api_name": "app.api_helpers.get_providers", "line_number": 115, "usage_type": "call"}, {"api_name": "app.schemas.Movie", "line_number": 95, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 131, "usage_type": "call"}, {"api_name": "app.schemas.TV", "line_number": 134, "usage_type": "call"}, {"api_name": "app.api_helpers.get_providers", "line_number": 143, "usage_type": "call"}, {"api_name": "app.schemas.TV", "line_number": 124, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 162, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 163, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 154, "usage_type": "name"}, {"api_name": "app.db.models.Media", "line_number": 176, "usage_type": "attribute"}, {"api_name": "app.db.models", "line_number": 176, "usage_type": "name"}, {"api_name": "app.config.get_settings", "line_number": 177, "usage_type": "call"}, {"api_name": "app.config.get_settings", "line_number": 178, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 189, "usage_type": "call"}, {"api_name": "app.api_helpers.get_providers", "line_number": 192, "usage_type": "call"}, {"api_name": "typing.Dict", "line_number": 199, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 199, "usage_type": "name"}]} +{"seq_id": "61303030", "text": "\r\n\r\nimport sys\r\nimport random\r\nimport math\r\n\r\nimport pygame\r\nimport pygame.gfxdraw\r\nfrom pygame.locals import *\r\n\r\n#Define some standard colors\r\nFUCHSIA = (255, 0, 255)\r\nPURPLE = (128, 0, 128)\r\nTEAL = (0, 128, 128)\r\nLIME = (0, 255, 0)\r\nGREEN = (0, 128, 0)\r\nOLIVE = (128, 128, 0)\r\nYELLOW = (255, 255, 0)\r\nORANGE = (255, 165, 0)\r\nRED = (255, 0, 0)\r\nMAROON = (128, 0, 0)\r\nSILVER = (192, 192, 192)\r\nGRAY = (128, 128, 128)\r\nBLUE = (0, 0, 255)\r\nNAVY = (0, 0, 128)\r\nAQUA = (0, 255, 255)\r\nWHITE = (255, 255, 255)\r\nBLACK = (0, 0, 0)\r\n\r\npygame.init()\r\n\r\nDISPLAY_WIDTH = 800\r\nDISPLAY_HEIGHT = 600\r\nDW_HALF = DISPLAY_WIDTH / 2\r\nDH_HALF = DISPLAY_HEIGHT / 2\r\nDISPLAY_AREA = DISPLAY_WIDTH * DISPLAY_HEIGHT\r\nDS = pygame.display.set_mode((DISPLAY_WIDTH, DISPLAY_HEIGHT))\r\n\r\n# FUNCTIONS ------------------------------------------------------------------------------------------------ FUNCTIONS\r\ndef event_handler():\r\n\tfor event in pygame.event.get():\r\n\t\tif event.type == QUIT or (event.type == KEYDOWN and event.key == K_ESCAPE):\r\n\t\t\tpygame.quit()\r\n\t\t\tsys.exit()\r\n\r\nPI = math.pi # simplifies the code\r\nrotation = 0.0 # create a floating point number for the line rotation\r\nrotation_vector = (PI * 2) / 360 # the line will rotate through 360 degrees\r\nrotation_radius = 250 # how big the rotation 'circle' radius is\r\n\r\nwhile True:\r\n\tevent_handler()\r\n\t\r\n\t# starting co-ordinates of the line\r\n\tx1 = DW_HALF + math.cos(rotation) * rotation_radius\r\n\ty1 = DH_HALF + math.sin(rotation) * rotation_radius\r\n\t\r\n\t\r\n\t# ending co-ordinates of the line\r\n\tx2 = DW_HALF + math.cos(rotation + PI) * rotation_radius\r\n\ty2 = DH_HALF + math.sin(rotation + PI) * rotation_radius\r\n\t\r\n\t# increase rotation by 1 degree\r\n\trotation += rotation_vector\r\n\t\r\n\t\r\n\t# draw a blue line on the primary display surface starting at x1,y1 and ending at x2, y2 with a thickness of 5pixels\r\n\tpygame.draw.line(DS, BLUE, (x1, y1), (x2, y2), 5)\r\n\t\r\n\tpygame.display.update()\r\n\tDS.fill([0,0,0])", "sub_path": "Tutorial_Backup/Tutorial_2/justcode_line.py", "file_name": "justcode_line.py", "file_ext": "py", "file_size_in_byte": 1930, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.init", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 37, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 37, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 41, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 43, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 44, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 46, "usage_type": "attribute"}, {"api_name": "math.cos", "line_number": 55, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 56, "usage_type": "call"}, {"api_name": "math.cos", "line_number": 60, "usage_type": "call"}, {"api_name": "math.sin", "line_number": 61, "usage_type": "call"}, {"api_name": "pygame.draw.line", "line_number": 68, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 70, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 70, "usage_type": "attribute"}]} +{"seq_id": "190140573", "text": "import av\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom PIL import Image\nfrom skimage import io\nimport os \nimport cv2\nimport glob\nimport imageio\nimport moviepy.editor as mp\n##########################################################################################\n##########################################################################################\n# \tInstructions:\n#\t\tThe goal of this file is to convert the image or movie into matrices.\n#\t\tThis is all done in this file to keep this step outside the main code. \n#\t\tThis may require some changes to accomodate different file types\n# Currently support file type: tif, tiff, avi, \n##########################################################################################\n##########################################################################################\ndef file_pre_processing(file_name,extension='avi'):\n\tfolder_name = 'ALL_MOVIES_MATRICES'\n\tif not os.path.exists(folder_name):\n\t\tos.makedirs(folder_name)\n\t##########################################################################################\n\tfile_names_list = [file_name]\n\n\tfor kk in range(0,len(file_names_list)):\n\t\n\t\tt1 = file_names_list[kk]\n\t\n\t\tfolder_output = folder_name + '/' + t1 + '_matrices'\n\t\tif not os.path.exists(folder_output):\n\t\t\tos.makedirs(folder_output)\n\t\t\n\t\tfolder_input = 'ALL_MOVIES_RAW/' + t1 + '/' + t1 \n\t\tcontainer = av.open(folder_input + '.' + extension)\n\t\tfor frame in container.decode(video=0):\n\t\t\tframe_img = frame.to_image()\n\t\t\tframe_npy = np.array(frame_img)\n\t\t\tmax_li = [] \n\t\t\tnum_frames = frame_npy.shape[2]\n\t\t\tif num_frames > 0:\n\t\t\t\tarr = 0.2989 * frame_npy[:,:,0] + 0.5870 * frame_npy[:,:,1] + 0.1140 * frame_npy[:,:,2]\n\t\t\telse:\n\t\t\t\tarr = frame_npy\n\t\n\t\t\tnp.save(folder_output + '/frame-%04d' % frame.index, arr)\n\t\n\t\t# add a png of the image just to see -- not necessary \n\t\tplt.figure()\n\t\tplt.imshow(arr)\n\t\tplt.axis('off')\n\t\tplt.title(t1)\n\t\tplt.savefig(folder_output + '/sample_image.png')\n\n##########################################################################################\ndef file_pre_processing_tif(path2file, file_name, file_format, video, rgb):\n\tfolder_output = create_output_folder(file_name)\n\t\n\tif file_format in ['tif', 'tiff']:\n\t\twith Image.open(path2file + file_name + '.' + file_format) as im:\n\t\t\timarray = np.array(im)\n\t\n\tif not video:\n\t\timarray = imarray.reshape(1,*imarray.shape)\n\tfor frame_num in range(imarray.shape[0]):\n\t\tif rgb:\n\t\t\tframe = 0.2989 * imarray[frame_num,:,:,0] + \\\n\t\t\t\t\t\t0.5870 * imarray[frame_num,:,:,1] + 0.1140 * imarray[frame_num,:,:,2]\n\t\telse:\n\t\t\tframe = imarray[frame_num,:,:]\n\t\tnp.savetxt(folder_output + '/frame-%04d' % (frame_num) + '.txt', frame, fmt='%.5e')\n\t\n\tplt.figure()\n\tplt.imshow(frame)\n\tplt.axis('off')\n\tplt.title(file_name)\n\tplt.savefig(folder_output + '/sample_image.png')\n\n##########################################################################################\ndef file_pre_processing_Kehan(file_name, file_source, is_avi):\n\tfolder_name = 'ALL_MOVIES_MATRICES'\n\tif not os.path.exists(folder_name):\n\t\tos.makedirs(folder_name)\t\n\t\n\tfolder_output = folder_name + '/' + file_name + '_matrices'\n\tif not os.path.exists(folder_output):\n\t\tos.makedirs(folder_output)\n\t\n\t# read image\n\tif is_avi:\n\t\tcontainer = av.open('Kehan_Movies/' + file_source + '.avi')\n\t\tfor frame in container.decode(video=0):\n\t\t\tframe_img = frame.to_image()\n\t\t\tframe_npy = np.array(frame_img)\n\t\t\tmax_li = [] \n\t\t\tnum_frames = frame_npy.shape[2]\n\t\t\tif num_frames > 0:\n\t\t\t\tarr = 0.2989 * frame_npy[:,:,0] + 0.5870 * frame_npy[:,:,1] + 0.1140 * frame_npy[:,:,2]\n\t\t\telse:\n\t\t\t\tarr = frame_npy\n\t\t\t\n\t\t\tnp.save(folder_output + '/frame-%04d' % frame.index, arr)\n\telse:\n\t\tim = io.imread('Kehan_Movies/' + file_source + '.tif')\n\t\tfor kk in range(0,im.shape[0]):\n\t\t\tframe_npy = im[kk,:,:]\n\t\t\tnp.save(folder_output + '/frame-%04d' % (kk), frame_npy)\n\t\n\tplt.figure()\n\tplt.imshow(frame_npy)\n\tplt.axis('off')\n\tplt.title(file_name)\n\tplt.savefig(folder_output + '/sample_image.png')\n\n##########################################################################################\ndef file_pre_processing_Kehan_timelapse(file_name, file_source, channel):\n\tfolder_name = 'ALL_MOVIES_MATRICES'\n\tif not os.path.exists(folder_name):\n\t\tos.makedirs(folder_name)\t\n\t\n\tfolder_output = folder_name + '/' + file_name + '_channel_%i'%(channel) + '_matrices'\n\tif not os.path.exists(folder_output):\n\t\tos.makedirs(folder_output)\n\t\n\tim = io.imread( file_source + '.tif')\n\t\n\tfor kk in range(0,im.shape[0]):\n\t\tframe_npy = im[kk,channel,:,:]\n\t\tnp.save(folder_output + '/frame-%04d' % (kk), frame_npy)\n\t\n\tplt.figure()\n\tplt.imshow(frame_npy)\n\tplt.axis('off')\n\tplt.title(file_name)\n\tplt.savefig(folder_output + '/sample_image.png')\n\t\n##########################################################################################\ndef get_frame_matrix(folder_name, frame):\n\t\"\"\"Get the npy matrix for a frame of the movie.\"\"\"\n\tif frame < 10: file_root = '_matrices/frame-000%i'%(frame)\n\telif frame < 100: file_root = '_matrices/frame-00%i'%(frame)\n\telse: file_root = '_matrices/frame-0%i'%(frame)\n\troot = 'ALL_MOVIES_MATRICES/' + folder_name + file_root + '.npy'\n\traw_img = np.load(root)\n\treturn raw_img\n\n##########################################################################################\ndef make_movie_from_npy(file_name,include_eps=False): \n\tfolder_name = 'ALL_MOVIES_MATRICES'\n\tfolder_output = folder_name + '/' + file_name + '_matrices'\n\tfolder_output_movie = folder_name + '/' + file_name + '_matrices/movie'\n\tif not os.path.exists(folder_output_movie):\n\t\tos.makedirs(folder_output_movie)\n\t\n\timg_list = [] \n\tnum_frames = len(glob.glob('ALL_MOVIES_MATRICES/' + file_name + '_matrices/*.npy'))\n\tfor kk in range(0,num_frames):\n\t\traw_img = get_frame_matrix(file_name, kk)\n\t\tplt.figure()\n\t\tplt.imshow(raw_img, cmap=plt.cm.gray)\n\t\tax = plt.gca()\n\t\tax.set_xticks([]); ax.set_yticks([])\n\t\tplt.savefig(folder_output_movie + '/' + 'frame_%04d.png'%(kk),bbox_inches = 'tight', pad_inches = 0)\n\t\tif include_eps:\n\t\t\tplt.savefig(folder_output_movie + '/' + 'frame_%i.eps'%(kk),bbox_inches = 'tight', pad_inches = 0)\n\t\tplt.close()\n\t\timg_list.append(imageio.imread(folder_output_movie + '/' + 'frame_%04d.png'%(kk)))\n\t\n\timageio.mimsave(folder_output_movie + '/contract_anim.gif', img_list)\t\n\tclip = mp.VideoFileClip(folder_output_movie + '/contract_anim.gif')\n\tclip.write_videofile( folder_output_movie + '/' + file_name + '.mp4')\n\t\n\treturn\n\t\n##########################################################################################\ndef create_output_folder(file_name):\n\tfolder_name = 'ALL_MOVIES_MATRICES'\n\tif not os.path.exists(folder_name):\n\t\tos.makedirs(folder_name)\t\n\t\n\tfolder_output = folder_name + '/' + file_name + '_matrices'\n\tif not os.path.exists(folder_output):\n\t\tos.makedirs(folder_output)\n\n\treturn folder_output\n", "sub_path": "file_pre_processing.py", "file_name": "file_pre_processing.py", "file_ext": "py", "file_size_in_byte": 6770, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.exists", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 32, "usage_type": "call"}, {"api_name": "os.path", "line_number": 32, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 33, "usage_type": "call"}, {"api_name": "av.open", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 47, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 61, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 61, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 72, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 74, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 74, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 75, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 75, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 76, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 76, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 77, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 77, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 78, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 84, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 87, "usage_type": "call"}, {"api_name": "os.path", "line_number": 87, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 88, "usage_type": "call"}, {"api_name": "av.open", "line_number": 92, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 95, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 103, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 105, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 105, "usage_type": "name"}, {"api_name": "numpy.save", "line_number": 108, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 110, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 110, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 111, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 111, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 112, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 112, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 113, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 113, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 114, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 114, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 119, "usage_type": "call"}, {"api_name": "os.path", "line_number": 119, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path", "line_number": 123, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 124, "usage_type": "call"}, {"api_name": "skimage.io.imread", "line_number": 126, "usage_type": "call"}, {"api_name": "skimage.io", "line_number": 126, "usage_type": "name"}, {"api_name": "numpy.save", "line_number": 130, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 132, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 132, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 133, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 133, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.axis", "line_number": 134, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 134, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 135, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 135, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 136, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 136, "usage_type": "name"}, {"api_name": "numpy.load", "line_number": 145, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 154, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 157, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 160, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 160, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 161, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 161, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.cm", "line_number": 161, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.gca", "line_number": 162, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 162, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 164, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 164, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 166, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 166, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.close", "line_number": 167, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 167, "usage_type": "name"}, {"api_name": "imageio.imread", "line_number": 168, "usage_type": "call"}, {"api_name": "imageio.mimsave", "line_number": 170, "usage_type": "call"}, {"api_name": "moviepy.editor.VideoFileClip", "line_number": 171, "usage_type": "call"}, {"api_name": "moviepy.editor", "line_number": 171, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 179, "usage_type": "call"}, {"api_name": "os.path", "line_number": 179, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 180, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 184, "usage_type": "call"}]} +{"seq_id": "139736563", "text": "import json\nimport base64\nimport socket\nimport sys\nimport uuid\nfrom logging import getLogger\n\nfrom common import common\n\nlogger = getLogger(__name__)\n\ndef callService(content, interest, host, service, revSize):\n\n logger.debug(\"[callService] start calling service function\")\n\n if service['input'] == \"only content\":\n if content != b'':\n contentName = common.getContentName(interest)\n temp = json.loads(content)\n temp = temp[contentName]['content']['value']\n inputData = base64.b64decode(temp.encode('utf-8'))\n\n else:\n if content != b'':\n inputData = content.encode()\n\n port = int(service['port'])\n\n with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:\n\n s.connect((host, port))\n\n #sending input data\n if content != b'':\n s.sendall(inputData)\n s.shutdown(1)\n ########\n\n #receiving data from service function\n revData = b''\n data = s.recv(revSize)\n revData += data\n\n if sys.getsizeof(data) > revSize:\n\n while True:\n data = s.recv(revSize)\n #revData += data\n #print(sys.getsizeof(data))\n #if sys.getsizeof(data) < revSize:\n # revData += data\n # break\n if not data:\n break \n revData += data\n\n result = revData.decode()\n #########\n\n logger.debug(\"[callService] complete\")\n\n return result\n\n\ndef concatData(rawData, procData, funcName):\n\n logger.debug(\"[concatData] data serialization\")\n\n BODY = []\n\n if funcName == \"yolo\" or funcName == \"yologpu\" or funcName == \"yolotiny\":\n\n YOLO = []\n if (procData != \"null\"):\n\n temp = procData.split(\"_\")\n numDetect = len(temp)\n\n for i in range(numDetect-1):\n\n temp2 = temp[i].split(\",\")\n\n left = temp2[2]\n right = temp2[3]\n top = temp2[4]\n bottom = temp2[5]\n width = right - left\n height = bottom - top\n\n dict_body = {'tagName': temp2[0],\n 'tagID': str(uuid.uuid4()),\n 'probability': float(temp2[1])/100.0,\n 'boundingBox': {'left': left,\n 'top': top,\n 'width': width,\n 'height': height\n }\n }\n YOLO.append(dict_body)\n\n procData = YOLO\n\n if type(procData) is str:\n procData = json.loads(procData)\n\n if type(rawData) is str:\n rawData = json.loads(rawData)\n\n BODY = rawData\n BODY.update({funcName: procData})\n\n BODY = json.dumps(BODY)\n\n #logger.info(\"[concatData] data {}\".format(BODY))\n\n logger.debug(\"[concatData] complete!\")\n\n return BODY\n", "sub_path": "Test/VirIoT-TVF/protocol/callSF.py", "file_name": "callSF.py", "file_ext": "py", "file_size_in_byte": 3051, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 10, "usage_type": "call"}, {"api_name": "common.common.getContentName", "line_number": 18, "usage_type": "call"}, {"api_name": "common.common", "line_number": 18, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 19, "usage_type": "call"}, {"api_name": "base64.b64decode", "line_number": 21, "usage_type": "call"}, {"api_name": "socket.socket", "line_number": 29, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 29, "usage_type": "attribute"}, {"api_name": "socket.SOCK_STREAM", "line_number": 29, "usage_type": "attribute"}, {"api_name": "sys.getsizeof", "line_number": 44, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 91, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 104, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 107, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 112, "usage_type": "call"}]} +{"seq_id": "371471303", "text": "import math\n\n\ndef log_incorrect1(x, b):\n return 1\n \n \ndef log_incorrect2(x, b):\n if b == 1:\n raise ValueError()\n return 1\n\n\ndef log_incorrect3(x, b):\n if b == 1 or x <= 0:\n raise ValueError()\n return 1\n\n\ndef log_incorrect4(x, b):\n if b == 1 or x <= 0:\n raise ValueError()\n if x == 1:\n return 0\n return 1\n\n\ndef log(x, b):\n # return log_incorrect1(x, b)\n # return log_incorrect2(x, b)\n # return log_incorrect3(x, b)\n # return log_incorrect4(x, b)\n return math.log(x, b)\n\n\nimport pytest \nimport hypothesis.strategies as st\nfrom hypothesis import given, assume, example, settings\n\n@pytest.mark.parametrize(\"x,b,o\", [(4,2,2), (27,3,3), (81,3,4), (1024,2,10)])\ndef test_examples(x, b, o):\n assert log(x, b) == o\n\n\ndef assert_almost_equal(v1, v2, accuracy=1e6):\n assert pytest.approx(v1, accuracy) == v2\n\n\nSMALLEST_FLOAT=1e-6\nLARGEST_FLOAT=1e6\n\n\nwith settings(max_examples=500, min_satisfying_examples=500):\n \n # log of x to base x is 1\n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p1(x):\n assert log(x, x) == 1\n \n \n # log of 1 at any base (except 1) is 0\n @given(st.floats(min_value=SMALLEST_FLOAT))\n def test_p2(b):\n assume(b != 1)\n assert log(1, b) == 0\n \n \n # logarithm of negative values is undefined\n @given(st.floats(max_value=0), \n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p3(x, b):\n assume(b != 1)\n with pytest.raises(ValueError):\n log(x, b)\n \n \n # log(x, b) = n implies b ** n = x \n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p4(x, b):\n assume(b != 1)\n assert_almost_equal(b ** log(x, b), x)\n \n \n # log(x * y) = log(x) + log(y)\n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p5(x, y, b):\n assume(b != 1)\n tmp1 = x * y\n assume(tmp1 > 0)\n tmp2 = log(tmp1, b)\n assert_almost_equal(tmp2, (log(x, b) + log(y, b)))\n \n \n # log(x / y) = log(x) - log(y)\n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT), \n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p6(x, y, b):\n assume(b != 1)\n tmp1 = x / y\n assume(tmp1 > 0)\n tmp2 = log(tmp1, b)\n assert_almost_equal(tmp2, (log(x, b) - log(y, b)))\n \n \n # log(x ** y) = y * log(x)\n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p7(x, y, b):\n assume(b != 1)\n try:\n tmp1 = x ** y\n assume(tmp1 > 0)\n tmp2 = log(tmp1, b)\n assert_almost_equal(tmp2, y * log(x, b))\n except OverflowError:\n print(\"Over flow\")\n pass \n \n \n # log(x, c) / log(b, c) = log(x, b)\n @given(st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT),\n st.floats(min_value=SMALLEST_FLOAT, max_value=LARGEST_FLOAT))\n def test_p8(x, b, c):\n assume(c != 1)\n tmp1 = (log(x, c) / log(b, c))\n assert_almost_equal(tmp1, log(x, b))\n", "sub_path": "content/Property Based Testing/code-snippets-2/test_logarithm.py", "file_name": "test_logarithm.py", "file_ext": "py", "file_size_in_byte": 3733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "math.log", "line_number": 33, "usage_type": "call"}, {"api_name": "pytest.mark.parametrize", "line_number": 40, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pytest.approx", "line_number": 46, "usage_type": "call"}, {"api_name": "hypothesis.settings", "line_number": 53, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 56, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 56, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 56, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 64, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 62, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 62, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 62, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 72, "usage_type": "call"}, {"api_name": "pytest.raises", "line_number": 73, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 69, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 69, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 69, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 70, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 70, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 81, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 78, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 78, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 78, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 79, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 79, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 90, "usage_type": "call"}, {"api_name": "hypothesis.assume", "line_number": 92, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 86, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 86, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 86, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 87, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 87, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 88, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 88, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 102, "usage_type": "call"}, {"api_name": "hypothesis.assume", "line_number": 104, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 98, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 98, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 98, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 99, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 99, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 100, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 100, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 114, "usage_type": "call"}, {"api_name": "hypothesis.assume", "line_number": 117, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 110, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 110, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 110, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 111, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 111, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 112, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 112, "usage_type": "name"}, {"api_name": "hypothesis.assume", "line_number": 130, "usage_type": "call"}, {"api_name": "hypothesis.given", "line_number": 126, "usage_type": "call"}, {"api_name": "hypothesis.strategies.floats", "line_number": 126, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 126, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 127, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 127, "usage_type": "name"}, {"api_name": "hypothesis.strategies.floats", "line_number": 128, "usage_type": "call"}, {"api_name": "hypothesis.strategies", "line_number": 128, "usage_type": "name"}]} +{"seq_id": "194707179", "text": "import tweets as twt\nfrom curator import DataCurator\n\ntweets = twt.getTweets()\n\ndc = DataCurator(tweets)\nclean_data = dc.basicCleanUp().removeRepeatedPunctuations().removePunctuationsAndReTweet()\\\n .expandPunctuatedWords().expandShortHands().build()\n\nf = open('cleaned_data.txt', 'w', 1024) \nfor tweet in clean_data:\n f.write(\"%s\\n\" % (tweet))\nf.close()\n", "sub_path": "q1.py", "file_name": "q1.py", "file_ext": "py", "file_size_in_byte": 379, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "tweets.getTweets", "line_number": 4, "usage_type": "call"}, {"api_name": "curator.DataCurator", "line_number": 6, "usage_type": "call"}]} +{"seq_id": "291192628", "text": "\"\"\"\nASGI spec conformance test suite.\n\nCalling the functions with an ASGI channel layer instance will return you a\nsingle TestCase instance that checks for conformity on that instance.\n\nYou MUST also pass along an expiry value to the sets of tests, to allow the\nsuite to wait correctly for expiry. It's suggested you configure the layer\nfor 1-second expiry during tests, and use a 1.1 second expiry delay.\n\nThe channel layers should be empty to start with, and discarded after use,\nas they'll be full of test data. If a layer supports the \"flush\" extension,\nit'll be flushed before every test.\n\"\"\"\n\nfrom __future__ import unicode_literals\nimport six\nimport time\nimport unittest\n\n\ndef make_tests(channel_layer, expiry_delay):\n\n class LayerTestCase(unittest.TestCase):\n \"\"\"\n Tests that core ASGI functionality is maintained.\n \"\"\"\n\n def setUp(self):\n if \"flush\" in channel_layer.extensions:\n channel_layer.flush()\n\n def test_send_recv(self):\n \"\"\"\n Tests that channels can send and receive messages right.\n \"\"\"\n channel_layer.send(\"sr_test\", {\"value\": \"blue\"})\n channel_layer.send(\"sr_test\", {\"value\": \"green\"})\n channel_layer.send(\"sr_test2\", {\"value\": \"red\"})\n # Get just one first\n channel, message = channel_layer.receive_many([\"sr_test\"])\n self.assertEqual(channel, \"sr_test\")\n self.assertEqual(message, {\"value\": \"blue\"})\n # And the second\n channel, message = channel_layer.receive_many([\"sr_test\"])\n self.assertEqual(channel, \"sr_test\")\n self.assertEqual(message, {\"value\": \"green\"})\n # And the other channel with multi select\n channel, message = channel_layer.receive_many([\"sr_test\", \"sr_test2\"])\n self.assertEqual(channel, \"sr_test2\")\n self.assertEqual(message, {\"value\": \"red\"})\n\n def test_unicode_channel_name(self):\n \"\"\"\n Makes sure channel names can handle unicode\n \"\"\"\n channel_layer.send(\"\\u00a3_test\", {\"value\": \"blue\"})\n # Get just one first\n channel, message = channel_layer.receive_many([\"\\u00a3_test\"])\n self.assertEqual(channel, \"\\u00a3_test\")\n self.assertEqual(message, {\"value\": \"blue\"})\n\n def test_message_expiry(self):\n \"\"\"\n Tests that messages expire correctly.\n \"\"\"\n channel_layer.send(\"me_test\", {\"value\": \"blue\"})\n time.sleep(expiry_delay)\n channel, message = channel_layer.receive_many([\"me_test\"])\n self.assertIs(channel, None)\n self.assertIs(message, None)\n\n def test_new_channel(self):\n \"\"\"\n Tests that new channel names are made correctly.\n \"\"\"\n pattern = \"test.?.foo.?\"\n name1 = channel_layer.new_channel(pattern)\n self.assertIsInstance(name1, six.text_type)\n # Send a message and make sure new_channel on second pass changes\n channel_layer.send(name1, {\"value\": \"blue\"})\n name2 = channel_layer.new_channel(pattern)\n # Make sure the two ?s are replaced by the same string\n bits = name2.split(\".\")\n self.assertEqual(bits[1], bits[3], \"New channel random strings don't match\")\n # Make sure we can consume off of that new channel\n channel, message = channel_layer.receive_many([name1, name2])\n self.assertEqual(channel, name1)\n self.assertEqual(message, {\"value\": \"blue\"})\n\n def test_strings(self):\n \"\"\"\n Ensures byte strings and unicode strings both make it through\n serialization properly.\n \"\"\"\n # Message. Double-nested to ensure serializers are recursing properly.\n message = {\n \"values\": {\n # UTF-8 sequence for british pound, but we want it not interpreted into that.\n \"utf-bytes\": b\"\\xc2\\xa3\",\n # Actual unicode for british pound, should come back as 1 char\n \"unicode\": \"\\u00a3\",\n # Emoji, in case someone is using 3-byte-wide unicode storage\n \"emoji\": \"\\u1F612\",\n # Random control characters and null\n \"control\": b\"\\x01\\x00\\x03\\x21\",\n }\n }\n # Send it and receive it\n channel_layer.send(\"str_test\", message)\n _, received = channel_layer.receive_many([\"str_test\"])\n # Compare\n self.assertIsInstance(received[\"values\"][\"utf-bytes\"], six.binary_type)\n self.assertIsInstance(received[\"values\"][\"unicode\"], six.text_type)\n self.assertIsInstance(received[\"values\"][\"emoji\"], six.text_type)\n self.assertIsInstance(received[\"values\"][\"control\"], six.binary_type)\n self.assertEqual(received[\"values\"][\"utf-bytes\"], message[\"values\"][\"utf-bytes\"])\n self.assertEqual(received[\"values\"][\"unicode\"], message[\"values\"][\"unicode\"])\n self.assertEqual(received[\"values\"][\"emoji\"], message[\"values\"][\"emoji\"])\n self.assertEqual(received[\"values\"][\"control\"], message[\"values\"][\"control\"])\n\n @unittest.skipIf(\"groups\" not in channel_layer.extensions, \"No groups extension\")\n def test_groups(self):\n \"\"\"\n Tests that basic group addition and send works\n \"\"\"\n # Make a group and send to it\n channel_layer.group_add(\"tgroup\", \"tg_test\")\n channel_layer.group_add(\"tgroup\", \"tg_test2\")\n channel_layer.group_add(\"tgroup\", \"tg_test3\")\n channel_layer.group_discard(\"tgroup\", \"tg_test3\")\n channel_layer.send_group(\"tgroup\", {\"value\": \"orange\"})\n # Receive from the two channels in the group and ensure messages\n channel, message = channel_layer.receive_many([\"tg_test\"])\n self.assertEqual(channel, \"tg_test\")\n self.assertEqual(message, {\"value\": \"orange\"})\n channel, message = channel_layer.receive_many([\"tg_test2\"])\n self.assertEqual(channel, \"tg_test2\")\n self.assertEqual(message, {\"value\": \"orange\"})\n # Make sure another channel does not get a message\n channel, message = channel_layer.receive_many([\"tg_test3\"])\n self.assertIs(channel, None)\n self.assertIs(message, None)\n\n @unittest.skipIf(\"flush\" not in channel_layer.extensions, \"No flush extension\")\n def test_flush(self):\n \"\"\"\n Tests that messages go away after a flush.\n \"\"\"\n channel_layer.send(\"fl_test\", {\"value\": \"blue\"})\n channel_layer.flush()\n channel, message = channel_layer.receive_many([\"fl_test\"])\n self.assertIs(channel, None)\n self.assertIs(message, None)\n\n @unittest.skipIf(\"flush\" not in channel_layer.extensions, \"No flush extension\")\n @unittest.skipIf(\"groups\" not in channel_layer.extensions, \"No groups extension\")\n def test_flush_groups(self):\n \"\"\"\n Tests that groups go away after a flush.\n \"\"\"\n channel_layer.send(\"fl_test\", {\"value\": \"blue\"})\n channel_layer.flush()\n channel, message = channel_layer.receive_many([\"fl_test\"])\n self.assertIs(channel, None)\n self.assertIs(message, None)\n\n return LayerTestCase\n", "sub_path": "Lib/site-packages/asgiref/conformance.py", "file_name": "conformance.py", "file_ext": "py", "file_size_in_byte": 7581, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "unittest.TestCase", "line_number": 24, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 68, "usage_type": "call"}, {"api_name": "six.text_type", "line_number": 79, "usage_type": "attribute"}, {"api_name": "six.binary_type", "line_number": 113, "usage_type": "attribute"}, {"api_name": "six.text_type", "line_number": 114, "usage_type": "attribute"}, {"api_name": "six.text_type", "line_number": 115, "usage_type": "attribute"}, {"api_name": "six.binary_type", "line_number": 116, "usage_type": "attribute"}, {"api_name": "unittest.skipIf", "line_number": 122, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 145, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 156, "usage_type": "call"}, {"api_name": "unittest.skipIf", "line_number": 157, "usage_type": "call"}]} +{"seq_id": "593716935", "text": "from models.base_model import db\nfrom marshmallow import Schema, fields, post_dump, EXCLUDE\n\n# Helper table to tag experiences with skills\nexperience_skills = db.Table('experience_skills',\n db.Column('experience_id', db.Integer, db.ForeignKey('experience.id')),\n db.Column('skill_id', db.String, index=True),\n db.Column('contact_id', db.Integer),\n db.PrimaryKeyConstraint(\n 'experience_id',\n 'skill_id',\n 'contact_id',\n name='experience_skills_pk'),\n db.ForeignKeyConstraint(['skill_id', 'contact_id'],\n ['skill_item.id', 'skill_item.contact_id']),\n)\n\nclass SkillItem(db.Model):\n __tablename__ = 'skill_item'\n\n #table columns\n id = db.Column(db.String, nullable=False)\n name = db.Column(db.String, nullable=False)\n contact_id = db.Column(db.Integer, db.ForeignKey('contact.id'), nullable=False)\n\n #relationships\n contact = db.relationship('Contact')\n experiences = db.relationship('Experience',\n secondary=experience_skills,\n lazy='subquery')\n\n __table_args__ = (\n db.PrimaryKeyConstraint('id', 'contact_id', name='skill_item_pk'),\n )\n\n\nclass SkillItemSchema(Schema):\n id = fields.String(dump_only=True)\n name = fields.String(required=True)\n contact_id = fields.Integer(dump_only=True)\n\n class Meta:\n unknown = EXCLUDE\n", "sub_path": "models/old_skill_model.py", "file_name": "old_skill_model.py", "file_ext": "py", "file_size_in_byte": 1411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "models.base_model.db.Table", "line_number": 5, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 5, "usage_type": "name"}, {"api_name": "models.base_model.db.Column", "line_number": 6, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 6, "usage_type": "name"}, {"api_name": "models.base_model.db.Integer", "line_number": 6, "usage_type": "attribute"}, {"api_name": "models.base_model.db.ForeignKey", "line_number": 6, "usage_type": "call"}, {"api_name": "models.base_model.db.Column", "line_number": 7, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 7, "usage_type": "name"}, {"api_name": "models.base_model.db.String", "line_number": 7, "usage_type": "attribute"}, {"api_name": "models.base_model.db.Column", "line_number": 8, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 8, "usage_type": "name"}, {"api_name": "models.base_model.db.Integer", "line_number": 8, "usage_type": "attribute"}, {"api_name": "models.base_model.db.PrimaryKeyConstraint", "line_number": 9, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 9, "usage_type": "name"}, {"api_name": "models.base_model.db.ForeignKeyConstraint", "line_number": 14, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 14, "usage_type": "name"}, {"api_name": "models.base_model.db.Model", "line_number": 18, "usage_type": "attribute"}, {"api_name": "models.base_model.db", "line_number": 18, "usage_type": "name"}, {"api_name": "models.base_model.db.Column", "line_number": 22, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 22, "usage_type": "name"}, {"api_name": "models.base_model.db.String", "line_number": 22, "usage_type": "attribute"}, {"api_name": "models.base_model.db.Column", "line_number": 23, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 23, "usage_type": "name"}, {"api_name": "models.base_model.db.String", "line_number": 23, "usage_type": "attribute"}, {"api_name": "models.base_model.db.Column", "line_number": 24, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 24, "usage_type": "name"}, {"api_name": "models.base_model.db.Integer", "line_number": 24, "usage_type": "attribute"}, {"api_name": "models.base_model.db.ForeignKey", "line_number": 24, "usage_type": "call"}, {"api_name": "models.base_model.db.relationship", "line_number": 27, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 27, "usage_type": "name"}, {"api_name": "models.base_model.db.relationship", "line_number": 28, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 28, "usage_type": "name"}, {"api_name": "models.base_model.db.PrimaryKeyConstraint", "line_number": 33, "usage_type": "call"}, {"api_name": "models.base_model.db", "line_number": 33, "usage_type": "name"}, {"api_name": "marshmallow.Schema", "line_number": 37, "usage_type": "name"}, {"api_name": "marshmallow.fields.String", "line_number": 38, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 38, "usage_type": "name"}, {"api_name": "marshmallow.fields.String", "line_number": 39, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 39, "usage_type": "name"}, {"api_name": "marshmallow.fields.Integer", "line_number": 40, "usage_type": "call"}, {"api_name": "marshmallow.fields", "line_number": 40, "usage_type": "name"}, {"api_name": "marshmallow.EXCLUDE", "line_number": 43, "usage_type": "name"}]} +{"seq_id": "213282810", "text": "import numpy as np\nimport scipy as sp\nimport scipy.constants\nimport cPickle\nfrom bunch import Bunch\n\nimport echolect as el\nimport radarmodel\n\nmfblksize = 5\nmfvoters = [1, 2, 4]\n\nbasefilename = 'equatorial_example'\nwith open(basefilename + '.pkl', 'rb') as f:\n data = cPickle.load(f)\n\nn = 1\nfreqs = np.fft.fftfreq(int(n), data.ts/np.timedelta64(1, 's'))\nv = freqs/data.f0*sp.constants.c/2\n\nrslc = el.slice_by_value(data.r, 80000, 140000)\nr = data.r[rslc]\nm = r.shape[0]\n\nfilts = []\nfor code, delay in zip(data.codes, data.code_delays):\n s = (code/np.linalg.norm(code)).astype(data.vlt.dtype)\n filt = el.filtering.MatchedDoppler(s, n, data.vlt.shape[-1], xdtype=data.vlt.dtype)\n filt.nodelay = slice(filt.L - 1 - delay, filt.L - 1 - delay + filt.M)\n filts.append(filt)\n\nvlt_mf_all = np.zeros((mfblksize, n, m), data.vlt.dtype)\nvlt_mf = np.zeros((data.vlt.shape[0], m), data.vlt.dtype)\nfreq = np.zeros(data.vlt.shape[0])\nfor kp in xrange(data.vlt.shape[0]):\n y = data.vlt[kp]\n filt = filts[kp % len(filts)]\n x = filt(y)\n xnodelay = x[:, filt.nodelay]\n vlt_mf_all[kp % mfblksize] = xnodelay[:, rslc]\n \n if ((kp + 1) % mfblksize) == 0:\n # get the frequency shift that gives max SNR for each pulse in data block\n shifts = np.zeros(len(mfvoters), 'int8')\n for ks, kmf in enumerate(mfvoters):\n vlt = vlt_mf_all[kmf]\n shifts[ks] = np.unravel_index(np.argmax(vlt.real**2 + vlt.imag**2), vlt.shape)[0]\n \n # wrap high positive shifts to negative, so median works near 0\n shifts = (shifts + n/2) % n - n/2\n shift = np.median(shifts)\n \n # store matched filter data for selected shift\n for ks in xrange(mfblksize):\n k = kp + 1 - mfblksize + ks\n vlt_mf[k] = vlt_mf_all[ks, shift]\n freq[k] = float(shift)/n*(np.timedelta64(1, 's')/data.ts)\n\nmf = Bunch(vlt=vlt_mf, t=data.t, r=r, freq=freq, n=n, ts=data.ts, \n ipp=data.ipp, f0=data.f0, noise_sigma=data.noise_sigma)\nwith open(basefilename + '_mf.pkl', 'wb') as f:\n cPickle.dump(mf, f, protocol=-1)", "sub_path": "code/equatorial_example_mf.py", "file_name": "equatorial_example_mf.py", "file_ext": "py", "file_size_in_byte": 2103, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "cPickle.load", "line_number": 15, "usage_type": "call"}, {"api_name": "numpy.fft.fftfreq", "line_number": 18, "usage_type": "call"}, {"api_name": "numpy.fft", "line_number": 18, "usage_type": "attribute"}, {"api_name": "numpy.timedelta64", "line_number": 18, "usage_type": "call"}, {"api_name": "scipy.constants", "line_number": 19, "usage_type": "attribute"}, {"api_name": "echolect.slice_by_value", "line_number": 21, "usage_type": "call"}, {"api_name": "numpy.linalg.norm", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.linalg", "line_number": 27, "usage_type": "attribute"}, {"api_name": "echolect.filtering.MatchedDoppler", "line_number": 28, "usage_type": "call"}, {"api_name": "echolect.filtering", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 32, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.unravel_index", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.timedelta64", "line_number": 57, "usage_type": "call"}, {"api_name": "bunch.Bunch", "line_number": 59, "usage_type": "call"}, {"api_name": "cPickle.dump", "line_number": 62, "usage_type": "call"}]} +{"seq_id": "31836814", "text": "from bson.objectid import ObjectId\nfrom flask import request, jsonify, json, Response\n\nfrom CommonUtilities import *\n\n\nclass CrudOperations:\n\n # Creating object of CommonUtilities class\n commonUtils = CommonUtilities()\n\n product_collection = None\n\n \"\"\"\n Read the values from property files and store it to the respective variables\n \"\"\"\n sDbName = commonUtils.getDbDetails('DB_DETAILS', 'db.name')\n sDbUsername = commonUtils.getDbDetails('DB_DETAILS', 'db.username')\n sDbPassword = commonUtils.getDbDetails('DB_DETAILS', 'db.password')\n sCollectionName = commonUtils.getDbDetails('DB_DETAILS', 'db.collection.name')\n\n\n \"\"\"\n This piece of code will run when the Object of the class is created. \n If no errors in reading values from Properties file then connect to the Mongo DB cloud else set the variable value \n \"product_collection\" to None.\n On successfull connection to Mongo DB cloud, we get the collection Object in return\n \"\"\"\n if (sDbName is not None) and (sDbUsername is not None) and (sDbPassword is not None) and (\n sCollectionName is not None):\n sConnectionString = \"mongodb://{0}:{1}@productcluster-shard-00-00.jcvab.mongodb.net:27017,\" \\\n \"productcluster-shard-00-01.jcvab.mongodb.net:27017,\" \\\n \"productcluster-shard-00-02.jcvab.mongodb.net:27017/\" \\\n \"{2}?ssl=true&replicaSet=atlas-14aexf-shard-0&authSource=admin\" \\\n \"&retryWrites=true&w=majority\".format(sDbUsername, sDbPassword, sDbName)\n connectMongoDbCloud = commonUtils.connectMongoDbCloud(sConnectionString=sConnectionString, sDbName=sDbName,\n sCollectionName=sCollectionName)\n if connectMongoDbCloud is not None:\n product_collection = connectMongoDbCloud\n else:\n product_collection = None\n else:\n product_collection = None\n print(\"FAILED to read the configuration data.\")\n\n def getAllProducts(self):\n \"\"\"\n This method is used to get all the documents from the connected product_collection object.\n :return: After getting documents from collection, It returns the entire document in json format.\n \"\"\"\n if request.method == 'GET':\n if self.product_collection is not None:\n all_docs = self.product_collection.find({})\n if all_docs.count() > 0:\n # print(\"Inside doc\", self.product_collection.count_documents({}))\n return Response(json.dumps(list(all_docs), default=str), mimetype=\"application/json\"), 200\n else:\n return jsonify(success=\"false\", message=\"No data found\"), 404\n else:\n return jsonify(success=\"false\", message=\"Unable to connect mongo DB :(\"), 500\n else:\n return jsonify(success=\"false\", message=\"Method not allowed.\"), 405\n\n def insertProduct(self):\n \"\"\"\n This method helps in inserting the data into the database\n :return: Returns id of newly created object\n \"\"\"\n if request.method == 'POST':\n if self.product_collection is not None:\n _id = ObjectId()\n try:\n new_product = {\n \"_id\": _id,\n \"brand_name\": request.json['brand_name'],\n \"classification_l1\": request.json['classification_l1'],\n \"classification_l2\": request.json['classification_l2'],\n \"classification_l3\": request.json['classification_l3'],\n \"classification_l4\": request.json['classification_l4'],\n \"currency\": request.json['currency'],\n \"image_url\": request.json['image_url'],\n \"name\": request.json['name'],\n \"offer_price_value\": int(request.json['offer_price_value']),\n \"regular_price_value\": int(request.json['regular_price_value'])\n }\n # print(new_product)\n self.product_collection.insert_one(new_product)\n return jsonify(success=True, message=\"Successfully inserted with id :\" + str(_id)), 201\n except:\n return jsonify(success=\"false\", message=\"Someting went wrong.\"), 500\n else:\n return jsonify(success=\"false\", message=\"Unable to connect mongo DB :(\"), 500\n else:\n return jsonify(success=\"false\", message=\"Method not allowed.\"), 405\n\n def operationOnSpecificProduct(self, idProduct):\n \"\"\"\n This method just calls respective method based on the requested method\n :param idProduct: ID of the product on which we need to manipulate\n :return: Incase of GET method - It return the object of specified ID, Incase of PUT and DELETE method - Just return success if manipulated successfully\n \"\"\"\n if request.method == 'GET':\n return self.getSingleProduct(idProduct)\n elif request.method == 'PUT':\n return self.updateSingleProduct(idProduct)\n elif request.method == 'DELETE':\n return self.deleteSingleProduct(idProduct)\n else:\n return jsonify(success=\"false\", message=\"Method not allowed.\"), 405\n\n def getSingleProduct(self, idProduct):\n \"\"\"\n This method find the product in the DB by id\n :param idProduct: ID of the product for which we need to send the detail of\n :return: return object of specified ID if found else No data found\n \"\"\"\n try:\n sProduct = self.product_collection.find({\"_id\": ObjectId(idProduct)})\n if sProduct.count() > 0:\n return Response(json.dumps(list(sProduct), default=str), mimetype=\"application/json\"), 200\n else:\n return jsonify(success=\"false\", message=\"No data found\")\n except:\n return jsonify(success=\"false\", message=\"Something went wrong.\"), 500\n\n def updateSingleProduct(self, idProduct):\n \"\"\"\n This method is used to update all the fields of particular object, first it finds the object in the DB\n if found then it updates it else No data found\n :param idProduct: id of the product on which we need to manipulate\n :return: returns success json if updated successfull else \"Something went wrong\" 500\n \"\"\"\n try:\n sProduct = self.product_collection.find({\"_id\": ObjectId(idProduct)})\n if sProduct.count() == 1:\n self.product_collection.update({\"_id\": ObjectId(idProduct)}, {\"$set\": {\n \"brand_name\": request.json['brand_name'],\n \"classification_l1\": request.json['classification_l1'],\n \"classification_l2\": request.json['classification_l2'],\n \"classification_l3\": request.json['classification_l3'],\n \"classification_l4\": request.json['classification_l4'],\n \"currency\": request.json['currency'],\n \"image_url\": request.json['image_url'],\n \"name\": request.json['name'],\n \"offer_price_value\": int(request.json['offer_price_value']),\n \"regular_price_value\": int(request.json['regular_price_value'])\n }})\n return jsonify(success=True, message=\"Successfully updated.\"), 200\n else:\n return jsonify(success=\"false\", message=\"No data found.\"), 404\n except:\n return jsonify(success=\"false\", message=\"Something went wrong.\"), 500\n\n def deleteSingleProduct(self, idProduct):\n \"\"\"\n This method is used to delete entire object from DB. First it will find the object in DB by ID\n if found then delete it else return No data found.\n :param idProduct: id of the product which we want to delete\n :return: return success json if deleted successfully.\n \"\"\"\n try:\n sProduct = self.product_collection.find({\"_id\": ObjectId(idProduct)})\n if sProduct.count() == 1:\n self.product_collection.delete_one({\"_id\": ObjectId(idProduct)})\n return jsonify(success=True, message=\"Successfully deleted.\"), 200\n else:\n return jsonify(success=\"false\", message=\"No data found.\"), 404\n except:\n return jsonify(success=\"false\", message=\"Something went wrong.\"), 500\n", "sub_path": "CrudOperations.py", "file_name": "CrudOperations.py", "file_ext": "py", "file_size_in_byte": 8630, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.request.method", "line_number": 51, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 51, "usage_type": "name"}, {"api_name": "flask.Response", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.json.dumps", "line_number": 56, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 56, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 58, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 60, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 62, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 69, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 69, "usage_type": "name"}, {"api_name": "bson.objectid.ObjectId", "line_number": 71, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 75, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 75, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 76, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 76, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 77, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 78, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 78, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 79, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 79, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 80, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 80, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 81, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 81, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 82, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 82, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 83, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 83, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 84, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 84, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 88, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 90, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 92, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 94, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 102, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 102, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 104, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.request.method", "line_number": 106, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 106, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 109, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 118, "usage_type": "call"}, {"api_name": "flask.Response", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.json.dumps", "line_number": 120, "usage_type": "call"}, {"api_name": "flask.json", "line_number": 120, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 122, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 124, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 134, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 136, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 137, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 137, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 138, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 138, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 139, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 139, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 140, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 140, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 143, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 144, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 144, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 148, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 150, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 152, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 162, "usage_type": "call"}, {"api_name": "bson.objectid.ObjectId", "line_number": 164, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 165, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 167, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 169, "usage_type": "call"}]} +{"seq_id": "187683868", "text": "\"\"\"\nEmail service\n\nAccept request from other controllers in the cluster to send email. The service doesn't accept request originates\noriginates out of the cluster therefore authentication is no needed.\n\"\"\"\n\nimport os\nimport logging\nfrom datetime import datetime\nfrom typing import Tuple\n\nimport requests\nimport coloredlogs\nfrom flask import Flask, jsonify, request\n\nfrom app.templates import render\nfrom app.util import is_healthy, send_email, http_get\n\napp = Flask(__name__) # pylint: disable=invalid-name\n\ncoloredlogs.install(level=logging.INFO)\nlogger = logging.getLogger('a01.svc.email') # pylint: disable=invalid-name\n\n\n@app.route('/health')\ndef health():\n status, remark = is_healthy()\n return jsonify({'status': status, 'time': datetime.utcnow(), 'remark': remark})\n\n\n@app.route('/report', methods=['POST'])\ndef send_report():\n logger.info('requested to send email')\n\n # parse input\n run_id = request.json['run_id']\n receivers = request.json['receivers']\n template_url = request.json.get('template', None)\n logger.info(f'run: {run_id} | receivers: {receivers} | template: {template_url or \"None\"}')\n\n # retrieve run and tasks\n run = http_get(f'run/{run_id}')\n tasks = sorted(http_get(f'run/{run_id}/tasks'), key=lambda t: t['status'])\n logger.info(f'successfully read run {run_id}.')\n logger.info(f'using template {template_url or \"unknown\"}.')\n\n # send email\n content, subject = render(run, tasks, template_url)\n send_email(receivers, subject, content)\n\n return jsonify({'status': 'done'})\n\n\ndef download_template(uri: str, product: str) -> Tuple[str, str, str]:\n template_dir = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'templates')\n if uri:\n template_local_path = os.path.join(template_dir, f'{product}.html')\n try:\n resp = requests.get(uri)\n resp.raise_for_status()\n with open(template_local_path, 'w') as handler:\n handler.write(resp.text)\n return template_dir, f'{product}.html', product\n except requests.HTTPError:\n logger.exception('Fail to request template file.')\n except IOError:\n logger.exception('Fail to write template file.')\n else:\n logger.warning(\"Template URI is empty\")\n\n return template_dir, 'generic.html', 'generic'\n", "sub_path": "services/email/app/app/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 2341, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "app.templates", "line_number": 20, "usage_type": "name"}, {"api_name": "flask.Flask", "line_number": 20, "usage_type": "call"}, {"api_name": "coloredlogs.install", "line_number": 22, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 22, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 23, "usage_type": "call"}, {"api_name": "app.util.is_healthy", "line_number": 28, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}, {"api_name": "app.templates.route", "line_number": 26, "usage_type": "call"}, {"api_name": "app.templates", "line_number": 26, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 37, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 37, "usage_type": "name"}, {"api_name": "flask.request.json", "line_number": 38, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 38, "usage_type": "name"}, {"api_name": "flask.request.json.get", "line_number": 39, "usage_type": "call"}, {"api_name": "flask.request.json", "line_number": 39, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 39, "usage_type": "name"}, {"api_name": "app.util.http_get", "line_number": 43, "usage_type": "call"}, {"api_name": "app.util.http_get", "line_number": 44, "usage_type": "call"}, {"api_name": "app.templates.render", "line_number": 49, "usage_type": "call"}, {"api_name": "app.util.send_email", "line_number": 50, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 52, "usage_type": "call"}, {"api_name": "app.templates.route", "line_number": 32, "usage_type": "call"}, {"api_name": "app.templates", "line_number": 32, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path", "line_number": 56, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.abspath", "line_number": 56, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "requests.get", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.HTTPError", "line_number": 65, "usage_type": "attribute"}, {"api_name": "typing.Tuple", "line_number": 55, "usage_type": "name"}]} +{"seq_id": "583280096", "text": "# -*- coding: utf-8 -*-\nimport os\nimport hashlib\nimport xml.etree.ElementTree as ET\n\n\ndef xmlchangemd5(xmlpath):\n tree = ET.ElementTree()\n tree.parse(xmlpath)\n temp = tree.getroot()\n luamd5 = md5('lua.zip')\n resmd5 = md5('resources.zip')\n luavesion = int(temp.find('LobbyLuaVersion').text)\n resvesion = int(temp.find('LobbyResVersion').text)\n if alter(xmlpath, str(temp.find('LobbyLuaMD5').text), luamd5):\n alter(xmlpath, '' + str(luavesion),\n '' + str(luavesion + 1))\n if alter(xmlpath, str(temp.find('LobbyResMD5').text), resmd5):\n alter(xmlpath, '' + str(resvesion),\n '' + str(resvesion + 1))\n\n\ndef md5(name):\n namepath = os.path.join('./', name)\n fp = open(namepath, 'rb')\n contents = fp.read()\n fp.close()\n filemd5 = hashlib.md5(contents).hexdigest()\n return filemd5\n\n\ndef alter(file, old_str, new_str):\n if old_str != new_str:\n a = open(file, 'rb')\n str1 = a.read().decode('utf-8')\n str1 = str1.replace(old_str, new_str)\n b = open(file, 'wb')\n b.write(str1.encode('utf-8'))\n b.close()\n return 1\n else:\n return 0\n\n\nif __name__ == '__main__':\n files = os.listdir('./')\n for i in files:\n if str(i).split('.')[-1] == 'xml':\n xmlpath = os.path.join('./', i)\n xmlchangemd5(xmlpath)\n", "sub_path": "md5_tools/game_md5/game_md5.py", "file_name": "game_md5.py", "file_ext": "py", "file_size_in_byte": 1432, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "xml.etree.ElementTree.ElementTree", "line_number": 8, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 8, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 24, "usage_type": "call"}, {"api_name": "os.path", "line_number": 24, "usage_type": "attribute"}, {"api_name": "hashlib.md5", "line_number": 28, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 46, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}]} +{"seq_id": "287919132", "text": "import matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom tensorflow.keras.optimizers import Adam\nfrom tensorflow.keras.losses import mean_squared_error\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras.layers import Dense, LSTM\nfrom sklearn.model_selection import train_test_split\n\ndata = pd.read_csv(\"sample_data.csv\")\nprices = data.Price\n\nSEQ_LEN = 30\n\nmax_ = prices.max()\nprint(max_)\n\ndef n_grams(xs, n):\n for start in range(len(xs) - n):\n yield xs[start:start+n]\n\ndef make_model():\n model = Sequential((\n LSTM(1, batch_input_shape=(1, SEQ_LEN, 1)),\n ))\n\n optim = Adam()\n model.compile(optim, loss=mean_squared_error, metrics=[mean_squared_error])\n\n return model\n\nif __name__ == \"__main__\":\n obs = list(n_grams(prices, SEQ_LEN+1))\n X = np.array([ob[:-1] for ob in obs])\n y = np.array([ob.tail(1).item() for ob in obs])\n\n x_train, x_test, y_train, y_test = train_test_split(X, y, test_size=.2)\n del X; del y;\n\n m = make_model()\n m.fit(np.expand_dims(x_train, 2), y_train)\n m.evaluate(np.expand_dims(x_test, 2), y_test)\n y_hat = m.predict(np.expand_dims(x_test, 2))\n\n pd.DataFrame({\"y_hat\": y_hat.squeeze()*max_, \"y\": y_test}).plot()\n plt.show()\n", "sub_path": "Predict/lstm.py", "file_name": "lstm.py", "file_ext": "py", "file_size_in_byte": 1250, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 10, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 23, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.LSTM", "line_number": 24, "usage_type": "call"}, {"api_name": "tensorflow.keras.optimizers.Adam", "line_number": 27, "usage_type": "call"}, {"api_name": "tensorflow.keras.losses.mean_squared_error", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 34, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 35, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 43, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.show", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}]} +{"seq_id": "538968229", "text": "#from interface.services.icontainer_agent import ContainerAgentClient\n#from pyon.ion.endpoint import ProcessRPCClient\nfrom pyon.public import Container, log, IonObject\nfrom pyon.util.containers import DotDict\nfrom pyon.util.int_test import IonIntegrationTestCase\n\nfrom interface.services.coi.iresource_registry_service import ResourceRegistryServiceClient\nfrom ion.services.sa.observatory.observatory_management_service import ObservatoryManagementService\nfrom interface.services.sa.iobservatory_management_service import IObservatoryManagementService, ObservatoryManagementServiceClient\n\nfrom pyon.util.context import LocalContextMixin\nfrom pyon.core.exception import BadRequest, NotFound, Conflict, Inconsistent\nfrom pyon.public import RT, PRED\n#from mock import Mock, patch\nfrom pyon.util.unit_test import PyonTestCase\nfrom nose.plugins.attrib import attr\nimport unittest\nfrom pyon.util.log import log\n\nfrom ion.services.sa.test.helpers import any_old\n\n\n\nclass FakeProcess(LocalContextMixin):\n name = ''\n\n \n@attr('INT', group='sa')\nclass TestObservatoryManagementServiceIntegration(IonIntegrationTestCase):\n\n def setUp(self):\n # Start container\n #print 'instantiating container'\n self._start_container()\n #container = Container()\n #print 'starting container'\n #container.start()\n #print 'started container'\n\n self.container.start_rel_from_url('res/deploy/r2deploy.yml')\n self.RR = ResourceRegistryServiceClient(node=self.container.node)\n self.OMS = ObservatoryManagementServiceClient(node=self.container.node)\n #print 'TestObservatoryManagementServiceIntegration: started services'\n\n @unittest.skip('this exists only for debugging the launch process')\n def test_just_the_setup(self):\n return\n\n #@unittest.skip('targeting')\n def test_resources_associations(self):\n self._make_associations()\n\n\n #@unittest.skip('targeting') \n def test_find_related_frames_of_reference(self):\n # finding subordinates gives a dict of obj lists, convert objs to ids\n def idify(adict):\n ids = {}\n for k, v in adict.iteritems():\n ids[k] = []\n for obj in v:\n ids[k].append(obj._id)\n\n return ids\n\n # a short version of the function we're testing, with id-ify\n def short(resource_id, output_types):\n ret = self.OMS.find_related_frames_of_reference(resource_id,\n output_types)\n return idify(ret)\n \n \n #set up associations first\n stuff = self._make_associations()\n #basic traversal of tree from instrument to platform\n ids = short(stuff.instrument_site_id, [RT.PlatformSite])\n self.assertIn(RT.PlatformSite, ids)\n self.assertIn(stuff.platform_site_id, ids[RT.PlatformSite])\n self.assertIn(stuff.platform_siteb_id, ids[RT.PlatformSite])\n self.assertNotIn(stuff.platform_siteb2_id, ids[RT.PlatformSite])\n\n #since this is the first search, just make sure the input inst_id got stripped\n if RT.InstrumentSite in ids:\n self.assertNotIn(stuff.instrument_site_id, ids[RT.InstrumentSite])\n\n #basic traversal of tree from platform to instrument\n ids = short(stuff.platform_siteb_id, [RT.InstrumentSite])\n self.assertIn(RT.InstrumentSite, ids)\n self.assertIn(stuff.instrument_site_id, ids[RT.InstrumentSite])\n self.assertNotIn(stuff.instrument_site2_id, ids[RT.InstrumentSite])\n\n\n #full traversal of tree from observatory down to instrument\n ids = short(stuff.observatory_id, [RT.InstrumentSite])\n self.assertIn(RT.InstrumentSite, ids)\n self.assertIn(stuff.instrument_site_id, ids[RT.InstrumentSite])\n\n\n #full traversal of tree from instrument to observatory\n ids = short(stuff.instrument_site_id, [RT.Observatory])\n self.assertIn(RT.Observatory, ids)\n self.assertIn(stuff.observatory_id, ids[RT.Observatory])\n\n\n #partial traversal, only down to platform\n ids = short(stuff.observatory_id, [RT.Subsite, RT.PlatformSite])\n self.assertIn(RT.PlatformSite, ids)\n self.assertIn(RT.Subsite, ids)\n self.assertIn(stuff.platform_site_id, ids[RT.PlatformSite])\n self.assertIn(stuff.platform_siteb_id, ids[RT.PlatformSite])\n self.assertIn(stuff.platform_siteb2_id, ids[RT.PlatformSite])\n self.assertIn(stuff.platform_site3_id, ids[RT.PlatformSite])\n self.assertIn(stuff.subsite_id, ids[RT.Subsite])\n self.assertIn(stuff.subsite2_id, ids[RT.Subsite])\n self.assertIn(stuff.subsitez_id, ids[RT.Subsite])\n self.assertIn(stuff.subsiteb_id, ids[RT.Subsite])\n self.assertNotIn(RT.InstrumentSite, ids)\n\n\n #partial traversal, only down to platform\n ids = short(stuff.instrument_site_id, [RT.Subsite, RT.PlatformSite])\n self.assertIn(RT.PlatformSite, ids)\n self.assertIn(RT.Subsite, ids)\n self.assertIn(stuff.platform_siteb_id, ids[RT.PlatformSite])\n self.assertIn(stuff.platform_site_id, ids[RT.PlatformSite])\n self.assertIn(stuff.subsite_id, ids[RT.Subsite])\n self.assertIn(stuff.subsiteb_id, ids[RT.Subsite])\n self.assertNotIn(stuff.subsite2_id, ids[RT.Subsite])\n self.assertNotIn(stuff.subsitez_id, ids[RT.Subsite])\n self.assertNotIn(stuff.platform_siteb2_id, ids[RT.PlatformSite])\n self.assertNotIn(RT.Observatory, ids)\n \n\n def _make_associations(self):\n \"\"\"\n create one of each resource and association used by OMS\n to guard against problems in ion-definitions\n \"\"\"\n\n #raise unittest.SkipTest(\"https://jira.oceanobservatories.org/tasks/browse/CISWCORE-41\")\n \n\n \"\"\"\n the tree we're creating (observatory, sites, platforms, instruments)\n\n rows are lettered, colums numbered. \n - first row is implied a\n - first column is implied 1\n - site Z, just because \n\n O--Sz\n |\n S--S2--P3--I4\n |\n Sb-Pb2-Ib3\n |\n P--I2\n |\n Pb\n |\n I\n\n \"\"\"\n\n #stuff we control\n observatory_id, _ = self.RR.create(any_old(RT.Observatory))\n subsite_id, _ = self.RR.create(any_old(RT.Subsite))\n subsite2_id, _ = self.RR.create(any_old(RT.Subsite))\n subsiteb_id, _ = self.RR.create(any_old(RT.Subsite))\n subsitez_id, _ = self.RR.create(any_old(RT.Subsite))\n platform_site_id, _ = self.RR.create(any_old(RT.PlatformSite))\n platform_siteb_id, _ = self.RR.create(any_old(RT.PlatformSite))\n platform_siteb2_id, _ = self.RR.create(any_old(RT.PlatformSite))\n platform_site3_id, _ = self.RR.create(any_old(RT.PlatformSite))\n instrument_site_id, _ = self.RR.create(any_old(RT.InstrumentSite))\n instrument_site2_id, _ = self.RR.create(any_old(RT.InstrumentSite))\n instrument_siteb3_id, _ = self.RR.create(any_old(RT.InstrumentSite))\n instrument_site4_id, _ = self.RR.create(any_old(RT.InstrumentSite))\n\n #stuff we associate to\n instrument_model_id, _ = self.RR.create(any_old(RT.InstrumentModel))\n instrument_device_id, _ = self.RR.create(any_old(RT.InstrumentDevice))\n platform_model_id, _ = self.RR.create(any_old(RT.PlatformModel))\n platform_device_id, _ = self.RR.create(any_old(RT.PlatformDevice))\n deployment_id, _ = self.RR.create(any_old(RT.Deployment))\n\n #observatory\n self.RR.create_association(observatory_id, PRED.hasSite, subsite_id)\n self.RR.create_association(observatory_id, PRED.hasSite, subsitez_id)\n\n #site\n self.RR.create_association(subsite_id, PRED.hasSite, subsite2_id)\n self.RR.create_association(subsite_id, PRED.hasSite, subsiteb_id)\n self.RR.create_association(subsite2_id, PRED.hasSite, platform_site3_id)\n self.RR.create_association(subsiteb_id, PRED.hasSite, platform_siteb2_id)\n self.RR.create_association(subsiteb_id, PRED.hasSite, platform_site_id)\n \n #platform_site\n self.RR.create_association(platform_site3_id, PRED.hasSite, instrument_site4_id)\n self.RR.create_association(platform_siteb2_id, PRED.hasSite, instrument_siteb3_id)\n self.RR.create_association(platform_site_id, PRED.hasSite, instrument_site2_id)\n self.RR.create_association(platform_site_id, PRED.hasSite, platform_siteb_id)\n self.RR.create_association(platform_siteb_id, PRED.hasSite, instrument_site_id)\n\n self.RR.create_association(platform_site_id, PRED.hasModel, platform_model_id)\n self.RR.create_association(platform_site_id, PRED.hasDevice, platform_device_id)\n self.RR.create_association(platform_site_id, PRED.hasDeployment, deployment_id)\n\n #instrument_site\n self.RR.create_association(instrument_site_id, PRED.hasModel, instrument_model_id)\n self.RR.create_association(instrument_site_id, PRED.hasDevice, instrument_device_id)\n self.RR.create_association(instrument_site_id, PRED.hasDeployment, deployment_id)\n\n\n ret = DotDict()\n ret.observatory_id = observatory_id\n ret.subsite_id = subsite_id\n ret.subsite2_id = subsite2_id\n ret.subsiteb_id = subsiteb_id\n ret.subsitez_id = subsitez_id\n ret.platform_site_id = platform_site_id\n ret.platform_siteb_id = platform_siteb_id\n ret.platform_siteb2_id = platform_siteb2_id\n ret.platform_site3_id = platform_site3_id\n ret.instrument_site_id = instrument_site_id\n ret.instrument_site2_id = instrument_site2_id\n ret.instrument_siteb3_id = instrument_siteb3_id\n ret.instrument_site4_id = instrument_site4_id\n\n \n return ret\n\n #@unittest.skip(\"targeting\")\n def test_create_observatory(self):\n observatory_obj = IonObject(RT.Observatory,\n name='TestFacility',\n description='some new mf')\n self.OMS.create_observatory(observatory_obj)\n\n #@unittest.skip('targeting')\n def test_find_observatory_org(self):\n org_obj = IonObject(RT.Org,\n name='TestOrg',\n description='some new mf org')\n\n org_id = self.OMS.create_marine_facility(org_obj)\n\n observatory_obj = IonObject(RT.Observatory,\n name='TestObservatory',\n description='some new obs')\n observatory_id = self.OMS.create_observatory(observatory_obj)\n\n #make association\n \n self.OMS.assign_resource_to_observatory_org(observatory_id, org_id)\n\n\n #find association\n\n org_objs = self.OMS.find_org_by_observatory(observatory_id)\n self.assertEqual(1, len(org_objs))\n self.assertEqual(org_id, org_objs[0]._id)\n print(\"org_id=<\" + org_id + \">\")\n\n #create a subsite with parent Observatory\n subsite_obj = IonObject(RT.Subsite,\n name= 'TestSubsite',\n description = 'sample subsite')\n subsite_id = self.OMS.create_subsite(subsite_obj, observatory_id)\n self.assertIsNotNone(subsite_id, \"Subsite not created.\")\n\n\n # verify that Subsite is linked to Observatory\n mf_subsite_assoc = self.RR.get_association(observatory_id, PRED.hasSite, subsite_id)\n self.assertIsNotNone(mf_subsite_assoc, \"Subsite not connected to Observatory.\")\n\n\n # add the Subsite as a resource of this Observatory\n self.OMS.assign_resource_to_observatory_org(resource_id=subsite_id, org_id=org_id)\n # verify that Subsite is linked to Org\n org_subsite_assoc = self.RR.get_association(org_id, PRED.hasResource, subsite_id)\n self.assertIsNotNone(org_subsite_assoc, \"Subsite not connected as resource to Org.\")\n\n\n #create a logical platform with parent Subsite\n platform_site_obj = IonObject(RT.PlatformSite,\n name= 'TestPlatformSite',\n description = 'sample logical platform')\n platform_site_id = self.OMS.create_platform_site(platform_site_obj, subsite_id)\n self.assertIsNotNone(platform_site_id, \"PlatformSite not created.\")\n\n\n # verify that PlatformSite is linked to Site\n site_lp_assoc = self.RR.get_association(subsite_id, PRED.hasSite, platform_site_id)\n self.assertIsNotNone(site_lp_assoc, \"PlatformSite not connected to Site.\")\n\n\n # add the PlatformSite as a resource of this Observatory\n self.OMS.assign_resource_to_observatory_org(resource_id=platform_site_id, org_id=org_id)\n # verify that PlatformSite is linked to Org\n org_lp_assoc = self.RR.get_association(org_id, PRED.hasResource, platform_site_id)\n self.assertIsNotNone(org_lp_assoc, \"PlatformSite not connected as resource to Org.\")\n\n\n\n #create a logical instrument with parent logical platform\n instrument_site_obj = IonObject(RT.InstrumentSite,\n name= 'TestInstrumentSite',\n description = 'sample logical instrument')\n instrument_site_id = self.OMS.create_instrument_site(instrument_site_obj, platform_site_id)\n self.assertIsNotNone(instrument_site_id, \"InstrumentSite not created.\")\n\n\n # verify that InstrumentSite is linked to PlatformSite\n li_lp_assoc = self.RR.get_association(platform_site_id, PRED.hasSite, instrument_site_id)\n self.assertIsNotNone(li_lp_assoc, \"InstrumentSite not connected to PlatformSite.\")\n\n\n # add the InstrumentSite as a resource of this Observatory\n self.OMS.assign_resource_to_observatory_org(resource_id=instrument_site_id, org_id=org_id)\n # verify that InstrumentSite is linked to Org\n org_li_assoc = self.RR.get_association(org_id, PRED.hasResource, instrument_site_id)\n self.assertIsNotNone(org_li_assoc, \"InstrumentSite not connected as resource to Org.\")\n\n\n # remove the InstrumentSite as a resource of this Observatory\n self.OMS.unassign_resource_from_observatory_org(instrument_site_id, org_id)\n # verify that InstrumentSite is linked to Org\n assocs,_ = self.RR.find_objects(org_id, PRED.hasResource, RT.InstrumentSite, id_only=True )\n self.assertEqual(len(assocs), 0)\n\n # remove the InstrumentSite\n self.OMS.delete_instrument_site(instrument_site_id)\n assocs, _ = self.RR.find_objects(platform_site_id, PRED.hasInstrument, RT.InstrumentSite, id_only=True )\n self.assertEqual(len(assocs), 0)\n\n\n # remove the PlatformSite as a resource of this Observatory\n self.OMS.unassign_resource_from_observatory_org(platform_site_id, org_id)\n # verify that PlatformSite is linked to Org\n assocs,_ = self.RR.find_objects(org_id, PRED.hasResource, RT.PlatformSite, id_only=True )\n self.assertEqual(len(assocs), 0)\n\n \n # remove the Site as a resource of this Observatory\n self.OMS.unassign_resource_from_observatory_org(subsite_id, org_id)\n # verify that Site is linked to Org\n assocs,_ = self.RR.find_objects(org_id, PRED.hasResource, RT.Subsite, id_only=True )\n self.assertEqual(len(assocs), 0)\n\n", "sub_path": "ion/services/sa/observatory/test/test_observatory_management_service_integration.py", "file_name": "test_observatory_management_service_integration.py", "file_ext": "py", "file_size_in_byte": 15563, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pyon.util.context.LocalContextMixin", "line_number": 24, "usage_type": "name"}, {"api_name": "pyon.util.int_test.IonIntegrationTestCase", "line_number": 29, "usage_type": "name"}, {"api_name": "interface.services.coi.iresource_registry_service.ResourceRegistryServiceClient", "line_number": 41, "usage_type": "call"}, {"api_name": "interface.services.sa.iobservatory_management_service.ObservatoryManagementServiceClient", "line_number": 42, "usage_type": "call"}, {"api_name": "unittest.skip", "line_number": 45, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 76, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 77, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 78, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 78, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 79, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 80, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 80, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 83, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 83, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 84, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 87, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 87, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 88, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 89, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 89, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 90, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 94, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 94, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 95, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 95, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 96, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 96, "usage_type": "name"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 100, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 100, "usage_type": "name"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 101, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 101, "usage_type": "name"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 102, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 102, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 106, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 106, "usage_type": "attribute"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 107, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 107, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 108, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 108, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 109, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 109, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 110, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 110, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 111, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 111, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 112, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 112, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 113, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 113, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 114, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 114, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 115, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 115, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 116, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 116, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 117, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 117, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 121, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 122, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 123, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 123, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 124, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 124, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 125, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 125, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 126, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 126, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 127, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 128, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 128, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 129, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 130, "usage_type": "name"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 131, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 166, "usage_type": "call"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 166, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 166, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 167, "usage_type": "call"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 167, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 167, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 168, "usage_type": "call"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 168, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 168, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 169, "usage_type": "call"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 169, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 169, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 170, "usage_type": "call"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 170, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 170, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 171, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 171, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 171, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 172, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 172, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 172, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 173, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 173, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 174, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 174, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 174, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 175, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 175, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 175, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 176, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 176, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 176, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 177, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 177, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 177, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 178, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 178, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 178, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 181, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentModel", "line_number": 181, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 181, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 182, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentDevice", "line_number": 182, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 182, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 183, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformModel", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 183, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 184, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformDevice", "line_number": 184, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 184, "usage_type": "name"}, {"api_name": "ion.services.sa.test.helpers.any_old", "line_number": 185, "usage_type": "call"}, {"api_name": "pyon.public.RT.Deployment", "line_number": 185, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 185, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 188, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 188, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 189, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 189, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 192, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 192, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 193, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 193, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 194, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 194, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 195, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 195, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 196, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 196, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 199, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 199, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 200, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 200, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 201, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 201, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 202, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 202, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 203, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 203, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasModel", "line_number": 205, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 205, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasDevice", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 206, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasDeployment", "line_number": 207, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 207, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasModel", "line_number": 210, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 210, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasDevice", "line_number": 211, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 211, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasDeployment", "line_number": 212, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 212, "usage_type": "name"}, {"api_name": "pyon.util.containers.DotDict", "line_number": 215, "usage_type": "call"}, {"api_name": "pyon.public.IonObject", "line_number": 235, "usage_type": "call"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 235, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 235, "usage_type": "name"}, {"api_name": "pyon.public.IonObject", "line_number": 242, "usage_type": "call"}, {"api_name": "pyon.public.RT.Org", "line_number": 242, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 242, "usage_type": "name"}, {"api_name": "pyon.public.IonObject", "line_number": 248, "usage_type": "call"}, {"api_name": "pyon.public.RT.Observatory", "line_number": 248, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 248, "usage_type": "name"}, {"api_name": "pyon.public.IonObject", "line_number": 266, "usage_type": "call"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 266, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 266, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 274, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 281, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 281, "usage_type": "name"}, {"api_name": "pyon.public.IonObject", "line_number": 286, "usage_type": "call"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 286, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 286, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 294, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 294, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 301, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 301, "usage_type": "name"}, {"api_name": "pyon.public.IonObject", "line_number": 307, "usage_type": "call"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 307, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 307, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasSite", "line_number": 315, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 315, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 322, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 322, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 329, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 329, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 329, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 329, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasInstrument", "line_number": 334, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 334, "usage_type": "name"}, {"api_name": "pyon.public.RT.InstrumentSite", "line_number": 334, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 334, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 341, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 341, "usage_type": "name"}, {"api_name": "pyon.public.RT.PlatformSite", "line_number": 341, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 341, "usage_type": "name"}, {"api_name": "pyon.public.PRED.hasResource", "line_number": 348, "usage_type": "attribute"}, {"api_name": "pyon.public.PRED", "line_number": 348, "usage_type": "name"}, {"api_name": "pyon.public.RT.Subsite", "line_number": 348, "usage_type": "attribute"}, {"api_name": "pyon.public.RT", "line_number": 348, "usage_type": "name"}, {"api_name": "nose.plugins.attrib.attr", "line_number": 28, "usage_type": "call"}]} +{"seq_id": "346456645", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat May 19 23:18:03 2018\n\n@author: litian\n\"\"\"\n\nfrom jqdatasdk import *\nimport pandas as pd\nimport datetime\nimport requests\nfrom bs4 import BeautifulSoup\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport pickle\nimport random\n\n\ndef acc_hold_update(start_date, end_date, account_holding_df, Transaction_df = None, price_offline_df = None, t_days = None, bonus_dict = None):\n \"\"\"\n Task:\n Update the account holding table\n Parameters:\n start_date:\n define the start date of account info to be updated\n end_date:\n define the end date of account info to be updated\n Transaction_df:\n the dataframe including all transaction records\n columns = [\"Account_ID\", \"Date\", \"Type\", \"Stock_Code\", \"Number\", \"Price\"]\n account_holding_df:\n the origin account holding dataframe\n columns = [\"Date\", \"Account_ID\", \"Stock_Code\", \"Number\", \"Cost\", \"Value\", \"Position\", \"MaxProfit\", \"InDate\"]\n Return:\n a new dataframe of the updated account holdings info\n \"\"\"\n \n \n ## transform the startdate, endate to the datetime type\n datestart = datetime.datetime.strptime(start_date, \"%Y-%m-%d\")\n early_date = datestart - datetime.timedelta(days=20)\n early_date_str = early_date.strftime('%Y-%m-%d')\n \n\n ## create a new account holding dataframe for funtion returning\n account_holding_new = account_holding_df.copy()\n \n if t_days is None:\n ## get trade days with jqdatasdk\n trade_days = get_trade_days(early_date_str, end_date)\n else:\n start_index = np.argwhere(np.array(t_days) == datestart.date())[0][0]\n end_index = np.argwhere(np.array(t_days) == datetime.datetime.strptime(end_date, \"%Y-%m-%d\").date())[0][0]\n trade_days = t_days[start_index:(end_index+1)]\n \n ## if the trasaction dataframe is none, create an empty dataframe\n if Transaction_df is None:\n Transaction_df = pd.DataFrame(columns = [\"Account_ID\", \"Date\", \"Type\", \"Stock_Code\", \"Number\", \"Price\"])\n\n ## collect all stock_code from account_holding_df and Transaction_df\n set1 = set(account_holding_df['Stock_Code'].value_counts().index)\n set2 = set(Transaction_df['Stock_Code'].value_counts().index)\n code_set = set1 | set2\n code_l = list(code_set)\n\n\n \n if len(account_holding_new)>0:\n ## get the last date in the account holding dataframe\n acc_last_date = datetime.datetime.strptime(account_holding_new.iloc[-1, 0], \"%Y-%m-%d\").date()\n else:\n acc_last_date = datetime.datetime.strptime(\"1900-01-01\", \"%Y-%m-%d\").date()\n \n ## for loop, every trade day's holdings info will be updated\n for j,datecurrent in enumerate(trade_days):\n \n ## data type transformtion, get the rows of the day before datecurrent \n today_str = datecurrent.strftime('%Y-%m-%d')\n \n if datecurrent > acc_last_date:\n ## at datestart to update every trade day\n if datecurrent >= datestart.date():\n \n if j > 0:\n ## get the day before current day\n yesterday = trade_days[j-1]\n yesterday_str = yesterday.strftime('%Y-%m-%d')\n \n ## get the data of yesterday\n yesterday_df = account_holding_new[account_holding_new['Date'] == yesterday_str]\n else:\n yesterday_df = account_holding_new[account_holding_new['Date'] == acc_last_date.strftime('%Y-%m-%d')]\n \n ## initialize a variable 'bonus_cash'\n bonus_cash = 0 \n \n ## initialize money_record_index\n money_record_index = None \n \n \n ## select the transaction info rows of datecurrent\n t_df = Transaction_df[Transaction_df['Date'] == today_str]\n \n \n ## for each transaction on current day\n for i in t_df.index:\n \n \n ## get the information of transaction: ID, code, number, price, operation type\n ID = t_df.loc[i, 'Account_ID']\n code = t_df.loc[i, 'Stock_Code']\n num = t_df.loc[i, 'Number']\n price = t_df.loc[i, 'Price']\n op = t_df.loc[i, 'Type']\n \n ## the criterion to find the stock record in accounting holdings table\n criterion = ((account_holding_new['Date'] == today_str) &\n (account_holding_new['Account_ID'] == ID) &\n (account_holding_new['Stock_Code'] == code))\n \n ## the criterion to find the money record in accounting holdings table\n criterion_m = ((account_holding_new['Date'] == today_str) &\n (account_holding_new['Account_ID'] == ID) &\n (account_holding_new['Stock_Code'] == '000000.RMB'))\n \n ## index the sub-dataframe for the stock record\n result = account_holding_new[criterion]\n \n ## index the sub-dataframe for the money record\n result_m = account_holding_new[criterion_m]\n\n\n \n ## if the stock_code of the trasaction exist in the account holdings table \n if len(result) > 0:\n \n ## get the number, cost of the stock in the account \n old_num = account_holding_new.loc[criterion, 'Number'].iloc[0]\n old_cost = account_holding_new.loc[criterion, 'Cost'].iloc[0]\n old_value = account_holding_new.loc[criterion, 'Value'].iloc[0]\n \n ## get the number of money in the account\n old_cost_m = account_holding_new.loc[criterion_m, 'Cost'].iloc[0]\n old_value_m = account_holding_new.loc[criterion_m, 'Value'].iloc[0]\n \n ## if the operation is on money\n if code == '000000.RMB':\n \n ## the criterion to find all record in accounting holdings table\n criterion_acc = ((account_holding_new['Date'] == today_str) & \\\n (account_holding_new['Account_ID'] == ID))\n \n result_acc = account_holding_new[criterion_acc]\n \n account_cost = sum(result_acc['Number'] * result_acc['Cost'])\n account_value = sum(result_acc['Number'] * result_acc['Value'])\n \n stock_cost = account_cost - old_cost\n stock_value = account_value - old_value\n return_ratio = account_value/account_cost\n\n ## if the transaction is money in\n if op == 'Buy':\n ## increase the number of money in the account\n account_holding_new.loc[criterion, 'Cost'] = (((old_value + price)+stock_value)/return_ratio) - stock_cost\n account_holding_new.loc[criterion, 'Value'] = old_value + price\n account_holding_new.loc[criterion, 'InDate'] = today_str\n \n ## if the transacion is money out\n else:\n ## decrease the number of money in the account, can be less than zero\n account_holding_new.loc[criterion, 'Cost'] = ((old_value - price)+stock_value)/return_ratio - stock_cost\n account_holding_new.loc[criterion, 'Value'] = old_value - price\n\n \n ## if the operation is on stock\n else:\n ## if buy stock\n if op == 'Buy':\n\n \n ## increase the number of stock in the account\n account_holding_new.loc[criterion, 'Number'] = old_num + num\n ## update the cost of stock in the account\n new_cost = (old_num * old_cost + num * price) / (old_num + num)\n account_holding_new.loc[criterion, 'Cost'] = new_cost\n ## update the number of position of stock in the account\n account_holding_new.loc[criterion, 'Position'] += 1\n \n ## get maxprofit\n maxprofit = account_holding_new.loc[criterion, 'MaxProfit'].iloc[0]\n \n \n ## update maxprofit\n old_maxvalue = old_cost * (1 + maxprofit) \n new_maxprofit = (old_maxvalue - new_cost) / new_cost \n account_holding_new.loc[criterion, 'MaxProfit'] = new_maxprofit\n \n ## update InDate\n account_holding_new.loc[criterion, 'InDate'] = today_str\n \n ## update the money in the account\n account_holding_new.loc[criterion_m, 'Cost'] = old_cost_m - num * price\n account_holding_new.loc[criterion_m, 'Value'] = old_value_m - num * price\n \n ## debug\n## print(op + code)\n\n \n \n ## if sell stock\n else:\n\n ## if the number of stock sold greater than the stock in the account, sell all of the stock\n if num >= old_num:\n ## update the money in the account\n account_holding_new.loc[criterion_m, 'Cost'] = old_cost_m + old_num * old_cost\n account_holding_new.loc[criterion_m, 'Value'] = old_value_m + old_num * price\n \n ## drop the stock record in the account \n account_holding_new = account_holding_new.drop(account_holding_new.loc[criterion].index)\n account_holding_new = account_holding_new.reset_index(drop=True)\n \n ## if the number of stock sold less than the stock in the account\n else:\n ## decrease the number of stock in the account\n account_holding_new.loc[criterion, 'Number'] = old_num - num \n \n ## update the cost of stock in the account\n new_cost = (old_num * old_cost - num * price) / (old_num - num)\n account_holding_new.loc[criterion, 'Cost'] = new_cost\n ## update the number of position of stock in the account\n account_holding_new.loc[criterion, 'Position'] -= 1\n \n ## update maxprofit\n old_maxvalue = old_cost * (1 + maxprofit) \n new_maxprofit = (old_maxvalue - new_cost) / new_cost \n account_holding_new.loc[criterion, 'MaxProfit'] = new_maxprofit\n \n ## increase the money in the account\n account_holding_new.loc[criterion_m, 'Cost'] = old_cost_m + old_num * price\n account_holding_new.loc[criterion_m, 'Value'] = old_value_m + old_num * price\n\n ## debug\n## print(op + code)\n\n\n \n ## if the stock_code of the trasaction is a new code \n else:\n\n \n ## if the operation is transaction on money\n if code == '000000.RMB':\n\n\n ## if money in \n if op == 'Buy':\n ## create a new money record in the account\n new_row = pd.DataFrame([[today_str, ID, code, 1, price, price, 1, 0, today_str]], \n columns = [\"Date\", \"Account_ID\", \"Stock_Code\", \"Number\", \"Cost\", \"Value\", \"Position\", \"MaxProfit\", \"InDate\"])\n account_holding_new = account_holding_new.append(new_row, ignore_index = True)\n \n ## if money out, do nothing\n \n\n\n ## if the operation is on stock and money record exist in the account\n elif len(result_m) > 0:\n\n\n \n ## get the number of money in the account\n old_cost_m = account_holding_new.loc[criterion_m, 'Cost'].iloc[0]\n old_value_m = account_holding_new.loc[criterion_m, 'Value'].iloc[0]\n ## if buy stock\n if op == 'Buy':\n\n ## get the close price of the stock\n if price_offline_df is None:\n stock_price = get_price(code, start_date = today_str, end_date = today_str, fq = None)\n else:\n stock_price = get_price_local(code, today_str, today_str, price_offline_df)\n\n ## decrease the money in the account\n account_holding_new.loc[criterion_m, 'Cost'] = old_cost_m - num * price\n account_holding_new.loc[criterion_m, 'Value'] = old_value_m - num * price\n \n ## create a new stock record in the account\n new_row = pd.DataFrame([[today_str, ID, code, num, price, stock_price.loc[today_str, 'close'], 1, ((stock_price.loc[today_str, 'close'] - price)/price), today_str]],\n columns = [\"Date\", \"Account_ID\", \"Stock_Code\", \"Number\", \"Cost\", \"Value\", \"Position\", \"MaxProfit\", \"InDate\"])\n account_holding_new = account_holding_new.append(new_row, ignore_index = True)\n\n ## debug\n## print(op + code)\n\n \n ## make sure that Number, Cost and Value are numeric\n account_holding_new['Number'] = pd.to_numeric(account_holding_new['Number'])\n account_holding_new['Cost'] = pd.to_numeric(account_holding_new['Cost'])\n account_holding_new['Value'] = pd.to_numeric(account_holding_new['Value'])\n \n return account_holding_new\n\n\ndef net_value_cal(account_holding_df, start_date = None, account_value_df = None, t_days = None):\n \"\"\"\n Task:\n Update the account net value table\n Parameters:\n account_holding_df:\n the origin account holding dataframe\n columns = [\"Date\", \"Account_ID\", \"Stock_Code\", \"Number\", \"Cost\", \"Value\", \"Position\", \"MaxProfit\", \"InDate\"]\n start_date:\n define the start date of account info to be updated\n account_value_df:\n the dataframe including all account net value\n columns = [\"Date\", \"Account_ID\", \"Cost\", \"Value\"]\n Return:\n a new dataframe of the updated account net value table\n \"\"\" \n ## if start_date is None, set start_date to an impossible date\n if start_date is None:\n start_date = '2081-10-25'\n \n ## transform the startdate to the datetime type\n datestart = datetime.datetime.strptime(start_date, \"%Y-%m-%d\")\n \n ## get the date of the first record in the account_holding_df, if fails throw an error\n try:\n \n daterecord = datetime.datetime.strptime(account_holding_df.iloc[0, 0], \"%Y-%m-%d\")\n \n except TypeError:\n print('account_holding_df should be a dataframe with columns [\"Date\", \"Account_ID\", \"Stock_Code\", \"Number\", \"Cost\", \"Value\", \"Position\", \"MaxProfit\", \"InDate\"]')\n \n ## if no account_value_df or start_date is too early\n if (account_value_df is None) | (datestart < daterecord):\n \n ## create a new account_value_df\n account_value_updated_df = pd.DataFrame(columns = [\"Date\", \"Account_ID\", \"Cost\", \"Value\"])\n \n ## update the new df with all records in the account_holding_df\n ## from the first record\n if start_date == '2081-10-25':\n datestart = daterecord\n\n \n ## if the start_date is None\n elif start_date == '2081-10-25':\n ## copy account_value_df to a new dataframe\n account_value_updated_df = account_value_df.copy()\n \n ## get the date of the last record in the account_value_df\n try:\n daterecord = datetime.datetime.strptime(account_value_df.iloc[-1, 0], \"%Y-%m-%d\")\n except TypeError:\n print('account_value_df should be a dataframe with columns [\"Date\", \"Account_ID\", \"Cost\", \"Value\"]')\n \n ## set the datestart to the daterecord + 1\n datestart = daterecord + datetime.timedelta(days=1)\n\n \n else:\n ## copy account_value_df to a new dataframe\n account_value_updated_df = account_value_df.copy() ### need to be modified\n\n ## set the dateend to the last record in the account_holding_df\n dateend = datetime.datetime.strptime(account_holding_df.iloc[-1, 0], \"%Y-%m-%d\")\n\n if t_days is None: \n ## get trade days with jqdatasdk\n trade_days = get_trade_days(datestart, dateend)\n else:\n start_index = np.argwhere(np.array(t_days) == datestart.date())[0][0]\n end_index = np.argwhere(np.array(t_days) == dateend.date())[0][0]\n trade_days = t_days[start_index:(end_index+1)]\n \n ## for loop, every trade day's holdings info will be updated\n for j,datecurrent in enumerate(trade_days):\n \n ## data type transformtion \n today_str = datecurrent.strftime('%Y-%m-%d')\n \n ## get the current day's record from account_holding_df\n current_holding_df = account_holding_df[account_holding_df['Date'] == today_str]\n \n ## for every record\n for i in current_holding_df.index:\n \n ## get the information in the record\n account_id = current_holding_df.loc[i, 'Account_ID']\n number = current_holding_df.loc[i, 'Number']\n cost = current_holding_df.loc[i, 'Cost']\n value = current_holding_df.loc[i, 'Value']\n\n ## the criterion to find the stock record in account value table\n criterion = ((account_value_updated_df['Date'] == today_str) &\n (account_value_updated_df['Account_ID'] == account_id))\n \n ## index the sub-dataframe for the stock record\n result = account_value_updated_df[criterion]\n\n ## if find the account & currentdate in the table\n if len(result) > 0:\n \n ## update the record in the account value table\n old_cost = account_value_updated_df.loc[criterion, 'Cost'].iloc[0]\n old_value = account_value_updated_df.loc[criterion, 'Value'].iloc[0]\n account_value_updated_df.loc[criterion, 'Cost'] = old_cost + number * cost\n account_value_updated_df.loc[criterion, 'Value'] = old_value + number * value\n \n ## if not, create a new row\n else:\n \n new_cost = number * cost\n new_value = number * value\n \n ## create a new money record in the account\n new_row = pd.DataFrame([[today_str, account_id, new_cost, new_value]], \n columns = [\"Date\", \"Account_ID\", \"Cost\", \"Value\"])\n \n account_value_updated_df = account_value_updated_df.append(new_row, ignore_index = True)\n \n ## make sure that cost and value is numeric\n account_value_updated_df['Cost'] = pd.to_numeric(account_value_updated_df['Cost'] )\n account_value_updated_df['Value'] = pd.to_numeric(account_value_updated_df['Value'] )\n \n return account_value_updated_df\n\n\ndef comp_growth(account_value_df, account_id_l, start_date, end_date, stock_id_l = None):\n \"\"\"\n Task:\n compare the growth rate between account and stock\n Parameters:\n account_value_df:\n a dataframe of account value information\n columns = [\"Date\", \"Account_ID\", \"Cost\", \"Value\"]\n account_id_l:\n a list of account ids\n example: ['001.GPZH', '002.GPZH']\n start_date:\n start date\n end_date:\n end date\n stock_id_l:\n a list of stock code, according to jqdata\n example: ['600487.XSHG', '6002222.XSHG']\n Return:\n a dataframe of compared growth rate, very easy to plot\n columns' type:\n index(Date): datetime.date\n 'growth1': start from 1\n 'growth2': start from 1\n .\n .\n .\n \"\"\" \n \n ## get trade days with jqdatasdk\n trade_days = get_trade_days(start_date, end_date)\n \n ## create an empty dataframe with column 'Date'\n comp_df = pd.DataFrame(index = trade_days)\n \n ## create a timeseries dataframe from account_value_df\n ts_av_df = account_value_df.set_index(pd.to_datetime(account_value_df['Date']).dt.date)\n\n for account_id in account_id_l:\n \n ## filter the account value by account_id\n id_condition = ts_av_df['Account_ID'] == account_id\n id_df = ts_av_df.loc[id_condition, ['Cost','Value']]\n \n ## initialize yesterday's cost, value\n yesterday_cost = -1\n yesterday_value = -1\n yesterday_growth = 1\n \n for index_day in trade_days:\n \n ## if index-day has no record, growth not change\n if index_day not in id_df.index:\n today_growth = yesterday_growth\n \n ## if find a record, update the growth\n else:\n ## get today's cost, value\n today_cost = id_df.loc[index_day,'Cost']\n today_value = id_df.loc[index_day, 'Value']\n \n ## if the first record, set growth to 1\n if yesterday_cost == -1:\n today_growth = 1\n \n ## else, update the growth\n else:\n \n ## calculate today's growth\n today_increase = (today_value/today_cost)/(yesterday_value/yesterday_cost) - 1\n today_growth = yesterday_growth * (1 + today_increase)\n \n ## set yesterday's cost, value, growth\n yesterday_cost = today_cost\n yesterday_value = today_value\n yesterday_growth = today_growth \n \n comp_df.loc[index_day, account_id] = today_growth\n\n if stock_id_l is not None:\n \n for stock_id in stock_id_l:\n \n ## get the price from jqdata\n stock_df = get_price(stock_id, start_date, end_date, fq = 'post')\n \n ## get the origin value\n origin_value = stock_df.iloc[0, 1]\n \n for index_day in trade_days:\n current_value = stock_df.loc[index_day, 'close'] \n comp_df.loc[index_day, stock_id] = current_value/origin_value\n \n return comp_df\n \n\ndef get_bonus_info(stock_code, site = 'SINA'):\n \"\"\"\n Task:\n Fetch the bonus information from the specified site\n Parameters:\n stock_code:\n the stock code you want to fetch the bonus information\n example: '600487'\n site:\n now only support data from SINA, will support more sites in the future\n Return:\n a dataframe of bonus information, time series indexed,\n columns' type:\n index(Date): datetime.date\n 'B_Shares': int64, 10 shares to bonus shares\n 'I_Shares': int64, 10 shares to into shares\n 'Cash': int64, 10 shares to cash\n 'ED_Date': datetime.date, Ex-Dividend Date\n 'RR_Date': datetime.date, equity rights registration date\n \"\"\" \n \n bonus_df = pd.DataFrame(columns = [\"Date\", \"B_Shares\", \"I_Shares\", \"Cash\", \"ED_Date\", \"RR_Date\"])\n \n if site == 'SINA':\n \n ## get the url from stock_code\n url_a = 'http://money.finance.sina.com.cn/corp/go.php/vISSUE_ShareBonus/stockid/'\n url_z = '.phtml'\n url = url_a + stock_code + url_z\n \n ## get the soup\n r = requests.get(url) \n html_doc = r.text \n soup = BeautifulSoup(html_doc, \"lxml\")\n\n ## get the data tags from the soup \n bonus = soup.select(\"#sharebonus_1 > tbody > tr > td\")\n\n ## the number of record \n record_num = len(bonus)//9\n \n ## parse the data record\n for i in range(0,record_num):\n \n ## filter the valid record\n if (bonus[i*9].text != '--') & \\\n (bonus[i*9 + 1].text != '--') & \\\n (bonus[i*9 + 2].text != '--') & \\\n (bonus[i*9 + 3].text != '--') & \\\n (bonus[i*9 + 5].text != '--') & \\\n (bonus[i*9 + 6].text != '--') :\n \n ## update the columns' value\n bonus_df.loc[i, 'Date'] = bonus[i*9].text\n bonus_df.loc[i, 'B_Shares'] = bonus[i*9+1].text\n bonus_df.loc[i, 'I_Shares'] = bonus[i*9+2].text\n bonus_df.loc[i, 'Cash'] = bonus[i*9+3].text\n bonus_df.loc[i, 'ED_Date'] = bonus[i*9+5].text\n bonus_df.loc[i, 'RR_Date'] = bonus[i*9+6].text\n \n if len(bonus_df) > 0:\n \n\n ## transform the columns to desired type\n bonus_df['Date'] = pd.to_datetime(bonus_df['Date']).dt.date\n bonus_df['ED_Date'] = pd.to_datetime(bonus_df['ED_Date']).dt.date\n bonus_df['RR_Date'] = pd.to_datetime(bonus_df['RR_Date']).dt.date\n bonus_df['B_Shares'] = pd.to_numeric(bonus_df['B_Shares'])\n bonus_df['I_Shares'] = pd.to_numeric(bonus_df['I_Shares'])\n bonus_df['Cash'] = pd.to_numeric(bonus_df['Cash'])\n \n ## time series indexed\n bonus_df = bonus_df.set_index('Date')\n \n ## time ascending sort\n bonus_df = bonus_df.sort_index()\n \n return bonus_df\n \n\ndef get_weekly_price(price_df):\n \"\"\"\n Task:\n transform the daily price dataframe to a weekly price dataframe\n Parameters:\n price_df:\n daily price dataframe, from JQDATA method 'get_price()'\n Return:\n weekly_price_df: \n a dataframe of weekly price, time series indexed,\n columns' type:\n index(Date): datetime.date\n open 119 non-null float64\n close 119 non-null float64\n high 119 non-null float64\n low 119 non-null float64\n volume 119 non-null float64\n money 119 non-null float64\n \"\"\" \n ## create an empty dataframe of weekly price\n weekly_price_df = pd.DataFrame(columns = ['open', 'close', 'high', 'low', 'volume', 'money'])\n \n ## for loop to read the daily price dataframe\n for lab,row in price_df.iterrows():\n \n ## get the date of friday in the same week\n w_day = lab.weekday()\n fri_date = lab.to_pydatetime() - datetime.timedelta(days=w_day) + datetime.timedelta(days=4)\n \n ## if it is the first record in the daily price dataframe\n if len(weekly_price_df) == 0:\n \n ## set the first record\n weekly_price_df.loc[fri_date, 'open'] = row['open']\n weekly_price_df.loc[fri_date, 'close'] = row['close']\n weekly_price_df.loc[fri_date, 'high'] = row['high']\n weekly_price_df.loc[fri_date, 'low'] = row['low']\n weekly_price_df.loc[fri_date, 'volume'] = row['volume']\n weekly_price_df.loc[fri_date, 'money'] = row['money']\n \n else: \n ## if the friday record already exists\n if weekly_price_df.index[-1] != fri_date:\n weekly_price_df.loc[fri_date, 'open'] = row['open']\n weekly_price_df.loc[fri_date, 'close'] = row['close']\n weekly_price_df.loc[fri_date, 'high'] = row['high']\n weekly_price_df.loc[fri_date, 'low'] = row['low']\n weekly_price_df.loc[fri_date, 'volume'] = row['volume']\n weekly_price_df.loc[fri_date, 'money'] = row['money']\n \n else: \n weekly_price_df.loc[fri_date, 'close'] = row['close']\n \n if weekly_price_df.loc[fri_date, 'high'] < row['high']:\n weekly_price_df.loc[fri_date, 'high'] = row['high']\n \n if weekly_price_df.loc[fri_date, 'low'] > row['low']:\n weekly_price_df.loc[fri_date, 'low'] = row['low']\n \n weekly_price_df.loc[fri_date, 'volume'] += row['volume']\n \n weekly_price_df.loc[fri_date, 'money'] += row['money']\n \n return weekly_price_df\n \n\ndef get_weekly_mean(weekly_price_df):\n \"\"\"\n Task:\n add '5W' '10W' '20W' mean to the weekly price dataframe\n Parameters:\n weekly_price_df:\n weekly price dataframe\n Return:\n weekly_price_mean_df: \n a dataframe of weekly price, time series indexed, with 5W,10W,20W mean\n columns' type:\n index(Date): datetime.date\n open 119 non-null float64\n close 119 non-null float64\n high 119 non-null float64\n low 119 non-null float64\n volume 119 non-null float64\n money 119 non-null float64\n 5W 119 non-null float64\n 10W 119 non-null float64\n 20W 119 non-null float64\n UP 179 non-null integer\n DOWN 179 non-null integer\n \"\"\" \n\n ## create a copy dataframe of the weekly price dataframe \n weekly_price_mean_df = weekly_price_df.copy()\n \n ## calculate mean value\n weekly_price_mean_df['5W'] = weekly_price_mean_df['close'].rolling(window=5).mean()\n weekly_price_mean_df['10W'] = weekly_price_mean_df['close'].rolling(window=10).mean()\n weekly_price_mean_df['20W'] = weekly_price_mean_df['close'].rolling(window=20).mean()\n \n ## up trend or down trend?\n weekly_price_mean_df['UP'] = ((weekly_price_mean_df['5W'] > weekly_price_mean_df['10W']) & (weekly_price_mean_df['10W'] > weekly_price_mean_df['20W'])).apply(int)\n weekly_price_mean_df['DOWN'] = ((weekly_price_mean_df['5W'] < weekly_price_mean_df['10W']) & (weekly_price_mean_df['10W'] < weekly_price_mean_df['20W'])).apply(int)\n \n \n return weekly_price_mean_df\n \n\n \n \n\n", "sub_path": "stock.py", "file_name": "stock.py", "file_ext": "py", "file_size_in_byte": 31740, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 41, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 41, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 42, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 54, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 54, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 71, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 71, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 261, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 291, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 300, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 301, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 302, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 328, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 328, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 333, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 333, "usage_type": "attribute"}, {"api_name": "pandas.DataFrame", "line_number": 342, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 357, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 357, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 362, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 370, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 370, "usage_type": "attribute"}, {"api_name": "numpy.argwhere", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 376, "usage_type": "call"}, {"api_name": "numpy.argwhere", "line_number": 377, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 377, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 421, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 427, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 428, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 466, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 469, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 550, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 560, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 562, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 593, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 594, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 595, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 596, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 597, "usage_type": "call"}, {"api_name": "pandas.to_numeric", "line_number": 598, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 629, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 636, "usage_type": "call"}]} +{"seq_id": "76199839", "text": "import os \r\nimport pygame #pygame 모듈을 import \r\nimport time # 00초 후에 플레이 종료하는 경우 \r\n########################################################################################\r\n#필수 초기화\r\npygame.init() #pygame 라이브러리 초기화\r\n#화면 설정\r\nscreen_width = 800\r\nscreen_height = 600\r\nscreen = pygame.display.set_mode((screen_width, screen_height)) \r\npygame.display.set_caption('Popping Popping game in time!')\r\n\r\n\r\n\r\n#FPS 초당 프레임 수 \r\nclock = pygame.time.Clock() \r\n#######################################################################################\r\n#1-사용자 게임 초기화: 배경화면, 캐릭터이미지, 좌표/속도/폰트 설정 등 \r\ncurrent_path = os.path.dirname(__file__) #현재 파일 위치 반환 \r\nimage_path = os.path.join(current_path,\"images\") #image 폴더 위치 반환 \r\n#music_path = os.path.join(current_path, \"music\") #music 폴더 위치 반환 \r\n \r\n #사운드 \r\npygame.mixer.init() \r\npygame.mixer.music.load (\"C:/Users/www/OneDrive/바탕 화면/PythonWorkSpace/poppinggame/Wake Up.mp3\") #배경음악 \r\npygame.mixer.music.play(-1,0.0) #사운드 재생 계속 \r\npygame.mixer.music.set_volume(pygame.mixer.music.get_volume() + 0.2 ) #음량 설정\r\ngame_over_sound = pygame.mixer.Sound(\"C:/Users/www/OneDrive/바탕 화면/PythonWorkSpace/poppinggame/Mario.wav\") #게임 종료 사운드 \r\n#pygame.mixer.Sound.play(game_over_sound)\r\n#game_over_sound.set_volume(0.1) #음량 \r\n\r\n##########################################################################################\r\nbackground = pygame.image.load(os.path.join(image_path, \"background.png\")) #배경 삽입 \r\nstage = pygame.image.load(os.path.join(image_path, \"stage.png\")) #무대 삽입 \r\nstage_size = stage.get_rect().size \r\nstage_height = stage_size[1] #무대 위에 캐릭터 위치시키기 위해 사용 \r\n \r\n\r\ncharacter = pygame.image.load(os.path.join(image_path, \"character.png\")) #캐릭터 삽입 \r\ncharacter_size = character.get_rect().size \r\ncharacter_width = character_size[0] #캐릭터 가로 세로 \r\ncharacter_height = character_size[1] \r\ncharacter_x_pos = (screen_width/2)-(character_width/2) #캐릭터 좌표 \r\ncharacter_y_pos = screen_height - character_height - (stage_height-30)\r\n\r\ncharacter_to_x=0 #캐릭터 이동 방향 \r\ncharacter_speed=8 #캐릭터 이동 속도 \r\n\r\nweapon = pygame.image.load(os.path.join(image_path, \"weapon.png\")) #무기 삽입\r\nweapon_size = weapon.get_rect().size \r\nweapon_width = weapon_size[0] #무기 가로 길이\r\n\r\n#무기 여러 발 발사 \r\nweapons = [] \r\nweapon_speed = 10\r\n\r\n\r\n#이벤트 루프 \r\nrunning = True \r\n\r\nwhile running:\r\n dt = clock.tick_busy_loop(80) \r\n print(\"fps는 \" + str(clock.get_fps()))\r\n\r\n\r\n #2-이벤트 처리: 키보드 \r\n for event in pygame.event.get(): \r\n if event.type == pygame.QUIT: \r\n running = False \r\n\r\n if event.type == pygame.KEYDOWN: #키 누를 때 \r\n #방향키 \r\n if event.key == pygame.K_LEFT: #왼쪽 이동 \r\n character_to_x -= character_speed\r\n elif event.key == pygame.K_RIGHT: #오른쪽 이동 \r\n character = pygame.image.load(os.path.join(image_path, \"characterRightKey.png\"))\r\n character_to_x += character_speed \r\n break \r\n elif event.key == pygame.K_SPACE: #스페이스 클릭 시 무기발사 \r\n weapon_x_pos = character_x_pos + (character_width/2) - (weapon_width/2)\r\n weapon_y_pos = character_y_pos\r\n weapons.append([weapon_x_pos, weapon_y_pos]) #무기 리스트에 생성한 무기 추가 \r\n\r\n\r\n if event.type == pygame.KEYUP: #키 동작 해제\r\n if event.key == pygame.K_LEFT or pygame.K_RIGHT: \r\n character_to_x = 0 #움직이지지 않음 \r\n\r\n\r\n if not pygame.mixer.music.get_busy(): #소리가 섞여있지 않을 경우 \r\n pygame.mixer.music.fadeout(2000) #배경음악 fadeout \r\n pygame.time.delay(1000)\r\n running = False \r\n\r\n\r\n\r\n\r\n #3-캐릭터 위치 정의 \r\n character_x_pos += character_to_x\r\n\r\n #화면 경계값 처리 \r\n if character_x_pos < 0: #가로 \r\n character_x_pos = 0 \r\n elif character_x_pos > screen_width - character_width: \r\n character_x_pos = screen_width - character_width\r\n \r\n #무기 위치 설정 \r\n weapons = [ [w[0], w[1] - weapon_speed ] for w in weapons ] #위로 쏘아올리기 \r\n #천장에 닿으면 소멸 \r\n weapons = [ [w[0], w[1]] for w in weapons if w[1] > 0 ] \r\n\r\n\r\n\r\n\r\n \r\n #4-충돌 처리 \r\n \r\n #5-화면 그리기 \r\n screen.blit(background, (0, 0))\r\n\r\n for weapon_x_pos, weapon_y_pos in weapons: \r\n screen.blit(weapon, [weapon_x_pos, weapon_y_pos])\r\n\r\n screen.blit(stage, (0, screen_height - stage_height))\r\n screen.blit(character, (character_x_pos, character_y_pos))\r\n\r\n pygame.display.update()\r\n\r\n\r\npygame.quit() #pygame 종료 \r\n", "sub_path": "PythonWorkSpace/poppinggame/2_weapon_keyevent.py", "file_name": "2_weapon_keyevent.py", "file_ext": "py", "file_size_in_byte": 5004, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.init", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 10, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 10, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 11, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 11, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 16, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 19, "usage_type": "call"}, {"api_name": "os.path", "line_number": 19, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 20, "usage_type": "call"}, {"api_name": "os.path", "line_number": 20, "usage_type": "attribute"}, {"api_name": "pygame.mixer.init", "line_number": 24, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 24, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.load", "line_number": 25, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 25, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.play", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.set_volume", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.get_volume", "line_number": 27, "usage_type": "call"}, {"api_name": "pygame.mixer.Sound", "line_number": 28, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 28, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 33, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 33, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 33, "usage_type": "call"}, {"api_name": "os.path", "line_number": 33, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 34, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 34, "usage_type": "call"}, {"api_name": "os.path", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 39, "usage_type": "call"}, {"api_name": "os.path", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 49, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 49, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 49, "usage_type": "call"}, {"api_name": "os.path", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 67, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 68, "usage_type": "attribute"}, {"api_name": "pygame.KEYDOWN", "line_number": 71, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 73, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pygame.image.load", "line_number": 76, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 76, "usage_type": "call"}, {"api_name": "os.path", "line_number": 76, "usage_type": "attribute"}, {"api_name": "pygame.K_SPACE", "line_number": 79, "usage_type": "attribute"}, {"api_name": "pygame.KEYUP", "line_number": 85, "usage_type": "attribute"}, {"api_name": "pygame.K_LEFT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.K_RIGHT", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.get_busy", "line_number": 90, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygame.mixer.music.fadeout", "line_number": 91, "usage_type": "call"}, {"api_name": "pygame.mixer", "line_number": 91, "usage_type": "attribute"}, {"api_name": "pygame.time.delay", "line_number": 92, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 92, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 127, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 127, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 130, "usage_type": "call"}]} +{"seq_id": "294502569", "text": "#-------------------------------------------------------------------------\r\n# AUTHOR: Daniel Yoon\r\n# FILENAME: knn.py\r\n# SPECIFICATION: fufilled assignment requirements as specified\r\n# FOR: CS 4210- Assignment #2\r\n# TIME SPENT: 10 minutes\r\n#-----------------------------------------------------------*/\r\n\r\n#IMPORTANT NOTE: DO NOT USE ANY ADVANCED PYTHON LIBRARY TO COMPLETE THIS CODE SUCH AS numpy OR pandas. You have to work here only with standard vectors and arrays\r\n\r\n#importing some Python libraries\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\nimport csv\r\n\r\ndb = []\r\n\r\n#reading the data in a csv file\r\nwith open('binary_points.csv', 'r') as csvfile:\r\n reader = csv.reader(csvfile)\r\n for i, row in enumerate(reader):\r\n if i > 0: #skipping the header\r\n db.append (row)\r\n\r\nerror = 0 \r\n#loop your data to allow each instance to be your test set\r\nfor i, instance in enumerate(db):\r\n\r\n #add the training features to the 2D array X removing the instance that will be used for testing in this iteration. For instance, X = [[1, 3], [2, 1,], ...]]. Convert each feature value to\r\n # float to avoid warning messages\r\n #--> add your Python code here\r\n X = []\r\n Y = []\r\n symbolDict = {\r\n '+' : 1,\r\n '-' : 2\r\n }\r\n for j in range(len(db)):\r\n if j != i:\r\n appending = []\r\n for k in range(2):\r\n appending.append(float(db[j][k]))\r\n X.append(appending)\r\n Y.append(float(symbolDict[db[j][2]]))\r\n\r\n #transform the original training classes to numbers and add to the vector Y removing the instance that will be used for testing in this iteration. For instance, Y = [1, 2, ,...]. Convert each\r\n # feature value to float to avoid warning messages\r\n #--> add your Python code here\r\n\r\n #store the test sample of this iteration in the vector testSample\r\n #--> add your Python code here\r\n testSample = []\r\n for y in range(2):\r\n testSample.append(float(instance[y]))\r\n testSample.append(float(symbolDict[instance[2]]))\r\n \r\n #fitting the knn to the data\r\n clf = KNeighborsClassifier(n_neighbors=1, p=2)\r\n clf = clf.fit(X, Y)\r\n\r\n #use your test sample in this iteration to make the class prediction. For instance:\r\n #class_predicted = clf.predict([[1, 2]])[0]\r\n class_predicted = clf.predict([testSample[:2]])[0]\r\n #compare the prediction with the true label of the test instance to start calculating the error rate.\r\n #--> add your Python code here\r\n if class_predicted != testSample[2]:\r\n error += 1\r\n\r\n#print the error rate\r\n#--> add your Python code here\r\nprint(error/10)\r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "Assignment2/knn.py", "file_name": "knn.py", "file_ext": "py", "file_size_in_byte": 2658, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "csv.reader", "line_number": 19, "usage_type": "call"}, {"api_name": "sklearn.neighbors.KNeighborsClassifier", "line_number": 57, "usage_type": "call"}]} +{"seq_id": "614393777", "text": "try:\r\n import argparse\r\nexcept ImportError:\r\n print(\"Please check if module 'argparse' is installed\")\r\n quit()\r\n\r\n\r\nparser = argparse.ArgumentParser()\r\nparser.add_argument('--gene_map', type=argparse.FileType('r'), required=True,\r\n help=\"Table with 3 columns: gene_ID, transcript_ID, protein_ID\")\r\nparser.add_argument('--clust', type=argparse.FileType('r'), required=True,\r\n help=\"Clust output file with cluster IDs (first row) and genes included in these clusters\")\r\nparser.add_argument('--eggnog', type=argparse.FileType('r'), required=True,\r\n help=\"eggNOG-mapper output file\")\r\nparser.add_argument('--kegg', type=str, required=True, help=\"The ID of the interesting pathway in the KEGG. \\n\"\r\n \"For instance: ko04310 or ko04350\")\r\nparser.add_argument('--out', type=str, required=True, help=\"Prefix for output files\")\r\nargs = parser.parse_args()\r\n\r\n\r\ndef gene_map_parsing(gene_map, gene_dict):\r\n header = gene_map.readline()\r\n for line in gene_map:\r\n description = line.strip().split(\"\\t\")\r\n gene_ID, transcript_ID, protein_ID = description[0], description[1], description[2]\r\n gene_dict[gene_ID] = {\"transcript\": transcript_ID, \"protein\": protein_ID, \"cluster\": [],\r\n \"pathways\": []}\r\n\r\n\r\ndef clusters(gene_dict, table, dict):\r\n header = table.readline().strip().split(\"\\t\")\r\n\r\n for el in header:\r\n dict[el] = []\r\n\r\n for line in table:\r\n genes = line.strip().split(\"\\t\")\r\n for gene in genes:\r\n if len(gene) != 0:\r\n dict[header[genes.index(gene)]].append(gene)\r\n\r\n for gene in gene_dict.keys():\r\n for cluster, genes in dict.items():\r\n if gene in genes:\r\n gene_dict[gene][\"cluster\"].append(cluster)\r\n\r\n for gene, values in gene_dict.items():\r\n if len(values[\"cluster\"]) == 0:\r\n values[\"cluster\"].append(\"-\")\r\n\r\n\r\ndef eggNOG_mapper_parsing(gene_dict, prot_2_gene_dict, eggNOG):\r\n for line in eggNOG:\r\n if not line.startswith(\"#\"):\r\n description = line.strip().split(\"\\t\")\r\n protein, annotation = description[0], description[1:]\r\n if len(annotation) >= 9 and len(annotation[8]) != 0:\r\n if protein in prot_2_gene_dict.keys():\r\n gene_dict[prot_2_gene_dict[protein]][\"pathways\"].extend(annotation[8].split(\",\"))\r\n\r\n for gene, values in gene_dict.items():\r\n if len(values[\"pathways\"]) == 0:\r\n values[\"pathways\"].append(\"-\")\r\n\r\n\r\ndef output_writing(out, gene_dict, kegg):\r\n with open(\"{out}.{kegg}_pathways_in_clusters.tsv\".format(out=out, kegg=kegg), 'a') as output:\r\n output.write(\"Gene_ID\\tCluster\\n\")\r\n for gene, values in gene_dict.items():\r\n if kegg in values[\"pathways\"]:\r\n output.write(\"{gene}\\t{cluster}\\n\".format(gene=gene, cluster=values[\"cluster\"][0]))\r\n\r\n\r\nif __name__ == \"__main__\":\r\n gene_dict, cluster_dict = {}, {}\r\n gene_map_parsing(args.gene_map, gene_dict)\r\n prot_2_gene_dict = {gene_dict[gene][\"protein\"]: gene for gene in gene_dict.keys()}\r\n clusters(gene_dict, args.clust, cluster_dict)\r\n eggNOG_mapper_parsing(gene_dict, prot_2_gene_dict, args.eggnog)\r\n output_writing(args.out, gene_dict, args.kegg)", "sub_path": "Coexpressed_genes_clusters_and_KEGG.py", "file_name": "Coexpressed_genes_clusters_and_KEGG.py", "file_ext": "py", "file_size_in_byte": 3369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 8, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 9, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 11, "usage_type": "call"}, {"api_name": "argparse.FileType", "line_number": 13, "usage_type": "call"}]} +{"seq_id": "43268955", "text": "from django.conf.urls import url\nfrom .views import *\n\nurlpatterns = [\n\n url(r'^create_comment/', CreateComment.as_view(), name='create_comment'),\n url(r'^list_comment/', ListComment.as_view(), name='list_comment'),\n url(r'^update_comment/(?P\\d+)$', UpdateComment.as_view(), name='update_comment'),\n url(r'^delete_comment/(?P\\d+)$', DeleteComment.as_view(), name='delete_comment'),\n\n url(r'^create_publication/', CreatePublication.as_view(), name='create_publication'),\n url(r'^list_publication/', ListPublication.as_view(), name='list_publication'),\n url(r'^update_publication/(?P\\d+)$', UpdatePublication.as_view(), name='update_publication'),\n url(r'^delete_publication/(?P\\d+)$', DeletePublication.as_view(), name='delete_publication'),\n\n url(r'^create_owner/', CreateOwner.as_view(), name='create_owner'),\n url(r'^list_owner/', ListOwner.as_view(), name='list_owner'),\n url(r'^update_owner/(?P\\d+)$', UpdateOwner.as_view(), name='update_owner'),\n url(r'^delete_owner/(?P\\d+)$', DeleteOwner.as_view(), name='delete_owner'),\n\n url(r'^$', home, name=\"index\"),\n]", "sub_path": "proyecto/paginaVerano/apps/application/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 1126, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.conf.urls.url", "line_number": 6, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 7, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 8, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 9, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 11, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 12, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 13, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 14, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 17, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 18, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "django.conf.urls.url", "line_number": 21, "usage_type": "call"}]} +{"seq_id": "29674965", "text": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n#\n# Copyright 2019 The FATE Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom arch.api import federation\nfrom arch.api.proto import feature_selection_param_pb2\nfrom arch.api.utils import log_utils\nfrom federatedml.feature import feature_selection\nfrom federatedml.feature.hetero_feature_selection.base_feature_selection import BaseHeteroFeatureSelection\nfrom federatedml.util import consts\n\nLOGGER = log_utils.getLogger()\n\n\nclass HeteroFeatureSelectionHost(BaseHeteroFeatureSelection):\n def __init__(self, params):\n super(HeteroFeatureSelectionHost, self).__init__(params)\n\n self.static_obj = None\n self.iv_attrs = None\n self.fit_iv = False\n self.binning_obj = None\n self.results = []\n self.header = []\n self.flowid = ''\n self.party_name = consts.HOST\n\n def fit_transform(self, data_instances):\n self._abnormal_detection(data_instances)\n self._init_cols(data_instances)\n LOGGER.debug(\"host data count: {}, host header: {}\".format(data_instances.count(), self.header))\n for method in self.filter_method:\n self.filter_one_method(data_instances, method)\n self._renew_left_col_names()\n\n new_data = self._transfer_data(data_instances)\n self._reset_header()\n new_data.schema['header'] = self.header\n return data_instances\n\n def transform(self, data_instances):\n self._abnormal_detection(data_instances)\n\n self._init_cols(data_instances)\n LOGGER.info(\"[Result][FeatureSelection][Host]In transform, Self left cols are: {}\".format(\n self.left_cols\n ))\n new_data = self._transfer_data(data_instances)\n self._reset_header()\n new_data.schema['header'] = self.header\n\n return new_data\n\n def fit(self, data_instances):\n\n self._abnormal_detection(data_instances)\n\n self._init_cols(data_instances)\n\n for method in self.filter_method:\n self.filter_one_method(data_instances, method)\n self._renew_left_col_names()\n\n data_instances.schema['header'] = self.header\n return data_instances\n\n def filter_one_method(self, data_instances, method):\n\n if method == consts.IV_VALUE_THRES:\n self._send_select_cols(consts.IV_VALUE_THRES)\n self._received_result_cols(filter_name=consts.IV_VALUE_THRES)\n LOGGER.info(\n \"[Result][FeatureSelection][Host]Finish iv value threshold filter. Current left cols are: {}\".format(\n self.left_cols))\n\n if method == consts.IV_PERCENTILE:\n self._send_select_cols(consts.IV_PERCENTILE)\n self._received_result_cols(filter_name=consts.IV_PERCENTILE)\n LOGGER.info(\"[Result][FeatureSelection][Host]Finish iv percentile filter. Current left cols are: {}\".format(\n self.left_cols))\n\n if method == consts.COEFFICIENT_OF_VARIATION_VALUE_THRES:\n coe_param = self.params.coe_param\n coe_filter = feature_selection.CoeffOfVarValueFilter(coe_param, self.left_col_names, self.static_obj)\n self.left_cols = coe_filter.fit(data_instances)\n self.static_obj = coe_filter.statics_obj\n self.coe_meta = coe_filter.get_meta_obj()\n self.results.append(coe_filter.get_param_obj())\n self._renew_left_col_names()\n coe_filter.display_feature_result(self.party_name)\n LOGGER.info(\n \"[Result][FeatureSelection][Host]Finish coeffiecient_of_variation value threshold filter.\"\n \" Current left cols are: {}\".format(\n self.left_cols))\n\n if method == consts.UNIQUE_VALUE:\n unique_param = self.params.unique_param\n unique_filter = feature_selection.UniqueValueFilter(unique_param, self.left_col_names, self.static_obj)\n self.left_cols = unique_filter.fit(data_instances)\n self.static_obj = unique_filter.statics_obj\n self.unique_meta = unique_filter.get_meta_obj()\n self.results.append(unique_filter.get_param_obj())\n self._renew_left_col_names()\n unique_filter.display_feature_result(self.party_name)\n LOGGER.info(\"[Result][FeatureSelection][Host]Finish unique value filter. Current left cols are: {}\".format(\n self.left_cols))\n\n if method == consts.OUTLIER_COLS:\n outlier_param = self.params.outlier_param\n outlier_filter = feature_selection.OutlierFilter(outlier_param, self.left_col_names)\n self.left_cols = outlier_filter.fit(data_instances)\n self.outlier_meta = outlier_filter.get_meta_obj()\n self.results.append(outlier_filter.get_param_obj())\n self._renew_left_col_names()\n outlier_filter.display_feature_result(self.party_name)\n LOGGER.info(\"[Result][FeatureSelection][Host]Finish outlier cols filter. Current left cols are: {}\".format(\n self.left_cols))\n\n def _received_result_cols(self, filter_name):\n result_cols_id = self.transfer_variable.generate_transferid(self.transfer_variable.result_left_cols,\n filter_name)\n left_cols = federation.get(name=self.transfer_variable.result_left_cols.name,\n tag=result_cols_id,\n idx=0)\n LOGGER.info(\"Received left columns from guest\")\n self.left_cols = left_cols\n self._renew_left_col_names()\n\n host_cols = list(left_cols.keys())\n left_col_obj = feature_selection_param_pb2.LeftCols(original_cols=host_cols,\n left_cols=self.left_cols)\n\n result_obj = feature_selection_param_pb2.FeatureSelectionFilterParam(feature_values={},\n left_cols=left_col_obj,\n filter_name=filter_name)\n self.results.append(result_obj)\n LOGGER.info(\"Received Left cols are {}\".format(self.left_cols))\n\n def _send_select_cols(self, filter_name):\n host_select_cols_id = self.transfer_variable.generate_transferid(self.transfer_variable.host_select_cols,\n filter_name)\n federation.remote(self.left_col_names,\n name=self.transfer_variable.host_select_cols.name,\n tag=host_select_cols_id,\n role=consts.GUEST,\n idx=0)\n LOGGER.info(\"Sent select cols to guest\")\n", "sub_path": "federatedml/feature/hetero_feature_selection/feature_selection_host.py", "file_name": "feature_selection_host.py", "file_ext": "py", "file_size_in_byte": 7342, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "arch.api.utils.log_utils.getLogger", "line_number": 26, "usage_type": "call"}, {"api_name": "arch.api.utils.log_utils", "line_number": 26, "usage_type": "name"}, {"api_name": "federatedml.feature.hetero_feature_selection.base_feature_selection.BaseHeteroFeatureSelection", "line_number": 29, "usage_type": "name"}, {"api_name": "federatedml.util.consts.HOST", "line_number": 40, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 40, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_VALUE_THRES", "line_number": 83, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 83, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_VALUE_THRES", "line_number": 84, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 84, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_VALUE_THRES", "line_number": 85, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 85, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_PERCENTILE", "line_number": 90, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 90, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_PERCENTILE", "line_number": 91, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 91, "usage_type": "name"}, {"api_name": "federatedml.util.consts.IV_PERCENTILE", "line_number": 92, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 92, "usage_type": "name"}, {"api_name": "federatedml.util.consts.COEFFICIENT_OF_VARIATION_VALUE_THRES", "line_number": 96, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 96, "usage_type": "name"}, {"api_name": "federatedml.feature.feature_selection.CoeffOfVarValueFilter", "line_number": 98, "usage_type": "call"}, {"api_name": "federatedml.feature.feature_selection", "line_number": 98, "usage_type": "name"}, {"api_name": "federatedml.util.consts.UNIQUE_VALUE", "line_number": 110, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 110, "usage_type": "name"}, {"api_name": "federatedml.feature.feature_selection.UniqueValueFilter", "line_number": 112, "usage_type": "call"}, {"api_name": "federatedml.feature.feature_selection", "line_number": 112, "usage_type": "name"}, {"api_name": "federatedml.util.consts.OUTLIER_COLS", "line_number": 122, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 122, "usage_type": "name"}, {"api_name": "federatedml.feature.feature_selection.OutlierFilter", "line_number": 124, "usage_type": "call"}, {"api_name": "federatedml.feature.feature_selection", "line_number": 124, "usage_type": "name"}, {"api_name": "arch.api.federation.get", "line_number": 136, "usage_type": "call"}, {"api_name": "arch.api.federation", "line_number": 136, "usage_type": "name"}, {"api_name": "arch.api.proto.feature_selection_param_pb2.LeftCols", "line_number": 144, "usage_type": "call"}, {"api_name": "arch.api.proto.feature_selection_param_pb2", "line_number": 144, "usage_type": "name"}, {"api_name": "arch.api.proto.feature_selection_param_pb2.FeatureSelectionFilterParam", "line_number": 147, "usage_type": "call"}, {"api_name": "arch.api.proto.feature_selection_param_pb2", "line_number": 147, "usage_type": "name"}, {"api_name": "arch.api.federation.remote", "line_number": 156, "usage_type": "call"}, {"api_name": "arch.api.federation", "line_number": 156, "usage_type": "name"}, {"api_name": "federatedml.util.consts.GUEST", "line_number": 159, "usage_type": "attribute"}, {"api_name": "federatedml.util.consts", "line_number": 159, "usage_type": "name"}]} +{"seq_id": "102635266", "text": "'''\r\nCreated on 2017. 4. 12.\r\n\r\n@author: Byoungho Kang\r\n'''\r\n\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom Common.mnist import load_mnist\r\nfrom Propagation.mytwolayernet import MyTwoLayerNet\r\n\r\n(trainImg, trainLbl),(testImg, testLbl) = load_mnist(one_hot_label=True)\r\nnetwork = MyTwoLayerNet(784, 50, 10)\r\n\r\n# hyper parameters\r\nitersNum = 1000 # 반복횟수\r\ntrainSize = trainImg.shape[0] # 60000\r\nbatchSize = 100 # mini-bach 크기\r\nlearningRate = 0.1 # 학습률\r\n\r\n# 누적기록\r\ntrainLossList = []\r\n\r\nprint(\"-- Start Learning -- \")\r\n\r\nfor i in range(itersNum):\r\n # mini-batch 획득\r\n miniBatchMask = np.random.choice(trainSize, batchSize)\r\n trainImgBatch = trainImg[miniBatchMask]\r\n trainLblBatch = trainLbl[miniBatchMask]\r\n \r\n # Gradient 계산\r\n grad = network.gradient(trainImgBatch, trainLblBatch)\r\n \r\n # 가중치, 편향 갱신\r\n for key in ('W1', 'W2', 'b1', 'b2'):\r\n network.params[key] -= learningRate*grad[key]\r\n \r\n # 비용함수(오차)의 변화 기록\r\n loss = network.loss(trainImgBatch, trainLblBatch)\r\n trainLossList.append(loss)\r\n \r\n print(\"iteration\", i, \":\", loss)\r\n\r\nprint(\"-- End Learning -- \") \r\n \r\n# 그래프 그리기\r\nx = np.arange(len(trainLossList))\r\nplt.plot(x, trainLossList, label=\"loss\")\r\nplt.xlabel(\"iteration\")\r\nplt.ylabel(\"loss\")\r\nplt.show()\r\n", "sub_path": "MyPythonDeepLearning/Propagation/mymnistlearning.py", "file_name": "mymnistlearning.py", "file_ext": "py", "file_size_in_byte": 1369, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "Common.mnist.load_mnist", "line_number": 12, "usage_type": "call"}, {"api_name": "Propagation.mytwolayernet.MyTwoLayerNet", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 28, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 28, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 48, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 49, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 49, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 50, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 50, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 51, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 51, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 52, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 52, "usage_type": "name"}]} +{"seq_id": "409704511", "text": "import argparse\nimport os\nimport requests\nimport tempfile\nimport subprocess, sys\n\nimport pandas as pd\nimport numpy as np\nfrom glob import glob\n\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.impute import SimpleImputer\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.preprocessing import StandardScaler, OneHotEncoder\n\nimport logging\nimport logging.handlers\n\ndef _get_logger():\n '''\n 로깅을 위해 파이썬 로거를 사용\n # https://stackoverflow.com/questions/17745914/python-logging-module-is-printing-lines-multiple-times\n '''\n loglevel = logging.DEBUG\n l = logging.getLogger(__name__)\n if not l.hasHandlers():\n l.setLevel(loglevel)\n logging.getLogger().addHandler(logging.StreamHandler(sys.stdout)) \n l.handler_set = True\n return l \n\nlogger = _get_logger()\n\n\ndef split_train_test(df, test_ratio=0.1):\n '''\n 두 개의 데이터 세트로 분리\n '''\n total_rows = df.shape[0]\n train_end = int(total_rows * (1 - test_ratio))\n \n train_df = df[0:train_end]\n test_df = df[train_end:]\n \n return train_df, test_df\n\n\ndef get_dataframe(base_preproc_input_dir, file_name_prefix ): \n '''\n 파일 이름이 들어가 있는 csv 파일을 모두 저장하여 데이ㅓ 프레임을 리턴\n '''\n \n input_files = glob('{}/{}*.csv'.format(base_preproc_input_dir, file_name_prefix))\n #claim_input_files = glob('{}/dataset*.csv'.format(base_preproc_input_dir)) \n logger.info(f\"input_files: \\n {input_files}\") \n \n if len(input_files) == 0:\n raise ValueError(('There are no files in {}.\\n' +\n 'This usually indicates that the channel ({}) was incorrectly specified,\\n' +\n 'the data specification in S3 was incorrectly specified or the role specified\\n' +\n 'does not have permission to access the data.').format(base_preproc_input_dir, \"train\"))\n \n raw_data = [ pd.read_csv(file, index_col=0) for file in input_files ]\n df = pd.concat(raw_data)\n \n logger.info(f\"dataframe shape \\n {df.shape}\") \n logger.info(f\"dataset sample \\n {df.head(2)}\") \n #logger.info(f\"df columns \\n {df.columns}\") \n \n return df\n\n\ndef convert_type(raw, cols, type_target):\n '''\n 해당 데이터 타입으로 변경\n '''\n df = raw.copy()\n \n for col in cols:\n df[col] = df[col].astype(type_target)\n \n return df\n \n\nif __name__ =='__main__':\n \n ################################\n #### 커맨드 인자 파싱 \n ################################# \n \n parser = argparse.ArgumentParser()\n parser.add_argument('--base_output_dir', type=str, default=\"/opt/ml/processing/output\")\n parser.add_argument('--base_preproc_input_dir', type=str, default=\"/opt/ml/processing/input\") \n parser.add_argument('--split_rate', type=float, default=0.1) \n parser.add_argument('--label_column', type=str, default=\"fraud\") \n # parse arguments\n args = parser.parse_args() \n \n logger.info(\"######### Argument Info ####################################\")\n logger.info(f\"args.base_output_dir: {args.base_output_dir}\")\n logger.info(f\"args.base_preproc_input_dir: {args.base_preproc_input_dir}\") \n logger.info(f\"args.label_column: {args.label_column}\") \n logger.info(f\"args.split_rate: {args.split_rate}\") \n\n base_output_dir = args.base_output_dir\n base_preproc_input_dir = args.base_preproc_input_dir\n label_column = args.label_column \n split_rate = args.split_rate\n\n ################################# \n #### 두개의 파일(claim, customer) 을 로딩하여 policy_id 로 조인함 ########\n ################################# \n \n logger.info(f\"\\n### Loading Claim Dataset\")\n claim_df = get_dataframe(base_preproc_input_dir,file_name_prefix='claim' ) \n \n logger.info(f\"\\n### Loading Customer Dataset\") \n customer_df = get_dataframe(base_preproc_input_dir,file_name_prefix='customer' ) \n \n df = customer_df.join(claim_df, how='left')\n logger.info(f\"### dataframe merged with customer and claim: {df.shape}\")\n\n\n ################################# \n #### 카테고리 피쳐를 원핫인코딩 \n ################################# \n \n logger.info(f\"\\n ### Encoding: Category Features\") \n categorical_features = df.select_dtypes(include=['object']).columns.values.tolist() \n #categorical_features = ['driver_relationship'] \n logger.info(f\"categorical_features: {categorical_features}\") \n\n categorical_transformer = Pipeline(\n steps=[\n (\"imputer\", SimpleImputer(strategy=\"constant\", fill_value=\"missing\")),\n (\"onehot\", OneHotEncoder(handle_unknown=\"ignore\"))\n ]\n )\n \n preprocess = ColumnTransformer(\n transformers=[\n (\"cat\", categorical_transformer, categorical_features)\n ],\n sparse_threshold = 0, # dense format 으로 제공\n )\n\n X_pre_category = preprocess.fit_transform(df)\n \n\n # 원핫인코딩한 컬럼의 이름 로딩\n # Ref: Sklearn Pipeline: Get feature names after OneHotEncode In ColumnTransformer, https://stackoverflow.com/questions/54646709/sklearn-pipeline-get-feature-names-after-onehotencode-in-columntransformer\n \n processed_category_features = preprocess.transformers_[0][1].named_steps['onehot'].get_feature_names(categorical_features)\n #logger.info(f\"processed_category_features: {processed_category_features}\")\n# print(X_pre)\n \n ###############################\n ### 숫자형 변수 전처리 \n ###############################\n \n logger.info(f\"\\n ### Encoding: Numeric Features\") \n \n float_cols = df.select_dtypes(include=['float64']).columns.values\n int_cols = df.select_dtypes(include=['int64']).columns.values\n numeric_features = np.concatenate((float_cols, int_cols), axis=0).tolist()\n \n logger.info(f\"int_cols: \\n{int_cols}\") \n logger.info(f\"float_cols: \\n{float_cols}\") \n #logger.info(f\"numeric_features: \\n{numeric_features}\")\n\n # 따로 스케일링은 하지 않고, 미싱 값만 중간값을 취함\n numeric_transformer = Pipeline(\n steps=[\n (\"imputer\", SimpleImputer(strategy=\"median\")),\n # (\"scaler\", StandardScaler())\n ]\n )\n\n numeric_preprocessor = ColumnTransformer(\n transformers=[\n (\"cat\", numeric_transformer, numeric_features)\n ],\n sparse_threshold = 0,\n )\n\n X_pre_numeric = numeric_preprocessor.fit_transform(df) \n\n \n ###############################\n ### 전처리 결과 결합 ####\n ###############################\n \n logger.info(f\"\\n ### Handle preprocess results\") \n \n # 전처리 결과를 데이터 프레임으로 생성\n category_df = pd.DataFrame(data=X_pre_category, columns=processed_category_features)\n numeric_df = pd.DataFrame(data=X_pre_numeric, columns=numeric_features) \n\n full_df = pd.concat([numeric_df, category_df ], axis=1)\n \n # float 타입을 int 로 변경\n full_df = convert_type(full_df, cols=int_cols, type_target='int')\n full_df = convert_type(full_df, cols=processed_category_features, type_target='int') \n \n # label_column을 맨 앞으로 이동 시킴\n full_df = pd.concat([full_df[label_column], full_df.drop(columns=[label_column])], axis=1)\n \n ############################### \n # 훈련, 검증 데이터 세트로 분리 및 저장\n ###############################\n \n train_df, test_df = split_train_test(full_df, test_ratio=split_rate) \n train_df.to_csv(f\"{base_output_dir}/train/train.csv\", index=False)\n test_df.to_csv(f\"{base_output_dir}/test/test.csv\", index=False) \n\n logger.info(f\"preprocessed train shape \\n {train_df.shape}\") \n logger.info(f\"preprocessed validation shape \\n {test_df.shape}\") \n\n # logger.info(f\"preprocessed train path \\n {base_output_dir}/train/train.csv\")\n logger.info(f\"\\n ### Final result for train dataset \") \n logger.info(f\"preprocessed train sample \\n {train_df.head(2)}\")\n\n\n \n", "sub_path": "sagemaker/recommendation/Neural-Collaborative-Filtering-On-SageMaker/3_MLOps/2_sm_serving_pipeline/src/backup/preprocessing.py", "file_name": "preprocessing.py", "file_ext": "py", "file_size_in_byte": 8242, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.DEBUG", "line_number": 24, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 25, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 28, "usage_type": "call"}, {"api_name": "logging.StreamHandler", "line_number": 28, "usage_type": "call"}, {"api_name": "sys.stdout", "line_number": 28, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 53, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 63, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 64, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 91, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 133, "usage_type": "call"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 135, "usage_type": "call"}, {"api_name": "sklearn.preprocessing.OneHotEncoder", "line_number": 136, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 165, "usage_type": "call"}, {"api_name": "sklearn.pipeline.Pipeline", "line_number": 172, "usage_type": "call"}, {"api_name": "sklearn.impute.SimpleImputer", "line_number": 174, "usage_type": "call"}, {"api_name": "sklearn.compose.ColumnTransformer", "line_number": 179, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 196, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 197, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 199, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 206, "usage_type": "call"}]} +{"seq_id": "623650610", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# Licensed to the Apache Software Foundation (ASF) under one or more\n# contributor license agreements. See the NOTICE file distributed with\n# this work for additional information regarding copyright ownership.\n# The ASF licenses this file to You under the Apache License, Version 2.0\n# (the \"License\"); you may not use this file except in compliance with\n# the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Endpoint for returning emails in mbox format as a single archive\"\"\"\nimport plugins.server\nimport plugins.session\nimport plugins.messages\nimport plugins.defuzzer\nimport re\nimport typing\nimport aiohttp.web\n\n\nasync def process(\n server: plugins.server.BaseServer, session: plugins.session.SessionObject, indata: dict,\n) -> typing.Union[dict, aiohttp.web.Response]:\n\n lid = indata.get(\"list\", \"_\")\n domain = indata.get(\"domain\", \"_\")\n \n try:\n query_defuzzed = plugins.defuzzer.defuzz(indata, list_override=\"@\" in lid and lid or None)\n except AssertionError as e: # If defuzzer encounters syntax errors, it will throw an AssertionError\n return aiohttp.web.Response(headers={\"content-type\": \"text/plain\",}, status=500, text=str(e))\n results = await plugins.messages.query(session, query_defuzzed, query_limit=server.config.database.max_hits,)\n\n sources = []\n for email in results:\n source = await plugins.messages.get_source(session, permalink=email[\"mid\"])\n if source:\n stext = source[\"_source\"][\"source\"]\n # Convert to mboxrd format\n mboxrd_source = \"\"\n line_no = 0\n for line in stext.split(\"\\n\"):\n line_no += 1\n if line_no > 1 and re.match(r\"^>*From\\s+\", line):\n line = \">\" + line\n mboxrd_source += line + \"\\n\"\n sources.append(mboxrd_source)\n\n # Figure out a sane filename\n xlist = re.sub(r\"[^-_.a-z0-9]+\", \"_\", lid)\n xdomain = re.sub(r\"[^-_.a-z0-9]+\", \"_\", domain)\n dlfile = f\"{xlist}-{xdomain}.mbox\"\n\n # Return mbox archive with filename\n return aiohttp.web.Response(\n headers={\"Content-Type\": \"application/mbox\", \"Content-Disposition\": f\"attachment; filename={dlfile}\",},\n status=200,\n text=\"\\n\\n\".join(sources),\n )\n\n\ndef register(server: plugins.server.BaseServer):\n return plugins.server.Endpoint(process)\n", "sub_path": "server/endpoints/mbox.py", "file_name": "mbox.py", "file_ext": "py", "file_size_in_byte": 2771, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "plugins.server.server", "line_number": 29, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 29, "usage_type": "name"}, {"api_name": "plugins.server.session", "line_number": 29, "usage_type": "attribute"}, {"api_name": "plugins.server.defuzzer.defuzz", "line_number": 36, "usage_type": "call"}, {"api_name": "plugins.server.defuzzer", "line_number": 36, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 36, "usage_type": "name"}, {"api_name": "aiohttp.web.web.Response", "line_number": 38, "usage_type": "call"}, {"api_name": "aiohttp.web.web", "line_number": 38, "usage_type": "attribute"}, {"api_name": "aiohttp.web", "line_number": 38, "usage_type": "name"}, {"api_name": "plugins.server.messages.query", "line_number": 39, "usage_type": "call"}, {"api_name": "plugins.server.messages", "line_number": 39, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 39, "usage_type": "name"}, {"api_name": "plugins.server.messages.get_source", "line_number": 43, "usage_type": "call"}, {"api_name": "plugins.server.messages", "line_number": 43, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 43, "usage_type": "name"}, {"api_name": "re.match", "line_number": 51, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 57, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 58, "usage_type": "call"}, {"api_name": "aiohttp.web.web.Response", "line_number": 62, "usage_type": "call"}, {"api_name": "aiohttp.web.web", "line_number": 62, "usage_type": "attribute"}, {"api_name": "aiohttp.web", "line_number": 62, "usage_type": "name"}, {"api_name": "typing.Union", "line_number": 30, "usage_type": "attribute"}, {"api_name": "aiohttp.web.web", "line_number": 30, "usage_type": "attribute"}, {"api_name": "aiohttp.web", "line_number": 30, "usage_type": "name"}, {"api_name": "plugins.server.server", "line_number": 69, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 69, "usage_type": "name"}, {"api_name": "plugins.server.server.Endpoint", "line_number": 70, "usage_type": "call"}, {"api_name": "plugins.server.server", "line_number": 70, "usage_type": "attribute"}, {"api_name": "plugins.server", "line_number": 70, "usage_type": "name"}]} +{"seq_id": "549794485", "text": "import pygame\r\nimport os\r\n#local Class\r\nimport tile\r\nimport ball\r\n\r\nwindow_size = 1080, 720\r\nbg_color = 0, 0, 0\r\nobj_color = 255, 255, 255\r\nobj_size = 5\r\ncaption = \"Pong By lazy_logic\"\r\ngame_end = False\r\ntiles_speed = 2\r\n\r\ndef main():\r\n # init game\r\n pygame.init()\r\n screen = pygame.display.set_mode(window_size)\r\n pygame.display.set_caption(caption)\r\n\r\n #Tiles\r\n tile1_pos1 = ((12), (window_size[1] / 2) - 90)\r\n tile1_pos2 = ((12), (window_size[1] / 2) + 90)\r\n tile1 = tile.Tile(screen, (tile1_pos1, tile1_pos2), obj_color, tiles_speed,10)\r\n\r\n tile2_pos1 = ((window_size[0] - 12), (window_size[1] / 2) - 90)\r\n tile2_pos2 = ((window_size[0] - 12), (window_size[1] / 2) + 90)\r\n tile2 = tile.Tile(screen, (tile2_pos1, tile2_pos2), obj_color, tiles_speed,10)\r\n\r\n ball_obj = ball.Ball(screen, (int(window_size[0]/2), int(window_size[1]/2)), obj_color, 2, 2, 10)\r\n\r\n while 1:\r\n #Event handler\r\n for event in pygame.event.get():\r\n #Close game event\r\n if event.type == 12:\r\n print(\"Exit\")\r\n exit()\r\n key = pygame.key.get_pressed()\r\n if key[pygame.K_UP]:\r\n tile2.goUp()\r\n if key[pygame.K_DOWN]:\r\n tile2.goDown()\r\n if key[pygame.K_w]:\r\n tile1.goUp()\r\n if key[pygame.K_s]:\r\n tile1.goDown()\r\n #draw background\r\n screen.fill(bg_color)\r\n #draw middle line\r\n pygame.draw.line(screen, obj_color, (window_size[0] / 2, 0), (window_size[0] / 2, window_size[1]), obj_size)\r\n if ball_obj.vector[0] < tile1.from_pos[0] or ball_obj.vector[0] > tile2.from_pos[0]:\r\n ball_obj.setVector((window_size[0]/2, window_size[1]/2))\r\n ball_obj.update(tile1, tile2)\r\n tile1.update()\r\n tile2.update()\r\n pygame.display.update()\r\n #Loop end\r\n#start here\r\nif __name__ == \"__main__\":\r\n main()\r\n\r\n", "sub_path": "Pong/Pong.py", "file_name": "Pong.py", "file_ext": "py", "file_size_in_byte": 1925, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygame.init", "line_number": 17, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 18, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 18, "usage_type": "attribute"}, {"api_name": "pygame.display.set_caption", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 19, "usage_type": "attribute"}, {"api_name": "tile.Tile", "line_number": 24, "usage_type": "call"}, {"api_name": "tile.Tile", "line_number": 28, "usage_type": "call"}, {"api_name": "ball.Ball", "line_number": 30, "usage_type": "call"}, {"api_name": "pygame.event.get", "line_number": 34, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygame.key.get_pressed", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.key", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.K_UP", "line_number": 40, "usage_type": "attribute"}, {"api_name": "pygame.K_DOWN", "line_number": 42, "usage_type": "attribute"}, {"api_name": "pygame.K_w", "line_number": 44, "usage_type": "attribute"}, {"api_name": "pygame.K_s", "line_number": 46, "usage_type": "attribute"}, {"api_name": "pygame.draw.line", "line_number": 51, "usage_type": "call"}, {"api_name": "pygame.draw", "line_number": 51, "usage_type": "attribute"}, {"api_name": "pygame.display.update", "line_number": 57, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 57, "usage_type": "attribute"}]} +{"seq_id": "329741605", "text": "\n\n\ntest_list = ['Gene Austry 39Frosty The Snowman 39',\\\n\t\t\t 'Pharrell Williams - Happy (Official Music Video)',\\\n\t\t\t 'Plane',\\\n\t\t\t 'Snoop Dogg - \"Sweat\" Snoop Dog vs D']\n\n\nimport wikipedia\n\nclass WikiSearch : \n\tdef __init__(self, search_item) :\n\t\tself.result = wikipedia.search(search_item)\n\n\n\tdef __str__(self) :\n\t\treturn self.result.__str__()\n\n\n\n\n\tdef chooseLink (self) :\n\t\t\"\"\"\n\t\tWill search through the list of search results and attempt to find the song\n\t\t1) (Song) only if all words present with (song) are in title (or maybe some threshold)\n\t\t2) (Album)\n\t\t3) (Discography)\n\n\t\tWill return False if nothing found\n\t\t\"\"\"\n\t\t\n\n\t\treturn False\n\n\nfor item in test_list :\n\tresult = WikiSearch(item)\n\n\tprint(result)\n\n\t\n\n\ntest_thingy = WikiSearch(test_list)", "sub_path": "queryGoogle.py", "file_name": "queryGoogle.py", "file_ext": "py", "file_size_in_byte": 751, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "wikipedia.search", "line_number": 14, "usage_type": "call"}]} +{"seq_id": "364012046", "text": "# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Mar 12 15:11:17 2018\n\n@author: tih\n\"\"\"\n# input paramters SurfWAT\n\ninput_nc = r\"F:\\Create_Sheets\\Wainganga\\Simulations\\Simulation_1\\test1.nc\"\noutput_nc = r\"F:\\Create_Sheets\\Wainganga\\Simulations\\Simulation_1\\test1_out.nc\"\ninclude_reservoirs = 1 # 1 = on, 0 = off\n\n\nimport time\nimport sys\nimport wa.General.raster_conversions as RC\nimport wa.General.data_conversions as DC\nimport numpy as np\nimport netCDF4\n\ntime1 = time.time()\n\n###############################################################################\n############################### Run Part 1 ####################################\n###############################################################################\n \nimport wa.Models.SurfWAT.Part1_Channel_Routing as Part1_Channel_Routing\nRouted_Array, Accumulated_Pixels, Rivers = Part1_Channel_Routing.Run(input_nc)\n \n###############################################################################\n################## Create NetCDF Part 1 results ###############################\n###############################################################################\n\n################### Get Example parameters for NetCDF #########################\n \n# Create NetCDF \ngeo_out_example, epsg_example, size_X_example, size_Y_example, size_Z_example, Time_example = RC.Open_nc_info(input_nc) \ngeo_out_example = np.array(geo_out_example)\n\ntime_or = RC.Open_nc_array(input_nc, Var = 'time') \n \n# Latitude and longitude\nlon_ls = np.arange(size_X_example)*geo_out_example[1]+geo_out_example[0] + 0.5 * geo_out_example[1]\nlat_ls = np.arange(size_Y_example)*geo_out_example[5]+geo_out_example[3] - 0.5 * geo_out_example[5]\n\nlat_n = len(lat_ls)\nlon_n = len(lon_ls)\n\n################################ Save NetCDF ##################################\n\n# Create NetCDF file\nnc_file = netCDF4.Dataset(output_nc, 'w', format = 'NETCDF4')\nnc_file.set_fill_on()\n\n# Create dimensions\nlat_dim = nc_file.createDimension('latitude', lat_n)\nlon_dim = nc_file.createDimension('longitude', lon_n)\n\n# Create NetCDF variables\ncrso = nc_file.createVariable('crs', 'i4')\ncrso.long_name = 'Lon/Lat Coords in WGS84'\ncrso.standard_name = 'crs'\ncrso.grid_mapping_name = 'latitude_longitude'\ncrso.projection = epsg_example\ncrso.longitude_of_prime_meridian = 0.0\ncrso.semi_major_axis = 6378137.0\ncrso.inverse_flattening = 298.257223563\ncrso.geo_reference = geo_out_example\n\n######################### Save Rasters in NetCDF ##############################\n\nlat_var = nc_file.createVariable('latitude', 'f8', ('latitude',))\nlat_var.units = 'degrees_north'\nlat_var.standard_name = 'latitude'\n\nlon_var = nc_file.createVariable('longitude', 'f8', ('longitude',))\nlon_var.units = 'degrees_east'\nlon_var.standard_name = 'longitude'\n\nnc_file.createDimension('time', None)\ntimeo = nc_file.createVariable('time', 'f4', ('time',))\ntimeo.units = 'Monthly'\ntimeo.standard_name = 'time'\n\n# Variables\nrivers_var = nc_file.createVariable('rivers', 'i',\n ('latitude', 'longitude'),\n fill_value=-9999)\nrivers_var.long_name = 'Rivers'\nrivers_var.grid_mapping = 'crs'\n\naccpix_var = nc_file.createVariable('accpix', 'f8',\n ('latitude', 'longitude'),\n fill_value=-9999)\naccpix_var.long_name = 'Accumulated Pixels'\naccpix_var.units = 'AmountPixels'\naccpix_var.grid_mapping = 'crs'\n\ndischarge_nat_var = nc_file.createVariable('discharge_natural', 'f8',\n ('time', 'latitude', 'longitude'),\n fill_value=-9999)\ndischarge_nat_var.long_name = 'Natural Discharge'\ndischarge_nat_var.units = 'm3/month'\ndischarge_nat_var.grid_mapping = 'crs'\n\n# Load data\nlat_var[:] = lat_ls\nlon_var[:] = lon_ls\ntimeo[:] = time_or\n\n# Static variables\nrivers_var[:, :] = Rivers[:, :]\naccpix_var[:, :] = Accumulated_Pixels[:, :]\nfor i in range(len(time_or)):\n discharge_nat_var[i,:,:] = Routed_Array[i,:,:] \n\ntime.sleep(1)\nnc_file.close() \ndel Routed_Array, Accumulated_Pixels, Rivers\n\n###############################################################################\n############################### Run Part 2 ####################################\n###############################################################################\n\nimport wa.Models.SurfWAT.Part2_Create_Dictionaries as Part2_Create_Dictionaries\nDEM_dict, River_dict, Distance_dict, Discharge_dict = Part2_Create_Dictionaries.Run(input_nc, output_nc)\n\n###############################################################################\n################## Create NetCDF Part 2 results ###############################\n###############################################################################\n\n# Create NetCDF file\nnc_file = netCDF4.Dataset(output_nc, 'r+', format = 'NETCDF4')\nnc_file.set_fill_on()\n\n###################### Save Dictionaries in NetCDF ############################\n\nparmsdem = nc_file.createGroup('demdict_static')\nfor k,v in DEM_dict.items():\n setattr(parmsdem, str(k), str(v.tolist()))\n\nparmsriver = nc_file.createGroup('riverdict_static')\nfor k,v in River_dict.items():\n setattr(parmsriver, str(k), str(v.tolist()))\n\nparmsdist = nc_file.createGroup('distancedict_static')\nfor k,v in Distance_dict.items():\n setattr(parmsdist, str(k), str(v.tolist()))\n \nparmsdis = nc_file.createGroup('dischargedict_dynamic')\nfor k,v in Discharge_dict.items():\n setattr(parmsdis, str(k), str(v.tolist()))\n \n# Close file\ntime.sleep(1)\nnc_file.close()\ndel DEM_dict, River_dict, Distance_dict, Discharge_dict\n\n###############################################################################\n############################### Run Part 3 ####################################\n###############################################################################\n'''\nif include_reservoirs == 1:\n import wa.Models.SurfWAT.Part2_Create_Dictionaries as Part2_Create_Dictionaries\n Discharge_dict_reservoirs, River_dict_res, Distance_dict_res, DEM_dict_res = Part2_Create_Dictionaries.Run(input_nc, output_nc) \n\n\n\n\n'''\n###############################################################################\n############################### Run Part 4 ####################################\n###############################################################################\n\nimport wa.Models.SurfWAT.Part4_Withdrawals as Part4_Withdrawals\nDischarge_dict_end = Part4_Withdrawals.Run(input_nc, output_nc)\n\n###############################################################################\n################## Create NetCDF Part 4 results ###############################\n###############################################################################\n\n# Create NetCDF file\nnc_file = netCDF4.Dataset(output_nc, 'r+', format = 'NETCDF4')\nnc_file.set_fill_on()\n\n###################### Save Dictionaries in NetCDF ############################\n\nparmsdisend = nc_file.createGroup('dischargedictend_dynamic')\nfor k,v in Discharge_dict_end.items():\n setattr(parmsdisend, str(k), str(v.tolist()))\n\n# Close file\ntime.sleep(1)\nnc_file.close()\ndel Discharge_dict_end\n\n###############################################################################\n############### Part 5 Convert dictionaries to rasters ########################\n###############################################################################\n\nRiver_dict = RC.Open_nc_dict(output_nc, 'riverdict_static')\n\n# End discharge dictionary to raster\nDischarge_dict_end = RC.Open_nc_dict(output_nc, 'dischargedictend_dynamic')\nDataCube_Discharge_end = DC.Convert_dict_to_array(River_dict, Discharge_dict_end, input_nc)\n\n###################### Save Dictionaries in NetCDF ############################\n\n# Create NetCDF file\nnc_file = netCDF4.Dataset(output_nc, 'r+', format = 'NETCDF4')\nnc_file.set_fill_on()\n\ndischarge_end_var = nc_file.createVariable('discharge_end', 'f8',\n ('time', 'latitude', 'longitude'),\n fill_value=-9999)\ndischarge_end_var.long_name = 'End Discharge'\ndischarge_end_var.units = 'm3/month'\ndischarge_end_var.grid_mapping = 'crs'\n\nfor i in range(len(time_or)):\n discharge_end_var[i,:,:] = DataCube_Discharge_end[i,:,:] \n\n# Close file\nnc_file.close()\ndel DataCube_Discharge_end\n\n\n\n\n\n\n\n\n\n\n\n\n \n \n \n \n \n \n \n \n \n \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nimport pandas as pd\nimport numpy as np\nimport ast\nimport netCDF4\n\ninput1 = np.load(\"F:\\Create_Sheets\\Wainganga\\Simulations\\Simulation_1\\Sheet_5\\River_dict_CR1_simulation1.npy\").item() \n\ninput1 = np.load(\"F:\\Create_Sheets\\Wainganga\\Simulations\\Simulation_1\\Sheet_5\\Discharge_dict_CR1_simulation1.npy\").item() \n\nAmount_months = 46\n\ninput2 = str(input1)\n\n\n\n\n\n\n\n\n\n\n\n\n\ninput_netcdf = output_nc\ngroup_name = 'dischargedict'\ngroup_name = 'demdict_static'\n\n\n\n\n \n \n\n\n\n\n\n\n\n\n", "sub_path": "Models/SurfWAT/Run_SurfWAT.py", "file_name": "Run_SurfWAT.py", "file_ext": "py", "file_size_in_byte": 9122, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "time.time", "line_number": 21, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part1_Channel_Routing.Run", "line_number": 28, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part1_Channel_Routing", "line_number": 28, "usage_type": "name"}, {"api_name": "wa.General.raster_conversions.Open_nc_info", "line_number": 37, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions", "line_number": 37, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 38, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions.Open_nc_array", "line_number": 40, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions", "line_number": 40, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 43, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 44, "usage_type": "call"}, {"api_name": "netCDF4.Dataset", "line_number": 52, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 117, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part2_Create_Dictionaries.Run", "line_number": 126, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part2_Create_Dictionaries", "line_number": 126, "usage_type": "name"}, {"api_name": "netCDF4.Dataset", "line_number": 133, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 155, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part4_Withdrawals.Run", "line_number": 176, "usage_type": "call"}, {"api_name": "wa.Models.SurfWAT.Part4_Withdrawals", "line_number": 176, "usage_type": "name"}, {"api_name": "netCDF4.Dataset", "line_number": 183, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 193, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions.Open_nc_dict", "line_number": 201, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions", "line_number": 201, "usage_type": "name"}, {"api_name": "wa.General.raster_conversions.Open_nc_dict", "line_number": 204, "usage_type": "call"}, {"api_name": "wa.General.raster_conversions", "line_number": 204, "usage_type": "name"}, {"api_name": "wa.General.data_conversions.Convert_dict_to_array", "line_number": 205, "usage_type": "call"}, {"api_name": "wa.General.data_conversions", "line_number": 205, "usage_type": "name"}, {"api_name": "netCDF4.Dataset", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 281, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 283, "usage_type": "call"}]} +{"seq_id": "490501801", "text": "import sys\nfrom ui_QWDialogHeaders import Ui_QWDialogHeaders\nfrom PyQt5.QtWidgets import QDialog,QApplication,QAbstractItemView\nfrom PyQt5.QtCore import QStringListModel,Qt\n# from PyQt5.QtGui import\n\nclass QmyDialogHeaders(QDialog):\n def __init__(self,parent =None):\n super().__init__(parent)\n self.ui = Ui_QWDialogHeaders()\n self.ui.setupUi(self)\n\n self.__model = QStringListModel()\n self.ui.listView.setModel(self.__model)\n\n self.ui.listView.setAlternatingRowColors(True)#开启交替颜色背景\n self.ui.listView.setDragDropMode(QAbstractItemView.InternalMove)#控制视图拖放事件的处理方式(不是拷贝)\n self.ui.listView.setDefaultDropAction(Qt.MoveAction)#视图放下事件的处理方式(从拖动点移动到鼠标释放处)\n def setHeaderList(self,headerStrList):\n self.__model.setStringList(headerStrList)#设置模型数据\n def headerList(self):\n return self.__model.stringList()#返回模型数据\n\n def __del__(self):\n print(\"QmyDialogHeaders 对象被删除了\")\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n form = QmyDialogHeaders()\n form.show()\n sys.exit(app.exec_())", "sub_path": "Demo6 Forms/Demo6_2CustomDialog/myDialogHeaders.py", "file_name": "myDialogHeaders.py", "file_ext": "py", "file_size_in_byte": 1210, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PyQt5.QtWidgets.QDialog", "line_number": 7, "usage_type": "name"}, {"api_name": "ui_QWDialogHeaders.Ui_QWDialogHeaders", "line_number": 10, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QStringListModel", "line_number": 13, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView.InternalMove", "line_number": 17, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QAbstractItemView", "line_number": 17, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.Qt.MoveAction", "line_number": 18, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.Qt", "line_number": 18, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QApplication", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 30, "usage_type": "call"}]} +{"seq_id": "92312104", "text": "from __future__ import print_function\nimport sys\nimport os\nimport argparse\nimport torch\nimport torch.nn as nn\nimport torch.backends.cudnn as cudnn\nimport torchvision.transforms as transforms\nfrom torch.autograd import Variable\nfrom data import WIDERFace_ROOT , WIDERFace_CLASSES as labelmap\nfrom PIL import Image\nfrom data import WIDERFaceDetection, WIDERFaceAnnotationTransform, WIDERFace_CLASSES, WIDERFace_ROOT, BaseTransform , TestBaseTransform\nfrom data import *\nimport torch.utils.data as data\nfrom face_ssd import build_ssd\n#from resnet50_ssd import build_sfd\nimport pdb\nimport numpy as np\nimport cv2\nimport math\n#import matplotlib.pyplot as plt\nimport time\n\nif torch.cuda.is_available():\n torch.set_default_tensor_type('torch.cuda.FloatTensor')\nelse:\n torch.set_default_tensor_type('torch.FloatTensor')\n\ndef bbox_vote(det):\n order = det[:, 4].ravel().argsort()[::-1]\n det = det[order, :]\n while det.shape[0] > 0:\n # IOU\n area = (det[:, 2] - det[:, 0] + 1) * (det[:, 3] - det[:, 1] + 1)\n xx1 = np.maximum(det[0, 0], det[:, 0])\n yy1 = np.maximum(det[0, 1], det[:, 1])\n xx2 = np.minimum(det[0, 2], det[:, 2])\n yy2 = np.minimum(det[0, 3], det[:, 3])\n w = np.maximum(0.0, xx2 - xx1 + 1)\n h = np.maximum(0.0, yy2 - yy1 + 1)\n inter = w * h\n o = inter / (area[0] + area[:] - inter)\n # get needed merge det and delete these det\n merge_index = np.where(o >= 0.3)[0]\n det_accu = det[merge_index, :]\n det = np.delete(det, merge_index, 0)\n if merge_index.shape[0] <= 1:\n continue\n det_accu[:, 0:4] = det_accu[:, 0:4] * np.tile(det_accu[:, -1:], (1, 4))\n max_score = np.max(det_accu[:, 4])\n det_accu_sum = np.zeros((1, 5))\n det_accu_sum[:, 0:4] = np.sum(det_accu[:, 0:4], axis=0) / np.sum(det_accu[:, -1:])\n det_accu_sum[:, 4] = max_score\n try:\n dets = np.row_stack((dets, det_accu_sum))\n except:\n dets = det_accu_sum\n dets = dets[0:750, :]\n return dets\n\ndef write_to_txt(f, det , event , im_name):\n f.write('{:s}\\n'.format(event + '/' + im_name))\n f.write('{:d}\\n'.format(det.shape[0]))\n for i in range(det.shape[0]):\n xmin = det[i][0]\n ymin = det[i][1]\n xmax = det[i][2]\n ymax = det[i][3]\n score = det[i][4]\n f.write('{:.1f} {:.1f} {:.1f} {:.1f} {:.3f}\\n'.\n format(xmin, ymin, (xmax - xmin + 1), (ymax - ymin + 1), score))\n\ndef infer(net , img , transform , thresh , cuda , shrink):\n if shrink != 1:\n img = cv2.resize(img, None, None, fx=shrink, fy=shrink, interpolation=cv2.INTER_LINEAR)\n x = torch.from_numpy(transform(img)[0]).permute(2, 0, 1)\n\n with torch.no_grad():\n x = Variable(x.unsqueeze(0))\n if cuda:\n x = x.cuda()\n #print (shrink , x.shape)\n y = net(x) # forward pass\n detections = y.data\n # scale each detection back up to the image\n scale = torch.Tensor([ img.shape[1]/shrink, img.shape[0]/shrink,\n img.shape[1]/shrink, img.shape[0]/shrink] )\n det = []\n for i in range(detections.size(1)):\n j = 0\n while detections[0, i, j, 0] >= thresh:\n score = detections[0, i, j, 0]\n #label_name = labelmap[i-1]\n pt = (detections[0, i, j, 1:]*scale).cpu().numpy()\n coords = (pt[0], pt[1], pt[2], pt[3])\n det.append([pt[0], pt[1], pt[2], pt[3], score])\n j += 1\n if (len(det)) == 0:\n det = [ [0.1,0.1,0.2,0.2,0.01] ]\n det = np.array(det)\n\n keep_index = np.where(det[:, 4] >= 0)[0]\n det = det[keep_index, :]\n return det\n\ndef infer_flip(net , img , transform , thresh , cuda , shrink):\n img = cv2.flip(img, 1)\n det = infer(net , img , transform , thresh , cuda , shrink)\n det_t = np.zeros(det.shape)\n det_t[:, 0] = img.shape[1] - det[:, 2]\n det_t[:, 1] = det[:, 1]\n det_t[:, 2] = img.shape[1] - det[:, 0]\n det_t[:, 3] = det[:, 3]\n det_t[:, 4] = det[:, 4]\n return det_t\n\n\ndef infer_multi_scale_sfd(net , img , transform , thresh , cuda , max_im_shrink):\n # shrink detecting and shrink only detect big face\n st = 0.5 if max_im_shrink >= 0.75 else 0.5 * max_im_shrink\n det_s = infer(net , img , transform , thresh , cuda , st)\n index = np.where(np.maximum(det_s[:, 2] - det_s[:, 0] + 1, det_s[:, 3] - det_s[:, 1] + 1) > 30)[0]\n det_s = det_s[index, :]\n # enlarge one times\n bt = min(2, max_im_shrink) if max_im_shrink > 1 else (st + max_im_shrink) / 2\n det_b = infer(net , img , transform , thresh , cuda , bt)\n # enlarge small iamge x times for small face\n if max_im_shrink > 2:\n bt *= 2\n while bt < max_im_shrink:\n det_b = np.row_stack((det_b, infer(net , img , transform , thresh , cuda , bt)))\n bt *= 2\n det_b = np.row_stack((det_b, infer(net , img , transform , thresh , cuda , max_im_shrink) ))\n # enlarge only detect small face\n if bt > 1:\n index = np.where(np.minimum(det_b[:, 2] - det_b[:, 0] + 1, det_b[:, 3] - det_b[:, 1] + 1) < 100)[0]\n det_b = det_b[index, :]\n else:\n index = np.where(np.maximum(det_b[:, 2] - det_b[:, 0] + 1, det_b[:, 3] - det_b[:, 1] + 1) > 30)[0]\n det_b = det_b[index, :]\n return det_s, det_b\n\ndef test_oneimage(net, img, model_path):\n cfg = widerface_640\n\n # evaluation\n cuda = torch.cuda.is_available()\n transform = TestBaseTransform((104, 117, 123))\n thresh=cfg['conf_thresh']\n\n max_im_shrink = ( (2000.0*2000.0) / (img.shape[0] * img.shape[1])) ** 0.5\n shrink = max_im_shrink if max_im_shrink < 1 else 1\n\n det0 = infer(net , img , transform , thresh , cuda , shrink)\n det1 = infer_flip(net , img , transform , thresh , cuda , shrink)\n\n # shrink detecting and shrink only detect big face\n st = 0.5 if max_im_shrink >= 0.75 else 0.5 * max_im_shrink\n det_s = infer(net , img , transform , thresh , cuda , st)\n index = np.where(np.maximum(det_s[:, 2] - det_s[:, 0] + 1, det_s[:, 3] - det_s[:, 1] + 1) > 30)[0]\n det_s = det_s[index, :]\n\n # enlarge one times\n factor = 2\n bt = min(factor, max_im_shrink) if max_im_shrink > 1 else (st + max_im_shrink) / 2\n det_b = infer(net , img , transform , thresh , cuda , bt)\n # enlarge small iamge x times for small face\n if max_im_shrink > factor:\n bt *= factor\n while bt < max_im_shrink:\n det_b = np.row_stack((det_b, infer(net , img , transform , thresh , cuda , bt)))\n bt *= factor\n det_b = np.row_stack((det_b, infer(net , img , transform , thresh , cuda , max_im_shrink) ))\n\n # enlarge only detect small face\n if bt > 1:\n index = np.where(np.minimum(det_b[:, 2] - det_b[:, 0] + 1, det_b[:, 3] - det_b[:, 1] + 1) < 100)[0]\n det_b = det_b[index, :]\n else:\n index = np.where(np.maximum(det_b[:, 2] - det_b[:, 0] + 1, det_b[:, 3] - det_b[:, 1] + 1) > 30)[0]\n det_b = det_b[index, :]\n det = np.row_stack((det0, det1, det_s, det_b))\n det = bbox_vote(det)\n\n \"\"\"\n for aDet in det:\n print(aDet)\n cv2.rectangle(img,(int(aDet[0]),int(aDet[1])),(int(aDet[2]),int(aDet[3])),(0,255,0),3)\n cv2.imshow('image',img)\n cv2.waitKey(0)\n cv2.destroyAllWindows()\n \"\"\"\n return det\n\nif __name__ == '__main__':\n test_oneimage()\n", "sub_path": "detection.py", "file_name": "detection.py", "file_ext": "py", "file_size_in_byte": 7367, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.cuda.is_available", "line_number": 24, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 24, "usage_type": "attribute"}, {"api_name": "torch.set_default_tensor_type", "line_number": 25, "usage_type": "call"}, {"api_name": "torch.set_default_tensor_type", "line_number": 27, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 35, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 36, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 37, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 38, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 39, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 40, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 50, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 51, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 52, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 55, "usage_type": "call"}, {"api_name": "cv2.resize", "line_number": 75, "usage_type": "call"}, {"api_name": "cv2.INTER_LINEAR", "line_number": 75, "usage_type": "attribute"}, {"api_name": "torch.from_numpy", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.autograd.Variable", "line_number": 79, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 86, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 100, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 102, "usage_type": "call"}, {"api_name": "cv2.flip", "line_number": 107, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 109, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 131, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 133, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 139, "usage_type": "call"}, {"api_name": "torch.cuda.is_available", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.cuda", "line_number": 147, "usage_type": "attribute"}, {"api_name": "data.TestBaseTransform", "line_number": 148, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 160, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 171, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 173, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 177, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 180, "usage_type": "call"}, {"api_name": "numpy.row_stack", "line_number": 182, "usage_type": "call"}]} +{"seq_id": "311447493", "text": "import torch\nimport torch.nn as nn\nimport torch.nn.init as init\nimport torch.nn.functional as F\nfrom torch.nn.parameter import Parameter\n\nimport numpy as np\nimport math\n\nfrom . import object_detector_network\nfrom . import scene_detector_network\nfrom . import action_detector_network\n\n\nclass Indomain_Dynamic_Attention(nn.Module):\n\n def __init__(self, classes, branches, T):\n super(Indomain_Dynamic_Attention, self).__init__()\n #### parameters\n self.reduce = 16\n self.len = 32\n self.branches = branches\n self.T = T\n self.classes = classes\n\n self.adaptiveavgpool = nn.AdaptiveAvgPool1d(1)\n\n self.atten_feat_reduce_bn = nn.BatchNorm1d(self.classes*self.branches)\n self.atten_feat_reduce = nn.Conv1d(in_channels=self.classes*self.branches, out_channels=self.classes, kernel_size=1, stride=1, bias=False)\n\n len = self.T // 2\n \n self.w_reduce = Parameter(torch.Tensor(len, self.T))\n stv = 1. / math.sqrt(self.w_reduce.size(1))\n self.w_reduce.data.uniform_(-stv, stv)\n\n self.w_atten = Parameter(torch.Tensor(self.branches, len))\n stv = 1. / math.sqrt(self.w_atten.size(1))\n self.w_atten.data.uniform_(-stv, stv)\n\n ## time attention\n len = max(self.classes // self.reduce, self.len)\n \n self.time_feature = Parameter(torch.Tensor(len, self.classes))\n stv = 1. / math.sqrt(self.time_feature.size(1))\n self.time_feature.data.uniform_(-stv, stv)\n \n self.time_atten = Parameter(torch.Tensor(1, len))\n stv = 1. / math.sqrt(self.time_atten.size(1))\n self.time_atten.data.uniform_(-stv, stv)\n\n for l in self.children():\n if isinstance(l, nn.Conv1d):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n elif isinstance(l, nn.BatchNorm1d):\n l.weight.data.fill_(1)\n l.bias.data.zero_()\n elif isinstance(l, nn.Linear):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n nn.init.constant_(l.bias, 0)\n\n def forward(self, feature):\n \n ## convs attention\n feature = torch.cat(feature, dim=1)\n feature = F.relu(self.atten_feat_reduce_bn(feature))\n feature = self.atten_feat_reduce(feature)\n\n batch_size = feature.size(0)\n atten_feat = feature.view(-1, self.T)\n atten_feat = torch.tanh(F.linear(atten_feat, self.w_reduce, None))\n atten_feat = F.linear(atten_feat, self.w_atten, None).view(batch_size, -1, self.branches).transpose(1,2).contiguous()\n atten_feat = F.softmax(atten_feat, dim=1).unsqueeze(3)\n\n ## time attention\n time_atten_feat = torch.tanh(F.linear(feature.permute(0,2,1).contiguous().view(-1, self.classes), self.time_feature, None))\n time_atten_feat = F.linear(time_atten_feat, self.time_atten, None).view(batch_size, -1, 1)\n time_atten_feat = F.softmax(time_atten_feat, dim=1)\n \n return [atten_feat, time_atten_feat]\n\n\nclass Crossdomain_Dynamic_Attention(nn.Module):\n\n def __init__(self, classes, branches, domains, T):\n super(Crossdomain_Dynamic_Attention, self).__init__()\n #### parameters\n self.classes = classes\n self.reduce = 16\n self.len = 32\n self.branches = branches\n self.T = T\n\n self.convs = nn.ModuleList([])\n params = [[1, 0, 1], [3, 1, 1], [3, 2, 2]] \n for i in range(branches):\n self.convs.append(nn.Conv1d(in_channels=self.classes, out_channels=self.classes, kernel_size=params[i][0], stride=1, padding=params[i][1], dilation=params[i][2]))\n \n self.adaptiveavgpool = nn.AdaptiveAvgPool1d(1)\n\n self.atten_feat_reduce_bn = nn.BatchNorm1d(self.classes*self.branches)\n self.atten_feat_reduce = nn.Conv1d(in_channels=self.classes*self.branches, out_channels=self.classes, kernel_size=1, stride=1, bias=False)\n\n len = self.T // 2\n self.w_reduce = Parameter(torch.Tensor(len, self.T))\n self.w_atten = Parameter(torch.Tensor(self.branches, len))\n\n stv = 1. / math.sqrt(self.w_reduce.size(1))\n self.w_reduce.data.uniform_(-stv, stv)\n\n stv = 1. / math.sqrt(self.w_atten.size(1))\n self.w_atten.data.uniform_(-stv, stv)\n\n ## time attention\n len = max(self.classes // self.reduce, self.len)\n \n self.time_feature = Parameter(torch.Tensor(len, self.classes))\n stv = 1. / math.sqrt(self.time_feature.size(1))\n self.time_feature.data.uniform_(-stv, stv)\n \n self.time_atten = Parameter(torch.Tensor(1, len))\n stv = 1. / math.sqrt(self.time_atten.size(1))\n self.time_atten.data.uniform_(-stv, stv)\n\n for l in self.children():\n if isinstance(l, nn.Conv1d):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n elif isinstance(l, nn.BatchNorm1d):\n l.weight.data.fill_(1)\n l.bias.data.zero_()\n elif isinstance(l, nn.Linear):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n nn.init.constant_(l.bias, 0)\n\n def forward(self, feature):\n \n ### multiple branch\n features = [conv(feature) for conv in self.convs]\n\n atten_feat = torch.cat(features, dim=1)\n atten_feat = F.relu(self.atten_feat_reduce_bn(atten_feat))\n atten_feat = self.atten_feat_reduce(atten_feat)\n\n features = torch.stack(features, dim=1)\n\n ### dynamic convolution\n atten_feat_b = torch.sum(features, dim=1)\n batch_size = atten_feat_b.size(0)\n\n atten_feat = atten_feat_b.view(-1, self.T)\n atten_feat = torch.tanh(F.linear(atten_feat, self.w_reduce, None))\n atten_feat = F.linear(atten_feat, self.w_atten, None).view(batch_size, -1, self.branches).transpose(1,2).contiguous()\n atten_feat = F.softmax(atten_feat, dim=1).unsqueeze(3)\n \n features = torch.sum(features*atten_feat, dim=1)\n\n ## time attention\n time_atten_feat = torch.tanh(F.linear(feature.permute(0,2,1).contiguous().view(-1, self.classes), self.time_feature, None))\n time_atten_feat = F.linear(time_atten_feat, self.time_atten, None).view(batch_size, -1, 1)\n time_atten_feat = F.softmax(time_atten_feat, dim=1)\n \n features = torch.matmul(features, time_atten_feat).squeeze(2)\n \n return features\n\n\nclass Event_Model(nn.Module):\n\n def __init__(self, opt):\n super(Event_Model, self).__init__()\n\n #### parameters\n self.opt = opt\n self.concept_number = opt.scene_classes + opt.object_classes + opt.action_classes\n self.scene_classes = opt.scene_classes\n self.object_classes = opt.object_classes\n self.action_classes = opt.action_classes\n self.T = opt.segment_number\n self.branches = 3\n self.domains = 3\n \n #### concept detectors\n self.scene_detector = scene_detector_network.Scene_Detector(opt)\n self.object_detector = object_detector_network.Object_Detector(opt)\n self.action_detector = action_detector_network.Action_Detector(opt)\n \n #### dynamic convolution\n ### indomain\n ## scene\n self.scene_indomain_bn1 = nn.BatchNorm1d(self.T)#scene_classes)\n self.scene_indomain_convs = nn.ModuleList([])\n params = [[1, 0, 1], [3, 1, 1], [3, 2, 2]] \n for i in range(self.branches):\n self.scene_indomain_convs.append(nn.Conv1d(in_channels=self.scene_classes, out_channels=self.scene_classes, kernel_size=params[i][0], stride=1, padding=params[i][1], dilation=params[i][2]))\n\n ## object\n self.object_indomain_bn1 = nn.BatchNorm1d(self.T)#object_classes)\n self.object_indomain_convs = nn.ModuleList([])\n params = [[1, 0, 1], [3, 1, 1], [3, 2, 2]] \n for i in range(self.branches):\n self.object_indomain_convs.append(nn.Conv1d(in_channels=self.object_classes, out_channels=self.object_classes, kernel_size=params[i][0], stride=1, padding=params[i][1], dilation=params[i][2]))\n \n ## action\n self.action_domain_bn1 = nn.BatchNorm1d(self.T)#object_classes)\n self.action_domain_convs = nn.ModuleList([])\n params = [[1, 0, 1], [3, 1, 1], [3, 2, 2]] \n for i in range(self.branches):\n self.action_domain_convs.append(nn.Conv1d(in_channels=self.action_classes, out_channels=self.action_classes, kernel_size=params[i][0], stride=1, padding=params[i][1], dilation=params[i][2]))\n \n self.crossdomain_bn1 = nn.BatchNorm1d(self.T)\n \n ## dynamic attentions\n # indomain\n self.scene_indomain_dynamic_attention = Indomain_Dynamic_Attention(classes=self.scene_classes, branches=self.branches, T=self.T)\n self.object_indomain_dynamic_attention = Indomain_Dynamic_Attention(classes=self.object_classes, branches=self.branches, T=self.T)\n self.action_indomain_dynamic_attention = Indomain_Dynamic_Attention(classes=self.action_classes, branches=self.branches, T=self.T)\n \n # cross domain\n self.crossdomain_dynamic_attention = Crossdomain_Dynamic_Attention(classes=self.concept_number, branches=self.branches, domains = self.domains, T=self.T)\n\n ## other\n self.adaptiveavgpool = nn.AdaptiveAvgPool1d(1)\n\n #### feature reduce dimentation & fusion\n self.reduce_dim = 1024\n self.concat_bn1 = nn.BatchNorm1d(self.concept_number*2)\n self.concat_reduce_dim = nn.Linear(self.concept_number*2, self.reduce_dim)\n \n #### classification\n self.final_bn1 = nn.BatchNorm1d(self.reduce_dim)\n self.final_classifier = nn.Linear(self.reduce_dim, opt.event_classes)\n self.dropout = nn.Dropout(0.5)\n self.relu = nn.ReLU(inplace=True)\n\n for l in self.children():\n if isinstance(l, nn.Conv1d):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n elif isinstance(l, nn.BatchNorm1d):\n l.weight.data.fill_(1)\n l.bias.data.zero_()\n elif isinstance(l, nn.Linear):\n nn.init.kaiming_normal_(l.weight, mode='fan_out', nonlinearity='relu')\n nn.init.constant_(l.bias, 0)\n\n def forward(self, sceobj_frame):\n \n #### extract initial concept representation\n action_frame = sceobj_frame.permute(0, 1, 3, 2, 4, 5)\n \n ## scene and object frame input size N T D C H W\n N, T, D, C, H, W = sceobj_frame.size()\n _, _, _, _, aH, aW = action_frame.size()\n \n # N T D C H W -> NTD C H W\n sceobj_frame = sceobj_frame.view(-1, C, H, W)\n # NTD C H W -> NTD F\n scene_feature = self.scene_detector(sceobj_frame)\n object_feature = self.object_detector(sceobj_frame)\n \n # NTD F -> N T D F\n scene_feature = scene_feature.view(N, T, D, -1)\n object_feature = object_feature.view(N, T, D, -1)\n # N T D F -> N T F\n scene_feature, _ = torch.max(scene_feature, dim=2)\n object_feature, _ = torch.max(object_feature, dim=2)\n \n ## action frame inpupt size N T C D aH aW\n # N T C D aH aW -> NT C D aH aW\n action_frame = action_frame.view(-1, C, D, aH, aW)\n # NT C D H W -> NT F\n action_feature = self.action_detector(action_frame)\n del action_frame\n # NT F -> N T F\n action_feature = action_feature.view(N, T, -1)\n \n scene_feature_s = scene_feature\n object_feature_s = object_feature\n action_feature_s = action_feature\n \n #### in-domain dynamic convolution\n ### multiple branches convolution\n ## bn & relu\n scene_feature = F.relu(self.scene_indomain_bn1(scene_feature_s))\n object_feature = F.relu(self.object_indomain_bn1(object_feature_s))\n action_feature = F.relu(self.action_domain_bn1(action_feature_s))\n \n ## permute\n # N T F -> N F T\n scene_feature = scene_feature.permute(0, 2, 1)\n object_feature = object_feature.permute(0, 2, 1)\n action_feature = action_feature.permute(0, 2, 1)\n \n ## scene\n scene_feature = [conv(scene_feature) for conv in self.scene_indomain_convs]\n scene_feature_b = torch.stack(scene_feature, dim=1)\n\n ## object\n object_feature = [conv(object_feature) for conv in self.object_indomain_convs]\n object_feature_b = torch.stack(object_feature, dim=1)\n \n ## action\n action_feature = [conv(action_feature) for conv in self.action_domain_convs]\n action_feature_b = torch.stack(action_feature, dim=1)\n\n ### dynamic attentions\n ## scene dynamic feature \n attentions = self.scene_indomain_dynamic_attention(scene_feature)\n indomain_scene_feature = torch.sum(scene_feature_b*attentions[0], dim=1)\n indomain_scene_feature = torch.matmul(indomain_scene_feature, attentions[1])\n\n ## object dynamic feature \n attentions = self.object_indomain_dynamic_attention(object_feature)\n indomain_object_feature = torch.sum(object_feature_b*attentions[0], dim=1)\n indomain_object_feature = torch.matmul(indomain_object_feature, attentions[1])\n\n ## action dynamic feature \n attentions = self.action_indomain_dynamic_attention(action_feature)\n indomain_action_feature = torch.sum(action_feature_b*attentions[0], dim=1)\n indomain_action_feature = torch.matmul(indomain_action_feature, attentions[1])\n\n #### cross-domain dynamic convolution\n ### multiple branches convolution\n concept_feature = torch.cat((scene_feature_s, object_feature_s), dim=2)\n concept_feature = torch.cat((concept_feature, action_feature_s), dim=2)\n concept_feature = F.relu(self.crossdomain_bn1(concept_feature))\n concept_feature = concept_feature.permute(0, 2, 1)\n \n crossdomain_feature = self.crossdomain_dynamic_attention(concept_feature)\n\n ## concat\n indomain_feature = torch.cat((indomain_scene_feature, indomain_object_feature), 1)\n indomain_feature = torch.cat((indomain_feature, indomain_action_feature), 1).squeeze(2)\n classification = torch.cat((indomain_feature, crossdomain_feature), 1)\n\n ## reduce dim\n classification = self.concat_bn1(classification)\n classification = self.relu(classification)\n classification = self.dropout(classification)\n classification = self.concat_reduce_dim(classification)\n \n ## classification\n classification = self.final_bn1(classification)\n classification = self.relu(classification)\n classification = self.dropout(classification)\n classification = self.final_classifier(classification)\n\n return classification\n", "sub_path": "networks/tdcmn_si_soa.py", "file_name": "tdcmn_si_soa.py", "file_ext": "py", "file_size_in_byte": 14986, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "torch.nn.Module", "line_number": 15, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 15, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 26, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 26, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 28, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 28, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 29, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 29, "usage_type": "name"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 33, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 33, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 34, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 37, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 37, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 38, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 44, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 44, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 45, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 48, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 48, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 49, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 53, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 53, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 54, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 54, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 54, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 55, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 55, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 58, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 58, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 59, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 59, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 59, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 60, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 60, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 60, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 65, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 66, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 66, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn.functional.linear", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 71, "usage_type": "name"}, {"api_name": "torch.nn.functional.linear", "line_number": 72, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 72, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 73, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 73, "usage_type": "name"}, {"api_name": "torch.tanh", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.functional.linear", "line_number": 76, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 76, "usage_type": "name"}, {"api_name": "torch.nn.functional.linear", "line_number": 77, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 77, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 78, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 78, "usage_type": "name"}, {"api_name": "torch.nn.Module", "line_number": 83, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 83, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 94, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 94, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 97, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 97, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 99, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 99, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 101, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 102, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 102, "usage_type": "name"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 105, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 106, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 106, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 108, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 111, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 117, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 117, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 118, "usage_type": "call"}, {"api_name": "torch.nn.parameter.Parameter", "line_number": 121, "usage_type": "call"}, {"api_name": "torch.Tensor", "line_number": 121, "usage_type": "call"}, {"api_name": "math.sqrt", "line_number": 122, "usage_type": "call"}, {"api_name": "torch.nn.Conv1d", "line_number": 126, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 126, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 127, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 127, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 127, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 128, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 128, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 131, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 131, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 132, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 132, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 132, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 133, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 133, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 133, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 140, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 141, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 141, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 144, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 147, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn.functional.linear", "line_number": 151, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 151, "usage_type": "name"}, {"api_name": "torch.nn.functional.linear", "line_number": 152, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 152, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 153, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 153, "usage_type": "name"}, {"api_name": "torch.sum", "line_number": 155, "usage_type": "call"}, {"api_name": "torch.tanh", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn.functional.linear", "line_number": 158, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 158, "usage_type": "name"}, {"api_name": "torch.nn.functional.linear", "line_number": 159, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 159, "usage_type": "name"}, {"api_name": "torch.nn.functional.softmax", "line_number": 160, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 160, "usage_type": "name"}, {"api_name": "torch.matmul", "line_number": 162, "usage_type": "call"}, {"api_name": "torch.nn.Module", "line_number": 167, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 167, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 190, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 190, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 191, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 191, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 194, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 194, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 197, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 197, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 198, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 198, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 201, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 201, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 204, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 204, "usage_type": "name"}, {"api_name": "torch.nn.ModuleList", "line_number": 205, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 205, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 208, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 208, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 210, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 210, "usage_type": "name"}, {"api_name": "torch.nn.AdaptiveAvgPool1d", "line_number": 222, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 222, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 226, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 226, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 227, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 227, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 230, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 230, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 231, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 231, "usage_type": "name"}, {"api_name": "torch.nn.Dropout", "line_number": 232, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 232, "usage_type": "name"}, {"api_name": "torch.nn.ReLU", "line_number": 233, "usage_type": "call"}, {"api_name": "torch.nn", "line_number": 233, "usage_type": "name"}, {"api_name": "torch.nn.Conv1d", "line_number": 236, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 236, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 237, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 237, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 237, "usage_type": "name"}, {"api_name": "torch.nn.BatchNorm1d", "line_number": 238, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 238, "usage_type": "name"}, {"api_name": "torch.nn.Linear", "line_number": 241, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 241, "usage_type": "name"}, {"api_name": "torch.nn.init.kaiming_normal_", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 242, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 242, "usage_type": "name"}, {"api_name": "torch.nn.init.constant_", "line_number": 243, "usage_type": "call"}, {"api_name": "torch.nn.init", "line_number": 243, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 243, "usage_type": "name"}, {"api_name": "torch.max", "line_number": 264, "usage_type": "call"}, {"api_name": "torch.max", "line_number": 265, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 283, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 283, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 284, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 284, "usage_type": "name"}, {"api_name": "torch.nn.functional.relu", "line_number": 285, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 285, "usage_type": "name"}, {"api_name": "torch.stack", "line_number": 295, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 299, "usage_type": "call"}, {"api_name": "torch.stack", "line_number": 303, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 308, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 313, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 314, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 318, "usage_type": "call"}, {"api_name": "torch.matmul", "line_number": 319, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 323, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 324, "usage_type": "call"}, {"api_name": "torch.nn.functional.relu", "line_number": 325, "usage_type": "call"}, {"api_name": "torch.nn.functional", "line_number": 325, "usage_type": "name"}, {"api_name": "torch.cat", "line_number": 331, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 332, "usage_type": "call"}, {"api_name": "torch.cat", "line_number": 333, "usage_type": "call"}]} +{"seq_id": "181739641", "text": "\"\"\"\nUtility functions for custom reports\n\"\"\"\nfrom datetime import datetime, timedelta\nfrom xformmanager.models import Metadata\n\ndef forms_submitted(chw_username, days=0, weeks=0, hours=0):\n \"\"\"Returns a count of the number of forms submitted over the given timeframe\n Timeframe - the period of time during which these forms were submitted\n \"\"\"\n metas = Metadata.objects.filter(username=chw_username)\n delta = timedelta(days=days, weeks=weeks, hours=hours)\n since = datetime.today() - delta\n metas = metas.filter(attachment__submission__submit_time__gte=since)\n return metas.count()\n\n\n", "sub_path": "apps/reports/sms/util.py", "file_name": "util.py", "file_ext": "py", "file_size_in_byte": 611, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "xformmanager.models.Metadata.objects.filter", "line_number": 11, "usage_type": "call"}, {"api_name": "xformmanager.models.Metadata.objects", "line_number": 11, "usage_type": "attribute"}, {"api_name": "xformmanager.models.Metadata", "line_number": 11, "usage_type": "name"}, {"api_name": "datetime.timedelta", "line_number": 12, "usage_type": "call"}, {"api_name": "datetime.datetime.today", "line_number": 13, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 13, "usage_type": "name"}]} +{"seq_id": "549389602", "text": "'''\nscript helps preview the annotated ground truth\n'''\n\nfrom PIL import Image\nimport numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport mplcursors\nimport sys\n\nimage = Image.open(sys.argv[1])\ndata = np.array(image)\nimg = plt.imshow(data, cmap=matplotlib.cm.plasma, norm=matplotlib.colors.Normalize(vmin=3, vmax=100))\n\npoints = []\n\ncursor = mplcursors.cursor(img, hover=False)\n@cursor.connect(\"add\")\ndef cursor_clicked(sel):\n # sel.annotation.set_visible(False)\n sel.annotation.set_text(\n f'Clicked on\\nx: {sel.target[0]:.2f} y: {sel.target[1]:.2f}\\nindex: {sel.target.index}')\n points.append(sel.target.index)\n print(\"Current list of points:\", points)\n\nplt.title(sys.argv[1])\nplt.show()\nprint(\"Selected points:\", points)", "sub_path": "Conversion4Seamseg/misc/click.py", "file_name": "click.py", "file_ext": "py", "file_size_in_byte": 757, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PIL.Image.open", "line_number": 12, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 12, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 12, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 13, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.imshow", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 14, "usage_type": "name"}, {"api_name": "matplotlib.cm", "line_number": 14, "usage_type": "attribute"}, {"api_name": "matplotlib.colors.Normalize", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.colors", "line_number": 14, "usage_type": "attribute"}, {"api_name": "mplcursors.cursor", "line_number": 18, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.show", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}]} +{"seq_id": "461334528", "text": "#!/usr/bin/env python\n# encoding: utf-8\n'''\n Get all tweet by user, Twitter API allows retrieval of only 3200 most recent tweets,\n Special handling involved when calling for full text of truncated tweets\n\n Stores 'latest_status_recorded_for_[username].txt' file to allow scrapper to only pull new tweets\n\n\n Output:\n ./../data/tweets/tweets_by_[]_.csv\n ./../data/users/user_info_[]_.csv\n ./../data/status_tracking/latest_status_recorded_for_[].csv\n\n'''\n\nimport tweepy\nimport os\nimport csv\nimport json\n\nfrom datetime import datetime as dt\nfrom collections import defaultdict\n\nfrom get_api import get_api\n\n\n# def user_data_to_csv(tweet, username, today):\ndef user_data_to_csv(user_data=None, username='', today=''):\n\n user_dict = defaultdict(str)\n\n\n user_details = ['id', 'id_str', 'name', 'screen_name', 'location',\n 'description', 'url', 'protected', 'followers_count',\n 'friends_count', 'listed_count', 'created_at',\n 'favourites_count', 'utc_offset', 'time_zone',\n 'geo_enabled', 'verified', 'statuses_count', 'lang',\n 'contributors_enabled', 'is_translator',\n 'is_translation_enabled', 'profile_background_color',\n 'profile_background_image_url',\n 'profile_background_image_url_https',\n 'profile_background_tile', 'profile_image_url',\n 'profile_image_url_https', 'profile_banner_url',\n 'profile_link_color', 'profile_sidebar_border_color',\n 'profile_sidebar_fill_color', 'profile_text_color',\n 'profile_use_background_image', 'has_extended_profile',\n 'default_profile', 'default_profile_image', 'following',\n 'follow_request_sent', 'notifications', 'translator_type']\n\n # collect user data\n for detail in user_details:\n try:\n user_dict[detail] = user_data[detail]\n except:\n pass\n\n # write to file\n\n # ensure destination directory exists\n destination_dir = f'./../data/users/{today[:6]}/{today[6:8]}//'\n if not os.path.exists(destination_dir):\n os.makedirs(destination_dir)\n\n # set column names dict keys\n csv_columns = user_dict.keys()\n\n # format file name\n csv_file = f'{destination_dir}user_info_[{username.lower()}]_{today}.csv'\n\n # write to csv file\n try:\n with open(csv_file, 'w') as output_file:\n writer = csv.writer(output_file)\n for key, value in user_dict.items():\n writer.writerow([key, value])\n\n print(f'[{dt.now().time()}] - File Written to: {csv_file}')\n\n except Exception as e:\n print(f'ERROR WRITING CSV FILE: {e}')\n pass\n\n return user_dict\n\ndef get_latest_record_for_user(username=''):\n result = None\n\n if username != '':\n # try to open record file\n try:\n with open(f'../data/status_tracking/latest_status_recorded_for_[{username}].txt', 'r') as input_file:\n result = str(input_file.read().replace('\\n', ''))\n print(f'Found a record tracking file!, collecting all available tweets after {result}')\n except:\n print('Could not find a record tracking file, collecting all available tweets')\n\n\n return result\n\ndef get_user_tweets(api = None, username='', collect_user = True):\n ''' username can also be passed as a user id string '''\n\n #get screen name as username\n username = api.get_user(username)._json['screen_name']\n print(f'Getting tweets and userdata for @{username}')\n\n today = f'{dt.now().year}{dt.now().month:02}{dt.now().day:02}{dt.now().hour:02}{dt.now().minute:02}'\n\n all_tweets = []\n latest_status_recorded = get_latest_record_for_user(username=username)\n\n # get screen_name from username\n\n for status in tweepy.Cursor(api.user_timeline, id=username, since_id=latest_status_recorded).items():\n\n # process status here\n tweet = status._json\n\n status_dict = defaultdict(str)\n\n status_details = ['created_at', 'id', 'id_str', 'text', 'full_text', 'truncated',\n 'source', 'in_reply_to_status_id',\n 'in_reply_to_status_id_str', 'in_reply_to_user_id',\n 'in_reply_to_user_id_str',\n 'in_reply_to_screen_name', 'geo', 'coordinates',\n 'place', 'contributors', 'is_quote_status',\n 'retweet_count', 'favorite_count', 'favorited',\n 'retweeted', 'lang']\n\n # collect user data\n for detail in status_details:\n try:\n status_dict[detail] = str(tweet[detail]).replace(',','')\n except:\n pass\n\n # if truncated, get full text, this is an extra API call...\n if status_dict['truncated'] == 'True':\n status_dict['full_text'] = str(api.get_status(status_dict['id'], tweet_mode='extended')._json['full_text']).replace(',', '')\n else:\n status_dict['full_text'] = status_dict['text']\n\n all_tweets.append(dict(status_dict))\n\n print(f'\\tcollected {len(all_tweets)} tweets...')\n if len(all_tweets) > 0:\n\n # get user date on just a single tweet, save time\n if collect_user:\n user_dict = user_data_to_csv(user_data=tweet['user'], username=username, today=today)\n\n # write to file\n\n # ensure destination directory exists\n destination_dir = f'./../data/tweets/{today[:6]}/{today[6:8]}//'\n if not os.path.exists(destination_dir):\n os.makedirs(destination_dir)\n\n # set column names dict keys\n csv_columns = list(all_tweets[0].keys())\n\n # format file name\n csv_file = f'{destination_dir}tweets_by_[{username.lower()}]_{today}.csv'\n\n # write to csv file\n try:\n with open(csv_file, 'w') as output_file:\n writer = csv.DictWriter(output_file, fieldnames=csv_columns)\n writer.writeheader()\n for data in all_tweets:\n writer.writerow(data)\n\n # for key, value in status_dict.items():\n # writer.writerow([key, value])\n\n print(f'[{dt.now().time()}] - File Written to: {csv_file}')\n\n except Exception as e:\n print(f'ERROR WRITING CSV FILE: {e}')\n pass\n\n # return latest status id\n\n latest_status_recorded = all_tweets[0]['id']\n\n # write file, overwrite if already exists\n # ensure destination directory exists\n destination_dir = './../data/status_tracking/'\n if not os.path.exists(destination_dir):\n os.makedirs(destination_dir)\n\n tracking_file = f'{destination_dir}latest_status_recorded_for_[{username.lower()}].txt'\n\n with open(tracking_file, 'w') as output_file:\n output_file.write(f'{latest_status_recorded}')\n print(f'[{dt.now().time()}] - File Written to: {tracking_file}')\n\n\n return latest_status_recorded\n\n else:\n print(f'{username} has not tweeted since we last checked. no new tweets to record')\n\n return latest_status_recorded\n\ndef main():\n username = 'samgutentag'\n latest_status = get_user_tweets(api = get_api(-1), username=username, collect_user = True)\n\n\nif __name__ == '__main__':\n main()\n# EOF\n", "sub_path": "scripts/get_user_tweets.py", "file_name": "get_user_tweets.py", "file_ext": "py", "file_size_in_byte": 7529, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "collections.defaultdict", "line_number": 31, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 62, "usage_type": "call"}, {"api_name": "os.path", "line_number": 62, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 63, "usage_type": "call"}, {"api_name": "csv.writer", "line_number": 74, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 78, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 78, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 108, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 108, "usage_type": "name"}, {"api_name": "tweepy.Cursor", "line_number": 115, "usage_type": "call"}, {"api_name": "collections.defaultdict", "line_number": 120, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 157, "usage_type": "call"}, {"api_name": "os.path", "line_number": 157, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 158, "usage_type": "call"}, {"api_name": "csv.DictWriter", "line_number": 169, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 177, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 177, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 190, "usage_type": "call"}, {"api_name": "os.path", "line_number": 190, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 191, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 197, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 197, "usage_type": "name"}, {"api_name": "get_api.get_api", "line_number": 209, "usage_type": "call"}]} +{"seq_id": "501279358", "text": "# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport mock\nfrom neutronclient.v2_0 import client\nfrom oslo_config import cfg\n\nfrom ironic.common import neutron\nfrom ironic.tests import base\n\n\nclass TestNeutronClient(base.TestCase):\n\n def setUp(self):\n super(TestNeutronClient, self).setUp()\n self.config(url='test-url',\n url_timeout=30,\n retries=2,\n group='neutron')\n self.config(insecure=False,\n certfile='test-file',\n admin_user='test-admin-user',\n admin_tenant_name='test-admin-tenant',\n admin_password='test-admin-password',\n auth_uri='test-auth-uri',\n group='keystone_authtoken')\n\n @mock.patch.object(client.Client, \"__init__\")\n def test_get_neutron_client_with_token(self, mock_client_init):\n token = 'test-token-123'\n expected = {'timeout': 30,\n 'retries': 2,\n 'insecure': False,\n 'ca_cert': 'test-file',\n 'token': token,\n 'endpoint_url': 'test-url',\n 'username': 'test-admin-user',\n 'tenant_name': 'test-admin-tenant',\n 'password': 'test-admin-password',\n 'auth_url': 'test-auth-uri'}\n\n mock_client_init.return_value = None\n neutron.get_client(token=token)\n mock_client_init.assert_called_once_with(**expected)\n\n @mock.patch.object(client.Client, \"__init__\")\n def test_get_neutron_client_without_token(self, mock_client_init):\n expected = {'timeout': 30,\n 'retries': 2,\n 'insecure': False,\n 'ca_cert': 'test-file',\n 'token': None,\n 'endpoint_url': 'test-url',\n 'username': 'test-admin-user',\n 'tenant_name': 'test-admin-tenant',\n 'password': 'test-admin-password',\n 'auth_url': 'test-auth-uri'}\n\n mock_client_init.return_value = None\n neutron.get_client(token=None)\n mock_client_init.assert_called_once_with(**expected)\n\n @mock.patch.object(client.Client, \"__init__\")\n def test_get_neutron_client_with_region(self, mock_client_init):\n expected = {'timeout': 30,\n 'retries': 2,\n 'insecure': False,\n 'ca_cert': 'test-file',\n 'token': None,\n 'endpoint_url': 'test-url',\n 'username': 'test-admin-user',\n 'tenant_name': 'test-admin-tenant',\n 'password': 'test-admin-password',\n 'auth_url': 'test-auth-uri',\n 'region_name': 'test-region'}\n\n self.config(region_name='test-region',\n group='keystone')\n mock_client_init.return_value = None\n neutron.get_client(token=None)\n mock_client_init.assert_called_once_with(**expected)\n\n @mock.patch.object(client.Client, \"__init__\")\n def test_get_neutron_client_noauth(self, mock_client_init):\n self.config(auth_strategy='noauth', group='neutron')\n expected = {'ca_cert': 'test-file',\n 'insecure': False,\n 'endpoint_url': 'test-url',\n 'timeout': 30,\n 'retries': 2,\n 'auth_strategy': 'noauth'}\n\n mock_client_init.return_value = None\n neutron.get_client(token=None)\n mock_client_init.assert_called_once_with(**expected)\n\n def test_out_range_auth_strategy(self):\n self.assertRaises(ValueError, cfg.CONF.set_override,\n 'auth_strategy', 'fake', 'neutron',\n enforce_type=True)\n", "sub_path": "ironic/tests/unit/common/test_neutron.py", "file_name": "test_neutron.py", "file_ext": "py", "file_size_in_byte": 4395, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "ironic.tests.base.TestCase", "line_number": 21, "usage_type": "attribute"}, {"api_name": "ironic.tests.base", "line_number": 21, "usage_type": "name"}, {"api_name": "ironic.common.neutron.get_client", "line_number": 52, "usage_type": "call"}, {"api_name": "ironic.common.neutron", "line_number": 52, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 37, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 37, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client.Client", "line_number": 37, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client", "line_number": 37, "usage_type": "name"}, {"api_name": "ironic.common.neutron.get_client", "line_number": 69, "usage_type": "call"}, {"api_name": "ironic.common.neutron", "line_number": 69, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 55, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 55, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client.Client", "line_number": 55, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client", "line_number": 55, "usage_type": "name"}, {"api_name": "ironic.common.neutron.get_client", "line_number": 89, "usage_type": "call"}, {"api_name": "ironic.common.neutron", "line_number": 89, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 72, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 72, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client.Client", "line_number": 72, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client", "line_number": 72, "usage_type": "name"}, {"api_name": "ironic.common.neutron.get_client", "line_number": 103, "usage_type": "call"}, {"api_name": "ironic.common.neutron", "line_number": 103, "usage_type": "name"}, {"api_name": "mock.patch.object", "line_number": 92, "usage_type": "call"}, {"api_name": "mock.patch", "line_number": 92, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client.Client", "line_number": 92, "usage_type": "attribute"}, {"api_name": "neutronclient.v2_0.client", "line_number": 92, "usage_type": "name"}, {"api_name": "oslo_config.cfg.CONF", "line_number": 107, "usage_type": "attribute"}, {"api_name": "oslo_config.cfg", "line_number": 107, "usage_type": "name"}]} +{"seq_id": "331784515", "text": "from params_container import Container\nfrom requests import get\nfrom bs4 import BeautifulSoup\nimport csv\nimport re\nimport os\nimport sys\nimport argparse\nimport unittest\n\n\nclass PageParser:\n def __init__(self,query):\n self.query = query\n self.page = None\n self.pagenum = 0\n \n def get_page(self):\n \"\"\"\n create a query url and return a page with results\n \"\"\"\n params = {'query': self.query,\n 'corpname': 'qirim',\n 'start':self.pagenum}\n s = get('http://korpus.juls.savba.sk:8080/manatee.ks/do_query', params=params)\n return s.text\n\n\n def parse_page(self):\n \"\"\"\n parse the page\n \"\"\"\n left_list = []\n center_list = []\n right_list = []\n soup = BeautifulSoup(self.page, 'lxml')\n for left in soup.select('td[class=\"lc\"]'):\n left_list.append(left.text)\n for center in soup.select('td[class=\"kwic\"]'):\n center_list.append(center.text)\n for right in soup.select('td[class=\"rc\"]'):\n right_list.append(right.text)\n res = [[l.strip(),c.strip(),r.strip()] for l,c,r in zip(left_list, center_list, right_list)]\n return res\n\n def extract_results(self):\n self.page = self.get_page()\n parsed_results = self.parse_page()\n return parsed_results\n\n \nclass Downloader(Container):\n def __init__(self,*args,**kwargs):\n super().__init__(*args,**kwargs)\n\n\n def download_all(self):\n \"\"\"\n get information and hits from first page and iterate until\n all hits are collected or the maximum set by user is achieved\n \"\"\"\n s = []\n parser = PageParser(self.query)\n for i in range(0, self.numResults-1, 10):\n try:\n parser.pagenum = i\n s += parser.extract_results()\n except:\n return []\n if len(s) > self.numResults:\n s = s[:self.numResults]\n return s\n \n\n# rewrite\nclass TestMethods(unittest.TestCase):\n\n def test1(self):\n self.assertEqual('', str(get_page(query='къырым')))\n\n def test2(self):\n self.assertIs(list, type(get_results(get_page(query='къырым'),\n n_results=50)))\n\n def test3(self):\n results = main(query='къырым', n_results=30, kwic=True, write=True)\n filelist = os.listdir('.')\n self.assertIn('crh_search_къырым.csv', filelist)\n os.remove('crh_search_къырым.csv')\n\nif __name__ == '__main__':\n unittest.main()\n args = sys.argv[1:]\n parser = argparse.ArgumentParser(prog='crh_corpus.py')\n parser.add_argument('query', type=str)\n parser.add_argument('n_results', type=int, default=10)\n parser.add_argument('kwic', type=bool, default=True)\n parser.add_argument('write', type=bool, default=False)\n args = parser.parse_args(args)\n main(**vars(args))\n", "sub_path": "corpora_for_refactor/crh_corpus.py", "file_name": "crh_corpus.py", "file_ext": "py", "file_size_in_byte": 3002, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "requests.get", "line_number": 25, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 36, "usage_type": "call"}, {"api_name": "params_container.Container", "line_number": 52, "usage_type": "name"}, {"api_name": "unittest.TestCase", "line_number": 76, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 87, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 89, "usage_type": "call"}, {"api_name": "unittest.main", "line_number": 92, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 93, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 94, "usage_type": "call"}]} +{"seq_id": "554796739", "text": "from typing import Any, Callable, Container, Dict, List, Optional, Tuple, TypedDict, cast\n\nfrom pydantic import Field\nfrom pydantic.fields import FieldInfo\nfrom sqlalchemy import Column, Enum, inspect\nfrom sqlalchemy.orm import ColumnProperty\nfrom sqlalchemy.types import TypeEngine\n\nfrom alchemista import func\n\n\nclass Info(TypedDict, total=False):\n alias: str\n allow_mutation: bool\n const: Any\n default: Any\n default_factory: Callable[[], Any]\n description: str\n example: str\n ge: float\n gt: float\n le: float\n lt: float\n max_items: int\n max_length: int\n min_items: int\n min_length: int\n multiple_of: float\n regex: str\n title: str\n\n\ndef _extract_python_type(type_engine: TypeEngine) -> type: # type: ignore[type-arg]\n try:\n # the `python_type` seems to always be a @property-decorated method,\n # so only checking its existence is not enough\n return type_engine.python_type\n except (AttributeError, NotImplementedError):\n return cast(type, type_engine.impl.python_type) # type: ignore[attr-defined]\n\n\ndef infer_python_type(column: Column) -> type: # type: ignore[type-arg]\n try:\n python_type = _extract_python_type(column.type)\n except (AttributeError, NotImplementedError) as ex:\n raise RuntimeError(\n f\"Could not infer the Python type for {column}.\"\n \" Check if the column type has a `python_type` in it or in `impl`\"\n ) from ex\n\n if python_type is list and hasattr(column.type, \"item_type\"):\n item_type = _extract_python_type(column.type.item_type)\n if column.nullable:\n return Optional[List[item_type]] # type: ignore[valid-type, return-value]\n return List[item_type] # type: ignore[valid-type]\n\n return python_type if not column.nullable else Optional[python_type] # type: ignore[return-value]\n\n\ndef _get_default_scalar(column: Column) -> Any: # type: ignore[type-arg]\n if column.default and column.default.is_scalar:\n return column.default.arg\n if column.nullable is False:\n return ...\n return None\n\n\ndef _maybe_set_max_length_from_column(field_kwargs: Info, column: Column) -> None: # type: ignore[type-arg]\n # some types have a length in the backend, but setting that interferes with the model generation\n # maybe we should list the types that we *should set* the length, instead of *not set* the length?\n if not isinstance(column.type, Enum):\n sa_type_length = getattr(column.type, \"length\", None)\n if sa_type_length is not None:\n field_kwargs[\"max_length\"] = sa_type_length\n\n\ndef make_field(column: Column) -> FieldInfo: # type: ignore[type-arg]\n info = Info()\n if column.info:\n for key in Info.__annotations__.keys(): # pylint: disable=no-member\n if key in column.info:\n info[key] = column.info[key] # type: ignore[misc]\n\n if \"max_length\" not in info:\n _maybe_set_max_length_from_column(info, column)\n\n if \"description\" not in info and column.doc:\n info[\"description\"] = column.doc\n\n if \"default\" in info and \"default_factory\" in info:\n raise ValueError(\n f\"Both `default` and `default_factory` were specified in info of column `{column.name}`.\"\n \" These two attributes are mutually-exclusive\"\n )\n\n if \"default\" not in info and \"default_factory\" not in info and column.default and column.default.is_callable:\n return cast(FieldInfo, Field(**info, default_factory=column.default.arg.__wrapped__)) # type: ignore[misc]\n\n if \"default_factory\" in info:\n return cast(FieldInfo, Field(**info))\n\n # pop `default` because it is not a keyword argument of `Field`\n default = info.pop(\"default\") if \"default\" in info else _get_default_scalar(column)\n return cast(FieldInfo, Field(default, **info)) # type: ignore[misc]\n\n\ndef fields_from(\n db_model: type,\n *,\n exclude: Optional[Container[str]] = None,\n include: Optional[Container[str]] = None,\n transform: Callable[[str, type, FieldInfo], Tuple[type, FieldInfo]] = func.unchanged,\n) -> Dict[str, Tuple[type, FieldInfo]]:\n if exclude and include:\n raise ValueError(\"`exclude` and `include` are mutually-exclusive\")\n mapper = inspect(db_model)\n candidate_attrs = mapper.attrs\n if exclude:\n candidate_attrs = (attr for attr in mapper.attrs if attr.key not in exclude)\n elif include:\n candidate_attrs = (attr for attr in mapper.attrs if attr.key in include)\n fields = {}\n for attr in candidate_attrs:\n if isinstance(attr, ColumnProperty) and attr.columns:\n name = attr.key\n column = attr.columns[0]\n python_type = infer_python_type(column)\n field = make_field(column)\n fields[name] = transform(name, python_type, field)\n return fields\n", "sub_path": "alchemista/field.py", "file_name": "field.py", "file_ext": "py", "file_size_in_byte": 4876, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "typing.TypedDict", "line_number": 12, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 15, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 16, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 17, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 17, "usage_type": "name"}, {"api_name": "sqlalchemy.types.TypeEngine", "line_number": 33, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 39, "usage_type": "call"}, {"api_name": "sqlalchemy.Column", "line_number": 42, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 54, "usage_type": "name"}, {"api_name": "typing.List", "line_number": 55, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 57, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 60, "usage_type": "name"}, {"api_name": "typing.Any", "line_number": 60, "usage_type": "name"}, {"api_name": "sqlalchemy.Column", "line_number": 68, "usage_type": "name"}, {"api_name": "sqlalchemy.Enum", "line_number": 71, "usage_type": "argument"}, {"api_name": "sqlalchemy.Column", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.cast", "line_number": 97, "usage_type": "call"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 97, "usage_type": "argument"}, {"api_name": "pydantic.Field", "line_number": 97, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 100, "usage_type": "call"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 100, "usage_type": "argument"}, {"api_name": "pydantic.Field", "line_number": 100, "usage_type": "call"}, {"api_name": "typing.cast", "line_number": 104, "usage_type": "call"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 104, "usage_type": "argument"}, {"api_name": "pydantic.Field", "line_number": 104, "usage_type": "call"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 77, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Container", "line_number": 110, "usage_type": "name"}, {"api_name": "typing.Optional", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Container", "line_number": 111, "usage_type": "name"}, {"api_name": "typing.Callable", "line_number": 112, "usage_type": "name"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 112, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 112, "usage_type": "name"}, {"api_name": "alchemista.func.unchanged", "line_number": 112, "usage_type": "attribute"}, {"api_name": "alchemista.func", "line_number": 112, "usage_type": "name"}, {"api_name": "sqlalchemy.inspect", "line_number": 116, "usage_type": "call"}, {"api_name": "sqlalchemy.orm.ColumnProperty", "line_number": 124, "usage_type": "argument"}, {"api_name": "typing.Dict", "line_number": 113, "usage_type": "name"}, {"api_name": "typing.Tuple", "line_number": 113, "usage_type": "name"}, {"api_name": "pydantic.fields.FieldInfo", "line_number": 113, "usage_type": "name"}]} +{"seq_id": "185417977", "text": "from django import forms\nfrom django.db.models import Q\nfrom django.utils.translation import gettext_lazy as _\n\nfrom common.util import email_user_async\nfrom funding.models import Attribution, FundingSource\nfrom institution.models import Institution\nfrom project.models import (\n Project, ProjectUserMembership, RSEAllocation, SystemAllocationRequest\n)\nfrom project.openldap import (\n update_openldap_project, update_openldap_project_membership\n)\nfrom users.models import CustomUser, Profile\n\nPROJECT_CODE_PREFIX = \"scw\"\n\n\nclass FileLinkWidget(forms.Widget):\n\n def __init__(self, obj, attrs=None):\n self.object = obj\n super(FileLinkWidget, self).__init__(attrs)\n\n def render(self, name, value, attrs=None, renderer=None):\n if self.object.pk:\n return u'Download'.format(\n self.object.id\n )\n else:\n return u''\n\n\nclass SelectMultipleTickbox(forms.widgets.CheckboxSelectMultiple):\n template_name = 'project/attributionwidget.html'\n\n def __init__(self, *args, project_id=None, **kwargs):\n self.project_id = project_id\n return super().__init__(*args, **kwargs)\n\n def get_context(self, name, value, attrs):\n context = super().get_context(name, value, attrs)\n if self.project_id:\n context['project_id'] = self.project_id\n return context\n\n\nclass ProjectAdminForm(forms.ModelForm):\n\n class Meta:\n model = Project\n fields = [\n 'title',\n 'description',\n 'legacy_hpcw_id',\n 'legacy_arcca_id',\n 'institution_reference',\n 'department',\n 'gid_number',\n 'supervisor_name',\n 'supervisor_position',\n 'supervisor_email',\n 'approved_by_supervisor',\n 'attributions',\n 'tech_lead',\n 'category',\n 'economic_user',\n 'custom_user_cap',\n ]\n\n def __init__(self, *args, **kwargs):\n super(ProjectAdminForm, self).__init__(*args, **kwargs)\n\n def clean_code(self):\n \"\"\"\n Ensure the project code is unique.\n \"\"\"\n current_code = self.instance.code\n updated_code = self.cleaned_data['code']\n if current_code != updated_code:\n if Project.objects.filter(code=updated_code).exists():\n raise forms.ValidationError(_('Project code must be unique.'))\n return updated_code\n\n def clean_legacy_hpcw_id(self):\n \"\"\"\n Ensure the project legacy HPCW id is unique.\n \"\"\"\n current_legacy_hpcw_id = self.instance.legacy_hpcw_id\n updated_legacy_hpcw_id = self.cleaned_data['legacy_hpcw_id']\n if updated_legacy_hpcw_id.startswith(PROJECT_CODE_PREFIX):\n raise forms.ValidationError(_('SCW Project codes are reserved.'))\n if current_legacy_hpcw_id != updated_legacy_hpcw_id:\n if Project.objects.filter(legacy_hpcw_id=updated_legacy_hpcw_id\n ).exists():\n raise forms.ValidationError(_('Project legacy HPCW id must be unique.'))\n return updated_legacy_hpcw_id\n\n def clean_legacy_arcca_id(self):\n \"\"\"\n Ensure the project legacy ARCCA id is unique.\n \"\"\"\n current_legacy_arcca_id = self.instance.legacy_arcca_id\n updated_legacy_arcca_id = self.cleaned_data['legacy_arcca_id']\n if updated_legacy_arcca_id.startswith(PROJECT_CODE_PREFIX):\n raise forms.ValidationError(_('SCW Project codes are reserved.'))\n if current_legacy_arcca_id != updated_legacy_arcca_id:\n if Project.objects.filter(legacy_arcca_id=updated_legacy_arcca_id\n ).exists():\n raise forms.ValidationError(_('Project legacy ARCCA id must be unique.'))\n return updated_legacy_arcca_id\n\n def save(self, commit=True):\n project = super(ProjectAdminForm, self).save(commit=False)\n if commit:\n project.save()\n return project\n\n\nclass SystemAllocationRequestAdminForm(forms.ModelForm):\n\n document_download = forms.CharField(\n label=_('Download Supporting Document'),\n required=False\n )\n\n class Meta:\n model = SystemAllocationRequest\n fields = [\n 'project',\n 'information',\n 'start_date',\n 'end_date',\n 'allocation_rse',\n 'allocation_cputime',\n 'allocation_memory',\n 'allocation_storage_home',\n 'allocation_storage_scratch',\n 'requirements_software',\n 'requirements_training',\n 'requirements_onboarding',\n 'document',\n 'document_download',\n 'status',\n 'reason_decision',\n 'notes',\n ]\n\n def __init__(self, *args, **kwargs):\n super(SystemAllocationRequestAdminForm, self).__init__(*args, **kwargs)\n self.initial_status = self.instance.status\n self.fields['document_download'].widget = FileLinkWidget(self.instance)\n self.fields['status'] = forms.ChoiceField(\n choices=self._get_status_choices(self.instance.status)\n )\n if self.instance.id is None:\n self.fields['status'] = forms.ChoiceField(\n choices=[\n SystemAllocationRequest.\n STATUS_CHOICES[SystemAllocationRequest.AWAITING_APPROVAL]\n ]\n )\n\n def _get_status_choices(self, status):\n # yapf: disable\n pre_approved_options = [\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.AWAITING_APPROVAL],\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.APPROVED],\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.DECLINED],\n ]\n post_approved_options = [\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.APPROVED],\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.REVOKED],\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.SUSPENDED],\n SystemAllocationRequest.STATUS_CHOICES[SystemAllocationRequest.CLOSED],\n ]\n if SystemAllocationRequest.STATUS_CHOICES[status] in post_approved_options:\n return post_approved_options\n else:\n return pre_approved_options\n # yapf: enable\n\n def save(self, commit=True):\n allocation = super(SystemAllocationRequestAdminForm,\n self).save(commit=False)\n allocation.previous_status = self.initial_status\n if self.initial_status != allocation.status:\n update_openldap_project(allocation)\n if commit:\n allocation.save()\n return allocation\n\n\nclass ProjectCreationForm(forms.ModelForm):\n\n class Meta:\n model = Project\n fields = [\n 'title',\n 'description',\n 'institution_reference',\n 'department',\n 'supervisor_name',\n 'supervisor_position',\n 'supervisor_email',\n 'attributions',\n ]\n\n def __init__(self, user, *args, **kwargs):\n super(ProjectCreationForm, self).__init__(*args, **kwargs)\n self.user = user\n if self.user.profile.institution is not None and not self.user.profile.institution.needs_funding_workflow:\n del self.fields['attributions']\n else:\n self.fields['attributions'] = forms.ModelMultipleChoiceField(\n label='',\n widget=SelectMultipleTickbox(),\n queryset=Attribution.objects.filter(created_by=self.user),\n required=False,\n )\n del self.fields['institution_reference']\n\n def clean_supervisor_email(self):\n cleaned_data = super().clean()\n try:\n email = cleaned_data['supervisor_email']\n domain = email.split('@')[1]\n if Institution.objects.filter(base_domain=domain).exists():\n return email\n except:\n pass\n raise forms.ValidationError(_(\n 'Please enter an institutional email address ending '\n 'with one of: ' + ', '.join(\n '@' + institution.base_domain\n for institution in Institution.objects.all()\n ) + '.'\n ))\n\n def clean(self):\n self.instance.tech_lead = self.user\n if self.instance.tech_lead.profile.institution is None:\n raise forms.ValidationError(_(\n 'Only users which belong to an institution can create projects.'\n ))\n\n\nclass ProjectAssociatedForm(forms.ModelForm):\n\n def __init__(\n self, user, include_project=True, project=None, *args, **kwargs\n ):\n super().__init__(*args, **kwargs)\n self.user = user\n if project:\n self.fields['project'].initial = project\n self.fields['project'].widget = forms.HiddenInput()\n elif include_project:\n self.fields['project'] = forms.ModelChoiceField(\n queryset=Project.objects.filter(tech_lead=user)\n )\n else:\n del self.fields['project']\n\n def clean_project(self):\n if self.cleaned_data['project'].tech_lead != self.user:\n raise forms.ValidationError(_('Selected project not found.'))\n return self.cleaned_data['project']\n\n\nclass ProjectManageAttributionForm(forms.ModelForm):\n\n class Meta:\n model = Project\n fields = ['attributions']\n\n def __init__(self, user, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.user = user\n owned_attributions = Attribution.objects.filter(owner=self.user,)\n # Also get funding sources the user is a member of\n fundingsources = Attribution.objects.filter(\n fundingsource__in=FundingSource.objects.\n filter(fundingsourcemembership__user=self.user,)\n )\n self.fields['attributions'] = forms.ModelMultipleChoiceField(\n label='',\n widget=SelectMultipleTickbox(project_id=kwargs['instance'].id),\n queryset=(owned_attributions | fundingsources),\n required=False,\n )\n\n\nclass ProjectSupervisorApproveForm(forms.ModelForm):\n\n class Meta:\n model = Project\n fields = ['approved_by_supervisor']\n\n def __init__(self, user, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.user = user\n\n def clean(self):\n if self.instance.supervisor_email != self.user.email:\n raise forms.ValidationError(_(\n 'You are not the supervisor of this project.'\n ))\n\n\nclass SystemAllocationRequestCreationForm(ProjectAssociatedForm):\n\n class Meta:\n model = SystemAllocationRequest\n fields = [\n 'information',\n 'start_date',\n 'end_date',\n 'allocation_cputime',\n 'allocation_memory',\n 'allocation_storage_home',\n 'allocation_storage_scratch',\n 'requirements_software',\n 'requirements_training',\n 'requirements_onboarding',\n 'document',\n 'project',\n ]\n widgets = {\n 'start_date': forms.DateInput(attrs={'class': 'datepicker'}),\n 'end_date': forms.DateInput(attrs={'class': 'datepicker'}),\n }\n\n\nclass RSEAllocationRequestCreationForm(ProjectAssociatedForm):\n\n class Meta:\n model = RSEAllocation\n fields = [\n 'title',\n 'duration',\n 'goals',\n 'software',\n 'outcomes',\n 'confidentiality',\n 'project',\n ]\n\n def save(self, *args, **kwargs):\n result = super().save(*args, **kwargs)\n if self.user.profile.institution.rse_notify_email:\n user_name = ' '.join((self.user.first_name, self.user.last_name))\n subject = _(\n 'New RSE time request from {}'.format(user_name)\n )\n context = {\n 'user': user_name,\n 'to': self.user.profile.institution.rse_notify_email,\n 'title': self.cleaned_data['title'],\n 'duration': self.cleaned_data['duration'],\n 'goals': self.cleaned_data['goals'],\n 'software': self.cleaned_data['software'],\n 'outcomes': self.cleaned_data['outcomes'],\n 'confidentiality': self.cleaned_data['confidentiality'],\n }\n text_template_path = 'notifications/project/rse_request.txt'\n html_template_path = 'notifications/project/rse_request.html'\n email_user_async.delay(subject, context, text_template_path, html_template_path)\n return result\n\n\nclass ProjectUserMembershipCreationForm(forms.Form):\n project_code = forms.CharField(max_length=20)\n\n def clean_project_code(self):\n # Verify the project code is valid and the project has been approved.\n project_code = self.cleaned_data['project_code']\n try:\n project = Project.objects.get(\n Q(code=project_code) | Q(legacy_hpcw_id=project_code) |\n Q(legacy_arcca_id=project_code)\n )\n user = self.initial.get('user', None)\n # The technical lead will automatically be added as a member of the of project.\n if project.tech_lead == user:\n raise forms.ValidationError(_(\"You are currently a member of the project.\"))\n if ProjectUserMembership.objects.filter(\n project=project, user=user\n ).exists():\n raise forms.ValidationError(_(\"A membership request for this project already exists.\"))\n except Project.DoesNotExist:\n raise forms.ValidationError(_(\"Invalid Project Code.\"))\n return project_code\n\n\nclass ProjectUserInviteForm(forms.Form):\n email = forms.CharField(max_length=50)\n initiated_by_user = False\n\n def clean_email(self):\n # Verify that the user exists and the project is approved\n email = self.cleaned_data['email']\n project = Project.objects.filter(id=self.initial['project_id']).first()\n user = CustomUser.objects.filter(email=email).first()\n # The technical lead will automatically be added as a member of the of project.\n if not user:\n raise forms.ValidationError(_(\"No user exists with given email.\"))\n if user.profile.account_status != Profile.APPROVED and user.profile.institution.needs_user_approval:\n raise forms.ValidationError( _('User is still awaiting authorisation'))\n if project.tech_lead == user:\n raise forms.ValidationError(_(\"You are currently a member of the project.\"))\n if ProjectUserMembership.objects.filter(\n project=project, user=user\n ).exists():\n raise forms.ValidationError(_(\"A membership request for this project already exists.\"))\n if not project.can_have_more_users():\n raise forms.ValidationError(_(\n \"This project has reached its membership cap. \"\n \"If you require more members, please contact support.\"\n ))\n return email\n\n\nclass ProjectUserMembershipAdminForm(forms.ModelForm):\n\n class Meta:\n model = ProjectUserMembership\n fields = [\n 'project',\n 'user',\n 'status',\n 'initiated_by_user',\n 'date_joined',\n 'date_left',\n ]\n\n def __init__(self, *args, **kwargs):\n super(ProjectUserMembershipAdminForm, self).__init__(*args, **kwargs)\n self.initial_status = self.instance.status\n self.fields['status'] = forms.ChoiceField(\n choices=self._get_status_choices(self.instance.status)\n )\n\n def _get_status_choices(self, status):\n # yapf: disable\n pre_approved_options = [\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.AWAITING_AUTHORISATION],\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.AUTHORISED],\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.DECLINED],\n ]\n post_approved_options = [\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.AUTHORISED],\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.REVOKED],\n ProjectUserMembership.STATUS_CHOICES[ProjectUserMembership.SUSPENDED],\n ]\n if ProjectUserMembership.STATUS_CHOICES[status] in post_approved_options:\n return post_approved_options\n else:\n return pre_approved_options\n # yapf: enable\n\n def clean_status(self):\n current_status = self.initial_status\n updated_status = self.cleaned_data['status']\n if current_status != updated_status:\n # Check for project user limit\n if not self.cleaned_data['project'].can_have_more_users():\n raise forms.ValidationError(_(\n \"This project has reached its membership cap. \"\n \"If you require more members, please contact support.\"\n ))\n return updated_status\n\n def save(self, commit=True):\n project_user_membership = super(ProjectUserMembershipAdminForm,\n self).save(commit=False)\n if self.initial_status != project_user_membership.status:\n if project_user_membership.status == ProjectUserMembership.AUTHORISED:\n user = project_user_membership.user\n if user.profile.institution and user.profile.institution.needs_user_approval:\n user.profile.activate()\n update_openldap_project_membership(project_user_membership)\n if commit:\n project_user_membership.save()\n return project_user_membership\n", "sub_path": "project/forms.py", "file_name": "forms.py", "file_ext": "py", "file_size_in_byte": 18074, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.forms.Widget", "line_number": 19, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 19, "usage_type": "name"}, {"api_name": "django.forms.widgets", "line_number": 34, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 34, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 48, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 48, "usage_type": "name"}, {"api_name": "project.models.Project", "line_number": 51, "usage_type": "name"}, {"api_name": "project.models.Project.objects.filter", "line_number": 81, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 81, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 81, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 82, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 82, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 82, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 92, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 92, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 92, "usage_type": "call"}, {"api_name": "project.models.Project.objects.filter", "line_number": 94, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 94, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 94, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 96, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 96, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 96, "usage_type": "call"}, {"api_name": "django.forms.ValidationError", "line_number": 106, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 106, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 106, "usage_type": "call"}, {"api_name": "project.models.Project.objects.filter", "line_number": 108, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 108, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 108, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 110, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 110, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 110, "usage_type": "call"}, {"api_name": "project.models", "line_number": 114, "usage_type": "name"}, {"api_name": "project.models.save", "line_number": 116, "usage_type": "call"}, {"api_name": "project.models", "line_number": 116, "usage_type": "name"}, {"api_name": "project.models", "line_number": 117, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 120, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 120, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 122, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 122, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 123, "usage_type": "call"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 128, "usage_type": "name"}, {"api_name": "django.forms.ChoiceField", "line_number": 153, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 153, "usage_type": "name"}, {"api_name": "django.forms.ChoiceField", "line_number": 157, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 157, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 159, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 159, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.AWAITING_APPROVAL", "line_number": 160, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 160, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 167, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 167, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.AWAITING_APPROVAL", "line_number": 167, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 168, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 168, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.APPROVED", "line_number": 168, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 169, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 169, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.DECLINED", "line_number": 169, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 172, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 172, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.APPROVED", "line_number": 172, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 173, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 173, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.REVOKED", "line_number": 173, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 174, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 174, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.SUSPENDED", "line_number": 174, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 175, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 175, "usage_type": "name"}, {"api_name": "project.models.SystemAllocationRequest.CLOSED", "line_number": 175, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest.STATUS_CHOICES", "line_number": 177, "usage_type": "attribute"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 177, "usage_type": "name"}, {"api_name": "project.openldap.update_openldap_project", "line_number": 188, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 194, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 194, "usage_type": "name"}, {"api_name": "project.models.Project", "line_number": 197, "usage_type": "name"}, {"api_name": "django.forms.ModelMultipleChoiceField", "line_number": 215, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 215, "usage_type": "name"}, {"api_name": "funding.models.Attribution.objects.filter", "line_number": 218, "usage_type": "call"}, {"api_name": "funding.models.Attribution.objects", "line_number": 218, "usage_type": "attribute"}, {"api_name": "funding.models.Attribution", "line_number": 218, "usage_type": "name"}, {"api_name": "institution.models.Institution.objects.filter", "line_number": 228, "usage_type": "call"}, {"api_name": "institution.models.Institution.objects", "line_number": 228, "usage_type": "attribute"}, {"api_name": "institution.models.Institution", "line_number": 228, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 232, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 232, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 232, "usage_type": "call"}, {"api_name": "institution.models.base_domain", "line_number": 235, "usage_type": "attribute"}, {"api_name": "institution.models", "line_number": 235, "usage_type": "name"}, {"api_name": "institution.models", "line_number": 236, "usage_type": "name"}, {"api_name": "institution.models.Institution.objects.all", "line_number": 236, "usage_type": "call"}, {"api_name": "institution.models.Institution.objects", "line_number": 236, "usage_type": "attribute"}, {"api_name": "institution.models.Institution", "line_number": 236, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 243, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 243, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 243, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 248, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 248, "usage_type": "name"}, {"api_name": "project.models", "line_number": 255, "usage_type": "name"}, {"api_name": "project.models", "line_number": 256, "usage_type": "name"}, {"api_name": "django.forms.HiddenInput", "line_number": 257, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 257, "usage_type": "name"}, {"api_name": "django.forms.ModelChoiceField", "line_number": 259, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 259, "usage_type": "name"}, {"api_name": "project.models.Project.objects.filter", "line_number": 260, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 260, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 260, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 267, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 267, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 267, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 271, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 271, "usage_type": "name"}, {"api_name": "project.models.Project", "line_number": 274, "usage_type": "name"}, {"api_name": "funding.models.Attribution.objects.filter", "line_number": 280, "usage_type": "call"}, {"api_name": "funding.models.Attribution.objects", "line_number": 280, "usage_type": "attribute"}, {"api_name": "funding.models.Attribution", "line_number": 280, "usage_type": "name"}, {"api_name": "funding.models.Attribution.objects.filter", "line_number": 282, "usage_type": "call"}, {"api_name": "funding.models.Attribution.objects", "line_number": 282, "usage_type": "attribute"}, {"api_name": "funding.models.Attribution", "line_number": 282, "usage_type": "name"}, {"api_name": "funding.models.FundingSource.objects.filter", "line_number": 283, "usage_type": "call"}, {"api_name": "funding.models.FundingSource.objects", "line_number": 283, "usage_type": "attribute"}, {"api_name": "funding.models.FundingSource", "line_number": 283, "usage_type": "name"}, {"api_name": "django.forms.ModelMultipleChoiceField", "line_number": 286, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 286, "usage_type": "name"}, {"api_name": "django.forms.ModelForm", "line_number": 294, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 294, "usage_type": "name"}, {"api_name": "project.models.Project", "line_number": 297, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 306, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 306, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 306, "usage_type": "call"}, {"api_name": "project.models.SystemAllocationRequest", "line_number": 314, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 330, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 330, "usage_type": "name"}, {"api_name": "django.forms.DateInput", "line_number": 331, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 331, "usage_type": "name"}, {"api_name": "project.models.RSEAllocation", "line_number": 338, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 353, "usage_type": "call"}, {"api_name": "common.util.email_user_async.delay", "line_number": 368, "usage_type": "call"}, {"api_name": "common.util.email_user_async", "line_number": 368, "usage_type": "name"}, {"api_name": "django.forms.Form", "line_number": 372, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 372, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 373, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 373, "usage_type": "name"}, {"api_name": "project.models", "line_number": 379, "usage_type": "name"}, {"api_name": "project.models.Project.objects.get", "line_number": 379, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 379, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 379, "usage_type": "name"}, {"api_name": "django.db.models.Q", "line_number": 380, "usage_type": "call"}, {"api_name": "django.db.models.Q", "line_number": 381, "usage_type": "call"}, {"api_name": "project.models.tech_lead", "line_number": 385, "usage_type": "attribute"}, {"api_name": "project.models", "line_number": 385, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 386, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 386, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 386, "usage_type": "call"}, {"api_name": "project.models.ProjectUserMembership.objects.filter", "line_number": 387, "usage_type": "call"}, {"api_name": "project.models.ProjectUserMembership.objects", "line_number": 387, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 387, "usage_type": "name"}, {"api_name": "project.models", "line_number": 388, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 390, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 390, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 390, "usage_type": "call"}, {"api_name": "project.models.Project.DoesNotExist", "line_number": 391, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 391, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 392, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 392, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 392, "usage_type": "call"}, {"api_name": "django.forms.Form", "line_number": 396, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 396, "usage_type": "name"}, {"api_name": "django.forms.CharField", "line_number": 397, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 397, "usage_type": "name"}, {"api_name": "project.models", "line_number": 403, "usage_type": "name"}, {"api_name": "project.models.Project.objects.filter", "line_number": 403, "usage_type": "call"}, {"api_name": "project.models.Project.objects", "line_number": 403, "usage_type": "attribute"}, {"api_name": "project.models.Project", "line_number": 403, "usage_type": "name"}, {"api_name": "users.models.CustomUser.objects.filter", "line_number": 404, "usage_type": "call"}, {"api_name": "users.models.CustomUser.objects", "line_number": 404, "usage_type": "attribute"}, {"api_name": "users.models.CustomUser", "line_number": 404, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 407, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 407, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 407, "usage_type": "call"}, {"api_name": "users.models.Profile.APPROVED", "line_number": 408, "usage_type": "attribute"}, {"api_name": "users.models.Profile", "line_number": 408, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 409, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 409, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 409, "usage_type": "call"}, {"api_name": "project.models.tech_lead", "line_number": 410, "usage_type": "attribute"}, {"api_name": "project.models", "line_number": 410, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 411, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 411, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 411, "usage_type": "call"}, {"api_name": "project.models.ProjectUserMembership.objects.filter", "line_number": 412, "usage_type": "call"}, {"api_name": "project.models.ProjectUserMembership.objects", "line_number": 412, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 412, "usage_type": "name"}, {"api_name": "project.models", "line_number": 413, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 415, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 415, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 415, "usage_type": "call"}, {"api_name": "project.models.can_have_more_users", "line_number": 416, "usage_type": "call"}, {"api_name": "project.models", "line_number": 416, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 417, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 417, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 417, "usage_type": "call"}, {"api_name": "django.forms.ModelForm", "line_number": 424, "usage_type": "attribute"}, {"api_name": "django.forms", "line_number": 424, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 427, "usage_type": "name"}, {"api_name": "django.forms.ChoiceField", "line_number": 440, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 440, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 447, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 447, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.AWAITING_AUTHORISATION", "line_number": 447, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 448, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 448, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.AUTHORISED", "line_number": 448, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 449, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 449, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.DECLINED", "line_number": 449, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 452, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 452, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.AUTHORISED", "line_number": 452, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 453, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 453, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.REVOKED", "line_number": 453, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 454, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 454, "usage_type": "name"}, {"api_name": "project.models.ProjectUserMembership.SUSPENDED", "line_number": 454, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership.STATUS_CHOICES", "line_number": 456, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 456, "usage_type": "name"}, {"api_name": "django.forms.ValidationError", "line_number": 468, "usage_type": "call"}, {"api_name": "django.forms", "line_number": 468, "usage_type": "name"}, {"api_name": "django.utils.translation.gettext_lazy", "line_number": 468, "usage_type": "call"}, {"api_name": "project.models.ProjectUserMembership.AUTHORISED", "line_number": 478, "usage_type": "attribute"}, {"api_name": "project.models.ProjectUserMembership", "line_number": 478, "usage_type": "name"}, {"api_name": "project.openldap.update_openldap_project_membership", "line_number": 482, "usage_type": "call"}]} +{"seq_id": "625803811", "text": "\"\"\"\nA collection of methods which support partitioning algorithms.\n\"\"\"\n\nimport logging\n\nfrom pacman.model.constraints.abstract_constraints\\\n .abstract_partitioner_constraint import AbstractPartitionerConstraint\nfrom spinn_machine.utilities.progress_bar import ProgressBar\n\nlogger = logging.getLogger(__name__)\n\n\ndef generate_sub_edges(subgraph, graph_to_subgraph_mapper, graph):\n \"\"\" Generate the sub edges for the subvertices in the graph\n\n :param subgraph: the partitioned graph to work with\n :type subgraph:\\\n :py:class:`pacman.model.partitioned_graph.partitioned_graph.PartitionedGraph`\n :param graph_to_subgraph_mapper: the mapper between the \\\n partitionable graph and the partitioned graph\n :type graph_to_subgraph_mapper:\\\n :py:class:`pacman.model.graph_mapper.GraphMapper`\n :param graph: the partitionable graph to work with\n :type graph:\\\n :py:class:`pacman.model.graph.partitionable_graph.PartitionableGraph`\n \"\"\"\n\n # start progress bar\n progress_bar = ProgressBar(len(subgraph.subvertices),\n \"Partitioning graph edges\")\n\n # Partition edges according to vertex partitioning\n for src_sv in subgraph.subvertices:\n\n # For each out edge of the parent vertex...\n vertex = graph_to_subgraph_mapper.get_vertex_from_subvertex(src_sv)\n outgoing_partitions = \\\n graph.outgoing_edges_partitions_from_vertex(vertex)\n for outgoing_partition_identifer in outgoing_partitions:\n partition = outgoing_partitions[outgoing_partition_identifer]\n out_edges = partition.edges\n partition_constraints = partition.constraints\n for edge in out_edges:\n\n # and create and store a new subedge for each post-subvertex\n post_vertex = edge.post_vertex\n post_subverts = (graph_to_subgraph_mapper\n .get_subvertices_from_vertex(post_vertex))\n for dst_sv in post_subverts:\n subedge = edge.create_subedge(src_sv, dst_sv)\n subgraph.add_subedge(subedge,\n outgoing_partition_identifer,\n partition_constraints)\n graph_to_subgraph_mapper.add_partitioned_edge(\n subedge, edge)\n progress_bar.update()\n progress_bar.end()\n\n\ndef get_remaining_constraints(vertex):\n \"\"\" Gets the rest of the constraints from a vertex after removing\\\n partitioning constraints\n \"\"\"\n constraints = list()\n for constraint in vertex.constraints:\n if not isinstance(constraint, AbstractPartitionerConstraint):\n constraints.append(constraint)\n return constraints\n", "sub_path": "src/spinnaker_ros_lsm/venv/lib/python2.7/site-packages/pacman/utilities/algorithm_utilities/partition_algorithm_utilities.py", "file_name": "partition_algorithm_utilities.py", "file_ext": "py", "file_size_in_byte": 2818, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "spinn_machine.utilities.progress_bar.ProgressBar", "line_number": 30, "usage_type": "call"}, {"api_name": "pacman.model.constraints.abstract_constraints.abstract_partitioner_constraint.AbstractPartitionerConstraint", "line_number": 67, "usage_type": "argument"}]} +{"seq_id": "240246241", "text": "import discord\nfrom discord.ext import commands\n\n\nclass Presence:\n def __init__(self, bot):\n self.bot = bot\n self.cmd = bot.get_command\n\n @commands.group(invoke_without_command=True, aliases=['pres'])\n async def presence(self, ctx, *, game_name: str=None):\n \"\"\" Change your discord presence\n Usage: provide a game name as a shortcut to presence game OR\n Provide no input and no subcommand to clear your playing status \"\"\"\n if game_name is not None:\n return await ctx.invoke(self.cmd('presence game'), game_name=game_name)\n\n return await ctx.invoke(self.cmd('presence clear'))\n\n @presence.command(aliases=['g', 'play', 'playing'])\n async def game(self, ctx, *, game_name: str):\n \"\"\" Set your \"playing\" status\n\n This is redundant. You can shortcut by calling presence \"\"\"\n await self.bot.change_presence(game=discord.Game(name=game_name))\n await ctx.message.edit(content=\"Changed presence!\")\n\n @presence.group(aliases=['stream', 'streaming'])\n async def twitch(self, ctx, game_name: str, twitch_channel_name: str):\n \"\"\" Set your status to streaming with a twitch channel name\n NOTE: Your game name must be double-quoted if it is longer than 1 word with spaces \"\"\"\n await self.bot.change_presence(game=discord.Game(name=game_name,\n type=1,\n url=f'https://www.twitch.tv/{twitch_channel_name}'))\n await ctx.message.edit(content=\"Changed to streaming!\")\n\n @presence.command(aliases=['cls', 'clean', 'remove', 'rmv'])\n async def clear(self, ctx):\n \"\"\" Reset your status, game, and streaming URL \"\"\"\n await self.bot.change_presence(game=discord.Game(name='', type=0, url=''))\n await ctx.message.edit(content=\"Presence cleared!\")\n\n\ndef setup(bot):\n bot.add_cog(Presence(bot))\n", "sub_path": "cogs/presence.py", "file_name": "presence.py", "file_ext": "py", "file_size_in_byte": 1971, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "discord.ext.commands.group", "line_number": 10, "usage_type": "call"}, {"api_name": "discord.ext.commands", "line_number": 10, "usage_type": "name"}, {"api_name": "discord.Game", "line_number": 25, "usage_type": "call"}, {"api_name": "discord.Game", "line_number": 32, "usage_type": "call"}, {"api_name": "discord.Game", "line_number": 40, "usage_type": "call"}]} +{"seq_id": "137877850", "text": "# ======================================================================\n# There are 5 questions in this test with increasing difficulty from 1-5\n# Please note that the weight of the grade for the question is relative\n# to its difficulty. So your Category 1 question will score much less\n# than your Category 5 question.\n# ======================================================================\n#\n# Computer Vision with CNNs\n#\n# For this exercise you will build a cats v dogs classifier\n# using the Cats v Dogs dataset from TFDS.\n# Be sure to use the final layer as shown\n# (Dense, 2 neurons, softmax activation)\n#\n# The testing infrastructre will resize all images to 224x224\n# with 3 bytes of color depth. Make sure your input layer trains\n# images to that specification, or the tests will fail.\n#\n# Make sure your output layer is exactly as specified here, or the\n# tests will fail.\n\n# =========== 합격 기준 가이드라인 공유 ============= #\n# val_loss 기준에 맞춰 주시는 것이 훨씬 더 중요 #\n# val_loss 보다 조금 높아도 상관없음. (언저리까지 OK) #\n# =================================================== #\n# 문제명: Category 3 - cats vs dogs\n# val_loss: 0.3158\n# val_acc: 0.8665\n# =================================================== #\n# =================================================== #\n\n\nimport tensorflow_datasets as tfds\nimport tensorflow as tf\n\nfrom tensorflow.keras.layers import Dense, Conv2D, MaxPooling2D, Flatten, Dropout\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras.callbacks import ModelCheckpoint\nfrom tensorflow.keras.applications import VGG16\n\ndataset_name = 'cats_vs_dogs'\n\n# 기존 코드 대신 사용\ntrain_dataset = tfds.load(name=dataset_name, split='train[:80%]')\nvalid_dataset = tfds.load(name=dataset_name, split='train[80%:]')\n\n\ndef preprocess(data):\n # YOUR CODE HERE\n x = data['image']\n y = data['label']\n x = x / 255\n x = tf.image.resize(x, size=(224, 224))\n\n return x, y\n\n\ndef solution_model():\n batch_size = 32\n train_data = train_dataset.map(preprocess).batch(batch_size)\n valid_data = valid_dataset.map(preprocess).batch(batch_size)\n\n transfer_model = VGG16(weights='imagenet', include_top=False, input_shape=(224, 224, 3))\n transfer_model.trainable = False\n\n model = Sequential([\n transfer_model,\n Flatten(),\n Dropout(0.5),\n Dense(512, activation='relu'),\n Dense(128, activation='relu'),\n\n # YOUR CODE HERE, BUT MAKE SURE YOUR LAST LAYER HAS 2 NEURONS ACTIVATED BY SOFTMAX\n tf.keras.layers.Dense(2, activation='softmax')\n ])\n\n model.summary()\n\n model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['acc'])\n\n checkpoint_path = \"my_checkpoint.ckpt\"\n checkpoint = ModelCheckpoint(filepath=checkpoint_path,\n save_weights_only=True,\n save_best_only=True,\n monitor='val_loss',\n verbose=1)\n\n model.fit(train_data,\n validation_data=(valid_data),\n epochs=20,\n callbacks=[checkpoint],\n )\n\n model.load_weights(checkpoint_path)\n\n return model\n\n\n# Note that you'll need to save your model as a .h5 like this\n# This .h5 will be uploaded to the testing infrastructure\n# and a score will be returned to you\nif __name__ == '__main__':\n model = solution_model()\n model.save(\"TF3-cats-vs-dogs.h5\")\n", "sub_path": "TF3/cats-vs-dogs.py", "file_name": "cats-vs-dogs.py", "file_ext": "py", "file_size_in_byte": 3509, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "tensorflow_datasets.load", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow_datasets.load", "line_number": 45, "usage_type": "call"}, {"api_name": "tensorflow.image.resize", "line_number": 53, "usage_type": "call"}, {"api_name": "tensorflow.image", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.applications.VGG16", "line_number": 63, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 66, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Flatten", "line_number": 68, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dropout", "line_number": 69, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 70, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 71, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 74, "usage_type": "call"}, {"api_name": "tensorflow.keras", "line_number": 74, "usage_type": "attribute"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 82, "usage_type": "call"}]} +{"seq_id": "623636066", "text": "from mpl_toolkits.mplot3d import Axes3D\r\nfrom sklearn.preprocessing import StandardScaler\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport os\r\nimport pandas as pd\r\nimport seaborn as sns\r\nimport csv\r\nfrom matplotlib.lines import Line2D\r\n\r\nDF2=pd.read_csv('C:/PythonProjects/AgentsTurboFan/Test/Test2/Scores.csv', delimiter=\",\", names=[\"Nodes\", \"Correlation\", \"RAE\", \"RSE\"])\r\nresult2 = DF2.sort_values(by=['Correlation'])\r\nresult2.to_csv('C:/PythonProjects/AgentsTurboFan/Test/Test2/ScoresSorted.csv', index=False)\r\nDF2S=pd.read_csv('C:/PythonProjects/AgentsTurboFan/Test/Test2/ScoresSorted.csv', delimiter=\",\")\r\nLN = []\r\nfor i in range(len(DF2S)):\r\n LN.append(DF2S['Nodes'][i])\r\nLC = []\r\nfor i in range(len(DF2S)):\r\n LC.append(DF2S['Correlation'][i])\r\nLRAE = []\r\nfor i in range(len(DF2S)):\r\n LRAE.append(DF2S['RAE'][i])\r\nLRSE = []\r\nfor i in range(len(DF2S)):\r\n LRSE.append(DF2S['RSE'][i])\r\nplt.rcParams['figure.figsize'] = [20,20]\r\nfig, ax1 = plt.subplots()\r\nax2 = ax1.twinx()\r\nx = np.arange(len(LN))\r\nax1.set_xticks(x)\r\nax1.set_xticklabels(LN, rotation=90)\r\nnodes = [x]\r\ncorrelation = [LC]\r\nrelativeabsoluteerror = [LRAE]\r\nrootrelativesquarederror = [LRSE]\r\nax1.scatter(nodes, correlation, color='green')\r\nax1.plot(nodes, correlation, color='green', label='Correlation', marker='o', markerfacecolor='green', markersize=12, linewidth=4)\r\nax2.scatter(nodes, relativeabsoluteerror, color='red')\r\nax2.plot(nodes, relativeabsoluteerror, color='red', label='Relative Absolute Error', marker='o', markerfacecolor='red', markersize=12, linewidth=4)\r\nax2.scatter(nodes, rootrelativesquarederror, color='yellow')\r\nax2.plot(nodes, rootrelativesquarederror, color='yellow', label='Root Relative Squared Error', marker='o', markerfacecolor='yellow', markersize=12, linewidth=4)\r\ncustom_lines = [Line2D([0], [0], marker='o', color='w', label='Correlation', markerfacecolor='g', markersize=15),\r\n Line2D([0], [0], marker='o', color='w', label='Relative Absolute Error', markerfacecolor='r', markersize=15),\r\n Line2D([0], [0], marker='o', color='w', label='Root Relative Squared Error', markerfacecolor='yellow', markersize=15)]\r\nax1.legend(custom_lines, ['Correlation', 'Relative Absolute Error', 'Root Relative Squared Error'], scatterpoints=1)\r\nax1.set_xlabel(\"Nodes\")\r\nax1.set_ylabel(\"Correlation\")\r\nax2.set_ylabel(\"Relative Absolute Error and Root Relative Squared Error\")\r\nax1.set_ylim(0, 1)\r\nax2.set_ylim(0, 120)\r\n\r\nplt.title('Scatter Plot of the second data set')\r\nplt.rc('axes', titlesize=30) # fontsize of the axes title\r\nplt.rc('axes', labelsize=30) # fontsize of the x and y labels\r\nplt.rc('xtick', labelsize=30) # fontsize of the tick labels\r\nplt.rc('ytick', labelsize=30) # fontsize of the tick labels\r\nplt.rc('legend', fontsize=30) # legend fontsize\r\nplt.rc('figure', titlesize=30)\r\n\r\nplt.show()\r\nfig.savefig(\"C:/PythonProjects/AgentsTurboFan/Test/Test2/Plot.pdf\", bbox_inches='tight')", "sub_path": "PlotOfSecondDataset.py", "file_name": "PlotOfSecondDataset.py", "file_ext": "py", "file_size_in_byte": 2955, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.read_csv", "line_number": 11, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 14, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.rcParams", "line_number": 27, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.lines.Line2D", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 53, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 53, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 54, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 55, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 55, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.rc", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}]} +{"seq_id": "34914307", "text": "\"\"\"\nDate sequence generation functions to be used by statistics apps\n\n\n\"\"\"\nfrom __future__ import absolute_import\n\nfrom dateutil.rrule import YEARLY, MONTHLY, rrule\nfrom dateutil.relativedelta import relativedelta\nfrom dateutil.parser import parse\n\nFREQS = {'y': YEARLY, 'm': MONTHLY}\nDURATIONS = {'y': 'years', 'm': 'months'}\n\n\ndef date_sequence(start, end, stats_duration, step_size):\n \"\"\"\n Generate a sequence of time span tuples\n\n :seealso:\n Refer to `dateutil.parser.parse` for details on date parsing.\n\n :param str start: Start date of first interval\n :param str end: End date. The end of the last time span may extend past this date.\n :param str stats_duration: What period of time shouuld be grouped\n :param str step_size: How far apart should the start dates be\n :return: sequence of (start_date, end_date) tuples\n \"\"\"\n step_size, freq = parse_interval(step_size)\n stats_duration = parse_duration(stats_duration)\n for start_date in rrule(freq, interval=step_size, dtstart=start, until=end):\n end_date = start_date + stats_duration\n if end_date <= end:\n yield start_date, start_date + stats_duration\n\n\ndef parse_interval(interval):\n count, units = split_duration(interval)\n try:\n return count, FREQS[units]\n except KeyError:\n raise ValueError('Invalid interval \"{}\", units not in of: {}'.format(interval, FREQS.keys))\n\n\ndef parse_duration(duration):\n count, units = split_duration(duration)\n try:\n delta = {DURATIONS[units]: count}\n except KeyError:\n raise ValueError('Duration \"{}\" not in months or years'.format(duration))\n\n return relativedelta(**delta)\n\n\ndef split_duration(duration):\n return int(duration[:-1]), duration[-1:]\n", "sub_path": "datacube/utils/dates.py", "file_name": "dates.py", "file_ext": "py", "file_size_in_byte": 1759, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "dateutil.rrule.YEARLY", "line_number": 12, "usage_type": "name"}, {"api_name": "dateutil.rrule.MONTHLY", "line_number": 12, "usage_type": "name"}, {"api_name": "dateutil.rrule.rrule", "line_number": 31, "usage_type": "call"}, {"api_name": "dateutil.relativedelta.relativedelta", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "317526180", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\nimport django.core.files.storage\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('pycupid', '0002_num_matches_included'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='people',\n name='gender',\n field=models.CharField(help_text='Gender identification', null=True, choices=[('m', 'Male'), ('f', 'Female'), ('o', 'Other / N/A')], max_length=1, default='o'),\n ),\n migrations.AlterField(\n model_name='people',\n name='preference',\n field=models.CharField(help_text='Gender preference', null=True, choices=[('m', 'Men'), ('f', 'Women'), ('a', 'All')], max_length=1, default='a'),\n ),\n migrations.AlterField(\n model_name='people',\n name='ukey',\n field=models.UUIDField(help_text='The unique id we use to differentiate users in links in the welcome email.', editable=False, null=True, default='0b45b27e13714004977f729f42d20ddb'),\n ),\n migrations.AlterField(\n model_name='people',\n name='usr_img',\n field=models.ImageField(help_text='Profile image.', upload_to='static/images/user/', null=True, storage=django.core.files.storage.FileSystemStorage(location='static/images/user'), default='static/images/default/cow_sm.jpg'),\n ),\n ]\n", "sub_path": "apps/pycupid/migrations/0003_auto_20160516_1035.py", "file_name": "0003_auto_20160516_1035.py", "file_ext": "py", "file_size_in_byte": 1459, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 8, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 8, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 15, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 15, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 20, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 20, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 23, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 23, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 25, "usage_type": "name"}, {"api_name": "django.db.models.UUIDField", "line_number": 28, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 28, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 30, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 30, "usage_type": "name"}, {"api_name": "django.db.models.ImageField", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 33, "usage_type": "name"}, {"api_name": "django.db.core.files.storage.FileSystemStorage", "line_number": 33, "usage_type": "call"}, {"api_name": "django.db.core", "line_number": 33, "usage_type": "attribute"}, {"api_name": "django.db", "line_number": 33, "usage_type": "name"}]} +{"seq_id": "80100343", "text": "# -*- coding: utf-8 -*-\n\nfrom scrapy.spiders import CrawlSpider, Rule\nfrom scrapy.linkextractors import LinkExtractor\nfrom scrapy.selector import Selector\nfrom scrapy.http.request import Request\nfrom crawler.items import GalleryCrawlerItem\n\n\nclass RedditSpider(CrawlSpider):\n name = 'reddit'\n allowed_domains = ['www.reddit.com']\n start_urls = ['http://www.reddit.com/r/pics/']\n\n rules = [\n Rule(\n LinkExtractor(allow=['/r/pics/\\?count=\\d*&after=\\w*']),\n callback='parse_item',\n follow=True\n ),\n ]\n def parse_item(self, response):\n images = response.xpath('//div[contains(@class, \"thing\")]')\n for image in images:\n item = GalleryCrawlerItem()\n item['url'] = image.xpath('@data-url')[0].extract()\n item['source'] = response.url\n item['source_pub']=image.xpath('div[contains(@class, \"entry\")]/div[contains(@class, \"top-matter\")]/p[contains(@class, \"tagline\")]/time/@datetime').extract(),\n yield item\n", "sub_path": "crawler/crawler/spiders/reddit.py", "file_name": "reddit.py", "file_ext": "py", "file_size_in_byte": 1032, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "scrapy.spiders.CrawlSpider", "line_number": 10, "usage_type": "name"}, {"api_name": "scrapy.spiders.Rule", "line_number": 16, "usage_type": "call"}, {"api_name": "scrapy.linkextractors.LinkExtractor", "line_number": 17, "usage_type": "call"}, {"api_name": "crawler.items.GalleryCrawlerItem", "line_number": 25, "usage_type": "call"}]} +{"seq_id": "261294010", "text": "# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('glad_plugins', '0011_auto_20170212_1029'),\n ]\n\n operations = [\n migrations.RemoveField(\n model_name='gladchartdataplugin',\n name='label',\n ),\n migrations.RemoveField(\n model_name='gladchartdatasetplugin',\n name='data',\n ),\n migrations.AlterField(\n model_name='gladchartdataplugin',\n name='color',\n field=models.CharField(verbose_name='цвет', default='parent', choices=[('255,0,0', 'красный'), ('0,150,0', 'зелёный'), ('0,0,150', 'синий'), ('220,220,0', 'жёлтый'), ('0,150,150', 'голубой'), ('150,0,150', 'пурпурный'), ('parent', 'как у родителя')], max_length=255),\n ),\n ]\n", "sub_path": "glad_plugins/migrations/0012_auto_20170212_1423.py", "file_name": "0012_auto_20170212_1423.py", "file_ext": "py", "file_size_in_byte": 937, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "django.db.migrations.Migration", "line_number": 7, "usage_type": "attribute"}, {"api_name": "django.db.migrations", "line_number": 7, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 14, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 14, "usage_type": "name"}, {"api_name": "django.db.migrations.RemoveField", "line_number": 18, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 18, "usage_type": "name"}, {"api_name": "django.db.migrations.AlterField", "line_number": 22, "usage_type": "call"}, {"api_name": "django.db.migrations", "line_number": 22, "usage_type": "name"}, {"api_name": "django.db.models.CharField", "line_number": 25, "usage_type": "call"}, {"api_name": "django.db.models", "line_number": 25, "usage_type": "name"}]} +{"seq_id": "100653129", "text": "import setuptools\nimport pathlib\n\nHERE = pathlib.Path(__file__).parent\n\nREADME = (HERE / 'README.md').read_text()\n\nsetuptools.setup(\n name='bifrost_py',\n version='0.0.5',\n author=\"Ignacio Althabe\",\n author_email=\"nacho.althabe@gmail.com \",\n description=\"Wallets based on SALT instead of Nonce.\",\n long_description=README,\n long_description_content_type=\"text/markdown\",\n url=\"https://gitlab.riaquest.com/nacho/bifrost\",\n packages=['bifrost_py'],\n package_data = {\n 'bifrost_py': ['**/*']\n },\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n ]\n)\n", "sub_path": "pypi_install_script/bifrost_py-0.0.5.tar/setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 701, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pathlib.Path", "line_number": 4, "usage_type": "call"}, {"api_name": "setuptools.setup", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "296369402", "text": "import pyautogui as pg\nimport os\nimport csv\nimport time\nimport sys\n\n\ndef tab():\n pg.press(['tab'], interval=0.10)\n\n\ndef MCQ(option):\n try:\n option = int(option)\n except ValueError:\n pass\n\n if option == 1 :\n pg.press(['space'])\n\n elif option > 1 :\n press_lst = ['right']\n press_lst *= option\n pg.press(press_lst)\n\n\ndef kill(sec):\n for rem in range(sec, 0, -1) :\n st = f'{rem} seconds remaining'\n print(st, end='')\n time.sleep(1)\n print('\\b'*len(st), end='', flush=True)\n\n\n# link address\nlnk_addrs = r'https://docs.google.com/forms/d/e/1FAIpQLScSVDFU76rZvbO_tiIwSt6d9sOK0CZyS9KKMCP6cP5O5W5lVQ/viewform'\n\n# # directory of CSV\n# os.chdir(r'.\\form project')\n# print('Dir changed')\n\n\nfile_in = open('form data.txt', 'r') # open CSV\nprint('file opened')\n\nprint('Make sure the form is in view')\n\nkill(5)\n\ndata = csv.reader(file_in) # getting CSV object\n\n# constants\nitr_no = 0 # for debug\nbrk = False # master break\nsintrvl = 0.04 # samll interval\nlintrvl = 0.05 # large interval\n\nwhile True: # true until file ends\n\n if brk: # master break\n break\n\n file_in.seek(0) # seek zero if run again is engaged\n\n try:\n for fill in data: # getting the data list from CSV\n\n if not fill: # if empty activate master break\n brk = True\n break\n\n # name\n name_pos = pg.locateOnScreen('1_name.PNG') # finding initial field\n pg.click(name_pos[0] + 10 , name_pos[1]+(name_pos[3] - 30)) # clicking initial field\n pg.typewrite(fill[0], interval=lintrvl) # filling initial field using csv[0]\n\n # greatest fears\n tab()\n pg.typewrite(fill[1], interval = lintrvl) #writing field\n\n # wizard power\n tab()\n pg.press(fill[2][0:1]) # pressing the first letter of the choice due to unique fields\n\n # robocop\n tab()\n MCQ(fill[3])\n\n # add comments\n tab()\n if fill[3] != '0' :\n tab() #to jump clear selection\n pg.typewrite(fill[4], interval = sintrvl)\n\n # submit\n tab()\n pg.press(['enter'])\n\n itr_no += 1 #iteration number inc.\n\n # reopen the form\n adrs_pos = pg.locateOnScreen('2_lock.PNG')\n pg.click(adrs_pos[0] + 60, adrs_pos[1] + 20 )\n pg.typewrite(lnk_addrs)\n pg.press(['enter'])\n\n kill(3)\n\n\n except Exception as error_name :\n\n exc_type, exc_obj, tb = sys.exc_info()\n print(f'Error {error_name} on line {tb.tb_lineno} of itr {itr_no}')\n\n choose = input('Run again(r), Quit(q), Stay(s) --> ')\n\n if choose == 'r' :\n continue\n elif choose == 'q' :\n break\n elif choose == 's' :\n while True : #to see input infinite loop\n pass\n else :\n while True : #to see input infinite loop\n pass\n", "sub_path": "form project/form_proj.py", "file_name": "form_proj.py", "file_ext": "py", "file_size_in_byte": 3044, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pyautogui.press", "line_number": 9, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 19, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 24, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 31, "usage_type": "call"}, {"api_name": "csv.reader", "line_number": 50, "usage_type": "call"}, {"api_name": "pyautogui.locateOnScreen", "line_number": 73, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 74, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 75, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 79, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 83, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 93, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 97, "usage_type": "call"}, {"api_name": "pyautogui.locateOnScreen", "line_number": 102, "usage_type": "call"}, {"api_name": "pyautogui.click", "line_number": 103, "usage_type": "call"}, {"api_name": "pyautogui.typewrite", "line_number": 104, "usage_type": "call"}, {"api_name": "pyautogui.press", "line_number": 105, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 112, "usage_type": "call"}]} +{"seq_id": "538124717", "text": "# -*- coding:utf-8 -*- \nimport sys\n\nreload(sys)\nsys.setdefaultencoding('utf8')\n\nfrom flask import render_template, redirect, url_for, jsonify, flash, request, current_app\nfrom flask_login import login_user, logout_user, login_required, current_user\nfrom datetime import datetime, timedelta\nfrom . import boss\nfrom .. import db\nfrom .forms import StatisticsWorkplanForm, ReviewAMWorkPlanForm, NotificationForm\nfrom ..models import WorkPlan, User, Notification, ReadNotification\nfrom ..decorators import CJsonEncoder, boss_required\nimport json\n\n#统计信息,单人时间段内三项数值,全体时间段内单项数值\n#时间段为了显示方便,不要弄太长,主要是保证列数不要太多就好\n\n@boss.route('/statistics_workplan', methods=[\"GET\", \"POST\"])\n@boss_required\ndef statistics_workplan():\n form = StatisticsWorkplanForm()\n return render_template('boss/statistics_workplan.html', form=form)\n\n#全体,一段时间,三项数值和\n@boss.route('/_statistics_allam_workplan_sum')\n@boss_required\ndef statistics_allam_workplan_sum():\n # am_id = request.args.get('am_id',0,type=int)\n startdate = request.args.get('start_date')\n enddate = request.args.get('end_date')\n\n searchdata=db.session.query(WorkPlan.am_id,\n db.func.sum(WorkPlan.client_contact),\n db.func.sum(WorkPlan.capital_increment),\n db.func.sum(WorkPlan.volume)\n ).filter(db.and_(WorkPlan.tommorrowdate.between(startdate,enddate),\n WorkPlan.flag==0)).group_by(WorkPlan.am_id).order_by(WorkPlan.am_id).all()\n am_id_name=db.session.query(User.name).filter(User.role_id==1).order_by(User.id).all()\n\n labels = []\n for am_name in am_id_name:\n labels.append(am_name[0])\n\n client_contact = []\n capital_increment = []\n volume =[]\n for data in searchdata:\n client_contact.append(data[1])\n capital_increment.append(data[2])\n volume.append(data[3])\n\n dict_client_contact={'fillColor' : \"rgba(220,220,220,0.5)\",\n 'strokeColor' : \"rgba(220,220,220,0.5)\",\n 'pointColor' : \"rgba(220,220,220,0.5)\",\n 'pointStrokeColor' : \"#fff\",\n 'data' : client_contact}\n dict_capital_increment={'fillColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'strokeColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'pointColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'pointStrokeColor' : \"#fff\",\n\t\t\t'data' : capital_increment}\n dict_volume={'fillColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'strokeColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'pointColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'pointStrokeColor' : \"#fff\",\n\t\t\t'data' : volume}\n client_contact = [dict_client_contact]\n capital_increment = [dict_capital_increment]\n volume = [dict_volume]\n\n return jsonify(labels=labels, client_contact=client_contact, capital_increment=capital_increment, volume=volume)\n\n#单人,指定日期,三项数值\n@boss.route('/review_am_workplan', methods=[\"GET\", \"POST\"])\n@boss_required\ndef review_am_workplan():\n form = ReviewAMWorkPlanForm()\n return render_template('boss/review_am_workplan.html', form=form)\n\n@boss.route('/_review_single_am_workplan')\n@boss_required\ndef review_single_am_workplan():\n amid = request.args.get('am_id',0,type=int)\n startdate = request.args.get('start_date')\n enddate = request.args.get('end_date')\n if startdate > enddate:\n return jsonify()\n\n searchdata=db.session.query(WorkPlan.client_contact,\n WorkPlan.capital_increment,\n WorkPlan.volume,\n WorkPlan.todaydate).filter(db.and_(WorkPlan.tommorrowdate.between(startdate,enddate),\n WorkPlan.flag==0,\n WorkPlan.am_id==amid)\n ).order_by(WorkPlan.todaydate).all()\n\n labels = []\n client_contact = []\n capital_increment = []\n volume =[]\n for data in searchdata:\n labels.append(json.dumps(data[3],cls=CJsonEncoder))\n client_contact.append(data[0])\n capital_increment.append(data[1])\n volume.append(data[2])\n\n dict_client_contact={'fillColor' : \"rgba(220,220,220,0.5)\",\n 'strokeColor' : \"rgba(220,220,220,0.5)\",\n 'pointColor' : \"rgba(220,220,220,0.5)\",\n 'pointStrokeColor' : \"#fff\",\n 'data' : client_contact}\n dict_capital_increment={'fillColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'strokeColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'pointColor' : \"rgba(151,187,205,0.5)\",\n\t\t\t'pointStrokeColor' : \"#fff\",\n\t\t\t'data' : capital_increment}\n dict_volume={'fillColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'strokeColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'pointColor' : \"rgba(101,117,205,0.5)\",\n\t\t\t'pointStrokeColor' : \"#fff\",\n\t\t\t'data' : volume}\n client_contact = [dict_client_contact]\n capital_increment = [dict_capital_increment]\n volume = [dict_volume]\n\n return jsonify(labels=labels, client_contact=client_contact, capital_increment=capital_increment, volume=volume)\n\n@boss.route('/notification', methods=[\"GET\", \"POST\"])\n@boss_required\ndef notification():\n form = NotificationForm()\n publish_datetime = datetime.now()\n target_role_id = form.target.data\n if form.validate_on_submit():\n newNotification=Notification(\n publish_id = current_user.id,\n target_role_id = target_role_id,\n title = form.title.data,\n body = form.title.data,\n publish_datetime = publish_datetime\n )\n db.session.add(newNotification)\n db.session.commit()\n\n notification_id=db.session.query(Notification.id).filter(\n Notification.publish_datetime==publish_datetime).first()[0]\n users=db.session.query(User.id).filter(User.flag==1).all() #默认发给所有人\n if target_role_id!=99:\n users=db.session.query(User.id).filter(\n db.and_(User.role_id==target_role_id,\n User.flag==1)).all()\n for reader in users:\n newReadNotification=ReadNotification(\n reader_id = reader.id,\n notification_id = notification_id\n )\n db.session.add(newReadNotification)\n db.session.commit()\n flash('通知已发布')\n return redirect(url_for('boss.notification'))\n return render_template('boss/notification.html', form=form)\n", "sub_path": "app/boss/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 6637, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.setdefaultencoding", "line_number": 5, "usage_type": "call"}, {"api_name": "forms.StatisticsWorkplanForm", "line_number": 23, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 24, "usage_type": "call"}, {"api_name": "decorators.boss_required", "line_number": 21, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 31, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 31, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 31, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 32, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 32, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 32, "usage_type": "name"}, {"api_name": "models.WorkPlan.am_id", "line_number": 34, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 34, "usage_type": "name"}, {"api_name": "models.WorkPlan.client_contact", "line_number": 35, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 35, "usage_type": "name"}, {"api_name": "models.WorkPlan.capital_increment", "line_number": 36, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 36, "usage_type": "name"}, {"api_name": "models.WorkPlan.volume", "line_number": 37, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 37, "usage_type": "name"}, {"api_name": "models.WorkPlan.tommorrowdate.between", "line_number": 38, "usage_type": "call"}, {"api_name": "models.WorkPlan.tommorrowdate", "line_number": 38, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 38, "usage_type": "name"}, {"api_name": "models.WorkPlan.flag", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 39, "usage_type": "name"}, {"api_name": "models.WorkPlan.am_id", "line_number": 39, "usage_type": "attribute"}, {"api_name": "models.User.name", "line_number": 40, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 40, "usage_type": "name"}, {"api_name": "models.User.role_id", "line_number": 40, "usage_type": "attribute"}, {"api_name": "models.User.id", "line_number": 40, "usage_type": "attribute"}, {"api_name": "flask.jsonify", "line_number": 73, "usage_type": "call"}, {"api_name": "decorators.boss_required", "line_number": 28, "usage_type": "name"}, {"api_name": "forms.ReviewAMWorkPlanForm", "line_number": 79, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 80, "usage_type": "call"}, {"api_name": "decorators.boss_required", "line_number": 77, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 85, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 85, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 85, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 86, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 86, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 86, "usage_type": "name"}, {"api_name": "flask.request.args.get", "line_number": 87, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 87, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 87, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 89, "usage_type": "call"}, {"api_name": "models.WorkPlan.client_contact", "line_number": 91, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 91, "usage_type": "name"}, {"api_name": "models.WorkPlan.capital_increment", "line_number": 92, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 92, "usage_type": "name"}, {"api_name": "models.WorkPlan.volume", "line_number": 93, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 93, "usage_type": "name"}, {"api_name": "models.WorkPlan.todaydate", "line_number": 94, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 94, "usage_type": "name"}, {"api_name": "models.WorkPlan.tommorrowdate.between", "line_number": 94, "usage_type": "call"}, {"api_name": "models.WorkPlan.tommorrowdate", "line_number": 94, "usage_type": "attribute"}, {"api_name": "models.WorkPlan.flag", "line_number": 95, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 95, "usage_type": "name"}, {"api_name": "models.WorkPlan.am_id", "line_number": 96, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 96, "usage_type": "name"}, {"api_name": "models.WorkPlan.todaydate", "line_number": 97, "usage_type": "attribute"}, {"api_name": "models.WorkPlan", "line_number": 97, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 104, "usage_type": "call"}, {"api_name": "decorators.CJsonEncoder", "line_number": 104, "usage_type": "name"}, {"api_name": "flask.jsonify", "line_number": 128, "usage_type": "call"}, {"api_name": "decorators.boss_required", "line_number": 83, "usage_type": "name"}, {"api_name": "forms.NotificationForm", "line_number": 133, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 134, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 134, "usage_type": "name"}, {"api_name": "models.Notification", "line_number": 137, "usage_type": "call"}, {"api_name": "flask_login.current_user.id", "line_number": 138, "usage_type": "attribute"}, {"api_name": "flask_login.current_user", "line_number": 138, "usage_type": "name"}, {"api_name": "models.Notification.id", "line_number": 147, "usage_type": "attribute"}, {"api_name": "models.Notification", "line_number": 147, "usage_type": "name"}, {"api_name": "models.Notification.publish_datetime", "line_number": 148, "usage_type": "attribute"}, {"api_name": "models.Notification", "line_number": 148, "usage_type": "name"}, {"api_name": "models.User.id", "line_number": 149, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 149, "usage_type": "name"}, {"api_name": "models.User.flag", "line_number": 149, "usage_type": "attribute"}, {"api_name": "models.User.id", "line_number": 151, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 151, "usage_type": "name"}, {"api_name": "models.User.role_id", "line_number": 152, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 152, "usage_type": "name"}, {"api_name": "models.User.flag", "line_number": 153, "usage_type": "attribute"}, {"api_name": "models.User", "line_number": 153, "usage_type": "name"}, {"api_name": "models.ReadNotification", "line_number": 155, "usage_type": "call"}, {"api_name": "flask.flash", "line_number": 161, "usage_type": "call"}, {"api_name": "flask.redirect", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.url_for", "line_number": 162, "usage_type": "call"}, {"api_name": "flask.render_template", "line_number": 163, "usage_type": "call"}, {"api_name": "decorators.boss_required", "line_number": 131, "usage_type": "name"}]} +{"seq_id": "451722595", "text": "import pytest\nfrom rest_framework.views import APIView\n\nimport winter\nfrom winter.http import MediaType\nfrom winter.routing import get_route\nfrom winter.schema.inspectors import SwaggerAutoSchema\n\n\n@pytest.fixture\ndef auto_schema():\n view = View()\n View.post.method = Controller.post\n route = get_route(Controller.post)\n try:\n yield SwaggerAutoSchema(view, 'path', route.http_method, 'components', 'request', {})\n finally:\n del View.post.method\n\n\ndef get_empty_swagger_auto_schema(method: str = 'post'):\n view = View()\n\n return SwaggerAutoSchema(view, 'path', method, 'components', 'request', {})\n\n\nclass View(APIView):\n\n def post(self):\n return\n\n\nclass Controller:\n\n @winter.route_post('/', produces=(MediaType.MULTIPART_FORM_DATA,), consumes=(MediaType.APPLICATION_JSON,))\n def post(self):\n pass\n\n\ndef test_get_produces(auto_schema):\n consumes = auto_schema.get_produces()\n assert consumes == [str(MediaType.MULTIPART_FORM_DATA)]\n\n\ndef test_get_consumes(auto_schema):\n consumes = auto_schema.get_consumes()\n assert consumes == [str(MediaType.APPLICATION_JSON)]\n\n\ndef test_get_produces_without_method():\n auto_schema = get_empty_swagger_auto_schema()\n consumes = auto_schema.get_produces()\n assert consumes == ['application/json']\n\n\n@pytest.mark.parametrize('method', ('post', 'patch'))\ndef test_get_consumes_without_method(method):\n auto_schema = get_empty_swagger_auto_schema(method)\n consumes = auto_schema.get_consumes()\n assert consumes == ['application/json']\n", "sub_path": "tests/test_swagger_auto_schema.py", "file_name": "test_swagger_auto_schema.py", "file_ext": "py", "file_size_in_byte": 1555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "winter.routing.get_route", "line_number": 14, "usage_type": "call"}, {"api_name": "winter.schema.inspectors.SwaggerAutoSchema", "line_number": 16, "usage_type": "call"}, {"api_name": "pytest.fixture", "line_number": 10, "usage_type": "attribute"}, {"api_name": "winter.schema.inspectors.SwaggerAutoSchema", "line_number": 24, "usage_type": "call"}, {"api_name": "rest_framework.views.APIView", "line_number": 27, "usage_type": "name"}, {"api_name": "winter.route_post", "line_number": 35, "usage_type": "call"}, {"api_name": "winter.http.MediaType.MULTIPART_FORM_DATA", "line_number": 35, "usage_type": "attribute"}, {"api_name": "winter.http.MediaType", "line_number": 35, "usage_type": "name"}, {"api_name": "winter.http.MediaType.APPLICATION_JSON", "line_number": 35, "usage_type": "attribute"}, {"api_name": "winter.http.MediaType.MULTIPART_FORM_DATA", "line_number": 42, "usage_type": "attribute"}, {"api_name": "winter.http.MediaType", "line_number": 42, "usage_type": "name"}, {"api_name": "winter.http.MediaType.APPLICATION_JSON", "line_number": 47, "usage_type": "attribute"}, {"api_name": "winter.http.MediaType", "line_number": 47, "usage_type": "name"}, {"api_name": "pytest.mark.parametrize", "line_number": 56, "usage_type": "call"}, {"api_name": "pytest.mark", "line_number": 56, "usage_type": "attribute"}]} +{"seq_id": "416694863", "text": "import requests\nfrom bs4 import BeautifulSoup\nimport spotipy\nfrom spotipy.oauth2 import SpotifyOAuth\nfrom spotipy.oauth2 import SpotifyClientCredentials\n\n\nSPOTIFY_USER_ID = 'to_replace'\n\nclass Billboard:\n\n def __init__(self, year):\n self.url = f'https://www.billboard.com/charts/hot-100/{year}'\n\n def fetch_tracks_title(self):\n tracks_title = []\n response = requests.get(self.url)\n soup = BeautifulSoup(response.text, 'html.parser')\n table_rows = soup.find_all(class_='chart-element__information__song') \n for track in table_rows:\n track_title = track.get_text()\n tracks_title.append(track_title)\n return tracks_title\n\nclass Anghami:\n\n def __init__(self, url, cookies):\n self.url = url\n self.cookies = cookies\n \n def fetch_tracks_title(self):\n tracks_title = []\n response = requests.get(self.url, cookies=self.cookies)\n soup = BeautifulSoup(response.text, 'html.parser')\n table_rows = soup.find_all(class_='table-row') \n for row in table_rows:\n track_title = row.select('.cell-title')[0].get_text()\n # artist_name = row.select('.cell-artist')[0].get_text()\n tracks_title.append(f'{track_title}')\n return tracks_title\n\n\nclass Spotify:\n\n def __init__(self, client_id, client_secret):\n self.client_id = client_id\n self.client_secret = client_secret\n self.sp = spotipy.Spotify(auth_manager=SpotifyOAuth(scope = \"playlist-modify-private\",redirect_uri = \"http://example.com\",\n client_id = client_id, client_secret = client_secret, show_dialog = True, cache_path='token.txt') \n )\n def create_playlist(self, user_id, playlist_name):\n playlist = self.sp.user_playlist_create(user=user_id, name=playlist_name, public=False)\n print(f'Playlist created.\\n')\n return playlist\n \n def fetch_track_uris(self, tracks_title):\n track_uris = []\n for track_name in tracks_title:\n response = self.sp.search(q=f\"track:{track_name}\", type=\"track\")\n try:\n uri = response[\"tracks\"][\"items\"][0][\"uri\"]\n track_uris.append(uri)\n except IndexError:\n print(f\"{track_name} doesn't exist in Spotify. Skipped.\")\n return track_uris\n\n def add_tracks_to_playlist(self, playlist_id, tracks_uris, user_id):\n self.sp.user_playlist_add_tracks(user=user_id, playlist_id=playlist_id, tracks=tracks_uris)\n\nif __name__ == \"__main__\":\n\n anghami_cookies = 'to_replace'\n\n # billboard = Billboard(year='2004-12-12')\n # billboard_tracks_title = billboard.fetch_tracks_title()\n # billboard_track_uris = spo.fetch_track_uris(billboard_tracks_title)\n anghami = Anghami('playlist_url', cookies=anghami_cookies)\n anghami_tracks_title = anghami.fetch_tracks_title()\n spo = Spotify(client_id='to_replace', client_secret='to_replace')\n playlist = spo.create_playlist(user_id=SPOTIFY_USER_ID, playlist_name='Anghami')\n anghami_track_uris = spo.fetch_track_uris(anghami_tracks_title)\n spo.add_tracks_to_playlist(playlist_id=playlist['id'], tracks_uris=anghami_track_uris, user_id=SPOTIFY_USER_ID)\n", "sub_path": "main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 3276, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "requests.get", "line_number": 17, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 18, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 33, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 34, "usage_type": "call"}, {"api_name": "spotipy.Spotify", "line_number": 48, "usage_type": "call"}, {"api_name": "spotipy.oauth2.SpotifyOAuth", "line_number": 48, "usage_type": "call"}]} +{"seq_id": "416931162", "text": "from django.conf.urls import url\n\nfrom mess.views.dialogs.auto_answer import AutoAnswersList, AutoAnswerDetail\nfrom mess.views.dialogs.dialogs import DialogsList, DialogDetail, DialogsSettingsList, DialogSettingsDetail\nfrom mess.views.dialogs.yes_no_answer import YesNoAnswerList, YesNoAnswerDetail\nfrom mess.views.mess.dispatch import DispatchList, DispatchDetail, DispatchTreeList, DispatchTreeDetail\nfrom mess.views.mess.mess import MessList, MessDetail\nfrom mess.views.mess.messages import MessagesList, MessageDetail, CreateListMessages, CreateInviteMessageWithLink\nfrom mess.views.mess.status import StatusList, StatusDetail, MarkAsReadMessStatus\nfrom mess.views.polls.answers import AnswersList, AnswerDetail\nfrom mess.views.polls.interview import InterviewList, InterviewDetail, InterviewStatisticsDetail\n\nurlpatterns = [\n # mess\n url(r'^mess/$', MessList.as_view()),\n url(r'^mess/(?P[0-9]+)/$', MessDetail.as_view()),\n\n # messages\n url(r'^mess/messages/$', MessagesList.as_view()),\n url(r'^mess/messages/(?P[0-9]+)/$', MessageDetail.as_view()),\n\n # create invite message\n url(r'^mess/messages/createinvite/', CreateInviteMessageWithLink.as_view()),\n\n # dispatch\n url(r'^mess/dispatch/$', DispatchList.as_view()),\n url(r'^mess/dispatch/(?P[0-9]+)/$', DispatchDetail.as_view()),\n\n # dispatchtree\n url(r'^mess/dispatchtree/$', DispatchTreeList.as_view()),\n url(r'^mess/dispatchtree/(?P[0-9]+)/$', DispatchTreeDetail.as_view()),\n\n # mark as read\n url(r'^mess/markasread/$', MarkAsReadMessStatus.as_view()),\n\n # create list message\n url(r'^mess/createlistmessages/$', CreateListMessages.as_view()),\n\n # status\n url(r'^mess/status/$', StatusList.as_view()),\n url(r'^mess/status/(?P[0-9]+)/$', StatusDetail.as_view()),\n\n # polls\n url(r'^mess/polls/answers/$', AnswersList.as_view()),\n url(r'^mess/polls/answers/(?P[0-9]+)/$', AnswerDetail.as_view()),\n\n url(r'^mess/polls/interview/$', InterviewList.as_view()),\n url(r'^mess/polls/interview/(?P[0-9]+)/$', InterviewDetail.as_view()),\n url(r'^mess/polls/interview/(?P[0-9]+)/statistics/$', InterviewStatisticsDetail.as_view()),\n\n # dialogs\n url(r'^mess/dialogs/dialogs/$', DialogsList.as_view()),\n url(r'^mess/dialogs/dialogs/(?P[0-9]+)/$', DialogDetail.as_view()),\n\n url(r'^mess/dialogs/autoanswers/$', AutoAnswersList.as_view()),\n url(r'^mess/dialogs/autoanswers/(?P[0-9]+)/$', AutoAnswerDetail.as_view()),\n\n url(r'^mess/dialogs/yesnoanswers/$', YesNoAnswerList.as_view()),\n url(r'^mess/dialogs/yesnoanswers/(?P[0-9]+)/$', YesNoAnswerDetail.as_view()),\n\n # dialogs settings\n url(r'^mess/dialogs/dialogsettings/$', DialogsSettingsList.as_view()),\n url(r'^mess/dialogs/dialogsettings/(?P[0-9]+)/$', DialogSettingsDetail.as_view()),\n]\n", "sub_path": "mess/urls.py", "file_name": "urls.py", "file_ext": "py", "file_size_in_byte": 2848, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "django.conf.urls.url", "line_number": 15, "usage_type": "call"}, {"api_name": "mess.views.mess.mess.MessList.as_view", "line_number": 15, "usage_type": "call"}, {"api_name": "mess.views.mess.mess.MessList", "line_number": 15, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 16, "usage_type": "call"}, {"api_name": "mess.views.mess.mess.MessDetail.as_view", "line_number": 16, "usage_type": "call"}, {"api_name": "mess.views.mess.mess.MessDetail", "line_number": 16, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 19, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.MessagesList.as_view", "line_number": 19, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.MessagesList", "line_number": 19, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 20, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.MessageDetail.as_view", "line_number": 20, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.MessageDetail", "line_number": 20, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 23, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.CreateInviteMessageWithLink.as_view", "line_number": 23, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.CreateInviteMessageWithLink", "line_number": 23, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 26, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchList.as_view", "line_number": 26, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchList", "line_number": 26, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 27, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchDetail.as_view", "line_number": 27, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchDetail", "line_number": 27, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 30, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchTreeList.as_view", "line_number": 30, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchTreeList", "line_number": 30, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 31, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchTreeDetail.as_view", "line_number": 31, "usage_type": "call"}, {"api_name": "mess.views.mess.dispatch.DispatchTreeDetail", "line_number": 31, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 34, "usage_type": "call"}, {"api_name": "mess.views.mess.status.MarkAsReadMessStatus.as_view", "line_number": 34, "usage_type": "call"}, {"api_name": "mess.views.mess.status.MarkAsReadMessStatus", "line_number": 34, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 37, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.CreateListMessages.as_view", "line_number": 37, "usage_type": "call"}, {"api_name": "mess.views.mess.messages.CreateListMessages", "line_number": 37, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 40, "usage_type": "call"}, {"api_name": "mess.views.mess.status.StatusList.as_view", "line_number": 40, "usage_type": "call"}, {"api_name": "mess.views.mess.status.StatusList", "line_number": 40, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 41, "usage_type": "call"}, {"api_name": "mess.views.mess.status.StatusDetail.as_view", "line_number": 41, "usage_type": "call"}, {"api_name": "mess.views.mess.status.StatusDetail", "line_number": 41, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 44, "usage_type": "call"}, {"api_name": "mess.views.polls.answers.AnswersList.as_view", "line_number": 44, "usage_type": "call"}, {"api_name": "mess.views.polls.answers.AnswersList", "line_number": 44, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 45, "usage_type": "call"}, {"api_name": "mess.views.polls.answers.AnswerDetail.as_view", "line_number": 45, "usage_type": "call"}, {"api_name": "mess.views.polls.answers.AnswerDetail", "line_number": 45, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 47, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewList.as_view", "line_number": 47, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewList", "line_number": 47, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 48, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewDetail.as_view", "line_number": 48, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewDetail", "line_number": 48, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 49, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewStatisticsDetail.as_view", "line_number": 49, "usage_type": "call"}, {"api_name": "mess.views.polls.interview.InterviewStatisticsDetail", "line_number": 49, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 52, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogsList.as_view", "line_number": 52, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogsList", "line_number": 52, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 53, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogDetail.as_view", "line_number": 53, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogDetail", "line_number": 53, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 55, "usage_type": "call"}, {"api_name": "mess.views.dialogs.auto_answer.AutoAnswersList.as_view", "line_number": 55, "usage_type": "call"}, {"api_name": "mess.views.dialogs.auto_answer.AutoAnswersList", "line_number": 55, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 56, "usage_type": "call"}, {"api_name": "mess.views.dialogs.auto_answer.AutoAnswerDetail.as_view", "line_number": 56, "usage_type": "call"}, {"api_name": "mess.views.dialogs.auto_answer.AutoAnswerDetail", "line_number": 56, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 58, "usage_type": "call"}, {"api_name": "mess.views.dialogs.yes_no_answer.YesNoAnswerList.as_view", "line_number": 58, "usage_type": "call"}, {"api_name": "mess.views.dialogs.yes_no_answer.YesNoAnswerList", "line_number": 58, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 59, "usage_type": "call"}, {"api_name": "mess.views.dialogs.yes_no_answer.YesNoAnswerDetail.as_view", "line_number": 59, "usage_type": "call"}, {"api_name": "mess.views.dialogs.yes_no_answer.YesNoAnswerDetail", "line_number": 59, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 62, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogsSettingsList.as_view", "line_number": 62, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogsSettingsList", "line_number": 62, "usage_type": "name"}, {"api_name": "django.conf.urls.url", "line_number": 63, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogSettingsDetail.as_view", "line_number": 63, "usage_type": "call"}, {"api_name": "mess.views.dialogs.dialogs.DialogSettingsDetail", "line_number": 63, "usage_type": "name"}]} +{"seq_id": "284371077", "text": "from sklearn.model_selection import train_test_split\nfrom sklearn.metrics import classification_report\nfrom sklearn.datasets import make_blobs\nimport matplotlib\nmatplotlib.use\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport argparse\n\ndef sigmoid_activation(x):\n return 1.0/(1+np.exp(-x))\n #returns b/w 0 to 1\n\ndef predict(X,W):\n #step function,if # >0.5 -> 1 otherwise 0\n #here # is the return from sigmoid_Activation\n #function, not the original #\n preds= sigmoid_activation(X.dot(W))\n preds[preds<=0.5]=0\n preds[preds>0]=1\n\n return preds\n\nap = argparse.ArgumentParser()\nap.add_argument(\"-e\", \"--epochs\", type=float, default=100,\nhelp=\"# of epochs\")\nap.add_argument(\"-a\", \"--alpha\", type=float, default=0.01,\nhelp=\"learning rate\")\nargs = vars(ap.parse_args()) \n\n# generate a 2-class classification problem with 1,000 data points,\n# where each data point is a 2D feature vector\n(X,y)=make_blobs(n_samples=1000, n_features=2,centers=2,cluster_std=1.5,\n random_state=1)\nprint(\"Original Shape of y was:\", y.shape)\ny=y.reshape((y.shape[0],1))#making it a 1d matrix rather than a vector\n\n#print(\"size of X is:\", X.shape)#(1000,2)\n#print(\"size of y is:\",y.shape)#(1000,1)\n\nX=np.c_[X, np.ones((X.shape[0]))]#bias trick\n#print(\"size of X after biasing is:\", X.shape)#(1000,3)\n\n(trainX, testX, trainY, testY)=train_test_split(X,y, test_size=0.5, random_state=42)\n\n#print(trainY) gives only 0 or 1.\n#print(\"shape of trainY is:\", trainY.shape) #(500,1)\nW=np.random.randn(X.shape[1],1) #our weighted matrix\n#print(\"shape of weighted matrix is:\",W.shape)\n#shape of weighted matrix is: (3, 1)\nlosses=[]\n\nfor epoch in np.arange(0, args[\"epochs\"]):\n preds=sigmoid_activation(trainX.dot(W))\n #i.e scoring_function=sigmoid_activation(linear_scoring_function)\n #and preds stores the value\n error=preds- trainY\n #the error matrix(here vector though) values will be b/w -1 and 1\n #print(\"dim of error matrix is:\",error.shape) (500,1)\n loss=np.sum(error**2)#compute least square error over our predictions,\n print(loss)\n print(loss.shape)\n #a simple loss typically used for binary classification problems\n losses.append(loss)\n \n '''\n the gradient descent update is the dot product between our\n features and the error of the predictions.\n in the update stage, all we need to do is \"nudge\" the weight\n matrix in the negative direction of the gradient (hence the\n term \"gradient descent\" by taking a small step towards a set\n of \"more optimal\" parameters\n '''\n gradient=trainX.T.dot(error) #trainX.T makes it transpose of trainX\n #derived mathematically, check notes \n\n W+= -args[\"alpha\"]*gradient\n \n\n #check to see if an update should be displayed\n if epoch==0 or (epoch+1)%5==0:\n print(\"[INFO] epoch={}, loss={:.7f}\".format(int(epoch+1),loss))\n\nprint(\"[INFO] evaluating\")\npreds=predict(testX, W)\nprint(classification_report(testY, preds))\n\n# plot the (testing) classification data\nplt.style.use(\"ggplot\")\nplt.figure()\nplt.title(\"Data\")\nplt.scatter(testX[:, 0], testX[:, 1], marker=\"o\", c=testY, s=30)\n\n# construct a figure that plots the loss over time\nplt.style.use(\"ggplot\")\nplt.figure()\nplt.plot(np.arange(0, args[\"epochs\"]), losses)\nplt.title(\"Training Loss\")\nplt.xlabel(\"Epoch #\")\nplt.ylabel(\"Loss\")\nplt.show()\n\n", "sub_path": "gradient_descent.py", "file_name": "gradient_descent.py", "file_ext": "py", "file_size_in_byte": 3332, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "matplotlib.use", "line_number": 5, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 11, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 24, "usage_type": "call"}, {"api_name": "sklearn.datasets.make_blobs", "line_number": 33, "usage_type": "call"}, {"api_name": "numpy.c_", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 41, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 44, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 48, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 60, "usage_type": "call"}, {"api_name": "sklearn.metrics.classification_report", "line_number": 86, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 89, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 89, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 89, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 90, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 90, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 91, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 91, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 92, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 92, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.style.use", "line_number": 95, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.style", "line_number": 95, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot", "line_number": 95, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 96, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 96, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 97, "usage_type": "name"}, {"api_name": "numpy.arange", "line_number": 97, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.title", "line_number": 98, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 98, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 99, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 99, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 100, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 100, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 101, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 101, "usage_type": "name"}]} +{"seq_id": "538563481", "text": "\"\"\"\n Implements PPO\n\n PPO: https://arxiv.org/abs/1707.06347\n Modified from policy Written by Patrick Coady (pat-coady.github.io) to implement\n latest version of PPO with pessimistic ratio clipping\n\n o Has an option to servo both the learning rate and the clip_param to keep KL \n within a specified range. This helps on some control tasks\n (i.e., Mujoco Humanid-v2)\n \n o Uses approximate KL \n\n o Models distribution of actions as a Gaussian with variance not conditioned on state\n\n o Has option to discretize sampled actions\n \n\"\"\"\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom rl_utils import Action_converter\nimport rl_utils\nfrom time import time\nimport sklearn.utils\n \nclass Policy(object):\n \"\"\" NN-based policy approximation \"\"\"\n def __init__(self, net, actions_per_dim=3, kl_targ=0.003,epochs=20,discretize=False, init_func=rl_utils.default_init,\n test_mode=False,shuffle=True, servo_kl=False, beta=0.1, max_grad_norm=999, \n obs_key='observes', scale_obs=True, verbose=False, rollout_limit=1,\n mpc_samples=1000, mpc_steps=10, use_rewards=True):\n \"\"\"\n Args:\n actions_per_dim: used when discretizing action space\n kl_targ: target KL divergence between pi_old and pi_new\n epochs: number of epochs per update\n discretize: boolean, True discretizes action space\n test_mode: boolean, True removes all exploration noise\n shuffle: boolean, shuffles data each epoch \n servo_kl: boolean: set to False to not servo beta to KL, which is original PPO implementation\n beta: clipping parameter for pessimistic loss ratio\n \n \"\"\"\n print('Policy with vectorized sample')\n net.apply(init_func)\n\n self.net = net\n \n self.servo_kl = servo_kl\n self.test_mode = test_mode\n self.discretize = discretize\n self.shuffle = shuffle\n if self.net.recurrent_steps > 1:\n print('Policy: recurrent steps > 1, disabling shuffle')\n self.shuffle = False\n self.actions_per_dim = actions_per_dim\n self.kl_stat = None\n self.entropy_stat = None\n self.kl_targ = kl_targ\n self.epochs = epochs \n self.lr_multiplier = 1.0 # dynamically adjust lr when D_KL out of control\n self.max_beta = 0.5\n self.min_beta = 0.01 \n self.max_grad_norm = max_grad_norm\n self.beta = beta\n self.obs_key = obs_key\n self.action_converter = Action_converter(1,actions_per_dim)\n self.grad_monitor = rl_utils.Grad_monitor('Policy', net)\n self.scaler = rl_utils.Scaler(net.obs_dim)\n self.scale_obs = scale_obs\n self.verbose = verbose \n self.rollout_limit = rollout_limit\n self.rollout_list = []\n\n self.mpc_samples = mpc_samples\n self.mpc_steps = mpc_steps\n self.use_rewards = use_rewards\n\n if self.net.recurrent_steps > 1:\n self.use_padding = True\n else:\n self.use_padding = False\n\n self.optimizer = torch.optim.Adam(self.net.parameters(), self.net.lr)\n\n print('\\tTest Mode: ',self.test_mode)\n print('\\tClip Param: ',self.beta)\n print('\\tShuffle : ',self.shuffle)\n print('\\tMax Grad Norm: ',self.max_grad_norm)\n print('\\tRecurrent Steps: ',self.net.recurrent_steps)\n print('\\tRollout Limit: ',self.rollout_limit)\n\n def save_params(self,fname):\n fname = 'policy_' + fname + '.pt'\n param_dict = {}\n param_dict['scaler_u'] = self.scaler.means\n param_dict['scaler_var'] = self.scaler.vars\n param_dict['net_state'] = self.net.state_dict()\n torch.save(param_dict, fname)\n\n def load_params(self,fname):\n fname = 'policy_' + fname + '.pt'\n param_dict = torch.load(fname)\n self.scaler.means = param_dict['scaler_u']\n self.scaler.vars = param_dict['scaler_var']\n self.net.load_state_dict(param_dict['net_state'])\n\n def _kl_entropy(self, logp, old_logp, log_vars, masks):\n \n \"\"\"\n We do approximate KL here\n https://en.wikipedia.org/wiki/Multivariate_normal_distribution#Entropy\n \"\"\"\n\n if self.use_padding:\n logp, old_logp = rl_utils.unpad_list([logp, old_logp],masks)\n\n kl = 0.5 * np.mean((logp - old_logp)**2)\n entropy = 0.5 * (self.net.act_dim * (np.log(2 * np.pi) + 1) +\n np.sum(log_vars))\n\n return kl, entropy\n\n\n def sample1(self, obs, state):\n \"\"\"Draw sample from policy distribution\"\"\"\n\n if self.scale_obs:\n obs = self.scaler.apply(obs)\n deterministic_action, log_vars, state = self.net.forward(obs, state, np.ones(1), np.zeros(1), return_tensor=False)\n if self.test_mode:\n action = deterministic_action \n else:\n sd = np.exp(log_vars / 2.0)\n action = deterministic_action + np.random.normal(scale=sd, size=(obs.shape[0], sd.shape[0]))\n\n return action, state \n\n \n def sample(self, obs, policy_state, model_state, model):\n t0 = time()\n observes = np.tile(obs,(self.mpc_samples,1))\n model_errors = np.zeros_like(observes)\n #print('MS: ',model_state.shape)\n model_state = np.tile(model_state, (self.mpc_samples,1))\n #print('MS: ',model_state.shape)\n \n masks = np.ones(self.mpc_samples)\n flags = np.zeros(self.mpc_samples)\n\n rewards = np.zeros(self.mpc_samples)\n \n for k in range(self.mpc_steps):\n actions, policy_state = self.sample1(observes, policy_state)\n if self.net.recurrent_steps == 1 or self.net.recurrent_steps is None:\n policy_state = np.zeros((self.mpc_samples, 1)) \n if k == 0:\n first_actions = actions.copy()\n first_states = policy_state.copy()\n observes, rpreds, model_state, _ = model.predict(observes, actions, observes, model_state, model_errors, masks, flags )\n rewards += rpreds\n\n if self.use_rewards:\n best = np.argmax(rewards)\n else:\n best = np.argmax(rpreds) # which should actually be values if we use correct model here\n act = np.expand_dims(first_actions[best],axis=0)\n state = first_states[best]\n if len(state.shape) ==1:\n state = np.expand_dims(state,axis=0)\n #print('ACT: ', act.shape)\n #print('PS: ', state.shape)\n if self.verbose:\n print('sample et: ',time()-t0) \n return act, act.copy(), state \n \n def update_scalers(self, rollouts):\n self.scaler.update(rollouts[self.obs_key])\n\n def update(self, rollouts, logger):\n if len(self.rollout_list) == self.rollout_limit:\n del self.rollout_list[0]\n self.rollout_list.append(rollouts)\n keys = self.rollout_list[0].keys()\n comb_rollouts = {}\n for k in keys:\n comb_rollouts[k] = np.concatenate([r[k] for r in self.rollout_list])\n self.update1(comb_rollouts, logger)\n \n def update1(self, rollouts, logger):\n \n if self.use_padding:\n key = 'padded_'\n else:\n key = '' \n observes = rollouts[key + self.obs_key]\n actions = rollouts[key + 'actions']\n states = rollouts[key + 'policy_states']\n vtarg = rollouts[key + 'disc_sum_rew']\n vpred = rollouts[key + 'vpreds']\n masks = rollouts[key + 'masks']\n flags = rollouts[key + 'flags']\n\n if self.scale_obs:\n observes = self.scaler.apply(observes)\n \n vtarg_unp = rollouts['disc_sum_rew']\n vpred_unp = rollouts['vpreds']\n \n actions_pt = torch.from_numpy(actions).float()\n with torch.no_grad():\n means_pt, logvars_pt, _ = self.net.forward(observes, states, masks, flags)\n\n old_logp_pt = self.calc_logp(actions_pt, means_pt, logvars_pt) \n old_logp = old_logp_pt.detach().numpy() \n loss, kl, entropy = 0, 0, 0\n\n advantages_unp = vtarg_unp - vpred_unp\n u_adv = advantages_unp.mean()\n std_adv = advantages_unp.std() + 1e-6\n\n advantages = vtarg - vpred \n advantages = (advantages - u_adv) / std_adv \n\n t0 = time()\n for e in range(self.epochs):\n\n if self.shuffle:\n observes, actions, advantages, states, masks, flags, old_logp = \\\n sklearn.utils.shuffle(observes, actions, advantages, states, masks, flags, old_logp)\n\n actions_pt = torch.from_numpy(actions).float()\n\n self.optimizer.zero_grad()\n means_pt, log_vars_pt, _ = self.net.forward(observes, states, masks, flags, unroll=True)\n logp_pt = self.calc_logp(actions_pt, means_pt, log_vars_pt)\n loss = self.calc_loss(logp_pt, torch.from_numpy(old_logp).float(), torch.from_numpy(advantages).float(), self.beta, masks)\n loss.backward()\n if self.max_grad_norm is not None:\n ng = nn.utils.clip_grad_norm_(self.net.parameters(), self.max_grad_norm)\n else:\n ng = None\n self.optimizer.step()\n self.grad_monitor.add(ng)\n \n log_vars = log_vars_pt.detach().numpy()\n kl, entropy = self._kl_entropy(logp_pt.detach().numpy(), old_logp, log_vars, masks)\n\n if kl > 4.0 * self.kl_targ and self.servo_kl:\n print(' *** BROKE ***')\n break \n\n t1 = time()\n \n if self.servo_kl:\n self.adjust_beta(kl)\n\n for g in self.optimizer.param_groups:\n g['lr'] = self.net.lr * self.lr_multiplier\n self.kl_stat = kl\n self.entropy_stat = entropy\n var_monitor = np.exp(log_vars/2.0)\n self.grad_monitor.show()\n\n if self.verbose:\n print('POLICY ROLLOUT LIST: ',len(self.rollout_list))\n print('POLICY Update: ',t1-t0,observes.shape)\n print('kl = ',kl, ' beta = ',self.beta,' lr_mult = ',self.lr_multiplier)\n print('var: ' ,var_monitor)\n print('u_adv: ',u_adv)\n print('std_adv: ',std_adv)\n\n logger.log({'PolicyLoss': loss,\n 'PolicyEntropy': entropy,\n 'KL': kl,\n 'Beta': self.beta,\n 'Variance' : np.max(var_monitor),\n 'lr_multiplier': self.lr_multiplier})\n\n def adjust_beta(self,kl):\n if kl < self.kl_targ / 2:\n self.beta = np.minimum(self.max_beta, 1.5 * self.beta) # max clip beta\n #print('too low')\n if self.beta > (self.max_beta/2) and self.lr_multiplier < 10:\n self.lr_multiplier *= 1.5\n elif kl > self.kl_targ * 2:\n #print('too high')\n self.beta = np.maximum(self.min_beta, self.beta / 1.5) # min clip beta\n if self.beta <= (2*self.min_beta) and self.lr_multiplier > 0.1:\n self.lr_multiplier /= 1.5\n\n def calc_loss(self,logp, old_logp, advantages, beta, masks):\n if self.use_padding:\n logp, old_logp, advantages = rl_utils.unpad_list([logp, old_logp, advantages], masks)\n\n ratio = torch.exp(logp - old_logp)\n surr1 = advantages * ratio\n surr2 = advantages * torch.clamp(ratio, 1.0 - beta, 1.0 + beta)\n \n loss = -torch.mean(torch.min(surr1,surr2)) \n return loss\n\n\n def calc_logp(self, act, means, log_vars):\n logp1 = -0.5 * torch.sum(log_vars)\n diff = act - means\n logp2 = -0.5 * torch.sum(torch.mul(diff, diff) / torch.exp(log_vars), 1)\n logp3 = -0.5 * np.log(2.0 * np.pi) * self.net.act_dim\n logp = logp1 + logp2 + logp3\n return logp\n\n", "sub_path": "RL_lib/Agents/PPO/policy_mpc.py", "file_name": "policy_mpc.py", "file_ext": "py", "file_size_in_byte": 12064, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "rl_utils.default_init", "line_number": 30, "usage_type": "attribute"}, {"api_name": "rl_utils.Action_converter", "line_number": 69, "usage_type": "call"}, {"api_name": "rl_utils.Grad_monitor", "line_number": 70, "usage_type": "call"}, {"api_name": "rl_utils.Scaler", "line_number": 71, "usage_type": "call"}, {"api_name": "torch.optim.Adam", "line_number": 86, "usage_type": "call"}, {"api_name": "torch.optim", "line_number": 86, "usage_type": "attribute"}, {"api_name": "torch.save", "line_number": 101, "usage_type": "call"}, {"api_name": "torch.load", "line_number": 105, "usage_type": "call"}, {"api_name": "rl_utils.unpad_list", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 121, "usage_type": "attribute"}, {"api_name": "numpy.sum", "line_number": 122, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 132, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 137, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 137, "usage_type": "attribute"}, {"api_name": "time.time", "line_number": 143, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 145, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 158, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 166, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 168, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 169, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 172, "usage_type": "call"}, {"api_name": "time.time", "line_number": 176, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 189, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 212, "usage_type": "call"}, {"api_name": "torch.no_grad", "line_number": 213, "usage_type": "call"}, {"api_name": "time.time", "line_number": 227, "usage_type": "call"}, {"api_name": "sklearn.utils.utils.shuffle", "line_number": 232, "usage_type": "call"}, {"api_name": "sklearn.utils.utils", "line_number": 232, "usage_type": "attribute"}, {"api_name": "sklearn.utils", "line_number": 232, "usage_type": "name"}, {"api_name": "torch.from_numpy", "line_number": 234, "usage_type": "call"}, {"api_name": "torch.from_numpy", "line_number": 239, "usage_type": "call"}, {"api_name": "torch.nn.utils.clip_grad_norm_", "line_number": 242, "usage_type": "call"}, {"api_name": "torch.nn.utils", "line_number": 242, "usage_type": "attribute"}, {"api_name": "torch.nn", "line_number": 242, "usage_type": "name"}, {"api_name": "time.time", "line_number": 255, "usage_type": "call"}, {"api_name": "numpy.exp", "line_number": 264, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 279, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 284, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 290, "usage_type": "call"}, {"api_name": "rl_utils.unpad_list", "line_number": 296, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 298, "usage_type": "call"}, {"api_name": "torch.clamp", "line_number": 300, "usage_type": "call"}, {"api_name": "torch.mean", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.min", "line_number": 302, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 307, "usage_type": "call"}, {"api_name": "torch.sum", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.mul", "line_number": 309, "usage_type": "call"}, {"api_name": "torch.exp", "line_number": 309, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 310, "usage_type": "attribute"}]} +{"seq_id": "298702028", "text": "import numpy as np\nfrom tensorflow.keras.preprocessing.image import ImageDataGenerator\n\nx_train = np.load('./_save/_npy/k59_manwoman_train_x.npy')\nx_test = np.load('./_save/_npy/k59_manwoman_test_x.npy')\nx_predic = np.load('./_save/_npy/k59_manwoman_predic_x.npy')\n\n\nx_train_noised = x_train + np.random.normal(0,0.1,size=x_train.shape)\nx_test_noised = x_test + np.random.normal(0,0.1,size=x_test.shape)\nx_predic_noised = x_predic + np.random.normal(0,0.1,size=x_test.shape)\nx_train_noised = np.clip(x_train_noised,a_min=0,a_max=1)\nx_test_noised = np.clip(x_test_noised,a_min=0,a_max=1)\nx_predic_noised = np.clip(x_test_noised,a_min=0,a_max=1)\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras.layers import Conv2D,Dense,Flatten\n\ndef autoencoder(hidden_layer_size):\n model = Sequential()\n model.add(Dense(units=hidden_layer_size, input_shape=(150,150,3),activation='relu'))\n model.add(Dense(3,activation='softmax'))\n return model\na=154\nmodel = autoencoder(a)\nmodel.compile(optimizer='adam',loss='mse')\nmodel.fit(x_train_noised,x_train,epochs=10)\n\ny_predic = model.predict([x_predic][-1])\n\nfrom matplotlib import pyplot as plt\nimport random\nfig, ((ax1,ax2,ax3,ax4,ax5),\n (ax11,ax12,ax13,ax14,ax15),\n (ax6,ax7,ax8,ax9,ax10))= \\\n plt.subplots(3,1,figsize=(20,7))\n# 이미지를 무작위로 5개 고른다\n#ran_image = random.sample(range(y_predic.shape[0]),5)\n\nfor i,ax in enumerate([ax1,ax2,ax3,ax4,ax5]):\n ax.imshow(x_predic[ran_image[i]].reshape(28,28),cmap='gray')\n if i == 0:\n ax.set_ylabel('INPUT',size=20)\n ax.grid(False)\n ax.set_xticks([])\n ax.set_yticks([])\n\nfor i,ax in enumerate([ax11,ax12,ax13,ax14,ax15]):\n ax.imshow(x_predic_noised[ran_image[i]].reshape(28,28),cmap='gray')\n if i == 0:\n ax.set_ylabel('NOISED_INPUT',size=20)\n ax.grid(False)\n ax.set_xticks([])\n ax.set_yticks([])\n\nfor i,ax in enumerate([ax6,ax7,ax8,ax9,ax10]):\n ax.imshow(y_predic[ran_image[i]].reshape(28,28),cmap='gray')\n if i == 0:\n ax.set_ylabel('OUTPUT',size=20)\n ax.grid(False)\n ax.set_xticks([])\n ax.set_yticks([]) \n\nplt.tight_layout()\nplt.show()", "sub_path": "AE/a08_noise3_mali_female.py", "file_name": "a08_noise3_mali_female.py", "file_ext": "py", "file_size_in_byte": 2172, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.load", "line_number": 4, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 5, "usage_type": "call"}, {"api_name": "numpy.load", "line_number": 6, "usage_type": "call"}, {"api_name": "numpy.random.normal", "line_number": 9, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 9, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 10, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 10, "usage_type": "attribute"}, {"api_name": "numpy.random.normal", "line_number": 11, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 11, "usage_type": "attribute"}, {"api_name": "numpy.clip", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 13, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 14, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.Sequential", "line_number": 19, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 20, "usage_type": "call"}, {"api_name": "tensorflow.keras.layers.Dense", "line_number": 21, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.subplots", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.tight_layout", "line_number": 63, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 63, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 64, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 64, "usage_type": "name"}]} +{"seq_id": "294802192", "text": "from tkinter import *\nfrom PIL import Image, ImageTk\n\nfrom tkinter import colorchooser\n\nIMAGE_PATH = 'background.jpg'\nWIDTH, HEIGTH = 500, 500\n\nroot = Tk()\nroot.geometry('{}x{}+100+100'.format(WIDTH, HEIGTH))\n\ncanvas = Canvas(root, width=WIDTH, height=HEIGTH)\ncanvas.pack()\n\nimg = ImageTk.PhotoImage(Image.open(IMAGE_PATH).resize((WIDTH, HEIGTH), Image.ANTIALIAS))\ncanvas.background = img # Keep a reference in case this code is put in a function.\nbg = canvas.create_image(0, 0, anchor=NW, image=img)\n\n\ndef colour():\n colour = colorchooser.askcolor()\n bg = colour[1]\n cBtn.config(bg=bg)\n button.config(bg=bg)\n lab.config(bg=bg)\n print(colour[1])\n\n\n# Put a tkinter widget on the canvas.\nbutton = Button(root, text=\"Start\", relief=GROOVE)\ncBtn = Button(root, text=\"Colour\", command=colour, relief=GROOVE)\nbutton_window = canvas.create_window(150, 200, anchor=NW, window=button)\ncBtn_window = canvas.create_window(200, 200, anchor=NW, window=cBtn)\n\nlab = Label(root, text='the colour')\nlab_wind = canvas.create_window(10, 200, anchor=NW, window=lab)\n\nroot.mainloop()\n", "sub_path": "Python Programs/Tkinter/TO_DO_List/Image as background.py", "file_name": "Image as background.py", "file_ext": "py", "file_size_in_byte": 1083, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "PIL.ImageTk.PhotoImage", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.ImageTk", "line_number": 15, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 15, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 15, "usage_type": "name"}, {"api_name": "PIL.Image.ANTIALIAS", "line_number": 15, "usage_type": "attribute"}, {"api_name": "tkinter.colorchooser.askcolor", "line_number": 21, "usage_type": "call"}, {"api_name": "tkinter.colorchooser", "line_number": 21, "usage_type": "name"}]} +{"seq_id": "10889412", "text": "\"\"\"\nCopyright (R) @huawei.com, all rights reserved\n-*- coding:utf-8 -*-\nCREATED: 2021-01-20 20:12:13\nMODIFIED: 2021-01-29 14:04:45\n\"\"\"\n\nimport sys\nimport os\nimport acl\nimport image_net_classes\npath = os.path.dirname(os.path.abspath(__file__))\n\nsys.path.append(os.path.join(path, \"..\"))\nsys.path.append(os.path.join(path, \"../../../../common/\"))\nsys.path.append(os.path.join(path, \"../../../../common/atlas_utils\"))\n\nfrom constants import ACL_MEM_MALLOC_HUGE_FIRST, ACL_MEMCPY_DEVICE_TO_DEVICE, IMG_EXT\nfrom acl_dvpp import Dvpp\nfrom acl_model import Model\nfrom acl_image import AclImage\nfrom acl_resource import AclResource\n\nfrom PIL import Image, ImageDraw, ImageFont\n\nclass Classify(object):\n \"\"\"\n\tdefine gesture class\n \"\"\"\n def __init__(self, acl_resource, model_path, model_width, model_height):\n self._model_path = model_path\n self._model_width = model_width\n self._model_height = model_height\n self._dvpp = Dvpp(acl_resource)\n self._model = Model(model_path)\n\n def __del__(self):\n if self._dvpp:\n del self._dvpp\n print(\"[Sample] class Samle release source success\")\n\n def pre_process(self, image):\n \"\"\"\n pre_precess\n \"\"\"\n yuv_image = self._dvpp.jpegd(image)\n resized_image = self._dvpp.resize(yuv_image, \n self._model_width, self._model_height)\n print(\"resize yuv end\")\n return resized_image\n\n def inference(self, resized_image):\n \"\"\"\n\t inference\n \"\"\"\n return self._model.execute([resized_image, ])\n\n def post_process(self, infer_output, image_file):\n \"\"\"\n\t post_process\n \"\"\"\n print(\"post process\")\n data = infer_output[0]\n vals = data.flatten()\n top_k = vals.argsort()[-1:-6:-1]\n print(\"images:{}\".format(image_file))\n print(\"======== top5 inference results: =============\")\n for n in top_k:\n object_class = image_net_classes.get_image_net_class(n)\n print(\"label:%d confidence: %f, class: %s\" % (n, vals[n], object_class))\n \n #using pillow, the category with the highest confidence is written on the image and saved locally\n if len(top_k):\n object_class = image_net_classes.get_image_net_class(top_k[0])\n output_path = os.path.join(os.path.join(SRC_PATH, \"../outputs\"), os.path.basename(image_file))\n origin_img = Image.open(image_file)\n draw = ImageDraw.Draw(origin_img)\n font = ImageFont.load_default()\n draw.text((10, 50), object_class, font=font, fill=255)\n origin_img.save(output_path)\n\nSRC_PATH = os.path.realpath(__file__).rsplit(\"/\", 1)[0]\nMODEL_PATH = os.path.join(SRC_PATH, \"../model/googlenet_yuv.om\")\nMODEL_WIDTH = 224\nMODEL_HEIGHT = 224\n\n\ndef main():\n\n \"\"\"\n Program execution with picture directory parameters\n \"\"\"\n if (len(sys.argv) != 2):\n print(\"The App arg is invalid\")\n exit(1)\n acl_resource = AclResource()\n acl_resource.init()\n #Instantiation classification detection, incoming om model path, model input width and height parameters\n classify = Classify(acl_resource, MODEL_PATH, MODEL_WIDTH, MODEL_HEIGHT)\n \n #Get the picture storage directory from the parameters, and infer picture by picture\n image_dir = sys.argv[1]\n images_list = [os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if os.path.splitext(img)[1] in IMG_EXT]\n \n #Create a directory and save the infer results\n if not os.path.isdir(os.path.join(SRC_PATH, \"../outputs\")):\n os.mkdir(os.path.join(SRC_PATH, \"../outputs\"))\n\n for image_file in images_list:\n #read the picture\n image = AclImage(image_file)\n image_dvpp = image.copy_to_dvpp()\n #preprocess image\n resized_image = classify.pre_process(image_dvpp)\n print(\"pre process end\")\n #inference\n result = classify.inference(resized_image)\n #post process\n classify.post_process(result, image_file)\n\nif __name__ == '__main__':\n main()\n \n", "sub_path": "python/level2_simple_inference/1_classification/googlenet_onnx_picture/src/classify.py", "file_name": "classify.py", "file_ext": "py", "file_size_in_byte": 4133, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.dirname", "line_number": 12, "usage_type": "call"}, {"api_name": "os.path", "line_number": 12, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 12, "usage_type": "call"}, {"api_name": "sys.path.append", "line_number": 14, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 14, "usage_type": "call"}, {"api_name": "os.path", "line_number": 14, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 15, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 15, "usage_type": "call"}, {"api_name": "os.path", "line_number": 15, "usage_type": "attribute"}, {"api_name": "sys.path.append", "line_number": 16, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 16, "usage_type": "call"}, {"api_name": "os.path", "line_number": 16, "usage_type": "attribute"}, {"api_name": "acl_dvpp.Dvpp", "line_number": 34, "usage_type": "call"}, {"api_name": "acl_model.Model", "line_number": 35, "usage_type": "call"}, {"api_name": "image_net_classes.get_image_net_class", "line_number": 69, "usage_type": "call"}, {"api_name": "image_net_classes.get_image_net_class", "line_number": 74, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 75, "usage_type": "call"}, {"api_name": "os.path", "line_number": 75, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 75, "usage_type": "call"}, {"api_name": "PIL.Image.open", "line_number": 76, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 76, "usage_type": "name"}, {"api_name": "PIL.ImageDraw.Draw", "line_number": 77, "usage_type": "call"}, {"api_name": "PIL.ImageDraw", "line_number": 77, "usage_type": "name"}, {"api_name": "PIL.ImageFont.load_default", "line_number": 78, "usage_type": "call"}, {"api_name": "PIL.ImageFont", "line_number": 78, "usage_type": "name"}, {"api_name": "os.path.realpath", "line_number": 82, "usage_type": "call"}, {"api_name": "os.path", "line_number": 82, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 83, "usage_type": "call"}, {"api_name": "os.path", "line_number": 83, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 93, "usage_type": "attribute"}, {"api_name": "acl_resource.AclResource", "line_number": 96, "usage_type": "call"}, {"api_name": "acl_resource.init", "line_number": 97, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 104, "usage_type": "call"}, {"api_name": "os.path.splitext", "line_number": 105, "usage_type": "call"}, {"api_name": "os.path", "line_number": 105, "usage_type": "attribute"}, {"api_name": "constants.IMG_EXT", "line_number": 105, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 108, "usage_type": "call"}, {"api_name": "os.path", "line_number": 108, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 108, "usage_type": "call"}, {"api_name": "os.mkdir", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 109, "usage_type": "call"}, {"api_name": "os.path", "line_number": 109, "usage_type": "attribute"}, {"api_name": "acl_image.AclImage", "line_number": 113, "usage_type": "call"}]} +{"seq_id": "35764923", "text": "from google.cloud import texttospeech\nfrom playsound import playsound\n\ntts_client = texttospeech.TextToSpeechClient()\nparams = texttospeech.VoiceSelectionParams(language_code='en-US', ssml_gender=texttospeech.SsmlVoiceGender.FEMALE)\naudio = texttospeech.AudioConfig(audio_encoding=texttospeech.AudioEncoding.MP3)\nsi = texttospeech.SynthesisInput(text='Peter Piper picked a peck of pickled peppers.')\nresponse = tts_client.synthesize_speech(input=si, voice=params, audio_config=audio)\nf = open('en_us_female.mp3', 'wb')\nf.write(response.audio_content)\nf.close()\nplaysound('en_us_female.mp3')\n", "sub_path": "cricket/tracker/soundTest.py", "file_name": "soundTest.py", "file_ext": "py", "file_size_in_byte": 591, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "google.cloud.texttospeech.TextToSpeechClient", "line_number": 4, "usage_type": "call"}, {"api_name": "google.cloud.texttospeech", "line_number": 4, "usage_type": "name"}, {"api_name": "google.cloud.texttospeech.VoiceSelectionParams", "line_number": 5, "usage_type": "call"}, {"api_name": "google.cloud.texttospeech", "line_number": 5, "usage_type": "name"}, {"api_name": "google.cloud.texttospeech.SsmlVoiceGender", "line_number": 5, "usage_type": "attribute"}, {"api_name": "google.cloud.texttospeech.AudioConfig", "line_number": 6, "usage_type": "call"}, {"api_name": "google.cloud.texttospeech", "line_number": 6, "usage_type": "name"}, {"api_name": "google.cloud.texttospeech.AudioEncoding", "line_number": 6, "usage_type": "attribute"}, {"api_name": "google.cloud.texttospeech.SynthesisInput", "line_number": 7, "usage_type": "call"}, {"api_name": "google.cloud.texttospeech", "line_number": 7, "usage_type": "name"}, {"api_name": "playsound.playsound", "line_number": 12, "usage_type": "call"}]} +{"seq_id": "8973694", "text": "# -*- coding: utf-8 -*-\nimport json\nimport re\n\nimport scrapy\n\nfrom locations.items import GeojsonPointItem\n\n\nclass FcBankingSpider(scrapy.Spider):\n name = \"fcbanking\"\n item_attributes = {\"brand\": \"First Commonwealth Bank\", \"brand_wikidata\": \"Q5452773\"}\n allowed_domains = [\"www.fcbanking.com\"]\n start_urls = [\n \"https://www.fcbanking.com/sitemap/branch-locations_0.xml\",\n \"https://www.fcbanking.com/sitemap/branch-locations_1.xml\",\n ]\n\n def parse(self, response):\n response.selector.remove_namespaces()\n urls = response.xpath(\"//loc/text()\").extract()\n\n for url in urls:\n if url == \"https://www.fcbanking.com/branch-locations/\":\n continue\n yield scrapy.Request(url, callback=self.parse_branch)\n\n def parse_branch(self, response):\n map_script = response.xpath('//script/text()[contains(., \"setLat\")]').get()\n ldjson = response.xpath('//script[@type=\"application/ld+json\"]/text()').get()\n data = json.loads(re.sub(r\"^//.*$\", \"\", ldjson, flags=re.M))\n\n lat = re.search(r'setLat\\(\"(.*)\"\\)', map_script)[1]\n lon = re.search(r'setLon\\(\"(.*)\"\\)', map_script)[1]\n address = data[\"address\"]\n\n properties = {\n \"lat\": lat,\n \"lon\": lon,\n \"ref\": response.url,\n \"name\": data[\"name\"],\n \"addr_full\": address[\"streetAddress\"],\n \"city\": address[\"addressLocality\"],\n \"state\": address[\"addressRegion\"],\n \"postcode\": address[\"postalCode\"],\n \"phone\": data[\"telephone\"],\n \"website\": response.url,\n \"opening_hours\": \",\".join(data[\"openingHours\"]),\n }\n\n return GeojsonPointItem(**properties)\n", "sub_path": "locations/spiders/fcbanking.py", "file_name": "fcbanking.py", "file_ext": "py", "file_size_in_byte": 1744, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "scrapy.Spider", "line_number": 10, "usage_type": "attribute"}, {"api_name": "scrapy.Request", "line_number": 26, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 31, "usage_type": "call"}, {"api_name": "re.sub", "line_number": 31, "usage_type": "call"}, {"api_name": "re.M", "line_number": 31, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 33, "usage_type": "call"}, {"api_name": "re.search", "line_number": 34, "usage_type": "call"}, {"api_name": "locations.items.GeojsonPointItem", "line_number": 51, "usage_type": "call"}]} +{"seq_id": "348158945", "text": "import numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\n#import tensorflow.contrib.eager as tfe # error in VsCode but works!!\n\ntf.enable_eager_execution()\ntf.set_random_seed(777) # for reproducibility\nprint(tf.__version__)\n\n# x_data is two-dimensional-array\n# test data is red dot in graph\nx_train = [[1., 2.],\n [2., 3.],\n [3., 1.],\n [4., 3.],\n [5., 3.],\n [6., 2.]]\ny_train = [[0.],\n [0.],\n [0.],\n [1.],\n [1.],\n [1.]]\n\nx_test = [[5., 2.]]\ny_test = [[1.]]\n\nx1 = [x[0] for x in x_train]\nx2 = [x[1] for x in x_train]\n\ncolors = [int(y[0] % 3) for y in y_train]\nplt.scatter(x1, x2, c = colors, marker = '^')\nplt.scatter(x_test[0][0], x_test[0][1], c = \"red\")\n\nplt.xlabel(\"x1\")\nplt.ylabel(\"x2\")\nplt.show()\n\n# use Tensorflow data API for data\ndataset = tf.data.Dataset.from_tensor_slices((x_train, y_train)).batch(len(x_train))\n\n# Weight and Bias (can be 0 or random(tf.random_normal([2,1])))\nW = tf.Variable(tf.zeros([2, 1]), name = 'weight')\nb = tf.Variable(tf.zeros([1]), name = 'bias')\n\n# def hypothesis, loss function and accuracy_check function\ndef logistic_regression(features):\n hypothesis = tf.div(1., 1. + tf.exp(tf.matmul(features, W) + b))\n return hypothesis\n\ndef loss_fn(hypothesis, features, labels):\n cost = -tf.reduce_mean(labels * tf.log(logistic_regression(features)) + (1 - labels) * tf.log(1 - hypothesis))\n return cost\n\noptimizer = tf.train.GradientDescentOptimizer(learning_rate = 0.01)\n\ndef accuracy_fn(hypothesis, labels):\n predicted = tf.cast(hypothesis > 0.5, dtype = tf.float32)\n accuracy = tf.reduce_mean(tf.cast(tf.equal(predicted, labels), dtype = tf.int32))\n return accuracy\n\n# calculate gradient by GradientTape\ndef grad(hypothesis, features, labels):\n with tf.GradientTape() as tape:\n loss_value = loss_fn(logistic_regression(features), features, labels)\n return tape.gradient(loss_value, [W, b])\n\n# train on Eager mode\nEPOCHS = 1001\n\nfor step in range(EPOCHS):\n #for features, labels in tfe.Iterator(dataset): # error in VsCode but works!!\n for features, labels in tf.contrib.eager.Iterator(dataset): \n grads = grad(logistic_regression(features), features, labels)\n optimizer.apply_gradients(grads_and_vars = zip(grads, [W, b]))\n if step % 100 == 0:\n print(\"Iter: {}, Loss: {:.4f}\".format(step, loss_fn(logistic_regression(features), features, labels)))\n\ntest_acc = accuracy_fn(logistic_regression(x_test), y_test)\nprint(\"Testset Accuracy: {:.4f}\".format(test_acc))\n", "sub_path": "03.Logistic_Classification.py", "file_name": "03.Logistic_Classification.py", "file_ext": "py", "file_size_in_byte": 2593, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "tensorflow.enable_eager_execution", "line_number": 6, "usage_type": "call"}, {"api_name": "tensorflow.set_random_seed", "line_number": 7, "usage_type": "call"}, {"api_name": "tensorflow.__version__", "line_number": 8, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.scatter", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 35, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 35, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 36, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 36, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 37, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 37, "usage_type": "name"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 40, "usage_type": "call"}, {"api_name": "tensorflow.data", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tensorflow.Variable", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 43, "usage_type": "call"}, {"api_name": "tensorflow.Variable", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.zeros", "line_number": 44, "usage_type": "call"}, {"api_name": "tensorflow.div", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.exp", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.matmul", "line_number": 48, "usage_type": "call"}, {"api_name": "tensorflow.reduce_mean", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.log", "line_number": 52, "usage_type": "call"}, {"api_name": "tensorflow.train.GradientDescentOptimizer", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.train", "line_number": 55, "usage_type": "attribute"}, {"api_name": "tensorflow.cast", "line_number": 58, "usage_type": "call"}, {"api_name": "tensorflow.float32", "line_number": 58, "usage_type": "attribute"}, {"api_name": "tensorflow.reduce_mean", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.cast", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.equal", "line_number": 59, "usage_type": "call"}, {"api_name": "tensorflow.int32", "line_number": 59, "usage_type": "attribute"}, {"api_name": "tensorflow.GradientTape", "line_number": 64, "usage_type": "call"}, {"api_name": "tensorflow.contrib.eager.Iterator", "line_number": 73, "usage_type": "call"}, {"api_name": "tensorflow.contrib", "line_number": 73, "usage_type": "attribute"}]} +{"seq_id": "251408743", "text": "\"\"\"this script finds all debtors and emails them a message.\"\"\"\n\nimport smtplib\nimport openpyxl\nimport os\n\nos.chdir('/Users/ilja/Dropbox/atbs/')\n\nwb = openpyxl.load_workbook('dues.xlsx')\ns = wb.active\n# print(s['A1'].value)\n\nnames = []\nemails = []\nfor row in range(2, s.max_row+1):\n if s['C'+str(row)].value != 'paid':\n email = s['B'+str(row)].value\n emails.append(email)\n name = s['A'+str(row)].value\n names.append(name)\n\nprint(names)\nprint(emails)\n\n# part 2 =======================================================================\ns = smtplib.SMTP('smtp.office365.com', 587)\ns.ehlo()\ns.starttls()\ns.login('x', 'x')\n\nfor name, email in zip(names, emails):\n body = f'Subject: Due date reminder. \\n\\nHey {name}! \\nThis is a gentle reminder that you\\'re past the due date on your payment. \\nHope youre well otherwise. \\nThanks.'\n status = s.sendmail('x', email, body)\n\n if status != {}:\n print(f'there was an error sending a message to {email} with {status} status')\ns.quit()\n", "sub_path": "18_pj_due_reminders.py", "file_name": "18_pj_due_reminders.py", "file_ext": "py", "file_size_in_byte": 1019, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "os.chdir", "line_number": 7, "usage_type": "call"}, {"api_name": "openpyxl.load_workbook", "line_number": 9, "usage_type": "call"}, {"api_name": "smtplib.SMTP", "line_number": 26, "usage_type": "call"}]} +{"seq_id": "79550551", "text": "import pandas_datareader.data as pdr\nimport pandas as pd\nfrom datetime import date, datetime\n\ndef sp500(years):\n \"\"\"\n Retrieves historical data from the coinmarketcap top 200 cryptocurrencies index.\n Does not return days the stockmarket is closed (May alter analysis but fine for initial purposes)\n Inputs:\n Years (an integer representing the number of years of historical data the user needs for analysis)\n Output:\n A dataframe with the adj_close price of the cmc200 index\n \"\"\"\n # Set and format date strings for the datareader call.\n end_date = date.today() - pd.Timedelta(days=1)\n start_date = (end_date - pd.Timedelta(weeks=52*(int(years))))\n\n # Using pandas datareader we get data for the cmc200 index\n df = pdr.DataReader('^GSPC', 'yahoo', start=str(start_date), end=str(end_date))\n df = df.drop(columns=['High', 'Low', 'Open', 'Volume', 'Close'])\n df = df.rename(columns={'Adj Close':'Close'})\n\n return df", "sub_path": "data_retrieval/sp500_index.py", "file_name": "sp500_index.py", "file_ext": "py", "file_size_in_byte": 968, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "datetime.date.today", "line_number": 15, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 15, "usage_type": "name"}, {"api_name": "pandas.Timedelta", "line_number": 15, "usage_type": "call"}, {"api_name": "pandas.Timedelta", "line_number": 16, "usage_type": "call"}, {"api_name": "pandas_datareader.data.DataReader", "line_number": 19, "usage_type": "call"}, {"api_name": "pandas_datareader.data", "line_number": 19, "usage_type": "name"}]} +{"seq_id": "171320431", "text": "#!/usr/bin/env python3.3\nfrom string import Template\nimport datetime\n\nnamespaces = {'HGNC': 'hgnc-human-genes.belns',\n 'MGI': 'mgi-mouse-genes.belns',\n 'RGD': 'rgd-rat-genes.belns'}\nannotations = {}\nbase_url = 'http://resources.openbel.org/belframework/latest/'\ntoday = datetime.date.today()\nversion = today.strftime('%Y%m%d')\n\n\ndef bel_term(value, ns, f):\n \"\"\" Create bel term given value, namespace id,\n and bel function string. \"\"\"\n must_quote_values = ['a', 'SET']\n must_quote_chars = [':', '(', ')', '<', '>', '.', '-', '/', '@', ' ']\n if any(char in value for char in must_quote_chars) or value in must_quote_values:\n s = Template('${f}(${ns}:\"${value}\")')\n else:\n s = Template('${f}(${ns}:${value})')\n term = s.substitute(f=f, ns=ns, value=value)\n return term\n\n\ndef write_bel_header(\n f,\n *,\n authors='OpenBEL',\n contact_info=None,\n doc_name=\"Document Name\",\n description=None,\n licenses=None,\n version=version,\n namespaces=namespaces,\n annotations=annotations,\n base_url=base_url):\n '''Write BEL document header to file object f, given a document name,\n description, namespace dictionary (prefixes to urls) and annotation\n dictionary.'''\n separator = 50 * '#' + '\\n'\n f.write(separator + '# Document Properties Section\\n')\n f.write('SET DOCUMENT Name = \"{0}\"\\n'.format(doc_name))\n if description:\n f.write('SET DOCUMENT Description = \"{0}\"\\n'.format(description))\n f.write('SET DOCUMENT Version = \"{0}\"\\n'.format(version))\n f.write('SET DOCUMENT Copyright = \"Copyright (c) {0}, OpenBEL Project. This work is licensed under a Creative Commons Attribution 3.0 Unported License.\"\\n'.format(\n str(today.year)))\n f.write('SET DOCUMENT Authors = {0}\\n'.format(authors))\n if licenses:\n f.write('SET DOCUMENT Licenses = {0}\\n'.format(licenses))\n if contact_info:\n f.write('SET Document ContactInfo = {0}\\n'.format(contact_info))\n f.write('\\n' + separator + '# Definitions Section\\n')\n for ns_prefix, ns_name in namespaces.items():\n base_url = base_url.rstrip('/')\n f.write(\n 'DEFINE NAMESPACE {0} AS URL \"{1}/namespace/{2}\"\\n'.format(ns_prefix, base_url, ns_name))\n f.write('\\n')\n for anno, url in annotations.items():\n f.write(\n 'DEFINE ANNOTATION {0} AS URL \"{1}/annotation/{2}\"\\n'.format(anno, base_url, url))\n f.write(separator)\n f.write('# Statements Section\\n\\n')\n return None\n\n", "sub_path": "bel_functions.py", "file_name": "bel_functions.py", "file_ext": "py", "file_size_in_byte": 2554, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "datetime.date.today", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 10, "usage_type": "attribute"}, {"api_name": "string.Template", "line_number": 20, "usage_type": "call"}, {"api_name": "string.Template", "line_number": 22, "usage_type": "call"}]} +{"seq_id": "154388529", "text": "# Copyright 2019, Hudson and Thames Quantitative Research\r\n# All rights reserved\r\n# Read more: https://github.com/hudson-and-thames/mlfinlab/blob/master/LICENSE.txt\r\n\r\n\"\"\"\r\nVolume classification methods (BVC and tick rule).\r\n\"\"\"\r\n# pylint: disable=invalid-name\r\n\r\nimport pandas as pd\r\nfrom scipy.stats import norm\r\nfrom scipy.stats import t\r\n\r\nfrom mlfinlab.util import devadarsh\r\n\r\n\r\ndef get_bvc_buy_volume(close: pd.Series, volume: pd.Series, window: int = 20, distribution: str = 'norm',\r\n df: float = 0.25) -> pd.Series:\r\n \"\"\"\r\n Calculates the BVC buy volume.\r\n\r\n :param close: (pd.Series): Close prices.\r\n :param volume: (pd.Series): Bar volumes.\r\n :param window: (int): Window for std estimation uses in BVC calculation.\r\n :param distribution: (str): Distribution function used to estimate. Either 'norm' or 't_student'.\r\n :param df: (float) If `distribution` = 't_student', df is a number of degrees of freedom.\r\n Common used values are: 0.1, 0.25.\r\n :return: (pd.Series) BVC buy volume.\r\n \"\"\"\r\n\r\n # .apply(norm.cdf) is used to omit Warning for norm.cdf(pd.Series with NaNs)\r\n devadarsh.track('get_bvc_buy_volume')\r\n\r\n if distribution == 'norm':\r\n buy_volume_frac = volume * (close.diff() / close.diff().rolling(window=window).std()).apply(norm.cdf)\r\n elif distribution == 't_student':\r\n buy_volume_frac = volume * (close.diff() / close.diff().rolling(window=window).std()).apply(\r\n lambda x: t.cdf(x, df=df))\r\n else:\r\n raise ValueError('Unknown value for `distribution`: use either `norm` or `t_student`')\r\n\r\n return buy_volume_frac\r\n", "sub_path": "src/collection/mlfinlab/util/volume_classifier.py", "file_name": "volume_classifier.py", "file_ext": "py", "file_size_in_byte": 1648, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pandas.Series", "line_number": 17, "usage_type": "attribute"}, {"api_name": "mlfinlab.util.devadarsh.track", "line_number": 32, "usage_type": "call"}, {"api_name": "mlfinlab.util.devadarsh", "line_number": 32, "usage_type": "name"}, {"api_name": "scipy.stats.norm.cdf", "line_number": 35, "usage_type": "attribute"}, {"api_name": "scipy.stats.norm", "line_number": 35, "usage_type": "name"}, {"api_name": "scipy.stats.t.cdf", "line_number": 38, "usage_type": "call"}, {"api_name": "scipy.stats.t", "line_number": 38, "usage_type": "name"}, {"api_name": "pandas.Series", "line_number": 18, "usage_type": "attribute"}]} +{"seq_id": "98905480", "text": "\nfrom kernel.operation import Operation\nimport logging\nimport os\nimport minibam_transfer\nimport shutil\nimport argparse\n\nclass SyncFiles(Operation):\n\n @staticmethod\n def name():\n return \"sync\"\n\n @staticmethod\n def description():\n return \"Synchronize mini-bam file transfer on JTracker with Github Repo\"\n\n def _parser(self, main_parser):\n main_parser.add_argument('-p', '--payload', dest='payload', required=True,\n type=argparse.FileType('r'), help=\"Song payload json file\")\n\n def _schema(self):\n return {\n \"jtracker_host\": {\"type\": \"string\"},\n \"jtracker_user\": {\"type\": \"string\"},\n \"queues\": {\n \"type\": \"array\",\n \"items\": {\"type\": \"string\"}\n },\n \"git_dirs\":{\n \"type\": \"object\",\n \"properties\": {\n \"queued\": {\"type\":\"string\"},\n \"failed\": {\"type\":\"string\"},\n \"completed\": {\"type\":\"string\"},\n \"running\": {\"type\":\"string\"},\n \"backlog\": {\"type\":\"string\"},\n \"resume\": {\"type\": \"string\"}\n },\n \"required\": [\"queued\",\"failed\",\"completed\",\"running\",\"backlog\"]\n },\n \"required\": [\"jtracker_host\",\"jtracker_user\",\"git_dirs\",\"queues\"]\n }\n\n def _run(self):\n logging.info(\"Sync mini-bam files with git repository\")\n jobname_state = minibam_transfer.get_jobnames_state(self._config.get('jtracker_host'),self._config.get('jtracker_user'),self._config.get('queues'))\n\n for job in jobname_state:\n for state in self._config.get('git_dirs'):\n for file in os.listdir(self._config.get('git_dirs')[state]):\n if job in file:\n found = True\n if state == jobname_state[job]:\n break\n else:\n self.mv_job_to_state(os.path.join(self._config.get('git_dirs')[state],file),file,jobname_state[job],self._config.get('git_dirs')[jobname_state[job]])\n\n def mv_job_to_state(self, file_path, file_name, state, dir_path):\n final_path = os.path.join(dir_path,file_name)\n if state == 'completed':\n final_path = os.path.join(dir_path,file_name.strip('.json'))\n if not os.path.isdir(final_path):\n os.mkdir(final_path)\n final_path = os.path.join(final_path,file_name)\n\n shutil.move(file_path,final_path)", "sub_path": "operations/minibam/sync_files/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2595, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "kernel.operation.Operation", "line_number": 9, "usage_type": "name"}, {"api_name": "argparse.FileType", "line_number": 21, "usage_type": "call"}, {"api_name": "logging.info", "line_number": 47, "usage_type": "call"}, {"api_name": "minibam_transfer.get_jobnames_state", "line_number": 48, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 52, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 58, "usage_type": "call"}, {"api_name": "os.path", "line_number": 58, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 61, "usage_type": "call"}, {"api_name": "os.path", "line_number": 61, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 63, "usage_type": "call"}, {"api_name": "os.path", "line_number": 63, "usage_type": "attribute"}, {"api_name": "os.path.isdir", "line_number": 64, "usage_type": "call"}, {"api_name": "os.path", "line_number": 64, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 66, "usage_type": "call"}, {"api_name": "os.path", "line_number": 66, "usage_type": "attribute"}, {"api_name": "shutil.move", "line_number": 68, "usage_type": "call"}]} +{"seq_id": "55312757", "text": "import bisect\nimport subprocess\nimport numpy as np\nimport time\nimport os\ntry:\n import ujson as json\nexcept:\n import json\nfrom functools import partial\nfrom itertools import chain\nfrom pprint import pprint\nfrom copy import deepcopy\nfrom aser.extract.discourse_parser import ConnectiveExtractor, ArgumentPositionClassifier, \\\n SSArgumentExtractor, PSArgumentExtractor, ExplicitSenseClassifier\nfrom aser.eventuality import Eventuality\nfrom aser.relation import Relation, relation_senses\nfrom aser.extract.rule import SEED_CONNECTIVE_DICT\nfrom aser.extract.utils import EMPTY_SENT_PARSED_RESULT\n\n\nclass BaseRelationExtractor(object):\n def __init__(self, **kw):\n pass\n\n def close(self):\n pass\n\n def __del__(self):\n self.close()\n\n def extract_from_parsed_result(self, parsed_result, para_eventualities, output_format=\"Relation\", in_order=True, **kw):\n \"\"\" This method extracts relations among extracted eventualities.\n\n :type parsed_result: list\n :type para_eventualities: list\n :type output_format: str\n :type in_order bool\n :param parsed_result: a list of dicts generated by `aser.extract.utils.parse_sentense_with_stanford`\n :param para_eventualities: a list of lists of `Eventuality` objects\n :param output_format: the specific output format\n :param in_order: in order or out of order\n :return: a list of lists of `Relation` objects, or a list of lists of triples\n\n .. highlight:: python\n .. code-block:: python\n\n Input:\n [\n {'text': 'The dog barks loudly because it is hungry.',\n 'dependencies': [(1, 'det', 0),\n (2, 'nsubj', 1),\n (2, 'advmod', 3),\n (2, 'punct', 8),\n (3, 'dep', 7),\n (7, 'mark', 4),\n (7, 'nsubj', 5),\n (7, 'cop', 6)],\n 'tokens': ['The', 'dog', 'barks', 'loudly', 'because', 'it', 'is', 'hungry', '.'],\n 'pos_tags': ['DT', 'NN', 'VBZ', 'RB', 'IN', 'PRP', 'VBZ', 'JJ', '.'],\n 'lemmas': ['the', 'dog', 'bark', 'loudly', 'because', 'it', 'be', 'hungry', '.'],\n 'ners': ['O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O'],\n 'mentions': [],\n 'parse': '(ROOT (S (NP (DT The) (NN dog)) (VP (VBZ barks) (ADVP (RB loudly) (SBAR (IN because) (S (NP (PRP it)) (VP (VBZ is) (ADJP (JJ hungry))))))) (. .)))'},\n {'text': 'The dog barks loudly because',\n 'dependencies': [(4, 'cc', 0),\n (4, 'nsubj', 1),\n (4, 'aux', 2),\n (4, 'neg', 3),\n (4, 'dobj', 6),\n (4, 'punct', 7),\n (6, 'compound', 5)],\n 'tokens': ['But', 'we', 'do', \"n't\", 'have', 'food', 'left', '.'],\n 'pos_tags': ['CC', 'PRP', 'VBP', 'RB', 'VB', 'NN', 'NN', '.'],\n 'lemmas': ['but', 'we', 'do', 'not', 'have', 'food', 'left', '.'],\n 'ners': ['O', 'O', 'O', 'O', 'O', 'O', 'O', 'O'],\n 'mentions': [],\n 'parse': \"(ROOT (S (CC But) (NP (PRP we)) (VP (VBP do) (RB n't) (VP (VB have) (NP (NN food) (NN left)))) (. .)))\"}],\n [\n [\n Eventuality(\n {'dependencies': [((1, 'dog', 'NN'), 'det', (0, 'the', 'DT')),\n ((2, 'bark', 'VBZ'), 'nsubj', (1, 'dog', 'NN')),\n ((2, 'bark', 'VBZ'), 'advmod', (3, 'loudly', 'RB'))],\n 'eid': 'b51425727182a0d25734a92ae16a456cb5e6351f',\n 'frequency': 1.0,\n 'mentions': {},\n 'ners': ['O', 'O', 'O', 'O'],\n 'pattern': 's-v',\n 'pos_tags': ['DT', 'NN', 'VBZ', 'RB'],\n 'skeleton_dependencies': [((2, 'bark', 'VBZ'), 'nsubj', (1, 'dog', 'NN'))],\n 'skeleton_words': ['dog', 'bark'],\n 'verbs': ['bark'],\n 'words': ['the', 'dog', 'bark', 'loudly']}),\n Eventuality(\n {'dependencies': [((2, 'hungry', 'JJ'), 'nsubj', (0, 'it', 'PRP')),\n ((2, 'hungry', 'JJ'), 'cop', (1, 'be', 'VBZ'))],\n 'eid': '8fbd35fcb293f526b54c5989969251d6a31e4893',\n 'frequency': 1.0,\n 'mentions': {},\n 'ners': ['O', 'O', 'O'],\n 'pattern': 's-be-a',\n 'pos_tags': ['PRP', 'VBZ', 'JJ'],\n 'skeleton_dependencies': [((2, 'hungry', 'JJ'), 'nsubj', (0, 'it', 'PRP')),\n ((2, 'hungry', 'JJ'), 'cop', (1, 'be', 'VBZ'))],\n 'skeleton_words': ['it', 'be', 'hungry'],\n 'verbs': ['be'],\n 'words': ['it', 'be', 'hungry']})],\n [\n Eventuality(\n {'dependencies': [((3, 'have', 'VB'), 'nsubj', (0, 'we', 'PRP')),\n ((3, 'have', 'VB'), 'aux', (1, 'do', 'VBP')),\n ((3, 'have', 'VB'), 'neg', (2, 'not', 'RB')),\n ((3, 'have', 'VB'), 'dobj', (5, 'left', 'NN')),\n ((5, 'left', 'NN'), 'compound', (4, 'food', 'NN'))],\n 'eid': '32bd10b7e116f7656b7424d3f3a47dab230d52de',\n 'frequency': 1.0,\n 'mentions': {},\n 'ners': ['O', 'O', 'O', 'O', 'O', 'O'],\n 'pattern': 's-v-o',\n 'pos_tags': ['PRP', 'VBP', 'RB', 'VB', 'NN', 'NN'],\n 'skeleton_dependencies': [((3, 'have', 'VB'), 'nsubj', (0, 'we', 'PRP')),\n ((3, 'have', 'VB'), 'dobj', (5, 'left', 'NN'))],\n 'skeleton_words': ['we', 'have', 'left'],\n 'verbs': ['do', 'have'],\n 'words': ['we', 'do', 'not', 'have', 'food', 'left']})]],\n \"Relation\",\n True\n Output:\n [\n [\n Relation(\n {'rid': 'fcdfcabe07446e8e0ba7950016a300ce00d1e4b9',\n 'hid': '8fbd35fcb293f526b54c5989969251d6a31e4893',\n 'tid': 'b51425727182a0d25734a92ae16a456cb5e6351f'}\n 'relations': \"{'Co_Occurrence': 1.0}\")],\n [],\n [\n Relation(\n {'rid': 'd75f64002dc742fdbadbc0465dbde50c22facf67',\n 'hid': '8fbd35fcb293f526b54c5989969251d6a31e4893',\n 'tid': '32bd10b7e116f7656b7424d3f3a47dab230d52de',\n 'relations': \"{'Contrast': 1.0}\",}),\n Relation(\n {'rid': '53edca79788c6df43212c034353305061c2619d3',\n 'hid': 'b51425727182a0d25734a92ae16a456cb5e6351f',\n 'tid': '32bd10b7e116f7656b7424d3f3a47dab230d52de',\n 'relations': \"{'Contrast': 1.0}\",})]]\n \"\"\"\n raise NotImplementedError\n\nclass SeedRuleRelationExtractor(BaseRelationExtractor):\n def __init__(self, **kw):\n super().__init__(**kw)\n\n def extract_from_parsed_result(self, parsed_result, para_eventualities, output_format=\"Relation\", in_order=True, **kw):\n if output_format not in [\"Relation\", \"triple\"]:\n raise NotImplementedError(\"Error: extract_from_parsed_result only supports Relation or triple.\")\n\n connective_dict = kw.get(\"connective_dict\", SEED_CONNECTIVE_DICT)\n\n para_relations = list()\n for sent_parsed_result, eventualities in zip(parsed_result, para_eventualities):\n relations_in_sent = list()\n for head_eventuality in eventualities:\n for tail_eventuality in eventualities:\n if not head_eventuality.position < tail_eventuality.position:\n continue\n heid = head_eventuality.eid\n teid = tail_eventuality.eid\n extracted_senses = self._extract_from_eventuality_pair_in_one_sentence(\n connective_dict, sent_parsed_result, head_eventuality, tail_eventuality)\n if len(extracted_senses) > 0:\n relations_in_sent.append(Relation(heid, teid, extracted_senses))\n para_relations.append(relations_in_sent)\n\n for i in range(len(parsed_result) - 1):\n eventualities1, eventualities2 = para_eventualities[i], para_eventualities[i+1]\n relations_between_sents = list()\n if len(eventualities1) == 1 and len(eventualities2) == 1:\n s1_tokens, s2_tokens = parsed_result[i][\"tokens\"], parsed_result[i+1][\"tokens\"]\n s1_eventuality, s2_eventuality = eventualities1[0], eventualities2[0]\n heid, teid = s1_eventuality.eid, s2_eventuality.eid\n extracted_senses = self._extract_from_eventuality_pair_in_two_sentence(\n connective_dict, s1_eventuality, s2_eventuality, s1_tokens, s2_tokens)\n if len(extracted_senses) > 0:\n relations_between_sents.append(Relation(heid, teid, extracted_senses))\n para_relations.append(relations_between_sents)\n\n if in_order:\n if output_format == \"Relation\":\n return para_relations\n elif output_format == \"triple\":\n return [sorted(chain.from_iterable([r.to_triples() for r in relations])) \\\n for relations in para_relations]\n else:\n if output_format == \"Relation\":\n rid2relation = dict()\n for relation in chain(*para_relations):\n if relation.rid not in rid2relation:\n rid2relation[relation.rid] = deeocopy(relation)\n else:\n rid2relation[relation.rid].update(relation)\n return sorted(rid2relation.values(), key=lambda r: r.rid)\n if output_format == \"triple\":\n return sorted([r.to_triples() for relations in para_relations for r in relations])\n \n\n def _extract_from_eventuality_pair_in_one_sentence(self, connective_dict, sent_parsed_result, head_eventuality, tail_eventuality):\n extracted_senses = ['Co_Occurrence']\n for sense in relation_senses:\n for connective_words in connective_dict[sense]:\n if self._verify_connective_in_one_sentence(\n connective_words, head_eventuality, tail_eventuality,\n sent_parsed_result[\"dependencies\"],\n sent_parsed_result[\"tokens\"]):\n extracted_senses.append(sense)\n break\n return extracted_senses\n\n def _extract_from_eventuality_pair_in_two_sentence(self, connective_dict, s1_eventuality, s2_eventuality, s1_tokens, s2_tokens):\n extracted_senses = list()\n for sense in relation_senses:\n for connective_words in connective_dict[sense]:\n if self._verify_connective_in_two_sentence(connective_words, s1_eventuality, s2_eventuality, s1_tokens, s2_tokens):\n extracted_senses.append(sense)\n break\n\n return extracted_senses\n\n def _verify_connective_in_one_sentence(self, connective_words, head_eventuality, tail_eventuality, sentence_dependencies, sentence_tokens):\n def get_connective_position(connective_words):\n tmp_positions = list()\n for w in connective_words:\n tmp_positions.append(sentence_tokens.index(w))\n return sum(tmp_positions) / len(tmp_positions) if tmp_positions else 0.0\n # Connective Words need to be presented in sentence\n if set(connective_words) - set(sentence_tokens):\n return False\n # Connective phrase need to be presented in sentence\n connective_string = \" \".join(connective_words)\n sentence_string = \" \".join(sentence_tokens)\n if connective_string not in sentence_string:\n return False\n shrinked_dependencies = self._shrink_sentence_dependencies(\n head_eventuality._raw_dependencies,\n tail_eventuality._raw_dependencies,\n sentence_dependencies)\n if not shrinked_dependencies:\n return False\n found_advcl = False\n for (governor, dep, dependent) in shrinked_dependencies:\n if governor == '_H_' and dependent == '_T_' and 'advcl' in dep:\n found_advcl = True\n break\n if not found_advcl:\n return False\n connective_position = get_connective_position(connective_words)\n e1_position, e2_position = head_eventuality.position, tail_eventuality.position\n if 'instead' not in connective_words:\n if e1_position < connective_position < e2_position:\n return True\n else:\n return False\n else:\n if e1_position < e2_position < connective_position:\n return True\n else:\n return False\n\n\n def _verify_connective_in_two_sentence(self, connective_words, s1_eventuality, s2_eventuality, s1_tokens, s2_tokens):\n def get_connective_position():\n tmp_positions = list()\n for w in connective_words:\n if w in s1_tokens:\n tmp_positions.append(s1_tokens.index(w))\n elif w in s2_tokens:\n tmp_positions.append(s2_tokens.index(w) + len(s1_tokens))\n return sum(tmp_positions) / len(tmp_positions) if tmp_positions else 0.0\n sentence_tokens = s1_tokens + s2_tokens\n # Connective Words need to be presented in sentence\n if set(connective_words) - set(sentence_tokens):\n return False\n # Connective phrase need to be presented in sentence\n connective_string = \" \".join(connective_words)\n sentence_string = \" \".join(sentence_tokens)\n if connective_string not in sentence_string:\n return False\n connective_position = get_connective_position()\n e1_position, e2_position = s1_eventuality.position, \\\n s2_eventuality.position + len(s1_tokens)\n if 'instead' not in connective_words:\n if e1_position < connective_position < e2_position and e2_position - e1_position < 10:\n return True\n else:\n return False\n else:\n if e1_position < e2_position < connective_position and e2_position - e1_position < 10:\n return True\n else:\n return False\n\n\n def _shrink_sentence_dependencies(self, head_dependencies, tail_dependencies,\n sentence_dependencies):\n head_nodes = set()\n for governor, _, dependent in head_dependencies:\n head_nodes.add(governor)\n head_nodes.add(dependent)\n tail_nodes = set()\n for governor, _, dependent in tail_dependencies:\n tail_nodes.add(governor)\n tail_nodes.add(dependent)\n if head_nodes & tail_nodes:\n return None\n\n new_dependencies = list()\n for governor, dep, dependent in sentence_dependencies:\n if governor in head_nodes:\n new_governor = '_H_'\n elif governor in tail_nodes:\n new_governor = '_T_'\n else:\n new_governor = governor\n if dependent in head_nodes:\n new_dependent = '_H_'\n elif dependent in tail_nodes:\n new_dependent = '_T_'\n else:\n new_dependent = dependent\n if new_governor != new_dependent:\n new_dependencies.append((new_governor, dep, new_dependent))\n return new_dependencies\n\n\nclass DiscourseRelationExtractor(BaseRelationExtractor):\n def __init__(self, **kw):\n super().__init__(**kw)\n self.conn_extractor = ConnectiveExtractor(**kw)\n self.argpos_classifier = ArgumentPositionClassifier(**kw)\n self.ss_extractor = SSArgumentExtractor(**kw)\n self.ps_extractor = PSArgumentExtractor(**kw)\n self.explicit_classifier = ExplicitSenseClassifier(**kw)\n\n def extract_from_parsed_result(self, parsed_result, para_eventualities, output_format=\"triple\", in_order=False, **kw):\n if output_format not in [\"Relation\", \"triple\"]:\n raise NotImplementedError(\"Error: extract_from_parsed_result only supports Relation or triple.\")\n \n similarity = kw.get(\"similarity\", \"simpson\").lower()\n threshold = kw.get(\"threshold\", 0.8)\n if threshold < 0.0 or threshold > 1.0:\n raise ValueError(\"Error: threshold should be between 0.0 and 1.0.\")\n if similarity == \"simpson\":\n similarity_func = self._match_argument_eventuality_by_Simpson\n elif similarity == \"jaccard\":\n similarity_func = self._match_argument_eventuality_by_Jaccard\n elif similarity == \"discourse\":\n similarity_func = self._match_argument_eventuality_by_dependencies\n else:\n raise NotImplementedError(\"Error: extract_from_parsed_result only supports Simpson or Jaccard.\")\n\n syntax_tree_cache = kw.get(\"syntax_tree_cache\", dict())\n\n len_sentences = len(parsed_result)\n if len_sentences == 0:\n if in_order:\n return [list()]\n else:\n return list()\n\n para_relations = [list() for _ in range(2*len_sentences-1)]\n\n # replace sentences that contains no eventuality with empty sentences\n filtered_parsed_result = list()\n for sent_idx, (sent_parsed_result, sent_eventualities) in enumerate(zip(parsed_result, para_eventualities)):\n if len(sent_eventualities) > 0:\n filtered_parsed_result.append(sent_parsed_result)\n relations_in_sent = para_relations[sent_idx]\n for head_e in sent_eventualities:\n heid = head_e.eid\n for tail_e in sent_eventualities:\n if not head_e.position < tail_e.position:\n continue\n teid = tail_e.eid\n relations_in_sent.append(Relation(heid, teid, [\"Co_Occurrence\"]))\n else:\n filtered_parsed_result.append(EMPTY_SENT_PARSED_RESULT) # empty sentence\n # filtered_parsed_result.append(sent_parsed_result)\n\n connectives = self.conn_extractor.extract(filtered_parsed_result, syntax_tree_cache)\n SS_connectives, PS_connectives = self.argpos_classifier.classify(filtered_parsed_result, connectives, syntax_tree_cache)\n SS_connectives = self.ss_extractor.extract(filtered_parsed_result, SS_connectives, syntax_tree_cache)\n PS_connectives = self.ps_extractor.extract(filtered_parsed_result, PS_connectives, syntax_tree_cache)\n connectives = self.explicit_classifier.classify(filtered_parsed_result, SS_connectives+PS_connectives, syntax_tree_cache)\n connectives.sort(key=lambda x: (x[\"sent_idx\"], x[\"indices\"][0] if len(x[\"indices\"]) > 0 else -1))\n \n # For CoNLL share task 2015\n # sent_offset = 0\n # for sent_parsed_result in parsed_result:\n # sent_parsed_result[\"sentence_offset\"] = sent_offset\n # sent_offset += len(sent_parsed_result[\"tokens\"])\n # with open(\"aser.json\", \"a\") as f:\n # for conn_idx, connective in enumerate(connectives):\n # sense = connective.get(\"sense\", None)\n # arg1 = connective.get(\"arg1\", None)\n # arg2 = connective.get(\"arg2\", None)\n # if arg1 and arg2 and sense and sense != \"None\":\n # x = {\n # \"DocID\": parsed_result[0][\"doc\"], \n # \"ID\": conn_idx, \n # \"Connective\": {\n # \"RawText\": connective[\"connective\"],\n # \"TokenList\": [i+parsed_result[connective[\"sent_idx\"]][\"sentence_offset\"] for i in connective[\"indices\"]],\n # \"Tokens\": [parsed_result[connective[\"sent_idx\"]][\"tokens\"][i] for i in connective[\"indices\"]]},\n # \"Arg1\": {\n # \"RawText\": \" \".join([parsed_result[arg1[\"sent_idx\"]][\"tokens\"][i] for i in arg1[\"indices\"]]),\n # \"TokenList\": [i+parsed_result[arg1[\"sent_idx\"]][\"sentence_offset\"] for i in arg1[\"indices\"]],\n # \"Tokens\": [parsed_result[arg1[\"sent_idx\"]][\"tokens\"][i] for i in arg1[\"indices\"]]},\n # \"Arg2\": {\n # \"RawText\": \" \".join([parsed_result[arg2[\"sent_idx\"]][\"tokens\"][i] for i in arg2[\"indices\"]]),\n # \"TokenList\": [i+parsed_result[arg2[\"sent_idx\"]][\"sentence_offset\"] for i in arg2[\"indices\"]],\n # \"Tokens\": [parsed_result[arg2[\"sent_idx\"]][\"tokens\"][i] for i in arg2[\"indices\"]]},\n # \"Type\": \"Explicit\",\n # \"Sense\": [connective[\"sense\"]]}\n # f.write(json.dumps(x))\n # f.write(\"\\n\")\n\n for connective in connectives:\n conn_indices = connective.get(\"indices\", None)\n arg1 = connective.get(\"arg1\", None)\n arg2 = connective.get(\"arg2\", None)\n sense = connective.get(\"sense\", None)\n if conn_indices and arg1 and arg2 and (sense and sense != \"None\"):\n arg1_sent_idx = arg1[\"sent_idx\"]\n arg2_sent_idx = arg2[\"sent_idx\"]\n relation_list_idx = arg1_sent_idx if arg1_sent_idx == arg2_sent_idx else arg1_sent_idx + len_sentences\n relations = para_relations[relation_list_idx]\n sent_parsed_result1, sent_eventualities1 = parsed_result[arg1_sent_idx], para_eventualities[arg1_sent_idx]\n sent_parsed_result2, sent_eventualities2 = parsed_result[arg2_sent_idx], para_eventualities[arg2_sent_idx]\n arg1_eventualities = [e for e in sent_eventualities1 if \\\n similarity_func(sent_parsed_result1, arg1, e, threshold=threshold, conn_indices=conn_indices)]\n arg2_eventualities = [e for e in sent_eventualities2 if \\\n similarity_func(sent_parsed_result2, arg2, e, threshold=threshold, conn_indices=conn_indices)]\n cnt = 0.0\n if len(arg1_eventualities) > 0 and len(arg2_eventualities) > 0:\n cnt = 1.0 / (len(arg1_eventualities) * len(arg2_eventualities))\n for e1 in arg1_eventualities:\n heid = e1.eid\n for e2 in arg2_eventualities:\n teid = e2.eid\n is_existed = False\n for relation in relations:\n if relation.hid == heid and relation.tid == teid:\n relation.update({sense: cnt})\n is_existed = True\n break\n if not is_existed:\n relations.append(Relation(heid, teid, {sense: cnt}))\n\n if in_order:\n if output_format == \"Relation\":\n return para_relations\n elif output_format == \"triple\":\n return [sorted(chain.from_iterable([r.to_triples() for r in relations])) \\\n for relations in para_relations]\n else:\n if output_format == \"Relation\":\n rid2relation = dict()\n for relation in chain(*para_relations):\n if relation.rid not in rid2relation:\n rid2relation[relation.rid] = deeocopy(relation)\n else:\n rid2relation[relation.rid].update(relation)\n return sorted(rid2relation.values(), key=lambda r: r.rid)\n if output_format == \"triple\":\n return sorted([r.to_triples() for relations in para_relations for r in relations])\n\n @staticmethod\n def _match_argument_eventuality_by_Simpson(sent_parsed_result, argument, eventuality, **kw):\n threshold = kw.get(\"threshold\", 0.8)\n match = False\n if eventuality.raw_sent_mapping:\n argument_indices = set(argument[\"indices\"])\n event_indices = set(eventuality.raw_sent_mapping.values())\n try:\n Simpson = len(argument_indices & event_indices) / min(len(argument_indices), len(event_indices))\n match = Simpson >= threshold\n except ZeroDivisionError:\n match = False\n else:\n argument_tokens = set([sent_parsed_result[\"lemmas\"][idx].lower() for idx in argument[\"indices\"]])\n event_tokens = set(eventuality.words)\n try:\n Simpson = len(argument_tokens & event_tokens) / min(len(argument_tokens), len(event_tokens))\n match = Simpson >= threshold\n except ZeroDivisionError:\n match = False\n return match\n \n @staticmethod\n def _match_argument_eventuality_by_Jaccard(sent_parsed_result, argument, eventuality, **kw):\n threshold = kw.get(\"threshold\", 0.8)\n match = False\n if eventuality.raw_sent_mapping:\n argument_indices = set(argument[\"indices\"])\n event_indices = set(eventuality.raw_sent_mapping.values())\n try:\n Jaccard = len(argument_indices & event_indices) / len(argument_indices | event_indices)\n match = Jaccard >= threshold\n except ZeroDivisionError:\n match = False\n else:\n argument_tokens = set([sent_parsed_result[\"lemmas\"][idx].lower() for idx in argument[\"indices\"]])\n event_tokens = set(eventuality.words)\n try:\n Jaccard = len(argument_tokens & event_tokens) / len(argument_tokens | event_tokens)\n match = Jaccard >= threshold\n except ZeroDivisionError:\n match = False\n return match\n\n @staticmethod\n def _match_argument_eventuality_by_dependencies(sent_parsed_result, argument, eventuality, **kw):\n conn_indices = kw.get(\"conn_indices\", list())\n match = False\n conn_indices = set(conn_indices)\n if eventuality.raw_sent_mapping:\n argument_indices = set(argument[\"indices\"])\n event_indices = set(eventuality.raw_sent_mapping.values())\n\n for (governor, dep, dependent) in sent_parsed_result[\"dependencies\"]:\n # find the word linked to the connective\n if dependent in conn_indices and governor in argument_indices and governor in event_indices:\n match = True\n break\n elif governor in conn_indices and dependent in argument_indices and dependent in event_indices:\n match = True\n break\n else:\n argument_tokens = set([sent_parsed_result[\"lemmas\"][idx].lower() for idx in argument[\"indices\"]])\n event_token_pos_tags = set(zip(eventuality.words, eventuality.pos_tags))\n\n argument_indices = set(argument[\"indices\"])\n for (governor, dep, dependent) in sent_parsed_result[\"dependencies\"]:\n # find the word linked to the connective\n if dependent in conn_indices and governor in argument_indices:\n token_pos_tag = (sent_parsed_result[\"lemmas\"][governor].lower(), sent_parsed_result[\"pos_tags\"][governor])\n if token_pos_tag in event_token_pos_tags:\n match = True\n break\n elif governor in conn_indices and dependent in argument_indices:\n token_pos_tag = (sent_parsed_result[\"lemmas\"][dependent].lower(), sent_parsed_result[\"pos_tags\"][dependent])\n if token_pos_tag in event_token_pos_tags:\n match = True\n break\n return match\n \n\nclass NeuralRelationExtractor(BaseRelationExtractor):\n def __init__(self, **kw):\n super().__init__(kw)\n raise NotImplementedError\n\n def extract(self, eventualtity_pair):\n raise NotImplementedError", "sub_path": "aser/extract/relation_extractor.py", "file_name": "relation_extractor.py", "file_ext": "py", "file_size_in_byte": 29555, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "aser.extract.rule.SEED_CONNECTIVE_DICT", "line_number": 160, "usage_type": "argument"}, {"api_name": "aser.relation.Relation", "line_number": 174, "usage_type": "call"}, {"api_name": "aser.relation.Relation", "line_number": 187, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 194, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 194, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 199, "usage_type": "call"}, {"api_name": "aser.relation.relation_senses", "line_number": 211, "usage_type": "name"}, {"api_name": "aser.relation.relation_senses", "line_number": 223, "usage_type": "name"}, {"api_name": "aser.extract.discourse_parser.ConnectiveExtractor", "line_number": 340, "usage_type": "call"}, {"api_name": "aser.extract.discourse_parser.ArgumentPositionClassifier", "line_number": 341, "usage_type": "call"}, {"api_name": "aser.extract.discourse_parser.SSArgumentExtractor", "line_number": 342, "usage_type": "call"}, {"api_name": "aser.extract.discourse_parser.PSArgumentExtractor", "line_number": 343, "usage_type": "call"}, {"api_name": "aser.extract.discourse_parser.ExplicitSenseClassifier", "line_number": 344, "usage_type": "call"}, {"api_name": "aser.relation.Relation", "line_number": 386, "usage_type": "call"}, {"api_name": "aser.extract.utils.EMPTY_SENT_PARSED_RESULT", "line_number": 388, "usage_type": "argument"}, {"api_name": "aser.relation.Relation", "line_number": 459, "usage_type": "call"}, {"api_name": "itertools.chain.from_iterable", "line_number": 465, "usage_type": "call"}, {"api_name": "itertools.chain", "line_number": 465, "usage_type": "name"}, {"api_name": "itertools.chain", "line_number": 470, "usage_type": "call"}]} +{"seq_id": "286876531", "text": "# matplotlib backtest for missing $DISPLAY\nimport matplotlib\nmatplotlib.use('Agg')\n\n# scientific computing\nimport numpy as np\n\nimport qtrader as qt\n\n# visualization tools\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\nstart_date = '2015-01-01'\nend_date = '2017-01-01'\n\nuniverse = ['AAPL', 'GOOGL', 'MSFT']\n\nqt.data.Market.start_date = start_date\nqt.data.Market.end_date = end_date\nqt.data.Market.source = 'yahoo'\n\nreturns = qt.data.Market.Returns(universe)\n\nreturns[universe[0]].hist(bins=50)\nplt.savefig('tests/tmp/test__qtrader_quandl_returns.pdf',\n format='pdf', dpi=300)\n\nplt.figure()\n\nprices = qt.data.Market.Prices(universe)\n\nprices[universe[0]].plot()\nplt.savefig('tests/tmp/test__qtrader_quandl_prices.pdf', format='pdf', dpi=300)\n", "sub_path": "tests/test__qtrader_quandl.py", "file_name": "test__qtrader_quandl.py", "file_ext": "py", "file_size_in_byte": 758, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.use", "line_number": 3, "usage_type": "call"}, {"api_name": "qtrader.data", "line_number": 19, "usage_type": "attribute"}, {"api_name": "qtrader.data", "line_number": 20, "usage_type": "attribute"}, {"api_name": "qtrader.data", "line_number": 21, "usage_type": "attribute"}, {"api_name": "qtrader.data.Market.Returns", "line_number": 23, "usage_type": "call"}, {"api_name": "qtrader.data", "line_number": 23, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "qtrader.data.Market.Prices", "line_number": 31, "usage_type": "call"}, {"api_name": "qtrader.data", "line_number": 31, "usage_type": "attribute"}, {"api_name": "matplotlib.pyplot.savefig", "line_number": 34, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 34, "usage_type": "name"}]} +{"seq_id": "425290816", "text": "#!/usr/bin/python\nimport json\nfrom lxml import etree\n\nimport requests\nfrom time import sleep\n# for soap interface\nimport zeep\nimport zeep.helpers\nfrom flask import Flask\nfrom flask import request\nfrom flask import make_response as fmake_response\nfrom threading import Thread\n# from flask_cors import CORS, cross_origin\n\nfrom base64 import b64encode\nimport enumList\n\nfrom controller.RequestError import RequestError\nimport controller.UserController as uc\n\nfrom ejbcaUtils import ejbcaServ, initicalConf, createXMLfromWSDL, \\\n returnHistory, renewCACRL\n\nfrom dojot.module import Messenger, Config\napp = Flask(__name__)\n# CORS(app)\napp.url_map.strict_slashes = False\n\n\ndef make_response(payload, status):\n resp = fmake_response(payload, status)\n resp.headers['content-type'] = 'application/json'\n return resp\n\n\ndef formatResponse(status, message=None):\n payload = None\n if message:\n payload = json.dumps({'message': message, 'status': status})\n elif status >= 200 and status < 300:\n payload = json.dumps({'message': 'ok', 'status': status})\n else:\n payload = json.dumps({'message': 'Request failed', 'status': status})\n\n return make_response(payload, status)\n\ndef receiver_kafka(tenant, message):\n message = json.loads(message)\n try:\n event = message.get(\"event\")\n device_id = message['meta']['service']+':'+message['data']['id']\n if event == \"create\" or event == \"update\":\n message['username'] = device_id\n uc.createOrEditUser(message)\n elif event == \"remove\":\n uc.deleteUser(device_id)\n except Exception as e:\n print(e)\n\n@app.route('/ejbca/version', methods=['GET'])\ndef checkVersion():\n version = ejbcaServ().getEjbcaVersion()\n return make_response(json.dumps({'version': version}), 200)\n\n\n@app.route('/ca', methods=['GET'])\ndef getAvalibleCA():\n caList = zeep.helpers.serialize_object(ejbcaServ().getAvailableCAs())\n return make_response(json.dumps({'CAs': caList}), 200)\n\n\n# retrieve CA certificate chain\n@app.route('/ca/', methods=['GET'])\ndef getCAChain(cacn):\n try:\n cert = zeep.helpers.serialize_object(ejbcaServ().getLastCAChain(cacn))\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n \n return make_response(json.dumps({'certificate': cert[0]['certificateData'].decode('ascii')}), 200)\n\n\n# receive the cert status\n@app.route('/ca//certificate//status', methods=['GET'])\ndef verifyCert(cacn, certsn):\n try:\n cert = zeep.helpers.serialize_object(ejbcaServ().checkRevokationStatus(cacn, certsn))\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n resp = {\n 'reason': enumList.REVOKATION_REASON(cert['reason']).name,\n 'date': cert['revocationDate'].isoformat()\n }\n return make_response(json.dumps({'status': resp}), 200)\n\n\n# receive the cert status\n@app.route('/ca//certificate/', methods=['GET'])\ndef getCert(cacn, certsn):\n try:\n cert = zeep.helpers.serialize_object(ejbcaServ().getCertificate(certsn, cacn))\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n if cert is None:\n return formatResponse(404, 'no certificates found')\n return make_response(json.dumps({'certificate': cert}), 200)\n\n\n# revoke a certificate by serial number\n@app.route('/ca//certificate/', methods=['DELETE'])\ndef revokeCert(cacn, certsn):\n reasonCode = enumList.REVOKATION_REASON['UNSPECIFIED'].value\n if len(request.args) > 0:\n if 'reason' in request.args:\n try:\n reasonCode = enumList.REVOKATION_REASON[request.args['reason']].value\n except KeyError:\n return formatResponse(400, 'invalid revokation reason ' + request.args['reason'])\n try:\n resp = zeep.helpers.serialize_object( ejbcaServ().revokeCert(cacn, certsn, reasonCode))\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n return formatResponse(200)\n\n\n# create or update CRL\n@app.route('/ca//crl', methods=['PUT'])\ndef createCRL(caname):\n try:\n ejbcaServ().createCRL(caname)\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n return formatResponse(200)\n\n\n# get CA CRL\n@app.route('/ca//crl', methods=['GET'])\ndef getLatestCRL(caname):\n delta = False\n if len(request.args) > 0:\n if 'delta' in request.args:\n delta = request.args['delta'] in ['True', 'true']\n\n if 'update' in request.args:\n if request.args['update'] in ['True', 'true']:\n # refresh the crl data\n renewCACRL(caname)\n\n try:\n resp = ejbcaServ().getLatestCRL(caname, delta)\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n encoded = b64encode(resp).decode(\"utf-8\")\n\n return make_response(json.dumps({'CRL': encoded}), 200)\n\n\n@app.route('/user', methods=['POST'])\ndef createOrEditUser():\n if request.mimetype != 'application/json':\n return formatResponse(400, 'invalid mimetype')\n\n try:\n userInfoJson = json.loads(request.data)\n except ValueError:\n return formatResponse(400, 'malformed JSON')\n\n try:\n uc.createOrEditUser(userInfoJson)\n except RequestError as err:\n return formatResponse(err.errorCode, err.message)\n return formatResponse(200)\n\ndef findUserandReset(username):\n query = {\n \"matchtype\": 0,\n \"matchvalue\": username,\n \"matchwith\": 0\n }\n\n try:\n user = zeep.helpers.serialize_object(ejbcaServ().findUser(query))\n except zeep.exceptions.Fault as error:\n print(str(error))\n return False\n if len(user) == 0:\n print(\"No certificate found\")\n return False\n \n form_user = json.loads(json.dumps(user))\n\n if form_user[0]['status'] != 10: \n form_user[0] ['status'] = 10 # NEW = 10\n ejbcaServ().editUser(form_user)\n\n return True\n\n@app.route('/user/', methods=['GET'])\ndef findUser(username):\n query = {\n \"matchtype\": 0,\n \"matchvalue\": username,\n \"matchwith\": 0\n }\n\n try:\n user = zeep.helpers.serialize_object(ejbcaServ().findUser(query))\n print(user)\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n if user:\n return formatResponse(404, 'no certificates found')\n return make_response(json.dumps({'user': user}), 200)\n\n\n@app.route('/user/', methods=['DELETE'])\ndef deleteUser(username):\n # default values\n deleteAfter = False\n reason = 'UNSPECIFIED'\n\n # URL param\n if len(request.args) > 0:\n if 'reason' in request.args:\n reason = request.args['reason']\n elif 'delete' in request.args:\n deleteAfter = request.args['delete'] in ['True', 'true']\n try:\n uc.deleteUser(username, reason, deleteAfter)\n except RequestError as err:\n return formatResponse(err.errorCode, err.message)\n return formatResponse(200)\n\n\n@app.route('/user//find', methods=['GET'])\ndef findCerts(username):\n onlyValid = True\n if len(request.args) > 0:\n if 'valid' in request.args:\n onlyValid = request.args['valid'] in ['True', 'true']\n\n try:\n certs = zeep.helpers.serialize_object(ejbcaServ().findCerts(username, onlyValid))\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n if len(certs) == 0:\n return formatResponse(404, 'no certificates found')\n\n return make_response(json.dumps({'certs': certs}), 200)\n\n\n# json parameters: 'passwd': the user password.\n# 'certificate' base64 pkcs10 csr\n@app.route('/sign//pkcs10', methods=['POST'])\ndef pkcs10Request(cname):\n if request.mimetype != 'application/json':\n return formatResponse(400, 'invalid mimetype')\n \n try:\n info = json.loads(request.data)\n keys = info.keys()\n if 'passwd' not in keys and 'certificate' not in keys:\n return formatResponse(400,\n 'Missing parameter.'\n ' Expected: passwd and certificate')\n except ValueError:\n return formatResponse(400, 'malformed JSON')\n\n #First we need to set the user status to new \n #(the cert can only be obtained if the user have NEW status)\n # reference: https://araschnia.unam.mx/doc/ws/index.html\n\n if findUserandReset(cname) is False:\n return formatResponse(400, 'User not found to renew..')\n\n try:\n resp = (\n zeep.helpers.serialize_object(ejbcaServ()\n .pkcs10Request(\n cname,\n info['passwd'],\n info['certificate'],\n None, \"CERTIFICATE\"\n ))\n )\n except zeep.exceptions.Fault as error:\n return formatResponse(400, 'soap message: ' + error.message)\n ret = dict(resp)\n ret['data'] = ret['data'].decode('utf-8')\n\n resp_obj = {\n 'status': {\n 'data': ret['data'],\n 'responseType': ret['responseType']\n }\n }\n return make_response(json.dumps(resp_obj), 200)\n\n\nif __name__ == '__main__':\n while True:\n try:\n # execute the EJBCA handshake and load SOAP API metadata\n initicalConf()\n break\n except requests.exceptions.RequestException:\n print(\"Cant connect to EJBCA server for initial configuration\")\n print(\"Chances are the server is not ready yet.\")\n print(\"Will retry in 30sec\")\n sleep(30)\n # Configure and initalize the messenger\n config = Config()\n messenger = Messenger(\"ejbca-rest\", config)\n messenger.init()\n # Subscribe to devices topics and register callback to process new events\n messenger.create_channel(config.dojot['subjects']['devices'], \"r\")\n messenger.on(config.dojot['subjects']['devices'], \"message\", receiver_kafka)\n # Gets all devices that are already active on dojot\n messenger.generate_device_create_event_for_active_devices()\n app.run(host='0.0.0.0', port=5583, threaded=True)\n \n", "sub_path": "RESTmain.py", "file_name": "RESTmain.py", "file_ext": "py", "file_size_in_byte": 10628, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "flask.Flask", "line_number": 26, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 32, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 40, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 42, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 44, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 49, "usage_type": "call"}, {"api_name": "controller.UserController.createOrEditUser", "line_number": 55, "usage_type": "call"}, {"api_name": "controller.UserController", "line_number": 55, "usage_type": "name"}, {"api_name": "controller.UserController.deleteUser", "line_number": 57, "usage_type": "call"}, {"api_name": "controller.UserController", "line_number": 57, "usage_type": "name"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 63, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 64, "usage_type": "call"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 69, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 69, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 69, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 70, "usage_type": "call"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 77, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 77, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 77, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 78, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 81, "usage_type": "call"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 88, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 88, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 88, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 89, "usage_type": "attribute"}, {"api_name": "enumList.REVOKATION_REASON", "line_number": 92, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 95, "usage_type": "call"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 102, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 102, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 102, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 103, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 107, "usage_type": "call"}, {"api_name": "enumList.REVOKATION_REASON", "line_number": 113, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 114, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 114, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 115, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 115, "usage_type": "name"}, {"api_name": "enumList.REVOKATION_REASON", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 117, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 117, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 119, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 119, "usage_type": "name"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 121, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 121, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 121, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 122, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 131, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 132, "usage_type": "attribute"}, {"api_name": "flask.request.args", "line_number": 141, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 141, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 142, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 142, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 143, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 143, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 145, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 145, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 146, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 146, "usage_type": "name"}, {"api_name": "ejbcaUtils.renewCACRL", "line_number": 148, "usage_type": "call"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 151, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 152, "usage_type": "attribute"}, {"api_name": "base64.b64encode", "line_number": 154, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 156, "usage_type": "call"}, {"api_name": "flask.request.mimetype", "line_number": 161, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 161, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 165, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 165, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 165, "usage_type": "name"}, {"api_name": "controller.UserController.createOrEditUser", "line_number": 170, "usage_type": "call"}, {"api_name": "controller.UserController", "line_number": 170, "usage_type": "name"}, {"api_name": "controller.RequestError.RequestError", "line_number": 171, "usage_type": "name"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 183, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 183, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 183, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 184, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 191, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 191, "usage_type": "call"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 195, "usage_type": "call"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 208, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 208, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 208, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 210, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 214, "usage_type": "call"}, {"api_name": "flask.request.args", "line_number": 224, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 224, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 225, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 225, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 226, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 226, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 227, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 227, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 228, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 228, "usage_type": "name"}, {"api_name": "controller.UserController.deleteUser", "line_number": 230, "usage_type": "call"}, {"api_name": "controller.UserController", "line_number": 230, "usage_type": "name"}, {"api_name": "controller.RequestError.RequestError", "line_number": 231, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 239, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 239, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 240, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 240, "usage_type": "name"}, {"api_name": "flask.request.args", "line_number": 241, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 241, "usage_type": "name"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 244, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 244, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 244, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 245, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 250, "usage_type": "call"}, {"api_name": "flask.request.mimetype", "line_number": 257, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 257, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 261, "usage_type": "call"}, {"api_name": "flask.request.data", "line_number": 261, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 261, "usage_type": "name"}, {"api_name": "zeep.helpers.serialize_object", "line_number": 279, "usage_type": "call"}, {"api_name": "zeep.helpers", "line_number": 279, "usage_type": "attribute"}, {"api_name": "ejbcaUtils.ejbcaServ", "line_number": 279, "usage_type": "call"}, {"api_name": "zeep.exceptions", "line_number": 287, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 298, "usage_type": "call"}, {"api_name": "ejbcaUtils.initicalConf", "line_number": 305, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 307, "usage_type": "attribute"}, {"api_name": "time.sleep", "line_number": 311, "usage_type": "call"}, {"api_name": "dojot.module.Config", "line_number": 313, "usage_type": "call"}, {"api_name": "dojot.module.Messenger", "line_number": 314, "usage_type": "call"}]} +{"seq_id": "106458158", "text": "from concurrent.futures import ThreadPoolExecutor\nfrom pprint import pprint\nfrom datetime import datetime\nimport time\nfrom itertools import repeat\n\nimport yaml\nfrom netmiko import ConnectHandler\n\n\nstart_msg = '===> {} Connection to device: {}'\nreceived_msg = '<=== {} Received result from device: {}'\n\n\ndef connect_ssh(device_dict, command):\n print(start_msg.format(datetime.now().time(), device_dict['ip']))\n if device_dict['ip'] == '192.168.100.1':\n time.sleep(10)\n with ConnectHandler(**device_dict) as ssh:\n ssh.enable()\n result = ssh.send_command(command)\n print(received_msg.format(datetime.now().time(), device_dict['ip']))\n return {device_dict['ip']: result}\n\n\n\ndef threads_conn(function, devices, limit=2, command=''):\n with ThreadPoolExecutor(max_workers=limit) as executor:\n f_result = executor.map(function, devices, repeat(command))\n return list(f_result)\n\n\nif __name__ == '__main__':\n devices = yaml.load(open('devices.yaml'))\n all_done = threads_conn(connect_ssh,\n devices['routers'],\n command='sh clock')\n pprint(all_done)\n\n\"\"\"\n$ python netmiko_threads_map_final.py\n===> 05:01:08.314962 Connection to device: 192.168.100.1\n===> 05:01:08.315114 Connection to device: 192.168.100.2\n<=== 05:01:13.693083 Received result from device: 192.168.100.2\n===> 05:01:13.799002 Connection to device: 192.168.100.3\n<=== 05:01:19.363250 Received result from device: 192.168.100.3\n<=== 05:01:23.685859 Received result from device: 192.168.100.1\n[{'192.168.100.1': '*05:01:23.513 UTC Mon Aug 28 2017'},\n{'192.168.100.2': '*05:01:13.522 UTC Mon Aug 28 2017'},\n{'192.168.100.3': '*05:01:19.189 UTC Mon Aug 28 2017'}]\n\"\"\"", "sub_path": "python/conspect/DEVICE/example/exeption_concurrent_futures/netmiko_threads_map_final.py", "file_name": "netmiko_threads_map_final.py", "file_ext": "py", "file_size_in_byte": 1733, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "datetime.datetime.now", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 18, "usage_type": "call"}, {"api_name": "netmiko.ConnectHandler", "line_number": 19, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 22, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 22, "usage_type": "name"}, {"api_name": "concurrent.futures.ThreadPoolExecutor", "line_number": 28, "usage_type": "call"}, {"api_name": "itertools.repeat", "line_number": 29, "usage_type": "call"}, {"api_name": "yaml.load", "line_number": 34, "usage_type": "call"}, {"api_name": "pprint.pprint", "line_number": 38, "usage_type": "call"}]} +{"seq_id": "537100483", "text": "# coding: utf-8\nimport json\nimport os\nimport sys\nimport requests\nimport socket\nimport threading\n\nfrom flask import Flask, Response, render_template, request\n\nsys.path.insert(0, os.getcwd()[:os.getcwd().rfind('/')])\n\nimport udpc\n\n\ns = udpc.socketCUDP(socket.AF_INET)\n\npiloto_automatico = False\n\ninfo = \"0\"\n\ndef restart_program():\n \"\"\"Restarts the current program.\n Note: this function does not return. Any cleanup action (like\n saving data) must be done before calling this function.\"\"\"\n python = sys.executable\n os.execl(python, python, * sys.argv)\n\napp = Flask(__name__)\n\n@app.after_request\ndef add_header(response):\n \"\"\"\n Add headers to both force latest IE rendering engine or Chrome Frame,\n and also to cache the rendered page for 10 minutes.\n \"\"\"\n response.headers['X-UA-Compatible'] = 'IE=Edge,chrome=1'\n response.headers['Cache-Control'] = 'public, max-age=0'\n return response\n\n@app.route('/', methods=['GET', 'POST', 'PUT', 'DELETE'])\ndef api():\n if request.method == 'GET':\n return render_template('index.html') \n else:\n return json.dumps({'erro': 'Método inválido'})\n\n@app.route('/conectar//', methods=['POST'])\ndef conectar(ip, porta):\n\ts.connect((ip, int(porta)))\n\treturn json.dumps({'resposta': 'Conectado'})\n\n@app.route('/desconectar', methods=['POST'])\ndef desconectar():\n\tglobal s\n\n\ts.finalize()\n\ts.close()\n\ts = udpc.socketCUDP(socket.AF_INET)\n\treturn json.dumps({'resposta': 'Desconectado'})\n\n@app.route('/comando/', methods=['POST'])\ndef comando(tipo):\n\tif tipo == 'cima':\n\t\ts.send(\"F\")\n\telif tipo == 'baixo':\n\t\ts.send(\"T\")\n\telif tipo == 'direita':\n\t\ts.send(\"D\")\n\telif tipo == 'esquerda':\n\t\ts.send(\"E\")\n\telif tipo == 'parar':\n\t\ts.send(\"p\")\n\telse:\n\t\treturn json.dumps({'resposta': 'desconhecido'})\n\n\treturn json.dumps({'resposta': tipo})\n\n@app.route('/dados', methods=['POST'])\ndef dados():\n\tglobal info\n\t\n\tdados_f = '5.00'\n\tdados_t = '5.00'\n\n\ts.send(\"d\")\n\tdados, dados_cli = s.recv(9)\n\tinfo = dados.replace(\" \", \"\").split('|')\n\tif(len(info) < 2):\n\t\tinfo = [0, 0]\n\treturn json.dumps({'dados': 'Proximidade Frente: '+str(info[0])+' cm
    Proximidade atrás: '+str(info[1])+' cm
    '})\n\n@app.route('/piloto', methods=['POST'])\ndef piloto():\n\tglobal piloto_automatico\n\n\tif piloto_automatico == False:\n\t\ts.send(\"P\")\n\t\tpiloto_automatico = True\n\t\treturn json.dumps({'resposta': 'Piloto Automatico'})\n\telse:\n\t\ts.send(\"M\")\n\t\tpiloto_automatico = False\n\t\treturn json.dumps({'resposta': 'Manual'})\n\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n", "sub_path": "Cliente_WS/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 2518, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.path.insert", "line_number": 11, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 11, "usage_type": "attribute"}, {"api_name": "os.getcwd", "line_number": 11, "usage_type": "call"}, {"api_name": "udpc.socketCUDP", "line_number": 16, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 16, "usage_type": "attribute"}, {"api_name": "sys.executable", "line_number": 26, "usage_type": "attribute"}, {"api_name": "os.execl", "line_number": 27, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 27, "usage_type": "attribute"}, {"api_name": "flask.Flask", "line_number": 29, "usage_type": "call"}, {"api_name": "flask.request.method", "line_number": 43, "usage_type": "attribute"}, {"api_name": "flask.request", "line_number": 43, "usage_type": "name"}, {"api_name": "flask.render_template", "line_number": 44, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 51, "usage_type": "call"}, {"api_name": "udpc.socketCUDP", "line_number": 59, "usage_type": "call"}, {"api_name": "socket.AF_INET", "line_number": 59, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 60, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 75, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 77, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 91, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 100, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 104, "usage_type": "call"}]} +{"seq_id": "308300612", "text": "import sys\nimport pygame as pg\n\n\npg.init()\nscreen = pg.display.set_mode((640, 480))\n\nBG_COLOR = pg.Color('darkslategray')\n# Here I just create an image with per-pixel alpha and draw\n# some shapes on it so that we can better see the rotation effects.\n#ORIG_IMAGE = pg.Surface((240, 180), pg.SRCALPHA)\nORIG_IMAGE = pg.image.load('12.png').convert()\n#pg.draw.rect(ORIG_IMAGE, pg.Color('aquamarine3'), (80, 0, 80, 180))\n#pg.draw.rect(ORIG_IMAGE, pg.Color('gray16'), (60, 0, 120, 40))\n#pg.draw.circle(ORIG_IMAGE, pg.Color('gray16'), (120, 180), 50)\n\n\ndef main():\n clock = pg.time.Clock()\n # The rect where we'll blit the image.\n rect = ORIG_IMAGE.get_rect(center=(300, 220))\n angle = 0\n\n done = False\n while not done:\n for event in pg.event.get():\n if event.type == pg.QUIT:\n done = True\n\n # Increment the angle, then rotate the image.\n angle += 2\n image = pg.transform.rotate(ORIG_IMAGE, angle) # rotate often looks ugly.\n #image = pg.transform.rotozoom(ORIG_IMAGE, angle, 1) # rotozoom is smoother.\n # The center of the new rect is the center of the old rect.\n rect = image.get_rect(center=rect.center)\n screen.fill(BG_COLOR)\n screen.blit(image, rect)\n\n pg.display.flip()\n clock.tick(30)\n\n\nif __name__ == '__main__':\n main()\n pg.quit()\n sys.exit()", "sub_path": "Pruebas/RotaImagen.py", "file_name": "RotaImagen.py", "file_ext": "py", "file_size_in_byte": 1376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "pygame.init", "line_number": 5, "usage_type": "call"}, {"api_name": "pygame.display.set_mode", "line_number": 6, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 6, "usage_type": "attribute"}, {"api_name": "pygame.Color", "line_number": 8, "usage_type": "call"}, {"api_name": "pygame.image.load", "line_number": 12, "usage_type": "call"}, {"api_name": "pygame.image", "line_number": 12, "usage_type": "attribute"}, {"api_name": "pygame.time.Clock", "line_number": 19, "usage_type": "call"}, {"api_name": "pygame.time", "line_number": 19, "usage_type": "attribute"}, {"api_name": "pygame.event.get", "line_number": 26, "usage_type": "call"}, {"api_name": "pygame.event", "line_number": 26, "usage_type": "attribute"}, {"api_name": "pygame.QUIT", "line_number": 27, "usage_type": "attribute"}, {"api_name": "pygame.transform.rotate", "line_number": 32, "usage_type": "call"}, {"api_name": "pygame.transform", "line_number": 32, "usage_type": "attribute"}, {"api_name": "pygame.display.flip", "line_number": 39, "usage_type": "call"}, {"api_name": "pygame.display", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygame.quit", "line_number": 45, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 46, "usage_type": "call"}]} +{"seq_id": "23886114", "text": "# Importing libraries\r\n#general stuff\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport pandas as pd\r\nimport seaborn as sns \r\n\r\n#apna time ayega\r\nfrom preprocessing import loadMat\r\nfrom util import *\r\n\r\n# Importing dataset (temp 24)\r\nB0005 = loadMat('B0005.mat')\r\nB0006 = loadMat('B0006.mat')\r\nB0007 = loadMat('B0007.mat')\r\nB0018 = loadMat('B0018.mat')\r\n\r\n# battery capacity\r\nB0005_capacity = getBatteryCapcity(B0005)\r\nB0006_capacity = getBatteryCapcity(B0006)\r\nB0007_capacity = getBatteryCapcity(B0007)\r\nB0018_capacity = getBatteryCapcity(B0018)\r\n\r\n# plotting disintegrating of battery capacity\r\nplt.plot(B0005_capacity[0], B0005_capacity[1], color='blue', label='Battery-05')\r\nplt.plot(B0006_capacity[0], B0006_capacity[1], color='green', label='Battery-06')\r\nplt.plot(B0007_capacity[0], B0007_capacity[1], color='red', label='Battery-07')\r\nplt.plot(B0018_capacity[0], B0018_capacity[1], color='purple', label='Battery-18')\r\nplt.xlabel('Discharge cycles')\r\nplt.ylabel('Capacity/Ah')\r\nplt.title('Capacity degradation at ambient temperature of 24°C')\r\nplt.legend()\r\nplt.show() \r\n\r\n\r\n\r\n# capacity retention percent\r\nB0005_capacity_retention = getBatteryCapacityRetention(B0005_capacity[1])\r\n\r\n#Plotting(battery retention % remains same for 1 ambient temperature) \r\nplt.plot(B0005_capacity_retention[0], B0005_capacity_retention[1], color='blue', label='Battery-05')\r\nplt.xlabel('Discharge cycles')\r\nplt.ylabel('Capacity retention percentage')\r\nplt.title('Capacity degradation at ambient temperature of 24°C')\r\nplt.legend()\r\nplt.show() \r\n\r\npred_range = []\r\nfor i in range(1, 270):\r\n pred_range.append(i)\r\n\r\n#brute force linear fitting\r\npoly = np.polyfit(B0005_capacity[0], B0005_capacity[1], 1)\r\nB0005_capacity_pred = np.polyval(poly, pred_range)\r\n\r\nplt.plot(B0005_capacity[0], B0005_capacity[1], color='orange', label='24°C (Battery-05)')\r\nplt.plot(pred_range, B0005_capacity_pred, color='orange')\r\nplt.xlabel('Discharge cycles')\r\nplt.ylabel('Capacity/Ah')\r\nplt.title('Capacity degradation ')\r\nplt.legend()\r\nplt.show()\r\n", "sub_path": "diagrams.py", "file_name": "diagrams.py", "file_ext": "py", "file_size_in_byte": 2040, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "preprocessing.loadMat", "line_number": 13, "usage_type": "call"}, {"api_name": "preprocessing.loadMat", "line_number": 14, "usage_type": "call"}, {"api_name": "preprocessing.loadMat", "line_number": 15, "usage_type": "call"}, {"api_name": "preprocessing.loadMat", "line_number": 16, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 25, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 25, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 26, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 26, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 27, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 27, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 28, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 28, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 29, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 29, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 30, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 30, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 31, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 31, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 32, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 32, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 33, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 33, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 41, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 41, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 42, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 42, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 43, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 43, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 44, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 44, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 45, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 45, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 46, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 46, "usage_type": "name"}, {"api_name": "numpy.polyfit", "line_number": 53, "usage_type": "call"}, {"api_name": "numpy.polyval", "line_number": 54, "usage_type": "call"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 56, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 56, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.plot", "line_number": 57, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 57, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.xlabel", "line_number": 58, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 58, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.ylabel", "line_number": 59, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 59, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.title", "line_number": 60, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 60, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.legend", "line_number": 61, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 61, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.show", "line_number": 62, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 62, "usage_type": "name"}]} +{"seq_id": "413731332", "text": "import apscheduler.schedulers.blocking\nimport datetime\nimport logging\nimport os\nimport pathlib\nimport requests\nimport requests.auth\nimport signal\nimport sys\n\nlog = logging.getLogger(__name__)\n\n\nclass Settings:\n _true_values = ('1', 'true', 'on', 'yes')\n\n def __init__(self):\n self.client_id = os.getenv('SEISMIC_CLIENT_ID')\n self.client_secret = os.getenv('SEISMIC_CLIENT_SECRET')\n self.interval = int(os.getenv('INTERVAL', '6'))\n self.log_format = os.getenv('LOG_FORMAT', '%(levelname)s [%(name)s] %(message)s')\n self.log_level = os.getenv('LOG_LEVEL', 'INFO')\n self.output_folder = pathlib.Path(os.getenv('OUTPUT_FOLDER', '/data'))\n self.password = os.getenv('SEISMIC_PASSWORD')\n self.run_and_exit = os.getenv('RUN_AND_EXIT', 'False').lower() in self._true_values\n self.tenant = os.getenv('SEISMIC_TENANT')\n self.username = os.getenv('SEISMIC_USERNAME')\n self.version = os.getenv('APP_VERSION')\n\n\nclass BearerAuth(requests.auth.AuthBase):\n def __init__(self, token: str):\n self.token = token\n\n def __call__(self, r: requests.Request) -> requests.Request:\n r.headers['Authorization'] = f'Bearer {self.token}'\n return r\n\n\nclass SeismicClient:\n def __init__(self, settings: Settings):\n self.settings = settings\n self.session = requests.Session()\n self._token = None\n self.token_expiration = None\n\n @property\n def token_expired(self) -> bool:\n if self.token_expiration is None:\n return True\n if self.token_expiration < datetime.datetime.utcnow():\n return True\n return False\n\n @property\n def token(self) -> str:\n if self._token is None or self.token_expired:\n log.debug('Getting a new access token')\n url = f'https://auth.seismic.com/tenants/{self.settings.tenant}/connect/token'\n data = {\n 'grant_type': 'password',\n 'client_id': self.settings.client_id,\n 'client_secret': self.settings.client_secret,\n 'username': self.settings.username,\n 'password': self.settings.password,\n 'scope': 'download library reporting'\n }\n resp = self.session.post(url, data=data)\n resp.raise_for_status()\n j = resp.json()\n self._token = j.get('access_token')\n expires_in = j.get('expires_in')\n self.token_expiration = datetime.datetime.utcnow() + datetime.timedelta(seconds=expires_in)\n log.debug(f'Access token {self._token[:6]}... will expire at {self.token_expiration}')\n return self._token\n\n def get_report_csv(self, report_name: str, **kwargs):\n log.info(f'Getting csv for {report_name}')\n url = f'https://api.seismic.com/reporting/v2/{report_name}'\n headers = {'Accept': 'text/csv'}\n resp = self.session.get(url, auth=BearerAuth(self.token), headers=headers, params=kwargs)\n resp.raise_for_status()\n return resp.text\n\n def get_report_json(self, report_name: str, **kwargs):\n log.info(f'Getting json for {report_name}')\n url = f'https://api.seismic.com/reporting/v2/{report_name}'\n headers = {'Accept': 'application/json'}\n resp = self.session.get(url, auth=BearerAuth(self.token), headers=headers, params=kwargs)\n resp.raise_for_status()\n return resp.json()\n\n\ndef main_job(settings: Settings):\n c = SeismicClient(settings)\n reports = [\n 'contentUsageHistory',\n 'contentViewHistory',\n 'searchHistory',\n 'livesendLinks',\n 'livesendLinkContents',\n 'livesendLinkMembers',\n 'livesendPageViews',\n 'generatedLiveDocs',\n 'generatedLiveDocComponents',\n 'generatedLiveDocSlides',\n 'generatedLiveDocFields',\n 'generatedLiveDocOutputFormats',\n 'libraryContents',\n 'contentPropertyAssignments',\n 'workspaceContents',\n 'users',\n 'userPropertyAssignments',\n 'groupMembers',\n ]\n for report_name in reports:\n try:\n result = c.get_report_csv(report_name)\n except requests.exceptions.HTTPError as e:\n log.error(e)\n continue\n output_file = settings.output_folder / f'{report_name}.csv'\n with output_file.open('w') as f:\n f.write(result)\n\n\ndef main():\n settings = Settings()\n logging.basicConfig(format=settings.log_format, level=logging.DEBUG, stream=sys.stdout)\n log.debug(f'seismic-etl {settings.version}')\n if not settings.log_level == 'DEBUG':\n log.debug(f'Changing log level to {settings.log_level}')\n logging.getLogger().setLevel(settings.log_level)\n\n log.info(f'RUN_AND_EXIT: {settings.run_and_exit}')\n if settings.run_and_exit:\n main_job(settings)\n else:\n scheduler = apscheduler.schedulers.blocking.BlockingScheduler()\n\n # Run it now ...\n scheduler.add_job(main_job, args=[settings])\n # ... and run it later!\n scheduler.add_job(main_job, 'interval', args=[settings], hours=settings.interval)\n\n scheduler.start()\n\n\ndef handle_sigterm(_signal, _frame):\n sys.exit()\n\n\nif __name__ == '__main__':\n signal.signal(signal.SIGTERM, handle_sigterm)\n main()\n", "sub_path": "seismic-etl.py", "file_name": "seismic-etl.py", "file_ext": "py", "file_size_in_byte": 5330, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 11, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 18, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 19, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 20, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 21, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 22, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 23, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 23, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 24, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 25, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 26, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 27, "usage_type": "call"}, {"api_name": "os.getenv", "line_number": 28, "usage_type": "call"}, {"api_name": "requests.auth", "line_number": 31, "usage_type": "attribute"}, {"api_name": "requests.Request", "line_number": 35, "usage_type": "attribute"}, {"api_name": "requests.Session", "line_number": 43, "usage_type": "call"}, {"api_name": "datetime.datetime.utcnow", "line_number": 51, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 51, "usage_type": "attribute"}, {"api_name": "datetime.datetime.utcnow", "line_number": 73, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 73, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 73, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 119, "usage_type": "attribute"}, {"api_name": "logging.basicConfig", "line_number": 129, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 129, "usage_type": "attribute"}, {"api_name": "sys.stdout", "line_number": 129, "usage_type": "attribute"}, {"api_name": "logging.getLogger", "line_number": 133, "usage_type": "call"}, {"api_name": "apscheduler.schedulers.blocking.schedulers.blocking.BlockingScheduler", "line_number": 139, "usage_type": "call"}, {"api_name": "apscheduler.schedulers.blocking.schedulers", "line_number": 139, "usage_type": "attribute"}, {"api_name": "apscheduler.schedulers.blocking", "line_number": 139, "usage_type": "name"}, {"api_name": "sys.exit", "line_number": 150, "usage_type": "call"}, {"api_name": "signal.signal", "line_number": 154, "usage_type": "call"}, {"api_name": "signal.SIGTERM", "line_number": 154, "usage_type": "attribute"}]} +{"seq_id": "505009146", "text": "import sys\nfrom gevent import event\nimport gevent\nimport logging\nimport time\n\nLOG = logging.getLogger(__name__)\n\nclass LoopingCallDone(Exception):\n def __init__(self, retvalue=True):\n self.retvalue = retvalue\n\n\nclass LoopingcallBase(object):\n def __init__(self, f=None, *args, **kw):\n self.args = args\n self.kw = kw\n self.f = f\n self._running = False\n self.done = None\n\n def stop(self):\n self._running = False\n\n def wait(self):\n return self.don.wait()\n\n\nclass FixedIntervalLoopingCall(LoopingcallBase):\n\n def start(self, interval, initial_delay=None):\n self._running = True\n done = event.AsyncResult()\n\n def _inner():\n if initial_delay:\n gevent.sleep(initial_delay)\n\n try:\n while self._running:\n start = time.time()\n self.f(*self.args, **self.kw)\n end = time.time()\n if not self._running:\n break\n delay = interval - (end - start)\n if delay <= 0:\n log.warn('task run outlasted interval by {delay} sec'.format(delay = -delay))\n gevent.sleep(delay if delay > 0 else 0)\n except Exception as e:\n LOG.exception('in fixed duration looping call')\n done.set_exception(e)\n return\n else:\n done.done()\n\n self.done = done\n green = gevent.spawn(_inner)\n green.run()\n return self.done\n\n\ndef test():\n print('test')\n\ndef hello():\n while True:\n print('Hello')\n gevent.sleep(1)\n\nif __name__ == '__main__':\n\n loop = FixedIntervalLoopingCall(f=test)\n #loop.interval(1)\n g1 = gevent.spawn(hello)\n loop.start(interval=2, initial_delay=60)\n print('sss')\n g1.join()\n", "sub_path": "loopingcall.py", "file_name": "loopingcall.py", "file_ext": "py", "file_size_in_byte": 1912, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 7, "usage_type": "call"}, {"api_name": "gevent.event.AsyncResult", "line_number": 33, "usage_type": "call"}, {"api_name": "gevent.event", "line_number": 33, "usage_type": "name"}, {"api_name": "gevent.sleep", "line_number": 37, "usage_type": "call"}, {"api_name": "time.time", "line_number": 41, "usage_type": "call"}, {"api_name": "time.time", "line_number": 43, "usage_type": "call"}, {"api_name": "gevent.sleep", "line_number": 49, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 58, "usage_type": "call"}, {"api_name": "gevent.sleep", "line_number": 69, "usage_type": "call"}, {"api_name": "gevent.spawn", "line_number": 75, "usage_type": "call"}]} +{"seq_id": "490102465", "text": "import datetime\nimport os\nfrom unittest import TestCase\nfrom unittest.mock import patch, Mock\n\nfrom pipelinewise.fastsync.commons.tap_s3_csv import FastSyncTapS3Csv, S3Helper\n\n\n# pylint: disable=missing-function-docstring,protected-access,invalid-name\nclass TestFastSyncTapS3Csv(TestCase):\n \"\"\"\n Unit tests for fastsync common functions for tap s3 csv\n \"\"\"\n\n def setUp(self) -> None:\n self.maxDiff = None\n con_config = {\n 'bucket': 'testBucket',\n 'aws_endpoint_url': 'https://aws.com/random-url',\n 'start_date': '2000-01-01',\n 'tables': [\n {\n 'table_name': 'table 1',\n 'key_properties': None\n }, {\n 'table_name': 'table 2',\n 'key_properties': []\n }, {\n 'table_name': 'table 3',\n 'key_properties': ['key_1']\n }, {\n 'table_name': 'table 4',\n 'key_properties': ['key_2', 'key_3']\n }\n ]\n }\n\n def tap_type_to_target_type(tap_type):\n return {\n 'boolean': 'boolean',\n 'integer': 'number',\n 'number': 'number',\n 'string': 'varchar'\n }.get(tap_type, 'varchar')\n\n with patch('pipelinewise.fastsync.commons.tap_s3_csv.S3Helper') as s3_helper_mock:\n s3_helper_mock.return_value.list_files_in_bucket.return_value = []\n self.fs_tap_s3_csv = FastSyncTapS3Csv(con_config, tap_type_to_target_type)\n\n def test_copy_table_given_an_invalid_file_path_throws_exception(self):\n with self.assertRaises(Exception):\n self.fs_tap_s3_csv.copy_table('table_1', 'invalid_file_path.csv')\n\n def test_copy_table_given_a_valid_file_path(self):\n with patch('pipelinewise.fastsync.commons.tap_s3_csv.S3Helper') as s3_helper_mock:\n s3_helper_mock.get_input_files_for_table.return_value = [\n {\n 'key': 'file_1.csv',\n 'last_modified': datetime.datetime.strptime('2001-07-13', '%Y-%m-%d')\n }, {\n 'key': 'file_2.csv',\n 'last_modified': datetime.datetime.strptime('2001-10-05', '%Y-%m-%d')\n }\n ]\n\n with patch.object(self.fs_tap_s3_csv, '_get_file_records') as get_file_rec_mock:\n get_file_rec_mock.return_value = 'test'\n\n with patch('pipelinewise.fastsync.commons.tap_s3_csv.gzip') as gzip_mock:\n mock_enter = Mock()\n mock_enter.return_value.open.return_value = ''\n\n gzip_mock.return_value.__enter__ = mock_enter\n gzip_mock.return_value.__exit__ = Mock()\n\n self.fs_tap_s3_csv.copy_table('table 2', 'file_path.csv.gz')\n\n self.assertEqual(2, get_file_rec_mock.call_count)\n self.assertIn('table 2', self.fs_tap_s3_csv.tables_last_modified)\n self.assertEqual('2001-10-05',\n self.fs_tap_s3_csv.tables_last_modified['table 2'].strftime('%Y-%m-%d'))\n\n def test_get_file_records(self):\n with patch.object(S3Helper, 'get_file_handle') as get_file_handle_mock:\n handle = Mock().return_value\n handle.configure_mock(**{\n '_raw_stream.return_value': 'file handle'\n })\n\n get_file_handle_mock.return_value = handle.return_value\n\n with patch('pipelinewise.fastsync.commons.tap_s3_csv.singer_encodings_csv') as singer_encodings_csv_mock:\n singer_encodings_csv_mock.get_row_iterator.return_value = [\n {\n 'id': 1,\n 'group': 'A',\n }, {\n 'id': 2,\n 'group': 'A',\n 'test': True\n }, {\n 'id': 3,\n 'group': 'B',\n }\n ]\n\n with patch('pipelinewise.fastsync.commons.tap_s3_csv.datetime') as datetime_mock:\n datetime_mock.utcnow.return_value.strftime.return_value = '2019-11-21'\n\n records = []\n headers = set()\n\n self.fs_tap_s3_csv._get_file_records('s3 path 1',\n {},\n records,\n headers)\n\n self.assertListEqual([\n {\n S3Helper.SDC_SOURCE_BUCKET_COLUMN: 'testBucket',\n S3Helper.SDC_SOURCE_FILE_COLUMN: 's3 path 1',\n S3Helper.SDC_SOURCE_LINENO_COLUMN: 1,\n '_SDC_EXTRACTED_AT': '2019-11-21',\n '_SDC_BATCHED_AT': '2019-11-21',\n '_SDC_DELETED_AT': None,\n '\"ID\"': 1,\n '\"GROUP\"': 'A',\n }, {\n S3Helper.SDC_SOURCE_BUCKET_COLUMN: 'testBucket',\n S3Helper.SDC_SOURCE_FILE_COLUMN: 's3 path 1',\n S3Helper.SDC_SOURCE_LINENO_COLUMN: 2,\n '_SDC_EXTRACTED_AT': '2019-11-21',\n '_SDC_BATCHED_AT': '2019-11-21',\n '_SDC_DELETED_AT': None,\n '\"ID\"': 2,\n '\"GROUP\"': 'A',\n '\"TEST\"': True,\n }, {\n S3Helper.SDC_SOURCE_BUCKET_COLUMN: 'testBucket',\n S3Helper.SDC_SOURCE_FILE_COLUMN: 's3 path 1',\n S3Helper.SDC_SOURCE_LINENO_COLUMN: 3,\n '_SDC_EXTRACTED_AT': '2019-11-21',\n '_SDC_BATCHED_AT': '2019-11-21',\n '_SDC_DELETED_AT': None,\n '\"ID\"': 3,\n '\"GROUP\"': 'B',\n }\n ], records)\n\n self.assertSetEqual(\n {\n '\"ID\"', '\"GROUP\"', '\"TEST\"',\n S3Helper.SDC_SOURCE_LINENO_COLUMN, S3Helper.SDC_SOURCE_FILE_COLUMN,\n S3Helper.SDC_SOURCE_BUCKET_COLUMN, '_SDC_EXTRACTED_AT',\n '_SDC_BATCHED_AT', '_SDC_DELETED_AT'\n },\n headers)\n\n def test_fetch_current_incremental_key_pos_with_no_tables_in_dictionary_returns_empty_dict(self):\n self.assertFalse(self.fs_tap_s3_csv.fetch_current_incremental_key_pos('table-x'))\n\n def test_fetch_current_incremental_key_pos_with_tables_in_dictionary_returns_empty_dict(self):\n dt = datetime.datetime.strptime('2019-11-21', '%Y-%m-%d')\n self.fs_tap_s3_csv.tables_last_modified['table-x'] = dt\n self.assertEqual({'modified_since': dt.isoformat()},\n self.fs_tap_s3_csv.fetch_current_incremental_key_pos('table-x', 'key'))\n\n def test_get_primary_keys_with_table_that_has_no_keys_returns_none(self):\n self.assertIsNone(self.fs_tap_s3_csv._get_primary_keys({}))\n\n def test_get_primary_keys_with_table_that_has_empty_keys_list_returns_none(self):\n self.assertIsNone(self.fs_tap_s3_csv._get_primary_keys({'key_properties': []}))\n\n def test_get_primary_keys_with_table_that_has_1_key_returns_one_safe_key(self):\n self.assertEqual(['\"KEY_1\"'], self.fs_tap_s3_csv._get_primary_keys({'key_properties': ['key_1']}))\n\n def test_get_primary_keys_with_table_that_has_2_keys_returns_concatenated_keys(self):\n self.assertIn(self.fs_tap_s3_csv._get_primary_keys({'key_properties': ['key_2', 'key_3']}),\n [['\"KEY_2\"', '\"KEY_3\"'], ['\"KEY_3\"', '\"KEY_2\"']])\n\n def test_get_table_columns(self):\n output = list(\n self.fs_tap_s3_csv._get_table_columns(f'{os.path.dirname(__file__)}/resources/dummy_data.csv.gz'))\n\n self.assertListEqual([\n ('Region', 'string'),\n ('Country', 'string'),\n ('Item Type', 'string'),\n ('Sales Channel', 'string'),\n ('Order Priority', 'string'),\n ('Order Date', 'string'),\n ('Order ID', 'integer'),\n ('Ship Date', 'string'),\n ('Units Sold', 'integer'),\n ('Unit Price', 'number'),\n ('Unit Cost', 'number'),\n ('Total Revenue', 'number'),\n ('Total Cost', 'number'),\n ('Total Profit', 'number'),\n ], output)\n\n def test_map_column_types_to_target(self):\n output = self.fs_tap_s3_csv.map_column_types_to_target(\n f'{os.path.dirname(__file__)}/resources/dummy_data.csv.gz', 'table 1')\n\n self.assertDictEqual({\n 'columns': [\n 'Region varchar',\n 'Country varchar',\n 'Item Type varchar',\n 'Sales Channel varchar',\n 'Order Priority varchar',\n 'Order Date varchar',\n 'Order ID number',\n 'Ship Date varchar',\n 'Units Sold number',\n 'Unit Price number',\n 'Unit Cost number',\n 'Total Revenue number',\n 'Total Cost number',\n 'Total Profit number',\n ],\n 'primary_key': None\n }, output)\n", "sub_path": "tests/units/fastsync/commons/test_fastsync_tap_s3_csv.py", "file_name": "test_fastsync_tap_s3_csv.py", "file_ext": "py", "file_size_in_byte": 9707, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "unittest.TestCase", "line_number": 10, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 46, "usage_type": "call"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.FastSyncTapS3Csv", "line_number": 48, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 55, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 59, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 59, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 62, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 62, "usage_type": "attribute"}, {"api_name": "unittest.mock.patch.object", "line_number": 66, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 66, "usage_type": "name"}, {"api_name": "unittest.mock.patch", "line_number": 69, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 70, "usage_type": "call"}, {"api_name": "unittest.mock.Mock", "line_number": 74, "usage_type": "call"}, {"api_name": "unittest.mock.patch.object", "line_number": 84, "usage_type": "call"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 84, "usage_type": "argument"}, {"api_name": "unittest.mock.patch", "line_number": 84, "usage_type": "name"}, {"api_name": "unittest.mock.Mock", "line_number": 85, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 92, "usage_type": "call"}, {"api_name": "unittest.mock.patch", "line_number": 107, "usage_type": "call"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_BUCKET_COLUMN", "line_number": 120, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 120, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_FILE_COLUMN", "line_number": 121, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 121, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_LINENO_COLUMN", "line_number": 122, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 122, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_BUCKET_COLUMN", "line_number": 129, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 129, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_FILE_COLUMN", "line_number": 130, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 130, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_LINENO_COLUMN", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 131, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_BUCKET_COLUMN", "line_number": 139, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 139, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_FILE_COLUMN", "line_number": 140, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 140, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_LINENO_COLUMN", "line_number": 141, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 141, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_LINENO_COLUMN", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 153, "usage_type": "name"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_FILE_COLUMN", "line_number": 153, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper.SDC_SOURCE_BUCKET_COLUMN", "line_number": 154, "usage_type": "attribute"}, {"api_name": "pipelinewise.fastsync.commons.tap_s3_csv.S3Helper", "line_number": 154, "usage_type": "name"}, {"api_name": "datetime.datetime.strptime", "line_number": 163, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 163, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 183, "usage_type": "call"}, {"api_name": "os.path", "line_number": 183, "usage_type": "attribute"}, {"api_name": "os.path.dirname", "line_number": 204, "usage_type": "call"}, {"api_name": "os.path", "line_number": 204, "usage_type": "attribute"}]} +{"seq_id": "524192329", "text": "# import google translate\nfrom ais.ai_lib.translate import *\n# print(translate('hola amigos'))\n# print(detect_language('hola amigos'))\n\n# NLP with spaCy https://spacy.io\nfrom spacy.en import English\nnlp = English()\n\n# Useful properties, summary of the docs from https://spacy.io\n\n# class Doc\n# properties: text, vector, vector_norm, ents, noun_chunks, sents\n# method: similarity\n# NER specs https://spacy.io/docs#annotation-ner\n# doc tokenization will preserve meaningful units together\n\n# class Token\n# token.doc -> parent sequence\n# string features: text, lemma, lower, shape\n# boolean flags: https://spacy.io/docs#token-booleanflags\n# POS: pos_, tag_\n# tree: https://spacy.io/docs#token-navigating \n# ner: ent_type, ent_iob\n\n# class Span\n# span.doc -> parent sequence\n# vector, vector_norm\n# string features: text, lemma\n# methods: similarity\n# syntactic parse: use root, lefts, rights, subtree https://spacy.io/docs#span-navigativing-parse\n\n\n# !more to implement:\n# also filter to prepare for tree\n# syntactic parse tree https://spacy.io/docs#span-navigativing-parse\n# word2vec, numpy array\n# similarity https://spacy.io/docs#examples-word-vectors https://spacy.io/docs#span-similarity\n\n# https://github.com/spacy-io/sense2vec/\n# tuts https://spacy.io/docs#tutorials\n# custom NER and intent arg parsing eg https://github.com/spacy-io/spaCy/issues/217\n\n\n# Helper methods\n##########################################\n\ndef merge_ents(doc):\n '''Helper: merge adjacent entities into single tokens; modifies the doc.'''\n for ent in doc.ents:\n ent.merge(ent.root.tag_, ent.text, ent.label_)\n return doc\n\ndef _NER_POS_lr_subtree(root):\n '''Helper: generate a NER_POS subtree with left/right for a root token. The doc must have merge_ents(doc) ran on it.'''\n subtree = {\n root.text: {\n \"edge\": root.dep_,\n \"tag\": root.ent_type_ or root.pos_,\n \"lefts\": [],\n \"rights\": []\n }\n }\n for l in root.lefts:\n subtree[root.text][\"lefts\"].append(_NER_POS_subtree(l))\n for r in root.rights:\n subtree[root.text][\"rights\"].append(_NER_POS_subtree(r))\n return subtree\n\ndef _NER_POS_subtree(root):\n '''Helper: generate a NER_POS subtree without left/right for a root token. The doc must have merge_ents(doc) ran on it.'''\n subtree = {\n root.text: {\n \"edge\": root.dep_,\n \"tag\": root.ent_type_ or root.pos_,\n \"children\": []\n }\n }\n for c in root.children:\n subtree[root.text][\"children\"].append(_NER_POS_subtree(c))\n return subtree\n\ndef _NER_POS_lr_tree(sent):\n '''Helper: generate the NER_POS tree (with left/right) for a sentence'''\n return _NER_POS_lr_subtree(sent.root)\n\ndef _NER_POS_tree(sent):\n '''Helper: generate the NER_POS tree for a sentence'''\n return _NER_POS_subtree(sent.root)\n\n\ndef NER_POS_lr_tree(doc):\n '''generate the NER_POS tree (with left/right) for all sentences in a doc'''\n merge_ents(doc) # merge the entities into single tokens first\n return [_NER_POS_lr_tree(sent) for sent in doc.sents]\n\ndef NER_POS_tree(doc):\n '''generate the NER_POS tree for all sentences in a doc'''\n merge_ents(doc) # merge the entities into single tokens first\n return [_NER_POS_tree(sent) for sent in doc.sents]\n\ndef NER_POS_tag(doc):\n '''tag the doc first by NER (merged as tokens) then POS. Can be seen as the flat version of NER_POS_tree'''\n merge_ents(doc) # merge the entities into single tokens first\n return [(token.text, token.ent_type_ or token.pos_) for token in doc]\n\n# s = \"find me flights from New York to London next month\"\n# doc = nlp(s)\n# NER_POS_tag(doc)\n# => [('find', 'VERB'), ('me', 'NOUN'), ('flights', 'NOUN'), ('from', 'ADP'), ('New York', 'GPE'), ('to', 'ADP'), ('London', 'GPE'), ('next month', 'DATE')]\n# => [{'find': {'edge': 'ROOT', 'children': [{'flights': {'edge': 'ccomp', 'children': [{'me': {'edge': 'nsubj', 'children': [], 'tag': 'NOUN'}}, {'from': {'edge': 'prep', 'children': [{'New York': {'edge': 'pobj', 'children': [], 'tag': 'GPE'}}], 'tag': 'ADP'}}, {'to': {'edge': 'prep', 'children': [{'London': {'edge': 'pobj', 'children': [], 'tag': 'GPE'}}], 'tag': 'ADP'}}], 'tag': 'NOUN'}}, {'next month': {'edge': 'npadvmod', 'children': [], 'tag': 'DATE'}}], 'tag': 'VERB'}}]\n\n\n\n# Primary methods\n##########################################\n\ndef parse(sentence):\n '''\n Main method: parse an input sentence and return the nlp properties.\n '''\n doc = nlp(sentence)\n reply = {\n \"text\": doc.text,\n \"len\": len(doc),\n\n \"tokens\": [token.text for token in doc],\n \"lemmas\": [token.lemma_ for token in doc],\n # \"lower\": [token.lower_ for token in doc],\n # \"shape\": [token.shape_ for token in doc],\n\n \"NER\": list(zip([token.text for token in doc.ents], [token.label_ for token in doc.ents])),\n \"noun_phrases\": [token.text for token in doc.noun_chunks],\n \"pos_coarse\": list(zip([token.text for token in doc], [token.pos_ for token in doc])),\n \"pos_fine\": list(zip([token.text for token in doc], [token.tag_ for token in doc])),\n \"NER_POS_tree\": NER_POS_tree(doc),\n \"NER_POS_tag\": NER_POS_tag(doc)\n }\n return reply\n\n# res = parse(\"find me flights from New York to London next month.\")\n\ndef parsedoc(input):\n '''\n parse for multi-sentences; split and apply parse in a list.\n '''\n doc = nlp(input, tag=False, entity=False)\n return [parse(sent.text) for sent in doc.sents]\n\n# res = parsedoc(\"find me flights from New York to London next month.\")\n\n", "sub_path": "lib/py/nlp.py", "file_name": "nlp.py", "file_ext": "py", "file_size_in_byte": 5376, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "spacy.en.English", "line_number": 8, "usage_type": "call"}]} +{"seq_id": "429725418", "text": "\"\"\"Bla.\"\"\"\n\nfrom rest_framework import serializers\nfrom saboteurs import (\n csv_to_groups_data,\n find_statistical_saboteurs,\n find_logical_saboteurs,\n statistics_report,\n)\nfrom ..base import AsyncWorker, StartJobView\nfrom ..tools import data_to_html_data, file_to_filelike_object\nfrom ..serializers import FileSerializer\n\n\nclass serializer_class(serializers.Serializer):\n \"\"\"Serializer.\"\"\"\n\n assemblies_data_file = FileSerializer()\n method = serializers.CharField()\n parts_input_type = serializers.CharField()\n constructs = serializers.ListField(\n child=FileSerializer(), allow_null=True, required=False\n )\n\n\nclass worker_class(AsyncWorker):\n def work(self):\n data = self.data\n self.logger(message=\"analyzing results...\")\n asm_data = file_to_filelike_object(data.assemblies_data_file).read()\n if data.method == \"statistical\":\n groups_data = csv_to_groups_data(csv_string=asm_data.decode())\n analysis_results = find_statistical_saboteurs(groups_data)\n self.logger(message=\"Writing a report...\")\n report_data = statistics_report(\n analysis_results,\n \"@memory\",\n replacements=[\n (\"groups\", \"assemblies\"),\n (\"group\", \"assembly\"),\n (\"member\", \"part\"),\n ],\n )\n return {\n \"report\": {\n \"data\": data_to_html_data(report_data, \"pdf\"),\n \"name\": \"saboteur_report.pdf\",\n \"mimetype\": \"application/pdf\",\n }\n }\n else:\n groups_data = csv_to_groups_data(csv_string=asm_data.decode())\n groups = {\n e['id']: e['members']\n for e in groups_data.values()\n }\n failed = [\n e['id']\n for e in groups_data.values()\n if e['attempts'] == e['failures']\n ]\n result = find_logical_saboteurs(\n groups, failed_groups=failed)\n return dict(\n saboteurs=result[\"saboteurs\"], suspicious=result[\"suspicious\"]\n )\n\n\nclass FindSaboteurPartsView(StartJobView):\n serializer_class = serializer_class\n worker_class = worker_class\n", "sub_path": "backend/app/views/find_saboteur_parts/FindSaboteurPartsView.py", "file_name": "FindSaboteurPartsView.py", "file_ext": "py", "file_size_in_byte": 2337, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "rest_framework.serializers.Serializer", "line_number": 15, "usage_type": "attribute"}, {"api_name": "rest_framework.serializers", "line_number": 15, "usage_type": "name"}, {"api_name": "serializers.FileSerializer", "line_number": 18, "usage_type": "call"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 19, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 19, "usage_type": "name"}, {"api_name": "rest_framework.serializers.CharField", "line_number": 20, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 20, "usage_type": "name"}, {"api_name": "rest_framework.serializers.ListField", "line_number": 21, "usage_type": "call"}, {"api_name": "rest_framework.serializers", "line_number": 21, "usage_type": "name"}, {"api_name": "serializers.FileSerializer", "line_number": 22, "usage_type": "call"}, {"api_name": "base.AsyncWorker", "line_number": 26, "usage_type": "name"}, {"api_name": "tools.file_to_filelike_object", "line_number": 30, "usage_type": "call"}, {"api_name": "saboteurs.csv_to_groups_data", "line_number": 32, "usage_type": "call"}, {"api_name": "saboteurs.find_statistical_saboteurs", "line_number": 33, "usage_type": "call"}, {"api_name": "saboteurs.statistics_report", "line_number": 35, "usage_type": "call"}, {"api_name": "tools.data_to_html_data", "line_number": 46, "usage_type": "call"}, {"api_name": "saboteurs.csv_to_groups_data", "line_number": 52, "usage_type": "call"}, {"api_name": "saboteurs.find_logical_saboteurs", "line_number": 62, "usage_type": "call"}, {"api_name": "base.StartJobView", "line_number": 69, "usage_type": "name"}]} +{"seq_id": "478533367", "text": "from setuptools import setup\n\nVERSION = '1.0' # hmmm... how to tag forks...\n\n# this plugin was originally by Jeff Hammel \n# but I've substantially altered and maintained it for a while now.\n# updated for Trac 1.4\n\nsetup(name='TicketMoverPlugin',\n version=VERSION,\n description=\"move tickets from one Trac to a sibling Trac\",\n author='Nathan Bird',\n author_email='nathan@acceleration.net',\n url='https://github.com/OnroerendErfgoed/TicketMoverPlugin',\n keywords='trac plugin',\n license=\"BSD\",\n py_modules=['ticketmoverplugin'],\n install_requires=[\n 'Trac>=1.4',\n 'TracSQLHelper==0.3.1'\n ],\n dependency_links=[\n \"svn+https://trac-hacks.org/svn/tracsqlhelperscript/0.12/#egg=TracSQLHelper-0.2.2\",\n ],\n entry_points={\n 'trac.plugins': [\n 'ticketmoverplugin=ticketmoverplugin'\n ]\n },\n)\n", "sub_path": "setup.py", "file_name": "setup.py", "file_ext": "py", "file_size_in_byte": 936, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "setuptools.setup", "line_number": 9, "usage_type": "call"}]} +{"seq_id": "469481088", "text": "#!/usr/bin/python3\nimport pandas\nimport glob\nimport os\nfrom datetime import datetime\n\ndef dateFunc(date_str, format=0):\n dateformats = [\"%Y-%m-%dT%H:%M:%S\", \"%Y-%m-%d %H:%M:%S\", \"%m/%d/%y %H:%M\", \"%m/%d/%Y %H:%M\"]\n try:\n return datetime.strptime(date_str, dateformats[format])\n except:\n return dateFunc(date_str, format+1)\n\ndef toTimestamp(x):\n try:\n return datetime.timestamp(x)\n except TypeError:\n return 0\n\ndef getDataForContry(country):\n cwd = os.getcwd()\n all_files = glob.glob(cwd + \"/csse_covid_19_data/csse_covid_19_daily_reports/*.csv\")\n li = []\n\n for filename in all_files:\n df = pandas.read_csv(filename, index_col=None, header=0)\n df = df.rename(columns={'Country/Region': 'Country_Region', 'Last Update': 'Last_Update'})\n df = df.loc[df['Country_Region'] == country].filter([\"Country_Region\", \"Last_Update\", \"Confirmed\", \"Deaths\", \"Recovered\", \"Active\"])\n li.append(df)\n\n data = pandas.concat(li, axis=0, ignore_index=True, sort=False)\n\n return data\n\ndef addTimeStampAndSortIt(data):\n germany['Timestamps'] = germany['Last_Update']\n germany['Timestamps'] = germany['Timestamps'].apply(dateFunc).apply(toTimestamp)\n return germany.sort_values(by=['Timestamps'])\n\ngermany = getDataForContry('Germany')\n\ngermany = addTimeStampAndSortIt(germany)\n\ngermany.to_csv('germany.csv', sep='\\t', encoding='utf-8')", "sub_path": "germany.py", "file_name": "germany.py", "file_ext": "py", "file_size_in_byte": 1411, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "datetime.datetime.strptime", "line_number": 10, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 10, "usage_type": "name"}, {"api_name": "datetime.datetime.timestamp", "line_number": 16, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 16, "usage_type": "name"}, {"api_name": "os.getcwd", "line_number": 21, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 22, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 26, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 31, "usage_type": "call"}]} +{"seq_id": "21510328", "text": "#\"\"\"\n#The flask application package.\n#\"\"\"\n\n#from flask import Flask\n#app = Flask(__name__)\n\n#import ai_api.views\n\n#!/usr/bin/python\n# Updated as of 13th April 2020\n\nimport os\nfrom flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom os import environ\nimport sys\nfrom flask_swagger_ui import get_swaggerui_blueprint\n\ndb = SQLAlchemy()\n\ndef create_app(test_config=None):\n # create and configure the app\n app = Flask(__name__, instance_relative_config=False)\n\n ### swagger specific ###\n SWAGGER_URL = '/dtiapi'\n API_URL = '/static/swagger.json'\n SWAGGERUI_BLUEPRINT = get_swaggerui_blueprint(\n SWAGGER_URL,\n API_URL,\n config={\n 'app_name': \"DTI APIs\"\n }\n )\n app.register_blueprint(SWAGGERUI_BLUEPRINT, url_prefix=SWAGGER_URL)\n ### end swagger specific ###\n \n env = sys.argv[1] if len(sys.argv) > 1 else 'development'\n config_dict = {\n 'local': 'config.LocalConfig',\n 'development': 'config.DevelopmentConfig',\n 'uat': 'config.UATConfig',\n 'production': 'config.ProductionConfig',\n 'demo' : 'config.DemoConfig'\n }\n config = config_dict[env]\n app.config.from_object(config)\n try:\n os.makedirs(app.instance_path)\n except OSError:\n pass\n db.init_app(app)\n\n\n with app.app_context():\n from ai_api.models import (User, Role)\n from ai_api.routes import (login, register, api_trigger)\n from ai_api.decorators import before_request, after_request\n \n db.create_all()\n #print('tables created')\n return app\n", "sub_path": "ai_api/ai_api/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 1481, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "flask_sqlalchemy.SQLAlchemy", "line_number": 20, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 24, "usage_type": "call"}, {"api_name": "flask_swagger_ui.get_swaggerui_blueprint", "line_number": 29, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 39, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 50, "usage_type": "call"}]} +{"seq_id": "468238660", "text": "# Python program to swap nodes \n\nfrom memory_profiler import profile\n\n# def func():\n# your function\n\n# A binary tree node \nclass Node: \n\n\t# constructor to create a new node \n\tdef __init__(self, data): \n\t\tself.data = data \n\t\tself.left = None\n\t\tself.right = None\n\n# A utility function swap left node and right node of tree \n# of every k'th level \n@profile(precision=4)\n\ndef swapEveryKLevelUtil(root, level, k): \n\t\n\t# Base Case \n\tif (root is None or (root.left is None and\n\t\t\t\t\t\troot.right is None ) ): \n\t\treturn\n\n\t# If current level+1 is present in swap vector \n\t# then we swap left and right node \n\tif (level+1)%k == 0: \n\t\troot.left, root.right = root.right, root.left \n\t\n\t# Recur for left and right subtree \n\tswapEveryKLevelUtil(root.left, level+1, k) \n\tswapEveryKLevelUtil(root.right, level+1, k) \n\n\t\n# This function mainly calls recursive function \n# swapEveryKLevelUtil \ndef swapEveryKLevel(root, k): \n\t\n\t# Call swapEveryKLevelUtil function with \n\t# initial level as 1 \n\tswapEveryKLevelUtil(root, 1, k) \n\n# Method to find the inorder tree travesal \ndef inorder(root): \n\t\n\t# Base Case \n\tif root is None: \n\t\treturn\n\tinorder(root.left) \n\tprint (root.data,) \n\tinorder(root.right) \n\n# Driver code \n\"\"\" \n\t\t1 \n\t\t/ \\ \n\t2\t 3 \n\t/\t / \\ \n\t4\t 7 8 \n\"\"\"\nroot = Node(1) \nroot.left = Node(2) \nroot.right = Node(3) \nroot.left.left = Node(4) \nroot.right.right = Node(8) \nroot.right.left = Node(7) \n\nk = 1\nprint (\"Before swap node :\")\ninorder(root) \n\nswapEveryKLevel(root, k) \n\nprint (\"\\nAfter swap Node : \")\ninorder(root) \n\n# This code is contributed by Nikhil Kumar Singh(nickzuck_007) \n", "sub_path": "data_structures_and_algorithms/challenges/training/swape_sides.py", "file_name": "swape_sides.py", "file_ext": "py", "file_size_in_byte": 1579, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "memory_profiler.profile", "line_number": 19, "usage_type": "call"}]} +{"seq_id": "499402936", "text": "import cv2\nimport time\nimport os\nimport sys\n\ndef getOutputsNames(net):\n # Get the names of all the layers in the network\n layersNames = net.getLayerNames()\n # Get the names of the output layers, i.e. the layers with unconnected outputs\n return [layersNames[i[0] - 1] for i in net.getUnconnectedOutLayers()]\n\ndef main():\n start_time = time.time()\n checkpoint_pb = '/media/common/DOWNLOADS/UBUNTU/OpenCV-IE/TEST/graph.pb'\n tensorflowNet = cv2.dnn.readNetFromTensorflow(checkpoint_pb)\n tensorflowNet.setPreferableBackend(cv2.dnn.DNN_BACKEND_INFERENCE_ENGINE)\n tensorflowNet.setPreferableTarget(cv2.dnn.DNN_TARGET_CPU)\n\n picture_path = '/media/common/DOWNLOADS/UBUNTU/OpenCV-IE/TEST/77428.png'\n picture = cv2.imread(picture_path)\n picture = cv2.cvtColor(picture, cv2.COLOR_BGR2GRAY)\n picture = cv2.resize(picture, (640, 360))\n\n blob = cv2.dnn.blobFromImage(picture, 1 / 255, (640, 360), [0, 0, 0], 1, crop=False)\n\n # Sets the input to the network\n tensorflowNet.setInput(blob)\n\n # Runs the forward pass to get output of the output layers\n outs = tensorflowNet.forward()\n\n print('Success!')\n print(outs)\n # print(f'run time {1000*(time.time() - start_time)} msec')\n\n\nmain()\n", "sub_path": "build/TEST/foo.py", "file_name": "foo.py", "file_ext": "py", "file_size_in_byte": 1232, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "time.time", "line_number": 13, "usage_type": "call"}, {"api_name": "cv2.dnn.readNetFromTensorflow", "line_number": 15, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 15, "usage_type": "attribute"}, {"api_name": "cv2.dnn", "line_number": 16, "usage_type": "attribute"}, {"api_name": "cv2.dnn", "line_number": 17, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 20, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 21, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 21, "usage_type": "attribute"}, {"api_name": "cv2.resize", "line_number": 22, "usage_type": "call"}, {"api_name": "cv2.dnn.blobFromImage", "line_number": 24, "usage_type": "call"}, {"api_name": "cv2.dnn", "line_number": 24, "usage_type": "attribute"}]} +{"seq_id": "472278110", "text": "# Copyright 2019 Red Hat\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nfrom __future__ import absolute_import\n\nfrom tobiko.openstack.heat import _client\nfrom tobiko.openstack.heat import _template\nfrom tobiko.openstack.heat import _resource\nfrom tobiko.openstack.heat import _stack\n\nheat_client = _client.heat_client\ndefault_heat_client = _client.default_heat_client\nget_heat_client = _client.get_heat_client\nHeatClient = _client.HeatClient\nHeatClientFixture = _client.HeatClientFixture\nHeatClientType = _client.HeatClientType\n\nheat_template = _template.heat_template\nheat_template_file = _template.heat_template_file\nHeatTemplateFixture = _template.HeatTemplateFixture\nHeatTemplateFileFixture = _template.HeatTemplateFileFixture\n\nRESOURCE_CLASSES = _resource.RESOURCE_CLASSES\nResourceType = _resource.ResourceType\nfind_resource = _resource.find_resource\nlist_resources = _resource.list_resources\n\nStackType = _stack.StackType\nHeatStackFixture = _stack.HeatStackFixture\nHeatStackNotFound = _stack.HeatStackNotFound\nheat_stack_parameters = _stack.heat_stack_parameters\nfind_stack = _stack.find_stack\nlist_stacks = _stack.list_stacks\nInvalidStackError = _stack.InvalidStackError\nINIT_IN_PROGRESS = _stack.INIT_IN_PROGRESS\nINIT_COMPLETE = _stack.INIT_COMPLETE\nCREATE_IN_PROGRESS = _stack.CREATE_IN_PROGRESS\nCREATE_COMPLETE = _stack.CREATE_COMPLETE\nCREATE_FAILED = _stack.CREATE_FAILED\nDELETE_IN_PROGRESS = _stack.DELETE_IN_PROGRESS\nDELETE_COMPLETE = _stack.DELETE_COMPLETE\nDELETE_FAILED = _stack.DELETE_FAILED\nSTACK_CLASSES = _stack.STACK_CLASSES\n", "sub_path": "tobiko/openstack/heat/__init__.py", "file_name": "__init__.py", "file_ext": "py", "file_size_in_byte": 2072, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "tobiko.openstack.heat._client.heat_client", "line_number": 21, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 21, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._client.default_heat_client", "line_number": 22, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 22, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._client.get_heat_client", "line_number": 23, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 23, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._client.HeatClient", "line_number": 24, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 24, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._client.HeatClientFixture", "line_number": 25, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 25, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._client.HeatClientType", "line_number": 26, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._client", "line_number": 26, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._template.heat_template", "line_number": 28, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._template", "line_number": 28, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._template.heat_template_file", "line_number": 29, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._template", "line_number": 29, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._template.HeatTemplateFixture", "line_number": 30, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._template", "line_number": 30, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._template.HeatTemplateFileFixture", "line_number": 31, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._template", "line_number": 31, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._resource.RESOURCE_CLASSES", "line_number": 33, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._resource", "line_number": 33, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._resource.ResourceType", "line_number": 34, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._resource", "line_number": 34, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._resource.find_resource", "line_number": 35, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._resource", "line_number": 35, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._resource.list_resources", "line_number": 36, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._resource", "line_number": 36, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.StackType", "line_number": 38, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 38, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.HeatStackFixture", "line_number": 39, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 39, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.HeatStackNotFound", "line_number": 40, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 40, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.heat_stack_parameters", "line_number": 41, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 41, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.find_stack", "line_number": 42, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 42, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.list_stacks", "line_number": 43, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 43, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.InvalidStackError", "line_number": 44, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 44, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.INIT_IN_PROGRESS", "line_number": 45, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 45, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.INIT_COMPLETE", "line_number": 46, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 46, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.CREATE_IN_PROGRESS", "line_number": 47, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 47, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.CREATE_COMPLETE", "line_number": 48, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 48, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.CREATE_FAILED", "line_number": 49, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 49, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.DELETE_IN_PROGRESS", "line_number": 50, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 50, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.DELETE_COMPLETE", "line_number": 51, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 51, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.DELETE_FAILED", "line_number": 52, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 52, "usage_type": "name"}, {"api_name": "tobiko.openstack.heat._stack.STACK_CLASSES", "line_number": 53, "usage_type": "attribute"}, {"api_name": "tobiko.openstack.heat._stack", "line_number": 53, "usage_type": "name"}]} +{"seq_id": "20427408", "text": "\nimport json\nfrom collections import OrderedDict\n\n\ndef check_name(name):\n \n check_name_dict = {'to': 'tohex'}\n if name in check_name_dict:\n return check_name_dict[name]\n else:\n return name\n\n\ndef make_php(method, params):\n result = '\\t//==================== \\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + '$' + check_name(key) + ', '\n result = result + '\\tfunction ' + method + '(' + tmp + ' &$data){ \\n'\n result = result + '\\t\\t$js = $this->make_json (\\'' + method + '\\', array( '\n tmp = '\\n'\n for key, value in params.items():\n tmp = tmp + '\\t\\t\\t\"' + key + '\" => $' + check_name(key) + ',\\n'\n\n result = result + tmp[:-2]\n result = result + '));\\n'\n result = result + '\\t\\t return $this->request($js, $data); \\n'\n result = result + '\\t} \\n'\n \n return result\n\n\ndef make_ruby(method, params):\n result = '\\t#==================== \\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + '' + check_name(key) + ', '\n tmp = tmp[:-2]\n result = result + '\\tdef ' + method + '(' + tmp + ' ) \\n'\n result = result + '\\t\\tjson=make_json(\\'' + method + '\\',{\\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + '\\t\\t\\t\\'' + key + '\\' : ' + check_name(key) + ',\\n'\n\n result = result + tmp[:-2] + '\\n'\n result = result + '\\t\\t})\\n'\n result = result + '\\t\\tputs (\\'' + method + ' method call\\' )\\n'\n result = result + '\\t\\tresult = make_request(json) \\n'\n result = result + '\\t\\treturn result\\n'\n result = result + '\\tend\\n'\n return result\n\n\ndef make_python(method, params):\n result = ' \\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + ', ' + check_name(key) + ''\n\n result = result + ' def ' + method + '(self' + tmp + '): \\n'\n result = result + ' data = {\"jsonrpc\": \"2.0\", \"method\": \"' + method + '\",\\n'\n result = result + ' \"params\": {\\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + ' \\'' + key + '\\': '+check_name(key) + ',\\n'\n\n result = result + tmp[:-2] + '\\n'\n result = result + ' },\\n'\n result = result + ' \"filter\": self.genericfilter\\n'\n result = result + ' }\\n'\n result = result + ' logging.info(u\\'' + method + ' method call\\')\\n'\n result = result + ' return self.send_request(data)\\n'\n return result\n\n\ndef make_perl(method, params):\n result = 'sub '+method+' { \\n'\n result = result + '\\tmy $self = shift;\\n'\n tmp = ''\n for key, value in params.items():\n tmp = tmp + '\\tmy $' + check_name(key) + '=shift;\\n'\n result = result+tmp\n tmp = ''\n for key, value in params.items():\n tmp = tmp + '\\'' + key + '\\' => $' + check_name(key) + ','\n\n result = result + '\\tmy %params = (' + tmp[:-1] + ');\\n'\n result = result + '\\tmy $result = $self->postApi(\\'' + method + \"\\', \\%params);\\n\"\n result = result + '\\treturn $result;\\n'\n result = result + '}\\n'\n return result\n\n\ndef make_bash(method, params):\n result = '\\n'\n result = result + '' + method + '() {\\n'\n result = result + 'param=$(./jq -n \"{'\n tmp = ''\n argument_id = 1\n for key, value in params.items():\n tmp = tmp + '' + key + ': \\\\\"$' + str(id) + '\\\\\",'\n argument_id = argument_id + 1\n\n result = result + tmp[:-1]\n result = result + '}\")\\n'\n result = result + 'jsondata=$(makejson \"' + method + '\" \"$param\")\\n'\n result = result + 'makerequest \"$jsondata\"\\n'\n result = result + '}\\n'\n return result\n\n\ndef make_asp(method, params):\n result = '\\'-----------'+method+' -------------\\n'\n\n tmp = ''\n for key, value in params.items():\n tmp = tmp + ' ' + check_name(key) + ','\n tmp = tmp[:-2]\n result = result + 'public Function api' + method + '(' + tmp + ' )\\n'\n \n tmp = ''\n for key, value in params.items():\n tmp = tmp + '\"\"' + key + '\"\":\"\"\"+' + check_name(key) + '\"\"\",'\n\n result = result + '\\tapi' + method + ' = apiRequest(Session(\"mySessionId\"), \"'\\\n + method + '\", \"{ ' + tmp[:-1] + '}\")\\n'\n result = result + 'end function\\n'\n return result\n\n\ndef generate_files(data, name, ext, filename):\n pre = open(name + '/pre.' + ext, 'r')\n post = open(name + '/post.' + ext, 'r')\n file = open(name + '/' + filename + '.' + ext, 'w')\n\n file.write(pre.read())\n for method in data:\n file.write(globals()[\"make_\"+lang](method['method'], method['params'] if 'params' in method else {}))\n \n file.write(post.read())\n file.close()\n pre.close()\n post.close()\n\n\nif __name__ == '__main__':\n \n with open('methods.json') as json_file: \n methods = json.load(json_file, object_pairs_hook=OrderedDict)\n\n api_name = \"sample\"\n\n languages = {'asp': 'inc', 'bash': 'sh', \"perl\": \"pm\",\n \"python\": \"py\", \"ruby\": \"rb\", \"php\": \"php\"}\n\n for lang in languages:\n generate_files(methods, lang, languages[lang], api_name)\n", "sub_path": "generator.py", "file_name": "generator.py", "file_ext": "py", "file_size_in_byte": 5056, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "json.load", "line_number": 150, "usage_type": "call"}, {"api_name": "collections.OrderedDict", "line_number": 150, "usage_type": "name"}]} +{"seq_id": "376339164", "text": "# -*- coding: utf-8 -*-\n# sample dialog\n\nimport wx\nimport globalVars\nimport views.ViewCreator\nfrom logging import getLogger\nfrom simpleDialog import *\nfrom views.baseDialog import *\n\nclass Dialog(BaseDialog):\n\tdef __init__(self):\n\t\tsuper().__init__()\n\t\tself.result = \"\"\n\t\tself.list = []\n\t\tself.tesseract_flag= False# Trueならスペース置換ボタンが有効になる\n\n\tdef Initialize(self):\n\t\tself.identifier=\"converted result dialog\"#このビューを表す文字列\n\t\tself.log=getLogger(self.identifier)\n\t\tself.log.debug(\"created\")\n\t\tsuper().Initialize(self.app.hMainView.hFrame,_(\"認識結果\"))\n\t\tself.InstallControls()\n\t\treturn True\n\n\tdef InstallControls(self):\n\t\t\"\"\"いろんなwidgetを設置する。\"\"\"\n\t\tself.creator=views.ViewCreator.ViewCreator(self.viewMode,self.panel,self.sizer,wx.VERTICAL,20)\n\t\tself.resultView,static = self.creator.inputbox(\"認識結果\", x=800,defaultValue=self.result, style=wx.TE_MULTILINE|wx.TE_READONLY|wx.TE_DONTWRAP)\n\t\tself.repButton = self.creator.button(_(\"スペースを置換\"), self.onRep)\n\t\tif self.tesseract_flag == False:\n\t\t\tself.repButton.Disable()\n\t\tself.bOk=self.creator.okbutton(_(\"OK\"), None)\n\n\tdef onRep(self, event):\n\t\tif qDialog(_(\"余計なスペースをすべて置換します。文章が崩れる可能性があります。よろしいですか?\")) == wx.ID_NO:\n\t\t\treturn\n\t\tsaved = \"\"\n\t\tfor path in self.list:\n\t\t\ttext=path.read_text().replace(\" \", \"\")\n\t\t\tpath.write_text(text)\n\t\t\tsaved += text\n\t\tdialog(_(\"置換が完了しました。\"), _(\"置換\"))\n\t\tself.resultView.SetValue(saved)\n\n\tdef Destroy(self, events = None):\n\t\tself.log.debug(\"destroy\")\n\t\tself.wnd.Destroy()\n\n\t#def GetData(self):\n\t#\treturn self.iText.GetLineText(0)\n", "sub_path": "views/converted.py", "file_name": "converted.py", "file_ext": "py", "file_size_in_byte": 1704, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "logging.getLogger", "line_number": 20, "usage_type": "call"}, {"api_name": "views.ViewCreator.ViewCreator.ViewCreator", "line_number": 28, "usage_type": "call"}, {"api_name": "views.ViewCreator.ViewCreator", "line_number": 28, "usage_type": "attribute"}, {"api_name": "views.ViewCreator", "line_number": 28, "usage_type": "name"}, {"api_name": "wx.VERTICAL", "line_number": 28, "usage_type": "attribute"}, {"api_name": "wx.TE_MULTILINE", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wx.TE_READONLY", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wx.TE_DONTWRAP", "line_number": 29, "usage_type": "attribute"}, {"api_name": "wx.ID_NO", "line_number": 36, "usage_type": "attribute"}]} +{"seq_id": "411736475", "text": "import argparse\n\n__author__ = 'koder'\n\nclass Opt(object):\n def __init__(self, tp, doc=\"\", default=None):\n self.tp = tp\n self.doc = doc\n self.name = None\n self.default = default\n\n def add_to_parser(self, parser):\n if self.default is not None:\n parser.add_argument(\"--\" + self.name,\n dest=self.name,\n default=self.default)\n else:\n parser.add_argument(\"--\" + self.name,\n dest=self.name)\n\nclass TypedOpt(Opt):\n def_tp = None\n\n def __init__(self, *dt, **mp):\n super(TypedOpt, self).__init__(self.def_tp, *dt, **mp)\n\nclass StrOpt(TypedOpt):\n def_tp = str\n\nclass IntOpt(TypedOpt):\n def_tp = int\n\nclass PyOptParser(object):\n _parser = None\n @classmethod\n def parse(cls):\n parser = cls.get_parser()\n\n @classmethod\n def fields(cls):\n for name, obj in cls.__dict__.items():\n if isinstance(obj, Opt):\n yield name, obj\n\n @classmethod\n def get_parser(cls):\n if cls._parser is None:\n parser = argparse.ArgumentParser()\n for name, obj in cls.fields():\n obj.name = name\n obj.add_to_parser(parser)\n cls._parser = parser\n return cls._parser\n\n @classmethod\n def parse_opts(cls, opts=None):\n obj = cls()\n parser = cls.get_parser()\n options = parser.parse_args(opts)\n for name, _ in cls.fields():\n setattr(obj, name, getattr(options, name))\n return obj\n\n\n", "sub_path": "disk_test/easy_opt.py", "file_name": "easy_opt.py", "file_ext": "py", "file_size_in_byte": 1557, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 48, "usage_type": "call"}]} +{"seq_id": "527658715", "text": "import time\nimport logging\nfrom flask import Flask, jsonify, make_response, request, abort, render_template\n\nimport sys\nimport os\nsys.path.append(os.path.abspath(os.path.join('..', '')))\n\nfrom db_manager import master_list\nfrom cache_manager import CacheManager\nfrom config import device_id\n\nformatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nlogging.basicConfig(level=logging.INFO)\nhandler = logging.FileHandler('webapp_log.log')\nhandler.setFormatter(formatter)\nlogger = logging.getLogger(__name__)\nlogger.addHandler(handler)\n\napp = Flask(__name__, static_folder=\"./static\", template_folder=\"./static\")\nch = CacheManager()\n\n@app.route(\"/list\")\ndef get_list():\n return \"Master List:{}\".format(master_list)\n #return render_template(\"index.html\")\n\n@app.route(\"/ping\")\ndef hello():\n return \"PONG\"\n\n@app.route(\"/mylist\", methods=[\"GET\"])\ndef get_my_list():\n keys = ch.get_keys(\"{}*\".format(device_id))\n for key in keys:\n data = {\"id\": str(key.split(\":\")[-1]), \"name\": ch.get_value(key)}\n return jsonify(str(data))\n\n@app.route(\"/mylist\", methods=[\"DELETE\"])\ndef delete_my_list():\n r = ch.delete_key(\"{}:{}\".format(device_id, \"mylist\"))\n if r != 0:\n print(\"Failed to delete\")\n return jsonify(\"\")\n\n@app.errorhandler(404)\ndef not_found(error):\n return make_response(jsonify({'error': 'Not found'}), 404)\n\n@app.errorhandler(400)\ndef not_found(error):\n return make_response(jsonify({'error': 'Bad request'}), 400)\n\n'''\nif __name__ == '__main__':\n app.run(debug=True, host='0.0.0.0')\n'''", "sub_path": "src/webbapp/app.py", "file_name": "app.py", "file_ext": "py", "file_size_in_byte": 1561, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "sys.path.append", "line_number": 7, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 7, "usage_type": "call"}, {"api_name": "os.path", "line_number": 7, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 7, "usage_type": "call"}, {"api_name": "logging.Formatter", "line_number": 13, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 14, "usage_type": "call"}, {"api_name": "logging.INFO", "line_number": 14, "usage_type": "attribute"}, {"api_name": "logging.FileHandler", "line_number": 15, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 17, "usage_type": "call"}, {"api_name": "flask.Flask", "line_number": 20, "usage_type": "call"}, {"api_name": "cache_manager.CacheManager", "line_number": 21, "usage_type": "call"}, {"api_name": "db_manager.master_list", "line_number": 25, "usage_type": "argument"}, {"api_name": "config.device_id", "line_number": 34, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 37, "usage_type": "call"}, {"api_name": "config.device_id", "line_number": 41, "usage_type": "argument"}, {"api_name": "flask.jsonify", "line_number": 44, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 48, "usage_type": "call"}, {"api_name": "flask.make_response", "line_number": 52, "usage_type": "call"}, {"api_name": "flask.jsonify", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "404980879", "text": "\nimport os\nimport argparse\nimport cv2\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport json\n\ndef load_poses_from_json(json_filename):\n\n with open(json_filename) as data_file:\n loaded = json.load(data_file)\n poses, conf = json_to_poses(loaded)\n\n if len(poses) != 1:\n return None, None\n else:\n return poses, conf\n\ndef json_to_poses(js_data): \n\n poses = []\n confidences = []\n\n for arr in js_data[\"people\"]:\n confidences.append(arr[\"pose_keypoints_2d\"][2::3])\n arr = np.delete(arr[\"pose_keypoints_2d\"], slice(2, None, 3))\n poses.append(list(zip(arr[::2],arr[1::2])))\n\n\n return poses, confidences\n\n# def draw_annotations(img, annot):\n\n# result = img.copy()\n\n# for a in annot:\n\n# point1 = np.array(list(map(float,a[6:8])))\n# point2 = np.array(list(map(float,a[8:10])))\n\n# point1[0] *= img.shape[1]\n# point1[1] *= img.shape[0]\n# point2[0] *= img.shape[1]\n# point2[1] *= img.shape[0]\n\n# vector = point2 - point1\n# versor = vector / np.linalg.norm(vector)\n\n# versor[0] *= 60\n# versor[1] *= 60\n\n# new_point = versor + point1\n\n# # eyes center\n# cv2.circle(result, (int(float(a[6])*img.shape[1]), int(float(a[7])*img.shape[0])), 10, (0,0,0), 6) \n# cv2.circle(result, (int(float(a[6])*img.shape[1]), int(float(a[7])*img.shape[0])), 10, (0,50,255), 3) \n\n# # gaze point\n# cv2.circle(result, (int(float(a[8])*img.shape[1]), int(float(a[9])*img.shape[0])), 10, (0,0,0), 6) \n# cv2.circle(result, (int(float(a[8])*img.shape[1]), int(float(a[9])*img.shape[0])), 10, (255,0,0), 3) \n\n# # arrow\n# cv2.arrowedLine(result, (int(float(a[6])*img.shape[1]),(int(float(a[7])*img.shape[0]))), (int(new_point[0]), int(new_point[1])), (0,0,0), thickness=7, tipLength=0.2)\n# cv2.arrowedLine(result, (int(float(a[6])*img.shape[1]),(int(float(a[7])*img.shape[0]))), (int(new_point[0]), int(new_point[1])), (0,255,0), thickness=3, tipLength=0.2)\n\n# return result\n\n\n\n# arguments\n\nparser = argparse.ArgumentParser(description='Classification')\n\nparser.add_argument('--dir', type=str, required=True)\n# parser.add_argument('--id', type=int, required=True)\n\nargs = parser.parse_args()\n\ncount = 0\n\nroot_dir = \"/media/damiano/Data/Datasets/GazeFollow/data\" \n\noutput_filename = root_dir + '/one_subject.txt'\n\nwith open(output_filename, 'w') as output_file:\n\n for identifier in range(0, 119125):\n\n # filename adaptation\n\n top_level_dir = \"%08d\" % (int(identifier/1000))\n file_id_8digit = \"%08d\" % (identifier)\n\n\n json_dir = os.path.join(root_dir, 'json', args.dir, top_level_dir)\n img_dir = os.path.join(root_dir, 'rendered', args.dir, top_level_dir)\n\n json_filename = json_dir + '/' + file_id_8digit + '_keypoints.json'\n\n if os.path.exists(json_filename):\n \n poses, conf = load_poses_from_json(json_filename)\n\n if poses != None:\n \n count += 1\n\n output_file.write('%d\\n' % identifier)\n\noutput_file.close()\n\nprint(\"Total number of samples with one subject: %d\" % count)\n\n# annotations loading\n\n# target_header = os.path.join(args.dir, top_level_dir, file_id_8digit + '.jpg')\n\n# train_annotations = '/media/damiano/Data/Datasets/GazeFollow/data/train_annotations.txt'\n# train_file = open(train_annotations, \"r\")\n\n# contents = train_file.readlines()\n# contents = [x.strip() for x in contents]\n# contents_split = [x.split(\",\") for x in contents]\n# contents_reduced = [line for line in contents_split if line[0] == target_header]\n\n# print(\"Total number of lines in the %s_annotations.txt file: %d\" % (args.dir,len(contents)) )\n# print(\"Number of annotations for the selected image: %d\" % len(contents_reduced))\n\n\n\n# # show image\n\n# img_filename = img_dir + '/' + file_id_8digit + '_rendered.png'\n\n# img = cv2.imread(img_filename)\n# img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)\n\n# img = draw_annotations(img, contents_reduced)\n\n# print(img_filename)\n\n# plt.figure()\n# plt.imshow(img)\n# plt.axis('off')\n# plt.ion()\n# plt.show()\n# plt.pause(5)\n\n", "sub_path": "src/gazefollow_count_one_subject.py", "file_name": "gazefollow_count_one_subject.py", "file_ext": "py", "file_size_in_byte": 4167, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "json.load", "line_number": 12, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 27, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 73, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 96, "usage_type": "call"}, {"api_name": "os.path", "line_number": 96, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 97, "usage_type": "call"}, {"api_name": "os.path", "line_number": 97, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 101, "usage_type": "call"}, {"api_name": "os.path", "line_number": 101, "usage_type": "attribute"}]} +{"seq_id": "205199772", "text": "\"\"\"Module for handling http requests.\"\"\"\n\nimport json\n\nfrom six.moves import BaseHTTPServer\nfrom ..models.response import ControllerResponse\n\n\nclass WebRequestHandler(BaseHTTPServer.BaseHTTPRequestHandler):\n \"\"\"Class for handling http requests.\"\"\"\n\n GET_METHOD = 'GET'\n POST_METHOD = 'POST'\n timeout = None\n\n def do_GET(self):\n \"\"\"Handling GET requests.\"\"\"\n # pylint: disable=invalid-name\n self.handle_method(self.GET_METHOD)\n\n def do_POST(self):\n \"\"\"Handling POST requests.\"\"\"\n # pylint: disable=invalid-name\n self.handle_method(self.POST_METHOD)\n\n def get_payload(self, max_size):\n \"\"\"Getting payload.\"\"\"\n # pylint: disable=protected-access\n self.timeout = (self.server.options.read_timeout_milliseconds/1000)\n payload_len = int(self.headers.get('content-length', 0))\n payload_len = min(payload_len, max_size)\n payload = self.rfile.read(payload_len)\n return payload\n\n def handle_method(self, method):\n \"\"\"Handling arbitrary requests.\"\"\"\n # pylint: disable=protected-access\n self.timeout = (self.server.options.write_timeout_milliseconds/1000)\n controller = self.server.routes_repo.get_controller(method, self.path, self.server.options)\n data = None\n if method == self.POST_METHOD:\n data = self.get_payload(self.server.options.max_request_size)\n result = controller.handle(data, None)\n if result.jsonize:\n result = ControllerResponse(\n response=json.dumps(result.response),\n status=result.status,\n mime=result.mime\n )\n self.send_response(result.status)\n self.send_header('Content-type', result.mime)\n self.end_headers()\n self.wfile.write(result.response.encode())\n", "sub_path": "activities_python/common/http/web_request.py", "file_name": "web_request.py", "file_ext": "py", "file_size_in_byte": 1842, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "six.moves.BaseHTTPServer.BaseHTTPRequestHandler", "line_number": 9, "usage_type": "attribute"}, {"api_name": "six.moves.BaseHTTPServer", "line_number": 9, "usage_type": "name"}, {"api_name": "models.response.ControllerResponse", "line_number": 45, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 46, "usage_type": "call"}]} +{"seq_id": "478207457", "text": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\nimport urllib\nimport xml.etree.ElementTree as ET\nimport re\nimport json\nimport requests\nimport time\nfrom datetime import datetime\nfrom bs4 import BeautifulSoup\nfrom elasticsearch import Elasticsearch\nfrom multiprocessing import Pool\nimport urllib.request\n\nes_client = Elasticsearch(['172.18.1.96:9200'])\n\n\ndef main():\n #testfile = urllib.URLopener()\n #testfile.retrieve('https://s3.amazonaws.com/medios.plazavip.com/mktclaroshop/catGoogle.xml', 'claro.xmls')\n testfile = urllib.request.urlretrieve('https://s3.amazonaws.com/medios.plazavip.com/mktclaroshop/catGoogle.xml', 'claro.xmls')\n\n tree = ET.parse('claro.xmls')\n root = tree.getroot()\n\n ns = {'tag_item': 'http://base.google.com/ns/1.0'}\n\n listItems = []\n\n for channel in root.findall('channel'):\n for item in channel.findall('item'):\n listItems.append(ET.tostring(item).decode())\n\n p = Pool(4)\n p.map(urlparser, listItems)\n p.terminate()\n p.join()\n\n\ndef urlparser(item):\n item = ET.fromstring(item)\n\n # item = tree.getroot()\n\n ns = {'tag_item': 'http://base.google.com/ns/1.0'}\n\n id = item.find('tag_item:id', ns)\n\n # print(id.text)\n\n title = item.find('title')\n\n # print(title.text)\n\n link = item.find('link')\n\n # print(link.text)\n\n try:\n page = requests.get(link.text)\n except requests.exceptions.RequestException as e:\n return\n else:\n statusCode = page.status_code\n if statusCode == 200:\n soup = BeautifulSoup(page.content, 'lxml')\n description = soup.find_all('ul', {'class': 'viewDescrip'})\n if len(description) != 0:\n li = description[0].find_all('li', {'class': 'laDescrip'})\n try:\n li[0].text\n except Exception as e:\n return\n else:\n #print(li[0].text)\n pass\n else:\n return\n else:\n\n # print(li[0].text)\n\n return\n image_link = item.find('tag_item:image_link', ns)\n\n # print(image_link.text)\n\n condition = item.find('tag_item:condition', ns)\n\n # print(condition.text)\n\n availability = item.find('tag_item:availability', ns)\n\n # print(availability.text)\n\n price = item.find('tag_item:price', ns)\n price_vector = re.findall(\"[-+]?\\d*\\.\\d+|\\d+\", price.text)\n\n # print(price_vector[0])\n\n sale_price = item.find('tag_item:sale_price', ns)\n sale_price_vector = re.findall(\"[-+]?\\d*\\.\\d+|\\d+\", sale_price.text)\n\n # print(sale_price_vector[0])\n\n brand = item.find('tag_item:brand', ns)\n\n # print(brand.text)\n\n product_type = item.find('tag_item:product_type', ns)\n product_type_vector = product_type.text.split(' > ')\n\n # print(product_type_vector)\n\n google_product_category = \\\n item.find('tag_item:google_product_category', ns)\n google_product_category = google_product_category.text\n google_product_category_vector = google_product_category.split(' > '\n )\n\n # print(google_product_category_vector)\n\n gtin = item.find('tag_item:gtin', ns)\n\n # print(gtin.text)\n\n product_digital = item.find('tag_item:product_digital', ns)\n\n # print(product_digital.text)\n\n installment_json = {}\n if item.findall('tag_item:installment', ns):\n for installment in item.findall('tag_item:installment', ns):\n months = installment.find('tag_item:months', ns)\n installment_json['months'] = months.text\n\n # print(months.text)\n\n amount = installment.find('tag_item:amount', ns)\n amount_vector = re.findall(\"[-+]?\\d*\\.\\d+|\\d+\", amount.text)\n\n # print(amount_vector[0])\n\n installment_json['amount'] = amount_vector[0]\n\n # print(json.dumps(installment_json))\n\n doc = {\n 'installment': installment_json,\n 'title': title.text,\n 'description': li[0].text,\n 'link': link.text,\n 'image_link': image_link.text,\n 'condition': condition.text,\n 'availability': availability.text,\n 'price': float(price_vector[0]),\n 'sale_price': float(sale_price_vector[0]),\n 'brand': brand.text,\n 'product_type': product_type_vector,\n 'google_product_category': google_product_category_vector,\n 'gtin': gtin.text,\n 'product_digital': float(product_digital.text),\n 'date': datetime.now(),\n }\n\n # print(json.dumps(doc))\n\n res = es_client.index(index='claro_shop', doc_type='default',\n id=int(id.text), body=doc)\n\n\nif __name__ == '__main__':\n main()\n\n\t\t\t", "sub_path": "scripts_python/claro.py", "file_name": "claro.py", "file_ext": "py", "file_size_in_byte": 4674, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "elasticsearch.Elasticsearch", "line_number": 15, "usage_type": "call"}, {"api_name": "urllib.request.urlretrieve", "line_number": 21, "usage_type": "call"}, {"api_name": "urllib.request", "line_number": 21, "usage_type": "attribute"}, {"api_name": "xml.etree.ElementTree.parse", "line_number": 23, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 23, "usage_type": "name"}, {"api_name": "xml.etree.ElementTree.tostring", "line_number": 32, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 32, "usage_type": "name"}, {"api_name": "multiprocessing.Pool", "line_number": 34, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree.fromstring", "line_number": 41, "usage_type": "call"}, {"api_name": "xml.etree.ElementTree", "line_number": 41, "usage_type": "name"}, {"api_name": "requests.get", "line_number": 60, "usage_type": "call"}, {"api_name": "requests.exceptions", "line_number": 61, "usage_type": "attribute"}, {"api_name": "bs4.BeautifulSoup", "line_number": 66, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 97, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 102, "usage_type": "call"}, {"api_name": "re.findall", "line_number": 140, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 163, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 163, "usage_type": "name"}]} +{"seq_id": "63606750", "text": "from model import create_model\nfrom sklearn.model_selection import train_test_split\nfrom tensorflow.keras.callbacks import CSVLogger\nfrom tensorflow.keras.callbacks import ModelCheckpoint\nfrom tensorflow.keras.callbacks import Callback as KCallback \nfrom datetime import datetime\nfrom numpy import concatenate\n##############################\n# Do not change this file :) #\n##############################\n\ndef getNormalization(normies=None):\n if not normies is None:\n # Get the importable function\n normies = normies.split(\".\")\n baseImport = __import__(normies[0], globals(), locals(), [normies[1]], 0)\n normies = getattr(baseImport, normies[1])\n return normies\n\ndef deltaToString(tme):\n sec = tme.total_seconds()\n hours = int(sec) // 60 // 60\n minutes = int(sec - hours* 60*60) // 60\n sec = sec - hours* 60*60 - minutes * 60\n return \"{:02d}:{:02d}:{:05.2f}\".format(hours, minutes, sec)\n\nclass TimeHistory(KCallback):\n def on_epoch_begin(self, epoch, logs):\n self.epoch_time_start = datetime.now()\n def on_epoch_end(self, epoch, logs):\n logs[\"epoch_time\"] = deltaToString(datetime.now() - self.epoch_time_start)\n\ndef getInputClass(x,y, iclass):\n from numpy import array, where, argmax, isin\n index = {\n 'triples': [0, 1, 2, 3, 4, 5],\n 'doubles': [6, 7],\n 'DtoT': [8, 9, 10, 11, 12, 13],\n 'false': [14]}\n if isinstance(iclass, list):\n i = list(map(lambda x: index[x], iclass))\n valid_index = []\n for d in i:\n valid_index.extend(d)\n valid_index = array(valid_index)\n else:\n valid_index = array(index[iclass])\n t = argmax(y, axis=1).reshape(y.shape[0], 1)\n z1, z2 = where(isin(t, valid_index))\n x, y = x[z1], y[z1]\n return x, y\n\ndef getDevices():\n from tensorflow import config\n devices = [device for device in config.list_physical_devices() if \"GPU\" == device.device_type]\n devices = [\"/gpu:{}\".format(i) for i, device in enumerate(devices)]\n return devices\n\ndef set_tf_config_mpi():\n from mpi4py import MPI\n comm = MPI.COMM_WORLD\n rank = comm.Get_rank()\n host = MPI.Get_processor_name()\n\n # Create the config dict\n tf_config = {}\n hosts = comm.allgather(host)\n import getPort\n prt = getPort.find_free_port()\n hosts = [\"{}:{}\".format(xhost, prt) for xhost in hosts]\n tf_config['cluster'] = {'worker': hosts}\n tf_config['task'] = {'type': 'worker', 'index': rank}\n print(rank, tf_config)\n\n # Dump into the environment for tensorflow to use\n # This is the suggested method on the docs for some reason\n from os import environ\n import json\n environ['TF_CONFIG'] = json.dumps(tf_config)\n\ndef createCallbacks(params, callbacks, rank, resume_training):\n callbacks.append(TimeHistory())\n if resume_training:\n csv = CSVLogger(\"training_log_{:02d}.csv\".format(rank), append=True)\n else:\n csv = CSVLogger(\"training_log_{:02d}.csv\".format(rank), append=False)\n callbacks.append(csv)\n checkpoints = ModelCheckpoint(\"checkpoints/model-\"+\"{epoch:05d}\", \n monitor='val_loss', period=64,verbose=0)\n callbacks.append(checkpoints)\n return callbacks\n\ndef removeDoubles(yold):\n from numpy import empty\n import numpy\n # Index 6 and 7 are the doubles to be removes\n ynew = empty((yold.shape[0],y.shape[1]-2))\n ynew[:,:6] = yold[:,:6]\n ynew[:,6:] = yold[:,8:]\n return ynew\n\ndef getInitialEpochsAndModelName(rank):\n from os import listdir\n modelName = None\n initial_epoch = 0\n items = listdir()\n # Check for an existing checkpoints folder\n if \"checkpoints\" in items:\n # Find the latest checkpoint\n checkpoints = listdir(\"checkpoints/\")\n epochNumbers = list(map(lambda s: int(s.split(\"-\")[1]), checkpoints))\n if len(epochNumbers) > 0:\n lastCheckpoint = max(epochNumbers)\n checkpointInd = epochNumbers.index(lastCheckpoint)\n modelName = \"checkpoints/{}\".format(checkpoints[checkpointInd])\n # Check if the latest checkpoint is the most recent line in the training CSV\n from pandas import read_csv\n # If this fails, you have to restart training from scratch anyways....\n CSVname = \"training_log_{:02d}.csv\".format(rank)\n trainingCSV = read_csv(CSVname)\n # If not, remove the extra lines\n trainingCSV_adj = trainingCSV[trainingCSV['epoch'] <= lastCheckpoint]\n trainingCSV_adj.to_csv(CSVname, index=False)\n initial_epoch = lastCheckpoint\n print(\"Resuming training, starting at epoch {}\".format(lastCheckpoint))\n if initial_epoch == params['epochs']:\n print(\"Fully trained? initial_epoch = {} = {} = params['epochs']\"\n .format(initial_epoch, params['epochs']))\n exit()\n # Return the latest epoch and modelName\n return initial_epoch+1, modelName\n return None, None\n\nif __name__ == \"__main__\":\n # Get mpi rank\n from getOneHot import getOneHot\n from mpi4py import MPI\n comm = MPI.COMM_WORLD\n rank = comm.Get_rank()\n \n # Load in the parameter files\n from json import load as loadf\n with open(\"params.json\", 'r') as inFile:\n params = loadf(inFile)\n\n # Get data files and prep them for the generator\n from tensorflow import distribute as D\n callbacks = []\n devices = getDevices()\n print(devices)\n set_tf_config_mpi()\n strat = D.experimental.MultiWorkerMirroredStrategy(\n communication=D.experimental.CollectiveCommunication.NCCL)\n # Create network\n from sys import argv\n resume_training = False\n print(argv)\n if \"resume_latest\" in argv:\n resume_training = True\n\n with strat.scope():\n # Scheduler\n if isinstance(params[\"learning_rate\"], str):\n # Get the string for the importable function\n lr = params[\"learning_rate\"]\n from tensorflow.keras.callbacks import LearningRateScheduler\n # Use a dummy learning rate\n params[\"learning_rate\"] = 0.1\n # model = create_model(**params)\n # Get the importable function\n lr = lr.split(\".\")\n baseImport = __import__(lr[0], globals(), locals(), [lr[1]], 0)\n lr = getattr(baseImport, lr[1])\n # Make a schedule\n lr = LearningRateScheduler(lr)\n callbacks.append(lr)\n # Resume Model?\n model_name = None\n if resume_training:\n initial_epoch, model_name = getInitialEpochsAndModelName(rank)\n if model_name is None:\n initial_epoch=0\n model = create_model(**params)\n resume_training = False\n else:\n from tensorflow.keras.models import load_model\n model = load_model(model_name)\n # Load data from disk\n import numpy\n if \"root\" in params.keys():\n root = params['root']\n else:\n root = \"./\"\n if \"filename\" in params.keys():\n filename = params[\"filename\"]\n else:\n filename = \"150MeV_all_shuffled_normed.csv\"\n\n restricted = [\n 'euc1', 'e1', 'x1', 'y1', 'z1',\n 'euc2', 'e2', 'x2', 'y2', 'z2',\n 'euc3', 'e3', 'x3', 'y3', 'z3',\n ]\n x, y = getOneHot(\"{}/{}\".format(root, filename), restricted=restricted, **params)\n # val_filename = \"150MeV_180kMUmin-stdCC_stitched_triples_dtot_trip_only.csv\"\n # val_x, val_y = getOneHot(\"{}/{}\".format(root, val_filename), restricted=restricted)\n val_x, val_y = None, None\n params[\"gbatch_size\"] = params['batch_size'] * len(devices)\n print(\"x.shape =\", x.shape)\n print(\"y.shape =\", y.shape)\n print(\"epochs =\", params['epochs'], type(params['epochs']))\n print(\"batch =\", params['batch_size'], type(params['batch_size']))\n print(\"gbatch =\", params['gbatch_size'], type(params['gbatch_size']))\n # Load data into a distributed dataset\n # Dataset object does nothing in place:\n # https://stackoverflow.com/questions/55645953/shape-of-tensorflow-dataset-data-in-keras-tensorflow-2-0-is-wrong-after-conver\n from tensorflow.data import Dataset\n data = Dataset.from_tensor_slices((x, y))\n\n # Create validation set\n v = params['validation']\n if val_x is not None:\n vrecord = val_x.shape[0]\n val = Dataset.from_tensor_slices((val_x, val_y))\n validation = val # data.take(vrecord)\n else:\n vrecord = int(x.shape[0]*v)\n validation = data.take(vrecord)\n validation = validation.batch(params['gbatch_size'])\n validation = validation.repeat(params['epochs'])\n # Validation -- need to do kfold one day\n # This set should NOT be distributed\n vsteps = vrecord // params['gbatch_size']\n if vrecord % params['gbatch_size'] != 0:\n vsteps += 1\n # Shuffle the data during preprocessing or suffer...\n # Parallel randomness == nightmare\n # data = data.shuffle(x.shape[0])\n # Ordering these two things is very important! \n # Consider 3 elements, batch size 2 repeat 2\n # [1 2 3] -> [[1 2] [3]] -> [[1 2] [3] [1 2] [3]] (correct) batch -> repeat\n # [1 2 3] -> [1 2 3 1 2 3] -> [[1 2] [3 1] [2 3]] (incorrect) repeat -> batch\n # data = data.skip(vrecord)\n data = data.batch(params['gbatch_size'])\n data = data.repeat(params['epochs'])\n records = x.shape[0] # - vrecord\n steps = records // params['gbatch_size']\n if records % params['gbatch_size']:\n steps += 1\n print(\"steps =\", steps)\n # Note that if we are resuming that the number of _remaining_ epochs has\n # changed!\n # The number of epochs * steps is the numbers of samples to drop\n print(\"initial cardinality = \", data.cardinality())\n print(\"initial v cardinality = \", data.cardinality())\n data = data.skip(initial_epoch*steps)\n validation = validation.skip(initial_epoch*vsteps)\n print(\"final cardinality = \", data.cardinality())\n print(\"final v cardinality = \", data.cardinality())\n # data = strat.experimental_distribute_dataset(data)\n # Split into validation and training\n callbacks = createCallbacks(params, callbacks, rank, resume_training)\n print(callbacks)\n\n history = model.fit(data, epochs=params['epochs'],\n batch_size=params['gbatch_size'],\n steps_per_epoch=steps,\n verbose=0, \n initial_epoch=initial_epoch,\n validation_data=validation,\n validation_steps=vsteps,\n callbacks=callbacks)\n if rank == 0:\n model.save(\"model-final\")\n else:\n model.save(\"checkpoints/model-tmp\")\n", "sub_path": "2021-projects/team-2/train/singletask/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 10657, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "tensorflow.keras.callbacks.Callback", "line_number": 27, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 29, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 29, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 31, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 31, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 47, "usage_type": "call"}, {"api_name": "numpy.argmax", "line_number": 48, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 49, "usage_type": "call"}, {"api_name": "numpy.isin", "line_number": 49, "usage_type": "call"}, {"api_name": "tensorflow.config.list_physical_devices", "line_number": 55, "usage_type": "call"}, {"api_name": "tensorflow.config", "line_number": 55, "usage_type": "name"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 61, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 61, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Get_processor_name", "line_number": 63, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 63, "usage_type": "name"}, {"api_name": "getPort.find_free_port", "line_number": 69, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 79, "usage_type": "name"}, {"api_name": "json.dumps", "line_number": 79, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.CSVLogger", "line_number": 84, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.CSVLogger", "line_number": 86, "usage_type": "call"}, {"api_name": "tensorflow.keras.callbacks.ModelCheckpoint", "line_number": 88, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 97, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 106, "usage_type": "call"}, {"api_name": "os.listdir", "line_number": 110, "usage_type": "call"}, {"api_name": "pandas.read_csv", "line_number": 120, "usage_type": "call"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 138, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 138, "usage_type": "name"}, {"api_name": "json.load", "line_number": 144, "usage_type": "call"}, {"api_name": "tensorflow.distribute.experimental.MultiWorkerMirroredStrategy", "line_number": 152, "usage_type": "call"}, {"api_name": "tensorflow.distribute.experimental", "line_number": 152, "usage_type": "attribute"}, {"api_name": "tensorflow.distribute", "line_number": 152, "usage_type": "name"}, {"api_name": "tensorflow.distribute.experimental", "line_number": 153, "usage_type": "attribute"}, {"api_name": "tensorflow.distribute", "line_number": 153, "usage_type": "name"}, {"api_name": "sys.argv", "line_number": 157, "usage_type": "argument"}, {"api_name": "sys.argv", "line_number": 158, "usage_type": "name"}, {"api_name": "tensorflow.keras.callbacks.LearningRateScheduler", "line_number": 175, "usage_type": "call"}, {"api_name": "model.create_model", "line_number": 183, "usage_type": "call"}, {"api_name": "tensorflow.keras.models.load_model", "line_number": 187, "usage_type": "call"}, {"api_name": "getOneHot.getOneHot", "line_number": 204, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 218, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset", "line_number": 218, "usage_type": "name"}, {"api_name": "tensorflow.data.Dataset.from_tensor_slices", "line_number": 224, "usage_type": "call"}, {"api_name": "tensorflow.data.Dataset", "line_number": 224, "usage_type": "name"}, {"api_name": "model.fit", "line_number": 265, "usage_type": "call"}, {"api_name": "model.save", "line_number": 274, "usage_type": "call"}, {"api_name": "model.save", "line_number": 276, "usage_type": "call"}]} +{"seq_id": "500869069", "text": "import bpy\nimport re\nimport os\n#import yafrayinterface\n\ndef noise2string(ntype):\n\n\tif ntype == 'BLENDER_ORIGINAL' : return \"blender\"\n\telif ntype == 'ORIGINAL_PERLIN' : return \"stdperlin\"\n\telif ntype == 'IMPROVED_PERLIN' : return \"newperlin\"\n\telif ntype == 'VORONOI_F1' : return \"voronoi_f1\"\n\telif ntype == 'VORONOI_F2' : return \"voronoi_f2\"\n\telif ntype == 'VORONOI_F3' : return \"voronoi_f3\"\n\telif ntype == 'VORONOI_F4' : return \"voronoi_f4\"\n\telif ntype == 'VORONOI_F2_F1' : return \"voronoi_f2f1\"\n\telif ntype == 'VORONOI_CRACKLE' : return \"voronoi_crackle\"\n\telif ntype == 'CELL_NOISE' : return \"cellnoise\"\n\treturn \"newperlin\"\n\n#this function is tested under linux\ndef get_image_filename(filepath):\n\tpath = filepath.replace('//',os.path.expanduser('~')+'/',1)\n\treturn os.path.abspath(path)\n\nclass yafTexture:\n\tdef __init__(self, interface):\n\t\tself.yi = interface\n\t\n\tdef writeTexture(self,scene,tex):\n\t\t\n\t\tname = tex.name\n\t\tyi = self.yi\n\t\tyi.paramsClearAll()\n\t\t\n\t\tif tex.yaf_tex_type == 'BLEND' :\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type BLEND\")\n\t\t\tyi.paramsSetString(\"type\", \"blend\")\n\t\t\tstype = \"lin\"\n\t\t\tif tex.progression == 'LINEAR' : stype = \"lin\"\n\t\t\telif tex.progression == 'QUADRATIC' :\tstype = \"quad\"\n\t\t\telif tex.progression == 'EASING' :\tstype = \"ease\"\n\t\t\telif tex.progression == 'DIAGONAL' :\tstype = \"diag\"\n\t\t\telif tex.progression == 'SPHERICAL' :\tstype = \"sphere\"\n\t\t\telif tex.progression == 'QUADRATIC_SPHERE' :\tstype = \"halo\"\n\t\t\tyi.paramsSetString(\"stype\", stype)\n\t\t\n\t\t\n\t\telif tex.yaf_tex_type == 'CLOUDS':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type CLOUDS\")\n\t\t\tyi.paramsSetString(\"type\", \"clouds\")\n\t\t\t\n\t\t\tnoise_scale = tex.noise_scale\n\t\t\tif noise_scale > 0: noise_scale = 1.0/noise_scale\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"size\", noise_scale)\n\t\t\t\n\t\t\tif tex.noise_type == 'HARD_NOISE' :\n\t\t\t\thard = True\n\t\t\telse:\n\t\t\t\thard = False\n\t\t\n\t\t\tyi.paramsSetBool(\"hard\", hard)\n\t\t\tyi.paramsSetInt(\"depth\", tex.noise_depth)\n\t\t\n\t\t\n\t\telif tex.yaf_tex_type == 'WOOD':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type WOOD\")\n\t\t\tyi.paramsSetString(\"type\", \"wood\")\n\t\t\t\n\t\t\tyi.paramsSetInt(\"depth\", 0)\n\t\t\t\n\t\t\tturb = 0.0\n\t\t\tnoise_size = 0.25\n\t\t\thard = True\n\t\t\t\n\t\t\tif tex.stype == 'BANDNOISE' or tex.stype == 'RINGNOISE':\n\t\t\t\t\n\t\t\t\tturb = tex.turbulence\n\t\t\t\tnoise_size = tex.noise_size\n\t\t\t\t\n\t\t\t\tif noise_size > 0:\n\t\t\t\t\tnoise_size = 1.0/noise_size\n\t\t\t\tif tex.noise_type == 'SOFT_NOISE' :\n\t\t\t\t\thard = False\n\t\t\n\t\t\tyi.paramsSetFloat(\"turbulence\", turb)\n\t\t\tyi.paramsSetFloat(\"size\", noise_size)\n\t\t\tyi.paramsSetBool(\"hard\", hard )\n\t\t\t\n\t\t\tts = \"bands\"\n\t\t\t\n\t\t\tif tex.stype == 'RINGS' or tex.stype == 'RINGNOISE':\n\t\t\t\tts = \"rings\"\n\t\t\t\n\t\t\tyi.paramsSetString(\"wood_type\", ts )\n\t\t\tyi.paramsSetString(\"noise_type\", noise2string(tex.noise_basis) )\n\t\t\t\n\t\t\t# shape parameter\n\t\t\t\n\t\t\tif tex.noisebasis2 == 'SAW' :\n\t\t\t\tts=\"saw\"\n\t\t\telif tex.noisebasis2 == 'TRI':\n\t\t\t\tts=\"tri\"\n\t\t\telse:\n\t\t\t\tts = \"sin\"\n\n\t\t\tyi.paramsSetString(\"shape\", ts )\n\t\t\n\t\telif tex.yaf_tex_type == 'MARBLE':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type MARBLE\")\n\t\t\tyi.paramsSetString(\"type\", \"marble\")\n\t\t\t\n\t\t\tyi.paramsSetInt(\"depth\", tex.noise_depth)\n\t\t\tyi.paramsSetFloat(\"turbulence\", tex.turbulence)\n\t\t\t\n\t\t\tnoise_size = tex.noise_size\n\t\t\tif noise_size > 0:\n\t\t\t\tnoise_size = 1.0/noise_size\n\t\t\t\n\t\t\tif tex.noise_type == 'HARD_NOISE' :\n\t\t\t\thard = True\n\t\t\telse:\n\t\t\t\thard = False\n\t\t\t\t\n\t\t\tyi.paramsSetFloat(\"size\", noise_size)\n\t\t\tyi.paramsSetBool(\"hard\", hard )\n\t\t\t\n\t\t\tsharp = 4.0\n\t\t\tif tex.stype == 'SOFT':\n\t\t\t\tsharp = 2.0\n\t\t\telif tex.stype == 'SHARP':\n\t\t\t\tsharp = 4.0\n\t\t\telif tex.stype == 'SHARPER':\n\t\t\t\tsharp = 8.0\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"sharpness\", sharp)\n\t\t\tyi.paramsSetString(\"noise_type\", noise2string(tex.noise_basis) )\n\t\t\t\n\t\t\tif tex.noisebasis2 == 'SAW' :\n\t\t\t\tts=\"saw\"\n\t\t\telif tex.noisebasis2 == 'TRI':\n\t\t\t\tts=\"tri\"\n\t\t\telse:\n\t\t\t\tts = \"sin\"\n\t\t\t\n\t\t\tyi.paramsSetString(\"shape\", ts)\n\t\t\n\t\telif tex.yaf_tex_type == 'VORONOI':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type VORONOI\")\n\t\t\tyi.paramsSetString(\"type\", \"voronoi\")\n\t\t\t\n\t\t\tif tex.coloring == 'POSITION':\n\t\t\t\tts = \"col1\" \n\t\t\telif tex.coloring == 'POSITION_OUTLINE':\n\t\t\t\tts = \"col2\"\n\t\t\telif tex.coloring == 'POSITION_OUTLINE_INTENSITY':\n\t\t\t\tts = \"col3\"\n\t\t\telse:\n\t\t\t\tts = \"int\"\n\t\t\n\t\t\tyi.paramsSetString(\"color_type\", ts)\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"weight1\", tex.weight_1)\n\t\t\tyi.paramsSetFloat(\"weight2\", tex.weight_2)\n\t\t\tyi.paramsSetFloat(\"weight3\", tex.weight_3)\n\t\t\tyi.paramsSetFloat(\"weight4\", tex.weight_4)\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"mk_exponent\", tex.minkovsky_exponent)\n\t\t\tyi.paramsSetFloat(\"intensity\", tex.noise_intensity)\n\t\t\t\n\t\t\tnoise_size = tex.noise_size\n\t\t\tif noise_size > 0:\n\t\t\t\tnoise_size = 1.0/noise_size\n\t\t\tyi.paramsSetFloat(\"size\", noise_size)\n\t\t\t\n\t\t\tts = \"actual\"\n\t\t\tif tex.distance_metric == 'DISTANCE_SQUARED':\n\t\t\t\tts = \"squared\"\n\t\t\telif tex.distance_metric == 'MANHATTAN':\n\t\t\t\tts = \"manhattan\"\n\t\t\telif tex.distance_metric == 'CHEBYCHEV':\n\t\t\t\tts = \"chebychev\"\n\t\t\telif tex.distance_metric == 'MINKOVSKY_HALF':\n\t\t\t\tts = \"minkovsky_half\"\n\t\t\telif tex.distance_metric == 'MINKOVSKY_FOUR':\n\t\t\t\tts = \"minkovsky_four\"\n\t\t\telif tex.distance_metric == 'MINKOVSKY':\n\t\t\t\tts = \"minkovsky\"\n\t\t\t\n\t\t\tyi.paramsSetString(\"distance_metric\", ts)\n\t\t\n\t\telif tex.yaf_tex_type == 'MUSGRAVE':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type MUSGRAVE\")\n\t\t\tyi.paramsSetString(\"type\", \"musgrave\")\n\t\t\t\n\t\t\tts = \"fBm\"\n\t\t\tif tex.musgrave_type == 'MULTIFRACTAL' :\n\t\t\t\tts = \"multifractal\"\n\t\t\telif tex.musgrave_type == 'RIDGED_MULTIFRACTAL':\n\t\t\t\tts = \"ridgedmf\"\n\t\t\telif tex.musgrave_type == 'HYBRID_MULTIFRACTAL':\n\t\t\t\tts = \"hybridmf\"\n\t\t\telif tex.musgrave_type == 'HETERO_TERRAIN':\n\t\t\t\tts = \"heteroterrain\"\n\t\t\t\n\t\t\tyi.paramsSetString(\"musgrave_type\", ts)\n\t\t\tyi.paramsSetString(\"noise_type\", noise2string(tex.noise_basis))\n\t\t\tyi.paramsSetFloat(\"H\", tex.highest_dimension)\n\t\t\tyi.paramsSetFloat(\"lacunarity\", tex.lacunarity)\n\t\t\tyi.paramsSetFloat(\"octaves\", tex.octaves)\n\n\t\t\tnoise_size = tex.noise_size\n\t\t\tif noise_size > 0:\n\t\t\t\tnoise_size = 1.0/noise_size\n\t\t\tyi.paramsSetFloat(\"size\", noise_size)\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"intensity\", tex.offset)\n\t\t\n\t\telif tex.yaf_tex_type == 'DISTORTED_NOISE':\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type DISTORTED NOISE\")\n\t\t\tyi.paramsSetString(\"type\", \"distorted_noise\")\n\t\t\t\n\t\t\tyi.paramsSetFloat(\"distort\", tex.distortion)\n\t\t\t\n\t\t\tnoise_size = tex.noise_size\n\t\t\tif noise_size > 0:\n\t\t\t\tnoise_size = 1.0/noise_size\n\t\t\tyi.paramsSetFloat(\"size\", noise_size)\n\t\t\t\n\t\t\tyi.paramsSetString(\"noise_type1\", noise2string(tex.noise_basis))\n\t\t\tyi.paramsSetString(\"noise_type2\", noise2string(tex.noise_distortion))\n\t\t\n\t\telif tex.yaf_tex_type == 'IMAGE':\n\t\t\t\n\t\t\t#ima = tex.image\n\t\t\t##print(str(ima))\n\t\t\t#if ima is not None:\n\t\t\t\t# get image full path\n\t\t\t\t#imagefile = get_image_filename(ima.filepath)\n\t\t\timport os\n\t\t\tif tex.tex_file_name != \"\" and not os.path.exists(tex.tex_file_name):\n\t\t\t\tyi.printInfo(\"Exporter: No valid texture image supplied.\")\n\t\t\t\treturn False\n\t\t\t\n\t\t\t\n\t\t\tyi.printInfo(\"Exporter: Creating Texture: \\\"\" + name + \"\\\" type IMAGE: \" + tex.tex_file_name)\n\n\t\t\tyi.paramsSetString(\"type\", \"image\")\n\t\t\tyi.paramsSetString(\"filename\", tex.tex_file_name)\n\t\t\t\t#yi.paramsSetString(\"filename\", imagefile)\n\n\t\t\tyi.paramsSetBool(\"use_alpha\", tex.use_alpha)\n\t\t\tyi.paramsSetBool(\"calc_alpha\", tex.use_calculate_alpha)\n\t\t\tyi.paramsSetBool(\"normalmap\", tex.use_normal_map)\n\t\t\t\t\t\t\n\t\t\t# repeat\n\t\t\trepeat_x = 1\n\t\t\trepeat_y = 1\n\t\t\t\t\n\t\t\tif tex.extension == 'REPEAT':\n\t\t\t\trepeat_x = tex.repeat_x\n\t\t\t\trepeat_y = tex.repeat_y\n\t\t\t\t\n\t\t\tyi.paramsSetInt(\"xrepeat\", repeat_x)\n\t\t\tyi.paramsSetInt(\"yrepeat\", repeat_y)\n\t\t\t\t\t\t\n\t\t\t# clipping\n\t\t\text = tex.extension\n\t\t\t\t\n\t\t\t#print tex.getExtend()\n\t\t\tif ext == 'EXTEND':\n\t\t\t\tyi.paramsSetString(\"clipping\", \"extend\")\n\t\t\telif ext == 'CLIP':\n\t\t\t\tyi.paramsSetString(\"clipping\", \"clip\")\n\t\t\telif ext == 'CLIP_CUBE':\n\t\t\t\tyi.paramsSetString(\"clipping\", \"clipcube\")\n\t\t\telif ext == \"CHECKER\": #Blender.Texture.ExtendModes.CHECKER:\n\t\t\t\tyi.paramsSetString(\"clipping\", \"checker\")\n\t\t\t\tyi.paramsSetBool(\"even_tiles\", tex.checker_even)\n\t\t\t\tyi.paramsSetBool(\"odd_tiles\", tex.checker_odd)\n\t\t\telse:\n\t\t\t\tyi.paramsSetString(\"clipping\", \"repeat\")\n\t\t\t\t\n\t\t\t# crop min/max\n\t\t\tyi.paramsSetFloat(\"cropmin_x\", tex.crop_min_x)\n\t\t\tyi.paramsSetFloat(\"cropmin_y\", tex.crop_min_y) \n\t\t\tyi.paramsSetFloat(\"cropmax_x\", tex.crop_max_x)\n\t\t\tyi.paramsSetFloat(\"cropmax_y\", tex.crop_max_y)\n\t\t\t\t\n\t\t\t# rot90 flag\n\t\t\t#if tex.rot90 != 0:\n\t\t\tyi.paramsSetBool(\"rot90\", tex.use_flip_axis)\n\t\tyi.createTexture(name)\n\t\n\tdef createTextures(self,yi,scene):\n\t\t#alternative option: bpy.data.textures\n\t\t#objects = scene.objects\n\t\t#for item in objects:\n\t\t#\tfor index in range(16):\n\t\t#\t\tif not item.active_material.texture_slots[index]:\n\t\t#\t\t\tbreak\n\t\t#\t\telif item.active_material.use_textures[index]:\n\t\t#\t\t\ttex = item.active_material.texture_slots[index].texture\n\t\t#\t\t\tself.writeTexture(scene, tex)\n\t\ttextures = bpy.data.textures\n\t\tfor tex in textures:\n\t\t\tself.writeTexture(scene,tex)\n\t\t\t\n", "sub_path": "io/yaf_texture.py", "file_name": "yaf_texture.py", "file_ext": "py", "file_size_in_byte": 9121, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "os.path.expanduser", "line_number": 22, "usage_type": "call"}, {"api_name": "os.path", "line_number": 22, "usage_type": "attribute"}, {"api_name": "os.path.abspath", "line_number": 23, "usage_type": "call"}, {"api_name": "os.path", "line_number": 23, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 248, "usage_type": "call"}, {"api_name": "os.path", "line_number": 248, "usage_type": "attribute"}, {"api_name": "bpy.data", "line_number": 312, "usage_type": "attribute"}]} +{"seq_id": "395040210", "text": "# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /tmp/pip-build-ed191__6/Pygments/pygments/lexers/jvm.py\n# Compiled at: 2020-01-10 16:25:35\n# Size of source mod 2**32: 70347 bytes\n\"\"\"\n pygments.lexers.jvm\n ~~~~~~~~~~~~~~~~~~~\n\n Pygments lexers for JVM languages.\n\n :copyright: Copyright 2006-2019 by the Pygments team, see AUTHORS.\n :license: BSD, see LICENSE for details.\n\"\"\"\nimport re\nfrom pygments.lexer import Lexer, RegexLexer, include, bygroups, using, this, combined, default, words\nfrom pygments.token import Text, Comment, Operator, Keyword, Name, String, Number, Punctuation\nfrom pygments.util import shebang_matches\nfrom pygments import unistring as uni\n__all__ = [\n 'JavaLexer', 'ScalaLexer', 'GosuLexer', 'GosuTemplateLexer',\n 'GroovyLexer', 'IokeLexer', 'ClojureLexer', 'ClojureScriptLexer',\n 'KotlinLexer', 'XtendLexer', 'AspectJLexer', 'CeylonLexer',\n 'PigLexer', 'GoloLexer', 'JasminLexer', 'SarlLexer']\n\nclass JavaLexer(RegexLexer):\n __doc__ = '\\n For `Java `_ source code.\\n '\n name = 'Java'\n aliases = ['java']\n filenames = ['*.java']\n mimetypes = ['text/x-java']\n flags = re.MULTILINE | re.DOTALL | re.UNICODE\n tokens = {'root':[\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '(assert|break|case|catch|continue|default|do|else|finally|for|if|goto|instanceof|new|return|switch|this|throw|try|while)\\\\b',\n Keyword),\n (\n '((?:(?:[^\\\\W\\\\d]|\\\\$)[\\\\w.\\\\[\\\\]$<>]*\\\\s+)+?)((?:[^\\\\W\\\\d]|\\\\$)[\\\\w$]*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Punctuation)),\n (\n '@[^\\\\W\\\\d][\\\\w.]*', Name.Decorator),\n (\n '(abstract|const|enum|extends|final|implements|native|private|protected|public|static|strictfp|super|synchronized|throws|transient|volatile)\\\\b',\n Keyword.Declaration),\n (\n '(boolean|byte|char|double|float|int|long|short|void)\\\\b',\n Keyword.Type),\n (\n '(package)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(class|interface)(\\\\s+)', bygroups(Keyword.Declaration, Text),\n 'class'),\n (\n '(var)(\\\\s+)', bygroups(Keyword.Declaration, Text),\n 'var'),\n (\n '(import(?:\\\\s+static)?)(\\\\s+)', bygroups(Keyword.Namespace, Text),\n 'import'),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'\\\\\\\\.'|'[^\\\\\\\\]'|'\\\\\\\\u[0-9a-fA-F]{4}'\", String.Char),\n (\n '(\\\\.)((?:[^\\\\W\\\\d]|\\\\$)[\\\\w$]*)',\n bygroups(Punctuation, Name.Attribute)),\n (\n '^\\\\s*([^\\\\W\\\\d]|\\\\$)[\\\\w$]*:', Name.Label),\n (\n '([^\\\\W\\\\d]|\\\\$)[\\\\w$]*', Name),\n (\n '([0-9][0-9_]*\\\\.([0-9][0-9_]*)?|\\\\.[0-9][0-9_]*)([eE][+\\\\-]?[0-9][0-9_]*)?[fFdD]?|[0-9][eE][+\\\\-]?[0-9][0-9_]*[fFdD]?|[0-9]([eE][+\\\\-]?[0-9][0-9_]*)?[fFdD]|0[xX]([0-9a-fA-F][0-9a-fA-F_]*\\\\.?|([0-9a-fA-F][0-9a-fA-F_]*)?\\\\.[0-9a-fA-F][0-9a-fA-F_]*)[pP][+\\\\-]?[0-9][0-9_]*[fFdD]?',\n Number.Float),\n (\n '0[xX][0-9a-fA-F][0-9a-fA-F_]*[lL]?', Number.Hex),\n (\n '0[bB][01][01_]*[lL]?', Number.Bin),\n (\n '0[0-7_]+[lL]?', Number.Oct),\n (\n '0|[1-9][0-9_]*[lL]?', Number.Integer),\n (\n '[~^*!%&\\\\[\\\\]<>|+=/?-]', Operator),\n (\n '[{}();:.,]', Punctuation),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '([^\\\\W\\\\d]|\\\\$)[\\\\w$]*', Name.Class, '#pop')], \n 'var':[\n (\n '([^\\\\W\\\\d]|\\\\$)[\\\\w$]*', Name, '#pop')], \n 'import':[\n (\n '[\\\\w.]+\\\\*?', Name.Namespace, '#pop')]}\n\n\nclass AspectJLexer(JavaLexer):\n __doc__ = '\\n For `AspectJ `_ source code.\\n\\n .. versionadded:: 1.6\\n '\n name = 'AspectJ'\n aliases = ['aspectj']\n filenames = ['*.aj']\n mimetypes = ['text/x-aspectj']\n aj_keywords = {\n 'aspect', 'pointcut', 'privileged', 'call', 'execution',\n 'initialization', 'preinitialization', 'handler', 'get', 'set',\n 'staticinitialization', 'target', 'args', 'within', 'withincode',\n 'cflow', 'cflowbelow', 'annotation', 'before', 'after', 'around',\n 'proceed', 'throwing', 'returning', 'adviceexecution', 'declare',\n 'parents', 'warning', 'error', 'soft', 'precedence', 'thisJoinPoint',\n 'thisJoinPointStaticPart', 'thisEnclosingJoinPointStaticPart',\n 'issingleton', 'perthis', 'pertarget', 'percflow', 'percflowbelow',\n 'pertypewithin', 'lock', 'unlock', 'thisAspectInstance'}\n aj_inter_type = {\n 'parents:', 'warning:', 'error:', 'soft:', 'precedence:'}\n aj_inter_type_annotation = {'@type', '@method', '@constructor', '@field'}\n\n def get_tokens_unprocessed(self, text):\n for index, token, value in JavaLexer.get_tokens_unprocessed(self, text):\n if token is Name and value in self.aj_keywords:\n yield (\n index, Keyword, value)\n elif token is Name.Label and value in self.aj_inter_type:\n yield (\n index, Keyword, value[:-1])\n yield (index, Operator, value[(-1)])\n elif token is Name.Decorator and value in self.aj_inter_type_annotation:\n yield (\n index, Keyword, value)\n else:\n yield (\n index, token, value)\n\n\nclass ScalaLexer(RegexLexer):\n __doc__ = '\\n For `Scala `_ source code.\\n '\n name = 'Scala'\n aliases = ['scala']\n filenames = ['*.scala']\n mimetypes = ['text/x-scala']\n flags = re.MULTILINE | re.DOTALL\n op = '[-~\\\\^\\\\*!%&\\\\\\\\<>\\\\|+=:/?@¦-§©¬®°-±¶×÷϶҂؆-؈؎-؏۩۽-۾߶৺୰௳-௸௺౿ೱ-ೲ൹༁-༃༓-༗༚-༟༴༶༸྾-࿅࿇-࿏႞-႟፠᎐-᎙᥀᧠-᧿᭡-᭪᭴-᭼⁄⁒⁺-⁼₊-₌℀-℁℃-℆℈-℉℔№-℘℞-℣℥℧℩℮℺-℻⅀-⅄⅊-⅍⅏←-⌨⌫-⑊⒜-ⓩ─-❧➔-⟄⟇-⟥⟰-⦂⦙-⧗⧜-⧻⧾-⭔⳥-⳪⺀-⿻〄〒-〓〠〶-〷〾-〿㆐-㆑㆖-㆟㇀-㇣㈀-㈞㈪-㉐㉠-㉿㊊-㊰㋀-㏿䷀-䷿꒐-꓆꠨-꠫﬩﷽﹢﹤-﹦+<->|~¬¦│-○-�]+'\n letter = '[a-zA-Z\\\\$_ªµºÀ-ÖØ-öø-ʯͰ-ͳͶ-ͷͻ-ͽΆΈ-ϵϷ-ҁҊ-Ֆա-ևא-ײء-ؿف-يٮ-ٯٱ-ۓەۮ-ۯۺ-ۼۿܐܒ-ܯݍ-ޥޱߊ-ߪऄ-हऽॐक़-ॡॲ-ॿঅ-হঽৎড়-ৡৰ-ৱਅ-ਹਖ਼-ਫ਼ੲ-ੴઅ-હઽૐ-ૡଅ-ହଽଡ଼-ୡୱஃ-ஹௐఅ-ఽౘ-ౡಅ-ಹಽೞ-ೡഅ-ഽൠ-ൡൺ-ൿඅ-ෆก-ะา-ำเ-ๅກ-ະາ-ຳຽ-ໄໜ-ༀཀ-ཬྈ-ྋက-ဪဿၐ-ၕၚ-ၝၡၥ-ၦၮ-ၰၵ-ႁႎႠ-ჺᄀ-ፚᎀ-ᎏᎠ-ᙬᙯ-ᙶᚁ-ᚚᚠ-ᛪᛮ-ᜑᜠ-ᜱᝀ-ᝑᝠ-ᝰក-ឳៜᠠ-ᡂᡄ-ᢨᢪ-ᤜᥐ-ᦩᧁ-ᧇᨀ-ᨖᬅ-ᬳᭅ-ᭋᮃ-ᮠᮮ-ᮯᰀ-ᰣᱍ-ᱏᱚ-ᱷᴀ-ᴫᵢ-ᵷᵹ-ᶚḀ-ᾼιῂ-ῌῐ-Ίῠ-Ῥῲ-ῼⁱⁿℂℇℊ-ℓℕℙ-ℝℤΩℨK-ℭℯ-ℹℼ-ℿⅅ-ⅉⅎⅠ-ↈⰀ-ⱼⲀ-ⳤⴀ-ⵥⶀ-ⷞ〆-〇〡-〩〸-〺〼ぁ-ゖゟァ-ヺヿ-ㆎㆠ-ㆷㇰ-ㇿ㐀-䶵一-ꀔꀖ-ꒌꔀ-ꘋꘐ-ꘟꘪ-ꙮꚀ-ꚗꜢ-ꝯꝱ-ꞇꞋ-ꠁꠃ-ꠅꠇ-ꠊꠌ-ꠢꡀ-ꡳꢂ-ꢳꤊ-ꤥꤰ-ꥆꨀ-ꨨꩀ-ꩂꩄ-ꩋ가-힣豈-יִײַ-ﬨשׁ-ﴽﵐ-ﷻﹰ-ﻼA-Za-zヲ-ッア-ンᅠ-ᅵ]'\n upper = '[A-Z\\\\$_À-ÖØ-ÞĀĂĄĆĈĊČĎĐĒĔĖĘĚĜĞĠĢĤĦĨĪĬĮİIJĴĶĹĻĽĿŁŃŅŇŊŌŎŐŒŔŖŘŚŜŞŠŢŤŦŨŪŬŮŰŲŴŶŸ-ŹŻŽƁ-ƂƄƆ-ƇƉ-ƋƎ-ƑƓ-ƔƖ-ƘƜ-ƝƟ-ƠƢƤƦ-ƧƩƬƮ-ƯƱ-ƳƵƷ-ƸƼDŽLJNJǍǏǑǓǕǗǙǛǞǠǢǤǦǨǪǬǮDZǴǶ-ǸǺǼǾȀȂȄȆȈȊȌȎȐȒȔȖȘȚȜȞȠȢȤȦȨȪȬȮȰȲȺ-ȻȽ-ȾɁɃ-ɆɈɊɌɎͰͲͶΆΈ-ΏΑ-ΫϏϒ-ϔϘϚϜϞϠϢϤϦϨϪϬϮϴϷϹ-ϺϽ-ЯѠѢѤѦѨѪѬѮѰѲѴѶѸѺѼѾҀҊҌҎҐҒҔҖҘҚҜҞҠҢҤҦҨҪҬҮҰҲҴҶҸҺҼҾӀ-ӁӃӅӇӉӋӍӐӒӔӖӘӚӜӞӠӢӤӦӨӪӬӮӰӲӴӶӸӺӼӾԀԂԄԆԈԊԌԎԐԒԔԖԘԚԜԞԠԢԱ-ՖႠ-ჅḀḂḄḆḈḊḌḎḐḒḔḖḘḚḜḞḠḢḤḦḨḪḬḮḰḲḴḶḸḺḼḾṀṂṄṆṈṊṌṎṐṒṔṖṘṚṜṞṠṢṤṦṨṪṬṮṰṲṴṶṸṺṼṾẀẂẄẆẈẊẌẎẐẒẔẞẠẢẤẦẨẪẬẮẰẲẴẶẸẺẼẾỀỂỄỆỈỊỌỎỐỒỔỖỘỚỜỞỠỢỤỦỨỪỬỮỰỲỴỶỸỺỼỾἈ-ἏἘ-ἝἨ-ἯἸ-ἿὈ-ὍὙ-ὟὨ-ὯᾸ-ΆῈ-ΉῘ-ΊῨ-ῬῸ-Ώℂℇℋ-ℍℐ-ℒℕℙ-ℝℤΩℨK-ℭℰ-ℳℾ-ℿⅅↃⰀ-ⰮⱠⱢ-ⱤⱧⱩⱫⱭ-ⱯⱲⱵⲀⲂⲄⲆⲈⲊⲌⲎⲐⲒⲔⲖⲘⲚⲜⲞⲠⲢⲤⲦⲨⲪⲬⲮⲰⲲⲴⲶⲸⲺⲼⲾⳀⳂⳄⳆⳈⳊⳌⳎⳐⳒⳔⳖⳘⳚⳜⳞⳠⳢꙀꙂꙄꙆꙈꙊꙌꙎꙐꙒꙔꙖꙘꙚꙜꙞꙢꙤꙦꙨꙪꙬꚀꚂꚄꚆꚈꚊꚌꚎꚐꚒꚔꚖꜢꜤꜦꜨꜪꜬꜮꜲꜴꜶꜸꜺꜼꜾꝀꝂꝄꝆꝈꝊꝌꝎꝐꝒꝔꝖꝘꝚꝜꝞꝠꝢꝤꝦꝨꝪꝬꝮꝹꝻꝽ-ꝾꞀꞂꞄꞆꞋA-Z]'\n idrest = '%s(?:%s|[0-9])*(?:(?<=_)%s)?' % (letter, letter, op)\n letter_letter_digit = '%s(?:%s|\\\\d)*' % (letter, letter)\n tokens = {'root':[\n (\n '(class|trait|object)(\\\\s+)', bygroups(Keyword, Text), 'class'),\n (\n '[^\\\\S\\\\n]+', Text),\n include('comments'),\n (\n '@%s' % idrest, Name.Decorator),\n (\n '(abstract|ca(?:se|tch)|d(?:ef|o)|e(?:lse|xtends)|f(?:inal(?:ly)?|or(?:Some)?)|i(?:f|mplicit)|lazy|match|new|override|pr(?:ivate|otected)|re(?:quires|turn)|s(?:ealed|uper)|t(?:h(?:is|row)|ry)|va[lr]|w(?:hile|ith)|yield)\\\\b|(<[%:-]|=>|>:|[#=@_⇒←])(\\\\b|(?=\\\\s)|$)',\n Keyword),\n (\n ':(?!%s)' % op, Keyword, 'type'),\n (\n '%s%s\\\\b' % (upper, idrest), Name.Class),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(import|package)(\\\\s+)', bygroups(Keyword, Text), 'import'),\n (\n '(type)(\\\\s+)', bygroups(Keyword, Text), 'type'),\n (\n '\"\"\".*?\"\"\"(?!\")', String),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'\\\\\\\\.'|'[^\\\\\\\\]'|'\\\\\\\\u[0-9a-fA-F]{4}'\", String.Char),\n (\n \"'%s\" % idrest, Text.Symbol),\n (\n '[fs]\"\"\"', String, 'interptriplestring'),\n (\n '[fs]\"', String, 'interpstring'),\n (\n 'raw\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n idrest, Name),\n (\n '`[^`]+`', Name),\n (\n '\\\\[', Operator, 'typeparam'),\n (\n '[(){};,.#]', Operator),\n (\n op, Operator),\n (\n '([0-9][0-9]*\\\\.[0-9]*|\\\\.[0-9]+)([eE][+-]?[0-9]+)?[fFdD]?',\n Number.Float),\n (\n '0x[0-9a-fA-F]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '(%s|%s|`[^`]+`)(\\\\s*)(\\\\[)' % (idrest, op),\n bygroups(Name.Class, Text, Operator), ('#pop', 'typeparam')),\n (\n '\\\\s+', Text),\n include('comments'),\n (\n '\\\\{', Operator, '#pop'),\n (\n '\\\\(', Operator, '#pop'),\n (\n '%s|%s|`[^`]+`' % (idrest, op), Name.Class, '#pop')], \n 'type':[\n (\n '\\\\s+', Text),\n include('comments'),\n (\n '<[%:]|>:|[#_]|\\\\bforSome\\\\b|\\\\btype\\\\b', Keyword),\n (\n '([,);}]|=>|=|⇒)(\\\\s*)', bygroups(Operator, Text), '#pop'),\n (\n '[({]', Operator, '#push'),\n (\n '((?:%s|%s|`[^`]+`)(?:\\\\.(?:%s|%s|`[^`]+`))*)(\\\\s*)(\\\\[)' % (\n idrest, op, idrest, op),\n bygroups(Keyword.Type, Text, Operator), ('#pop', 'typeparam')),\n (\n '((?:%s|%s|`[^`]+`)(?:\\\\.(?:%s|%s|`[^`]+`))*)(\\\\s*)$' % (\n idrest, op, idrest, op),\n bygroups(Keyword.Type, Text), '#pop'),\n (\n '\\\\.|%s|%s|`[^`]+`' % (idrest, op), Keyword.Type)], \n 'typeparam':[\n (\n '\\\\s+', Text),\n include('comments'),\n (\n ',+', Punctuation),\n (\n '<[%:]|=>|>:|[#_⇒]|\\x08forSome\\x08|\\x08type\\x08', Keyword),\n (\n '([\\\\])}])', Operator, '#pop'),\n (\n '[(\\\\[{]', Operator, '#push'),\n (\n '\\\\.|%s|%s|`[^`]+`' % (idrest, op), Keyword.Type)], \n 'comments':[\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*', Comment.Multiline, 'comment')], \n 'comment':[\n (\n '[^/*]+', Comment.Multiline),\n (\n '/\\\\*', Comment.Multiline, '#push'),\n (\n '\\\\*/', Comment.Multiline, '#pop'),\n (\n '[*/]', Comment.Multiline)], \n 'import':[\n (\n '(%s|\\\\.)+' % idrest, Name.Namespace, '#pop')], \n 'interpstringcommon':[\n (\n '[^\"$\\\\\\\\]+', String),\n (\n '\\\\$\\\\$', String),\n (\n '\\\\$' + letter_letter_digit, String.Interpol),\n (\n '\\\\$\\\\{', String.Interpol, 'interpbrace'),\n (\n '\\\\\\\\.', String)], \n 'interptriplestring':[\n (\n '\"\"\"(?!\")', String, '#pop'),\n (\n '\"', String),\n include('interpstringcommon')], \n 'interpstring':[\n (\n '\"', String, '#pop'),\n include('interpstringcommon')], \n 'interpbrace':[\n (\n '\\\\}', String.Interpol, '#pop'),\n (\n '\\\\{', String.Interpol, '#push'),\n include('root')]}\n\n\nclass GosuLexer(RegexLexer):\n __doc__ = '\\n For Gosu source code.\\n\\n .. versionadded:: 1.5\\n '\n name = 'Gosu'\n aliases = ['gosu']\n filenames = ['*.gs', '*.gsx', '*.gsp', '*.vark']\n mimetypes = ['text/x-gosu']\n flags = re.MULTILINE | re.DOTALL\n tokens = {'root':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_]\\\\w*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '@[a-zA-Z_][\\\\w.]*', Name.Decorator),\n (\n '(in|as|typeof|statictypeof|typeis|typeas|if|else|foreach|for|index|while|do|continue|break|return|try|catch|finally|this|throw|new|switch|case|default|eval|super|outer|classpath|using)\\\\b',\n Keyword),\n (\n '(var|delegate|construct|function|private|internal|protected|public|abstract|override|final|static|extends|transient|implements|represents|readonly)\\\\b',\n Keyword.Declaration),\n (\n '(property\\\\s+)(get|set)?', Keyword.Declaration),\n (\n '(boolean|byte|char|double|float|int|long|short|void|block)\\\\b',\n Keyword.Type),\n (\n '(package)(\\\\s+)', bygroups(Keyword.Namespace, Text)),\n (\n '(true|false|null|NaN|Infinity)\\\\b', Keyword.Constant),\n (\n '(class|interface|enhancement|enum)(\\\\s+)([a-zA-Z_]\\\\w*)',\n bygroups(Keyword.Declaration, Text, Name.Class)),\n (\n '(uses)(\\\\s+)([\\\\w.]+\\\\*?)',\n bygroups(Keyword.Namespace, Text, Name.Namespace)),\n (\n '\"', String, 'string'),\n (\n '(\\\\??[.#])([a-zA-Z_]\\\\w*)',\n bygroups(Operator, Name.Attribute)),\n (\n '(:)([a-zA-Z_]\\\\w*)',\n bygroups(Operator, Name.Attribute)),\n (\n '[a-zA-Z_$]\\\\w*', Name),\n (\n 'and|or|not|[\\\\\\\\~^*!%&\\\\[\\\\](){}<>|+=:;,./?-]', Operator),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float),\n (\n '[0-9]+', Number.Integer),\n (\n '\\\\n', Text)], \n 'templateText':[\n (\n '(\\\\\\\\<)|(\\\\\\\\\\\\$)', String),\n (\n '(<%@\\\\s+)(extends|params)',\n bygroups(Operator, Name.Decorator), 'stringTemplate'),\n (\n '<%!--.*?--%>', Comment.Multiline),\n (\n '(<%)|(<%=)', Operator, 'stringTemplate'),\n (\n '\\\\$\\\\{', Operator, 'stringTemplateShorthand'),\n (\n '.', String)], \n 'string':[\n (\n '\"', String, '#pop'),\n include('templateText')], \n 'stringTemplate':[\n (\n '\"', String, 'string'),\n (\n '%>', Operator, '#pop'),\n include('root')], \n 'stringTemplateShorthand':[\n (\n '\"', String, 'string'),\n (\n '\\\\{', Operator, 'stringTemplateShorthand'),\n (\n '\\\\}', Operator, '#pop'),\n include('root')]}\n\n\nclass GosuTemplateLexer(Lexer):\n __doc__ = '\\n For Gosu templates.\\n\\n .. versionadded:: 1.5\\n '\n name = 'Gosu Template'\n aliases = ['gst']\n filenames = ['*.gst']\n mimetypes = ['text/x-gosu-template']\n\n def get_tokens_unprocessed(self, text):\n lexer = GosuLexer()\n stack = ['templateText']\n for item in lexer.get_tokens_unprocessed(text, stack):\n yield item\n\n\nclass GroovyLexer(RegexLexer):\n __doc__ = '\\n For `Groovy `_ source code.\\n\\n .. versionadded:: 1.5\\n '\n name = 'Groovy'\n aliases = ['groovy']\n filenames = ['*.groovy', '*.gradle']\n mimetypes = ['text/x-groovy']\n flags = re.MULTILINE | re.DOTALL\n tokens = {'root':[\n (\n '#!(.*?)$', Comment.Preproc, 'base'),\n default('base')], \n 'base':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_]\\\\w*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '@[a-zA-Z_][\\\\w.]*', Name.Decorator),\n (\n '(assert|break|case|catch|continue|default|do|else|finally|for|if|goto|instanceof|new|return|switch|this|throw|try|while|in|as)\\\\b',\n Keyword),\n (\n '(abstract|const|enum|extends|final|implements|native|private|protected|public|static|strictfp|super|synchronized|throws|transient|volatile)\\\\b',\n Keyword.Declaration),\n (\n '(def|boolean|byte|char|double|float|int|long|short|void)\\\\b',\n Keyword.Type),\n (\n '(package)(\\\\s+)', bygroups(Keyword.Namespace, Text)),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(class|interface)(\\\\s+)', bygroups(Keyword.Declaration, Text),\n 'class'),\n (\n '(import)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n '\"\"\".*?\"\"\"', String.Double),\n (\n \"'''.*?'''\", String.Single),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String.Double),\n (\n \"'(\\\\\\\\\\\\\\\\|\\\\\\\\'|[^'])*'\", String.Single),\n (\n '\\\\$/((?!/\\\\$).)*/\\\\$', String),\n (\n '/(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^/])*/', String),\n (\n \"'\\\\\\\\.'|'[^\\\\\\\\]'|'\\\\\\\\u[0-9a-fA-F]{4}'\", String.Char),\n (\n '(\\\\.)([a-zA-Z_]\\\\w*)', bygroups(Operator, Name.Attribute)),\n (\n '[a-zA-Z_]\\\\w*:', Name.Label),\n (\n '[a-zA-Z_$]\\\\w*', Name),\n (\n '[~^*!%&\\\\[\\\\](){}<>|+=:;,./?-]', Operator),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float),\n (\n '0x[0-9a-fA-F]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '[a-zA-Z_]\\\\w*', Name.Class, '#pop')], \n 'import':[\n (\n '[\\\\w.]+\\\\*?', Name.Namespace, '#pop')]}\n\n def analyse_text(text):\n return shebang_matches(text, 'groovy')\n\n\nclass IokeLexer(RegexLexer):\n __doc__ = '\\n For `Ioke `_ (a strongly typed, dynamic,\\n prototype based programming language) source.\\n\\n .. versionadded:: 1.4\\n '\n name = 'Ioke'\n filenames = ['*.ik']\n aliases = ['ioke', 'ik']\n mimetypes = ['text/x-iokesrc']\n tokens = {'interpolatableText':[\n (\n '(\\\\\\\\b|\\\\\\\\e|\\\\\\\\t|\\\\\\\\n|\\\\\\\\f|\\\\\\\\r|\\\\\\\\\"|\\\\\\\\\\\\\\\\|\\\\\\\\#|\\\\\\\\\\\\Z|\\\\\\\\u[0-9a-fA-F]{1,4}|\\\\\\\\[0-3]?[0-7]?[0-7])',\n String.Escape),\n (\n '#\\\\{', Punctuation, 'textInterpolationRoot')], \n 'text':[\n (\n '(?>|\\\\|\\\\|>>|\\\\*\\\\*>>|:::|::|\\\\.\\\\.\\\\.|===|\\\\*\\\\*>|\\\\*\\\\*=|&&>|&&=|\\\\|\\\\|>|\\\\|\\\\|=|\\\\->>|\\\\+>>|!>>|<>>>|<>>|&>>|%>>|#>>|@>>|/>>|\\\\*>>|\\\\?>>|\\\\|>>|\\\\^>>|~>>|\\\\$>>|=>>|<<=|>>=|<=>|<\\\\->|=~|!~|=>|\\\\+\\\\+|\\\\-\\\\-|<=|>=|==|!=|&&|\\\\.\\\\.|\\\\+=|\\\\-=|\\\\*=|\\\\/=|%=|&=|\\\\^=|\\\\|=|<\\\\-|\\\\+>|!>|<>|&>|%>|#>|\\\\@>|\\\\/>|\\\\*>|\\\\?>|\\\\|>|\\\\^>|~>|\\\\$>|<\\\\->|\\\\->|<<|>>|\\\\*\\\\*|\\\\?\\\\||\\\\?&|\\\\|\\\\||>|<|\\\\*|\\\\/|%|\\\\+|\\\\-|&|\\\\^|\\\\||=|\\\\$|!|~|\\\\?|#|≠|∘|∈|∉)',\n Operator),\n (\n '(and|nand|or|xor|nor|return|import)(?![\\\\w!?])',\n Operator),\n (\n \"(\\\\`\\\\`|\\\\`|\\\\'\\\\'|\\\\'|\\\\.|\\\\,|@@|@|\\\\[|\\\\]|\\\\(|\\\\)|\\\\{|\\\\})\", Punctuation),\n (\n '[A-Z][\\\\w!:?]*', Name.Class),\n (\n '[a-z_][\\\\w!:?]*', Name)]}\n\n\nclass ClojureLexer(RegexLexer):\n __doc__ = '\\n Lexer for `Clojure `_ source code.\\n\\n .. versionadded:: 0.11\\n '\n name = 'Clojure'\n aliases = ['clojure', 'clj']\n filenames = ['*.clj']\n mimetypes = ['text/x-clojure', 'application/x-clojure']\n special_forms = ('.', 'def', 'do', 'fn', 'if', 'let', 'new', 'quote', 'var', 'loop')\n declarations = ('def-', 'defn', 'defn-', 'defmacro', 'defmulti', 'defmethod', 'defstruct',\n 'defonce', 'declare', 'definline', 'definterface', 'defprotocol',\n 'defrecord', 'deftype', 'defproject', 'ns')\n builtins = ('*', '+', '-', '->', '/', '<', '<=', '=', '==', '>', '>=', '..', 'accessor',\n 'agent', 'agent-errors', 'aget', 'alength', 'all-ns', 'alter', 'and',\n 'append-child', 'apply', 'array-map', 'aset', 'aset-boolean', 'aset-byte',\n 'aset-char', 'aset-double', 'aset-float', 'aset-int', 'aset-long',\n 'aset-short', 'assert', 'assoc', 'await', 'await-for', 'bean', 'binding',\n 'bit-and', 'bit-not', 'bit-or', 'bit-shift-left', 'bit-shift-right',\n 'bit-xor', 'boolean', 'branch?', 'butlast', 'byte', 'cast', 'char',\n 'children', 'class', 'clear-agent-errors', 'comment', 'commute',\n 'comp', 'comparator', 'complement', 'concat', 'conj', 'cons', 'constantly',\n 'cond', 'if-not', 'construct-proxy', 'contains?', 'count', 'create-ns',\n 'create-struct', 'cycle', 'dec', 'deref', 'difference', 'disj', 'dissoc',\n 'distinct', 'doall', 'doc', 'dorun', 'doseq', 'dosync', 'dotimes',\n 'doto', 'double', 'down', 'drop', 'drop-while', 'edit', 'end?', 'ensure',\n 'eval', 'every?', 'false?', 'ffirst', 'file-seq', 'filter', 'find',\n 'find-doc', 'find-ns', 'find-var', 'first', 'float', 'flush', 'for',\n 'fnseq', 'frest', 'gensym', 'get-proxy-class', 'get', 'hash-map',\n 'hash-set', 'identical?', 'identity', 'if-let', 'import', 'in-ns',\n 'inc', 'index', 'insert-child', 'insert-left', 'insert-right', 'inspect-table',\n 'inspect-tree', 'instance?', 'int', 'interleave', 'intersection',\n 'into', 'into-array', 'iterate', 'join', 'key', 'keys', 'keyword',\n 'keyword?', 'last', 'lazy-cat', 'lazy-cons', 'left', 'lefts', 'line-seq',\n 'list*', 'list', 'load', 'load-file', 'locking', 'long', 'loop',\n 'macroexpand', 'macroexpand-1', 'make-array', 'make-node', 'map',\n 'map-invert', 'map?', 'mapcat', 'max', 'max-key', 'memfn', 'merge',\n 'merge-with', 'meta', 'min', 'min-key', 'name', 'namespace', 'neg?',\n 'new', 'newline', 'next', 'nil?', 'node', 'not', 'not-any?', 'not-every?',\n 'not=', 'ns-imports', 'ns-interns', 'ns-map', 'ns-name', 'ns-publics',\n 'ns-refers', 'ns-resolve', 'ns-unmap', 'nth', 'nthrest', 'or', 'parse',\n 'partial', 'path', 'peek', 'pop', 'pos?', 'pr', 'pr-str', 'print',\n 'print-str', 'println', 'println-str', 'prn', 'prn-str', 'project',\n 'proxy', 'proxy-mappings', 'quot', 'rand', 'rand-int', 'range', 're-find',\n 're-groups', 're-matcher', 're-matches', 're-pattern', 're-seq',\n 'read', 'read-line', 'reduce', 'ref', 'ref-set', 'refer', 'rem',\n 'remove', 'remove-method', 'remove-ns', 'rename', 'rename-keys',\n 'repeat', 'replace', 'replicate', 'resolve', 'rest', 'resultset-seq',\n 'reverse', 'rfirst', 'right', 'rights', 'root', 'rrest', 'rseq',\n 'second', 'select', 'select-keys', 'send', 'send-off', 'seq', 'seq-zip',\n 'seq?', 'set', 'short', 'slurp', 'some', 'sort', 'sort-by', 'sorted-map',\n 'sorted-map-by', 'sorted-set', 'special-symbol?', 'split-at', 'split-with',\n 'str', 'string?', 'struct', 'struct-map', 'subs', 'subvec', 'symbol',\n 'symbol?', 'sync', 'take', 'take-nth', 'take-while', 'test', 'time',\n 'to-array', 'to-array-2d', 'tree-seq', 'true?', 'union', 'up', 'update-proxy',\n 'val', 'vals', 'var-get', 'var-set', 'var?', 'vector', 'vector-zip',\n 'vector?', 'when', 'when-first', 'when-let', 'when-not', 'with-local-vars',\n 'with-meta', 'with-open', 'with-out-str', 'xml-seq', 'xml-zip', 'zero?',\n 'zipmap', 'zipper')\n valid_name = '(?!#)[\\\\w!$%*+<=>?/.#|-]+'\n tokens = {'root': [\n (\n ';.*$', Comment.Single),\n (\n '[,\\\\s]+', Text),\n (\n '-?\\\\d+\\\\.\\\\d+', Number.Float),\n (\n '-?\\\\d+', Number.Integer),\n (\n '0x-?[abcdef\\\\d]+', Number.Hex),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'\" + valid_name, String.Symbol),\n (\n '\\\\\\\\(.|[a-z]+)', String.Char),\n (\n '::?#?' + valid_name, String.Symbol),\n (\n \"~@|[`\\\\'#^~&@]\", Operator),\n (\n words(special_forms, suffix=' '), Keyword),\n (\n words(declarations, suffix=' '), Keyword.Declaration),\n (\n words(builtins, suffix=' '), Name.Builtin),\n (\n '(?<=\\\\()' + valid_name, Name.Function),\n (\n valid_name, Name.Variable),\n (\n '(\\\\[|\\\\])', Punctuation),\n (\n '(\\\\{|\\\\})', Punctuation),\n (\n '(\\\\(|\\\\))', Punctuation)]}\n\n\nclass ClojureScriptLexer(ClojureLexer):\n __doc__ = '\\n Lexer for `ClojureScript `_\\n source code.\\n\\n .. versionadded:: 2.0\\n '\n name = 'ClojureScript'\n aliases = ['clojurescript', 'cljs']\n filenames = ['*.cljs']\n mimetypes = ['text/x-clojurescript', 'application/x-clojurescript']\n\n\nclass TeaLangLexer(RegexLexer):\n __doc__ = '\\n For `Tea `_ source code. Only used within a\\n TeaTemplateLexer.\\n\\n .. versionadded:: 1.5\\n '\n flags = re.MULTILINE | re.DOTALL\n tokens = {'root':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w\\\\.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_]\\\\w*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '@[a-zA-Z_][\\\\w\\\\.]*', Name.Decorator),\n (\n '(and|break|else|foreach|if|in|not|or|reverse)\\\\b',\n Keyword),\n (\n '(as|call|define)\\\\b', Keyword.Declaration),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(template)(\\\\s+)', bygroups(Keyword.Declaration, Text), 'template'),\n (\n '(import)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"\\\\'(\\\\\\\\\\\\\\\\|\\\\\\\\\\\\'|[^\\\\'])*\\\\'\", String),\n (\n '(\\\\.)([a-zA-Z_]\\\\w*)', bygroups(Operator, Name.Attribute)),\n (\n '[a-zA-Z_]\\\\w*:', Name.Label),\n (\n '[a-zA-Z_\\\\$]\\\\w*', Name),\n (\n '(isa|[.]{3}|[.]{2}|[=#!<>+-/%&;,.\\\\*\\\\\\\\\\\\(\\\\)\\\\[\\\\]\\\\{\\\\}])', Operator),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float),\n (\n '0x[0-9a-fA-F]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text)], \n 'template':[\n (\n '[a-zA-Z_]\\\\w*', Name.Class, '#pop')], \n 'import':[\n (\n '[\\\\w.]+\\\\*?', Name.Namespace, '#pop')]}\n\n\nclass CeylonLexer(RegexLexer):\n __doc__ = '\\n For `Ceylon `_ source code.\\n\\n .. versionadded:: 1.6\\n '\n name = 'Ceylon'\n aliases = ['ceylon']\n filenames = ['*.ceylon']\n mimetypes = ['text/x-ceylon']\n flags = re.MULTILINE | re.DOTALL\n _ws = '(?:\\\\s|//.*?\\\\n|/[*].*?[*]/)+'\n tokens = {'root':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_]\\\\w*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*', Comment.Multiline, 'comment'),\n (\n '(shared|abstract|formal|default|actual|variable|deprecated|small|late|literal|doc|by|see|throws|optional|license|tagged|final|native|annotation|sealed)\\\\b',\n Name.Decorator),\n (\n '(break|case|catch|continue|else|finally|for|in|if|return|switch|this|throw|try|while|is|exists|dynamic|nonempty|then|outer|assert|let)\\\\b',\n Keyword),\n (\n '(abstracts|extends|satisfies|super|given|of|out|assign)\\\\b',\n Keyword.Declaration),\n (\n '(function|value|void|new)\\\\b',\n Keyword.Type),\n (\n '(assembly|module|package)(\\\\s+)', bygroups(Keyword.Namespace, Text)),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(class|interface|object|alias)(\\\\s+)',\n bygroups(Keyword.Declaration, Text), 'class'),\n (\n '(import)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'\\\\\\\\.'|'[^\\\\\\\\]'|'\\\\\\\\\\\\{#[0-9a-fA-F]{4}\\\\}'\", String.Char),\n (\n '\".*``.*``.*\"', String.Interpol),\n (\n '(\\\\.)([a-z_]\\\\w*)',\n bygroups(Operator, Name.Attribute)),\n (\n '[a-zA-Z_]\\\\w*:', Name.Label),\n (\n '[a-zA-Z_]\\\\w*', Name),\n (\n '[~^*!%&\\\\[\\\\](){}<>|+=:;,./?-]', Operator),\n (\n '\\\\d{1,3}(_\\\\d{3})+\\\\.\\\\d{1,3}(_\\\\d{3})+[kMGTPmunpf]?', Number.Float),\n (\n '\\\\d{1,3}(_\\\\d{3})+\\\\.[0-9]+([eE][+-]?[0-9]+)?[kMGTPmunpf]?',\n Number.Float),\n (\n '[0-9][0-9]*\\\\.\\\\d{1,3}(_\\\\d{3})+[kMGTPmunpf]?', Number.Float),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][+-]?[0-9]+)?[kMGTPmunpf]?',\n Number.Float),\n (\n '#([0-9a-fA-F]{4})(_[0-9a-fA-F]{4})+', Number.Hex),\n (\n '#[0-9a-fA-F]+', Number.Hex),\n (\n '\\\\$([01]{4})(_[01]{4})+', Number.Bin),\n (\n '\\\\$[01]+', Number.Bin),\n (\n '\\\\d{1,3}(_\\\\d{3})+[kMGTP]?', Number.Integer),\n (\n '[0-9]+[kMGTP]?', Number.Integer),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '[A-Za-z_]\\\\w*', Name.Class, '#pop')], \n 'import':[\n (\n '[a-z][\\\\w.]*',\n Name.Namespace, '#pop')], \n 'comment':[\n (\n '[^*/]', Comment.Multiline),\n (\n '/\\\\*', Comment.Multiline, '#push'),\n (\n '\\\\*/', Comment.Multiline, '#pop'),\n (\n '[*/]', Comment.Multiline)]}\n\n\nclass KotlinLexer(RegexLexer):\n __doc__ = '\\n For `Kotlin `_\\n source code.\\n\\n .. versionadded:: 1.5\\n '\n name = 'Kotlin'\n aliases = ['kotlin']\n filenames = ['*.kt']\n mimetypes = ['text/x-kotlin']\n flags = re.MULTILINE | re.DOTALL | re.UNICODE\n kt_name = '@?[_' + uni.combine('Lu', 'Ll', 'Lt', 'Lm', 'Nl') + ']' + '[' + uni.combine('Lu', 'Ll', 'Lt', 'Lm', 'Nl', 'Nd', 'Pc', 'Cf', 'Mn', 'Mc') + ']*'\n kt_space_name = '@?[_' + uni.combine('Lu', 'Ll', 'Lt', 'Lm', 'Nl') + ']' + '[' + uni.combine('Lu', 'Ll', 'Lt', 'Lm', 'Nl', 'Nd', 'Pc', 'Cf', 'Mn', 'Mc', 'Zs') + ',-]*'\n kt_id = '(' + kt_name + '|`' + kt_space_name + '`)'\n tokens = {'root':[\n (\n '^\\\\s*\\\\[.*?\\\\]', Name.Attribute),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '\\\\s+', Text),\n (\n '\\\\\\\\\\\\n', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/[*].*?[*]/', Comment.Multiline),\n (\n '\"\"\".*?\"\"\"', String),\n (\n '\\\\n', Text),\n (\n '::|!!|\\\\?[:.]', Operator),\n (\n '[~!%^&*()+=|\\\\[\\\\]:;,.<>/?-]', Punctuation),\n (\n '[{}]', Punctuation),\n (\n '@\"(\"\"|[^\"])*\"', String),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"\\\\n])*[\"\\\\n]', String),\n (\n \"'\\\\\\\\.'|'[^\\\\\\\\]'\", String.Char),\n (\n '[0-9](\\\\.[0-9]*)?([eE][+-][0-9]+)?[flFL]?|0[xX][0-9a-fA-F]+[Ll]?',\n Number),\n (\n '(object)(\\\\s+)(:)(\\\\s+)', bygroups(Keyword, Text, Punctuation, Text), 'class'),\n (\n '(companion)(\\\\s+)(object)', bygroups(Keyword, Text, Keyword)),\n (\n '(class|interface|object)(\\\\s+)', bygroups(Keyword, Text), 'class'),\n (\n '(package|import)(\\\\s+)', bygroups(Keyword, Text), 'package'),\n (\n '(val|var)(\\\\s+)([(])', bygroups(Keyword, Text, Punctuation), 'property_dec'),\n (\n '(val|var)(\\\\s+)', bygroups(Keyword, Text), 'property'),\n (\n '(fun)(\\\\s+)', bygroups(Keyword, Text), 'function'),\n (\n '(inline fun)(\\\\s+)', bygroups(Keyword, Text), 'function'),\n (\n '(abstract|annotation|as|break|by|catch|class|companion|const|constructor|continue|crossinline|data|do|dynamic|else|enum|external|false|final|finally|for|fun|get|if|import|in|infix|inline|inner|interface|internal|is|lateinit|noinline|null|object|open|operator|out|override|package|private|protected|public|reified|return|sealed|set|super|tailrec|this|throw|true|try|val|var|vararg|when|where|while)\\\\b',\n Keyword),\n (\n kt_id, Name)], \n 'package':[\n (\n '\\\\S+', Name.Namespace, '#pop')], \n 'class':[\n (\n kt_id, Name.Class, '#pop')], \n 'property':[\n (\n kt_id, Name.Property, '#pop')], \n 'property_dec':[\n (\n '(,)(\\\\s*)', bygroups(Punctuation, Text)),\n (\n '(:)(\\\\s*)', bygroups(Punctuation, Text)),\n (\n '<', Punctuation, 'generic'),\n (\n '([)])', Punctuation, '#pop'),\n (\n kt_id, Name.Property)], \n 'function':[\n (\n '<', Punctuation, 'generic'),\n (\n '' + kt_id + '([.])' + kt_id, bygroups(Name.Class, Punctuation, Name.Function), '#pop'),\n (\n kt_id, Name.Function, '#pop')], \n 'generic':[\n (\n '(>)(\\\\s*)', bygroups(Punctuation, Text), '#pop'),\n (\n ':', Punctuation),\n (\n '(reified|out|in)\\\\b', Keyword),\n (\n ',', Text),\n (\n '\\\\s+', Text),\n (\n kt_id, Name)]}\n\n\nclass XtendLexer(RegexLexer):\n __doc__ = '\\n For `Xtend `_ source code.\\n\\n .. versionadded:: 1.6\\n '\n name = 'Xtend'\n aliases = ['xtend']\n filenames = ['*.xtend']\n mimetypes = ['text/x-xtend']\n flags = re.MULTILINE | re.DOTALL\n tokens = {'root':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_$][\\\\w$]*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '@[a-zA-Z_][\\\\w.]*', Name.Decorator),\n (\n '(assert|break|case|catch|continue|default|do|else|finally|for|if|goto|instanceof|new|return|switch|this|throw|try|while|IF|ELSE|ELSEIF|ENDIF|FOR|ENDFOR|SEPARATOR|BEFORE|AFTER)\\\\b',\n Keyword),\n (\n '(def|abstract|const|enum|extends|final|implements|native|private|protected|public|static|strictfp|super|synchronized|throws|transient|volatile)\\\\b',\n Keyword.Declaration),\n (\n '(boolean|byte|char|double|float|int|long|short|void)\\\\b',\n Keyword.Type),\n (\n '(package)(\\\\s+)', bygroups(Keyword.Namespace, Text)),\n (\n '(true|false|null)\\\\b', Keyword.Constant),\n (\n '(class|interface)(\\\\s+)', bygroups(Keyword.Declaration, Text),\n 'class'),\n (\n '(import)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n \"(''')\", String, 'template'),\n (\n '(»)', String, 'template'),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'(\\\\\\\\\\\\\\\\|\\\\\\\\'|[^'])*'\", String),\n (\n '[a-zA-Z_]\\\\w*:', Name.Label),\n (\n '[a-zA-Z_$]\\\\w*', Name),\n (\n '[~^*!%&\\\\[\\\\](){}<>\\\\|+=:;,./?-]', Operator),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float),\n (\n '0x[0-9a-fA-F]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '[a-zA-Z_]\\\\w*', Name.Class, '#pop')], \n 'import':[\n (\n '[\\\\w.]+\\\\*?', Name.Namespace, '#pop')], \n 'template':[\n (\n \"'''\", String, '#pop'),\n (\n '«', String, '#pop'),\n (\n '.', String)]}\n\n\nclass PigLexer(RegexLexer):\n __doc__ = '\\n For `Pig Latin `_ source code.\\n\\n .. versionadded:: 2.0\\n '\n name = 'Pig'\n aliases = ['pig']\n filenames = ['*.pig']\n mimetypes = ['text/x-pig']\n flags = re.MULTILINE | re.IGNORECASE\n tokens = {'root':[\n (\n '\\\\s+', Text),\n (\n '--.*', Comment),\n (\n '/\\\\*[\\\\w\\\\W]*?\\\\*/', Comment.Multiline),\n (\n '\\\\\\\\\\\\n', Text),\n (\n '\\\\\\\\', Text),\n (\n \"\\\\'(?:\\\\\\\\[ntbrf\\\\\\\\\\\\']|\\\\\\\\u[0-9a-f]{4}|[^\\\\'\\\\\\\\\\\\n\\\\r])*\\\\'\", String),\n include('keywords'),\n include('types'),\n include('builtins'),\n include('punct'),\n include('operators'),\n (\n '[0-9]*\\\\.[0-9]+(e[0-9]+)?[fd]?', Number.Float),\n (\n '0x[0-9a-f]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text),\n (\n '([a-z_]\\\\w*)(\\\\s*)(\\\\()',\n bygroups(Name.Function, Text, Punctuation)),\n (\n '[()#:]', Text),\n (\n '[^(:#\\\\\\'\")\\\\s]+', Text),\n (\n '\\\\S+\\\\s+', Text)], \n 'keywords':[\n (\n '(assert|and|any|all|arrange|as|asc|bag|by|cache|CASE|cat|cd|cp|%declare|%default|define|dense|desc|describe|distinct|du|dump|eval|exex|explain|filter|flatten|foreach|full|generate|group|help|if|illustrate|import|inner|input|into|is|join|kill|left|limit|load|ls|map|matches|mkdir|mv|not|null|onschema|or|order|outer|output|parallel|pig|pwd|quit|register|returns|right|rm|rmf|rollup|run|sample|set|ship|split|stderr|stdin|stdout|store|stream|through|union|using|void)\\\\b',\n Keyword)], \n 'builtins':[\n (\n '(AVG|BinStorage|cogroup|CONCAT|copyFromLocal|copyToLocal|COUNT|cross|DIFF|MAX|MIN|PigDump|PigStorage|SIZE|SUM|TextLoader|TOKENIZE)\\\\b',\n Name.Builtin)], \n 'types':[\n (\n '(bytearray|BIGINTEGER|BIGDECIMAL|chararray|datetime|double|float|int|long|tuple)\\\\b',\n Keyword.Type)], \n 'punct':[\n (\n '[;(){}\\\\[\\\\]]', Punctuation)], \n 'operators':[\n (\n '[#=,./%+\\\\-?]', Operator),\n (\n '(eq|gt|lt|gte|lte|neq|matches)\\\\b', Operator),\n (\n '(==|<=|<|>=|>|!=)', Operator)]}\n\n\nclass GoloLexer(RegexLexer):\n __doc__ = '\\n For `Golo `_ source code.\\n\\n .. versionadded:: 2.0\\n '\n name = 'Golo'\n filenames = ['*.golo']\n aliases = ['golo']\n tokens = {'root':[\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '#.*$', Comment),\n (\n '(\\\\^|\\\\.\\\\.\\\\.|:|\\\\?:|->|==|!=|=|\\\\+|\\\\*|%|/|<=|<|>=|>|=|\\\\.)',\n Operator),\n (\n '(?<=[^-])(-)(?=[^-])', Operator),\n (\n '(?<=[^`])(is|isnt|and|or|not|oftype|in|orIfNull)\\\\b', Operator.Word),\n (\n '[]{}|(),[]', Punctuation),\n (\n '(module|import)(\\\\s+)',\n bygroups(Keyword.Namespace, Text),\n 'modname'),\n (\n '\\\\b([a-zA-Z_][\\\\w$.]*)(::)', bygroups(Name.Namespace, Punctuation)),\n (\n '\\\\b([a-zA-Z_][\\\\w$]*(?:\\\\.[a-zA-Z_][\\\\w$]*)+)\\\\b', Name.Namespace),\n (\n '(let|var)(\\\\s+)',\n bygroups(Keyword.Declaration, Text),\n 'varname'),\n (\n '(struct)(\\\\s+)',\n bygroups(Keyword.Declaration, Text),\n 'structname'),\n (\n '(function)(\\\\s+)',\n bygroups(Keyword.Declaration, Text),\n 'funcname'),\n (\n '(null|true|false)\\\\b', Keyword.Constant),\n (\n '(augment|pimp|if|else|case|match|return|case|when|then|otherwise|while|for|foreach|try|catch|finally|throw|local|continue|break)\\\\b',\n Keyword),\n (\n '(map|array|list|set|vector|tuple)(\\\\[)',\n bygroups(Name.Builtin, Punctuation)),\n (\n '(print|println|readln|raise|fun|asInterfaceInstance)\\\\b',\n Name.Builtin),\n (\n '(`?[a-zA-Z_][\\\\w$]*)(\\\\()',\n bygroups(Name.Function, Punctuation)),\n (\n '-?[\\\\d_]*\\\\.[\\\\d_]*([eE][+-]?\\\\d[\\\\d_]*)?F?', Number.Float),\n (\n '0[0-7]+j?', Number.Oct),\n (\n '0[xX][a-fA-F0-9]+', Number.Hex),\n (\n '-?\\\\d[\\\\d_]*L', Number.Integer.Long),\n (\n '-?\\\\d[\\\\d_]*', Number.Integer),\n (\n '`?[a-zA-Z_][\\\\w$]*', Name),\n (\n '@[a-zA-Z_][\\\\w$.]*', Name.Decorator),\n (\n '\"\"\"', String, combined('stringescape', 'triplestring')),\n (\n '\"', String, combined('stringescape', 'doublestring')),\n (\n \"'\", String, combined('stringescape', 'singlestring')),\n (\n '----((.|\\\\n)*?)----', String.Doc)], \n 'funcname':[\n (\n '`?[a-zA-Z_][\\\\w$]*', Name.Function, '#pop')], \n 'modname':[\n (\n '[a-zA-Z_][\\\\w$.]*\\\\*?', Name.Namespace, '#pop')], \n 'structname':[\n (\n '`?[\\\\w.]+\\\\*?', Name.Class, '#pop')], \n 'varname':[\n (\n '`?[a-zA-Z_][\\\\w$]*', Name.Variable, '#pop')], \n 'string':[\n (\n '[^\\\\\\\\\\\\\\'\"\\\\n]+', String),\n (\n '[\\\\\\'\"\\\\\\\\]', String)], \n 'stringescape':[\n (\n '\\\\\\\\([\\\\\\\\abfnrtv\"\\\\\\']|\\\\n|N\\\\{.*?\\\\}|u[a-fA-F0-9]{4}|U[a-fA-F0-9]{8}|x[a-fA-F0-9]{2}|[0-7]{1,3})',\n String.Escape)], \n 'triplestring':[\n (\n '\"\"\"', String, '#pop'),\n include('string'),\n (\n '\\\\n', String)], \n 'doublestring':[\n (\n '\"', String.Double, '#pop'),\n include('string')], \n 'singlestring':[\n (\n \"'\", String, '#pop'),\n include('string')], \n 'operators':[\n (\n '[#=,./%+\\\\-?]', Operator),\n (\n '(eq|gt|lt|gte|lte|neq|matches)\\\\b', Operator),\n (\n '(==|<=|<|>=|>|!=)', Operator)]}\n\n\nclass JasminLexer(RegexLexer):\n __doc__ = '\\n For `Jasmin `_ assembly code.\\n\\n .. versionadded:: 2.0\\n '\n name = 'Jasmin'\n aliases = ['jasmin', 'jasminxt']\n filenames = ['*.j']\n _whitespace = ' \\\\n\\\\t\\\\r'\n _ws = '(?:[%s]+)' % _whitespace\n _separator = '%s:=' % _whitespace\n _break = '(?=[%s]|$)' % _separator\n _name = '[^%s]+' % _separator\n _unqualified_name = '(?:[^%s.;\\\\[/]+)' % _separator\n tokens = {'default':[\n (\n '\\\\n', Text, '#pop'),\n (\n \"'\", String.Single, ('#pop', 'quote')),\n (\n '\"', String.Double, 'string'),\n (\n '=', Punctuation),\n (\n ':', Punctuation, 'label'),\n (\n _ws, Text),\n (\n ';.*', Comment.Single),\n (\n '(\\\\$[-+])?0x-?[\\\\da-fA-F]+%s' % _break, Number.Hex),\n (\n '(\\\\$[-+]|\\\\+)?-?\\\\d+%s' % _break, Number.Integer),\n (\n '-?(\\\\d+\\\\.\\\\d*|\\\\.\\\\d+)([eE][-+]?\\\\d+)?[fFdD]?[\\\\x00-\\\\x08\\\\x0b\\\\x0c\\\\x0e-\\\\x1f]*%s' % _break, Number.Float),\n (\n '\\\\$%s' % _name, Name.Variable),\n (\n '\\\\.annotation%s' % _break, Keyword.Reserved, 'annotation'),\n (\n '(\\\\.attribute|\\\\.bytecode|\\\\.debug|\\\\.deprecated|\\\\.enclosing|\\\\.interface|\\\\.line|\\\\.signature|\\\\.source|\\\\.stack|\\\\.var|abstract|annotation|bridge|class|default|enum|field|final|fpstrict|interface|native|private|protected|public|signature|static|synchronized|synthetic|transient|varargs|volatile)%s' % _break,\n Keyword.Reserved),\n (\n '\\\\.catch%s' % _break, Keyword.Reserved, 'caught-exception'),\n (\n '(\\\\.class|\\\\.implements|\\\\.inner|\\\\.super|inner|invisible|invisibleparam|outer|visible|visibleparam)%s' % _break,\n Keyword.Reserved, 'class/convert-dots'),\n (\n '\\\\.field%s' % _break, Keyword.Reserved,\n ('descriptor/convert-dots', 'field')),\n (\n '(\\\\.end|\\\\.limit|use)%s' % _break, Keyword.Reserved,\n 'no-verification'),\n (\n '\\\\.method%s' % _break, Keyword.Reserved, 'method'),\n (\n '\\\\.set%s' % _break, Keyword.Reserved, 'var'),\n (\n '\\\\.throws%s' % _break, Keyword.Reserved, 'exception'),\n (\n '(from|offset|to|using)%s' % _break, Keyword.Reserved, 'label'),\n (\n 'is%s' % _break, Keyword.Reserved,\n ('descriptor/convert-dots', 'var')),\n (\n '(locals|stack)%s' % _break, Keyword.Reserved, 'verification'),\n (\n 'method%s' % _break, Keyword.Reserved, 'enclosing-method'),\n (\n words(('aaload', 'aastore', 'aconst_null', 'aload', 'aload_0', 'aload_1', 'aload_2',\n 'aload_3', 'aload_w', 'areturn', 'arraylength', 'astore', 'astore_0', 'astore_1',\n 'astore_2', 'astore_3', 'astore_w', 'athrow', 'baload', 'bastore', 'bipush',\n 'breakpoint', 'caload', 'castore', 'd2f', 'd2i', 'd2l', 'dadd', 'daload',\n 'dastore', 'dcmpg', 'dcmpl', 'dconst_0', 'dconst_1', 'ddiv', 'dload', 'dload_0',\n 'dload_1', 'dload_2', 'dload_3', 'dload_w', 'dmul', 'dneg', 'drem', 'dreturn',\n 'dstore', 'dstore_0', 'dstore_1', 'dstore_2', 'dstore_3', 'dstore_w', 'dsub',\n 'dup', 'dup2', 'dup2_x1', 'dup2_x2', 'dup_x1', 'dup_x2', 'f2d', 'f2i', 'f2l',\n 'fadd', 'faload', 'fastore', 'fcmpg', 'fcmpl', 'fconst_0', 'fconst_1', 'fconst_2',\n 'fdiv', 'fload', 'fload_0', 'fload_1', 'fload_2', 'fload_3', 'fload_w', 'fmul',\n 'fneg', 'frem', 'freturn', 'fstore', 'fstore_0', 'fstore_1', 'fstore_2', 'fstore_3',\n 'fstore_w', 'fsub', 'i2b', 'i2c', 'i2d', 'i2f', 'i2l', 'i2s', 'iadd', 'iaload',\n 'iand', 'iastore', 'iconst_0', 'iconst_1', 'iconst_2', 'iconst_3', 'iconst_4',\n 'iconst_5', 'iconst_m1', 'idiv', 'iinc', 'iinc_w', 'iload', 'iload_0', 'iload_1',\n 'iload_2', 'iload_3', 'iload_w', 'imul', 'ineg', 'int2byte', 'int2char', 'int2short',\n 'ior', 'irem', 'ireturn', 'ishl', 'ishr', 'istore', 'istore_0', 'istore_1',\n 'istore_2', 'istore_3', 'istore_w', 'isub', 'iushr', 'ixor', 'l2d', 'l2f',\n 'l2i', 'ladd', 'laload', 'land', 'lastore', 'lcmp', 'lconst_0', 'lconst_1',\n 'ldc2_w', 'ldiv', 'lload', 'lload_0', 'lload_1', 'lload_2', 'lload_3', 'lload_w',\n 'lmul', 'lneg', 'lookupswitch', 'lor', 'lrem', 'lreturn', 'lshl', 'lshr',\n 'lstore', 'lstore_0', 'lstore_1', 'lstore_2', 'lstore_3', 'lstore_w', 'lsub',\n 'lushr', 'lxor', 'monitorenter', 'monitorexit', 'nop', 'pop', 'pop2', 'ret',\n 'ret_w', 'return', 'saload', 'sastore', 'sipush', 'swap'),\n suffix=_break), Keyword.Reserved),\n (\n '(anewarray|checkcast|instanceof|ldc|ldc_w|new)%s' % _break,\n Keyword.Reserved, 'class/no-dots'),\n (\n 'invoke(dynamic|interface|nonvirtual|special|static|virtual)%s' % _break, Keyword.Reserved,\n 'invocation'),\n (\n '(getfield|putfield)%s' % _break, Keyword.Reserved,\n ('descriptor/no-dots', 'field')),\n (\n '(getstatic|putstatic)%s' % _break, Keyword.Reserved,\n ('descriptor/no-dots', 'static')),\n (\n words(('goto', 'goto_w', 'if_acmpeq', 'if_acmpne', 'if_icmpeq', 'if_icmpge', 'if_icmpgt',\n 'if_icmple', 'if_icmplt', 'if_icmpne', 'ifeq', 'ifge', 'ifgt', 'ifle', 'iflt',\n 'ifne', 'ifnonnull', 'ifnull', 'jsr', 'jsr_w'),\n suffix=_break),\n Keyword.Reserved, 'label'),\n (\n '(multianewarray|newarray)%s' % _break, Keyword.Reserved,\n 'descriptor/convert-dots'),\n (\n 'tableswitch%s' % _break, Keyword.Reserved, 'table')], \n 'quote':[\n (\n \"'\", String.Single, '#pop'),\n (\n '\\\\\\\\u[\\\\da-fA-F]{4}', String.Escape),\n (\n \"[^'\\\\\\\\]+\", String.Single)], \n 'string':[\n (\n '\"', String.Double, '#pop'),\n (\n '\\\\\\\\([nrtfb\"\\\\\\'\\\\\\\\]|u[\\\\da-fA-F]{4}|[0-3]?[0-7]{1,2})',\n String.Escape),\n (\n '[^\"\\\\\\\\]+', String.Double)], \n 'root':[\n (\n '\\\\n+', Text),\n (\n \"'\", String.Single, 'quote'),\n include('default'),\n (\n '(%s)([ \\\\t\\\\r]*)(:)' % _name,\n bygroups(Name.Label, Text, Punctuation)),\n (\n _name, String.Other)], \n 'annotation':[\n (\n '\\\\n', Text, ('#pop', 'annotation-body')),\n (\n 'default%s' % _break, Keyword.Reserved,\n ('#pop', 'annotation-default')),\n include('default')], \n 'annotation-body':[\n (\n '\\\\n+', Text),\n (\n '\\\\.end%s' % _break, Keyword.Reserved, '#pop'),\n include('default'),\n (\n _name, String.Other, ('annotation-items', 'descriptor/no-dots'))], \n 'annotation-default':[\n (\n '\\\\n+', Text),\n (\n '\\\\.end%s' % _break, Keyword.Reserved, '#pop'),\n include('default'),\n default(('annotation-items', 'descriptor/no-dots'))], \n 'annotation-items':[\n (\n \"'\", String.Single, 'quote'),\n include('default'),\n (\n _name, String.Other)], \n 'caught-exception':[\n (\n 'all%s' % _break, Keyword, '#pop'),\n include('exception')], \n 'class/convert-dots':[\n include('default'),\n (\n '(L)((?:%s[/.])*)(%s)(;)' % (_unqualified_name, _name),\n bygroups(Keyword.Type, Name.Namespace, Name.Class, Punctuation),\n '#pop'),\n (\n '((?:%s[/.])*)(%s)' % (_unqualified_name, _name),\n bygroups(Name.Namespace, Name.Class), '#pop')], \n 'class/no-dots':[\n include('default'),\n (\n '\\\\[+', Punctuation, ('#pop', 'descriptor/no-dots')),\n (\n '(L)((?:%s/)*)(%s)(;)' % (_unqualified_name, _name),\n bygroups(Keyword.Type, Name.Namespace, Name.Class, Punctuation),\n '#pop'),\n (\n '((?:%s/)*)(%s)' % (_unqualified_name, _name),\n bygroups(Name.Namespace, Name.Class), '#pop')], \n 'descriptor/convert-dots':[\n include('default'),\n (\n '\\\\[+', Punctuation),\n (\n '(L)((?:%s[/.])*)(%s?)(;)' % (_unqualified_name, _name),\n bygroups(Keyword.Type, Name.Namespace, Name.Class, Punctuation),\n '#pop'),\n (\n '[^%s\\\\[)L]+' % _separator, Keyword.Type, '#pop'),\n default('#pop')], \n 'descriptor/no-dots':[\n include('default'),\n (\n '\\\\[+', Punctuation),\n (\n '(L)((?:%s/)*)(%s)(;)' % (_unqualified_name, _name),\n bygroups(Keyword.Type, Name.Namespace, Name.Class, Punctuation),\n '#pop'),\n (\n '[^%s\\\\[)L]+' % _separator, Keyword.Type, '#pop'),\n default('#pop')], \n 'descriptors/convert-dots':[\n (\n '\\\\)', Punctuation, '#pop'),\n default('descriptor/convert-dots')], \n 'enclosing-method':[\n (\n _ws, Text),\n (\n '(?=[^%s]*\\\\()' % _separator, Text, ('#pop', 'invocation')),\n default(('#pop', 'class/convert-dots'))], \n 'exception':[\n include('default'),\n (\n '((?:%s[/.])*)(%s)' % (_unqualified_name, _name),\n bygroups(Name.Namespace, Name.Exception), '#pop')], \n 'field':[\n (\n 'static%s' % _break, Keyword.Reserved, ('#pop', 'static')),\n include('default'),\n (\n '((?:%s[/.](?=[^%s]*[/.]))*)(%s[/.])?(%s)' % (\n _unqualified_name, _separator, _unqualified_name, _name),\n bygroups(Name.Namespace, Name.Class, Name.Variable.Instance),\n '#pop')], \n 'invocation':[\n include('default'),\n (\n '((?:%s[/.](?=[^%s(]*[/.]))*)(%s[/.])?(%s)(\\\\()' % (\n _unqualified_name, _separator, _unqualified_name, _name),\n bygroups(Name.Namespace, Name.Class, Name.Function, Punctuation),\n ('#pop', 'descriptor/convert-dots', 'descriptors/convert-dots', 'descriptor/convert-dots'))], \n 'label':[\n include('default'),\n (\n _name, Name.Label, '#pop')], \n 'method':[\n include('default'),\n (\n '(%s)(\\\\()' % _name, bygroups(Name.Function, Punctuation),\n ('#pop', 'descriptor/convert-dots', 'descriptors/convert-dots', 'descriptor/convert-dots'))], \n 'no-verification':[\n (\n '(locals|method|stack)%s' % _break, Keyword.Reserved, '#pop'),\n include('default')], \n 'static':[\n include('default'),\n (\n '((?:%s[/.](?=[^%s]*[/.]))*)(%s[/.])?(%s)' % (\n _unqualified_name, _separator, _unqualified_name, _name),\n bygroups(Name.Namespace, Name.Class, Name.Variable.Class), '#pop')], \n 'table':[\n (\n '\\\\n+', Text),\n (\n 'default%s' % _break, Keyword.Reserved, '#pop'),\n include('default'),\n (\n _name, Name.Label)], \n 'var':[\n include('default'),\n (\n _name, Name.Variable, '#pop')], \n 'verification':[\n include('default'),\n (\n '(Double|Float|Integer|Long|Null|Top|UninitializedThis)%s' % _break, Keyword, '#pop'),\n (\n 'Object%s' % _break, Keyword, ('#pop', 'class/no-dots')),\n (\n 'Uninitialized%s' % _break, Keyword, ('#pop', 'label'))]}\n\n def analyse_text(text):\n score = 0\n if re.search('^\\\\s*\\\\.class\\\\s', text, re.MULTILINE):\n score += 0.5\n if re.search('^\\\\s*[a-z]+_[a-z]+\\\\b', text, re.MULTILINE):\n score += 0.3\n if re.search('^\\\\s*\\\\.(attribute|bytecode|debug|deprecated|enclosing|inner|interface|limit|set|signature|stack)\\\\b', text, re.MULTILINE):\n score += 0.6\n return score\n\n\nclass SarlLexer(RegexLexer):\n __doc__ = '\\n\\tFor `SARL `_ source code.\\n\\t\\n\\t.. versionadded:: 2.4\\n\\t'\n name = 'SARL'\n aliases = ['sarl']\n filenames = ['*.sarl']\n mimetypes = ['text/x-sarl']\n flags = re.MULTILINE | re.DOTALL\n tokens = {'root':[\n (\n '^(\\\\s*(?:[a-zA-Z_][\\\\w.\\\\[\\\\]]*\\\\s+)+?)([a-zA-Z_$][\\\\w$]*)(\\\\s*)(\\\\()',\n bygroups(using(this), Name.Function, Text, Operator)),\n (\n '[^\\\\S\\\\n]+', Text),\n (\n '//.*?\\\\n', Comment.Single),\n (\n '/\\\\*.*?\\\\*/', Comment.Multiline),\n (\n '@[a-zA-Z_][\\\\w.]*', Name.Decorator),\n (\n '(as|break|case|catch|default|do|else|extends|extension|finally|fires|for|if|implements|instanceof|new|on|requires|return|super|switch|throw|throws|try|typeof|uses|while|with)\\\\b',\n Keyword),\n (\n '(abstract|def|dispatch|final|native|override|private|protected|public|static|strictfp|synchronized|transient|val|var|volatile)\\\\b', Keyword.Declaration),\n (\n '(boolean|byte|char|double|float|int|long|short|void)\\\\b',\n Keyword.Type),\n (\n '(package)(\\\\s+)', bygroups(Keyword.Namespace, Text)),\n (\n '(false|it|null|occurrence|this|true|void)\\\\b', Keyword.Constant),\n (\n '(agent|annotation|artifact|behavior|capacity|class|enum|event|interface|skill|space)(\\\\s+)', bygroups(Keyword.Declaration, Text),\n 'class'),\n (\n '(import)(\\\\s+)', bygroups(Keyword.Namespace, Text), 'import'),\n (\n '\"(\\\\\\\\\\\\\\\\|\\\\\\\\\"|[^\"])*\"', String),\n (\n \"'(\\\\\\\\\\\\\\\\|\\\\\\\\'|[^'])*'\", String),\n (\n '[a-zA-Z_]\\\\w*:', Name.Label),\n (\n '[a-zA-Z_$]\\\\w*', Name),\n (\n '[~^*!%&\\\\[\\\\](){}<>\\\\|+=:;,./?-]', Operator),\n (\n '[0-9][0-9]*\\\\.[0-9]+([eE][0-9]+)?[fd]?', Number.Float),\n (\n '0x[0-9a-fA-F]+', Number.Hex),\n (\n '[0-9]+L?', Number.Integer),\n (\n '\\\\n', Text)], \n 'class':[\n (\n '[a-zA-Z_]\\\\w*', Name.Class, '#pop')], \n 'import':[\n (\n '[\\\\w.]+\\\\*?', Name.Namespace, '#pop')]}", "sub_path": "pycfiles/libopenstorage_openstorage-0.42.24.1-py3-none-any/jvm.cpython-36.py", "file_name": "jvm.cpython-36.py", "file_ext": "py", "file_size_in_byte": 59800, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pygments.lexer.RegexLexer", "line_number": 28, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 34, "usage_type": "attribute"}, {"api_name": "re.UNICODE", "line_number": 34, "usage_type": "attribute"}, {"api_name": "pygments.token.Text", "line_number": 37, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 39, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 39, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 41, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 41, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 44, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 47, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 47, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 47, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 47, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 47, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 47, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 47, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 49, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 49, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 52, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 52, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 55, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 55, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 57, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 57, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 57, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 57, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 59, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 59, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 61, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 61, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 61, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 61, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 64, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 64, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 64, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 64, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 67, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 67, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 67, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 67, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 70, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 72, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 72, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 75, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 75, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 75, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 75, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 77, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 77, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 79, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 82, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 82, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 84, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 84, "usage_type": "name"}, {"api_name": "pygments.token.Number.Bin", "line_number": 86, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 86, "usage_type": "name"}, {"api_name": "pygments.token.Number.Oct", "line_number": 88, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 88, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 90, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 90, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 92, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 94, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 96, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 99, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 99, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 102, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 105, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 105, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 130, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 132, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 133, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 135, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 136, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 137, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 137, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 139, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 145, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 151, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 151, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 159, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 159, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 159, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 161, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 162, "usage_type": "call"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 164, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 164, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 167, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 169, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 171, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 171, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 173, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 173, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 175, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 175, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 175, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 177, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 177, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 177, "usage_type": "argument"}, {"api_name": "pygments.token.String", "line_number": 179, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 181, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 183, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 183, "usage_type": "name"}, {"api_name": "pygments.token.Text.Symbol", "line_number": 185, "usage_type": "attribute"}, {"api_name": "pygments.token.Text", "line_number": 185, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 187, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 189, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 191, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 193, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 195, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 197, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 199, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 201, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 204, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 204, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 206, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 206, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 208, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 208, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 210, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 214, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 214, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 214, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Class", "line_number": 214, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 214, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 216, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 217, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 219, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 221, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 223, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 223, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 226, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 227, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 229, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 231, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 231, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 231, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 233, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 237, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 237, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 237, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 237, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 237, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 241, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 241, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 241, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 241, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 243, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 243, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 246, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 247, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 249, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 251, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 253, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 255, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 257, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 257, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 260, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 260, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 262, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 262, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 265, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 265, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 267, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 267, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 269, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 269, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 271, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 271, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 274, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 274, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 277, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 279, "usage_type": "name"}, {"api_name": "pygments.token.String.Interpol", "line_number": 281, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 281, "usage_type": "name"}, {"api_name": "pygments.token.String.Interpol", "line_number": 283, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 283, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 285, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 288, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 290, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 291, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 294, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 295, "usage_type": "call"}, {"api_name": "pygments.token.String.Interpol", "line_number": 298, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 298, "usage_type": "name"}, {"api_name": "pygments.token.String.Interpol", "line_number": 300, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 300, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 301, "usage_type": "call"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 304, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 310, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 310, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 314, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 314, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 314, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 314, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 314, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 314, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 314, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 316, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 318, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 318, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 320, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 320, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 322, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 322, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 325, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 328, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 328, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 330, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 330, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 333, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 333, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 335, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 335, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 335, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 335, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 337, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 337, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 340, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 340, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 340, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 340, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 340, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 340, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 343, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 343, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 343, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 343, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 343, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 343, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 345, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 348, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 348, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 348, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 348, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 351, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 351, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 351, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 351, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 353, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 355, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 357, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 357, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 359, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 359, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 361, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 364, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 367, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 367, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 367, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 367, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 369, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 369, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 371, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 373, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 375, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 378, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 379, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 382, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 384, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 385, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 388, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 390, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 392, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 393, "usage_type": "call"}, {"api_name": "pygments.lexer.Lexer", "line_number": 396, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 410, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 416, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 416, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment.Preproc", "line_number": 419, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 419, "usage_type": "name"}, {"api_name": "pygments.lexer.default", "line_number": 420, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 424, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 424, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 424, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 424, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 424, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 424, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 424, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 426, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 428, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 428, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 430, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 430, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 432, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 432, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 435, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 438, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 438, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 441, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 441, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 443, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 443, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 443, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 443, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 445, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 445, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 447, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 447, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 447, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 447, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 450, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 450, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 450, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 450, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 452, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 452, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 454, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 454, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 456, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 456, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 458, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 458, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 460, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 462, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 464, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 464, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 466, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 466, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 466, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 466, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 468, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 468, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 470, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 472, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 474, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 474, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 476, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 476, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 478, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 478, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 480, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 483, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 483, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 486, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 486, "usage_type": "name"}, {"api_name": "pygments.util.shebang_matches", "line_number": 489, "usage_type": "call"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 492, "usage_type": "name"}, {"api_name": "pygments.token.String.Escape", "line_number": 501, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 501, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 503, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 506, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 507, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 509, "usage_type": "name"}, {"api_name": "pygments.token.String.Doc", "line_number": 512, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 512, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 513, "usage_type": "call"}, {"api_name": "pygments.token.String.Doc", "line_number": 515, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 515, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 518, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 519, "usage_type": "call"}, {"api_name": "pygments.token.String.Regex", "line_number": 522, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 522, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 523, "usage_type": "call"}, {"api_name": "pygments.token.String.Regex", "line_number": 525, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 525, "usage_type": "name"}, {"api_name": "pygments.token.String.Regex", "line_number": 527, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 527, "usage_type": "name"}, {"api_name": "pygments.token.String.Regex", "line_number": 530, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 530, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 531, "usage_type": "call"}, {"api_name": "pygments.token.String.Regex", "line_number": 533, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 533, "usage_type": "name"}, {"api_name": "pygments.token.String.Regex", "line_number": 535, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 535, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 538, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 539, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 541, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 544, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 546, "usage_type": "name"}, {"api_name": "pygments.token.Comment", "line_number": 548, "usage_type": "name"}, {"api_name": "pygments.token.Comment", "line_number": 550, "usage_type": "name"}, {"api_name": "pygments.token.String.Regex", "line_number": 552, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 552, "usage_type": "name"}, {"api_name": "pygments.token.String.Regex", "line_number": 554, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 554, "usage_type": "name"}, {"api_name": "pygments.token.String.Symbol", "line_number": 556, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 556, "usage_type": "name"}, {"api_name": "pygments.token.String.Other", "line_number": 558, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 558, "usage_type": "name"}, {"api_name": "pygments.token.String.Symbol", "line_number": 560, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 560, "usage_type": "name"}, {"api_name": "pygments.token.String.Doc", "line_number": 563, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 563, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 565, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 567, "usage_type": "name"}, {"api_name": "pygments.token.Name.Entity", "line_number": 569, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 569, "usage_type": "name"}, {"api_name": "pygments.token.Name.Variable", "line_number": 572, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 572, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 575, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 575, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 577, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 580, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 582, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 584, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 584, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 587, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 587, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 590, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 592, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 592, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 595, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 597, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 597, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 600, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 603, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 603, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 606, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 606, "usage_type": "name"}, {"api_name": "pygments.token.Name.Constant", "line_number": 608, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 608, "usage_type": "name"}, {"api_name": "pygments.token.Name.Builtin", "line_number": 611, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 611, "usage_type": "name"}, {"api_name": "pygments.token.Name.Function", "line_number": 614, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 614, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 616, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 616, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 618, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 618, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 620, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 620, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 622, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 625, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 628, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 630, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 632, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 632, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 634, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 637, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 696, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 696, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 698, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 700, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 700, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 702, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 702, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 704, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 704, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 706, "usage_type": "name"}, {"api_name": "pygments.token.String.Symbol", "line_number": 708, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 708, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 710, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 710, "usage_type": "name"}, {"api_name": "pygments.token.String.Symbol", "line_number": 712, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 712, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 714, "usage_type": "name"}, {"api_name": "pygments.lexer.words", "line_number": 716, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 716, "usage_type": "name"}, {"api_name": "pygments.lexer.words", "line_number": 718, "usage_type": "call"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 718, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 718, "usage_type": "name"}, {"api_name": "pygments.lexer.words", "line_number": 720, "usage_type": "call"}, {"api_name": "pygments.token.Name.Builtin", "line_number": 720, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 720, "usage_type": "name"}, {"api_name": "pygments.token.Name.Function", "line_number": 722, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 722, "usage_type": "name"}, {"api_name": "pygments.token.Name.Variable", "line_number": 724, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 724, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 726, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 728, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 730, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 741, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 743, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 743, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 747, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 747, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 747, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 747, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 747, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 747, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 747, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 749, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 751, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 751, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 753, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 753, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 755, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 755, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 758, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 760, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 760, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 762, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 762, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 764, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 764, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 764, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 764, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 766, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 766, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 766, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 766, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 768, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 770, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 772, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 772, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 772, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 772, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 774, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 774, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 776, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 778, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 780, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 780, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 782, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 782, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 784, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 784, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 786, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 789, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 789, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 792, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 792, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 795, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 801, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 801, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 806, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 806, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 806, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 806, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 806, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 806, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 806, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 808, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 810, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 810, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 812, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 812, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 815, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 815, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 818, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 821, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 821, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 824, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 824, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 826, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 826, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 826, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 826, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 828, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 828, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 831, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 831, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 831, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 831, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 833, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 833, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 833, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 833, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 835, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 837, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 837, "usage_type": "name"}, {"api_name": "pygments.token.String.Interpol", "line_number": 839, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 839, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 842, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 842, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 842, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 842, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 844, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 844, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 846, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 848, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 850, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 850, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 853, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 853, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 855, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 855, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 858, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 858, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 860, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 860, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 862, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 862, "usage_type": "name"}, {"api_name": "pygments.token.Number.Bin", "line_number": 864, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 864, "usage_type": "name"}, {"api_name": "pygments.token.Number.Bin", "line_number": 866, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 866, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 868, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 868, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 870, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 870, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 872, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 875, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 875, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 879, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 879, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 882, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 882, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 884, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 884, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 886, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 886, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 888, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 888, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 891, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 897, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 897, "usage_type": "attribute"}, {"api_name": "re.UNICODE", "line_number": 897, "usage_type": "attribute"}, {"api_name": "pygments.unistring.combine", "line_number": 898, "usage_type": "call"}, {"api_name": "pygments.unistring", "line_number": 898, "usage_type": "name"}, {"api_name": "pygments.unistring.combine", "line_number": 899, "usage_type": "call"}, {"api_name": "pygments.unistring", "line_number": 899, "usage_type": "name"}, {"api_name": "pygments.token.Name.Attribute", "line_number": 903, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 903, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 905, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 907, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 909, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 911, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 911, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 913, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 913, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 915, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 917, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 919, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 921, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 923, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 925, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 927, "usage_type": "name"}, {"api_name": "pygments.token.String.Char", "line_number": 929, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 929, "usage_type": "name"}, {"api_name": "pygments.token.Number", "line_number": 932, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 934, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 934, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 934, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 934, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 936, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 936, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 936, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 938, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 938, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 938, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 940, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 940, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 940, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 942, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 942, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 942, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 942, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 944, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 944, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 944, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 946, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 946, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 946, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 948, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 948, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 948, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword", "line_number": 951, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 953, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 956, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 956, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 959, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 959, "usage_type": "name"}, {"api_name": "pygments.token.Name.Property", "line_number": 962, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 962, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 965, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 965, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 965, "usage_type": "argument"}, {"api_name": "pygments.lexer.bygroups", "line_number": 967, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 967, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 967, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 969, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 971, "usage_type": "name"}, {"api_name": "pygments.token.Name.Property", "line_number": 973, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 973, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 976, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 978, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 978, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Class", "line_number": 978, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 978, "usage_type": "name"}, {"api_name": "pygments.token.Name.Function", "line_number": 978, "usage_type": "attribute"}, {"api_name": "pygments.token.Name.Function", "line_number": 980, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 980, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 983, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 983, "usage_type": "argument"}, {"api_name": "pygments.token.Text", "line_number": 983, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 985, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 987, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 989, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 991, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 993, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 996, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 1002, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 1002, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1006, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1006, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 1006, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 1006, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 1006, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 1006, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1006, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1008, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 1010, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1010, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 1012, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1012, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 1014, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1014, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1017, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1020, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1020, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1023, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1023, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1025, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1025, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 1025, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1025, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 1027, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1027, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1029, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1029, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1029, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1029, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1032, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1032, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 1032, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1032, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1034, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1036, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1038, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1040, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 1042, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1042, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 1044, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1046, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 1048, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1048, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 1050, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1050, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1052, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1052, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1054, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1057, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1057, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1060, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1060, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1063, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1065, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1067, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 1070, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 1076, "usage_type": "attribute"}, {"api_name": "re.IGNORECASE", "line_number": 1076, "usage_type": "attribute"}, {"api_name": "pygments.token.Text", "line_number": 1079, "usage_type": "name"}, {"api_name": "pygments.token.Comment", "line_number": 1081, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 1083, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1083, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1085, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1087, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1089, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1090, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1091, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1092, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1093, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1094, "usage_type": "call"}, {"api_name": "pygments.token.Number.Float", "line_number": 1096, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1096, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 1098, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1098, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1100, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1100, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1102, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1105, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1105, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 1105, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 1105, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1105, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1107, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1109, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1111, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1115, "usage_type": "name"}, {"api_name": "pygments.token.Name.Builtin", "line_number": 1119, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1119, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1123, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1123, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 1126, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1129, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1131, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1133, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 1136, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1143, "usage_type": "name"}, {"api_name": "pygments.token.Comment", "line_number": 1145, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1148, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1150, "usage_type": "name"}, {"api_name": "pygments.token.Operator.Word", "line_number": 1152, "usage_type": "attribute"}, {"api_name": "pygments.token.Operator", "line_number": 1152, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 1154, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1157, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1157, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 1157, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1157, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1160, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1160, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1160, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1160, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1162, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1162, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1165, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1165, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1165, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1165, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1169, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1169, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1169, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1169, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1173, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1173, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1173, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1173, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 1176, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1176, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1179, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1182, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1182, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Builtin", "line_number": 1182, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1182, "usage_type": "name"}, {"api_name": "pygments.token.Name.Builtin", "line_number": 1185, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1185, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1188, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1188, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 1188, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1188, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 1190, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1190, "usage_type": "name"}, {"api_name": "pygments.token.Number.Oct", "line_number": 1192, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1192, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 1194, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1194, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1196, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1196, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1198, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1198, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 1200, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 1202, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1202, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1204, "usage_type": "name"}, {"api_name": "pygments.lexer.combined", "line_number": 1204, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 1206, "usage_type": "name"}, {"api_name": "pygments.lexer.combined", "line_number": 1206, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 1208, "usage_type": "name"}, {"api_name": "pygments.lexer.combined", "line_number": 1208, "usage_type": "call"}, {"api_name": "pygments.token.String.Doc", "line_number": 1210, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1210, "usage_type": "name"}, {"api_name": "pygments.token.Name.Function", "line_number": 1213, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1213, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1216, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1216, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1219, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1219, "usage_type": "name"}, {"api_name": "pygments.token.Name.Variable", "line_number": 1222, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1222, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1225, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1227, "usage_type": "name"}, {"api_name": "pygments.token.String.Escape", "line_number": 1231, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1231, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1234, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1235, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 1237, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 1240, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1240, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1241, "usage_type": "call"}, {"api_name": "pygments.token.String", "line_number": 1244, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1245, "usage_type": "call"}, {"api_name": "pygments.token.Operator", "line_number": 1248, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1250, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1252, "usage_type": "name"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 1255, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1268, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 1270, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1270, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 1272, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1272, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 1274, "usage_type": "name"}, {"api_name": "pygments.token.Punctuation", "line_number": 1276, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1278, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 1280, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1280, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 1282, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1282, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1284, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1284, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 1286, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1286, "usage_type": "name"}, {"api_name": "pygments.token.Name.Variable", "line_number": 1288, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1288, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1290, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1290, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1293, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1293, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1295, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1295, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1298, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1298, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1300, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1300, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1303, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1303, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1306, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1306, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1308, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1308, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1310, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1310, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1312, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1312, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1314, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1314, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1317, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1317, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1319, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1319, "usage_type": "name"}, {"api_name": "pygments.lexer.words", "line_number": 1321, "usage_type": "call"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1344, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1344, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1347, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1347, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1349, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1349, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1352, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1352, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1355, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1355, "usage_type": "name"}, {"api_name": "pygments.lexer.words", "line_number": 1358, "usage_type": "call"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1362, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1362, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1364, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1364, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1367, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1367, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 1370, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1370, "usage_type": "name"}, {"api_name": "pygments.token.String.Escape", "line_number": 1372, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1372, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 1374, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1374, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 1377, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1377, "usage_type": "name"}, {"api_name": "pygments.token.String.Escape", "line_number": 1380, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1380, "usage_type": "name"}, {"api_name": "pygments.token.String.Double", "line_number": 1382, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1382, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1385, "usage_type": "name"}, {"api_name": "pygments.token.String.Single", "line_number": 1387, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1387, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1388, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1391, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1391, "usage_type": "argument"}, {"api_name": "pygments.token.Punctuation", "line_number": 1391, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Label", "line_number": 1391, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1391, "usage_type": "name"}, {"api_name": "pygments.token.String.Other", "line_number": 1393, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1393, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1396, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1398, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1398, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1400, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1403, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1405, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1405, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1406, "usage_type": "call"}, {"api_name": "pygments.token.String.Other", "line_number": 1408, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1408, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1411, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1413, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1413, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1414, "usage_type": "call"}, {"api_name": "pygments.lexer.default", "line_number": 1415, "usage_type": "call"}, {"api_name": "pygments.token.String.Single", "line_number": 1418, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1418, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1419, "usage_type": "call"}, {"api_name": "pygments.token.String.Other", "line_number": 1421, "usage_type": "attribute"}, {"api_name": "pygments.token.String", "line_number": 1421, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1424, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1425, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1427, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1430, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1430, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1430, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1430, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1430, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1430, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1430, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1434, "usage_type": "call"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1434, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1434, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1434, "usage_type": "attribute"}, {"api_name": "pygments.lexer.include", "line_number": 1436, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1438, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1441, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1441, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1441, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1441, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1441, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1441, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1441, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1445, "usage_type": "call"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1445, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1445, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1445, "usage_type": "attribute"}, {"api_name": "pygments.lexer.include", "line_number": 1447, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1449, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1452, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1452, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1452, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1452, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1452, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1452, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1452, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1455, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1455, "usage_type": "name"}, {"api_name": "pygments.lexer.default", "line_number": 1456, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1458, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1460, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1463, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1463, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1463, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1463, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1463, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1466, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1466, "usage_type": "name"}, {"api_name": "pygments.lexer.default", "line_number": 1467, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1470, "usage_type": "name"}, {"api_name": "pygments.lexer.default", "line_number": 1471, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1474, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1476, "usage_type": "name"}, {"api_name": "pygments.lexer.default", "line_number": 1477, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1479, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1482, "usage_type": "call"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1482, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1482, "usage_type": "name"}, {"api_name": "pygments.token.Name.Exception", "line_number": 1482, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1485, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1485, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1486, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1490, "usage_type": "call"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1490, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1490, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1490, "usage_type": "attribute"}, {"api_name": "pygments.token.Name.Variable", "line_number": 1490, "usage_type": "attribute"}, {"api_name": "pygments.lexer.include", "line_number": 1493, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1497, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1497, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1497, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1497, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1497, "usage_type": "attribute"}, {"api_name": "pygments.token.Name.Function", "line_number": 1497, "usage_type": "attribute"}, {"api_name": "pygments.lexer.include", "line_number": 1500, "usage_type": "call"}, {"api_name": "pygments.token.Name.Label", "line_number": 1502, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1502, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1504, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1506, "usage_type": "call"}, {"api_name": "pygments.token.Punctuation", "line_number": 1506, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 1506, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1506, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1510, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1510, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1511, "usage_type": "call"}, {"api_name": "pygments.lexer.include", "line_number": 1513, "usage_type": "call"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1517, "usage_type": "call"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1517, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1517, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1517, "usage_type": "attribute"}, {"api_name": "pygments.token.Name.Variable", "line_number": 1517, "usage_type": "attribute"}, {"api_name": "pygments.token.Text", "line_number": 1520, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Reserved", "line_number": 1522, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1522, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1523, "usage_type": "call"}, {"api_name": "pygments.token.Name.Label", "line_number": 1525, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1525, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1527, "usage_type": "call"}, {"api_name": "pygments.token.Name.Variable", "line_number": 1529, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1529, "usage_type": "name"}, {"api_name": "pygments.lexer.include", "line_number": 1531, "usage_type": "call"}, {"api_name": "pygments.token.Keyword", "line_number": 1533, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1535, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1537, "usage_type": "name"}, {"api_name": "re.search", "line_number": 1541, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 1541, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 1543, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 1543, "usage_type": "attribute"}, {"api_name": "re.search", "line_number": 1545, "usage_type": "call"}, {"api_name": "re.MULTILINE", "line_number": 1545, "usage_type": "attribute"}, {"api_name": "pygments.lexer.RegexLexer", "line_number": 1550, "usage_type": "name"}, {"api_name": "re.MULTILINE", "line_number": 1556, "usage_type": "attribute"}, {"api_name": "re.DOTALL", "line_number": 1556, "usage_type": "attribute"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1560, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1560, "usage_type": "argument"}, {"api_name": "pygments.token.Operator", "line_number": 1560, "usage_type": "argument"}, {"api_name": "pygments.lexer.using", "line_number": 1560, "usage_type": "call"}, {"api_name": "pygments.lexer.this", "line_number": 1560, "usage_type": "argument"}, {"api_name": "pygments.token.Name.Function", "line_number": 1560, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1560, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1562, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Single", "line_number": 1564, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1564, "usage_type": "name"}, {"api_name": "pygments.token.Comment.Multiline", "line_number": 1566, "usage_type": "attribute"}, {"api_name": "pygments.token.Comment", "line_number": 1566, "usage_type": "name"}, {"api_name": "pygments.token.Name.Decorator", "line_number": 1568, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1568, "usage_type": "name"}, {"api_name": "pygments.token.Keyword", "line_number": 1571, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1573, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1573, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Type", "line_number": 1576, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1576, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1578, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1578, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 1578, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1578, "usage_type": "name"}, {"api_name": "pygments.token.Keyword.Constant", "line_number": 1580, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1580, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1582, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1582, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Declaration", "line_number": 1582, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1582, "usage_type": "name"}, {"api_name": "pygments.lexer.bygroups", "line_number": 1585, "usage_type": "call"}, {"api_name": "pygments.token.Text", "line_number": 1585, "usage_type": "argument"}, {"api_name": "pygments.token.Keyword.Namespace", "line_number": 1585, "usage_type": "attribute"}, {"api_name": "pygments.token.Keyword", "line_number": 1585, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1587, "usage_type": "name"}, {"api_name": "pygments.token.String", "line_number": 1589, "usage_type": "name"}, {"api_name": "pygments.token.Name.Label", "line_number": 1591, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1591, "usage_type": "name"}, {"api_name": "pygments.token.Name", "line_number": 1593, "usage_type": "name"}, {"api_name": "pygments.token.Operator", "line_number": 1595, "usage_type": "name"}, {"api_name": "pygments.token.Number.Float", "line_number": 1597, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1597, "usage_type": "name"}, {"api_name": "pygments.token.Number.Hex", "line_number": 1599, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1599, "usage_type": "name"}, {"api_name": "pygments.token.Number.Integer", "line_number": 1601, "usage_type": "attribute"}, {"api_name": "pygments.token.Number", "line_number": 1601, "usage_type": "name"}, {"api_name": "pygments.token.Text", "line_number": 1603, "usage_type": "name"}, {"api_name": "pygments.token.Name.Class", "line_number": 1606, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1606, "usage_type": "name"}, {"api_name": "pygments.token.Name.Namespace", "line_number": 1609, "usage_type": "attribute"}, {"api_name": "pygments.token.Name", "line_number": 1609, "usage_type": "name"}]} +{"seq_id": "201357429", "text": "from django.shortcuts import render, get_object_or_404\nfrom django.http import HttpResponse\nfrom datetime import timezone\n\n# Create your views here.\nfrom account.models import Professor, Usage\n\ndef index(request):\n professors = Professor.objects.all()\n content = {'professors':professors}\n return render(request,'account/index.html', content)\n\ndef list(request):\n usages = Usage.objects.all()\n content = {'usages':usages}\n return render(request, 'account/list.html', content)\n\ndef save(request):\n if request.method == \"GET\":\n try:\n\n u_professor= Professor.objects.get(name=request.GET['professor_name'])\n #return render(request,'account/test.html', {'professor_name':u_professor})\n new_usage = Usage.objects.create(user=request.GET['user_name'],professor=u_professor)\n except:\n return HttpResponse(\"Can't create new usage account\")\n #new_usage.user = request.user_name\n #new_usage.professor = request.professor_name\n # 최종적으로 아래 문들은 모두 위쪽의 create 안쪽에 넣어야 함 2015.07.08\n new_usage.page_count = int(request.GET['page_count'])\n new_usage.page_size = request.GET['page_size']\n new_usage.page_media = request.GET['page_media']\n u_professor.usage_sum = u_professor.usage_sum + int(request.GET['page_count'])\n\n new_usage.save()\n u_professor.save()\n else: return HttpResponse('Error')\n return HttpResponse('ok')\n\ndef usage(request):\n professors = Professor.objects.all()\n content = {'professors': professors}\n return render(request, 'account/usage.html', content)\n\ndef detail(request,professor_id):\n professor = get_object_or_404(Professor, pk=professor_id)\n return render(request, 'account/detail.html', {'professor':professor})\n\ndef test(request):\n #content = {'test':'test'}\n #return render(request,'account/test.html', content)\n return HttpResponse('test ok')", "sub_path": "account/views.py", "file_name": "views.py", "file_ext": "py", "file_size_in_byte": 1969, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "account.models.Professor.objects.all", "line_number": 9, "usage_type": "call"}, {"api_name": "account.models.Professor.objects", "line_number": 9, "usage_type": "attribute"}, {"api_name": "account.models.Professor", "line_number": 9, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 11, "usage_type": "call"}, {"api_name": "account.models.Usage.objects.all", "line_number": 14, "usage_type": "call"}, {"api_name": "account.models.Usage.objects", "line_number": 14, "usage_type": "attribute"}, {"api_name": "account.models.Usage", "line_number": 14, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 16, "usage_type": "call"}, {"api_name": "account.models.Professor.objects.get", "line_number": 22, "usage_type": "call"}, {"api_name": "account.models.Professor.objects", "line_number": 22, "usage_type": "attribute"}, {"api_name": "account.models.Professor", "line_number": 22, "usage_type": "name"}, {"api_name": "account.models.Usage.objects.create", "line_number": 24, "usage_type": "call"}, {"api_name": "account.models.Usage.objects", "line_number": 24, "usage_type": "attribute"}, {"api_name": "account.models.Usage", "line_number": 24, "usage_type": "name"}, {"api_name": "django.http.HttpResponse", "line_number": 26, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 37, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 38, "usage_type": "call"}, {"api_name": "account.models.Professor.objects.all", "line_number": 41, "usage_type": "call"}, {"api_name": "account.models.Professor.objects", "line_number": 41, "usage_type": "attribute"}, {"api_name": "account.models.Professor", "line_number": 41, "usage_type": "name"}, {"api_name": "django.shortcuts.render", "line_number": 43, "usage_type": "call"}, {"api_name": "django.shortcuts.get_object_or_404", "line_number": 46, "usage_type": "call"}, {"api_name": "account.models.Professor", "line_number": 46, "usage_type": "argument"}, {"api_name": "django.shortcuts.render", "line_number": 47, "usage_type": "call"}, {"api_name": "django.http.HttpResponse", "line_number": 52, "usage_type": "call"}]} +{"seq_id": "557519378", "text": "#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on %(date)s\n\n@author: %(username)s\n\"\"\"\nimport sys\nimport argparse\nimport logging\nimport requests\nimport re\nimport json\nimport os\n\n\n# set argument parser\nparser = argparse.ArgumentParser(description='Download CV.')\nparser.add_argument(\"-cvdir\", type = str,\n help = \"Directory to store output of this file.\",\n default = \"data/faculty_cv\")\nparser.add_argument(\"-namesdir\", type = str,\n help = \"Directory storing faculty roster.\",\n default = \"data/faculty_names\")\nparser.add_argument(\"-v\", \"--verbose\", \n help = \"Set logging level to DEBUG\",\n action = \"store_true\")\nargs = parser.parse_args()\n\n# set logging\nlog = logging.getLogger(__name__)\nlog.setLevel(logging.ERROR)\nif args.verbose:\n log.setLevel(logging.DEBUG)\nloghandler = logging.StreamHandler(sys.stderr)\nloghandler.setFormatter(logging.Formatter(\"[%(asctime)s %(message)s]\"))\nlog.addHandler(loghandler)\n\n\nuser_agent = \"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.109 Safari/537.36\"\n\n\n\n\n\n\n# download pdf and save meta information in json\ndef download_cv(name, url):\n \"\"\"\n name (str, name of the junior faculty)\n url (str, url of the professor's CV)\n \"\"\"\n r = requests.get(url, headers={\n 'Accept': 'text/html; charset=iso-8859-1',\n 'User-Agent': user_agent}) \n filename = name.replace(\" \", \"\").replace(\".\", \"\")\n \n # save meta information in json\n with open(os.path.join(args.cvdir, filename + \".json\"), \"w\") as j:\n json.dump(dict(r.headers), j)\n if r.status_code != 200:\n r.raise_for_status()\n else:\n if url.endswith(\".pdf\"):\n with open(os.path.join(args.cvdir, filename + \".pdf\"), \"wb\") as f:\n f.write(r.content)\n # if not downloadable, write message\n if url.endswith(\".html\"):\n with open(os.path.join(args.cvdir, filename + \".html\"), \"w\") as h:\n h.write(r.text)\n log.info(\"CV accessed for {}\".format(name))\n \n\n\nif __name__ == \"__main__\":\n \n files = [file for file in os.listdir(args.namesdir) if file.endswith(\".json\")]\n for file in files:\n with open(os.path.join(args.namesdir, file), \"r\") as f:\n names = json.load(f)\n for name in names:\n if re.search(\"people\", name, re.IGNORECASE):\n continue\n else:\n input_url = str(input(\"Enter url for {}. If none, enter 'none': \".format(name)))\n if input_url != 'none':\n download_cv(name, input_url)\n else:\n log.info(\"Unable to find stand alone CV\")\n sys.exit()", "sub_path": "script/download_cv.py", "file_name": "download_cv.py", "file_ext": "py", "file_size_in_byte": 2801, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "argparse.ArgumentParser", "line_number": 18, "usage_type": "call"}, {"api_name": "logging.getLogger", "line_number": 31, "usage_type": "call"}, {"api_name": "logging.ERROR", "line_number": 32, "usage_type": "attribute"}, {"api_name": "logging.DEBUG", "line_number": 34, "usage_type": "attribute"}, {"api_name": "logging.StreamHandler", "line_number": 35, "usage_type": "call"}, {"api_name": "sys.stderr", "line_number": 35, "usage_type": "attribute"}, {"api_name": "logging.Formatter", "line_number": 36, "usage_type": "call"}, {"api_name": "requests.get", "line_number": 53, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 59, "usage_type": "call"}, {"api_name": "os.path", "line_number": 59, "usage_type": "attribute"}, {"api_name": "json.dump", "line_number": 60, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 65, "usage_type": "call"}, {"api_name": "os.path", "line_number": 65, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 69, "usage_type": "call"}, {"api_name": "os.path", "line_number": 69, "usage_type": "attribute"}, {"api_name": "os.listdir", "line_number": 77, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 79, "usage_type": "call"}, {"api_name": "os.path", "line_number": 79, "usage_type": "attribute"}, {"api_name": "json.load", "line_number": 80, "usage_type": "call"}, {"api_name": "re.search", "line_number": 82, "usage_type": "call"}, {"api_name": "re.IGNORECASE", "line_number": 82, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 90, "usage_type": "call"}]} +{"seq_id": "362211403", "text": "import pyperclip\n# types:\n# f float\n# F double\n# i int\n# I unsigned int\n# q i64\n# Q u64\n# s i16\n# S u16\n# u i8\n# U u8\n# B bool\n# G = fF\n# L = iqsu\n# P = IQSU\n# A = GLPB\n\nG = \"fF\"\nL = \"iqsu\"\nP = \"IQSU\"\n\n\n# Also: S = sqmat\n\ntypes = []\n\ntype_name_dict = {\"f\" : \"float\", \"F\" : \"double\", \"i\" : \"int\", \"I\" : \"glm::uint\", \"q\" : \"glm::i64\", \"Q\" : \"glm::u64\", \"s\" : \"glm::i16\", \"S\" : \"glm::u16\", \"u\" : \"glm::i8\", \"U\" : \"glm::u8\", \"B\" : \"bool\"}\n\ndef get_types(accepted_types):\n out = []\n accepted_types = accepted_types.replace(\"A\", \"GLPB\").replace(\"P\", \"IQSU\").replace(\"L\", \"iqsu\").replace(\"G\", \"fF\")\n for short_name in type_name_dict:\n if short_name in accepted_types:\n out.append(type_name_dict[short_name])\n return out\n\ndef get_return_function(return_type):\n if return_type == \"N\":\n return \"pack\"\n if return_type == \"V\":\n return \"pack\"\n if return_type == \"Q\":\n return \"pack\"\n\ndef gen_define(arg_name):\n return \"#define PyGLM_MAKE_GLM_FUNC_{}(NAME)\\\\\\n\".format(arg_name)\n\ndef gen_end_O():\n return \"\"\"\\tPyGLM_TYPEERROR_O(\"invalid argument type for \" #NAME \"(): \", arg);\\\\\\n\\treturn NULL;\\\\\\n}\"\"\"\n\ndef gen_end_2O():\n return \"\"\"\\tPyGLM_TYPEERROR_2O(\"invalid argument type(s) for \" #NAME \"(): \", arg1, arg2);\\\\\\n\\treturn NULL;\\\\\\n}\"\"\"\n\ndef gen_end_nO():\n return \"\"\"\\tPyErr_SetString(PyExc_TypeError, \"invalid argument type(s) for \" #NAME \"()\");\\\\\\n\\treturn NULL;\\\\\\n}\"\"\"\n\nMETH_O = \"static PyObject*\\\\\\nNAME##_(PyObject*, PyObject* arg) {\\\\\\n\"\nMETH_VA = \"static PyObject*\\\\\\nNAME##_(PyObject*, PyObject* args) {\\\\\\n\"\nVA_2O = \"\\tPyObject *arg1, *arg2;\\\\\\n\\tPyGLM_Arg_Unpack_2O(args, #NAME, arg1, arg2);\\\\\\n\"\nVA_3O = \"\\tPyObject *arg1, *arg2, *arg3;\\\\\\n\\tPyGLM_Arg_Unpack_3O(args, #NAME, arg1, arg2, arg3);\\\\\\n\"\nVA_4O = \"\\tPyObject *arg1, *arg2, *arg3, *arg4;\\\\\\n\\tPyGLM_Arg_Unpack_4O(args, #NAME, arg1, arg2, arg3, arg4);\\\\\\n\"\nVA_6O = \"\\tPyObject *arg1, *arg2, *arg3, *arg4, *arg5, *arg6;\\\\\\n\\tPyGLM_Arg_Unpack_6O(args, #NAME, arg1, arg2, arg3, arg4, arg5, arg6);\\\\\\n\"\n\n\ndef make_V():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_M():\n out = \"\"\n for C in (2, 3, 4):\n for R in (2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Mat_Check({C}, {R}, {T}, arg)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_mat<{C}, {R}, {T}>(arg)));\\\\\\n\\t}}\\\\\\n\".format(C=C, R=R, T=T)\n return out\n\ndef make_S():\n out = \"\"\n for L in (2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Mat_Check({C}, {R}, {T}, arg)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_mat<{C}, {R}, {T}>(arg)));\\\\\\n\\t}}\\\\\\n\".format(C=L, R=L, T=T)\n return out\n\ndef make_Q():\n out = \"\"\n q_types = []\n if \"float\" in types:\n q_types.append(\"float\")\n if \"double\" in types:\n q_types.append(\"double\")\n for T in q_types:\n out += \"\\tif (PyGLM_Qua_Check({T}, arg)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_qua<{T}>(arg)));\\\\\\n\\t}}\\\\\\n\".format(T=T)\n return out\n\ndef make_VV():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg1) && PyGLM_Vec_Check({L}, {T}, arg2)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg1), unpack_vec<{L}, {T}>(arg2)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_MM():\n out = \"\"\n for C in (2, 3, 4):\n for R in (2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Mat_Check({C}, {R}, {T}, arg1) && PyGLM_Mat_Check({C}, {R}, {T}, arg2)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_mat<{C}, {R}, {T}>(arg1), unpack_mat<{C}, {R}, {T}>(arg2)));\\\\\\n\\t}}\\\\\\n\".format(C=C, R=R, T=T)\n return out\n\ndef make_VN():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg1) && PyGLM_Number_Check(arg2)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg1), PyGLM_Number_FromPyObject<{T}>(arg2)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_VVV():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg1) && PyGLM_Vec_Check({L}, {T}, arg2) && PyGLM_Vec_Check({L}, {T}, arg3)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg1), unpack_vec<{L}, {T}>(arg2), unpack_vec<{L}, {T}>(arg3)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_VVN():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg1) && PyGLM_Vec_Check({L}, {T}, arg2) && PyGLM_Number_Check(arg3)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg1), unpack_vec<{L}, {T}>(arg2), PyGLM_Number_FromPyObject<{T}>(arg3)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_QQN():\n out = \"\"\n for T in types:\n out += \"\\tif (PyGLM_Qua_Check({T}, arg1) && PyGLM_Qua_Check({T}, arg2) && PyGLM_Number_Check(arg3)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_qua<{T}>(arg1), unpack_qua<{T}>(arg2), PyGLM_Number_FromPyObject<{T}>(arg3)));\\\\\\n\\t}}\\\\\\n\".format(T=T)\n return out\n\ndef make_VNN():\n out = \"\"\n for L in (1, 2, 3, 4):\n for T in types:\n out += \"\\tif (PyGLM_Vec_Check({L}, {T}, arg1) && PyGLM_Number_Check(arg2) && PyGLM_Number_Check(arg3)) {{\\\\\\n\\t\\treturn pack(glm::NAME(unpack_vec<{L}, {T}>(arg1), PyGLM_Number_FromPyObject<{T}>(arg2), PyGLM_Number_FromPyObject<{T}>(arg3)));\\\\\\n\\t}}\\\\\\n\".format(L=L, T=T)\n return out\n\ndef make_Nn(n, T):\n out = \"\\tif (\"\n if n == 1:\n out += \"PyGLM_Number_Check(arg)\"\n else:\n out += \"PyGLM_Number_Check(arg1)\" + \"\".join([\" && PyGLM_Number_Check(arg{})\".format(i) for i in range(2, n+1)])\n\n out += \") {\\\\\\n\\t\\treturn pack(glm::NAME(\"\n\n if n == 1:\n out += \"PyGLM_Number_FromPyObject<{T}>(arg)\".format(T=T)\n else:\n out += \"PyGLM_Number_FromPyObject<{T}>(arg1)\".format(T=T) + \"\".join([\", PyGLM_Number_FromPyObject<{T}>\".format(T=T) + \"(arg{})\".format(i) for i in range(2, n+1)])\n\n out += \"));\\\\\\n\\t}\\\\\\n\"\n\n return out\n\ndef make_function(arg_name, accepted_types = \"A\"):\n global types\n out = \"\"\n underscore_count = arg_name.count(\"_\")\n\n arg_count = len(arg_name[:arg_name.find(\"_\")]) if underscore_count else len(arg_name)\n \n func_name = arg_name\n if accepted_types != \"A\":\n func_name += \"__t{}\".format(accepted_types)\n## if return_types:\n## func_name += \"__r{}\".format(return_types)\n out += gen_define(func_name)\n\n types = get_types(accepted_types)\n\n argtypes = list(arg_name.split(\"_\"))\n \n if arg_count == 1:\n out += METH_O\n type_count = len(arg_name)\n for i in range(underscore_count+1):\n type_ = argtypes[i]\n## rtype = return_types[(i%len(return_types))-1]\n if type_ == \"N\":\n if \"double\" in types:\n out += make_Nn(1, \"double\")\n elif \"float\" in types:\n out += make_Nn(1, \"float\")\n elif type_ == \"V\":\n out += make_V()\n elif type_ == \"M\":\n out += make_M()\n elif type_ == \"S\":\n out += make_S()\n elif type_ == \"Q\":\n out += make_Q()\n else:\n print(\"{} not defined yet\".format(type_))\n out += gen_end_O()\n else:\n \n if arg_count == 2:\n out += METH_VA + VA_2O\n for i in range(underscore_count+1):\n type_ = argtypes[i]\n if type_ == \"VV\":\n out += make_VV()\n elif type_ == \"MM\":\n out += make_MM()\n elif type_ == \"VN\":\n out += make_VN()\n elif type_ == \"NN\":\n if \"double\" in types:\n out += make_Nn(2, \"double\")\n elif \"float\" in types:\n out += make_Nn(2, \"float\")\n else:\n print(\"{} not defined yet\".format(type_))\n out += gen_end_2O()\n elif arg_count == 3:\n out += METH_VA + VA_3O\n for i in range(underscore_count+1):\n type_ = argtypes[i]\n if type_ == \"NNN\":\n if \"double\" in types:\n out += make_Nn(3, \"double\")\n elif \"float\" in types:\n out += make_Nn(3, \"float\")\n elif type_ == \"VVV\":\n out += make_VVV()\n elif type_ == \"VVN\":\n out += make_VVN()\n elif type_ == \"QQN\":\n out += make_QQN()\n elif type_ == \"VNN\":\n out += make_VNN()\n else:\n print(\"{} not defined yet\".format(type_))\n out += gen_end_nO()\n elif arg_count == 4:\n out += METH_VA + VA_4O\n for i in range(underscore_count + 1):\n type_ = argtypes[i]\n if type_ == \"NNNN\":\n if \"double\" in types:\n out += make_Nn(4, \"double\")\n elif \"float\" in types:\n out += make_Nn(4, \"float\")\n out += gen_end_nO()\n elif arg_count == 6:\n out += METH_VA + VA_6O\n for i in range(underscore_count + 1):\n type_ = argtypes[i]\n if type_ == \"NNNNNN\":\n if \"double\" in types:\n out += make_Nn(6, \"double\")\n elif \"float\" in types:\n out += make_Nn(6, \"float\")\n out += gen_end_nO()\n return out\n \ncopy = pyperclip.copy\n\ncmf = lambda x, *y, **z: copy(make_function(x, *y, **z))\n", "sub_path": "gen_macro_function.py", "file_name": "gen_macro_function.py", "file_ext": "py", "file_size_in_byte": 9875, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "pyperclip.copy", "line_number": 268, "usage_type": "attribute"}]} +{"seq_id": "106704723", "text": "import os\nimport json\nfrom mock import patch\n\nfrom gtmcore.fixtures import ENV_UNIT_TEST_REPO, ENV_UNIT_TEST_BASE, ENV_UNIT_TEST_REV\n\nfrom snapshottest import snapshot\nfrom lmsrvlabbook.tests.fixtures import fixture_working_dir_env_repo_scoped, fixture_working_dir, mock_create_labbooks\nfrom gtmcore.dispatcher import Dispatcher\n\nimport pytest\nfrom gtmcore.files import FileOperations\nfrom gtmcore.fixtures import remote_labbook_repo, mock_config_file, flush_redis_repo_cache\n\nfrom gtmcore.inventory.inventory import InventoryManager\n\n\nclass TestLabBookServiceMutations(object):\n\n def test_create_labbook(self, fixture_working_dir_env_repo_scoped, snapshot):\n \"\"\"Test listing labbooks\"\"\"\n # Mock the configuration class it it returns the same mocked config file\n # Create LabBook\n query = \"\"\"\n mutation myCreateLabbook($name: String!, $desc: String!, $repository: String!, \n $base_id: String!, $revision: Int!) {\n createLabbook(input: {name: $name, description: $desc, \n repository: $repository, \n baseId: $base_id, revision: $revision}) {\n labbook {\n id\n name\n description\n }\n }\n }\n \"\"\"\n variables = {\"name\": \"test-lab-book1\", \"desc\": \"my test description\",\n \"base_id\": ENV_UNIT_TEST_BASE, \"repository\": ENV_UNIT_TEST_REPO,\n \"revision\": ENV_UNIT_TEST_REV}\n snapshot.assert_match(fixture_working_dir_env_repo_scoped[2].execute(query, variable_values=variables))\n\n # Get LabBook you just created\n query = \"\"\"\n {\n labbook(name: \"test-lab-book1\", owner: \"default\") { \n activityRecords {\n edges{\n node{\n message\n type\n show\n importance\n tags\n username\n email\n detailObjects{\n type\n data\n show\n importance\n tags\n }\n } \n } \n }\n }\n }\n \"\"\"\n snapshot.assert_match(fixture_working_dir_env_repo_scoped[2].execute(query))\n\n def test_delete_labbook_with_linked_dataset_exists(self, fixture_working_dir_env_repo_scoped):\n \"\"\"Test deleting a LabBook with a linked dataset, while the dataset still exists (shouldn't clean up)\"\"\"\n def dispatcher_mock(self, function_ref, kwargs, metadata):\n # If you get here, a cleanup job was scheduled, which shouldn't have happened since dataset still there\n assert \"CLEANUP SHOULD NOT HAVE BEEN SCHEDULED\"\n\n im = InventoryManager()\n lb = im.create_labbook(\"default\", \"default\", \"labbook11\", description=\"Cats labbook 11\")\n lb_root_dir = lb.root_dir\n assert os.path.exists(lb_root_dir)\n\n ds = im.create_dataset('default', 'default', \"dataset2\", storage_type=\"gigantum_object_v1\", description=\"test\")\n im.link_dataset_to_labbook(f\"{ds.root_dir}/.git\", \"default\", \"dataset2\", lb, 'default')\n\n delete_query = f\"\"\"\n mutation delete {{\n deleteLabbook(input: {{\n owner: \"default\",\n labbookName: \"labbook11\",\n confirm: true\n }}) {{\n success\n }}\n }}\n \"\"\"\n with patch.object(Dispatcher, 'dispatch_task', dispatcher_mock):\n r = fixture_working_dir_env_repo_scoped[2].execute(delete_query)\n\n assert 'errors' not in r\n assert r['data']['deleteLabbook']['success'] is True\n assert not os.path.exists(lb_root_dir)\n assert os.path.exists(ds.root_dir)\n\n def test_delete_labbook_with_linked_dataset(self, fixture_working_dir_env_repo_scoped):\n \"\"\"Test deleting a LabBook with a linked dataset that has been deleted as well, should clean up\"\"\"\n class JobResponseMock(object):\n def __init__(self, key):\n self.key_str = key\n\n def dispatcher_mock(self, function_ref, kwargs, metadata):\n assert kwargs['logged_in_username'] == 'default'\n assert kwargs['dataset_owner'] == 'default'\n assert kwargs['dataset_name'] == 'dataset22'\n assert \".labmanager/datasets/test-gigantum-com/default/default/dataset22\" in kwargs['cache_location']\n assert metadata['method'] == 'clean_dataset_file_cache'\n\n with open(\"/tmp/mock_reached\", 'wt') as tf:\n tf.write(\"reached\")\n\n return JobResponseMock(\"rq:job:00923477-d46b-479c-ad0c-2dffcfdfb6b10\")\n\n im = InventoryManager()\n lb = im.create_labbook(\"default\", \"default\", \"labbook111\", description=\"Cats labbook 111\")\n lb_root_dir = lb.root_dir\n assert os.path.exists(lb_root_dir)\n assert os.path.exists(\"/tmp/mock_reached\") is False\n\n ds = im.create_dataset('default', 'default', \"dataset22\", storage_type=\"gigantum_object_v1\", description=\"test\")\n ds_root_dir = ds.root_dir\n im.link_dataset_to_labbook(f\"{ds.root_dir}/.git\", \"default\", \"dataset22\", lb, 'default')\n im.delete_dataset('default', 'default', \"dataset22\")\n\n delete_query = f\"\"\"\n mutation delete {{\n deleteLabbook(input: {{\n owner: \"default\",\n labbookName: \"labbook111\",\n confirm: true\n }}) {{\n success\n }}\n }}\n \"\"\"\n try:\n with patch.object(Dispatcher, 'dispatch_task', dispatcher_mock):\n r = fixture_working_dir_env_repo_scoped[2].execute(delete_query)\n\n assert 'errors' not in r\n assert r['data']['deleteLabbook']['success'] is True\n assert not os.path.exists(lb_root_dir)\n assert not os.path.exists(ds_root_dir)\n assert os.path.exists(\"/tmp/mock_reached\") is True\n finally:\n if os.path.exists(\"/tmp/mock_reached\"):\n os.remove(\"/tmp/mock_reached\")\n\n def test_create_labbook_already_exists(self, fixture_working_dir_env_repo_scoped, snapshot):\n \"\"\"Test listing labbooks\"\"\"\n query = \"\"\"\n mutation myCreateLabbook($name: String!, $desc: String!, $repository: String!, \n $base_id: String!, $revision: Int!) {\n createLabbook(input: {name: $name, description: $desc, \n repository: $repository, \n baseId: $base_id, revision: $revision}) {\n labbook {\n id\n name\n description\n }\n }\n }\n \"\"\"\n variables = {\"name\": \"test-lab-duplicate\", \"desc\": \"my test description\",\n \"base_id\": ENV_UNIT_TEST_BASE, \"repository\": ENV_UNIT_TEST_REPO,\n \"revision\": ENV_UNIT_TEST_REV}\n snapshot.assert_match(fixture_working_dir_env_repo_scoped[2].execute(query, variable_values=variables))\n\n # Get LabBook you just created\n check_query = \"\"\"\n {\n labbook(name: \"test-lab-duplicate\", owner: \"default\") { \n name\n description\n }\n }\n \"\"\"\n snapshot.assert_match(fixture_working_dir_env_repo_scoped[2].execute(check_query))\n\n # Second should fail with an error message\n snapshot.assert_match(fixture_working_dir_env_repo_scoped[2].execute(query, variable_values=variables))\n\n def test_delete_labbook(self, fixture_working_dir_env_repo_scoped):\n \"\"\"Test deleting a LabBook off disk. \"\"\"\n lb = InventoryManager().create_labbook(\"default\", \"default\", \"labbook-delete\", description=\"Cats labbook 1\")\n labbook_dir = lb.root_dir\n\n assert os.path.exists(labbook_dir)\n\n delete_query = f\"\"\"\n mutation delete {{\n deleteLabbook(input: {{\n owner: \"default\",\n labbookName: \"labbook-delete\",\n confirm: true\n }}) {{\n success\n }}\n }}\n \"\"\"\n\n r = fixture_working_dir_env_repo_scoped[2].execute(delete_query)\n assert 'errors' not in r\n assert r['data']['deleteLabbook']['success'] is True\n assert not os.path.exists(labbook_dir)\n\n def test_delete_labbook_dry_run(self, fixture_working_dir_env_repo_scoped):\n \"\"\"Test deleting a LabBook off disk. \"\"\"\n lb = InventoryManager().create_labbook(\"default\", \"default\", \"labbook-delete-dry\", description=\"Cats labbook 1\")\n labbook_dir = lb.root_dir\n assert os.path.exists(labbook_dir)\n\n delete_query = f\"\"\"\n mutation delete {{\n deleteLabbook(input: {{\n owner: \"default\",\n labbookName: \"labbook-delete-dry\",\n confirm: false\n }}) {{\n success\n }}\n }}\n \"\"\"\n\n r = fixture_working_dir_env_repo_scoped[2].execute(delete_query)\n assert 'errors' not in r\n assert r['data']['deleteLabbook']['success'] is False\n assert os.path.exists(labbook_dir)\n\n def test_update_labbook_description(self, fixture_working_dir_env_repo_scoped):\n lb = InventoryManager().create_labbook(\"default\", \"default\", \"labbook-description\",\n description=\"Cats labbook 1\")\n labbook_dir = lb.root_dir\n assert os.path.exists(labbook_dir)\n\n desc_md = f\"# Titłe\\n ## \\\"Subtitle\\\"\\n{'æbčdęfghį:*&^&%$%$@!_t ' * 200}. ## Ænother Sübtitle's\\n{'xyz.?/<>č ' * 300}.\\n\"\n description_query = f\"\"\"\n mutation setDesc($content: String!) {{\n setLabbookDescription(input: {{\n owner: \"default\",\n labbookName: \"labbook-description\",\n descriptionContent: $content\n }}) {{\n success\n }}\n }}\n \"\"\"\n variables = {'content': desc_md}\n r = fixture_working_dir_env_repo_scoped[2].execute(description_query, variable_values=variables)\n assert 'errors' not in r\n assert r['data']['setLabbookDescription']['success'] is True\n\n # Get LabBook you just created\n query = \"\"\"\n {\n labbook(name: \"labbook-description\", owner: \"default\") {\n description\n isRepoClean\n }\n }\n \"\"\"\n r = fixture_working_dir_env_repo_scoped[2].execute(query)\n assert 'errors' not in r\n # There's a lot of weird characters getting filtered out, make sure the bulk of the text remains\n assert abs(1.0 * len(r['data']['labbook']['description']) / len(desc_md)) > 0.75\n assert r['data']['labbook']['isRepoClean'] is True\n\n def test_move_file(self, mock_create_labbooks):\n \"\"\"Test moving a file\"\"\"\n labbook_dir = os.path.join(mock_create_labbooks[1], 'default', 'default', 'labbooks', 'labbook1')\n os.makedirs(os.path.join(labbook_dir, 'code', 'subdir'))\n\n query = \"\"\"\n mutation MoveLabbookFile {\n moveLabbookFile(\n input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n srcPath: \"sillyfile\",\n dstPath: \"subdir/sillyfile\"\n }) {\n updatedEdges {\n node {\n id\n key\n isDir\n size\n section\n } \n }\n }\n }\n \"\"\"\n result_1 = mock_create_labbooks[2].execute(query)\n assert 'errors' not in result_1\n nodes = result_1['data']['moveLabbookFile']['updatedEdges']\n assert len(nodes) == 1\n assert nodes[0]['node']['key'] == 'subdir/sillyfile'\n assert nodes[0]['node']['isDir'] is False\n assert nodes[0]['node']['size'] == '7'\n assert nodes[0]['node']['section'] == 'code'\n\n os.makedirs(os.path.join(labbook_dir, 'code', 'subdir2'))\n query = \"\"\"\n mutation MoveLabbookFile {\n moveLabbookFile(\n input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n srcPath: \"subdir\",\n dstPath: \"subdir2\"\n }) {\n updatedEdges {\n node {\n id\n section\n key\n isDir\n size\n } \n }\n }\n }\n \"\"\"\n result_2 = mock_create_labbooks[2].execute(query)\n assert 'errors' not in result_2\n nodes = result_2['data']['moveLabbookFile']['updatedEdges']\n assert len(nodes) == 2\n assert nodes[0]['node']['key'] == 'subdir2/subdir/'\n assert nodes[0]['node']['isDir'] is True\n assert nodes[0]['node']['section'] == 'code'\n assert nodes[1]['node']['key'] == 'subdir2/subdir/sillyfile'\n assert nodes[1]['node']['isDir'] is False\n assert nodes[1]['node']['section'] == 'code'\n\n assert os.path.exists(os.path.join(labbook_dir, 'code', 'subdir2', 'subdir', 'sillyfile')) is True\n\n def test_move_file_many_times(self, mock_create_labbooks):\n \"\"\"Test moving a file around a bunch\"\"\"\n labbook_dir = os.path.join(mock_create_labbooks[1], 'default', 'default', 'labbooks', 'labbook1', 'code')\n os.makedirs(os.path.join(labbook_dir, 'subdir'))\n\n query1 = \"\"\"\n mutation MoveLabbookFile {\n moveLabbookFile(input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n srcPath: \"sillyfile\",\n dstPath: \"subdir/sillyfile\"\n }) {\n updatedEdges {\n node {\n section\n key\n isDir\n size\n }\n }\n }\n }\n \"\"\"\n\n query2 = \"\"\"\n mutation MoveLabbookFile {\n moveLabbookFile(input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n srcPath: \"subdir/sillyfile\",\n dstPath: \"sillyfile\"\n }) {\n updatedEdges {\n node {\n section\n key\n isDir\n size\n }\n }\n }\n }\n \"\"\"\n result1 = mock_create_labbooks[2].execute(query1)\n assert 'errors' not in result1\n assert len(result1['data']['moveLabbookFile']['updatedEdges']) == 1\n assert result1['data']['moveLabbookFile']['updatedEdges'][0]['node']['key'] == 'subdir/sillyfile'\n assert result1['data']['moveLabbookFile']['updatedEdges'][0]['node']['isDir'] == False\n assert os.path.exists(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n assert os.path.isfile(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n\n result2 = mock_create_labbooks[2].execute(query2)\n assert 'errors' not in result2\n assert len(result2['data']['moveLabbookFile']['updatedEdges']) == 1\n assert result2['data']['moveLabbookFile']['updatedEdges'][0]['node']['key'] == 'sillyfile'\n assert result2['data']['moveLabbookFile']['updatedEdges'][0]['node']['isDir'] == False\n assert os.path.exists(os.path.join(labbook_dir, 'sillyfile'))\n assert os.path.isfile(os.path.join(labbook_dir, 'sillyfile'))\n\n result3 = mock_create_labbooks[2].execute(query1)\n assert 'errors' not in result3\n assert len(result3['data']['moveLabbookFile']['updatedEdges']) == 1\n assert result3['data']['moveLabbookFile']['updatedEdges'][0]['node']['key'] == 'subdir/sillyfile'\n assert result3['data']['moveLabbookFile']['updatedEdges'][0]['node']['isDir'] == False\n assert os.path.exists(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n assert os.path.isfile(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n\n result4 = mock_create_labbooks[2].execute(query2)\n assert len(result4['data']['moveLabbookFile']['updatedEdges']) == 1\n assert result4['data']['moveLabbookFile']['updatedEdges'][0]['node']['key'] == 'sillyfile'\n assert result4['data']['moveLabbookFile']['updatedEdges'][0]['node']['isDir'] == False\n assert os.path.exists(os.path.join(labbook_dir, 'sillyfile'))\n assert os.path.isfile(os.path.join(labbook_dir, 'sillyfile'))\n\n result5 = mock_create_labbooks[2].execute(query1)\n assert len(result5['data']['moveLabbookFile']['updatedEdges']) == 1\n assert result5['data']['moveLabbookFile']['updatedEdges'][0]['node']['key'] == 'subdir/sillyfile'\n assert result5['data']['moveLabbookFile']['updatedEdges'][0]['node']['isDir'] == False\n assert os.path.exists(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n assert os.path.isfile(os.path.join(labbook_dir, 'subdir', 'sillyfile'))\n\n def test_delete_file(self, mock_create_labbooks):\n query = \"\"\"\n mutation deleteLabbookFiless {\n deleteLabbookFiles(\n input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n filePaths: [\"sillyfile\"]\n }) {\n success\n }\n }\n \"\"\"\n filepath = os.path.join(mock_create_labbooks[1], 'default', 'default', 'labbooks', 'labbook1',\n 'code', 'sillyfile')\n assert os.path.exists(filepath) is True\n\n res = mock_create_labbooks[2].execute(query)\n assert res['data']['deleteLabbookFiles']['success'] is True\n\n assert os.path.exists(filepath) is False\n\n def test_delete_dir(self, mock_create_labbooks):\n im = InventoryManager()\n lb = im.load_labbook('default', 'default', 'labbook1')\n FileOperations.makedir(lb, 'code/subdir')\n\n test_file = os.path.join(lb.root_dir, 'code', 'subdir', 'test.txt')\n with open(test_file, 'wt') as tf:\n tf.write(\"puppers\")\n\n lb.git.add_all('code/')\n lb.git.commit(\"blah\")\n\n dir_path = os.path.join(lb.root_dir, 'code', 'subdir')\n assert os.path.exists(dir_path) is True\n assert os.path.exists(test_file) is True\n\n # Note, deleting a file should work with and without a trailing / at the end.\n query = \"\"\"\n mutation deleteLabbookFiles {\n deleteLabbookFiles(\n input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"code\",\n filePaths: [\"subdir/\"]\n }) {\n success\n }\n }\n \"\"\"\n res = mock_create_labbooks[2].execute(query)\n print(res)\n assert res['data']['deleteLabbookFiles']['success'] is True\n\n assert os.path.exists(dir_path) is False\n assert os.path.exists(test_file) is False\n assert os.path.exists(os.path.join(lb.root_dir, 'code')) is True\n\n def test_makedir(self, mock_create_labbooks, snapshot):\n query = \"\"\"\n mutation makeLabbookDirectory {\n makeLabbookDirectory(\n input: {\n owner: \"default\",\n labbookName: \"labbook1\",\n section: \"output\",\n directory: \"new_folder\",\n }) {\n newLabbookFileEdge {\n node{\n key\n isDir\n size\n }\n }\n }}\"\"\"\n snapshot.assert_match(mock_create_labbooks[2].execute(query))\n\n def test_write_readme(self, mock_create_labbooks, snapshot):\n flush_redis_repo_cache()\n content = json.dumps('##Overview\\n\\nThis is my readme\\n :df,a//3p49kasdf')\n\n query = f\"\"\"\n mutation writeReadme {{\n writeLabbookReadme(\n input: {{\n owner: \"default\",\n labbookName: \"labbook1\",\n content: {content},\n }}) {{\n updatedLabbook{{\n name\n description\n overview{{\n readme\n }}\n }}\n }}\n }}\n \"\"\"\n snapshot.assert_match(mock_create_labbooks[2].execute(query))\n\n def test_fetch_labbook_edge(self, mock_create_labbooks):\n query = f\"\"\"\n mutation f {{\n fetchLabbookEdge(input:{{\n owner: \"default\",\n labbookName: \"labbook1\"\n }}) {{\n newLabbookEdge {{\n node {{\n name\n owner\n }}\n }}\n }}\n }}\n \"\"\"\n r = mock_create_labbooks[2].execute(query)\n assert 'errors' not in r\n assert r['data']['fetchLabbookEdge']['newLabbookEdge']['node']['owner'] == 'default'\n assert r['data']['fetchLabbookEdge']['newLabbookEdge']['node']['name'] == 'labbook1'", "sub_path": "packages/gtmapi/lmsrvlabbook/tests/test_labbook_mutations.py", "file_name": "test_labbook_mutations.py", "file_ext": "py", "file_size_in_byte": 21646, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_BASE", "line_number": 39, "usage_type": "name"}, {"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_REPO", "line_number": 39, "usage_type": "name"}, {"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_REV", "line_number": 40, "usage_type": "name"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 41, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 41, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 41, "usage_type": "name"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 70, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 70, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 70, "usage_type": "name"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 78, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 81, "usage_type": "call"}, {"api_name": "os.path", "line_number": 81, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 97, "usage_type": "call"}, {"api_name": "gtmcore.dispatcher.Dispatcher", "line_number": 97, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 97, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 98, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 102, "usage_type": "call"}, {"api_name": "os.path", "line_number": 102, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 103, "usage_type": "call"}, {"api_name": "os.path", "line_number": 103, "usage_type": "attribute"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 123, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 126, "usage_type": "call"}, {"api_name": "os.path", "line_number": 126, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 127, "usage_type": "call"}, {"api_name": "os.path", "line_number": 127, "usage_type": "attribute"}, {"api_name": "mock.patch.object", "line_number": 146, "usage_type": "call"}, {"api_name": "gtmcore.dispatcher.Dispatcher", "line_number": 146, "usage_type": "argument"}, {"api_name": "mock.patch", "line_number": 146, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 147, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 151, "usage_type": "call"}, {"api_name": "os.path", "line_number": 151, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 152, "usage_type": "call"}, {"api_name": "os.path", "line_number": 152, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 153, "usage_type": "call"}, {"api_name": "os.path", "line_number": 153, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 155, "usage_type": "call"}, {"api_name": "os.path", "line_number": 155, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 156, "usage_type": "call"}, {"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_BASE", "line_number": 175, "usage_type": "name"}, {"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_REPO", "line_number": 175, "usage_type": "name"}, {"api_name": "gtmcore.fixtures.ENV_UNIT_TEST_REV", "line_number": 176, "usage_type": "name"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 177, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 177, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 177, "usage_type": "name"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 188, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 188, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 188, "usage_type": "name"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 191, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 191, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 191, "usage_type": "name"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 195, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 198, "usage_type": "call"}, {"api_name": "os.path", "line_number": 198, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 212, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 215, "usage_type": "call"}, {"api_name": "os.path", "line_number": 215, "usage_type": "attribute"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 219, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 221, "usage_type": "call"}, {"api_name": "os.path", "line_number": 221, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 235, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 238, "usage_type": "call"}, {"api_name": "os.path", "line_number": 238, "usage_type": "attribute"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 241, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 244, "usage_type": "call"}, {"api_name": "os.path", "line_number": 244, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 259, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.fixture_working_dir_env_repo_scoped", "line_number": 272, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 280, "usage_type": "call"}, {"api_name": "os.path", "line_number": 280, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 280, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 281, "usage_type": "call"}, {"api_name": "os.path", "line_number": 281, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 305, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 314, "usage_type": "call"}, {"api_name": "os.path", "line_number": 314, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 337, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path", "line_number": 348, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 348, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 352, "usage_type": "call"}, {"api_name": "os.path", "line_number": 352, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 352, "usage_type": "name"}, {"api_name": "os.makedirs", "line_number": 353, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 353, "usage_type": "call"}, {"api_name": "os.path", "line_number": 353, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 396, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path", "line_number": 401, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 402, "usage_type": "call"}, {"api_name": "os.path", "line_number": 402, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 402, "usage_type": "call"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 404, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 409, "usage_type": "call"}, {"api_name": "os.path", "line_number": 409, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 409, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 410, "usage_type": "call"}, {"api_name": "os.path", "line_number": 410, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 410, "usage_type": "call"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 412, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path", "line_number": 417, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 417, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 418, "usage_type": "call"}, {"api_name": "os.path", "line_number": 418, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 418, "usage_type": "call"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 420, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path", "line_number": 424, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 424, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 425, "usage_type": "call"}, {"api_name": "os.path", "line_number": 425, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 425, "usage_type": "call"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 427, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 431, "usage_type": "call"}, {"api_name": "os.path", "line_number": 431, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 431, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path", "line_number": 432, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 432, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 448, "usage_type": "call"}, {"api_name": "os.path", "line_number": 448, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 448, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 450, "usage_type": "call"}, {"api_name": "os.path", "line_number": 450, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 452, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 455, "usage_type": "call"}, {"api_name": "os.path", "line_number": 455, "usage_type": "attribute"}, {"api_name": "gtmcore.inventory.inventory.InventoryManager", "line_number": 458, "usage_type": "call"}, {"api_name": "gtmcore.files.FileOperations.makedir", "line_number": 460, "usage_type": "call"}, {"api_name": "gtmcore.files.FileOperations", "line_number": 460, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 462, "usage_type": "call"}, {"api_name": "os.path", "line_number": 462, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 469, "usage_type": "call"}, {"api_name": "os.path", "line_number": 469, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 470, "usage_type": "call"}, {"api_name": "os.path", "line_number": 470, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 471, "usage_type": "call"}, {"api_name": "os.path", "line_number": 471, "usage_type": "attribute"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 487, "usage_type": "name"}, {"api_name": "os.path.exists", "line_number": 491, "usage_type": "call"}, {"api_name": "os.path", "line_number": 491, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 492, "usage_type": "call"}, {"api_name": "os.path", "line_number": 492, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 493, "usage_type": "call"}, {"api_name": "os.path", "line_number": 493, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 493, "usage_type": "call"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 513, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 513, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 513, "usage_type": "name"}, {"api_name": "gtmcore.fixtures.flush_redis_repo_cache", "line_number": 516, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 517, "usage_type": "call"}, {"api_name": "snapshottest.snapshot.assert_match", "line_number": 537, "usage_type": "call"}, {"api_name": "snapshottest.snapshot", "line_number": 537, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 537, "usage_type": "name"}, {"api_name": "lmsrvlabbook.tests.fixtures.mock_create_labbooks", "line_number": 555, "usage_type": "name"}]} +{"seq_id": "42085824", "text": "import numpy as np\r\nfrom mpi4py import MPI\r\nfrom batch_helper import scatter_data, all_reduce_data\r\nfrom model_helper import all_gather_data\r\nfrom integrated_helper import scatter_broadcast\r\nfrom layers import l2_loss, fully_connected_layer, softmax_loss\r\nfrom sklearn import preprocessing\r\n \r\nimport sys\r\n\r\nclass NeuralNetwork:\r\n\r\n def __init__(self, nodes_model = 1, nodes_batch = 1):\r\n \"\"\"\r\n Initialize the NeuralNetwork\r\n\r\n :param nodes_model: int, number of nodes to use for model parallelism \r\n :param nodes_batch: int, number of nodes to use for batch parallelism\r\n \"\"\"\r\n self.layers = []\r\n self.loss = None\r\n self.nodes_model = nodes_model\r\n self.nodes_batch = nodes_batch\r\n pass\r\n\r\n def add_layer(self, layer_type, size_input=0, size_output=0):\r\n \"\"\"\r\n Add a layer to the NeuralNetwork. \r\n \"\"\"\r\n self.layers.append((layer_type, size_input, size_output))\r\n \r\n def add_loss(self, loss_function):\r\n self.loss = loss_function\r\n\r\n def train_model_parallelism(self, x, y, epochs, mini_batch_size, eta, test_data=None):\r\n \"\"\"\r\n TODO: Training procedure for model parallelism\r\n \"\"\"\r\n x_shape = x.shape\r\n y_shape = y.shape\r\n\r\n mini_batch_shapes = [len(x[k:k + mini_batch_size]) for k in range(0, len(x), mini_batch_size)]\r\n #print(mini_batch_shapes, mini_batch_size)\r\n\r\n # mpi init\r\n comm = MPI.COMM_WORLD\r\n rank = comm.Get_rank()\r\n size = comm.Get_size()\r\n \r\n #TODO\r\n # Create the layers themselves\r\n layers, loss = self._init_layers(rank, size)\r\n \r\n start = MPI.Wtime()\r\n epochTimes = []\r\n \r\n for e in range(epochs):\r\n print(\"starting epoch:\",e)\r\n eStart = MPI.Wtime()\r\n training_data = list(zip(list(x), list(y)))\r\n n = len(training_data)\r\n np.random.shuffle(training_data)\r\n mini_batches =[training_data[k:k+mini_batch_size] for k in range(0, n, mini_batch_size)]\r\n \r\n for i in range(len(mini_batch_shapes)):\r\n # Naming convention\r\n # variable_all means that variable is common to all processes\r\n # all_variable referes to a lsit of variables from each layer\r\n x_all = np.array([j[0] for j in mini_batches[i]])\r\n y_all = np.array([j[1] for j in mini_batches[i]])\r\n all_zs_gathered = [x_all]\r\n\r\n # Remember the dimensions for the backward pass\r\n # to handle the case where layer.w is empty in the backward pass\r\n dims = []\r\n for layer in layers:\r\n if layer.w.size != 0:\r\n z_rank = layer.forward(x_all)\r\n else: \r\n z_rank = np.array([]).reshape((0,0))\r\n\r\n #z_reduced = all_reduce_data(z_rank, comm, rank, size)\r\n z_gathered = np.transpose(all_gather_data(np.transpose(z_rank), comm, rank, size))\r\n all_zs_gathered.append(z_gathered)\r\n dims.append(x_all.shape)\r\n x_all = z_gathered\r\n \r\n loss_value, dy = loss.loss(all_zs_gathered[-1], y_all)\r\n \r\n j=-1 \r\n for layer in reversed(layers):\r\n if layer.w.size != 0:\r\n dy_slice = slice(dy, rank , size)\r\n dx_rank, dw_rank, db_rank = layer.backward(dy_slice)\r\n\r\n else:\r\n dx_rank = np.zeros(dims[j])\r\n j = j-1\r\n\r\n\r\n dx_reduced = all_reduce_data([dx_rank], comm, rank, size)[0]\r\n dy = dx_reduced\r\n \r\n if layer.w.size != 0:\r\n layer.apply_gradient(dw_rank, db_rank, eta, mini_batch_shapes[i])\r\n \r\n evaluation = self.evaluate(test_data, layers, loss, \"model\",comm, rank, size)\r\n if rank == 0: \r\n if test_data:\r\n print (\"Epoch {0}/{1} complete - loss: {2}\".format(e+1, epochs, evaluation))\r\n else:\r\n print (\"Epoch {0}/{1} complete - last training loss: {2}\".format(e+1, epochs, loss_value))\r\n eEnd = MPI.Wtime()\r\n epochTimes.append(eEnd - eStart)\r\n \r\n if rank == 0:\r\n end = MPI.Wtime()\r\n print(\"Finished model in\", end - start)\r\n for i in range(len(epochTimes)):\r\n print(\" \", i, \" \", epochTimes[i]) \r\n print()\r\n \r\n \r\n\r\n def train_batch_parallelism(self, x, y, epochs, mini_batch_size, eta, test_data=None):\r\n \"\"\"\r\n Training procedure for batch parallelism\r\n \"\"\"\r\n x_shape = x.shape\r\n y_shape = y.shape\r\n\r\n mini_batch_shapes = [len(x[k:k + mini_batch_size]) for k in range(0, len(x), mini_batch_size)]\r\n \r\n\r\n # mpi init\r\n comm = MPI.COMM_WORLD\r\n rank = comm.Get_rank()\r\n size = comm.Get_size()\r\n \r\n if rank != 0:\r\n del x\r\n del y\r\n \r\n # Create the layers themselves\r\n layers, loss = self._init_layers()\r\n \r\n start = MPI.Wtime()\r\n epochTimes = [] \r\n\r\n time_scatter_total = 0\r\n time_all_reduce_total = 0\r\n \r\n #print(x)\r\n #print(y)\r\n #print()\r\n \r\n # for 1 ... epoch:\r\n for e in range(epochs):\r\n print(\"starting epoch:\",e)\r\n eStart = MPI.Wtime()\r\n if(rank==0):\r\n #print(e, eStart)\r\n training_data = list(zip(list(x), list(y)))\r\n n = len(training_data)\r\n np.random.shuffle(training_data)\r\n mini_batches =[training_data[k:k+mini_batch_size] for k in range(0, n, mini_batch_size)]\r\n \r\n #for mini_batch in mini_batches:\r\n for i in range(len(mini_batch_shapes)):\r\n \r\n all_x = None\r\n all_y = None\r\n if rank == 0:\r\n all_x = np.array([j[0] for j in mini_batches[i]])\r\n all_y = np.array([j[1] for j in mini_batches[i]])\r\n \r\n time_scatter_total_start = MPI.Wtime()\r\n x_rank = scatter_data(all_x, (mini_batch_shapes[i], x_shape[1]) , comm, rank, size)\r\n #print('x',x_rank)\r\n \r\n y_rank = scatter_data(all_y, (mini_batch_shapes[i], y_shape[1]) , comm, rank, size)\r\n #print('y', y_rank)\r\n time_scatter_total += MPI.Wtime() - time_scatter_total_start\r\n #print(\"rank:\", rank, \"xshape\", x_rank.shape, \"yshape\", y_rank.shape)\r\n\r\n # The following if statement solves the problem when there is one single data to scatter on 2 processes. The second process will receive an empty data...\r\n if x_rank.size != 0: \r\n all_zs = [x_rank]\r\n for layer in layers:\r\n z = layer.forward(x_rank)\r\n #print(\" -- \", z.shape, z)\r\n all_zs.append(z)\r\n x_rank = z\r\n \r\n loss_value, dy = loss.loss(all_zs[-1], y_rank)\r\n #print(all_zs[-1])\r\n \r\n #print(all_zs[-1], dy.shape) \r\n dws, dbs = [], []\r\n for layer in reversed(layers):\r\n dx, dw, db = layer.backward(dy)\r\n \r\n dy = dx\r\n dws.append(dw)\r\n dbs.append(db)\r\n \r\n else: \r\n dws, dbs = [], []\r\n for layer in reversed(layers):\r\n dw = np.zeros(layer.w.shape)\r\n db = np.zeros(layer.b.shape)\r\n dws.append(dw)\r\n dbs.append(db)\r\n \r\n time_all_reduce_time_start = MPI.Wtime() \r\n reduced_dws = all_reduce_data(dws, comm, rank, size)\r\n\r\n reduced_dbs = all_reduce_data(dbs, comm, rank, size)\r\n time_all_reduce_total += MPI.Wtime() - time_all_reduce_time_start\r\n \r\n L = len(layers)\r\n\r\n for j in range(L):\r\n layer = layers[L-1-j]\r\n layer.apply_gradient(reduced_dws[j],reduced_dbs[j], eta, mini_batch_shapes[i])\r\n \r\n \"\"\"\r\n if rank == 0:\r\n print(\"weights first layer\", layers[0].w)\r\n \"\"\"\r\n if rank == 0:\r\n eEnd = MPI.Wtime()\r\n #print(\" \", eEnd)\r\n epochTimes.append(eEnd - eStart)\r\n '''\r\n count = 1\r\n print(\"number of layers:\", len(layers))\r\n for l in layers:\r\n print(\" -- \", count, \" --\")\r\n print(rank, l.w)\r\n print(\"++\")\r\n print()\r\n count += 1\r\n '''\r\n #if test_data:\r\n # print (\"Epoch {0}/{1} complete - loss: {2}\".format(e+1, epochs, self.evaluate(test_data, layers, loss, \"batch\")))\r\n #else:\r\n # print (\"Epoch {0}/{1} complete\".format(e+1, epochs))\r\n if rank == 0:\r\n end = MPI.Wtime()\r\n print(\"Total time was:\", end - start)\r\n print(\"Scatter time:\", time_scatter_total)\r\n print(\"All reduce time:\", time_all_reduce_total)\r\n for i in range(len(epochTimes)):\r\n print(\" (\" + str(i) + \")\", epochTimes[i]) \r\n print()\r\n\r\n def train_integrated_parallelism(self, x, y, epochs, mini_batch_size, eta, test_data=None):\r\n \"\"\"\r\n Training procedure for integrated batch and model parallelism\r\n \"\"\"\r\n x_shape = x.shape\r\n y_shape = y.shape \r\n\r\n mini_batch_shapes = [len(x[k:k + mini_batch_size]) for k in range(0, len(x), mini_batch_size)]\r\n\r\n # mpi init\r\n comm = MPI.COMM_WORLD\r\n rank = comm.Get_rank()\r\n size = comm.Get_size()\r\n\r\n # Convention: \"model-wise\" major order\r\n \"\"\"\r\n Example:\r\n nodes_model = 3\r\n nodes_batch = 2 \r\n \r\n -> WORLD\r\n comm_world:\r\n |comm_world|comm_world|comm_world|\r\n |comm_world|comm_world|comm_world|\r\n\r\n ranks: \r\n |0|1|2|\r\n |3|4|5|\r\n\r\n -> BATCH\r\n batch_split:\r\n |0|0|0|\r\n |1|1|1|\r\n \r\n batch_communicators:\r\n |comm_world0|comm_world1|comm_world2|\r\n |comm_world0|comm_world1|comm_world2|\r\n\r\n -> MODEL\r\n batch_split:\r\n |0|1|2|\r\n |0|1|2|\r\n \r\n batch_communicators:\r\n |comm_word0|comm_word0|comm_word0|\r\n |comm_word1|comm_word1|comm_word1|\r\n\r\n \"\"\"\r\n color_batch = rank % self.nodes_model\r\n comm_batch = MPI.Comm.Split(comm, color_batch, rank)\r\n batch_split = rank // self.nodes_model\r\n\r\n color_model = rank // self.nodes_model \r\n comm_model = MPI.Comm.Split(comm, color_model, rank)\r\n model_split = rank % self.nodes_model\r\n\r\n if rank != 0:\r\n del x\r\n del y\r\n\r\n layers, loss = self._init_layers(model_split, self.nodes_model)\r\n \r\n start = MPI.Wtime()\r\n epochTimes = [] \r\n\r\n\r\n for e in range(epochs):\r\n eStart = MPI.Wtime()\r\n if(rank==0):\r\n training_data = list(zip(list(x), list(y)))\r\n n = len(training_data)\r\n np.random.shuffle(training_data)\r\n mini_batches =[training_data[k:k+mini_batch_size] for k in range(0, n, mini_batch_size)]\r\n\r\n for i in range(len(mini_batch_shapes)):\r\n all_x = None\r\n all_y = None\r\n if rank == 0:\r\n all_x = np.array([j[0] for j in mini_batches[i]])\r\n all_y = np.array([j[1] for j in mini_batches[i]])\r\n \r\n x_batch_split = scatter_broadcast(all_x, (mini_batch_shapes[i], x_shape[1]), batch_split, self.nodes_model, self.nodes_batch, comm, rank, size)\r\n y_batch_split = scatter_broadcast(all_y, (mini_batch_shapes[i], y_shape[1]), batch_split, self.nodes_model, self.nodes_batch, comm, rank, size)\r\n\r\n if x_batch_split.size != 0 :\r\n all_zs_batch_split = [x_batch_split]\r\n\r\n dims = []\r\n for layer in layers:\r\n if layer.w.size != 0:\r\n z_rank = layer.forward(x_batch_split)\r\n else:\r\n z_rank = np.array([]).reshape((0,0))\r\n\r\n z_batch_split = np.transpose(all_gather_data(np.transpose(z_rank), comm_model, model_split, self.nodes_model))\r\n all_zs_batch_split.append(z_batch_split)\r\n dims.append(x_batch_split.shape)\r\n x_batch_split = z_batch_split\r\n\r\n loss_value, dy_batch_split = loss.loss(all_zs_batch_split[-1], y_batch_split)\r\n j = -1\r\n for layer in reversed(layers):\r\n if layer.w.size != 0:\r\n dy_rank = slice(dy_batch_split, model_split, self.nodes_model ) \r\n dx_rank, dw_rank, db_rank = layer.backward(dy_rank)\r\n dw_model_split = all_reduce_data([dw_rank], comm_batch, batch_split, self.nodes_batch)[0]\r\n db_model_split = all_reduce_data([db_rank], comm_batch, batch_split, self.nodes_batch)[0]\r\n layer.apply_gradient(dw_model_split, db_model_split, eta, mini_batch_shapes[i])\r\n \r\n else:\r\n dx_rank = np.zeros(dims[j])\r\n\r\n j = j - 1 \r\n\r\n dx_batch_split = all_reduce_data([dx_rank], comm_model, model_split, self.nodes_model)[0]\r\n dy_batch_split = dx_batch_split\r\n\r\n \r\n \r\n \r\n\r\n else:# (x_batch_split.size == 0)\r\n for layer in reversed(layers):\r\n if layer.w.size != 0:\r\n dw_rank = np.zeros(layer.w.shape)\r\n db_rank = np.zeros(layer.b.shape)\r\n dw_model_split = all_reduce_data([dw_rank], comm_batch, batch_split, self.nodes_batch)[0]\r\n db_model_split = all_reduce_data([db_rank], comm_batch, batch_split, self.nodes_batch)[0]\r\n layer.apply_gradient(dw_model_split, db_model_split, eta, mini_batch_shapes[i])\r\n\r\n if rank == 0:\r\n eEnd = MPI.Wtime()\r\n epochTimes.append(eEnd - eStart)\r\n print (\"Epoch {0}/{1} complete, time: {2}\".format(e+1, epochs, epochTimes[-1]))\r\n\r\n if rank == 0:\r\n end = MPI.Wtime()\r\n print(\" Total time:\", end - start)\r\n \r\n \r\n def evaluate(self, test_data, layers, loss, parr_type,comm, rank, size):\r\n test_results = [(self.feedforward(np.array([x_test]), layers, parr_type,comm, rank, size), y_test)\r\n for (x_test, y_test) in test_data]\r\n n_test = len(test_data)\r\n return (1.0/(1.0*n_test) * sum(loss.loss(y_predicted, y_truth)[0] for (y_predicted, y_truth) in test_results))\r\n \r\n def feedforward(self, a, layers, parr_type,comm, rank, size):\r\n \"\"\"Return the output of the network if ``a`` is input.\"\"\"\r\n if parr_type == \"serial\" or parr_type == \"batch\":\r\n for layer in layers:\r\n a = layer.forward(a)\r\n\r\n else:\r\n for layer in layers:\r\n if layer.w.size != 0:\r\n z_rank = layer.forward(a)\r\n else: \r\n z_rank = np.array([]).reshape((0,0))\r\n \r\n a = layer.forward(a)\r\n a = np.transpose(all_gather_data(np.transpose(a), comm, rank, size))\r\n\r\n return a\r\n\r\n\r\n def test(self, x):\r\n \"\"\"\r\n TODO: Test procedure\r\n \"\"\"\r\n pass\r\n\r\n def train(self, x, y):\r\n \"\"\"\r\n TODO: Combined training procedure for model and batch parallelism\r\n \"\"\"\r\n pass\r\n\r\n def train_serial(self, x, y):\r\n layers, loss = self.init_layers() \r\n \r\n \r\n\r\n def _init_layers(self, model_split=1, model_nodes=1):\r\n layers = []\r\n loss = None \r\n l = 0\r\n for layer in self.layers:\r\n if layer[0] == \"fc\":\r\n # Important note here: every layer is initialized on each process.\r\n # The initialization is random so: either we broadcast the weights and biases\r\n # Or we add a seed in layer. For now, we chose the latter.\r\n output_size = layer[2] // model_nodes\r\n if model_split < layer[2] % model_nodes:\r\n output_size +=1\r\n\r\n layers.append(fully_connected_layer(layer[1], output_size, l*model_nodes+model_split))\r\n \r\n else:\r\n print(layer[0], \"is not valid\")\r\n return []\r\n l += 1\r\n if self.loss == \"l2\":\r\n loss = l2_loss()\r\n elif self.loss == \"softmax\":\r\n loss = softmax_loss()\r\n else : \r\n print(\"invalid loss layer\")\r\n return []\r\n return layers, loss\r\n\r\n\r\ndef _load_data(f, delimiter=\",\"):\r\n data = []\r\n count = 0\r\n for l in f:\r\n split = l.split(delimiter)\r\n sub = []\r\n for s in split:\r\n try:\r\n sub.append(float(s))\r\n except Exception as e:\r\n sub.append(\"Error\")\r\n \r\n data.append(sub)\r\n count += 1\r\n f.close()\r\n\r\n return np.array(data).reshape((len(data), len(data[0])))\r\n\r\n\r\ndef slice(dy, model_split, model_nodes):\r\n shape = dy.shape\r\n features_dim = shape[1]\r\n window_size = features_dim // model_nodes #3\r\n start_ind = window_size * model_split #3*2 = 6\r\n start_ind = start_ind + min(features_dim % (model_nodes),model_split)\r\n end_ind = start_ind + window_size\r\n if model_split < features_dim%model_nodes: \r\n end_ind = end_ind + 1\r\n return(dy[:,start_ind:end_ind])\r\n \r\n\r\n\r\n\r\n# rank: curr_rank, size: number of ranks\r\n# assumes size is divisible by size_output\r\ndef _create_mask(model_split, model_nodes, size_input, size_output):\r\n if size == 1:\r\n mask = None\r\n else:\r\n mask_base = np.zeros(size_output)\r\n local_length = size_output / size\r\n start_ind = int(local_length * rank)\r\n end_ind = int(local_length * (rank + 1))\r\n # print(\"length\", local_length, \"start\", start_ind, \"end\", end_ind)\r\n mask_base[start_ind:end_ind] = 1\r\n mask = np.repeat(np.array([mask_base]), size_input, axis=0)\r\n\r\n # print(mask_base)\r\n # print(mask)\r\n return mask\r\n\r\ndef _fetchData(dataset):\r\n x, y = None, None\r\n if dataset.lower() == \"large\":\r\n data = _load_data(open(\"Data/ethylene_methane.csv\", \"r\"), delimiter=\",\")\r\n x = data[:,3:]\r\n y = data[:,1:3]\r\n elif dataset.lower() == \"medium\":\r\n data = _load_data(open(\"Data/airfoil_self_noise.dat\", \"r\"), delimiter=\"\\t\")\r\n x = data[:,1:5]\r\n y = data[:,5].reshape(len(x), 1)\r\n elif dataset.lower() == \"toy\":\r\n x = np.random.randn(100, 2)\r\n y = np.transpose([np.sin(x[:,0])])\r\n elif dataset.lower() == \"basic\":\r\n x = np.array([[1,2,3],[3,4,5],[5,6,7]])\r\n y = np.array([[6],[12],[18]]) \r\n \r\n #scaler = preprocessing.StandardScaler()\r\n #scaler.fit(x)\r\n #x = scaler.transform(x)\r\n \r\n x_train = x[:int(len(x)*.8)]\r\n y_train = y[:int(len(y)*.8)]\r\n \r\n x_test = x[int(len(x)*.8):]\r\n y_test = y[int(len(y)*.8):]\r\n \r\n return x_train, y_train, x_test, y_test\r\n \r\n\r\n# has all of the testing code\r\ndef main():\r\n if len(sys.argv) < 6:\r\n print(\"Input error, needs to be: python neuralnet.py \")\r\n return\r\n\r\n batch_nodes = int(sys.argv[1])\r\n model_nodes = int(sys.argv[2])\r\n typeParallel = sys.argv[3]\r\n dataset = sys.argv[4]\r\n epochs = int(sys.argv[5])\r\n mini_batch_size = int(sys.argv[6])\r\n eta = float(sys.argv[7])\r\n neurons = [int(x) for x in sys.argv[8].split(\",\")]\r\n \r\n rank = MPI.COMM_WORLD.Get_rank()\r\n\r\n if rank == 0:\r\n print(\"configured to run:\")\r\n print(\" batch nodes\", batch_nodes)\r\n print(\" model nodes\", model_nodes)\r\n print(\" type:\", typeParallel)\r\n print(\" dataset:\", dataset)\r\n print(\" epochs:\", epochs)\r\n print(\" mini batch size:\", mini_batch_size)\r\n print(\" eta:\", eta)\r\n print(\" neurons:\", neurons)\r\n \r\n print(\"Starting test...\")\r\n print(\" Fetching data...\")\r\n x_train, y_train, x_test, y_test = _fetchData(dataset)\r\n test_data = list(zip(list(x_test), list(y_test)))\r\n\r\n if rank == 0:\r\n print(\" succesfully fetched the data.\")\r\n print(\" x_train shape:\", x_train.shape)\r\n print(\" y_train shape:\", y_train.shape)\r\n print(\" x_test shape:\", x_test.shape)\r\n print(\" y_test shape:\", y_test.shape)\r\n\r\n print(\" Creating the neural network...\")\r\n nn = NeuralNetwork(nodes_model=model_nodes, nodes_batch=batch_nodes)\r\n \r\n prevSize = x_train.shape[1]\r\n for s in neurons:\r\n nn.add_layer(\"fc\", prevSize, s)\r\n if rank == 0:\r\n print(\" added layer |\", prevSize,\"->\", s)\r\n prevSize = s\r\n nn.add_layer(\"fc\", prevSize, y_train.shape[1]) \r\n \r\n if rank == 0: \r\n print(\" added output layer |\", prevSize,\"->\",y_train.shape[1])\r\n nn.add_loss(\"l2\")\r\n if rank == 0:\r\n print(\" added loss layer\")\r\n print(\" Finished creating network.\")\r\n \r\n print(\" Beginning training the network...\")\r\n if typeParallel.lower() == \"model\": \r\n nn.train_model_parallelism(x_train, y_train, epochs, mini_batch_size,eta, test_data = test_data)\r\n elif typeParallel.lower() == \"batch\":\r\n nn.train_batch_parallelism(x_train, y_train, epochs, mini_batch_size, eta, test_data=test_data)\r\n elif typeParallel.lower() == \"both\":\r\n nn.train_integrated_parallelism(x_train, y_train, epochs, mini_batch_size, eta, test_data=test_data)\r\n else:\r\n print(\"ERROR: need to use either model, batch, or both\")\r\n\r\n if rank == 0:\r\n print(\"Finished test\")\r\n\r\nif __name__==\"__main__\":\r\n main() \r\n\r\n\r\n\r\n\r\n\r\n", "sub_path": "neuralnet.py", "file_name": "neuralnet.py", "file_ext": "py", "file_size_in_byte": 26302, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 46, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 46, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 54, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 54, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 59, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 59, "usage_type": "name"}, {"api_name": "numpy.random.shuffle", "line_number": 62, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 62, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 69, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 70, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 83, "usage_type": "call"}, {"api_name": "model_helper.all_gather_data", "line_number": 83, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 97, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 101, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 113, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 113, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 117, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 117, "usage_type": "name"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 136, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 136, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 147, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 147, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 160, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 160, "usage_type": "name"}, {"api_name": "numpy.random.shuffle", "line_number": 165, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 165, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 174, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 175, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 177, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 177, "usage_type": "name"}, {"api_name": "batch_helper.scatter_data", "line_number": 178, "usage_type": "call"}, {"api_name": "batch_helper.scatter_data", "line_number": 181, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 183, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 183, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 210, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 211, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 215, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 215, "usage_type": "name"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 216, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 218, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 219, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 219, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 232, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 232, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 250, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 250, "usage_type": "name"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 268, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 268, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Comm.Split", "line_number": 307, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Comm", "line_number": 307, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 307, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Comm.Split", "line_number": 311, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Comm", "line_number": 311, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 311, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 320, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 320, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 325, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 325, "usage_type": "name"}, {"api_name": "numpy.random.shuffle", "line_number": 329, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 329, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 336, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 337, "usage_type": "call"}, {"api_name": "integrated_helper.scatter_broadcast", "line_number": 339, "usage_type": "call"}, {"api_name": "integrated_helper.scatter_broadcast", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 350, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 352, "usage_type": "call"}, {"api_name": "model_helper.all_gather_data", "line_number": 352, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 363, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 364, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 368, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 372, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 382, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 383, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 384, "usage_type": "call"}, {"api_name": "batch_helper.all_reduce_data", "line_number": 385, "usage_type": "call"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 389, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 389, "usage_type": "name"}, {"api_name": "mpi4py.MPI.Wtime", "line_number": 394, "usage_type": "call"}, {"api_name": "mpi4py.MPI", "line_number": 394, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 399, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 415, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 418, "usage_type": "call"}, {"api_name": "model_helper.all_gather_data", "line_number": 418, "usage_type": "call"}, {"api_name": "layers.append", "line_number": 453, "usage_type": "call"}, {"api_name": "layers.fully_connected_layer", "line_number": 453, "usage_type": "call"}, {"api_name": "layers.l2_loss", "line_number": 460, "usage_type": "call"}, {"api_name": "layers.softmax_loss", "line_number": 462, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 485, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 508, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 514, "usage_type": "call"}, {"api_name": "numpy.random.randn", "line_number": 531, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 531, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 532, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 534, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 535, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 552, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 556, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 557, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 558, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 559, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 560, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 561, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 562, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 563, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI.COMM_WORLD.Get_rank", "line_number": 565, "usage_type": "call"}, {"api_name": "mpi4py.MPI.COMM_WORLD", "line_number": 565, "usage_type": "attribute"}, {"api_name": "mpi4py.MPI", "line_number": 565, "usage_type": "name"}]} +{"seq_id": "613730115", "text": "import math\nfrom opentrons.types import Point\nfrom opentrons import protocol_api\nimport subprocess\nimport time\nimport numpy as np\nfrom timeit import default_timer as timer\nimport json\nfrom datetime import datetime\nimport csv\n\n\n# metadata\nmetadata = {\n 'protocolName': 'Station B - RNA extraction - Magmax (Viral Pathogen)',\n 'author': 'Aitor Gastaminza & José Luis Villanueva & Alex Gasulla & Manuel Alba & Daniel Peñil & David Martínez',\n 'source': 'HU Marqués de Valdecilla',\n 'apiLevel': '2.6',\n 'description': 'Protocol for RNA extraction'\n}\n\n################################################\n# CHANGE THESE VARIABLES ONLY\n################################################\nNUM_SAMPLES = 96 # Must be multiple of 8\nUSE_300_TIPS = True # Check that TIP_RECYCLING variables have desired values \n\nVOLUME_SAMPLE = 410 # Volume received from station A\nBEADS_VOLUME_PER_SAMPLE = 200\nWASH_1_VOLUME_PER_SAMPLE = 200\nWASH_2_VOLUME_PER_SAMPLE = 200\nELUTION_VOLUME_PER_SAMPLE = 50\nELUTION_FINAL_VOLUME_PER_SAMPLE = 50 # Volume transfered to final plates\n\nBEADS_WELL_FIRST_TIME_NUM_MIXES = 15\nBEADS_WELL_NUM_MIXES = 2\nBEADS_NUM_MIXES = 10 # 20\nWASH_NUM_MIXES = 10 # 10 \nEHTANOL_NUM_MIXES = 10 # 10\nELUTION_NUM_MIXES = 5 # 5\n\nTIP_RECYCLING_IN_WASH = False\nTIP_RECYCLING_IN_ELUTION = False\n\nSET_TEMP_ON = False # Do you want to start temperature module?\nTEMPERATURE = 4 # Set temperature. It will be uesed if set_temp_on is set to True\n\nPHOTOSENSITIVE = False # True if it has photosensitive reagents\nSOUND_NUM_PLAYS = 0\n################################################\n\nrun_id = 'B-Extraccion_total-Bikop'\npath_sounds = '/var/lib/jupyter/notebooks/sonidos/'\n\nrecycle_tip = True # Do you want to recycle tips? It shoud only be set True for testing\nmag_height = 6 # Height needed for NEST deepwell in magnetic deck\nwaste_drop_height = 0\ndeepwell_top_drop_height = 10\nmulti_well_rack_area = 8 * 71 #Cross section of the 12 well reservoir\nnext_well_index = 0 # First reservoir well to use\n\npipette_allowed_capacity = 280 if USE_300_TIPS else 180\ntxt_tip_capacity = '300 uL' if USE_300_TIPS else '200 uL'\n\nx_offset_rs_mv = 2 if USE_300_TIPS else 2.5\nx_offset_rs_sn = 1.1 if USE_300_TIPS else 2\n\nnum_cols = math.ceil(NUM_SAMPLES / 8) # Columns we are working on\nswitch_off_lights = False # Switch of the lights when the program finishes\n\ndef run(ctx: protocol_api.ProtocolContext):\n w1_tip_pos_list = []\n w2_tip_pos_list = []\n elution_tip_pos_list = []\n\n STEP = 0\n STEPS = { #Dictionary with STEP activation, description, and times\n 1:{'Execute': False, 'description': 'Transferir bolas magnéticas'},\n 2:{'Execute': False, 'description': 'Incubación con el imán ON', 'wait_time': 600},\n 3:{'Execute': False, 'description': 'Desechar sobrenadante'},\n 4:{'Execute': False, 'description': 'Imán OFF'},\n 5:{'Execute': False, 'description': 'Transferir primer lavado'},\n 6:{'Execute': False, 'description': 'Incubación con el imán ON', 'wait_time': 300},\n 7:{'Execute': False, 'description': 'Desechar sobrenadante'},\n 8:{'Execute': False, 'description': 'Imán OFF'},\n 9:{'Execute': False, 'description': 'Transferir segundo lavado'},\n 10:{'Execute': False, 'description': 'Incubación con el imán ON', 'wait_time': 300},\n 11:{'Execute': False, 'description': 'Desechar sobrenadante'},\n 12:{'Execute': False, 'description': 'Secado', 'wait_time': 300},\n 13:{'Execute': False, 'description': 'Imán OFF'},\n 14:{'Execute': True, 'description': 'Transferir elución'},\n 15:{'Execute': False, 'description': 'Incubación con el imán ON', 'wait_time': 300},\n 16:{'Execute': True, 'description': 'Transferir elución a la placa'},\n }\n\n #Folder and file_path for log time\n import os\n folder_path = '/var/lib/jupyter/notebooks/' + run_id\n if not ctx.is_simulating():\n if not os.path.isdir(folder_path):\n os.mkdir(folder_path)\n file_path = folder_path + '/Station_B_Extraccion_total_time_log.txt'\n\n #Define Reagents as objects with their properties\n class Reagent:\n def calc_vol_well(self):\n if(self.name == 'Sample'):\n self.num_wells = num_cols\n return VOLUME_SAMPLE\n elif self.placed_in_multi:\n trips = math.ceil(self.reagent_volume / self.max_volume_allowed)\n vol_trip = self.reagent_volume / trips * 8\n max_trips_well = math.floor(18000 / vol_trip)\n total_trips = num_cols * trips\n self.num_wells = math.ceil(total_trips / max_trips_well)\n return math.ceil(total_trips / self.num_wells) * vol_trip + self.dead_vol\n else:\n self.num_wells = 1\n return self.reagent_volume * NUM_SAMPLES\n\n def __init__(self, name, flow_rate_aspirate, flow_rate_dispense, flow_rate_aspirate_mix, flow_rate_dispense_mix,\n air_gap_vol_bottom, air_gap_vol_top, disposal_volume, max_volume_allowed, reagent_volume, v_fondo,\n dead_vol = 700, first_well = None, placed_in_multi = False):\n self.name = name\n self.flow_rate_aspirate = flow_rate_aspirate\n self.flow_rate_dispense = flow_rate_dispense\n self.flow_rate_aspirate_mix = flow_rate_aspirate_mix\n self.flow_rate_dispense_mix = flow_rate_dispense_mix\n self.air_gap_vol_bottom = air_gap_vol_bottom\n self.air_gap_vol_top = air_gap_vol_top\n self.disposal_volume = disposal_volume\n self.max_volume_allowed = max_volume_allowed\n self.reagent_volume = reagent_volume\n self.col = 0\n self.vol_well = 0\n self.v_cono = v_fondo\n self.dead_vol = dead_vol\n self.first_well = first_well\n self.placed_in_multi = placed_in_multi\n self.vol_well_original = self.calc_vol_well() if reagent_volume * NUM_SAMPLES > 0 else 0\n self.vol_well = self.vol_well_original\n\n #Reagents and their characteristics\n Beads = Reagent(name = 'Beads', \n flow_rate_aspirate = 25,\n flow_rate_dispense = 100,\n flow_rate_aspirate_mix = 25,\n flow_rate_dispense_mix = 100,\n air_gap_vol_bottom = 5,\n air_gap_vol_top = 0,\n disposal_volume = 1,\n max_volume_allowed = pipette_allowed_capacity,\n reagent_volume = BEADS_VOLUME_PER_SAMPLE,\n placed_in_multi = True,\n v_fondo = 695) #1.95 * multi_well_rack_area / 2, #Prismatic\n\n Wash_1 = Reagent(name = 'Wash 1',\n flow_rate_aspirate = 25,\n flow_rate_dispense = 100,\n flow_rate_aspirate_mix = 25,\n flow_rate_dispense_mix = 100,\n air_gap_vol_bottom = 5,\n air_gap_vol_top = 0,\n disposal_volume = 1,\n max_volume_allowed = pipette_allowed_capacity,\n reagent_volume = WASH_1_VOLUME_PER_SAMPLE,\n placed_in_multi = True,\n v_fondo = 695) #1.95 * multi_well_rack_area / 2, #Prismatic)\n\n Wash_2 = Reagent(name = 'Wash 2',\n flow_rate_aspirate = 25,\n flow_rate_dispense = 100,\n flow_rate_aspirate_mix = 25,\n flow_rate_dispense_mix = 100,\n air_gap_vol_bottom = 5,\n air_gap_vol_top = 0,\n disposal_volume = 1,\n max_volume_allowed = pipette_allowed_capacity, \n reagent_volume = WASH_2_VOLUME_PER_SAMPLE,\n placed_in_multi = True,\n v_fondo = 695) #1.95 * multi_well_rack_area / 2, #Prismatic)\n\n Elution = Reagent(name = 'Elution',\n flow_rate_aspirate = 25,\n flow_rate_dispense = 100,\n flow_rate_aspirate_mix = 25,\n flow_rate_dispense_mix = 100,\n air_gap_vol_bottom = 5,\n air_gap_vol_top = 0,\n disposal_volume = 1,\n max_volume_allowed = pipette_allowed_capacity,\n reagent_volume = ELUTION_VOLUME_PER_SAMPLE,\n placed_in_multi = True,\n v_fondo = 695) #1.95*multi_well_rack_area/2) #Prismatic\n\n Sample = Reagent(name = 'Sample',\n flow_rate_aspirate = 5, # Original 0.5\n flow_rate_dispense = 100, # Original 1\n flow_rate_aspirate_mix = 1,\n flow_rate_dispense_mix = 1,\n air_gap_vol_bottom = 5,\n air_gap_vol_top = 0,\n disposal_volume = 1,\n max_volume_allowed = pipette_allowed_capacity,\n reagent_volume = VOLUME_SAMPLE,\n v_fondo = 4 * math.pi * 4**3 / 3) #Sphere\n\n ctx.comment(' ')\n ctx.comment('###############################################')\n ctx.comment('VALORES DE VARIABLES')\n ctx.comment(' ')\n ctx.comment('Número de muestras: ' + str(NUM_SAMPLES) + ' (' + str(num_cols) + ' ' + ('columna' if num_cols == 1 else 'columnas') + ')')\n ctx.comment('Capacidad de puntas: ' + txt_tip_capacity)\n ctx.comment(' ')\n ctx.comment('Volumen de muestra en el deepwell: ' + str(VOLUME_SAMPLE) + ' ul')\n ctx.comment('Volumen de beads por muestra: ' + str(BEADS_VOLUME_PER_SAMPLE) + ' ul')\n ctx.comment('Volumen del lavado por muestra: ' + str(WASH_1_VOLUME_PER_SAMPLE) + ' ul')\n ctx.comment('Volumen del etanol por muestra: ' + str(WASH_2_VOLUME_PER_SAMPLE) + ' ul')\n ctx.comment('Volumen de elución por muestra: ' + str(ELUTION_VOLUME_PER_SAMPLE) + ' ul')\n ctx.comment('Volumen de elución a retirar del deepwell: ' + str(ELUTION_FINAL_VOLUME_PER_SAMPLE) + ' ul')\n ctx.comment(' ')\n ctx.comment('Número de mezclas en la primera recogida de un canal con bolas magnéticas: ' + str(BEADS_WELL_FIRST_TIME_NUM_MIXES))\n ctx.comment('Número de mezclas en el resto de recogidas de un canal con bolas magnéticas: ' + str(BEADS_WELL_NUM_MIXES)) \t\n ctx.comment('Número de mezclas con la solución de bolas magnéticas: ' + str(BEADS_NUM_MIXES))\n ctx.comment('Número de mezclas con el lavado: ' + str(WASH_NUM_MIXES))\n ctx.comment('Número de mezclas con el etanol lavado: ' + str(EHTANOL_NUM_MIXES))\n ctx.comment('Número de mezclas con la elución: ' + str(ELUTION_NUM_MIXES))\n ctx.comment(' ')\n ctx.comment('Reciclado de puntas en los lavados activado: ' + str(TIP_RECYCLING_IN_WASH))\n ctx.comment('Reciclado de puntas en la elución activado: ' + str(TIP_RECYCLING_IN_ELUTION))\n ctx.comment(' ')\n ctx.comment('Activar módulo de temperatura: ' + str(SET_TEMP_ON))\n ctx.comment('Valor objetivo módulo de temepratura: ' + str(TEMPERATURE) + ' ºC')\n ctx.comment(' ')\n ctx.comment('Foto-sensible: ' + str(PHOTOSENSITIVE))\n ctx.comment('Repeticiones del sonido final: ' + str(SOUND_NUM_PLAYS))\n ctx.comment(' ')\n\n #########\n def str_rounded(num):\n return str(int(num + 0.5))\n\n ###################\n #Custom functions\n def custom_mix(pipet, reagent, location, vol, rounds, blow_out, mix_height, offset, wait_time = 0, drop_height = -1, two_thirds_mix_bottom = False):\n '''\n Function for mix in the same location a certain number of rounds. Blow out optional. Offset\n can set to 0 or a higher/lower value which indicates the lateral movement\n '''\n if mix_height <= 0:\n mix_height = 1\n\n pipet.aspirate(1, location = location.bottom(z = mix_height), rate = reagent.flow_rate_aspirate_mix)\n\n for i in range(rounds):\n pipet.aspirate(vol, location = location.bottom(z = mix_height), rate = reagent.flow_rate_aspirate_mix)\n if two_thirds_mix_bottom and i < ((rounds / 3) * 2):\n pipet.dispense(vol, location = location.bottom(z = 5).move(Point(x = offset)), rate = reagent.flow_rate_dispense_mix)\n else:\n pipet.dispense(vol, location = location.top(z = drop_height).move(Point(x = offset)), rate = reagent.flow_rate_dispense_mix)\n \n pipet.dispense(1, location = location.bottom(z = mix_height), rate = reagent.flow_rate_dispense_mix)\n \n if blow_out == True:\n pipet.blow_out(location.top(z = -2)) # Blow out\n \n if wait_time != 0:\n ctx.delay(seconds = wait_time, msg = 'Esperando durante ' + str(wait_time) + ' segundos.')\n\n def calc_height(reagent, cross_section_area, aspirate_volume, min_height = 0.4):\n nonlocal ctx\n ctx.comment('¿Volumen útil restante ' + str(reagent.vol_well - reagent.dead_vol) +\n ' uL < volumen necesario ' + str(aspirate_volume - reagent.disposal_volume * 8) + ' uL?')\n if (reagent.vol_well - reagent.dead_vol + 1) < (aspirate_volume - reagent.disposal_volume * 8):\n ctx.comment('Se debe utilizar el siguiente canal')\n ctx.comment('Canal anterior: ' + str(reagent.col))\n # column selector position; intialize to required number\n reagent.col = reagent.col + 1\n ctx.comment('Nuevo canal: ' + str(reagent.col))\n reagent.vol_well = reagent.vol_well_original\n ctx.comment('Nuevo volumen: ' + str(reagent.vol_well) + ' uL')\n height = (reagent.vol_well - aspirate_volume - reagent.v_cono) / cross_section_area\n reagent.vol_well = reagent.vol_well - (aspirate_volume - reagent.disposal_volume * 8)\n ctx.comment('Volumen restante: ' + str(reagent.vol_well) + ' uL')\n if height < min_height:\n height = min_height\n col_change = True\n else:\n height = (reagent.vol_well - aspirate_volume - reagent.v_cono) / cross_section_area\n reagent.vol_well = reagent.vol_well - (aspirate_volume - (reagent.disposal_volume * 8))\n ctx.comment('La altura calculada es ' + str(round(height, 2)) + ' mm')\n if height < min_height:\n height = min_height\n ctx.comment('La altura utilizada es ' + str(round(height, 2)) + ' mm')\n col_change = False\n return height, col_change\n\n def move_vol_multi(pipet, reagent, source, dest, vol, x_offset_source, x_offset_dest, pickup_height,\n blow_out, wait_time = 0, touch_tip = False, touch_tip_v_offset = 0, drop_height = -5,\n aspirate_with_x_scroll = False, dispense_bottom_air_gap_before = False):\n\n # SOURCE\n if dispense_bottom_air_gap_before and reagent.air_gap_vol_bottom:\n pipet.dispense(reagent.air_gap_vol_bottom, source.top(z = -2), rate = reagent.flow_rate_dispense)\n\n if reagent.air_gap_vol_top != 0: #If there is air_gap_vol, switch pipette to slow speed\n pipet.air_gap(reagent.air_gap_vol_top, height = 0) #air gap\n\n if aspirate_with_x_scroll:\n aspirate_with_x_scrolling(pip = pipet, volume = vol, src = source, pickup_height = pickup_height, rate = reagent.flow_rate_aspirate, start_x_offset_src = 0, stop_x_offset_src = x_offset_source)\n else:\n s = source.bottom(pickup_height).move(Point(x = x_offset_source))\n pipet.aspirate(vol, s, rate = reagent.flow_rate_aspirate) # aspirate liquid\n\n if reagent.air_gap_vol_bottom != 0: #If there is air_gap_vol, switch pipette to slow speed\n pipet.air_gap(reagent.air_gap_vol_bottom, height = 0) #air gap\n\n # if wait_time != 0:\n # ctx.delay(seconds=wait_time, msg='Esperando durante ' + str(wait_time) + ' segundos.')\n\n # GO TO DESTINATION\n d = dest.top(z = drop_height).move(Point(x = x_offset_dest))\n pipet.dispense(vol - reagent.disposal_volume + reagent.air_gap_vol_bottom, d, rate = reagent.flow_rate_dispense)\n\n if reagent.air_gap_vol_top != 0:\n pipet.dispense(reagent.air_gap_vol_top, dest.top(z = 0), rate = reagent.flow_rate_dispense)\n\n if blow_out == True:\n pipet.blow_out(dest.top(z = drop_height))\n\n if touch_tip == True:\n pipet.touch_tip(speed = 20, v_offset = touch_tip_v_offset, radius=0.7)\n\n if wait_time != 0:\n ctx.delay(seconds=wait_time, msg='Esperando durante ' + str(wait_time) + ' segundos.')\n\n #if reagent.air_gap_vol_bottom != 0:\n #pipet.move_to(dest.top(z = 0))\n #pipet.air_gap(reagent.air_gap_vol_bottom) #air gap\n #pipet.aspirate(air_gap_vol_bottom, dest.top(z = 0),rate = reagent.flow_rate_aspirate) #air gap\n\n def aspirate_with_x_scrolling(pip, volume, src, pickup_height = 0, rate = 1, start_x_offset_src = 0, stop_x_offset_src = 0):\n\n max_asp = volume/pip.min_volume\n inc_step = (start_x_offset_src - stop_x_offset_src) / max_asp\n\n for x in reversed(np.arange(stop_x_offset_src, start_x_offset_src, inc_step)):\n s = src.bottom(pickup_height).move(Point(x = x))\n pip.aspirate(volume = pip.min_volume, location = s, rate = rate)\n\n ##########\n # pick up tip and if there is none left, prompt user for a new rack\n def pick_up_tip(pip, position = None):\n nonlocal tip_track\n #if not ctx.is_simulating():\n if recycle_tip:\n pip.pick_up_tip(tips300[0].wells()[0])\n else:\n if tip_track['counts'][pip] >= tip_track['maxes'][pip]:\n for i in range(3):\n ctx._hw_manager.hardware.set_lights(rails=False)\n ctx._hw_manager.hardware.set_lights(button=(1, 0 ,0))\n time.sleep(0.3)\n ctx._hw_manager.hardware.set_lights(rails=True)\n ctx._hw_manager.hardware.set_lights(button=(0, 0 ,1))\n time.sleep(0.3)\n ctx._hw_manager.hardware.set_lights(button=(0, 1 ,0))\n ctx.pause('Reemplaza las cajas de puntas de ' + str(pip.max_volume) + 'µl antes de continuar.')\n pip.reset_tipracks()\n tip_track['counts'][pip] = 0\n tip_track['num_refills'][pip] += 1\n if position is None:\n pip.pick_up_tip()\n else:\n pip.pick_up_tip(position)\n\n def drop_tip(pip, recycle = False, increment_count = True):\n nonlocal tip_track\n #if not ctx.is_simulating():\n if recycle or recycle_tip:\n pip.return_tip()\n else:\n pip.drop_tip(home_after = False)\n if increment_count:\n tip_track['counts'][pip] += 8\n\n def start_run():\n ctx.comment(' ')\n ctx.comment('###############################################')\n ctx.comment('Empezando protocolo')\n if PHOTOSENSITIVE == False:\n ctx._hw_manager.hardware.set_lights(button = True, rails = True)\n else:\n ctx._hw_manager.hardware.set_lights(button = True, rails = False)\n now = datetime.now()\n\n # dd/mm/YY H:M:S\n start_time = now.strftime(\"%Y/%m/%d %H:%M:%S\")\n return start_time\n\n def run_quiet_process(command):\n subprocess.check_output('{} &> /dev/null'.format(command), shell=True)\n\n def play_sound(filename):\n print('Speaker')\n print('Next\\t--> CTRL-C')\n try:\n run_quiet_process('mpg123 {}'.format(path_sounds + filename + '.mp3'))\n except KeyboardInterrupt:\n pass\n print()\n\n def finish_run(switch_off_lights = False):\n ctx.comment('###############################################')\n ctx.comment('Protocolo finalizado')\n ctx.comment(' ')\n #Set light color to blue\n ctx._hw_manager.hardware.set_lights(button = True, rails = False)\n now = datetime.now()\n # dd/mm/YY H:M:S\n finish_time = now.strftime(\"%Y/%m/%d %H:%M:%S\")\n if PHOTOSENSITIVE==False:\n for i in range(10):\n ctx._hw_manager.hardware.set_lights(button = False, rails = False)\n time.sleep(0.3)\n ctx._hw_manager.hardware.set_lights(button = True, rails = True)\n time.sleep(0.3)\n else:\n for i in range(10):\n ctx._hw_manager.hardware.set_lights(button = False, rails = False)\n time.sleep(0.3)\n ctx._hw_manager.hardware.set_lights(button = True, rails = False)\n time.sleep(0.3)\n if switch_off_lights:\n ctx._hw_manager.hardware.set_lights(button = True, rails = False)\n\n used_tips = tip_track['num_refills'][m300] * 96 * len(m300.tip_racks) + tip_track['counts'][m300]\n ctx.comment('Puntas de ' + txt_tip_capacity + ' utilizadas: ' + str(used_tips) + ' (' + str(round(used_tips / 96, 2)) + ' caja(s))')\n ctx.comment('###############################################')\n\n if not ctx.is_simulating():\n for i in range(SOUND_NUM_PLAYS):\n if i > 0:\n time.sleep(60)\n play_sound('finalizado')\n\n return finish_time\n\n def log_step_start():\n ctx.comment(' ')\n ctx.comment('###############################################')\n ctx.comment('PASO '+str(STEP)+': '+STEPS[STEP]['description'])\n ctx.comment('###############################################')\n ctx.comment(' ')\n return datetime.now()\n\n def log_step_end(start):\n end = datetime.now()\n time_taken = (end - start)\n STEPS[STEP]['Time:'] = str(time_taken)\n\n ctx.comment(' ')\n ctx.comment('Paso ' + str(STEP) + ': ' +STEPS[STEP]['description'] + ' hizo un tiempo de ' + str(time_taken))\n ctx.comment(' ')\n\n ##########\n def find_side(col):\n if col%2 == 0:\n side = -1 # left\n else:\n side = 1 # right\n return side\n\n\n def assign_wells(reagent, first_well_pos = None):\n global next_well_index\n if first_well_pos is not None and first_well_pos > next_well_index:\n reagent.first_well = first_well_pos\n else:\n reagent.first_well = next_well_index + 1\n\n next_well_index = reagent.first_well - 1 + reagent.num_wells\n reagent.reagent_reservoir = reagent_res.rows()[0][reagent.first_well - 1:next_well_index]\n ctx.comment(reagent.name + ': ' + str(reagent.num_wells) + (' canal' if reagent.num_wells == 1 else ' canales') + ' desde el canal '+ str(reagent.first_well) +' en el reservorio de 12 canales con un volumen de ' + str_rounded(reagent.vol_well_original) + ' uL cada uno')\n\n####################################\n # load labware and modules\n ######## 12 well rack\n reagent_res = ctx.load_labware('usascientific_12_reservoir_22ml', '5','Reagent 12 Well Reservoir')\n\n##################################\n ########## tempdeck\n tempdeck = ctx.load_module('Temperature Module Gen2', '1')\n\n ####### Elution plate - final plate, goes to C\n elution_plate = tempdeck.load_labware('kingfisher_96_aluminumblock_200ul', 'Kingfisher 96 Aluminum Block 200 uL')\n\n############################################\n ######## Deepwell - comes from A\n magdeck = ctx.load_module('Magnetic Module Gen2', '4')\n deepwell_plate = magdeck.load_labware('kingfisher_96_wellplate_2000ul', 'KingFisher 96 Well Plate 2mL')\n\n####################################\n ######## Waste reservoir\n waste_reservoir = ctx.load_labware('nest_1_reservoir_195ml', '7', 'waste reservoir') # Change to our waste reservoir\n waste = waste_reservoir.wells()[0] # referenced as reservoir\n\n####################################\n ######### Load tip_racks\n tips300 = [ctx.load_labware('opentrons_96_tiprack_300ul', slot, txt_tip_capacity + ' filter tiprack')\n for slot in ['2', '3', '6', '8', '9', '10', '11']]\n\n###############################################################################\n #Declare which reagents are in each reservoir as well as deepwell and elution plate\n ctx.comment(' ')\n ctx.comment('###############################################')\n ctx.comment('VOLÚMENES PARA ' + str(NUM_SAMPLES) + ' MUESTRAS')\n ctx.comment(' ')\n\n assign_wells(Beads, 1)\n assign_wells(Wash_1, 5)\n assign_wells(Wash_2, 9)\n assign_wells(Elution, 12)\n\n ctx.comment('###############################################')\n ctx.comment(' ')\n\n work_destinations = deepwell_plate.rows()[0][:Sample.num_wells]\n final_destinations = elution_plate.rows()[0][:Sample.num_wells]\n\n # pipettes.\n m300 = ctx.load_instrument('p300_multi_gen2', 'right', tip_racks = tips300) # Load multi pipette\n\n #### used tip counter and set maximum tips available\n tip_track = {\n 'counts': {m300: 0},\n 'maxes': {m300: 96 * len(m300.tip_racks)}, #96 tips per tiprack * number or tipracks in the layout\n 'num_refills' : {m300 : 0},\n 'tips': { m300: [tip for rack in tips300 for tip in rack.rows()[0]]}\n }\n\n###############################################################################\n start_run()\n magdeck.disengage()\n\n ###############################################################################\n # STEP 1 Transferir bolas magnéticas\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n beads_trips = math.ceil(Beads.reagent_volume / Beads.max_volume_allowed)\n beads_volume = Beads.reagent_volume / beads_trips\n beads_transfer_vol = []\n for i in range(beads_trips):\n beads_transfer_vol.append(beads_volume + Beads.disposal_volume)\n x_offset_source = 0\n x_offset_dest = 0\n\n for i in range(num_cols):\n ctx.comment(\"Column: \" + str(i))\n\n \n pick_up_tip(m300)\n\n for j,transfer_vol in enumerate(beads_transfer_vol):\n [pickup_height, change_col] = calc_height(Beads, multi_well_rack_area, transfer_vol * 8) \n \n if change_col == True or (i == 0 and j == 0): #If we switch column because there is not enough volume left in current reservoir column we mix new column\n ctx.comment('Mezclando nuevo canal del reservorio: ' + str(Beads.col + 1))\n custom_mix(m300, Beads, Beads.reagent_reservoir[Beads.col],\n vol = Beads.max_volume_allowed, rounds = BEADS_WELL_FIRST_TIME_NUM_MIXES, \n blow_out = False, mix_height = 1.5, offset = 0)\n first_mix_done = True\n else:\n ctx.comment('Mezclando canal del reservorio: ' + str(Beads.col + 1))\n mix_height = 1.5 if pickup_height > 1.5 else pickup_height\n custom_mix(m300, Beads, Beads.reagent_reservoir[Beads.col],\n vol = Beads.max_volume_allowed, rounds = BEADS_WELL_NUM_MIXES, \n blow_out = False, mix_height = mix_height, offset = 0)\n\n ctx.comment('Aspirando desde la columna del reservorio: ' + str(Beads.col + 1))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n move_vol_multi(m300, reagent = Beads, source = Beads.reagent_reservoir[Beads.col],\n dest = work_destinations[i], vol = transfer_vol, x_offset_source = x_offset_source, x_offset_dest = x_offset_dest,\n pickup_height = pickup_height, blow_out = True, drop_height = deepwell_top_drop_height)\n \n \n if BEADS_NUM_MIXES > 0:\n ctx.comment(' ')\n ctx.comment('Mezclando muestra ')\n custom_mix(m300, Beads, location = work_destinations[i], vol = Beads.max_volume_allowed,\n rounds = BEADS_NUM_MIXES, blow_out = False, mix_height = 1, offset = 0, wait_time = 2)\n\n m300.air_gap(Beads.air_gap_vol_bottom, height = 0) #air gap\n\n drop_tip(m300)\n\n log_step_end(start)\n ###############################################################################\n # STEP 1 Mezclar en deepwell\n ########\n\n ###############################################################################\n # STEP 2 Incubación con el imán ON\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n magdeck.engage(height = mag_height)\n ctx.delay(seconds = STEPS[STEP]['wait_time'], msg = 'Incubación con el imán ON durante ' + format(STEPS[STEP]['wait_time']) + ' segundos.')\n\n log_step_end(start)\n ####################################################################\n # STEP 2 Incubación con el imán ON\n ########\n\n ###############################################################################\n # STEP 3 Desechar sobrenadante\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n total_supernatant_volume = Sample.reagent_volume + Beads.reagent_volume\n\n supernatant_trips = math.ceil((total_supernatant_volume) / Sample.max_volume_allowed)\n supernatant_volume = Sample.max_volume_allowed # We try to remove an exceeding amount of supernatant to make sure it is empty\n supernatant_transfer_vol = []\n for i in range(supernatant_trips):\n supernatant_transfer_vol.append(supernatant_volume + Sample.disposal_volume)\n\n pickup_height = 0.5 # Original 0.5\n\n for i in range(num_cols):\n x_offset_source = find_side(i) * x_offset_rs_sn\n x_offset_dest = 0\n\n if not m300.hw_pipette['has_tip']:\n pick_up_tip(m300)\n for j, transfer_vol in enumerate(supernatant_transfer_vol):\n ctx.comment('Aspirando de la columna del deepwell: ' + str(i+1))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Beads, source = work_destinations[i], dest = waste, vol = transfer_vol,\n x_offset_source = x_offset_source, x_offset_dest = x_offset_dest, pickup_height = pickup_height,\n wait_time = 2, blow_out = True, drop_height = waste_drop_height,\n dispense_bottom_air_gap_before = not (i == 0 and j == 0))\n\n m300.air_gap(Sample.air_gap_vol_bottom, height = 0)\n\n drop_tip(m300)\n\n log_step_end(start)\n ###############################################################################\n # STEP 3 Desechar sobrenadante\n ########\n\n ###############################################################################\n # STEP 4 Imán OFF\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n # Imán OFF\n magdeck.disengage()\n\n log_step_end(start)\n ###############################################################################\n # STEP 4 Imán OFF\n ########\n\n ###############################################################################\n # STEP 5 Transferir primer lavado\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n wash_trips = math.ceil(Wash_1.reagent_volume / Wash_1.max_volume_allowed)\n wash_volume = Wash_1.reagent_volume / wash_trips #136.66\n wash_transfer_vol = []\n for i in range(wash_trips):\n wash_transfer_vol.append(wash_volume + Wash_1.disposal_volume)\n\n for i in range(num_cols):\n x_offset_source = 0\n x_offset_dest = -1 * find_side(i) * x_offset_rs_mv\n if not m300.hw_pipette['has_tip']:\n pick_up_tip(m300)\n if TIP_RECYCLING_IN_WASH:\n w1_tip_pos_list += [tip_track['tips'][m300][int(tip_track['counts'][m300] / 8)]]\n for transfer_vol in wash_transfer_vol:\n [pickup_height, change_col] = calc_height(Wash_1, multi_well_rack_area, transfer_vol*8)\n ctx.comment('Aspirando desde la columna del reservorio: ' + str(Wash_1.first_well + Wash_1.col))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Wash_1, source = Wash_1.reagent_reservoir[Wash_1.col], dest = work_destinations[i],\n vol = transfer_vol, x_offset_source = x_offset_source, x_offset_dest = x_offset_dest,\n pickup_height = pickup_height, drop_height = deepwell_top_drop_height, blow_out = False)\n\n if WASH_NUM_MIXES > 0:\n custom_mix(m300, Wash_1, location = work_destinations[i], vol = Wash_1.max_volume_allowed, two_thirds_mix_bottom = True,\n rounds = WASH_NUM_MIXES, blow_out = False, mix_height = 1.5, offset = x_offset_dest)\n\n m300.air_gap(Wash_1.air_gap_vol_bottom, height = 0) #air gap\n\n drop_tip(m300, recycle = TIP_RECYCLING_IN_WASH)\n\n log_step_end(start)\n ###############################################################################\n # STEP 5 Transferir primer lavado\n ########\n\n ###############################################################################\n # STEP 6 Incubación con el imán ON\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n magdeck.engage(mag_height)\n ctx.delay(seconds = STEPS[STEP]['wait_time'], msg = 'Incubación con el imán ON durante ' + format(STEPS[STEP]['wait_time']) + ' segundos.')\n\n log_step_end(start)\n ####################################################################\n # STEP 6 Incubación con el imán ON\n ########\n\n ###############################################################################\n # STEP 7 Desechar sobrenadante\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n supernatant_trips = math.ceil(Wash_1.reagent_volume / Wash_1.max_volume_allowed)\n supernatant_volume = Wash_1.max_volume_allowed # We try to remove an exceeding amount of supernatant to make sure it is empty\n supernatant_transfer_vol = []\n for i in range(supernatant_trips):\n supernatant_transfer_vol.append(supernatant_volume + Sample.disposal_volume)\n\n pickup_height = 0.5 # Original 0.5\n\n for i in range(num_cols):\n x_offset_source = find_side(i) * x_offset_rs_sn\n x_offset_dest = 0\n\n if not m300.hw_pipette['has_tip']:\n if TIP_RECYCLING_IN_WASH:\n pick_up_tip(m300, w1_tip_pos_list[i])\n m300.dispense(Wash_1.air_gap_vol_top, work_destinations[i].top(z = 0), rate = Wash_1.flow_rate_dispense)\n else:\n pick_up_tip(m300)\n for j, transfer_vol in enumerate(supernatant_transfer_vol):\n ctx.comment('Aspirando de la columna del deepwell: ' + str(i+1))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Sample, source = work_destinations[i], dest = waste, vol = transfer_vol,\n x_offset_source = x_offset_source, x_offset_dest = x_offset_dest, pickup_height = pickup_height,\n wait_time = 2, blow_out = False, drop_height = waste_drop_height,\n dispense_bottom_air_gap_before = not (i == 0 and j == 0))\n\n m300.air_gap(Sample.air_gap_vol_bottom, height = 0)\n\n drop_tip(m300, increment_count = not TIP_RECYCLING_IN_WASH)\n\n log_step_end(start)\n ###############################################################################\n # STEP 7 Desechar sobrenadante\n ########\n\n ###############################################################################\n # STEP 8 Imán OFF\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n # Imán OFF\n magdeck.disengage()\n\n log_step_end(start)\n ###############################################################################\n # STEP 8 Imán OFF\n ########\n\n ###############################################################################\n # STEP 9 Transferir segundo lavado\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n wash_trips = math.ceil(Wash_2.reagent_volume / Wash_2.max_volume_allowed)\n wash_volume = Wash_2.reagent_volume / wash_trips #136.66\n wash_transfer_vol = []\n for i in range(wash_trips):\n wash_transfer_vol.append(wash_volume + Wash_2.disposal_volume)\n pickup_height = 0.5\n\n for i in range(num_cols):\n x_offset_source = 0\n x_offset_dest = -1 * find_side(i) * x_offset_rs_mv\n if not m300.hw_pipette['has_tip']:\n pick_up_tip(m300)\n if TIP_RECYCLING_IN_WASH:\n w2_tip_pos_list += [tip_track['tips'][m300][int(tip_track['counts'][m300] / 8)]]\n for transfer_vol in wash_transfer_vol:\n [pickup_height, change_col] = calc_height(Wash_2, multi_well_rack_area, transfer_vol*8)\n ctx.comment('Aspirando desde la columna del reservorio: ' + str(Wash_2.first_well + Wash_2.col))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Wash_2, source = Wash_2.reagent_reservoir[Wash_2.col], dest = work_destinations[i],\n vol = transfer_vol, x_offset_source = x_offset_source, x_offset_dest = x_offset_dest,\n pickup_height = pickup_height, drop_height = deepwell_top_drop_height, blow_out = False)\n\n if EHTANOL_NUM_MIXES > 0:\n custom_mix(m300, Wash_2, location = work_destinations[i], vol = Wash_2.max_volume_allowed, two_thirds_mix_bottom = True,\n rounds = EHTANOL_NUM_MIXES, blow_out = False, mix_height = 1.5, offset = x_offset_dest)\n\n m300.air_gap(Wash_2.air_gap_vol_bottom, height = 0) #air gap\n\n drop_tip(m300, recycle = TIP_RECYCLING_IN_WASH)\n\n log_step_end(start)\n ###############################################################################\n # STEP 9 Transferir segundo lavado\n ########\n\n ###############################################################################\n # STEP 10 Incubación con el imán ON\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n magdeck.engage(mag_height)\n ctx.delay(seconds = STEPS[STEP]['wait_time'], msg = 'Incubación con el imán ON durante ' + format(STEPS[STEP]['wait_time']) + ' segundos.')\n\n log_step_end(start)\n ####################################################################\n # STEP 10 Incubación con el imán ON\n ########\n\n ###############################################################################\n # STEP 11 Desechar sobrenadante\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n supernatant_trips = math.ceil(Wash_2.reagent_volume / Wash_2.max_volume_allowed)\n supernatant_volume = Wash_2.max_volume_allowed # We try to remove an exceeding amount of supernatant to make sure it is empty\n supernatant_transfer_vol = []\n for i in range(supernatant_trips):\n supernatant_transfer_vol.append(supernatant_volume + Sample.disposal_volume)\n\n pickup_height = 0.5 # Original 0.5\n\n for i in range(num_cols):\n x_offset_source = find_side(i) * x_offset_rs_sn\n x_offset_dest = 0\n\n if not m300.hw_pipette['has_tip']:\n if TIP_RECYCLING_IN_WASH:\n pick_up_tip(m300, w2_tip_pos_list[i])\n m300.dispense(Wash_2.air_gap_vol_top, work_destinations[i].top(z = 0), rate = Wash_2.flow_rate_dispense)\n else:\n pick_up_tip(m300)\n for j, transfer_vol in enumerate(supernatant_transfer_vol):\n ctx.comment('Aspirando de la columna del deepwell: ' + str(i+1))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Sample, source = work_destinations[i], dest = waste, vol = transfer_vol,\n x_offset_source = x_offset_source, x_offset_dest = x_offset_dest, pickup_height = pickup_height,\n wait_time = 2, blow_out = False, dispense_bottom_air_gap_before = not (i == 0 and j == 0),\n drop_height = waste_drop_height)\n\n m300.air_gap(Sample.air_gap_vol_bottom, height = 0)\n\n drop_tip(m300, increment_count = not TIP_RECYCLING_IN_WASH)\n\n log_step_end(start)\n ###############################################################################\n # STEP 11 Desechar sobrenadante\n ########\n\n ###############################################################################\n # STEP 12 Secado\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n ctx.comment(' ')\n ctx.delay(seconds=STEPS[STEP]['wait_time'], msg='Secado durante ' + format(STEPS[STEP]['wait_time']) + ' segundos.') #\n ctx.comment(' ')\n\n log_step_end(start)\n ###############################################################################\n # STEP 12 Secado\n ########\n\n ###############################################################################\n # STEP 13 Imán OFF\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n # Imán OFF\n magdeck.disengage()\n\n log_step_end(start)\n ###############################################################################\n # STEP 13 Imán OFF\n ########\n\n ###############################################################################\n # STEP 14 Transferir elución\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n elution_trips = math.ceil(Elution.reagent_volume / Elution.max_volume_allowed)\n elution_volume = Elution.reagent_volume / elution_trips\n elution_wash_vol = []\n for i in range(elution_trips):\n elution_wash_vol.append(elution_volume + Sample.disposal_volume)\n\n ########\n # Water or elution buffer\n for i in range(num_cols):\n x_offset_source = 0\n x_offset_dest = -1 * find_side(i) * x_offset_rs_mv # Original 0\n if not m300.hw_pipette['has_tip']:\n pick_up_tip(m300)\n if TIP_RECYCLING_IN_ELUTION:\n elution_tip_pos_list += [tip_track['tips'][m300][int(tip_track['counts'][m300] / 8)]]\n for transfer_vol in elution_wash_vol:\n #Calculate pickup_height based on remaining volume and shape of container\n [pickup_height, change_col] = calc_height(Elution, multi_well_rack_area, transfer_vol*8)\n ctx.comment('Aspirando desde la columna del reservorio: ' + str(Elution.first_well + Elution.col))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm')\n\n move_vol_multi(m300, reagent = Elution, source = Elution.reagent_reservoir[Elution.col], dest = work_destinations[i],\n vol = transfer_vol, x_offset_source = x_offset_source, x_offset_dest = x_offset_dest,\n pickup_height = pickup_height, blow_out = False, drop_height = -35)\n\n if ELUTION_NUM_MIXES > 0:\n ctx.comment(' ')\n ctx.comment('Mezclando muestra con Elution')\n custom_mix(m300, Elution, work_destinations[i], vol = Elution.reagent_volume, rounds = ELUTION_NUM_MIXES,\n blow_out = False, mix_height = 1, offset = x_offset_dest, drop_height = -35)\n\n m300.air_gap(Elution.air_gap_vol_bottom, height = 0) #air gap\n\n drop_tip(m300, recycle = TIP_RECYCLING_IN_ELUTION)\n\n log_step_end(start)\n ###############################################################################\n # STEP 14 Transferir elución\n ########\n\n ###############################################################################\n # STEP 15 Incubación con el imán ON\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n magdeck.engage(mag_height)\n ctx.delay(seconds = STEPS[STEP]['wait_time'], msg = 'Incubación con el imán ON durante ' + format(STEPS[STEP]['wait_time']) + ' segundos.')\n\n log_step_end(start)\n ####################################################################\n # STEP 15 Incubación con el imán ON\n ########\n\n ###############################################################################\n # STEP 16 Transferir elución a la placa\n ########\n STEP += 1\n if STEPS[STEP]['Execute']==True:\n start = log_step_start()\n\n elution_trips = math.ceil(ELUTION_FINAL_VOLUME_PER_SAMPLE / Elution.max_volume_allowed)\n elution_volume = ELUTION_FINAL_VOLUME_PER_SAMPLE / elution_trips\n elution_vol = []\n for i in range(elution_trips):\n elution_vol.append(elution_volume + Elution.disposal_volume)\n\n for i in range(num_cols):\n x_offset_source = find_side(i) * x_offset_rs_sn\n x_offset_dest = 0\n if not m300.hw_pipette['has_tip']:\n if TIP_RECYCLING_IN_ELUTION:\n pick_up_tip(m300, elution_tip_pos_list[i])\n m300.dispense(Elution.air_gap_vol_top, work_destinations[i].top(z = 0), rate = Elution.flow_rate_dispense)\n else:\n pick_up_tip(m300)\n for transfer_vol in elution_vol:\n #Pickup_height is fixed here\n pickup_height = 1\n ctx.comment('Aspirando de la columna del deepwell: ' + str(i+1))\n ctx.comment('La altura de recogida es ' + str(round(pickup_height, 2)) + ' mm' )\n\n move_vol_multi(m300, reagent = Sample, source = work_destinations[i], dest = final_destinations[i],\n vol = transfer_vol, x_offset_source = x_offset_source, x_offset_dest = x_offset_dest,\n pickup_height = pickup_height, blow_out = True, touch_tip = False, drop_height = -1)\n\n m300.air_gap(Sample.air_gap_vol_bottom, height = 0) #air gap\n\n drop_tip(m300, increment_count = not TIP_RECYCLING_IN_ELUTION)\n\n if SET_TEMP_ON == True:\n tempdeck.set_temperature(TEMPERATURE)\n\n log_step_end(start)\n ###############################################################################\n # STEP 16 Transferir elución a la placa\n ########\n\n\n magdeck.disengage()\n ctx.comment(' ')\n ctx.comment('###############################################')\n ctx.comment('Homing robot')\n ctx.comment('###############################################')\n ctx.comment(' ')\n ctx.home()\n###############################################################################\n # Export the time log to a tsv file\n if not ctx.is_simulating():\n with open(file_path, 'w') as f:\n f.write('STEP\\texecution\\tdescription\\twait_time\\texecution_time\\n')\n for key in STEPS.keys():\n row = str(key)\n for key2 in STEPS[key].keys():\n row += '\\t' + format(STEPS[key][key2])\n f.write(row + '\\n')\n f.close()\n\n ############################################################################\n finish_run(switch_off_lights)", "sub_path": "Repository/Station B - 1 y 2 - Extracción total/Protocolos en desarrollo/NEW_B-Extraccion_total_Bikop_300.py", "file_name": "NEW_B-Extraccion_total_Bikop_300.py", "file_ext": "py", "file_size_in_byte": 48850, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "math.ceil", "line_number": 68, "usage_type": "call"}, {"api_name": "opentrons.protocol_api.ProtocolContext", "line_number": 71, "usage_type": "attribute"}, {"api_name": "opentrons.protocol_api", "line_number": 71, "usage_type": "name"}, {"api_name": "os.path.isdir", "line_number": 100, "usage_type": "call"}, {"api_name": "os.path", "line_number": 100, "usage_type": "attribute"}, {"api_name": "os.mkdir", "line_number": 101, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 111, "usage_type": "call"}, {"api_name": "math.floor", "line_number": 113, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 115, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 116, "usage_type": "call"}, {"api_name": "math.pi", "line_number": 206, "usage_type": "attribute"}, {"api_name": "opentrons.types.Point", "line_number": 258, "usage_type": "call"}, {"api_name": "opentrons.types.Point", "line_number": 260, "usage_type": "call"}, {"api_name": "opentrons.types.Point", "line_number": 312, "usage_type": "call"}, {"api_name": "opentrons.types.Point", "line_number": 322, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 347, "usage_type": "call"}, {"api_name": "opentrons.types.Point", "line_number": 348, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 363, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 366, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 395, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 395, "usage_type": "name"}, {"api_name": "subprocess.check_output", "line_number": 402, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 419, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 419, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 425, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 427, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 431, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 433, "usage_type": "call"}, {"api_name": "time.sleep", "line_number": 444, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 455, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 455, "usage_type": "name"}, {"api_name": "datetime.datetime.now", "line_number": 458, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 458, "usage_type": "name"}, {"api_name": "math.ceil", "line_number": 553, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 629, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 683, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 740, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 798, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 856, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 930, "usage_type": "call"}, {"api_name": "math.ceil", "line_number": 992, "usage_type": "call"}]} +{"seq_id": "269139890", "text": "#!/usr/bin/env python\n# external\nimport os\nimport sys\nimport argparse\nimport logging\nimport time\nimport pickle # the same as _pickle\nimport itertools\nimport importlib\nimport shutil\nimport numpy as np\nimport pandas as pd\nimport multiprocessing as mpi\nfrom datetime import datetime\nfrom inspect import getmembers, isclass\nfrom pathlib import Path as path\nfrom functools import partial as func_partial\nfrom tqdm import tqdm\n\nimport paddle as mi\nimport paddle.nn as nn\nfrom paddle.io import Dataset\nimport paddle.nn.functional as F\nfrom sklearn.model_selection import train_test_split\n\n# homebrew\nimport misc\nimport gwio\nimport mol_stru\nimport paddle_nets as MiNets\n\nlogger = logging.getLogger(__name__)\n\ndef parse_args2(*argv):\n \"\"\" argv is in the format of sys.argv[1:]\n Also serves the purpose to get default args \n \"\"\"\n parser = argparse.ArgumentParser(description='''\n Launch Pad for PaddlePaddle''',\n formatter_class=argparse.RawTextHelpFormatter)\n\n paparser = argparse.ArgumentParser(add_help=False)\n # paparser.add_argument('--action', type=str, nargs='+', default=[], metavar='', help=\"the action to take\")\n paparser.add_argument(\"--argv\", type=str, default='-h', help='commandline argv (auto defined)')\n paparser.add_argument('-v', '--verbose', choices=[0,1,2], default=1, type=int, metavar='', help=\"('_')\")\n paparser.add_argument(\"--resume\", action='store_true', help='whether to load model state dicts')\n paparser.add_argument('--load_dir', type=str, default=None, metavar='', help=\"directory for loading model args and states\")\n paparser.add_argument('--save_dir', type=str, default=None, metavar='', help=\"directory for saving model and results\")\n paparser.add_argument('--save_level', type=int, default=2, metavar='', help=\"0: no save, 1: final only, 2: interim\")\n paparser.add_argument('--save_grpby', type=str, nargs='+', default=['epoch', 'batch'], metavar='', help=\"groupby columns, e.g., batch, epoch\")\n paparser.add_argument('--log', type=str, default=path(__file__).stem + f'-{datetime.now().strftime(\"%b%d\")}.log', metavar='', help=\"the log file\")\n\n # data\n paparser.add_argument('--data_args', type=str, default='======= data args =======', metavar='', help=\"======= data args =======\")\n paparser.add_argument('--data_dir', type=str, default='data', metavar='', help=\"data directory\")\n paparser.add_argument('--data_name', type=str, default=None, metavar='', help=\"data file name\")\n paparser.add_argument('--data_suffix', type=str, default='.pkl', metavar='', help=\"data file suffix\")\n paparser.add_argument('--data_size', type=int, default=0, metavar='', help=\"the number of data to use if > 0\")\n paparser.add_argument('--test_size', type=float, default=0.1, metavar='', help=\"for train_test_split()\")\n paparser.add_argument('--split_seed', type=int, default=None, metavar='', help=\"not used yet\")\n\n paparser.add_argument('--input_genre', type=str, default='Seq', metavar='', help=\"not used yet\")\n paparser.add_argument('--input_fmt', type=str, default='NLC', metavar='', help=\"\")\n paparser.add_argument('--seq_length', type=int, nargs='+', default=[0, 512, -1], metavar='', help=\"<0: use max data len, 0: no padding\")\n paparser.add_argument('--residue_fmt', type=str, default='vector', metavar='', help=\"embed/vector or quant/scalar\")\n paparser.add_argument('--residue_nn', type=int, default=0, metavar='', help=\"# of nearest neighbors to use\")\n paparser.add_argument('--residue_dbn', action='store_true', help=\"use residue dbn in data\")\n paparser.add_argument('--residue_attr', action='store_true', help=\"use residue attribute in data\")\n paparser.add_argument('--residue_extra', action='store_true', help=\"use residue extra features in data\")\n\n paparser.add_argument('--label_genre', type=str, default='upp', metavar='', help=\"data type: upp/ct/dist/...\")\n paparser.add_argument('--label_fmt', type=str, default='NL', help=\"not yet used\")\n paparser.add_argument('--label_tone', type=str, default='none', help=\"none/hard/soft\")\n paparser.add_argument('--label_ntype', type=int, default=2, help=\"number of label types\")\n paparser.add_argument('--label_smooth', action='store_true', help=\"whether to smoothen labels\")\n\n # net\n paparser.add_argument('--net_args', type=str, default='======= net args =======', metavar='', help=\"======= net args =======\")\n paparser.add_argument('--net_src_file', type=str, default=path(MiNets.__file__), metavar='', help=\"\")\n paparser.add_argument('--net', type=str, default='lazylinear', metavar='', help=\"the name of the net class\")\n paparser.add_argument('--resnet', action='store_true', help=\"whether to use residual net\")\n # paparser.add_argument('--net_return', type=str, default='upp', metavar='', help=\"\")\n paparser.add_argument('--act_fn', type=str, default='relu', metavar='', help=\"activation: relu/sigmoid/...\")\n paparser.add_argument('--norm_fn', type=str, default='none', metavar='', help=\"normalization: batch/instance/layer/none\")\n paparser.add_argument('--norm_axis', type=int, default=-1, metavar='', help=\"the axis # when norm_fn is axis\")\n paparser.add_argument('--dropout', type=float, default=0.2, metavar='', help=\"the dropout fraction\")\n\n paparser.add_argument('--feature_dim', type=int, default=1, metavar='', help=\"the size of in_features\")\n paparser.add_argument('--embed_dim', type=int, default=32, metavar='', help=\"specific for embedding\")\n paparser.add_argument('--embed_num', type=int, default=1, metavar='', help=\"0: no, 1: yes, if residue_fmt is scalar/quant\")\n\n paparser.add_argument('--linear_num', type=int, default=2, metavar='', help=\"# of linear layers\")\n paparser.add_argument('--linear_dim', type=int, nargs='+', default=[32], metavar='', help=\"dims of linear layers\")\n paparser.add_argument('--linear_resnet', action='store_true', help=\"whether to use residual net\")\n\n paparser.add_argument('--conv1d_num', type=int, default=1, metavar='', help=\"# of Conv1D layers\")\n paparser.add_argument('--conv1d_dim', type=int, nargs='+', default=[32], metavar='', help=\"channels of Conv1D layers\")\n paparser.add_argument('--conv1d_resnet', action='store_true', help=\"whether to use residual net\")\n paparser.add_argument('--conv1d_stride', type=int, default=1, metavar='', help=\"stride in 1D convolution\")\n\n paparser.add_argument('--conv2d_num', type=int, default=1, metavar='', help=\"# of Conv2D layers\")\n paparser.add_argument('--conv2d_dim', type=int, nargs='+', default=[32], metavar='', help=\"channels of Conv2D layers \")\n paparser.add_argument('--conv2d_resnet', action='store_true', help=\"whether to use residual net\")\n\n paparser.add_argument('--attn_num', type=int, default=2, metavar='', help=\"# of Attention Encoder layers\")\n paparser.add_argument('--attn_nhead', type=int, default=2, metavar='', help=\"# of heads of Attention Encoders\")\n paparser.add_argument('--attn_act', type=str, default='relu', metavar='', help=\"activation of Attention Encoders\")\n paparser.add_argument('--attn_dropout', type=float, default=None, metavar='', help=\"dropout of Attention Encoders\")\n paparser.add_argument('--attn_ffdim', type=int, default=32, metavar='', help=\"feedforward dims of Attention Encoders\")\n paparser.add_argument('--attn_ffdropout', type=float, default=None, metavar='', help=\"feedforward dropout of Attention Encoders\")\n\n paparser.add_argument('--lstm_num', type=int, default=2, metavar='', help=\"# of LSTM layers\")\n paparser.add_argument('--lstm_dim', type=int, nargs='+', default=[32], metavar='', help=\"dims of LSTM layers\")\n paparser.add_argument('--lstm_direct', type=int, default=2, metavar='', help=\"direction of LSTM layers\")\n paparser.add_argument('--lstm_resnet', action='store_true', help=\"whether to use residual net\")\n\n # paparser.add_argument('--output_net', type=str, default='linear', metavar='', help=\"not used yet\")\n paparser.add_argument('--output_num', type=int, default=1, metavar='', help=\"# of output layers\")\n paparser.add_argument('--output_dim', type=int, nargs='+', default=[32,32,2], metavar='', help=\"output hidden dimensions\")\n paparser.add_argument('--output_resnet', action='store_true', help=\"whether to use residual net (should never set it!)\")\n\n ## optimization\n paparser.add_argument('--optim_args', type=str, default='======= optim args =======', metavar='', help=\"======= optim args =======\")\n paparser.add_argument('--optim', type=str, default='adam', metavar='', help=\"optimizer type: adam/sgd/...\")\n paparser.add_argument('--learning_rate', type=float, default=0.003, metavar='', help=\"change to 3e-4???\")\n paparser.add_argument('--beta1', type=float, default=0.9, metavar='', help=\"beta1 for Adam\")\n paparser.add_argument('--beta2', type=float, default=0.999, metavar='', help=\"beta2 for Adam\")\n paparser.add_argument('--epsilon', type=float, default=1e-8, metavar='', help=\"epsilon for Adam\")\n\n paparser.add_argument('--lr_scheduler', type=str, default='reduced', metavar='', help=\"learning rate scheduler\")\n paparser.add_argument('--lr_factor', type=float, default=0.9, metavar='', help=\"learning rate relative change factor\")\n paparser.add_argument('--lr_patience', type=int, default=10, metavar='', help=\"learning rate patience\")\n\n paparser.add_argument('--weight_decay', type=str, default='none', metavar='', help=\"weight decay: l1 or l2\")\n paparser.add_argument('--l1decay', type=float, default=1e-4, metavar='', help=\"L1Decay rate\")\n paparser.add_argument('--l2decay', type=float, default=1e-4, metavar='', help=\"L2Decay rate\")\n\n paparser.add_argument('--train_args', type=str, default='======= train/loss args =======', metavar='', help=\"======= train/loss args =======\")\n paparser.add_argument('--batch_size', type=int, default=4, metavar='', help=\"for train/validate/predict\")\n # paparser.add_argument('--dynamic_loader', action='store_true', help=\"turn on dynamic loading\")\n paparser.add_argument('--num_epochs', type=int, default=777, metavar='', help=\"# of maximum epochs (may be terminated early)\")\n paparser.add_argument('--num_recaps_per_epoch', type=int, default=30, metavar='', help=\"# of recaps/summaries per epoch\")\n paparser.add_argument('--num_callbacks_per_epoch', type=int, default=10, metavar='', help=\"# of validation callbacks per epoch\")\n\n paparser.add_argument('--loss_fn', type=str, nargs='+', default=['mse'], metavar='', help=\"loss function type: mse/bce/...\")\n paparser.add_argument('--loss_fn_scale', type=float, nargs='+', default=[1.0], metavar='', help=\"\")\n paparser.add_argument('--loss_sqrt', action='store_true', help=\"take sqrt before summing losses in a batch\")\n paparser.add_argument('--loss_padding', action='store_true', help=\"include padded seqs/zeros in the loss\")\n\n paparser.add_argument('--validate_callback', type=str, default=None, metavar='', help=\"not used yet\")\n paparser.add_argument('--trainloss_rdiff', type=float, default=1e-3, metavar='', help=\"relative difference\")\n paparser.add_argument('--validloss_rdiff', type=float, default=1e-3, metavar='', help=\"relative difference\")\n paparser.add_argument('--trainloss_patience', type=int, default=11, metavar='', help=\"train loss change patience\")\n paparser.add_argument('--validloss_patience', type=int, default=11, metavar='', help=\"valid loss change patience\")\n\n # pre-defined settings\n paparser.add_argument('--mood_args', type=str, default='======= mood args =======', metavar='', help=\"======= mood args =======\")\n paparser.add_argument('--debug', action='store_true', help=\"minimal hidden units\")\n paparser.add_argument('--lucky', action='store_true', help=\"[256, 256]\")\n paparser.add_argument('--lazy', action='store_true', help=\"[64]*3\")\n paparser.add_argument('--sharp', action='store_true', help=\"[64]*3\")\n paparser.add_argument('--comfort', action='store_true', help=\"[128]*3]\")\n paparser.add_argument('--explore', action='store_true', help=\"[512]\")\n paparser.add_argument('--exploit', action='store_true', help=\"[32]*5]\")\n paparser.add_argument('--diehard', action='store_true', help=\"[512]*7\")\n paparser.add_argument('--tune', action='store_true', help=\"('_')\")\n\n paparser.add_argument('--action_args', type=str, default='======= action args =======', metavar='', help=\"======= action args =======\")\n\n # action as subparsers\n subparsers = parser.add_subparsers(dest='action', required=True)\n\n subparser = subparsers.add_parser('summary', parents=[paparser], description='', help='only view net/loss')\n subparser = subparsers.add_parser('summarize', parents=[paparser], description='', help='alias for summary')\n subparser = subparsers.add_parser('view', parents=[paparser], description='', help='alias for summary')\n\n subparser = subparsers.add_parser('train', parents=[paparser], description='', help='just do it')\n subparser.set_defaults() # it will be overwritten by later set_defaults!\n\n subparser = subparsers.add_parser('dynamic_train', parents=[paparser], description='', help='not implemented')\n subparser.set_defaults() # it will be overwritten by later set_defaults!\n subparser.add_argument(\"--seq_lengths\", nargs='+', type=int, default=[300, 800, 2000, 5000], metavar='', help='a list of integers')\n subparser.add_argument(\"--batch_sizes\", nargs='+', type=int, default=[8, 4, 2, 1], metavar='', help='a list of integers')\n\n subparser = subparsers.add_parser('cross_validate', parents=[paparser], description='', help='cross validate the model')\n subparser.set_defaults()\n subparser.add_argument(\"--num_cvs\", type=int, default=5, metavar='', help='# of cross validations')\n\n subparser = subparsers.add_parser('validate', parents=[paparser], description='', help='predict and calculate loss')\n subparser.set_defaults()\n\n subparser = subparsers.add_parser('predict', parents=[paparser], description='', help='predict and save')\n subparser.set_defaults(data_name='predict')\n\n subparser = subparsers.add_parser('average_model', parents=[paparser], description='average multiple models', help='')\n subparser.set_defaults()\n subparser.add_argument(\"--model_dirs\", type=str, nargs='+', default=[], metavar='', help='directories of the models to average')\n subparser.add_argument(\"--best_earlystop\", action='store_true', help='use the best earlystop states in each directory')\n\n subparser = subparsers.add_parser('scan_data', parents=[paparser], description='', help='scan data_size and batch_size')\n subparser.set_defaults()\n subparser.add_argument(\"--data_sizes\", nargs='+', type=int, default=[0], metavar='', help='a list of integers')\n subparser.add_argument(\"--batch_sizes\", nargs='+', type=int, default=[1,2,4,8], metavar='', help='a list of integers')\n\n subparser = subparsers.add_parser('scout_args', parents=[paparser], description='', help='scout model args')\n subparser.set_defaults()\n subparser.add_argument(\"--rebake_midata\", action='store_true', help='whether new midata need to be obtained for each iter')\n subparser.add_argument(\"--grid_search\", action='store_true', help='perform grid search')\n subparser.add_argument(\"--spawn_search\", action='store_true', help='perform spawn search')\n\n subparser.add_argument(\"--num_scouts\", type=int, default=7, metavar='', help='only need for grid_search=False')\n subparser.add_argument(\"--num_spawns\", type=int, default=7, metavar='', help='only need for spawn_search=True')\n\n subparser.add_argument(\"--arg_names\", nargs='+', type=str, default=['learning_rate', 'dropout'], metavar='', help='a list of strings')\n subparser.add_argument(\"--arg_values\", nargs='+', type=str, default=['0.0001,0.001,0.01', '0.1,0.3,0.5'],\n metavar='', help='a list of STRINGs with values separated by \",\" for each arg\\n' +\n 'if grid_search is true, each string contains all values for the arg\\n' +\n 'if not grid_search, the string format is \"min,max\"')\n subparser.add_argument(\"--arg_scales\", nargs='+', type=int, default=[0], metavar='',\n help='the scale for each arg_names, 0 for linear, log otherwise')\n\n # args are managed in loosely defined three tiers\n # 1) the default values in parse_args([]), which runs without any user args\n # 2) the loaded args from args.load_dir (if applicable)\n # 3) the user_args from command line, parsed by and extracted from parse_args()\n # the ruling order is the 3 overwrites 2 overwrites 1, which is why user_args needs to be returned!\n #\n # Note: ONLY ONE set of args is maintained for each model\n if isinstance(argv, str): argv = [argv]\n argv = misc.unpack_list_tuple(argv)\n args = misc.Struct(vars(parser.parse_args(argv)))\n argv_dict = vars(misc.argv_optargs(argv, args)) # remember command line args\n\n return args, argv_dict\n\n\ndef autoconfig_args(args):\n \"\"\" set default arg values not easily done with argparser \"\"\"\n ####### DATA #######\n # directories should be path\n if args.save_dir: args.save_dir = path(args.save_dir)\n if args.data_dir: args.data_dir = path(args.data_dir)\n\n if args.data_name is None:\n if args.action in ['validate']:\n args.data_name = 'valid'\n elif args.action in ['predict', 'average_model']:\n args.data_name = 'predict'\n else:\n args.data_name = 'train'\n\n # net determines residue_fmt\n # if 'embed' in args.net.lower():\n # args.residue_fmt = 'quant'\n # else:\n # args.residue_fmt = 'vector'\n\n # residue_fmt determine feature_dim\n args.residue_fmt = args.residue_fmt.lower()\n if args.residue_fmt in ['vector', 'embed']:\n args.feature_dim = 4 * (1 + 2 * args.residue_nn)\n if args.residue_dbn: args.feature_dim += 4 * (1 + 2 * args.residue_nn)\n if args.residue_attr: args.feature_dim += 8\n if args.residue_extra: args.feature_dim += 2\n elif args.residue_fmt in ['quant', 'scalar']:\n args.feature_dim = 5 ** (1 + 2 * args.residue_nn) # the number of possible values\n if args.residue_dbn:\n args.feature_dim *= 4 ** (1 + 2 * args.residue_nn)\n\n # >1 batch_size requires the same sequence length\n if args.seq_length[-1] == 0 and args.batch_size > 1:\n args.seq_length[-1] = -1\n\n args.label_genre = args.label_genre.lower()\n if args.label_genre in ['upp']:\n pass\n elif args.label_genre in ['ct']:\n args.residue_dbn = False\n\n # set smaller numbers for debug\n if args.debug:\n args.embed_dim = min([8, args.embed_dim])\n args.linear_dim = [min([16, i]) for i in args.linear_dim]\n args.linear_num = min([2, args.linear_num])\n args.conv1d_dim = [min([8, i]) for i in args.conv1d_dim]\n args.conv1d_num = min([1, args.conv1d_num])\n args.conv2d_dim = [min([8, i]) for i in args.conv2d_dim]\n args.conv2d_num = min([1, args.conv2d_num])\n args.lstm_num = min([1, args.lstm_num])\n args.lstm_dim = [min([8, i]) for i in args.lstm_dim]\n args.attn_num = min([2, args.attn_num])\n args.attn_ffdim = min([16, args.attn_num])\n args.output_dim = [min([8, i]) for i in args.output_dim]\n args.output_num = 1\n args.batch_size = min([2, args.batch_size])\n args.num_epochs = min([3, args.num_epochs])\n\n if args.lucky:\n args.linear_dim = [256, 256]\n args.linear_num = 1\n\n if args.sharp:\n args.linear_dim = [32]\n args.linear_num = 5\n\n if args.explore:\n args.dropout = 0.5\n args.learning_rate = 1e-3\n args.l1decay = 0\n args.l2decay = 0\n\n if args.tune:\n args.learning_rate = 5e-5\n args.l1decay = 0\n args.l2decay = 1e-4\n if args.exploit:\n pass\n\n return args\n\n\ndef random_sample(midata, size=1, replace=False):\n \"\"\" midata can be a list/tuple/np.ndarray, replace=True will yield repeated elements \"\"\"\n return [midata[i] for i in np.random.choice(len(midata), size, replace=replace)]\n\n\ndef random_split_dict(dict_in, size=0.1):\n \"\"\" split each key values like train_test_split \"\"\"\n num_data = len(dict_in[list(dict_in.keys())[0]])\n\n # size can be a number or a fraction\n if 0.0 < size < 1.0:\n size = int(num_data * size)\n else:\n size = int(size)\n\n # make sure size < num_data / 2\n if size > num_data / 2:\n size = num_data - size\n reverse_order = True\n else:\n reverse_order = False\n\n indices = np.sort(np.random.choice(num_data, size, replace=False))\n\n dict_out1, dict_out2 = dict(), dict()\n\n for key, val in dict_in.items():\n dict_out1[key] = []\n dict_out2[key] = []\n\n dict_out2[key].extend(val[:indices[0]])\n\n for i in range(size - 1):\n dict_out1[key].append(val[indices[i]])\n dict_out2[key].extend(val[indices[i] + 1:indices[i + 1]])\n\n dict_out1[key].append(val[indices[-1]])\n dict_out2[key].extend(val[indices[-1] + 1:])\n\n if reverse_order:\n return dict_out2, dict_out1\n else:\n return dict_out1, dict_out2\n\n\ndef fix_length1d(data, length, **kwargs):\n \"\"\" np.pad is used for padding, the same kwargs \"\"\"\n data_len = len(data)\n\n if data_len >= length:\n return data[:length]\n else:\n return np.pad(data, (0, length - data_len), 'constant', **kwargs)\n\n\ndef fix_length2d(data, length, **kwargs):\n \"\"\" np.pad is used for padding, the same kwargs \"\"\"\n data_len = data.shape\n\n if isinstance(length, int) or isinstance(length, np.integer):\n length = [length]\n\n len2pad = [0, 0]\n\n if data_len[0] >= length[0]: # check 1st dimension\n data = data[:length[0], :]\n else:\n len2pad[0] = length[0] - data_len[0]\n\n if data_len[1] >= length[-1]: # check 2nd dimension\n data = data[:, :length[-1]]\n else:\n len2pad[1] = length[-1] - data_len[1]\n\n if any(len2pad): # pad if needed\n return np.pad(data, ((0, len2pad[0]), (0, len2pad[1])), 'constant', **kwargs)\n else:\n return data\n\ndef cut_padding(data, seq_len):\n \"\"\" assume the first dimension is batch_size and all data in the batch has the same seq_en\"\"\"\n if data.ndim == 1:\n data = data[:seq_len]\n elif data.ndim == 2:\n data = data[:, :seq_len]\n elif data.ndim == 3:\n data = data[:, :seq_len, :seq_len]\n elif data.ndim == 4:\n data = data[:, :seq_len, :seq_len, :seq_len]\n elif data.ndim == 5:\n data = data[:, :seq_len, :seq_len, :seq_len, :seq_len]\n \n return data\n\ndef soft2hard_label(data, keep_dim=False, np=np):\n \"\"\" convert soft to hard labels \"\"\"\n hard_data = data.argmax(axis=-1)\n if keep_dim:\n return np.expand_dims(hard_data, -1)\n else:\n return hard_data\n\n\ndef hard2soft_label(data, nlabel=2, discrete=False, np=np):\n \"\"\" true label starts from zero\n accept non-integers for hard labels, in which weights are assigned to\n two neighboring classes depending on the distance\n \"\"\"\n \n soft_data = np.zeros(list(data.shape) + [nlabel], dtype='float32')\n \n for i in range(nlabel): # there must be a better way...\n soft_data[..., i] = np.clip(np.abs(data - i), 0.0, 1.0)\n\n soft_data = 1.0 - soft_data\n\n return soft_data\n\n\ndef load_pkldata(args=misc.Struct(), **kwargs):\n \"\"\" kwargs > args > my_args, return pkldata, a dict \"\"\"\n def_args = misc.Struct(dict(\n data_dir = '',\n data_name = 'train',\n data_suffix = '.pkl',\n verbose = 1,\n ))\n def_args.update(vars(args))\n def_args.update(kwargs)\n args.update(vars(def_args))\n\n # data_dir must be path\n data_dir = args.data_dir if isinstance(args.data_dir, path) \\\n else path(args.data_dir)\n\n # fname must be string\n fname = args.data_name.name if isinstance(args.data_name, path) \\\n else args.data_name\n\n # read the pkl file\n pkldata_file = (data_dir / fname).with_suffix(args.data_suffix)\n logger.info(f'Loading data: {pkldata_file}')\n with pkldata_file.open('rb') as hfile:\n pkldata = pickle.load(hfile) # it is a dictionary of id, seq, dbn, ct... when available\n\n # logger.info(f' # of data: {len(pkldata[\"seq\"])}, min len: {pkldata[\"len\"].min()}, max len: {pkldata[\"len\"].max()}')\n\n return pkldata\n\n\ndef bake_midata(in_pkldata, args=misc.Struct(), **kwargs):\n \"\"\" kwargs > args > my_args\n return midata, a list of dataset for each sample \"\"\"\n def_args = misc.Struct(dict(\n label_genre = 'upp',\n label_tone = 'none',\n label_ntype = 2,\n seq_length = [-1],\n residue_fmt = 'embed', # \"embed\" a residue as a vector\n residue_nn = 0, # do not include nearest neighbor\n residue_dbn = False,\n residue_attr = False,\n residue_extra = False,\n verbose = 1,\n ))\n def_args.update(vars(args))\n def_args.update(kwargs)\n args.update(vars(def_args))\n\n # collect basic info about data\n num_seqs = len(in_pkldata['seq'])\n in_pkldata['len'] = np.array(in_pkldata['len'])\n args.max_seqlen = in_pkldata['len'].max()\n\n logger.info(f' # of data: {num_seqs}, max seqlen: {args.max_seqlen}, user seq_length: {args.seq_length}')\n logger.info(f' residue fmt: {args.residue_fmt}, nn: {args.residue_nn}, dbn: {args.residue_dbn},' + \\\n f' attr: {args.residue_attr}, genre: {args.label_genre}')\n\n # deal with sequence length\n # 1) seq_length as [min, max, pad_flag] or [max, pad_flag] (min would be zero)\n # only select sequences/samples with len between [min, max]\n # 2) seq_length as [pad_flag]\n # all sequences used\n # pad_flag is always the last element: >0: cut/pad to the length, 0: keep original, <0: use max_seqlen\n if type(args.seq_length) not in (list, tuple):\n args.seq_length = [args.seq_length]\n if len(args.seq_length) == 2: args.seq_length = [0] + args.seq_length\n if len(args.seq_length) > 2: # [min, max, pad_flag]\n pkldata = dict()\n seqs_idx = np.array([_i for _i, _len in enumerate(in_pkldata['len']) \n if args.seq_length[0] <= _len <= args.seq_length[1]], dtype=np.int32)\n for _key, _val in in_pkldata.items():\n pkldata[_key] = [_val[_i] for _i in seqs_idx]\n pkldata['len'] = np.array(pkldata['len'], dtype=np.int32)\n num_seqs = len(seqs_idx)\n args.max_seqlen = pkldata['len'].max()\n logger.info(f'Selected {num_seqs} data sets with length range: {args.seq_length}')\n else:\n pkldata = in_pkldata\n seqs_idx = np.arange(num_seqs, dtype=np.int32)\n\n if args.seq_length[-1] < 0: \n seq_length = args.max_seqlen\n elif args.seq_length[-1] == 0:\n seq_length = 0\n else:\n seq_length = args.seq_length[-1]\n\n # process to get midata\n logger.debug('Using multiprocessing pool...')\n mpool = mpi.Pool(processes=int(mpi.cpu_count() * 0.8))\n\n # get \"x\": sequence data\n if args.residue_fmt in ['vector', 'embed']:\n embed_func = func_partial(mol_stru.vector_rna_seq, \n use_nn=args.residue_nn,\n use_attr=args.residue_attr,\n use_dbn=args.residue_dbn,\n length=seq_length)\n elif args.residue_fmt in ['scalar', 'quant']:\n embed_func = func_partial(mol_stru.quant_rna_seq, \n use_nn=args.residue_nn,\n use_dbn=args.residue_dbn,\n length=seq_length)\n else:\n logger.critical(f'Unknown residue format: {args.residue_fmt}')\n\n if args.residue_dbn and 'dbn' in pkldata:\n seqdata = mpool.starmap(embed_func, zip(pkldata['seq'], pkldata['dbn']))\n else:\n seqdata = mpool.map(embed_func, pkldata['seq'])\n\n # add residue_extra to the seqdata (aka input)\n if args.residue_extra and 'extra' in pkldata:\n seqdata = [np.concatenate((seqdata[_i], \n fix_length2d(pkldata['extra'][_i], [seqdata[_i].shape[0], pkldata['extra'][_i].shape[1]])),\n axis=1) for _i in range(len(seqdata))]\n\n # get length, idx (START FROM 1!!!)\n if seq_length > 0:\n seqs_len = pkldata['len'].copy()\n seqs_len[seqs_len > seq_length] = seq_length\n else:\n seqs_len = pkldata['len']\n\n lendata = np.concatenate((seqs_len.reshape((-1, 1)),\n seqs_idx.reshape((-1, 1)) + 1), axis=1)\n\n # get \"y\": upp/ct/...\n uppdata = None\n if 'upp' in pkldata:\n logger.info('Processing upp data...')\n if seq_length == 0:\n uppdata = pkldata['upp']\n else:\n pad_func = func_partial(fix_length1d, constant_values=(0,0))\n uppdata = mpool.starmap(pad_func, zip(pkldata['upp'], [seq_length] * num_seqs))\n\n ctdata = None\n if 'ct' in pkldata:\n logger.info('Processing ct data...')\n ctdata = mpool.starmap(mol_stru.ct2mat, zip(pkldata['ct'], pkldata['len']))\n\n if seq_length > 0:\n pad_func = func_partial(fix_length2d, constant_values=((0, 0), (0, 0)))\n ctdata = mpool.starmap(pad_func, zip(ctdata, [seq_length] * num_seqs))\n\n # expand the last dimension\n # ctdata = [np.expand_dims(_ct, -1) for _ct in ctdata]\n\n mpool.close()\n\n # return\n midata = None\n args.label_genre = args.label_genre.lower()\n if args.label_genre == 'upp':\n if uppdata is None:\n midata = list(zip(seqdata, lendata))\n else:\n midata = list(zip(seqdata, lendata, uppdata))\n elif args.label_genre == 'ct':\n if ctdata is None:\n midata = list(zip(seqdata, lendata))\n else:\n midata = list(zip(seqdata, lendata, ctdata))\n\n if args.verbose > 1:\n shapes = [data.shape for data in midata[0]]\n print(f'Number of datasets: {len(midata)}')\n print(f'Number of items in each set: {len(midata[0])}, with shapes: {shapes}')\n print('Checking for consistent dimensions...')\n for i, data in enumerate(midata):\n for j, item in enumerate(data):\n if shapes[j] != item.shape:\n print(f'The shape of dataset #{i} item #{j}: {item.shape} differs from the first: {shapes[j]}')\n print('Done!')\n\n return midata\n\n\ndef get_midata(args=misc.Struct(), **kwargs):\n \"\"\" kwargs > args > my_args\n return midata, meta_data (a pandas dataframe) \"\"\"\n pkldata = load_pkldata(args, **kwargs)\n return bake_midata(pkldata, args, **kwargs)\n\n\nclass MyDataset(Dataset):\n def __init__(self, data):\n super(MyDataset, self).__init__()\n self.data = data\n\n def __len__(self):\n return len(self.data)\n\n def __getitem__(self, idx):\n return self.data[idx]\n\n\ndef get_dataloader(midata, **kwargs):\n \"\"\" \"\"\"\n loader_opts = dict(\n shuffle = False,\n batch_size = 1,\n timeout = 0,\n num_workers = int(mpi.cpu_count() * 0.5),\n use_buffer_reader = True, # False will corrupt cuda version somehow\n )\n loader_opts.update(kwargs)\n\n return mi.io.DataLoader(MyDataset(midata), **loader_opts)\n\n\ndef get_net(args, quiet=False):\n \"\"\" \"\"\"\n\n if isinstance(args.load_dir, str): args.load_dir = path(args.load_dir)\n if isinstance(args.net_src_file, str): args.net_src_file = path(args.net_src_file)\n\n # get the local src code path\n if args.load_dir and args.net_src_file:\n local_src_code = args.load_dir / args.net_src_file.name\n else:\n local_src_code = path(args.net_src_file)\n\n # reload the net if a local copy exits\n if local_src_code.exists() and os.path.exists(MiNets.__file__) and not local_src_code.samefile(MiNets.__file__):\n args.net_src_file = local_src_code\n logger.info(f'Found local net definition: {local_src_code}')\n sys.path.insert(0, local_src_code.parent.absolute().as_posix())\n if local_src_code.name == path(MiNets.__file__).name:\n LocalNets = importlib.reload(MiNets)\n else:\n LocalNets = importlib.import_module(local_src_code.stem)\n sys.path.remove(local_src_code.parent.absolute().as_posix())\n else:\n LocalNets = globals()['MiNets']\n\n logger.info(f'Used net definition: {misc.str_color(LocalNets.__file__, bkg=\"cyan\")}')\n # locate net classes by name\n net_classes = getmembers(LocalNets, isclass)\n net_names = [_s[0].lower() for _s in net_classes]\n idx_net = misc.get_list_index(net_names, args.net.lower())\n if not idx_net:\n idx_net = misc.get_list_index(net_names, args.net.lower() + 'net')\n if not idx_net:\n logger.error(f'No net definition with name: {args.net} found!')\n return None\n\n # use the first match\n net_init = net_classes[idx_net[0]][1]\n upp_net = net_init(args)\n if not quiet and hasattr(upp_net, 'summary'):\n args.params = upp_net.summary()\n logger.info(f'{args.params}')\n\n return upp_net\n\n\ndef save_net_pycode(net_src_file, save_dir):\n \"\"\" not yet to get model save to work, so save the code! \"\"\"\n if isinstance(net_src_file, str): net_src_file = path(net_src_file)\n if isinstance(save_dir, str): save_dir = path(save_dir)\n\n # net_src_file = path(MyNets.__file__)\n net_des_file = save_dir / net_src_file.name\n if net_src_file.resolve().as_posix() == net_des_file.resolve().as_posix():\n logger.info(f'Net python code: {net_des_file} aleady exists...')\n else:\n net_des_file.write_text(net_src_file.read_text())\n logger.info(f'Saved net python code: {net_des_file}')\n return net_des_file\n\n\ndef get_optimizer(upp_net, args):\n \"\"\" two returns: the optimizer and [lr_scheduler or learning_rate] \"\"\"\n weight_decay = None\n if args.weight_decay.lower().startswith('l1'):\n weight_decay = mi.regularizer.L1Decay(args.l1decay)\n elif args.weight_decay.lower().startswith('l2'):\n weight_decay = mi.regularizer.L2Decay(args.l2decay)\n\n if args.lr_scheduler.lower().startswith('non'):\n learning_rate = args.learning_rate\n else:\n learning_rate = mi.optimizer.lr.ReduceOnPlateau(\n learning_rate = args.learning_rate,\n factor = args.lr_factor,\n patience = args.lr_patience,\n verbose = True,\n )\n\n upp_opt = mi.optimizer.Adam(\n parameters = upp_net.parameters(),\n weight_decay = weight_decay,\n learning_rate = learning_rate,\n beta1 = args.beta1,\n beta2 = args.beta2,\n epsilon = args.epsilon,\n )\n\n logger.info(f'Optimizer method: {args.optim}')\n logger.info(f' learning rate: {args.learning_rate}')\n logger.info(f' lr_scheduler: {args.lr_scheduler}')\n logger.info(f' weight decay: {args.weight_decay}')\n logger.info(f' l1decay: {args.l1decay}')\n logger.info(f' l2decay: {args.l2decay}')\n\n return upp_opt, learning_rate\n\n\ndef sigmoid_mse(input, label, reduction='none'):\n input = F.sigmoid(input)\n loss = F.mse_loss(input, label, reduction=reduction)\n return loss\n\n\ndef softmax_mse(input, label, label_col=1, reduction='none'):\n \"\"\" this only makes sense for input.shape[-1]=2 \n label_col is only used if input.ndim == label.ndim + 1\n \"\"\"\n\n input = F.softmax(input, axis=-1)\n \n # only take one axis for loss calculation\n if input.ndim == label.ndim + 1:\n if input.ndim == 2: # yet to find a better way, tensor doesn't accept [...,label_col]\n input = input[:, label_col].squeeze(-1)\n elif input.ndim == 3:\n input = input[:, :, label_col].squeeze(-1)\n elif input.ndim == 4:\n input = input[:, :, :, label_col].squeeze(-1)\n elif input.ndim == 5:\n input = input[:, :, :, :, label_col].squeeze(-1)\n elif input.ndim == 6:\n input = input[:, :, :, :, :, label_col].squeeze(-1)\n else:\n logger.critical(f'Feeling dizzy with too many dimensions: {input.ndim}!')\n\n loss = F.mse_loss(input, label, reduction=reduction)\n \n return loss\n\n\ndef softmax_bce(input, label, label_col=1, reduction='none'):\n \"\"\" this only makes sense for input.shape[-1]=2 \"\"\"\n assert input.ndim == label.ndim + 1, \\\n f\"input.ndim:{input.ndim}- label.ndim:{label.ndim} != 1!\"\n assert input.shape[-1] == 2, \\\n f\"input.shape[-1]:{input.shape[-1]} != 2!\"\n\n y0, y1 = mi.unstack(input, axis=-1)\n\n if label_col == 1: # which index to use to compare with the label\n y_delta = y1 - y0\n else:\n y_delta = y0 - y1\n\n # this is a reduced formula, please derive to check\n loss = mi.log(1.0 + mi.exp(y_delta)) - label * y_delta\n \n if reduction == 'mean':\n loss = loss.mean()\n \n return loss\n\n\nclass SeqLossFn_P2P(nn.Layer):\n \"\"\" Returns a scalar loss, loss_vs_seq: [N], std_vs_seq: [N].\n Designed for functions that can calculate loss with padded zeros.\n One exception is f_score as padding affects both fp and tn counts\n std_vs_seq in only available if loss functions calculate losses point to point along \"L\"\n \"\"\"\n\n def __init__(self, fn=F.mse_loss, name='mse', **kwargs):\n super(SeqLossFn_P2P, self).__init__()\n\n self.name = name.lower()\n self.fn = fn\n self.kwargs = kwargs\n\n def as_label(self, input):\n \"\"\" convert input to the same form as label, as model outputs may need\n to go through sigmod, softmax, etc.\n \"\"\"\n if not isinstance(input, mi.Tensor):\n input = mi.to_tensor(input)\n\n if self.name in ['mse', 'bce']:\n pass\n \n elif self.name in ['sigmoid+mse']:\n input = F.sigmoid(input)\n\n elif self.name in ['softmax+mse', 'softmax+bce']:\n input = mi.unstack(F.softmax(input, axis=-1), axis=-1)\n input = input[self.kwargs['label_col']]\n\n elif self.name in ['ce', 'crossentropy']:\n logger.critical('Need test! Help is the same as softmax+ce')\n\n elif self.name in ['softmax+ce', 'softmax+crossentropy']:\n if self.kwargs['soft_label']:\n input = F.softmax(input)\n else:\n input = F.softmax(input)\n input = mi.argmax(input, axis=-1)\n else:\n logger.critical(f'Cannot recognize loss_fn name: {self.name}!')\n\n return input\n\n def forward(self, input, label, seqs_len=None, loss_padding=False,\n loss_sqrt=False, **kwargs):\n \"\"\" return the loss with the same shape as input \"\"\"\n\n batch_size = input.shape[0]\n\n fn_kwargs = self.kwargs.copy()\n fn_kwargs.update(kwargs)\n\n # deal with the specific requirements of the loss functions\n if self.name in ['softmax+crossentropy', 'crossentropy']:\n\n if self.kwargs.get('soft_label', None): # soft label\n if input.ndim > label.ndim:\n label = hard2soft_label(label, nlabel=input.shape[-1], np=mi)\n else: # hard label \n if not isinstance(label, mi.Tensor) or label.dtype.name != 'INT64':\n label = mi.to_tensor(label, dtype='int64')\n \n if input.ndim > label.ndim:\n label = mi.unsqueeze(label, axis=-1)\n\n elif self.name in ['mse', 'sigmoid+mse']:\n if input.ndim == label.ndim + 1 and input.shape[-1] == 1:\n input = input.squeeze(-1)\n \n logger.debug(f' num_dims: {input.ndim}')\n logger.debug(f' data size: {batch_size}')\n logger.debug(f' seqs_len: {seqs_len is not None}')\n logger.debug(f' loss_padding: {loss_padding}')\n logger.debug(f' loss_sqrt: {loss_sqrt}')\n logger.debug(f' loss kwargs: {fn_kwargs}')\n\n # if not isinstance(y_truth, mi.Tensor) or y_truth.dtype.name != 'FP32':\n # y_truth = mi.to_tensor(y_truth, dtype='float32')\n\n # calculate all anyway, maybe more efficient for GPU\n loss_mat = self.fn(input, label, **fn_kwargs)\n\n if loss_padding: # process loss_mat as a whole\n if loss_mat.ndim == 1:\n # the std of the errors for each instance\n std_vs_seq = np.zeros_like(loss_mat, dtype=np.float32)\n # may need to squeeze the loss_mat\n loss_vs_seq = mi.squeeze(loss_mat, -1)\n else:\n # the axes for each instance, from the 2nd to the last\n inst_axes = tuple(range(1, loss_mat.ndim))\n std_vs_seq = loss_mat.numpy().std(axis=inst_axes)\n loss_vs_seq = mi.mean(loss_mat, axis=inst_axes)\n\n if loss_sqrt:\n loss_vs_seq = mi.sqrt(loss_vs_seq)\n std_vs_seq = np.sqrt(std_vs_seq)\n\n loss_for_backprop = mi.sum(loss_vs_seq)\n loss_vs_seq = loss_vs_seq.numpy()\n\n else: # deal each instance/sequence separately\n loss_for_backprop = mi.to_tensor(0.0, dtype='float32', stop_gradient=False)\n # loss_vs_seq = mi.zeros((batch_size,), dtype='float32')\n loss_vs_seq = np.zeros((batch_size), dtype=np.float32)\n std_vs_seq = np.zeros((batch_size), dtype=np.float32)\n for i in range(batch_size):\n seq_len = int(seqs_len[i])\n if loss_mat.ndim == 1:\n seq_loss = loss_mat[i] # self.loss_fn(input[i], label[i], **kwargs)\n elif loss_mat.ndim == 2:\n seq_loss = loss_mat[i, :seq_len]\n elif loss_mat.ndim == 3:\n seq_loss = loss_mat[i, :seq_len, :seq_len]\n elif loss_mat.ndim == 4:\n seq_loss = loss_mat[i, :seq_len, :seq_len, :seq_len]\n elif loss_mat.ndim == 5:\n seq_loss = loss_mat[i, :seq_len, :seq_len, :seq_len, :seq_len]\n else:\n logger.critical('too many dimensions for y_model, unsupported!')\n\n if loss_sqrt:\n loss_for_backprop += mi.sqrt(mi.mean(seq_loss))\n loss_vs_seq[i] = np.sqrt(seq_loss.numpy().mean())\n std_vs_seq[i] = np.sqrt(seq_loss.numpy().std())\n else:\n loss_for_backprop += mi.mean(seq_loss)\n loss_vs_seq[i] = seq_loss.numpy().mean()\n std_vs_seq[i] = seq_loss.numpy().std()\n\n # calculate the next loss_fn as needed\n # if self.loss_fn_next is not None:\n # loss_for_backprop2, loss_vs_seq2, std_vs_seq2 = self.loss_fn_next(input, label,\n # seqs_len=seqs_len, loss_padding=loss_padding, loss_sqrt=loss_sqrt, **self.loss_fn_next_kwargs)\n\n # loss_for_backprop += loss_for_backprop2\n # loss_vs_seq += loss_vs_seq2\n # std_vs_seq += std_vs_seq2\n\n return loss_for_backprop, loss_vs_seq, std_vs_seq\n\n\ndef get_loss_fn(args):\n \"\"\" \"\"\"\n if isinstance(args.loss_fn, str):\n args.loss_fn = [args.loss_fn]\n args.loss_fn = [_s.lower() for _s in args.loss_fn]\n\n loss_fn = []\n logger.info(f'Getting loss function: {args.loss_fn}')\n\n if 'mse' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(F.mse_loss, name='mse',\n reduction='none'))\n \n if 'bce' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(F.binary_cross_entropy, name='bce',\n reduction='none'))\n\n if 'sigmoid+mse' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(sigmoid_mse, name='sigmoid+mse',\n reduction='none'))\n\n if 'softmax+mse' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(softmax_mse, name='softmax+mse',\n label_col=1, reduction='none'))\n \n if 'softmax+bce' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(softmax_bce, name='softmax+bce',\n label_col=1, reduction='none'))\n \n if 'crossentropy' in args.loss_fn or 'ce' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(F.cross_entropy, name='crossentropy', \n soft_label=args.label_tone.lower() == 'soft'))\n \n if 'softmax+crossentropy' in args.loss_fn or 'softmax+ce' in args.loss_fn:\n loss_fn.append(SeqLossFn_P2P(F.softmax_with_cross_entropy, name='softmax+crossentropy',\n soft_label=args.label_tone.lower() == 'soft'))\n\n if len(loss_fn) == 0:\n logger.critical(f'not supported loss functions found in: {args.loss_fn}!')\n\n return loss_fn\n\n\ndef save_loss_csv(save_file, loss_df, groupby=None):\n \"\"\" \"\"\"\n df = loss_df\n if groupby is not None:\n col = [_col for _col in groupby if _col in loss_df.columns]\n if col:\n logger.info(f'Grouping data by: {col} before saving')\n df = loss_df.groupby(col).mean().reset_index()\n\n if df is not None:\n df.to_csv(save_file, index=False, float_format='%.4g')\n # np.savetxt(save_file, train_loss, fmt='%6d ' + '%8.4f ' * 4 + ' %5d'*4)\n\ndef save_model_prediction(y_model, save_dir='./', seqs_len=None, istart=1, stem='predict'):\n \"\"\" \"\"\"\n num_seqs = len(y_model)\n save_dir = path(save_dir)\n\n ndim = y_model[0].ndim\n for i in range(num_seqs):\n if seqs_len is None:\n y_save = y_model[i]\n else:\n seq_len = int(seqs_len[i])\n if ndim == 0:\n y_save = y_model[i]\n elif ndim == 1:\n y_save = y_model[i][:seq_len]\n elif ndim == 2:\n y_save = y_model[i][:seq_len, :seq_len]\n elif ndim == 3:\n y_save = y_model[i][:seq_len, :seq_len, :seq_len]\n elif ndim == 4:\n y_save = y_model[i][:seq_len, :seq_len, :seq_len, :seq_len]\n else:\n logger.critical('too many dimensions to save')\n \n if ndim <= 2:\n np.savetxt(save_dir / f'{istart + i}.{stem}.txt', y_save, fmt='%10.8f')\n else:\n np.save(save_dir / f'{istart + i}.{stem}', y_save)\n\n\ndef compute_loss(loss_fn, input, label, seqs_len=None, shuffle=False, batch_size=23, **kwargs):\n \"\"\" both input and label can be list/array/tensor\n But the first dimension must be the batch_size\n \"\"\"\n if type(loss_fn) not in (list, tuple): loss_fn = [loss_fn]\n num_data = len(input)\n if seqs_len is None:\n midata = list(zip(input, label))\n else:\n midata = list(zip(input, seqs_len, label))\n\n miloader = get_dataloader(midata, batch_size=batch_size, shuffle=shuffle)\n\n loss_vs_seq = np.zeros((num_data), dtype=np.float32)\n std_vs_seq = np.zeros((num_data), dtype=np.float32)\n\n for ibatch, data in enumerate(miloader()):\n num_seqs = data[0].shape[0]\n istart = ibatch * batch_size\n iend = istart + num_seqs\n if seqs_len is not None:\n seqlen_batch = data[1]\n else:\n seqlen_batch = None\n\n for one_loss_fn in loss_fn:\n _, _loss_vs_seq, _std_vs_seq = one_loss_fn(data[0], data[-1],\n seqs_len=seqlen_batch, **kwargs)\n loss_vs_seq[istart:iend] = loss_vs_seq[istart:iend] + _loss_vs_seq\n std_vs_seq[istart:iend] = std_vs_seq[istart:iend] + _std_vs_seq\n\n return loss_vs_seq, std_vs_seq\n\n\ndef train(model, midata, **kwargs):\n \"\"\" Bad practices:\n 1) args procesing is odd (kwargs > model.args > args )\n 2) fields are added to model structure\n \"\"\"\n # default settings\n args = misc.Struct(dict(\n trainloss_patience = 5, trainloss_rdiff = 1e-3,\n validloss_patience = 3, validloss_rdiff = 1e-3,\n validate_callback = None,\n num_callbacks_per_epoch = 10,\n lr_scheduler = 'none',\n num_recaps_per_epoch = 30,\n num_epochs = 2,\n batch_size = 2,\n loss_padding = False,\n shuffle = True,\n save_dir = None,\n save_level = 1,\n verbose = 1,\n ))\n args.update(vars(model.args)) # model.args overwrite default args\n args.update(kwargs) # kwargs overwrite all\n if isinstance(args.save_dir, str): args.save_dir = path(args.save_dir)\n if args.save_dir: args.save_dir.mkdir(parents=True, exist_ok=True)\n model.args.update(vars(args)) # args should not change anymore\n\n if args.data_size > 0:\n if args.data_size < len(midata):\n midata = random_sample(midata, size=args.data_size, replace=False)\n elif args.data_size == len(midata):\n logger.warning(f'Specified data size: {args.data_size} == data length: {len(midata)}.')\n else:\n logger.warning(f'Specified data size: {args.data_size} > data length: {len(midata)}!')\n\n miloader = get_dataloader(midata, batch_size=args.batch_size, shuffle=args.shuffle)\n model.num_batches = len(miloader)\n model.num_data = len(midata)\n\n model.train_loss = [] # a list of DataFrames (concatenated at the end)\n validate_hist = misc.Struct(valid_loss=[]) # consider to include this in the model structure?\n model.validate_hist = validate_hist\n\n callback_interval = max([1, model.num_batches // args.num_callbacks_per_epoch])\n recap_interval = model.num_batches // args.num_recaps_per_epoch + 1\n # num_recaps = int(np.ceil(model.num_batches / recap_interval))\n logger.info(f'Training, data size: {len(midata)}')\n logger.info(f' batch size: {args.batch_size}')\n logger.info(f' shuffle: {args.shuffle}')\n logger.info(f' # of batches: {model.num_batches}')\n logger.info(f' recap interval: {recap_interval}')\n logger.info(f' validate interval: {callback_interval}')\n logger.info(f' # of epochs: {args.num_epochs}')\n logger.info(f' loss padding: {args.loss_padding}')\n\n # temporary vars for journaling, not saved to files\n loss_vs_epoch = pd.DataFrame() # np.array([], dtype=np.float32)\n loss_for_recap, std_for_recap = [], [] # accumulate results between recaps\n\n model.net.train()\n for model.epoch in range(args.num_epochs):\n loss_one_epoch = np.empty((model.num_data, 7), dtype=np.float32)\n for model.batch, data in enumerate(miloader()):\n model.optim.clear_grad()\n\n # data: [seq_in, upp_truth, [seq_len, idx]]\n x, y_truth = data[0], data[-1]\n seqs_len, seqs_idx = data[1][:, 0], data[1][:, 1]\n\n if x.ndim > 1 and x.shape[0] == 1 and x.shape[1] > seqs_len[0] and not args.loss_padding:\n x = cut_padding(x, int(seqs_len[0]))\n y_truth = cut_padding(y_truth, int(seqs_len[0]))\n\n y_model = model.net(x, seqs_len)\n\n num_seqs = y_model.shape[0]\n seqs_len, seqs_idx = seqs_len.numpy(), seqs_idx.numpy()\n\n loss_for_backprop = mi.to_tensor(0.0, dtype='float32', stop_gradient=False)\n loss_vs_seq = np.zeros((num_seqs), dtype=np.float32)\n std_vs_seq = np.zeros((num_seqs), dtype=np.float32)\n for loss_fn in model.loss_fn:\n _loss_for_backprop, _loss_vs_seq, _std_vs_seq = loss_fn(y_model, y_truth,\n seqs_len=seqs_len, loss_padding=args.loss_padding,\n loss_sqrt=args.loss_sqrt)\n loss_for_backprop += _loss_for_backprop\n loss_vs_seq += _loss_vs_seq\n std_vs_seq += _std_vs_seq\n\n loss_for_backprop.backward() # loss is loss_per_batch by convention\n if args.verbose > 1:\n print(\"Current state of net parameters:\")\n for par_n, par_v in model.net.named_parameters():\n\n print(f'{par_n:28s} - min: {par_v.min().numpy()[0]:11.6f}, max: {par_v.max().numpy()[0]:11.6f}, ' + \\\n f'grad_min: {par_v.grad.min().item():11.6f}, grad_max: {par_v.grad.max().item():11.6f}')\n\n # save&display progress\n ibatch = model.epoch * model.num_batches + model.batch\n istart = model.batch * args.batch_size # model.epoch * model.num_data +\n iend = istart + num_seqs\n loss_one_epoch[istart:iend, 0] = ibatch\n loss_one_epoch[istart:iend, 1] = loss_vs_seq\n loss_one_epoch[istart:iend, 2] = std_vs_seq\n loss_one_epoch[istart:iend, 3] = seqs_len\n loss_one_epoch[istart:iend, 4] = seqs_idx\n loss_one_epoch[istart:iend, 5] = model.epoch\n loss_one_epoch[istart:iend, 6] = model.batch\n\n loss_for_recap.extend(loss_vs_seq)\n std_for_recap.extend(std_vs_seq)\n\n if model.batch % recap_interval == 0: # recap the 0th!\n loss_for_recap = np.array(loss_for_recap).mean()\n std_for_recap = np.array(std_for_recap).mean()\n logger.info(f'Epoch/batch: {model.epoch:d}/{model.batch:4d}, ibatch: {ibatch:4d}, ' + \\\n f'loss: \\033[0;36m{loss_for_recap:6.4f}\\033[0m, std: {std_for_recap:6.4f}')\n\n # learning_rate tuning\n if not isinstance(model.lr_scheduler, float): # mi.optimizer.lr.LRScheduler):\n model.lr_scheduler.step(loss_for_recap)\n\n loss_for_recap, std_for_recap = [], []\n\n # callback (usually for early stopping)\n if args.validate_callback is not None and (model.batch % callback_interval == 0\n or model.batch == model.num_batches - 1):\n validate_hist.update(epoch=model.epoch, batch=model.batch, ibatch=ibatch)\n validate_hist = args.validate_callback(model=model, history=validate_hist)\n model.net.train()\n\n model.optim.step()\n\n # post-epoch\n loss_one_epoch = pd.DataFrame(loss_one_epoch, columns=['ibatch', 'loss', 'loss_std', 'seq_len', 'idx','epoch', 'batch'])\n model.train_loss.append(loss_one_epoch) # this will be saved as csv\n\n loss_vs_epoch = loss_vs_epoch.append(loss_one_epoch.mean(), ignore_index=True)\n\n logger.info(f'Epoch {model.epoch} average training loss: ' +\n f'\\033[0;46m{loss_vs_epoch.loss.iat[-1]:6.4f}\\033[0m' +\n f' std: {loss_vs_epoch.loss_std.iat[-1]:6.4f}')\n\n valid_vs_epoch = validate_hist.loss_per_call.groupby('epoch').mean().reset_index()\n logger.info(f'Epoch {model.epoch} average validate loss: ' +\n f'\\033[0;46m{valid_vs_epoch.loss.iat[-1]:6.4f}\\033[0m' +\n f' std: {valid_vs_epoch.loss_std.iat[-1]:6.4f}')\n\n if args.save_dir and args.save_level >= 2:\n epoch_save_dir = args.save_dir / 'epoch_log'\n epoch_save_dir.mkdir(parents=True, exist_ok=True)\n save_loss_csv(epoch_save_dir / f'train_epo{model.epoch:03d}_{loss_vs_epoch.loss.iat[-1]:6.4f}.csv', loss_one_epoch)\n save_loss_csv(epoch_save_dir / f'valid_epo{model.epoch:03d}_{valid_vs_epoch.loss.iat[-1]:6.4f}.csv', validate_hist.valid_loss[-1])\n\n # stop the train if needed\n if model.epoch >= args.trainloss_patience and model.epoch >= args.validloss_patience:\n\n if all((loss_vs_epoch.loss.diff() > 0)[-args.trainloss_patience:]):\n logger.warning(f'Training loss increased {args.trainloss_patience} consecutive epochs, stopping!!!')\n break\n if all(loss_vs_epoch.loss.pct_change().abs()[-args.trainloss_patience:] < args.trainloss_rdiff):\n logger.warning(f'Training loss changed < {args.trainloss_rdiff} for {args.trainloss_patience} consecutive epochs, stopping!!!')\n break\n if all((valid_vs_epoch.loss.diff() > 0)[-args.validloss_patience:]):\n logger.warning(f'Validate loss increased {args.validloss_patience} consecutive epochs, stopping!!!')\n break\n if all(valid_vs_epoch.loss.pct_change().abs()[-args.validloss_patience:] < args.validloss_rdiff):\n logger.warning(f'Validate loss changed < {args.validloss_rdiff} for {args.validloss_patience} consecutive epochs, stopping!!!')\n break\n\n # post-train\n model.train_loss = pd.concat(model.train_loss)\n model.train_loss_vs_epoch = loss_vs_epoch\n\n model.valid_loss = pd.concat(validate_hist.valid_loss)\n model.valid_loss_vs_epoch = validate_hist.loss_per_call.groupby('epoch').min().reset_index()\n\n model.validate_hist = validate_hist\n\n if args.save_dir and args.save_level >= 1: # save the final model (maybe unnecessary)\n logger.info(f'Saving final results in <{args.save_dir}>...')\n state_dict_save(model, fdir=args.save_dir)\n args.net_src_file = save_net_pycode(args.net_src_file, args.save_dir)\n save_loss_csv(args.save_dir / 'train_log.csv', model.train_loss, groupby=args.save_grpby)\n save_loss_csv(args.save_dir / 'valid_log.csv', model.valid_loss, groupby=args.save_grpby)\n\n return model.train_loss, validate_hist.valid_loss\n\n\ndef validate(model, midata, **kwargs):\n \"\"\" model structure is not changed during this call \"\"\"\n args = misc.Struct(dict(batch_size = 512,\n shuffle = False,\n num_recaps_per_epoch = 10,\n save_dir = None,\n verbose = 1,\n ))\n args.update(vars(model.args))\n args.update(kwargs) # kwargs rule all\n if isinstance(args.save_dir, str): args.save_dir = path(args.save_dir)\n model.args.update(vars(args)) # args should not change anymore\n\n miloader = get_dataloader(midata, batch_size=args.batch_size, shuffle=args.shuffle)\n # return: [ibatch, rmsd, std, seq_len, idx (in the original seq data)]\n valid_loss = np.zeros((len(midata), 5), dtype=np.float32)\n\n recap_interval = len(miloader) // args.num_recaps_per_epoch + 1\n logger.info(f'Validating, data size: {len(midata)}')\n logger.info(f' batch size: {args.batch_size}')\n logger.info(f' shuffle: {args.shuffle}')\n logger.info(f' # of batches: {len(miloader)}')\n logger.info(f' recap interval: {recap_interval}')\n logger.info(f' loss padding: {args.loss_padding}')\n\n model.net.eval()\n with mi.no_grad():\n loss_for_recap, std_for_recap = [], []\n for ibatch, data in enumerate(miloader()):\n # data: [seq_in, upp_truth, [seq_len, idx]]\n x, y_truth,= data[0], data[-1]\n seqs_len, seqs_idx = data[1][:,0], data[1][:,1]\n\n if x.ndim > 1 and x.shape[0] == 1 and x.shape[1] > seqs_len[0] and not args.loss_padding:\n x = cut_padding(x, int(seqs_len[0]))\n y_truth = cut_padding(y_truth, int(seqs_len[0]))\n\n y_model = model.net(x, seqs_len)\n\n num_seqs = y_model.shape[0]\n seqs_len, seqs_idx = seqs_len.numpy(), seqs_idx.numpy()\n\n loss_vs_seq = np.zeros((num_seqs), dtype=np.float32)\n std_vs_seq = np.zeros((num_seqs), dtype=np.float32)\n for loss_fn in model.loss_fn:\n _, _loss_vs_seq, _std_vs_seq = loss_fn(y_model, y_truth,\n seqs_len=seqs_len, loss_padding=args.loss_padding, loss_sqrt=args.loss_sqrt)\n loss_vs_seq += _loss_vs_seq\n std_vs_seq += _std_vs_seq\n\n istart = ibatch * args.batch_size\n iend = istart + num_seqs\n valid_loss[istart:iend, 0] = ibatch\n valid_loss[istart:iend, 1] = loss_vs_seq\n valid_loss[istart:iend, 2] = std_vs_seq\n valid_loss[istart:iend, 3] = seqs_len\n valid_loss[istart:iend, 4] = seqs_idx\n\n loss_for_recap.extend(loss_vs_seq)\n std_for_recap.extend(std_vs_seq)\n if ibatch % recap_interval == 0:\n loss_for_recap = np.array(loss_for_recap).mean()\n std_for_recap = np.array(std_for_recap).mean()\n logger.info(f'ibatch: {ibatch:4d}, loss: {loss_for_recap:6.4f}, std: {std_for_recap:6.4f}')\n loss_for_recap, std_for_recap = [], []\n\n valid_loss = pd.DataFrame(valid_loss, columns=['ibatch', 'loss', 'loss_std', 'seq_len', 'idx'])\n logger.info(f'Validate mean: \\033[0;46m{valid_loss.loss.mean():6.4f}\\033[0m' +\n f', std: {valid_loss.loss.std():6.4f}')\n\n if args.save_dir and args.save_level >= 1:\n if not args.save_dir.exists(): args.save_dir.mkdir(parents=True)\n save_loss_csv(args.save_dir / 'valid_log.csv', valid_loss)\n return valid_loss\n\n\ndef predict(model, midata, **kwargs):\n \"\"\" \"\"\"\n args = misc.Struct(dict(\n batch_size = 128,\n shuffle = False, # the first two not used yet\n num_recaps_per_epoch = 10,\n save_dir = path.cwd() / 'predict.files',\n ))\n args.update(vars(model.args))\n args.update(kwargs) # kwargs rule all\n if args.save_dir and not isinstance(args.save_dir, path):\n args.save_dir = path(args.save_dir)\n model.args.update(vars(args))\n\n miloader = get_dataloader(midata, batch_size=args.batch_size, shuffle=args.shuffle)\n\n recap_interval = len(miloader) // args.num_recaps_per_epoch + 1\n data_size = len(midata)\n\n logger.info(f'Predicting, data size: {data_size}')\n logger.info(f' batch size: {args.batch_size}')\n logger.info(f' shuffle: {args.shuffle}')\n logger.info(f' # of batches: {len(miloader)}')\n logger.info(f' recap interval: {recap_interval}')\n\n if args.save_dir and args.save_level >= 1:\n args.save_dir.mkdir(parents=True, exist_ok=True)\n logger.info(f'Predicted files will be saved in: {args.save_dir}')\n\n # two returned values\n y_model_all = [] # np.empty((data_size, midata[0][0].shape[0]), dtype=np.float32)\n seqs_len_all = np.empty((data_size), dtype=np.int32)\n\n model.net.eval()\n with mi.no_grad(), tqdm(total=data_size, disable=False) as prog_bar:\n for ibatch, data in enumerate(miloader()):\n x, y_truth = data[0], data[-1]\n seqs_len, seqs_idx = data[1][:, 0], data[1][:, 1]\n\n if x.ndim > 1 and x.shape[0] == 1 and x.shape[1] > seqs_len[0] and not args.loss_padding:\n x = cut_padding(x, int(seqs_len[0]))\n y_truth = cut_padding(y_truth, int(seqs_len[0]))\n\n y_model = model.net(x, seqs_len)\n num_seqs = y_model.shape[0]\n\n istart = ibatch * args.batch_size\n iend = istart + num_seqs\n y_model_all.extend(y_model.numpy())\n seqs_len_all[istart:iend] = seqs_len.numpy()\n\n prog_bar.update(num_seqs)\n if args.save_dir:\n y_model = model.loss_fn[0].as_label(y_model)\n save_model_prediction(y_model.numpy(), args.save_dir, seqs_len=seqs_len,\n istart=ibatch * args.batch_size + 1, stem='predict')\n\n logger.info(f'Completed prediction of {data_size} samples')\n return y_model_all, seqs_len_all\n\n\ndef state_dict_save(upp_model, fdir=path.cwd()):\n \"\"\" \"\"\"\n if isinstance(fdir, str): fdir = path(fdir)\n if not fdir.exists(): fdir.mkdir(parents=True)\n\n net_state_file = fdir / 'net.state'\n opt_state_file = fdir / 'opt.state'\n\n mi.save(upp_model.net.state_dict(), net_state_file)\n mi.save(upp_model.optim.state_dict(), opt_state_file)\n\n # mi.jit.save(model.net, (fdir / 'model').as_posix())\n logger.info(f'Saved model states in: {fdir}')\n\n\ndef state_dict_load(upp_model, fdir=path.cwd()):\n\n if isinstance(fdir, str): fdir = path(fdir)\n logger.info(f'Loading model states from: {fdir}')\n\n net_state_file = fdir / 'net.state'\n opt_state_file = fdir / 'opt.state'\n\n try:\n if net_state_file.exists():\n net_state_dict = mi.load(net_state_file)\n upp_model.net.set_state_dict(net_state_dict)\n logger.info(f'Loaded net state: {net_state_file}')\n except:\n logger.warning('Error in net state_dict loading!')\n\n try:\n if opt_state_file.exists():\n opt_state_dict = mi.load(opt_state_file)\n upp_model.optim.set_state_dict(opt_state_dict, use_structured_name=False)\n logger.info(f'Loaded optim state: {opt_state_file}')\n except:\n logger.warning('Error in optim state_dict loading!')\n\n\ndef get_model(args, quiet=False):\n \"\"\" \"\"\"\n upp_model = misc.Struct()\n upp_model.args = args\n upp_model.net = get_net(args, quiet=quiet)\n upp_model.optim, upp_model.lr_scheduler = get_optimizer(upp_model.net, args)\n upp_model.loss_fn = get_loss_fn(args)\n return upp_model\n\n\ndef validate_in_train(model=None, midata=None, history=misc.Struct(), **kwargs):\n \"\"\" \"\"\"\n configs = misc.Struct(\n epoch = 0, # the following three are from the train()\n batch = 0, # so as to align with the training curve\n ibatch = 0,\n batch_size = 128, # for early stop only\n times_called = 0,\n # valid_loss = pd.DataFrame(),\n valid_loss = [],\n loss_per_call = pd.DataFrame(columns=['ibatch', 'loss', 'loss_std', 'epoch', 'batch']),\n save_dir = None,\n saved_idx = [], # this points to loss_hist\n saved_dirs = [],\n valid_loss_best = 1000.0,\n keep_best_only = True,\n verbose = 1,\n )\n configs.update(vars(history))\n configs.update(kwargs)\n configs.times_called += 1\n\n # this is different from args.save_dir\n if configs.save_dir and isinstance(configs.save_dir, str):\n configs.save_dir = path(configs.save_dir)\n\n misc.logger_setlevel(logger, 0)\n valid_loss = validate(model, midata, save_dir=None, shuffle=False, batch_size=configs.batch_size)\n misc.logger_setlevel(logger, configs.verbose)\n\n valid_loss_avg = valid_loss.loss.mean()\n valid_loss_std = valid_loss.loss_std.mean()\n logger.info(f'loss: \\033[0;32m{valid_loss_avg:6.4f}\\033[0m, std: {valid_loss_std:6.4f}')\n\n configs.valid_loss.append(valid_loss.assign(ibatch=configs.ibatch,\n epoch=configs.epoch, batch=configs.batch))\n\n configs.loss_per_call = configs.loss_per_call.append(dict(ibatch=configs.ibatch,\n loss=valid_loss_avg, loss_std=valid_loss_std,\n epoch=configs.epoch, batch=configs.batch), ignore_index=True)\n\n if configs.save_dir and valid_loss_avg < configs.valid_loss_best and configs.times_called > 1:\n configs.valid_loss_best = valid_loss_avg\n\n # call \"next_backup_path\" just in case it exists\n new_save_dir = gwio.next_backup_path(configs.save_dir / f'earlystop_{valid_loss_avg:6.4f}')\n\n state_dict_save(model, fdir=new_save_dir)\n model.args.net_src_file = save_net_pycode(model.args.net_src_file, new_save_dir)\n gwio.dict2json(vars(model.args), new_save_dir / 'args.json')\n save_loss_csv(new_save_dir / 'valid_loss.csv', valid_loss)\n logger.info(f'Saved best model: {new_save_dir}')\n\n configs.saved_dirs.append(new_save_dir)\n configs.saved_idx.append(configs.loss_per_call.shape[0] - 1)\n\n if configs.keep_best_only:\n for old_save_dir in configs.saved_dirs[-2:-1]:\n if not old_save_dir.exists() or new_save_dir.samefile(old_save_dir): continue\n logger.info(f'Removing earlystop model: {old_save_dir}')\n shutil.rmtree(old_save_dir)\n\n return configs\n\n\ndef scan_data_args(args, midata, data_sizes=None, batch_sizes=None, **kwargs):\n \"\"\" data/batch_sizes='auto'|None|int|list/array \"\"\"\n args.update(kwargs)\n\n if args.save_dir and isinstance(args.save_dir, str):\n args.save_dir = path(args.save_dir)\n if args.save_dir: args.save_dir.mkdir(parents=True, exist_ok=True)\n\n num_data = len(midata)\n\n # get data grids, the default goes up from 1 by a factor 2\n def get_data_grid(val_in, default=1):\n if val_in is None:\n val_out = [default]\n elif isinstance(val_in, str):\n if val_in.lower() == 'auto':\n num_grids = int(np.log2(num_data)) + 1\n val_out = np.logspace(0, num_grids - 1, num=num_grids, base=2, dtype=int)\n elif val_in.lower() == 'all': # all\n val_out = [num_data]\n else:\n val_out = [1, num_data]\n elif isinstance(val_in, int):\n val_out = [val_in]\n else:\n val_out = np.array(val_in, dtype=int)\n return val_out\n\n data_sizes = get_data_grid(data_sizes, default=num_data)\n batch_sizes = get_data_grid(batch_sizes, default=args.batch_size)\n\n scan_best_loss = pd.DataFrame() # np.zeros((len(data_sizes)*len(batch_sizes), 6), dtype=float)\n # the lists here store the train_loss, etc. from each scan\n scan_train_loss = []\n scan_valid_loss = [] # this is the callback return\n\n data_indices = np.linspace(0, num_data-1, num=num_data, dtype=int)\n for i, (data_size, batch_size) in enumerate(itertools.product(data_sizes, batch_sizes)):\n logger.info(f'scan #{i}, data_size: {data_size}, batch_size: {batch_size}')\n batch_size = int(batch_size) # somehow neede for paddle\n\n # get train data\n if 0 < data_size < num_data:\n data_indices = np.random.permutation(data_indices)\n train_data = [midata[data_indices[_i]] for _i in range(data_size)]\n else:\n train_data = midata\n\n # train with chosen data and batch_size\n logger.info('Creating a new model...') # or re-initialize the model\n model = get_model(args)\n\n # both returns are pd.DataFrames of\n save_dir = args.save_dir\n train(model, train_data, batch_size=batch_size, save_dir=None, validate_callback=args.validate_callback)\n args.save_dir = save_dir\n\n # should consier to reduce the level to batch, at least\n scan_train_loss.append(model.train_loss.assign(data_size=data_size, batch_size=batch_size))\n scan_valid_loss.append(model.valid_loss.assign(data_size=data_size, batch_size=batch_size))\n\n scan_best_loss = scan_best_loss.append(dict(\n data_size = data_size,\n batch_size = batch_size,\n train_loss = model.train_loss_vs_epoch.loss.min(),\n valid_loss = model.valid_loss_vs_epoch.loss.min(),\n ), ignore_index=True)\n\n if args.save_dir: # save all the train and valid curves from each call\n save_prefix = f'scan_data_size{data_sizes[0]}-{data_sizes[-1]}' + \\\n f'_batch{batch_sizes[0]}-{batch_sizes[-1]}'\n\n save_file = args.save_dir / (save_prefix + '_train.csv')\n save_loss_csv(save_file, pd.concat(scan_train_loss), groupby=['data_size', 'batch_size', 'ibatch'])\n logger.info(f'Saved train curve: {save_file}')\n\n save_file = args.save_dir / (save_prefix + '_valid.csv')\n save_loss_csv(save_file, pd.concat(scan_valid_loss), groupby=['data_size', 'batch_size', 'ibatch'])\n logger.info(f'Saved valid curve: {save_file}')\n\n save_file = args.save_dir / (save_prefix + '_best.csv')\n scan_best_loss.to_csv(save_file, index=False, float_format='%.4g')\n logger.info(f'Saved scan summary: {save_file}')\n\n return scan_best_loss\n\n\ndef scout_args(args, train_set, valid_set=None, arg_names=None, arg_values=None,\n grid_search=False, **kwargs):\n \"\"\" both arg_names and argvalues are lists of MATCHING names/values \"\"\"\n # take care of args\n args.update(kwargs)\n\n if args.save_dir and isinstance(args.save_dir, str):\n args.save_dir = path(args.save_dir)\n if args.save_dir: args.save_dir.mkdir(parents=True, exist_ok=True)\n\n # scan_best_loss = np.zeros((np.prod(arg_lens), 6), dtype=float)\n scan_best_loss = pd.DataFrame(columns=arg_names + ['train_loss', 'valid_loss'], dtype=float)\n\n scan_train_loss = []\n scan_valid_loss = []\n\n if grid_search:\n arg_sets = list(itertools.product(*arg_values))\n else:\n arg_sets = list(zip(*arg_values))\n\n for i, value_set in enumerate(arg_sets):\n scan_args = dict(zip(arg_names, value_set))\n scan_best_loss.loc[i, arg_names] = value_set\n logger.info(f'args set #: {i}/{len(arg_sets)}, {scan_args}')\n\n args.update(scan_args)\n\n if args.rebake_midata: # midata should be the pkldata!!!\n args = autoconfig_args(args)\n train_data = bake_midata(train_set, args)\n if valid_set is not None:\n valid_data = bake_midata(valid_set, args)\n args.validate_callback = func_partial(validate_in_train, midata=valid_data,\n save_dir=args.save_dir, verbose=args.verbose)\n else:\n train_data = train_set\n\n model = get_model(args)\n\n save_dir = args.save_dir # train() overwrites save_dir\n train(model, train_data, save_dir=None, validate_callback=args.validate_callback)\n args.save_dir = save_dir\n\n # reduce the train_loss???\n scan_train_loss.append(model.train_loss.assign(**scan_args)) # append scan_args\n scan_valid_loss.append(model.valid_loss.assign(**scan_args))\n\n scan_best_loss.loc[i, 'train_loss'] = model.train_loss_vs_epoch.loss.min()\n scan_best_loss.loc[i, 'valid_loss'] = model.valid_loss_vs_epoch.loss.min()\n\n # Saving results\n if args.save_dir: # save all the train and valid curves from each call\n save_prefix = 'scan_args_' + '-'.join(arg_names)\n\n save_file = args.save_dir / (save_prefix + '_train.csv')\n save_loss_csv(save_file, pd.concat(scan_train_loss), groupby=arg_names + ['ibatch'])\n logger.info(f'Saved train curve: {save_file}')\n\n save_file = args.save_dir / (save_prefix + '_valid.csv')\n save_loss_csv(save_file, pd.concat(scan_valid_loss), groupby=arg_names + ['ibatch'])\n logger.info(f'Saved valid curve: {save_file}')\n\n save_file = args.save_dir / (save_prefix + '_best.csv')\n scan_best_loss.to_csv(save_file, index=False, float_format='%.4g')\n logger.info(f'Saved scan summary: {save_file}')\n else:\n logger.info(f'scan results are not saved with args.save_dir: {args.save_dir}')\n\n return scan_best_loss\n\n\ndef fly(args, **kwargs):\n \"\"\" \"\"\"\n if not isinstance(args, misc.Struct):\n args = misc.Struct(vars(args))\n\n # check local configuration json\n config_json = path.cwd() / 'config.json'\n if config_json.exists():\n logger.info(f'Loading local configuration json: {config_json}')\n args_local = gwio.json2dict(config_json)\n # print(json.dumps(args_local, indent=4))\n print(gwio.json_str(args_local))\n args.update(args_local)\n\n args.update(kwargs)\n logger.info(f'Applying kwargs:')\n print(gwio.json_str(kwargs))\n\n args = autoconfig_args(args) # resolve inconsistencies\n args.update(kwargs) # reapply...\n\n action_list = args.action\n if isinstance(action_list, str): action_list = [action_list]\n\n # resolve the load_dir use for loading purpose only\n if args.load_dir:\n if isinstance(args.load_dir, str): args.load_dir = path(args.load_dir)\n if not args.load_dir.exists():\n logger.critical(f'Model directory: {args.load_dir} does not exist, fail to load!')\n logger.critical(f'Use --save_dir {args.load_dir} if intended for saving...')\n sys.exit(1)\n elif 'predict' in sys.argv or 'validate' in sys.argv: # not really needed any more\n # beter load a model for validation and test/predict, set default\n args.load_dir = gwio.last_backup_path(args.net)\n\n if args.load_dir: # load args.json\n logger.info(f'Loading model args from directory: {args.load_dir}')\n if (args.load_dir / 'args.json').exists():\n args.update(gwio.json2dict(fname='args.json', fdir=args.load_dir))\n args.update(kwargs) # kwargs overwrite the model args!!!\n if isinstance(args.load_dir, str): args.load_dir = path(args.load_dir)\n else:\n logger.warning(f'args.json not found in {args.load_dir}, ' + \\\n 'using default/command line args!!')\n\n if not args.save_dir: # set up save_dir\n args.save_dir = args.load_dir if args.load_dir else gwio.next_backup_path(args.net)\n if isinstance(args.save_dir, str): args.save_dir = path(args.save_dir)\n if not args.save_dir.exists(): args.save_dir.mkdir(parents=True)\n\n logger.info(f'Results will be saved in: \\033[0;46m{args.save_dir}\\033[0m')\n gwio.dict2json(vars(args), fname='last.json', fdir=path.cwd())\n\n if 'summary' in action_list or 'summarize' in action_list or 'view' in action_list:\n model = get_model(args)\n\n if 'train' in action_list:\n model = get_model(args)\n if args.resume and args.load_dir:\n state_dict_load(model, fdir=args.load_dir)\n\n args.net_src_file = save_net_pycode(args.net_src_file, args.save_dir)\n gwio.dict2json(vars(args), fname='args.json', fdir=args.save_dir)\n\n midata = get_midata(args)\n train_data, valid_data = train_test_split(midata, \n test_size=args.test_size, random_state=args.split_seed)\n\n callback_func = func_partial(validate_in_train, midata=valid_data,\n save_dir=args.save_dir, verbose=args.verbose)\n\n train(model, train_data, validate_callback=callback_func)\n\n # plot or do visualDL for paddle\n\n if 'cross_validate' in action_list:\n model = get_model(args)\n if args.resume and args.load_dir:\n state_dict_load(model, fdir=args.load_dir)\n\n args.net_src_file = save_net_pycode(args.net_src_file, args.save_dir)\n gwio.dict2json(vars(args), fname='args.json', fdir=args.save_dir)\n\n midata = get_midata(args)\n train_data, valid_data = train_test_split(midata, test_size=args.test_size, \n random_state=args.split_seed)\n\n # train as usual first\n callback_func = func_partial(validate_in_train, midata=valid_data,\n save_dir=args.save_dir, verbose=args.verbose)\n\n train(model, train_data, validate_callback=callback_func)\n\n # cross-validation training\n num_seqs = len(train_data)\n data_indices = np.linspace(0, num_seqs-1, num=num_seqs, dtype=int)\n data_indices = np.random.permutation(data_indices)\n\n num_xvalids = num_seqs // args.num_cvs\n xvalid_dirs = []\n for i in range(args.num_cvs):\n valid_data_xv = [train_data[data_indices[j]] for j in\n range(i*num_xvalids, (i+1)*num_xvalids)]\n train_data_xv = [train_data[data_indices[j]] for j in\n itertools.chain(range(0, i*num_xvalids), range((i+1)*num_xvalids, num_seqs))]\n\n save_dir_xv = args.save_dir / f'xvalid_{i}'\n xvalid_dirs.append(save_dir_xv)\n\n callback_func = func_partial(validate_in_train, midata=valid_data_xv,\n save_dir=save_dir_xv, verbose=args.verbose)\n model = get_model(args)\n\n train(model, train_data_xv, save_dir=save_dir_xv, validate_callback=callback_func)\n\n if 'scan_data' in action_list:\n\n gwio.dict2json(vars(args), fname='args.json', fdir=args.save_dir)\n args.net_src_file = save_net_pycode(args.net_src_file, args.save_dir)\n\n midata = get_midata(args)\n train_data, valid_data = train_test_split(midata, test_size=args.test_size,\n random_state=args.split_seed)\n\n args.validate_callback = func_partial(validate_in_train, midata=valid_data,\n save_dir=args.save_dir, verbose=args.verbose)\n\n scan_report = scan_data_args(args, train_data,\n valid_data = None,\n data_sizes = args.data_sizes, # [1,2,4], # 'auto',\n batch_sizes = args.batch_sizes, #[1,2,4], # 'auto',\n )\n\n if 'scout_args' in action_list:\n\n if len(args.arg_values) != len(args.arg_names):\n logger.critical('arg names and values must of the same length!')\n\n args.net_src_file = save_net_pycode(args.net_src_file, args.save_dir)\n gwio.dict2json(vars(args), fname='args.json', fdir=args.save_dir)\n\n if args.rebake_midata:\n midata = load_pkldata(args)\n valid_data, train_data = random_split_dict(midata, size=args.test_size)\n else:\n midata = get_midata(args)\n train_data, valid_data = train_test_split(midata, test_size=args.test_size, \n random_state=args.split_seed)\n\n # will be overwritten in scout_args() if args.rebake_midata == True\n args.validate_callback = func_partial(validate_in_train, midata=valid_data,\n save_dir=args.save_dir, verbose=args.verbose)\n\n num_args = len(args.arg_names)\n # arg_values is a list of strings, e.g, ['1,2,3', '4,5,6']\n args.arg_values = [[float(_s) for _s in _v.split(',')] for _v in args.arg_values]\n if isinstance(args.arg_scales, int):\n args.arg_scales = [args.arg_scales] * num_args\n # pad with edge values if needed\n args.arg_scales = np.pad(args.arg_scales, (0, num_args - len(args.arg_scales)), mode='edge')\n\n passed_arg_values = args.arg_values\n if args.grid_search:\n args.arg_values = [np.sort(_v) for _v in args.arg_values]\n else: # generate a list of values\n args.arg_values = []\n for _i, min_max in enumerate(passed_arg_values):\n if args.arg_scales[_i] == 0:\n args.arg_values.append(min_max[0] +\n np.sort(np.random.random_sample(args.num_scouts)) * (min_max[-1] - min_max[0])\n )\n else:\n args.arg_values.append(np.exp(np.log(min_max[0]) +\n np.sort(np.random.random_sample(args.num_scouts)) * np.log(min_max[-1] / min_max[0])\n ))\n\n best_valid_loss = np.inf\n passed_arg_values = args.arg_values\n arg_values = args.arg_values\n master_save_dir = args.save_dir.resolve().as_posix()\n ispawn = 0\n while True:\n if ispawn == 0:\n save_dir = path(master_save_dir)\n else:\n save_dir = path(master_save_dir) / f'spawn_{ispawn}'\n\n scout_best_loss = scout_args(args, train_data, valid_data,\n arg_names = args.arg_names,\n arg_values = arg_values,\n grid_search = args.grid_search,\n save_dir = save_dir,\n # arg_names = ['learning_rate', 'l2decay'],\n # arg_values = [[1e-5, 1e-4, 1e-3, 1e-2], [1e-4, 1e-2]],\n )\n\n if not args.spawn_search: break #!!!! stop here uness spawn_search\n ispawn += 1\n\n # get arg_values giving the best valid_loss\n imin = scout_best_loss.valid_loss.argmin()\n if best_valid_loss < scout_best_loss.valid_loss[imin]:\n logger.info(f'Best spawned args found with valid_loss: {best_valid_loss}')\n break\n else:\n best_valid_loss = scout_best_loss.valid_loss[imin]\n\n # get the arg values giving the best loss\n arg_values_best = scout_best_loss.loc[imin, args.arg_names].to_numpy()\n # mutate from the best arg_values\n arg_values = []\n for _i in range(args.num_spawns):\n arg_values.append(arg_values_best.copy())\n\n iarg = np.random.randint(low=0, high=num_args)\n # make a random change, better stay within the confines of the input\n if args.arg_scales[iarg] == 0: # linear\n arg_values[-1][iarg] = passed_arg_values[iarg][0] + \\\n np.random.rand() * (passed_arg_values[iarg][-1] - passed_arg_values[iarg][0])\n else:\n arg_values[-1][iarg] = np.exp(np.log(passed_arg_values[iarg][0]) + \\\n np.random.rand() * np.log(passed_arg_values[iarg][-1] / passed_arg_values[iarg][0]))\n\n # get ready for the next scout run\n arg_values = list(zip(*arg_values))\n\n if 'average_model' in action_list:\n pkldata = load_pkldata(args)\n\n num_seqs = len(pkldata['seq'])\n num_models = len(args.model_dirs)\n\n valid_loss_models = np.ones((num_models,), dtype=np.float32)\n y_output_models = [] # np.array((num_models, num_data, midata[0][0].shape[0]), dtype=np.float32)\n as_label_models = []\n loss_vs_seq_models = np.zeros((num_models, num_seqs), dtype=np.float32)\n std_vs_seq_models = np.zeros((num_models, num_seqs), dtype=np.float32)\n \n logger.info(f'Averaging models in directories:{args.model_dirs}')\n for imodel, load_dir in enumerate(args.model_dirs):\n logger.info(f'Loading model from directory: {str(load_dir)} ...')\n \n if isinstance(load_dir, str): load_dir = path(load_dir)\n if args.best_earlystop: # use the best earlystop model\n load_dir = list(load_dir.glob('earlystop_*'))\n # directory name contains the validation loss\n loss_values = np.array([float(_dir.name.split('_')[-1]) for _dir in load_dir])\n idx_min = loss_values.argmin()\n load_dir = load_dir[idx_min]\n valid_loss_models[imodel] = loss_values[idx_min]\n logger.info(f'Found best earlystop: {load_dir} with valid loss: {valid_loss_models[imodel]}')\n\n # create and restore the model\n model_args, _ = parse_args2(['average_model'])\n model_args.update(gwio.json2dict(fname='args.json', fdir=load_dir))\n model_args.update(kwargs) # kwargs overwrite the model args!!!\n model_args.load_dir = path(load_dir)\n model = get_model(model_args)\n state_dict_load(model, fdir=load_dir)\n\n midata = bake_midata(pkldata, model_args)\n y_truth = [midata[i][-1] for i in range(num_seqs)]\n\n # compute ymodel, a list of predicted numpy arrays for each\n y_output, seqs_len = predict(model, midata, shuffle=False, batch_size=args.batch_size,\n save_dir=None)\n y_output_models.append(y_output)\n as_label_models.append(model.loss_fn[0].as_label(y_output).numpy())\n\n # compute loss\n if np.array_equal(as_label_models[-1][0].shape, y_truth[0].shape): \n loss_vs_seq, std_vs_seq = compute_loss(model.loss_fn, y_output, y_truth,\n batch_size=args.batch_size, seqs_len=seqs_len, shuffle=False,\n loss_padding=args.loss_padding, loss_sqrt=args.loss_sqrt)\n\n logger.info(f'Model: {imodel} loss: \\033[0;46m{loss_vs_seq.mean():6.4f}\\033[0m' +\n f' std: {std_vs_seq.mean():6.4f}')\n\n loss_vs_seq_models[imodel] = loss_vs_seq\n std_vs_seq_models[imodel] = std_vs_seq\n else:\n loss_vs_seq_models[imodel] = 1\n std_vs_seq_models[imodel] = 1\n\n # zip to get each item to be a list of model predictions for one seqence\n y_output_models = zip(*y_output_models)\n y_output_models = zip(*as_label_models)\n # stack to get one numpy array for each seq, then get the averaged model predictions!\n model_weights = 1.0 / valid_loss_models\n model_weights = model_weights / model_weights.sum()\n if num_seqs > 23:\n mpool = mpi.Pool(processes=int(mpi.cpu_count() * 0.8))\n y_output_models = mpool.map(np.stack, y_output_models)\n\n y_model_aver = mpool.map(func_partial(np.average, axis=0, weights=model_weights),\n y_output_models)\n mpool.close()\n else:\n y_output_models = list(map(np.stack, y_output_models))\n y_model_aver = list(map(func_partial(np.average, axis=0, weights=model_weights),\n y_output_models))\n\n if np.array_equal(y_model_aver[0].shape, y_truth[0].shape):\n loss_model_aver, std_model_aver = compute_loss( #model.loss_fn\n SeqLossFn_P2P(F.mse_loss, name='mse', reduction='none'),\n y_model_aver, y_truth, batch_size=args.batch_size, seqs_len=seqs_len,\n shuffle=False, loss_padding=args.loss_padding, loss_sqrt=args.loss_sqrt)\n\n logger.info(f'Model losses: {loss_vs_seq_models.mean(axis=1)}')\n logger.info(f'Model averaged loss: \\033[0;46m{loss_vs_seq_models.mean():6.4f}\\033[0m' +\n f' std: {std_vs_seq_models.mean():6.4f}')\n logger.info(f'Averaged model loss: \\033[0;46m{loss_model_aver.mean():6.4f}\\033[0m' +\n f' std: {std_model_aver.mean():6.4f}')\n\n # save the average model results\n logger.info(f'Saving the averaged model estimates in: {str(args.save_dir)}')\n\n save_model_prediction(y_model_aver, args.save_dir, seqs_len=seqs_len, istart=1)\n\n if 'validate' in action_list:\n model = get_model(args)\n if args.load_dir:\n state_dict_load(model, fdir=args.load_dir)\n valid_data = get_midata(args)\n rmsd_curve = validate(model, valid_data, shuffle=False, save_dir=args.save_dir)\n\n if 'predict' in action_list:\n model = get_model(args)\n if args.load_dir:\n state_dict_load(model, fdir=args.load_dir)\n predict_data = get_midata(args)\n predict(model, predict_data, shuffle=False, save_dir=args.save_dir / 'predict.files')\n\n\nif __name__ == '__main__' :\n \"\"\" \"\"\"\n sys.setrecursionlimit(int(1e4))\n # if sys.stdin.isatty(): # if running from terminal\n args, argv_dict = parse_args2(sys.argv[1:])\n # else: # run from vscode # os.chdir()\n # args, argv_dict = parse_args(['-h'])\n\n misc.logging_config(logging, logfile=args.log, lineno=False, level=args.verbose)\n args.argv = \" \".join(sys.argv[1:])\n logger.info(f'{path(__file__).name} {args.argv}')\n\n fly(args, **argv_dict)", "sub_path": "work/code/fly_paddle.py", "file_name": "fly_paddle.py", "file_ext": "py", "file_size_in_byte": 90513, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "logging.getLogger", "line_number": 33, "usage_type": "call"}, {"api_name": "argparse.ArgumentParser", "line_number": 39, "usage_type": "call"}, {"api_name": "argparse.RawTextHelpFormatter", "line_number": 41, "usage_type": "attribute"}, {"api_name": "argparse.ArgumentParser", "line_number": 43, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime.now", "line_number": 52, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 52, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 80, "usage_type": "call"}, {"api_name": "paddle_nets.__file__", "line_number": 80, "usage_type": "attribute"}, {"api_name": "misc.unpack_list_tuple", "line_number": 231, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 232, "usage_type": "call"}, {"api_name": "misc.argv_optargs", "line_number": 233, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 242, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 243, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 325, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 325, "usage_type": "attribute"}, {"api_name": "numpy.sort", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.random.choice", "line_number": 345, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.pad", "line_number": 375, "usage_type": "call"}, {"api_name": "numpy.integer", "line_number": 382, "usage_type": "attribute"}, {"api_name": "numpy.pad", "line_number": 398, "usage_type": "call"}, {"api_name": "numpy.expand_dims", "line_number": 421, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 432, "usage_type": "call"}, {"api_name": "numpy.clip", "line_number": 435, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 435, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 442, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 444, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 455, "usage_type": "argument"}, {"api_name": "pathlib.Path", "line_number": 456, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 459, "usage_type": "argument"}, {"api_name": "pickle.load", "line_number": 466, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 473, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 476, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 494, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 512, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 513, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 516, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 516, "usage_type": "attribute"}, {"api_name": "numpy.arange", "line_number": 522, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 522, "usage_type": "attribute"}, {"api_name": "multiprocessing.Pool", "line_number": 533, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 533, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 537, "usage_type": "call"}, {"api_name": "mol_stru.vector_rna_seq", "line_number": 537, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 543, "usage_type": "call"}, {"api_name": "mol_stru.quant_rna_seq", "line_number": 543, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 557, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 568, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 578, "usage_type": "call"}, {"api_name": "mol_stru.ct2mat", "line_number": 584, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 587, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 623, "usage_type": "call"}, {"api_name": "paddle.io.Dataset", "line_number": 630, "usage_type": "name"}, {"api_name": "multiprocessing.cpu_count", "line_number": 648, "usage_type": "call"}, {"api_name": "paddle.io.DataLoader", "line_number": 653, "usage_type": "call"}, {"api_name": "paddle.io", "line_number": 653, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 659, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 660, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 666, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 669, "usage_type": "call"}, {"api_name": "os.path", "line_number": 669, "usage_type": "attribute"}, {"api_name": "paddle_nets.__file__", "line_number": 669, "usage_type": "attribute"}, {"api_name": "sys.path.insert", "line_number": 672, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 672, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 673, "usage_type": "call"}, {"api_name": "paddle_nets.__file__", "line_number": 673, "usage_type": "attribute"}, {"api_name": "importlib.reload", "line_number": 674, "usage_type": "call"}, {"api_name": "importlib.import_module", "line_number": 676, "usage_type": "call"}, {"api_name": "sys.path.remove", "line_number": 677, "usage_type": "call"}, {"api_name": "sys.path", "line_number": 677, "usage_type": "attribute"}, {"api_name": "misc.str_color", "line_number": 681, "usage_type": "call"}, {"api_name": "inspect.getmembers", "line_number": 683, "usage_type": "call"}, {"api_name": "inspect.isclass", "line_number": 683, "usage_type": "argument"}, {"api_name": "misc.get_list_index", "line_number": 685, "usage_type": "call"}, {"api_name": "misc.get_list_index", "line_number": 687, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 704, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 705, "usage_type": "call"}, {"api_name": "paddle.regularizer.L1Decay", "line_number": 721, "usage_type": "call"}, {"api_name": "paddle.regularizer", "line_number": 721, "usage_type": "attribute"}, {"api_name": "paddle.regularizer.L2Decay", "line_number": 723, "usage_type": "call"}, {"api_name": "paddle.regularizer", "line_number": 723, "usage_type": "attribute"}, {"api_name": "paddle.optimizer.lr.ReduceOnPlateau", "line_number": 728, "usage_type": "call"}, {"api_name": "paddle.optimizer", "line_number": 728, "usage_type": "attribute"}, {"api_name": "paddle.optimizer.Adam", "line_number": 735, "usage_type": "call"}, {"api_name": "paddle.optimizer", "line_number": 735, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional.sigmoid", "line_number": 755, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 755, "usage_type": "name"}, {"api_name": "paddle.nn.functional.mse_loss", "line_number": 756, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 756, "usage_type": "name"}, {"api_name": "paddle.nn.functional.softmax", "line_number": 765, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 765, "usage_type": "name"}, {"api_name": "paddle.nn.functional.mse_loss", "line_number": 782, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 782, "usage_type": "name"}, {"api_name": "paddle.unstack", "line_number": 794, "usage_type": "call"}, {"api_name": "paddle.log", "line_number": 802, "usage_type": "call"}, {"api_name": "paddle.exp", "line_number": 802, "usage_type": "call"}, {"api_name": "paddle.nn.Layer", "line_number": 810, "usage_type": "attribute"}, {"api_name": "paddle.nn", "line_number": 810, "usage_type": "name"}, {"api_name": "paddle.nn.functional.mse_loss", "line_number": 817, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 817, "usage_type": "name"}, {"api_name": "paddle.Tensor", "line_number": 828, "usage_type": "attribute"}, {"api_name": "paddle.to_tensor", "line_number": 829, "usage_type": "call"}, {"api_name": "paddle.nn.functional.sigmoid", "line_number": 835, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 835, "usage_type": "name"}, {"api_name": "paddle.unstack", "line_number": 838, "usage_type": "call"}, {"api_name": "paddle.nn.functional.softmax", "line_number": 838, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 838, "usage_type": "name"}, {"api_name": "paddle.nn.functional.softmax", "line_number": 846, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 846, "usage_type": "name"}, {"api_name": "paddle.nn.functional.softmax", "line_number": 848, "usage_type": "call"}, {"api_name": "paddle.nn.functional", "line_number": 848, "usage_type": "name"}, {"api_name": "paddle.argmax", "line_number": 849, "usage_type": "call"}, {"api_name": "paddle.Tensor", "line_number": 871, "usage_type": "attribute"}, {"api_name": "paddle.to_tensor", "line_number": 872, "usage_type": "call"}, {"api_name": "paddle.unsqueeze", "line_number": 875, "usage_type": "call"}, {"api_name": "numpy.zeros_like", "line_number": 897, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 897, "usage_type": "attribute"}, {"api_name": "paddle.squeeze", "line_number": 899, "usage_type": "call"}, {"api_name": "paddle.mean", "line_number": 904, "usage_type": "call"}, {"api_name": "paddle.sqrt", "line_number": 907, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 908, "usage_type": "call"}, {"api_name": "paddle.sum", "line_number": 910, "usage_type": "call"}, {"api_name": "paddle.to_tensor", "line_number": 914, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 916, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 916, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 917, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 917, "usage_type": "attribute"}, {"api_name": "paddle.sqrt", "line_number": 934, "usage_type": "call"}, {"api_name": "paddle.mean", "line_number": 934, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 935, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 936, "usage_type": "call"}, {"api_name": "paddle.mean", "line_number": 938, "usage_type": "call"}, {"api_name": "paddle.nn.functional.mse_loss", "line_number": 964, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 964, "usage_type": "name"}, {"api_name": "paddle.nn.functional.binary_cross_entropy", "line_number": 968, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 968, "usage_type": "name"}, {"api_name": "paddle.nn.functional.cross_entropy", "line_number": 984, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 984, "usage_type": "name"}, {"api_name": "paddle.nn.functional.softmax_with_cross_entropy", "line_number": 988, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 988, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 1013, "usage_type": "call"}, {"api_name": "numpy.savetxt", "line_number": 1035, "usage_type": "call"}, {"api_name": "numpy.save", "line_number": 1037, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1053, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1053, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1054, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1054, "usage_type": "attribute"}, {"api_name": "misc.Struct", "line_number": 1080, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1097, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1114, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1130, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 1135, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1135, "usage_type": "attribute"}, {"api_name": "paddle.to_tensor", "line_number": 1152, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1153, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1153, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1154, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1154, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1187, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1188, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1208, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1245, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1248, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1265, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1273, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1278, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1278, "usage_type": "attribute"}, {"api_name": "paddle.no_grad", "line_number": 1289, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 1305, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1305, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1306, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1306, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 1324, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1325, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1329, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1341, "usage_type": "call"}, {"api_name": "pathlib.Path.cwd", "line_number": 1345, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1345, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 1349, "usage_type": "argument"}, {"api_name": "pathlib.Path", "line_number": 1350, "usage_type": "call"}, {"api_name": "numpy.empty", "line_number": 1370, "usage_type": "call"}, {"api_name": "numpy.int32", "line_number": 1370, "usage_type": "attribute"}, {"api_name": "paddle.no_grad", "line_number": 1373, "usage_type": "call"}, {"api_name": "tqdm.tqdm", "line_number": 1373, "usage_type": "call"}, {"api_name": "pathlib.Path.cwd", "line_number": 1400, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1400, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 1402, "usage_type": "call"}, {"api_name": "paddle.save", "line_number": 1408, "usage_type": "call"}, {"api_name": "paddle.save", "line_number": 1409, "usage_type": "call"}, {"api_name": "pathlib.Path.cwd", "line_number": 1415, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1415, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 1417, "usage_type": "call"}, {"api_name": "paddle.load", "line_number": 1425, "usage_type": "call"}, {"api_name": "paddle.load", "line_number": 1433, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1442, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1450, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1452, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1460, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1474, "usage_type": "call"}, {"api_name": "misc.logger_setlevel", "line_number": 1476, "usage_type": "call"}, {"api_name": "misc.logger_setlevel", "line_number": 1478, "usage_type": "call"}, {"api_name": "gwio.next_backup_path", "line_number": 1495, "usage_type": "call"}, {"api_name": "gwio.dict2json", "line_number": 1499, "usage_type": "call"}, {"api_name": "shutil.rmtree", "line_number": 1510, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1520, "usage_type": "call"}, {"api_name": "numpy.log2", "line_number": 1531, "usage_type": "call"}, {"api_name": "numpy.logspace", "line_number": 1532, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1540, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1546, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1551, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 1552, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 1558, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1558, "usage_type": "attribute"}, {"api_name": "pandas.concat", "line_number": 1588, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1592, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1609, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1613, "usage_type": "call"}, {"api_name": "itertools.product", "line_number": 1619, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1635, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1658, "usage_type": "call"}, {"api_name": "pandas.concat", "line_number": 1662, "usage_type": "call"}, {"api_name": "misc.Struct", "line_number": 1676, "usage_type": "attribute"}, {"api_name": "misc.Struct", "line_number": 1677, "usage_type": "call"}, {"api_name": "pathlib.Path.cwd", "line_number": 1680, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1680, "usage_type": "name"}, {"api_name": "gwio.json2dict", "line_number": 1683, "usage_type": "call"}, {"api_name": "gwio.json_str", "line_number": 1685, "usage_type": "call"}, {"api_name": "gwio.json_str", "line_number": 1690, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1700, "usage_type": "call"}, {"api_name": "sys.exit", "line_number": 1704, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 1705, "usage_type": "attribute"}, {"api_name": "gwio.last_backup_path", "line_number": 1707, "usage_type": "call"}, {"api_name": "gwio.json2dict", "line_number": 1712, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1714, "usage_type": "call"}, {"api_name": "gwio.next_backup_path", "line_number": 1720, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1721, "usage_type": "call"}, {"api_name": "gwio.dict2json", "line_number": 1725, "usage_type": "call"}, {"api_name": "pathlib.Path.cwd", "line_number": 1725, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1725, "usage_type": "name"}, {"api_name": "gwio.dict2json", "line_number": 1736, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 1739, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1742, "usage_type": "call"}, {"api_name": "gwio.dict2json", "line_number": 1755, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 1758, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1762, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 1769, "usage_type": "call"}, {"api_name": "numpy.random.permutation", "line_number": 1770, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1770, "usage_type": "attribute"}, {"api_name": "itertools.chain", "line_number": 1778, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1783, "usage_type": "call"}, {"api_name": "gwio.dict2json", "line_number": 1791, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 1795, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1798, "usage_type": "call"}, {"api_name": "gwio.dict2json", "line_number": 1813, "usage_type": "call"}, {"api_name": "sklearn.model_selection.train_test_split", "line_number": 1820, "usage_type": "call"}, {"api_name": "functools.partial", "line_number": 1824, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 1833, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 1837, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 1843, "usage_type": "call"}, {"api_name": "numpy.random.random_sample", "line_number": 1843, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1843, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 1846, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 1846, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 1847, "usage_type": "call"}, {"api_name": "numpy.random.random_sample", "line_number": 1847, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1847, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 1847, "usage_type": "call"}, {"api_name": "numpy.inf", "line_number": 1850, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 1857, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1859, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 1888, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1888, "usage_type": "attribute"}, {"api_name": "numpy.random.rand", "line_number": 1892, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1892, "usage_type": "attribute"}, {"api_name": "numpy.exp", "line_number": 1894, "usage_type": "call"}, {"api_name": "numpy.log", "line_number": 1894, "usage_type": "call"}, {"api_name": "numpy.random.rand", "line_number": 1895, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 1895, "usage_type": "attribute"}, {"api_name": "numpy.log", "line_number": 1895, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 1906, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1906, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1909, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1909, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 1910, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 1910, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 1916, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1920, "usage_type": "call"}, {"api_name": "gwio.json2dict", "line_number": 1928, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 1930, "usage_type": "call"}, {"api_name": "numpy.array_equal", "line_number": 1944, "usage_type": "call"}, {"api_name": "multiprocessing.Pool", "line_number": 1965, "usage_type": "call"}, {"api_name": "multiprocessing.cpu_count", "line_number": 1965, "usage_type": "call"}, {"api_name": "numpy.stack", "line_number": 1966, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 1968, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 1968, "usage_type": "attribute"}, {"api_name": "numpy.stack", "line_number": 1972, "usage_type": "attribute"}, {"api_name": "functools.partial", "line_number": 1973, "usage_type": "call"}, {"api_name": "numpy.average", "line_number": 1973, "usage_type": "attribute"}, {"api_name": "numpy.array_equal", "line_number": 1976, "usage_type": "call"}, {"api_name": "paddle.nn.functional.mse_loss", "line_number": 1978, "usage_type": "attribute"}, {"api_name": "paddle.nn.functional", "line_number": 1978, "usage_type": "name"}, {"api_name": "sys.setrecursionlimit", "line_number": 2010, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 2012, "usage_type": "attribute"}, {"api_name": "misc.logging_config", "line_number": 2016, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 2017, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 2018, "usage_type": "call"}]} +{"seq_id": "72855408", "text": "import sys\nimport os\nimport numpy as np\nfrom scipy import interpolate\nfrom scipy.ndimage import morphology\nfrom scipy.stats import multivariate_normal\nfrom PIL import Image, ImageDraw\nfrom cv2 import GaussianBlur, blur, getPerspectiveTransform, warpPerspective\n\nfrom hashlib import blake2s\n\nfrom deconvolution import Deconvolution\nimport deconvolution.pixeloperations as po\n\ndef rand_spline(dim, inPts = None, nPts = 5, random_seed = None, startEdge = True, endEdge = True):\n # splXY = rand_spline(dim, inPts= None, nPts = 5, random_seed =None, startEdge = True, endEdge = True)\n # builds a randomized spline from a set of randomized handle points\n # \n # ###\n # Inputs: Required\n # dim: a 2 element vector Width by Height\n # Inputs: Optional\n # inPts: n x 2 numpy arr Used to prespecify the handle points of the spline\n # note: this is not random\n # nPts: int The number of random handle points in the spline\n # random_seed: int The random seed for numpy for consistent generation\n # startEdge: bool Whether or not the start of the spline should be on the edge of the image\n # int(0,1,2,3) If startEdge is an int, it specifies which edge the spline starts on\n # 0 = Left, 1 = Top, 2 = Right, 3 = Bottom\n # endEdge: bool Whether or not the start of the spline should be on the edge of the image\n # int(0,1,2,3) If endEdge is a nonnegative int, it specifies which edge the spline stops on\n # 0 = Left, 1 = Top, 2 = Right, 3 = Bottom\n # int(-4,-3,-2,-1) If endEdge is a negative int, it specifies which edge the spline stops on \n # relative to the start\n # -4 = Same, -3 = End is 1 step clockwise (e.g. Bottom -> Left)\n # -2 = Opposite side, -1 = End is 1 step counterclockwise (e.g. Bottom -> Right)\n # ###\n # Output:\n # splXY: m x 2 numpy array Spline array sampled at a 1-pixel interval (distance between m points is ~1px)\n \n np.random.seed(seed=random_seed)\n\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n if inPts is None:\n inPts = np.concatenate((np.random.randint((dim[0]-1),size=(nPts,1)),\n np.random.randint((dim[1]-1),size=(nPts,1))),\n axis=1)\n \n startEdgeFlag = (startEdge == True) or (startEdge in range(0,4))\n if startEdgeFlag == True:\n if (startEdge in range(0,4)) and (type(startEdge)!=bool): # allow for manual specification of edge\n edgeNum = startEdge\n else:\n edgeNum = np.random.randint(4)\n LR_v_TB = edgeNum % 2 # left/right vs top/bottom\n LT_V_RB = edgeNum // 2 # left/top vs right/bottom\n \n inPts[0,LR_v_TB] = LT_V_RB * (dim[LR_v_TB]-1) # one edge or the other\n if endEdge == True or (endEdge in range(0,4)) or (endEdge in range(-4,0) and startEdgeFlag):\n if (endEdge in range(0,4)): # allow for manual specification of edge\n edgeNum = endEdge\n elif(endEdge in range(-4,0) and startEdgeFlag): \n # allow for relative specification of end edge compared to the start edge\n # -2 is opposite side, -4 is the same side\n edgeNum = ((endEdge + edgeNum + 4) % 4)\n else:\n edgeNum = np.random.randint(4)\n LR_v_TB = edgeNum % 2 # left/right vs top/bottom\n LT_V_RB = edgeNum // 2 # left/top vs right/bottom \n \n inPts[nPts-1,LR_v_TB] = LT_V_RB * (dim[LR_v_TB]-1) # one edge or the other\n \n else:\n if isinstance(inPts,list):\n inPts = np.array(inPts)\n nPts = inPts.shape[0]\n\n distXY = np.sqrt(np.sum(np.diff(inPts,axis=0)**2,axis=1))\n cdXY = np.concatenate((np.zeros((1)),np.cumsum(distXY)),axis=0)\n iDist = np.arange(np.floor(cdXY[-1])+1)\n splXY = interpolate.pchip_interpolate(cdXY,inPts,iDist)\n return splXY\n\ndef rand_gauss(dim, nNorms = 25, maxCov = 50, random_seed = None,centXY = None, zeroToOne = False,\n minMaxX = None, minMaxY = None, minCovScale = .1,minDiagCovScale = .25, maxCrCovScale = .7):\n # sumMap = rand_gauss(dim, nNorms = 25, maxCov = 50, random_seed = None,centXY = None, zeroToOne = False,\n # minMaxX = None, minMaxY = None, minCovScale = .1,minDiagCovScale = .25, maxCrCovScale = .7):\n # Builds a set of randomized Gaussians within a set range of properties, and adds them together\n #\n # ###\n # Inputs: Required\n # dim: a 2 element vector Width by Height\n # Inputs: Optional\n # nNorms: int The number of Gaussian distributions to generate\n # maxCov: float (+) The maximum covariance of each Gaussian\n # - Related to the size of each Gaussian distribution in pixels\n # random_seed: int The random seed for numpy for consistent generation\n # centXY: n x 2 float arr The locations of the Gaussians can be specified manually if desired, instead of randomly\n # zeroToOne: bool Whether or not the coordinates are scaled zeroToOne (requires retuning the other sizes)\n # minMaxX: 2 float vector The range of where the gaussians are generated in the image in the X dimension\n # - Defaults to the range of the image, could be off screen if desired\n # minMaxY: 2 float vector The range of where the gaussians are generated in the image in the Y dimension\n # - Defaults to the range of the image, could be off screen if desired\n # minCovScale: float The minimum on the range of sizes across the set of distributions\n # Recommend >0 & ≤1\n # minDiagCovScale: float Affects the diagonal of the covariance matrix, and the minimum relative size \n # Recommend >0 & ≤1 of the two components compared to the scaled max covariance\n # maxCrCovScale: float The maximum relative cross covariance (1 = straight line, 0 = uncorrelated)\n # Recommend >0 & ≤1 Affects the shape of the distributions, a higher number means more eccentricity\n # ###\n # Output:\n # sumMap: numpy array (dim) Creates a numpy array of the input size, where all the scaled Gaussian \n # distributions have been added together\n \n \n np.random.seed(seed=random_seed)\n invDim = (dim[1],dim[0])\n \n if zeroToOne == True:\n xV = np.linspace(0,1,num= dim[0])\n yV = np.linspace(0,1,num= dim[1])\n if minMaxX is None:\n minMaxX = [0,1]\n if minMaxY is None:\n minMaxY = [0,1]\n else:\n xV = np.arange(dim[0])\n yV = np.arange(dim[1])\n \n if minMaxX is None:\n minMaxX = [0,dim[0]]\n if minMaxY is None:\n minMaxY = [0,dim[1]]\n\n xM, yM = np.meshgrid(xV,yV)\n pos = np.dstack((xM, yM))\n \n if centXY is None:\n centXY = np.concatenate((np.random.uniform(minMaxX[0],minMaxX[1],size=(nNorms,1)),\n np.random.uniform(minMaxY[0],minMaxY[1],size=(nNorms,1))), axis=1)\n else:\n nNorms = centXY.shape[0]\n \n sumMap = np.zeros(invDim)\n for i in range(nNorms):\n cent = centXY[i,:]\n cov = np.zeros((2,2))\n # need to make a symmetric positive semidefinite covariance matrix\n cMaxCov = np.random.uniform(maxCov* minCovScale,maxCov,size=(1,1))\n cov = np.diag(np.random.uniform(cMaxCov * minDiagCovScale,cMaxCov,size=(1,2)).flatten())\n maxCrCov = np.sqrt(np.product(np.diag(cov)))\n cov[[0,1],[1,0]] = np.random.uniform(-maxCrCov* maxCrCovScale,maxCrCov*maxCrCovScale) \n rv = multivariate_normal(cent.flatten(), cov)\n \n sumMap += (rv.pdf(pos) * (cMaxCov * 2 * np.pi))\n\n sumMap = sumMap * maxCov\n return sumMap\n\ndef add_marker(inputIm,random_seed = None,nPts = 3, sampSpl = None, inPts = None, \n width = 100, alpha = .75, rgbVal= None,\n rgbRange = np.array([[0,50],[0,50],[0,100]])):\n # comp_im = add_marker(inputIm,random_seed = None,nPts = 3, sampSpl = None, inPts = None, \n # width = 100, alpha = .75, rgbVal= None,\n # rgbRange = np.array([[0,50],[0,50],[0,100]])):\n # adds a marker line onto the image of a fixed width and color\n #\n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # random_seed: int The random seed for numpy for consistent generation\n # nPts: int The number of random handle points in the spline\n # sampSpl: n x 2 numpy arr You can optionally specify the sampled spline (non-random)\n # - Note: should be sampled densely enough (i.e. at least every pixel)\n # inPts: n x 2 numpy arr Used to prespecify the handle points of the spline\n # - Note: this is not random\n # width: float (+) The width of the marker line, in pixels\n # alpha: float (0-1) The alpha transparency of the marker layer (1 = opaque, 0 = transparent)\n # rgbVal: 3 uint8 vector The RGB color of the marker can be optionally specified\n # >=0 <=255\n # rgbRange: 3 x 2 uint8 arr The RGB range of the randomized color [[minR,maxR],[minG,maxG],[minB,maxB]]\n # >=0 <=255 - Leans more blue heavy by default\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with the marker layer on top of the original image\n \n \n np.random.seed(seed=random_seed)\n if rgbVal is None:\n rgbVal = np.zeros((3,1))\n for i in range(3):\n rgbVal[i] = np.random.randint(rgbRange[i,0],rgbRange[i,1])\n rgbVal = rgbVal.flatten()\n\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n if sampSpl is None:\n if inPts is None:\n sampSpl = rand_spline(dim, nPts = nPts,random_seed = random_seed)\n else:\n sampSpl = rand_spline(dim, inPts = inPts,random_seed = random_seed)\n \n mask = np.ones(invDim)\n mask[(np.round(sampSpl[:,1])).astype(int),np.round(sampSpl[:,0]).astype(int)] = 0\n # create a distance map to the points on the spline\n bwDist = morphology.distance_transform_edt(mask)\n\n # use the distance map to build a fixed width region\n bwReg = bwDist <= width/2\n im_rgba = inputIm.convert(\"RGBA\")\n # build up the semi-transparent colored layer\n alpha_mask = Image.fromarray((bwReg*alpha*255).astype(np.uint8),'L')\n color_arr = np.zeros((invDim[0],invDim[1],3),dtype=np.uint8)\n for i in range(len(rgbVal)):\n color_arr[:,:,i] = rgbVal[i]\n color_layer = Image.fromarray(color_arr,'RGB')\n\n comp_im = Image.composite(color_layer, im_rgba, alpha_mask)\n comp_im = comp_im.convert(\"RGB\")\n return comp_im\n\n\ndef add_fold(inputIm, sampArr = None, sampSpl=None, inPts = None, random_seed =None, scaleXY =[1,1], width = 200,\n sampShiftXY = None, randEdge=False, nLayers = 2, nPts = 3,endEdge = -2):\n # comp_im = add_fold(inputIm,samp_arr =None, sampSpl=None, inPts = None,random_seed =None,scaleXY =[1,1], width = 200,\n # sampShiftXY = None,randEdge=False, nLayers = 2, nPts = 3,endEdge = -2):\n # adds a tissue fold to the input image along a spline path\n # based on sampling from the input image\n #\n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # sampArr: numpy arr A numpy array the size of the input image\n # - used for recursion - if the input image is not where the tissue should be sampled from\n # sampSpl: n x 2 numpy arr You can optionally specify the sampled spline (non-random)\n # - used for recursion \n # inPts: n x 2 numpy arr Used to prespecify the handle points of the spline\n # - Note: this is not random\n # random_seed: int The random seed for numpy for consistent generation\n # scaleXY: 2 float vector Used to scale the sampling bounding box, if the sample region should be resized\n # Defaults to no change between original and sampling\n # large scale = larger sample region\n # width: float (+) The width of the tissue fold region, in pixels\n # sampShiftXY: 2 int vec You can optionally specify the direction to shift the spline region\n # Defaults to a random direction at most half the size of the image\n # randEdge: bool Whether to add some randomness to the edge of the tissue fold region\n # Defaults to off\n # nLayers: int (+) Number of tissue layers to add to the image\n # Runs the function recursively, defaults to 2 layers\n # nPts: int (+) The number of random handle points in the spline\n # endEdge: bool Whether or not the start of the spline should be on the edge of the image\n # int(0,1,2,3) If endEdge is a nonnegative int, it specifies which edge the spline stops on\n # 0 = Left, 1 = Top, 2 = Right, 3 = Bottom\n # int(-4,-3,-2,-1) If endEdge is a negative int, it specifies which edge the spline stops on \n # relative to the start\n # -4 = Same, -3 = End is 1 step clockwise (e.g. Bottom -> Left)\n # -2 = Opposite side, -1 = End is 1 step counterclockwise (e.g. Bottom -> Right)\n # Defaults to -2\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with the tissue fold layers on top of the original image\n \n np.random.seed(seed=random_seed)\n if nLayers < 1: # if someone handed in an invalid # of layers, return back the original image\n return inputIm\n \n im_arr = np.array(inputIm)\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n\n if sampSpl is None:\n if inPts is None:\n sampSpl = rand_spline(dim, nPts = nPts,random_seed = random_seed, endEdge = endEdge)\n else:\n sampSpl = rand_spline(dim, inPts = inPts,random_seed = random_seed)\n \n if sampArr is None:\n sampArr = np.copy(im_arr)\n \n if sampShiftXY is None: # randomly initialized if empty\n shiftXY = np.random.randint(-int(dim[0]/2),int(dim[0]/2),size=(2,1))\n else:\n shiftXY = sampShiftXY\n \n\n pad_szXY = (max(dim),max(dim),0) # pad x, pad y, no pad z (have to reshape for np.pad, which takes y,x,z)\n sampBlur = (((width//40)*2)+1,((width//40)*2)+1) # has to be odd kernel\n \n # pad the array to allow for sampling, mirror tiles outside of range\n pad_amt = np.transpose(np.tile(np.array(pad_szXY)[[1,0,2]],(2,1)))\n sampPadArr = np.pad(sampArr,pad_amt,mode='symmetric')\n\n sampSplBBox = np.vstack((np.amin(sampSpl,axis=0),np.amax(sampSpl,axis=0)))\n sampSplBBSz = np.diff(sampSplBBox,axis=0)\n rsSplBBox = np.zeros((2,2))\n\n # build up the bounding box of the region to be sampled from\n signTup = (-1,1)\n for di in range(2):\n rsSplBBox[di,:] = np.mean(sampSplBBox,axis=0) + (((sampSplBBSz/2) * scaleXY) * signTup[di])\n \n rsSplBBSz = np.diff(rsSplBBox,axis=0)\n\n # allow for a random shift to the sampling region, to each of the corners of the sample region\n sampSplBBPts = np.zeros((4,2),dtype=np.float32)\n outBBPts = np.zeros((4,2),dtype=np.float32)\n # maximum change is ± 1/4 of the size of the sampling bounding box \n randShiftX = np.random.randint(-int(rsSplBBSz[0,0]/4),int(rsSplBBSz[0,0]/4),size=(4,1))\n randShiftY = np.random.randint(-int(rsSplBBSz[0,1]/4),int(rsSplBBSz[0,1]/4),size=(4,1))\n \n for di in range(sampSplBBPts.shape[0]): \n LR_v_TB = di % 2 # left/right vs top/bottom\n LT_V_RB = di // 2 # left/top vs right/bottom\n sampSplBBPts[di,0] = sampSplBBox[LR_v_TB,0]\n sampSplBBPts[di,1] = sampSplBBox[LT_V_RB,1]\n outBBPts[di,0] = rsSplBBox[LR_v_TB,0] + pad_szXY[0] + shiftXY[0] + randShiftX[di]\n outBBPts[di,1] = rsSplBBox[LT_V_RB,1] + pad_szXY[1] + shiftXY[1] + randShiftY[di]\n\n # generate mapping matrix from original bounding box to the sampled bounding box\n M = getPerspectiveTransform(outBBPts,sampSplBBPts)\n # warp the padded array based on this transform\n warp_im = warpPerspective(sampPadArr,M,dim)\n\n # find the distance to the spline\n mask = np.ones(invDim)\n mask[(sampSpl[:,1].astype(int)),sampSpl[:,0].astype(int)] = 0\n bwDist = morphology.distance_transform_edt(mask)\n if randEdge == True:\n distRand = np.random.randint(-int(width/4),int(width/4),size=invDim)\n bwDist = blur(bwDist+distRand,(5,5))\n \n im_L = inputIm.convert(\"L\")\n im_L_arr = np.array(im_L)\n bwReg = bwDist <= width/2\n \n\n # multiplicative combination. Makes things darker\n unit_dst_arr = np.ones(warp_im.shape)\n for i in range(warp_im.shape[2]):\n unit_dst_arr[:,:,i] = np.where(bwReg,warp_im[:,:,i]/255,1)\n unit_dst_arr = GaussianBlur(unit_dst_arr,sampBlur,0)\n comp_arr = unit_dst_arr * im_arr\n comp_im = Image.fromarray(comp_arr.astype(np.uint8),'RGB')\n \n if nLayers > 1: # recursive addition\n comp_im = add_fold(comp_im,sampArr=sampArr, sampSpl=sampSpl,inPts=inPts,random_seed = random_seed+1,\n scaleXY=scaleXY,width=width,sampShiftXY=sampShiftXY,randEdge=randEdge,\n nLayers=nLayers-1)\n return comp_im\n\ndef add_sectioning(inputIm, width = 240, random_seed = None, scaleMin = .5, scaleMax = .8, randEdge = True,\n sampSpl = None, inPts = None, nPts = 2, endEdge = -2):\n # comp_im = add_sectioning(inputIm, sliceWidth = 120, random_seed = None, scaleMin = .5, scaleMax = .8, randEdge = True,\n # sampSpl = None, inPts = None, nPts = 2, endEdge = -2):\n # Add a region of uneven (thinner) sectioning due to different thicknesses of slide\n # Saturation of the region is decreased by a randomized factor within a range\n # Value of the region is increased by half of the percentage change of the saturation\n #\n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # width: float The width of the sectioning region, in pixels\n # random_seed: int The random seed for numpy for consistent generation\n # scaleMin: float The minimum level of saturation allowed at random\n # -Note: this scales based off of the distance from the spline\n # scaleMax: float The maximum level of saturtation allowed at random\n # -Note: this scales based off of the distance from the spline\n # randEdge: bool Whether to add some randomness to the edge of the sectioning region\n # Defaults to on\n # sampSpl: n x 2 numpy arr You can optionally specify the sampled spline (non-random)\n # - used for recursion \n # inPts: n x 2 numpy arr Used to prespecify the handle points of the spline\n # - Note: this is not random\n # nPts: int The number of random handle points in the spline\n # Defaults to 2\n # endEdge: bool Whether or not the start of the spline should be on the edge of the image\n # int(0,1,2,3) If endEdge is a nonnegative int, it specifies which edge the spline stops on\n # 0 = Left, 1 = Top, 2 = Right, 3 = Bottom\n # int(-4,-3,-2,-1) If endEdge is a negative int, it specifies which edge the spline stops on \n # relative to the start\n # -4 = Same, -3 = End is 1 step clockwise (e.g. Bottom -> Left)\n # -2 = Opposite side, -1 = End is 1 step counterclockwise (e.g. Bottom -> Right)\n # Defaults to -2\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with the sectioning artifact applied to the original image\n \n np.random.seed(seed=random_seed)\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n \n if sampSpl is None:\n if inPts is None:\n sampSpl = rand_spline(dim, inPts = inPts,random_seed = random_seed)\n else:\n sampSpl = rand_spline(dim, nPts = nPts, endEdge = endEdge, random_seed = random_seed)\n\n mask = np.ones(invDim)\n mask[(sampSpl[:,1].astype(int)),sampSpl[:,0].astype(int)] = 0\n bw_dist = morphology.distance_transform_edt(mask)\n if randEdge == True:\n distRand = np.random.randint(-int(width/2),int(width/2),size=invDim)\n bw_dist = blur(bw_dist+distRand,(5,5))\n\n # scale the distance map from the randomized min to max, sectioning effect is stronger in the center\n bw_reg = bw_dist <= width/2\n nDistRng = bw_dist / (width/2)\n halfScale = (scaleMin + scaleMax)/2\n scaleRandMin = np.random.uniform(scaleMin,(halfScale+scaleMin)/2,size=(1,1))\n scaleRandMax = np.random.uniform((halfScale+scaleMax)/2,scaleMax,size=(1,1))\n scaleRMinMax = np.concatenate((scaleRandMin,scaleRandMax),axis = 1).flatten()\n\n nDistRng = np.interp(nDistRng,np.array([0,1],dtype=np.float64),scaleRMinMax)\n\n nDistRng[np.logical_not(bw_reg)] = 1\n\n imHSV = inputIm.convert(\"HSV\")\n imHSV_arr = np.array(imHSV)\n # increase the lightness by half the factor of decreased saturation\n imHSV_arr[:,:,1] = np.minimum(255,np.multiply(imHSV_arr[:,:,1],nDistRng))\n imHSV_arr[:,:,2] = np.minimum(255,np.divide(imHSV_arr[:,:,2],(nDistRng+1)/2))\n # minimum function is to stop integer overflow\n imSatHSV = Image.fromarray(imHSV_arr,\"HSV\")\n comp_im = imSatHSV.convert(\"RGB\")\n return comp_im\n\ndef add_bubbles(inputIm,random_seed = None,nBubbles = 25, maxWidth = 50,alpha = .75, edgeWidth = 2,\n edgeColorMult = (.75,.75,.75), rgbVal = (225,225,225)):\n # comp_im = add_bubbles(inputIm,random_seed = None,nBubbles = 25, maxWidth = 50,alpha = .75, edgeWidth = 2,\n # edgeColorMult = (.75,.75,.75), rgbVal = (225,225,225)):\n # adds bubbles in the mold of nuclear bubbling randomly throughout the image\n # \n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # random_seed: int The random seed for numpy for consistent generation\n # nBubbles: int (+) The number of bubbles to generate in the image\n # maxWidth: float (+) The maximum width of the randomized bubbles (roughly), in pixels\n # alpha: float (0-1) The alpha transparency of the bubble layer (1 = opaque, 0 = transparent)\n # edgeWidth: float (+) The width of the darker edge of the bubble, in pixels\n # edgeColorMult: The RGB multiplier of the edge of the bubble \n # 3 float vector -Relative to the mean RGB color of the image\n # rgbVal: 3 float vector The RGB color of the bubbles\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with the bubbles added to the original image\n \n np.random.seed(seed=random_seed)\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n \n # use the randomized gaussian function\n sumMap = rand_gauss(dim,random_seed = random_seed, nNorms=nBubbles, maxCov = maxWidth, zeroToOne = False,\n minCovScale = .1,minDiagCovScale = .25, maxCrCovScale = .7)\n \n bwReg = sumMap >= 1\n bwDist = morphology.distance_transform_edt(bwReg)\n edgeArea = np.logical_and(bwDist <= edgeWidth,bwReg)\n\n alphaMask = Image.fromarray((bwReg*alpha*255).astype(np.uint8),'L')\n colorArr = np.zeros((invDim[0],invDim[1],3),dtype=np.uint8)\n\n # set the colors for the bubbles & edges\n meanColor = np.mean(np.array(inputIm),axis=(0,1))\n for i in range(len(rgbVal)):\n colorArr[:,:,i] = rgbVal[i]\n colorArr[edgeArea,i] = np.uint8(meanColor[i] * edgeColorMult[i])\n\n color_layer = Image.fromarray(colorArr,'RGB')\n comp_im = Image.composite(color_layer, inputIm, alphaMask)\n return comp_im\n\ndef add_illumination(inputIm,random_seed = None, maxCov = 15, nNorms = 3,scaleMin = .8,scaleMax = 1.1,\n minCovScale = .5,minDiagCovScale = .1, maxCrCovScale = .2):\n # comp_im = add_illumination(inputIm,random_seed = None, maxCov = 15, nNorms = 3,scaleMin = .8,scaleMax = 1.1,\n # minCovScale = .5,minDiagCovScale = .1, maxCrCovScale = .2):\n # add uneven illumination artifact to the input image\n # \n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # random_seed: int The random seed for numpy for consistent generation\n # maxCov: float (+) The maximum covariance (governs the size of the distributions)\n # nNorms: int (+) The number of Gaussian distributions used to build the uneven illumination\n # scaleMin: float (<1) The minimum for the random factor used to adjust the illumination\n # scaleMax: float (>1) The maximum for the random factor used to adjust the illumination\n # minCovScale: float The minimum on the range of sizes across the set of distributions\n # Recommend >0 & ≤1\n # minDiagCovScale: float Affects the diagonal of the covariance matrix, and the minimum relative size \n # Recommend >0 & ≤1 of the two components compared to the scaled max covariance\n # maxCrCovScale: float The maximum relative cross covariance (1 = straight line, 0 = uncorrelated)\n # Recommend >0 & ≤1 Affects the shape of the distributions, a higher number means more eccentricity\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with the uneven illumination added to the original image\n \n np.random.seed(seed=random_seed)\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0])\n\n xV = np.linspace(0,1,num= dim[0])\n yV = np.linspace(0,1,num= dim[1])\n xM, yM = np.meshgrid(xV,yV)\n pos = np.dstack((xM, yM))\n \n sumMap = rand_gauss(dim,random_seed = random_seed, nNorms=nNorms, maxCov = maxCov, \n zeroToOne = True, minMaxX = [-.5,1.5],minMaxY = [-.5,1.5],\n minCovScale = minCovScale,minDiagCovScale = minDiagCovScale, maxCrCovScale = maxCrCovScale)\n\n nSumMap = (sumMap - np.amin(sumMap,axis=(0,1)))\n \n divFac = np.amax(nSumMap,axis=(0,1))\n nSumMap = nSumMap /divFac\n \n scaleRandMin = np.random.uniform(scaleMin,(1+scaleMin)/2,size=(1,1))\n scaleRandMax = np.random.uniform((1+scaleMax)/2,scaleMax,size=(1,1))\n scaleRMinMax = np.concatenate((scaleRandMin,scaleRandMax),axis = 1).flatten()\n nSumMap = np.interp(nSumMap,np.array([0,1],dtype=np.float64),scaleRMinMax)\n \n imHSV = inputIm.convert(\"HSV\")\n imHSV_arr = np.array(imHSV)\n imHSV_arr[:,:,2] = np.minimum(255,np.multiply(imHSV_arr[:,:,2],nSumMap))\n imLumHSV = Image.fromarray(imHSV_arr,\"HSV\")\n comp_im = imLumHSV.convert(\"RGB\")\n return comp_im\n\ndef adjust_stain(inputIm,adjFactor = [1,1,1]):\n # (rgbOut,rgb1,rgb2,rgb3) = adjust_stain(inputIm,adjFactor = [1,1,1])\n # adjust the stain levels of the H&E image\n # based on the Deconvolution package: \n # https://deconvolution.readthedocs.io/en/latest/readme.html#two-stain-deconvolution \n #\n # ###\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # adjFactor: 3 float vec The adjustment factor for each of the three basis vectors \n # (<1 = less stain, 1 = same, >1 = more stain)\n # Element 1: Eosin\n # Element 2: Hematoxylin\n # Element 3: Null (the remaining structure)\n # ###\n # Outputs:\n # rgbOut: m x n x 3 array A 2D RGB image (H&E) with the stain levels adjusted\n # rgb1: m x n x 3 numpy arr A 2D RGB image of the Eosin layer only\n # rgb2: m x n x 3 numpy arr A 2D RGB image of the Hematoxylin layer only\n # rgb3: m x n x 3 numpy arr A 2D RGB image of the null layer only\n \n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0]) # have to invert the size dim because rows cols is yx vs xy\n iDimRGB = (invDim[0],invDim[1],3)\n stain_dict = {'eosin':[0.91, 0.38, 0.71], 'null': [0.0, 0.0, 0.0],\n 'hematoxylin': [0.39, 0.47, 0.85]}\n \n ## https://deconvolution.readthedocs.io/en/latest/readme.html#two-stain-deconvolution\n# dec = Deconvolution(image=inputIm, basis=[[0.91, 0.38, 0.71], [0.39, 0.47, 0.85],[0.0, 0.0, 0.0]])\n dec = Deconvolution(image=inputIm, basis=[stain_dict['eosin'], stain_dict['hematoxylin'],stain_dict['null']])\n\n ## this section is extracted from the deconvolution package, but adjusted to allow for altering the stain levels\n pxO= dec.pixel_operations\n _white255 = np.array([255, 255, 255], dtype=float)\n \n v, u, w = pxO.get_basis()\n vf, uf, wf = np.zeros(iDimRGB), np.zeros(iDimRGB), np.zeros(iDimRGB)\n vf[:], uf[:], wf[:] = v, u, w\n \n # Produce density matrices for both colors + null. Be aware, as Beer's law do not always hold.\n a, b, c = map(po._array_positive, dec.out_scalars())\n af = np.repeat(a, 3).reshape(iDimRGB) * adjFactor[0] # Adjusting the exponential coefficient\n bf = np.repeat(b, 3).reshape(iDimRGB) * adjFactor[1] # For the different stain components\n cf = np.repeat(c, 3).reshape(iDimRGB) * adjFactor[2]\n\n # exponential map, for changing stain levels into RGB\n rgbOut = po._array_to_colour_255(_white255 * (vf ** af) * (uf ** bf) * (wf ** cf))\n rgb1 = po._array_to_colour_255(_white255 * (vf ** af))\n rgb2 = po._array_to_colour_255(_white255 * (uf ** bf))\n rgb3 = po._array_to_colour_255(_white255 * (wf ** cf))\n \n return rgbOut,rgb1,rgb2,rgb3\n\ndef add_stain(inputIm,adjFactor = None,scaleMax = [3,3,1.5], scaleMin = [1.25,1.25,1],random_seed = None):\n # comp_im = add_stain(inputIm,adjFactor = None,scaleMax = [3,3,1.5], scaleMin = [1.25,1.25,1],random_seed = None):\n # randomly adjust the stain levels of the H&E image\n # based on the Deconvolution package: \n # https://deconvolution.readthedocs.io/en/latest/readme.html#two-stain-deconvolution \n #\n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # adjFactor: 3 float vec The adjustment factor for each of the three basis vectors \n # (<1 = less stain, 1 = same, >1 = more stain)\n # Element 1: Eosin\n # Element 2: Hematoxylin\n # Element 3: Null (the remaining structure)\n # If set the change won't be random\n # scaleMax: 3 float vector The maximum amount of change (increase or decrease) to the stain levels\n # (>=1)\n # scaleMin: 3 float vector The minimum amount of change (increase or decrease) to the stain levels\n # (>=1)\n # random_seed: int The random seed for numpy for consistent generation\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image (H&E) with the stain levels adjusted\n \n \n if adjFactor is None:\n np.random.seed(seed=random_seed) \n adjFactor = np.ones((1,3))\n for stI in range(len(scaleMax)):\n adjFactor[0,stI] = np.random.uniform(scaleMin[stI],scaleMax[stI]) ** np.random.choice((-1,1))\n adjFactor = adjFactor.flatten().tolist()\n rgbOut,rgb1,rgb2,rgb3 = adjust_stain(inputIm,adjFactor = adjFactor)\n comp_im = Image.fromarray(rgbOut,'RGB')\n return comp_im\n\n\ndef add_tear(inputIm,sampSpl = None, random_seed = None, nPts = 2,\n minSpacing = 20, maxSpacing = 40, tearStartFactor = [-.15,.15],tearEndFactor = [.85,1.15],\n dirMin = 10, dirMax = 30, inLineMax = None, perpMax = None, ptRadius = 2.25, tearAlpha = 1,\n inLinePercs = np.array([(-.5,-.3,-.2),(.5,.3,.2)]),perpPercs = np.array([(-.5,-.3,-.2),(.5,.3,.2)]),\n l1MinCt = 3, l1MaxCt = 8, minDensity = [.5,.5], maxDensity = [1.5,1.5],\n edgeWidth = 2, edgeAlpha = .75, edgeColorMult = [.85,.7,.85],rgbVal = (245,245,245),\n randEdge = True):\n # comp_im = add_tear(inputIm,sampSpl = None, random_seed = None, nPts = 2,\n # minSpacing = 20, maxSpacing = 40, tearStartFactor = [-.15,.15],tearEndFactor = [.85,1.15]\n # dirMin = 10, dirMax = 30, inLineMax = None, perpMax = None, ptRadius = 2.25, tearAlpha = 1,\n # inLinePercs = np.array([(-.5,-.3,-.2),(.5,.3,.2)]),perpPercs = np.array([(-.5,-.3,-.2),(.5,.3,.2)]),\n # t1MinCt = 3, t1MaxCt = 8, minDensity = [.5,.5], maxDensity = [1.5,1.5],\n # edgeWidth = 2, edgeAlpha = .75, edgeColorMult = [.85,.7,.85], rgbVal = (245,245,245),\n # randEdge = True):\n # Adds a tear to the tissue as an artifact.\n # These tears are seeded along a spline, with a randomized distance between the center of each tear.\n # Each tear is built up in layers (3 by default), with a randomized uniform distribution at each level.\n # The first layer has a small number of points, with a larger percentage of distance between them.\n # The next layer uses the previous layers points as a starting point at random\n # then adds a smaller amount of distance in a uniform distribution\n # This is repeated again for each of the remaining layers.\n # The result is then used to feed a distance function, so that any pixel within a radius of any of the points\n # is added to the tear mask\n # \n # ### \n # Inputs: Required\n # inputIm: a PIL Image A 2D RGB image\n # Inputs: Optional\n # sampSpl: n x 2 numpy arr You can optionally specify the sampled spline (non-random)\n # random_seed: int The random seed for numpy for consistent generation\n # nPts: int (+) The number of random handle points in the spline\n # minSpacing: float (+) The minimum for the random spacing between tears, in pixels\n # maxSpacing: float (+) The maximum for the random spacing between tears, in pixels\n # tearStartFactor: The min for where the tear randomly starts along the spline, in percentage & \n # 2 float vec The max for where the tear randomly starts along the spline, in percentage\n # tearEndFactor: The min for where the tear randomly end along the spline, in percentage & \n # 2 float vec The max for where the tear randomly ends along the spline, in percentage\n # dirMin: float (+) The minimum for randomized inline and perpendicular direction max distance in pixels\n # dirMax: float (+) The maximum for randomized inline and perpendicular direction max distance in pixels\n # inLineMax: float (+) You can optionally set the maximum size of the tear in the spline direction\n # perpMax: float (+) You can optionally set the maximum size of the tear in the perpendicular direction\n # ptRadius: float (+) The size of each point's effect on the tear in pixels\n # tearAlpha: float (0-1) The alpha transparency of the tear layer (1 = opaque, 0 = transparent)\n #\n # inLinePercs: You can optionally set your own tear layer structure\n # 2 x n numpy float arr The values are the percentage of distance in the in line direction that the tears take up\n # n = number of layers [.5,.3,.2] means most of the structure of the tear is set early\n # rec. [[-,-,-],[+,+,+]] and later layers fill it out\n # rec. each row should Making the matrix asymmetric betw. the + and -, could give a force component to the tear\n # add up to 1 or -1 Should match the number of layers in the perpendicular side\n # \n # perpPercs: You can optionally set your own tear layer structure\n # 2 x n numpy float arr The values are the % of distance in the perpendicular direction that the tears take up\n # n = number of layers [.5,.3,.2] means most of the structure of the tear is set early\n # rec. [[-,-,-],[+,+,+]] and later layers fill it out\n # rec. each row should Making the matrix asymmetric betw. the + and -, could give a sided-ness to the tear\n # add up to 1 or -1 Should match the number of layers in the inline side\n # \n # l1MinCt: int (+) The first layer is set by number instead of density in the later layers\n # - Minimum # of pts in the first layer\n # l1MaxCt: int (+) The first layer is set by number instead of density in the later layers\n # - Maximum # of pts in the first layer\n # minDensity: The second layer and beyond are set by density instead of #\n # n-1 int (+) vector - Minimum density of points in the 2nd, 3rd, etc. layers\n # n = number of layers\n # maxDensity: The second layer and beyond are set by density instead of #\n # n-1 int (+) vector - Minimum density of points in the 2nd, 3rd, etc. layers\n # edgeAlpha: float (0-1) The alpha transparency of the edge layer (1 = opaque, 0 = transparent)\n # edgeColorMult: The RGB multiplier of the edge of the tear \n # 3 float vector -Relative to the mean RGB color of the image\n # rgbVal: 3 float vector The RGB color of the tear (i.e. background)\n # randEdge: bool Whether to add some randomness to the edge of the tear\n # Defaults to on\n # ###\n # Output:\n # comp_im: a PIL Image A 2D RGB image with tear artifact added\n \n np.random.seed(seed=random_seed)\n dim = inputIm.size # width by height\n invDim = (dim[1],dim[0])\n if sampSpl is None:\n sampSpl = rand_spline(dim, nPts = nPts,random_seed = random_seed,endEdge=-2)\n \n # determine where the tears are located\n tearSpacing = np.random.uniform(minSpacing,maxSpacing,size=(sampSpl.shape[0],1))\n splLen = sampSpl.shape[0]-1\n minTearStartPx = splLen * 0\n maxTearEndPx = splLen * 1\n # randomly trim the start and end\n tearStEnd = np.zeros((2,1))\n tearStEnd[0] = np.random.uniform(tearStartFactor[0],tearStartFactor[1],size=(1,1)) * splLen\n tearStEnd[1] = np.random.uniform(tearEndFactor[0],tearEndFactor[1],size=(1,1)) * splLen\n tearStEnd = (np.round(tearStEnd)).astype(int)\n\n\n tearStEnd[tearStEnd > maxTearEndPx] = maxTearEndPx\n tearStEnd[tearStEnd < minTearStartPx] = minTearStartPx\n cdTS = np.round(np.cumsum(tearSpacing)).astype(int)\n cdTS = cdTS[(cdTS >= tearStEnd[0]) & (cdTS < tearStEnd[1])]\n\n tearCents = sampSpl[cdTS,:]\n splDer = sampSpl[:-1,:]- sampSpl[1:,:]\n\n if inLineMax is None:\n inLineMax = np.random.uniform(dirMin,dirMax,size=(1,1))\n if perpMax is None:\n perpMax = np.random.uniform(dirMin,dirMax,size=(1,1))\n \n splDer = np.concatenate((splDer[[0],:],splDer))\n tearDer = splDer[cdTS,:]\n areaMax = inLineMax * perpMax\n tearDensity = areaMax/ ((ptRadius**2)*np.pi)\n\n nTears = tearCents.shape[0]\n \n tearCts = np.random.randint(l1MinCt,l1MaxCt,size=(nTears,1))\n for tNo in range(len(minDensity)): # build up the layer matrix\n tearCts = np.append(tearCts, np.random.randint(np.ceil(tearDensity*minDensity[tNo]),\n np.ceil(tearDensity*maxDensity[tNo]),size=(nTears,1)),\n axis = 1)\n\n tearCtIdxs = np.concatenate((np.zeros((1)),np.cumsum(np.sum(tearCts,axis=1))),axis=0)\n\n tearXY = np.zeros((np.sum(tearCts),2))\n layerMats = {}\n # generate tears by using random points in layers\n for tIdx in range(len(cdTS)):\n layerMats[tIdx] = {}\n for layer in range(tearCts.shape[1]): # work in layers, each layer builds off of the last, gradually filling out the space\n # each layer builds off the last with a uniform distribution\n nTPts = tearCts[tIdx,layer]\n if layer == 0:\n centPts = np.repeat(np.reshape(tearCents[tIdx,:],(1,2)),nTPts,axis=0)\n else:\n centIdxs = np.random.randint(0,tearCts[tIdx,layer-1],size=(nTPts))\n centPts = layerMats[tIdx][layer-1][centIdxs,:]\n inLineFactor = np.random.uniform(inLinePercs[0,layer]*inLineMax,inLinePercs[1,layer]*inLineMax,size=(nTPts,1))\n perpFactor = np.random.uniform(perpPercs[0,layer]*perpMax,perpPercs[1,layer]*perpMax,size=(nTPts,1))\n cDerIL = tearDer[tIdx,:]\n cDerP = np.array([tearDer[tIdx,1], -tearDer[tIdx,0]])\n totVec = (inLineFactor * cDerIL) + (perpFactor * cDerP)\n newPts = centPts + totVec\n layerMats[tIdx][layer] = newPts.copy()\n idxRng = range(tearCtIdxs[tIdx].astype(int),tearCtIdxs[tIdx+1].astype(int))\n tearXY[idxRng,:] = np.vstack(list(layerMats[tIdx].values()))\n \n # rectify the points so we don't go out of bounds\n tearXY = np.maximum(tearXY,0)\n tearXY[:,0] = np.minimum(tearXY[:,0],dim[0]-1)\n tearXY[:,1] = np.minimum(tearXY[:,1],dim[1]-1)\n\n # turn these points into a distance mask\n tearMask = np.ones(invDim)\n tearMask[(np.round(tearXY[:,1])).astype(int),np.round(tearXY[:,0]).astype(int)] = 0\n tearDist = morphology.distance_transform_edt(tearMask)\n \n if randEdge == True:\n distRand = np.random.uniform(-int(ptRadius*.5),int(ptRadius*.5),size=invDim)\n tearDist = blur(tearDist+distRand,(5,5))\n tearBW = tearDist <= ptRadius\n\n alphaArr = (tearBW*tearAlpha*255).astype(np.uint8)\n colorArr = np.zeros((invDim[0],invDim[1],3),dtype=np.uint8)\n edgeArea = np.logical_and(tearDist > ptRadius,tearDist <= ptRadius+edgeWidth)\n \n # determine the color of the edge area and the \n meanColor = np.mean(np.array(inputIm),axis=(0,1))\n for i in range(len(rgbVal)):\n colorArr[:,:,i] = rgbVal[i]\n colorArr[edgeArea,i] = np.uint8(np.minimum(meanColor[i] * edgeColorMult[i],255))\n\n alphaArr[edgeArea] = edgeAlpha * 255\n alphaMask = Image.fromarray(alphaArr,'L')\n colorLayer = Image.fromarray(colorArr,'RGB')\n comp_im = Image.composite(colorLayer, inputIm, alphaMask)\n return comp_im\n \ndef apply_artifact(inputImName,artifactType,outputImName = None, outputDir = None,randAdd = 0, ext = None, perTileRand = None):\n # outputIm = apply_artifact(inputImName,artifactType,outputImName = None, outputDir = None,\n # randAdd = 0, ext = None, perTileRand = None):\n # Commmand line version of this package\n # Applies the default settings for each of the artifacts for this package\n # Handles per tile/slide randomization via hashing the file name into a random seed\n # Leans on the file structure to determine the tile name & slide name\n # \n # ###\n # Inputs: Required\n # inputImName: string The fully qualified name of the input image (include path)\n # (filename)\n # artifactType: string The type of artifact to add\n # Currently implemented artifacts:\n # 'marker', 'fold', 'sectioning', 'illumination', 'bubbles', 'stain', 'tear'\n # Inputs: Optional\n # outputImName: string Optional output filename\n # currently defaults to original name + '_' + first 4 chars of artifact\n # e.g. im1.jpeg -> im1_mark.jpeg\n # outputDir: string Optional output directory\n # defaults to current directory\n # randAdd: string Optional number to add to the random seed, e.g. if additional trials are desired\n # ext: string Extension to output the file as (defaults to same as input)\n # no period (e.g. 'jpeg', 'png')\n # perTileRand: Whether to do randomization by tile or by slide (True = tile, False = slide)\n # None or True or False Default (None) = based on the type of artifact\n # {'marker' : True, 'fold': True, 'sectioning': True, 'illumination': True, \n # 'bubbles': True, 'stain' : False, 'tear': True}\n # ###\n # Output:\n # outputIm: a PIL Image A 2D RGB image with artifact added\n # File Output:\n # Altered image saved to outputImName\n \n \n \n artifactType = artifactType.lower()\n # to remove any linkage between the different types of random addition (e.g. marker vs fold)\n typeSeedAdd = {'marker' : 1, 'fold': 2, 'sectioning': 3, 'illumination': 4, 'bubbles': 5, 'stain' : 6, 'tear': 7}\n # to randomize slide/tile based on type of artifact\n typeTileRand = {'marker' : True, 'fold': True, 'sectioning': True, 'illumination': True, 'bubbles': True, \n 'stain' : False, 'tear': True}\n \n inputIm = Image.open(inputImName)\n\n inputImDir,fName = os.path.split(inputImName)\n oPath1, rDir1 = os.path.split(inputImDir)\n _, rDir2 = os.path.split(oPath1)\n fNameNoExt = os.path.splitext(fName)[0]\n if ext is None:\n ext = os.path.splitext(fName)[-1]\n \n if perTileRand is None:\n perTileRand = typeTileRand[artifactType]\n if perTileRand == True: # take into account the tile name\n fID = os.path.join(rDir2,rDir1,fNameNoExt)\n else: # only take into account the slide name\n fID = os.path.join(rDir2,rDir1)\n\n randMax = (2**32) -1 # max size of the random seed\n # there's potentially some concern about the difference in 32 bit vs 64 bit systems\n h = blake2s()\n h.update(fID.encode('utf-8'))\n h_int = int(h.hexdigest(), 16)\n\n random_hash = h_int + randAdd + typeSeedAdd[artifactType]\n random_seed = random_hash % randMax\n \n if artifactType == \"marker\":\n outputIm = add_marker(inputIm,random_seed = random_seed)\n elif artifactType == \"fold\":\n outputIm = add_fold(inputIm,random_seed = random_seed)\n elif artifactType == \"sectioning\":\n outputIm = add_sectioning(inputIm,random_seed = random_seed)\n elif artifactType == \"illumination\":\n outputIm = add_illumination(inputIm,random_seed = random_seed)\n elif artifactType == \"bubbles\":\n outputIm = add_bubbles(inputIm,random_seed = random_seed)\n elif artifactType == \"stain\":\n outputIm = add_stain(inputIm,random_seed = random_seed)\n elif artifactType == \"tear\":\n outputIm = add_tear(inputIm,random_seed = random_seed)\n outputSuffix = artifactType[0:4]\n if outputImName is None:\n outputImName = \"%s_%s.%s\" % (fNameNoExt, outputSuffix, ext)\n if outputDir is not None:\n if not os.path.exists(outputDir):\n os.makedirs(outputDir)\n outputImName = os.path.join(outputDir,outputImName)\n outputIm.save(outputImName)\n return outputIm\n\nif __name__ == '__main__':\n # Map command line arguments to function arguments.\n apply_artifact(*sys.argv[1:])\n", "sub_path": "image_manipulation/img_manip.py", "file_name": "img_manip.py", "file_ext": "py", "file_size_in_byte": 48899, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "numpy.random.seed", "line_number": 41, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 41, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 45, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 45, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 46, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 46, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 54, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 54, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 67, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 67, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 75, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 78, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 79, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 80, "usage_type": "call"}, {"api_name": "numpy.floor", "line_number": 80, "usage_type": "call"}, {"api_name": "scipy.interpolate.pchip_interpolate", "line_number": 81, "usage_type": "call"}, {"api_name": "scipy.interpolate", "line_number": 81, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 116, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 116, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 121, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 127, "usage_type": "call"}, {"api_name": "numpy.arange", "line_number": 128, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 135, "usage_type": "call"}, {"api_name": "numpy.dstack", "line_number": 136, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 139, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 139, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 140, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 140, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 144, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 147, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 149, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 149, "usage_type": "attribute"}, {"api_name": "numpy.diag", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 150, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 150, "usage_type": "attribute"}, {"api_name": "numpy.sqrt", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.product", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.diag", "line_number": 151, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 152, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 152, "usage_type": "attribute"}, {"api_name": "scipy.stats.multivariate_normal", "line_number": 153, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 155, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 162, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 189, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 189, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 191, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 193, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 193, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 204, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 205, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology.distance_transform_edt", "line_number": 207, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology", "line_number": 207, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 213, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 213, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 213, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 214, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 214, "usage_type": "attribute"}, {"api_name": "PIL.Image.fromarray", "line_number": 217, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 217, "usage_type": "name"}, {"api_name": "PIL.Image.composite", "line_number": 219, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 219, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 265, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 265, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 269, "usage_type": "call"}, {"api_name": "numpy.copy", "line_number": 280, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 283, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 283, "usage_type": "attribute"}, {"api_name": "numpy.transpose", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.tile", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 292, "usage_type": "call"}, {"api_name": "numpy.pad", "line_number": 293, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 295, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 296, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 297, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 302, "usage_type": "call"}, {"api_name": "numpy.diff", "line_number": 304, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 307, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 307, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 308, "usage_type": "call"}, {"api_name": "numpy.float32", "line_number": 308, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 310, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 310, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 311, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 311, "usage_type": "attribute"}, {"api_name": "cv2.getPerspectiveTransform", "line_number": 322, "usage_type": "call"}, {"api_name": "cv2.warpPerspective", "line_number": 324, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 327, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology.distance_transform_edt", "line_number": 329, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology", "line_number": 329, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 331, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 331, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 332, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 335, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 340, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 342, "usage_type": "call"}, {"api_name": "cv2.GaussianBlur", "line_number": 343, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 345, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 345, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 345, "usage_type": "attribute"}, {"api_name": "numpy.random.seed", "line_number": 391, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 391, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 401, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology.distance_transform_edt", "line_number": 403, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology", "line_number": 403, "usage_type": "name"}, {"api_name": "numpy.random.randint", "line_number": 405, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 405, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 406, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 412, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 412, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 413, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 413, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 414, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 416, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 416, "usage_type": "attribute"}, {"api_name": "numpy.logical_not", "line_number": 418, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 421, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 423, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 424, "usage_type": "call"}, {"api_name": "numpy.divide", "line_number": 424, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 426, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 426, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 452, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 452, "usage_type": "attribute"}, {"api_name": "scipy.ndimage.morphology.distance_transform_edt", "line_number": 461, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology", "line_number": 461, "usage_type": "name"}, {"api_name": "numpy.logical_and", "line_number": 462, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 464, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 464, "usage_type": "name"}, {"api_name": "numpy.uint8", "line_number": 464, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 465, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 465, "usage_type": "attribute"}, {"api_name": "numpy.mean", "line_number": 468, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 468, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 471, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 473, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 473, "usage_type": "name"}, {"api_name": "PIL.Image.composite", "line_number": 474, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 474, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 502, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 502, "usage_type": "attribute"}, {"api_name": "numpy.linspace", "line_number": 506, "usage_type": "call"}, {"api_name": "numpy.linspace", "line_number": 507, "usage_type": "call"}, {"api_name": "numpy.meshgrid", "line_number": 508, "usage_type": "call"}, {"api_name": "numpy.dstack", "line_number": 509, "usage_type": "call"}, {"api_name": "numpy.amin", "line_number": 515, "usage_type": "call"}, {"api_name": "numpy.amax", "line_number": 517, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 520, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 520, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 521, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 521, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 522, "usage_type": "call"}, {"api_name": "numpy.interp", "line_number": 523, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 523, "usage_type": "call"}, {"api_name": "numpy.float64", "line_number": 523, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 526, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 527, "usage_type": "call"}, {"api_name": "numpy.multiply", "line_number": 527, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 528, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 528, "usage_type": "name"}, {"api_name": "deconvolution.Deconvolution", "line_number": 562, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 566, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 569, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations._array_positive", "line_number": 573, "usage_type": "attribute"}, {"api_name": "deconvolution.pixeloperations", "line_number": 573, "usage_type": "name"}, {"api_name": "numpy.repeat", "line_number": 574, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 575, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 576, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations._array_to_colour_255", "line_number": 579, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations", "line_number": 579, "usage_type": "name"}, {"api_name": "deconvolution.pixeloperations._array_to_colour_255", "line_number": 580, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations", "line_number": 580, "usage_type": "name"}, {"api_name": "deconvolution.pixeloperations._array_to_colour_255", "line_number": 581, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations", "line_number": 581, "usage_type": "name"}, {"api_name": "deconvolution.pixeloperations._array_to_colour_255", "line_number": 582, "usage_type": "call"}, {"api_name": "deconvolution.pixeloperations", "line_number": 582, "usage_type": "name"}, {"api_name": "numpy.random.seed", "line_number": 612, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 612, "usage_type": "attribute"}, {"api_name": "numpy.ones", "line_number": 613, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 615, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 615, "usage_type": "attribute"}, {"api_name": "numpy.random.choice", "line_number": 615, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 618, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 618, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 625, "usage_type": "call"}, {"api_name": "numpy.random.seed", "line_number": 699, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 699, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 706, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 706, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 711, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 712, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 712, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 713, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 713, "usage_type": "attribute"}, {"api_name": "numpy.round", "line_number": 714, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 719, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 719, "usage_type": "call"}, {"api_name": "numpy.random.uniform", "line_number": 726, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 726, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 728, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 728, "usage_type": "attribute"}, {"api_name": "numpy.concatenate", "line_number": 730, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 733, "usage_type": "attribute"}, {"api_name": "numpy.random.randint", "line_number": 737, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 737, "usage_type": "attribute"}, {"api_name": "numpy.append", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 739, "usage_type": "attribute"}, {"api_name": "numpy.ceil", "line_number": 739, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 740, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.cumsum", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 743, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 745, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 745, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 754, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 754, "usage_type": "call"}, {"api_name": "numpy.random.randint", "line_number": 756, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 756, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 758, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 758, "usage_type": "attribute"}, {"api_name": "numpy.random.uniform", "line_number": 759, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 759, "usage_type": "attribute"}, {"api_name": "numpy.array", "line_number": 761, "usage_type": "call"}, {"api_name": "numpy.vstack", "line_number": 766, "usage_type": "call"}, {"api_name": "numpy.maximum", "line_number": 769, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 770, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 771, "usage_type": "call"}, {"api_name": "numpy.ones", "line_number": 774, "usage_type": "call"}, {"api_name": "numpy.round", "line_number": 775, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology.distance_transform_edt", "line_number": 776, "usage_type": "call"}, {"api_name": "scipy.ndimage.morphology", "line_number": 776, "usage_type": "name"}, {"api_name": "numpy.random.uniform", "line_number": 779, "usage_type": "call"}, {"api_name": "numpy.random", "line_number": 779, "usage_type": "attribute"}, {"api_name": "cv2.blur", "line_number": 780, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 783, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 784, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 784, "usage_type": "attribute"}, {"api_name": "numpy.logical_and", "line_number": 785, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 788, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 788, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 791, "usage_type": "call"}, {"api_name": "numpy.minimum", "line_number": 791, "usage_type": "call"}, {"api_name": "PIL.Image.fromarray", "line_number": 794, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 794, "usage_type": "name"}, {"api_name": "PIL.Image.fromarray", "line_number": 795, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 795, "usage_type": "name"}, {"api_name": "PIL.Image.composite", "line_number": 796, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 796, "usage_type": "name"}, {"api_name": "PIL.Image.open", "line_number": 842, "usage_type": "call"}, {"api_name": "PIL.Image", "line_number": 842, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 844, "usage_type": "call"}, {"api_name": "os.path", "line_number": 844, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 845, "usage_type": "call"}, {"api_name": "os.path", "line_number": 845, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 846, "usage_type": "call"}, {"api_name": "os.path", "line_number": 846, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 847, "usage_type": "call"}, {"api_name": "os.path", "line_number": 847, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 849, "usage_type": "call"}, {"api_name": "os.path", "line_number": 849, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 854, "usage_type": "call"}, {"api_name": "os.path", "line_number": 854, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 856, "usage_type": "call"}, {"api_name": "os.path", "line_number": 856, "usage_type": "attribute"}, {"api_name": "hashlib.blake2s", "line_number": 860, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 885, "usage_type": "call"}, {"api_name": "os.path", "line_number": 885, "usage_type": "attribute"}, {"api_name": "os.makedirs", "line_number": 886, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 887, "usage_type": "call"}, {"api_name": "os.path", "line_number": 887, "usage_type": "attribute"}, {"api_name": "sys.argv", "line_number": 893, "usage_type": "attribute"}]} +{"seq_id": "108258875", "text": "# -*- coding: = utf-8 -*-\n# __author__ = 'lucile'\nfrom app.Recepteur.nda import nda\nfrom app.Serveur.opentsdb import opentsdb\nimport translate\nimport ctypes\nfrom struct import *\nimport logging\nimport array\nimport binascii\nimport datetime\nimport pytz\nimport base64\nimport uuid\nimport numpy as np\nimport time\nimport json\nimport sys\nimport traceback\nimport pidly\nimport re\nimport fnmatch\nimport pandas as pd\nimport requests\nimport binascii\nimport simplejson\nimport gzip\nimport paramiko, base64\n#import astropy\nfrom astropy.io import fits\n#from couchbase import Couchbase, views\n#from couchbase.views.params import Query\nfrom matplotlib import *\nimport os.path\nimport uuid\nimport shutil\n#import h5py\nimport strict_rfc3339\n#import cv2\n#from PIL import Image,ImageOps\n#from astropy.table import Table\nimport matplotlib.pyplot as plt\n#import SimpleCV\nfrom elasticsearch import Elasticsearch\nimport av\nimport time\nimport cv2\nimport re\nfrom robobrowser import RoboBrowser\nfrom bs4 import BeautifulSoup\nimport lxml\nfrom selenium import webdriver\nimport time\n#from selenium import webdriver\nfrom selenium.common.exceptions import TimeoutException\nfrom selenium.webdriver.support.ui import WebDriverWait # available since 2.4.0\n# available since 2.26.0\nfrom selenium.webdriver.support import expected_conditions as EC\n\n\nclass nrh(object):\n\n DATA_DIR = \"/data/data_nrh/\"\n\n def translate(self, text):\n\n #session = requests.Session()\n headers = {\n #\"Content-Type\": \"application/json\",\n \"accept-encoding\": \"gzip, deflate, sdch\",\n \"accept-language\": \"fr-FR,fr;q=0.8,en-US;q=0.6,en;q=0.4,es;q=0.2,de;q=0.2,it;q=0.2\",\n \"accept\": \"*/*\",\n #\"x-client-data\": \"CKW2yQEIkpTKAQj9lcoBCMWYygE=\",\n \"referer\": \"https://translate.google.com/\",\n \"user_agent\": 'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.1) Gecko/2008071615 Fedora/3.0.1-1.fc9 Firefox/3.0.1'\n }\n # session.headers.update(headers)\n print(text)\n\n # url=\"https://translate.google.com/translate_a/single?client=t&sl=fr&tl=en&hl=fr&dt=at&dt=bd&dt=ex&dt=ld&dt=md&dt=qca&dt=rw&dt=rm&dt=ss&dt=t&ie=UTF-8&oe=UTF-8&source=btn&ssel=0&tsel=0&kc=0&tk=345478.223371&q=salut\"\n # url =\n # 'https://translate.google.com/#fr/en/single?client=t&sl=fr&tl=en&hl=fr&dt=at&dt=bd&dt=ex&dt=ld&dt=md&dt=qca&dt=rw&dt=rm&dt=ss&dt=t&ie=UTF-8&oe=UTF-8&source=bh&ssel=0&tsel=0&kc=1&tk=266742.146682&q=bonjour'\n url = \"https://translate.google.com/\"\n\n # Create a new instance of the Firefox driver\n driver = webdriver.Firefox()\n\n # go to the google home page\n driver.get(\"http://www.google.com\")\n\n print(driver.title)\n\n print(url)\n driver.get(\"http://pythonscraping.com/pages/javascript/ajaxDemo.html\")\n time.sleep(3)\n print(driver.find_element_by_id(\"content\").text)\n driver.close()\n print(url)\n resultat = session.get(url, headers=headers)\n # print(resultat.text)\n soup = BeautifulSoup(resultat.text, 'html.parser')\n\n print(soup.prettify())\n\n def getRequest(self, filtre, type, option):\n \"\"\"\n :param filtre:\n :param type:\n :param option:\n :return:\n \"\"\"\n\n if type == \"canvas\":\n return self.getCanvas(filtre, option)\n elif type == \"setImport\":\n return self.setImport(filtre, option)\n elif type == \"setImport1d\":\n return self.setImport1d(filtre, option)\n elif type == \"setImportLog\":\n return self.setImportLog(filtre, option)\n elif type == \"view\":\n return self.getView(filtre, option)\n elif type == \"getImage\":\n return self.getImage(filtre, option)\n elif type == \"fits\":\n return self.getFits(filtre, option)\n elif type == \"1d\":\n return self.get1d(filtre, option)\n elif type == \"getVideo\":\n return self.getVideo(filtre, option)\n elif type == \"getIntegre\":\n return self.getIntegre(filtre, option)\n\n def getImage(self, f, sel):\n print(\"+++++++++\", sel)\n\n try:\n o = opentsdb()\n #print (o)\n #resp = o.Query(f, sel)\n print(f)\n resp = o.getImage(f, sel)\n\n except:\n print(\"Unexpected error:\", sys.exc_info()[0])\n resp ={\n \"status\": \"error\",\n \"message\": \"server prob.\",\n \"type\": \"brut\"\n\n }\n\n return resp\n\n def getIntegre(self, f, sel):\n \"\"\"\n\n PRO rh_integre, tb_fich, fichier_itg, integ, hdd, hfd, TOUT = TOUT\n ;+ ------------------------------------------------------------------\n ; NAME:\n ; RH_INTEGRE\n ; PURPOSE:\n ; Cette procedure integre les donnees sur un delta T donne.\n ; Elle tourne sur des fichiers Nancay natifs et cree en sortie des\n ; fichiers Nancay natifs.\n ; CALLING SEQUENCE:\n ; RH_INTEGRE, ['/data/extraits/vilmer/2d020220.01B'],'./2i020220.01B',\n ; 4.,[11,13,59,87],[11,17,47,50]\n ;\n ; INPUTS:\n ; tb_fich :\tListe des fichiers a integrer\n ; fichier_itg:\tnom du fichier integre\n ; integ:\ttemps d'integration en secondes\n ; hdd:\t\tdebut d'integration demande(h,mn,sec,ct)\n ; hfd:\t\tfin d'integration demande \"\n ;\n ; KEYWORD:\n ; TOUT:\t\tintegration de tout le fichier\n\n ; MODIFICATIONS:\n ;\tjuin 1999 : Passage aux fichiers 8 images/sec\n ;\t\t lecture acq par acq\n ;\tnov 1999 : Demarrage a la freq zero\n ; jan 2006 : bug recherche heure de debut corrige\n ; accepte nint=1 : fait alors une extraction simple\n ; dec 2010 : bug corrigé (flag i_stop_fich) on ne lit pas les fichiers\n ; suivants quand on a rencontre l'heure de fin demandee.\n ;- ------------------------------------------------------------------\n \"\"\"\n\n #filtre = json.loads(f)\n filtre =f\n # date = datetime.datetime.strptime(filtre[\"date\"], \"%Y-%m-%dT%H:%M:%S.%f\")\n\n # \"2005-12-19T09:18:25.060+0100\"\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n\n datefin = datetime.datetime.strptime(\n filtre[\"datefin\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n\n #duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n print(filtre[\"integration\"])\n\n #fRH = [\"2d\" + datedeb.strftime(\"%y%m%d\") +\".01\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01Z\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01ZZ\"]\n\n sel = filtre[\"sel\"]\n\n \"\"\"\n fichier à utiliser en fonction de la date\n \"\"\"\n\n find = False\n files = []\n timehdeb = []\n timehfin = []\n unixdatedeb = time.mktime(datedeb.timetuple())\n unixdatefin = time.mktime(datefin.timetuple())\n\n for res in sel:\n nrhfile = res[\"file\"]\n #nrhfrq = res[\"frq\"]\n hdeb = (strict_rfc3339.rfc3339_to_timestamp(res[\"hdeb\"]))\n\n # 2005-12-01T10:07:42.085000+00:00\n thdeb = datetime.datetime.fromtimestamp(hdeb)\n if filtre[\"integration\"] != 0:\n integration = (filtre[\"integration\"])\n else:\n integration = (res[\"itg\"]/1000)\n hfin = (strict_rfc3339.rfc3339_to_timestamp(res[\"hfin\"]))\n thfin = datetime.datetime.fromtimestamp(hfin)\n files.append(\"/data/data_nrh/nrh/\" + datedeb.strftime(\"%Y\") + \"/\" +\n datedeb.strftime(\"%m\") + \"/\" + datedeb.strftime(\"%d\") + \"/\" + nrhfile)\n timehdeb.append(hdeb)\n timehfin.append(hfin)\n\n if (unixdatedeb > hdeb and unixdatedeb < hfin):\n find = True\n\n minhdeb = min(timehdeb)\n\n timeheuredebut = max([minhdeb, unixdatedeb])\n heuredebut = datetime.datetime.fromtimestamp(timeheuredebut)\n print(\"de \", heuredebut)\n maxhfin = max(timehfin)\n\n timeheurefin = min([maxhfin, unixdatefin])\n heurefin = datetime.datetime.fromtimestamp(timeheurefin)\n print(\"à \", heurefin)\n #fname = \"/data/data_nrh/rh/\"+datedeb.strftime(\"%Y\")+\"/\"+datedeb.strftime(\"%m\")+\"/\"+ datedeb.strftime(\"%d\")+\"/\"+nrhfile\n\n # print(fname)\n uid = str(uuid.uuid4())\n os.mkdir('/var/www/html/Public/Brut/' + uid)\n idl = pidly.IDL('/usr/local/bin/idl')\n #$IDL = \"/home/user/idl83/bin/idl \".self::$image_rh.\" -quiet -args \".$file.\" \".json_encode($time,JSON_NUMERIC_CHECK).\" \".$frequence.\" \".$polarite.\" \".$integration.\" \".self::$uri.\" \";\n\n idl.fichier = files\n\n heurededebut = list(heuredebut.utctimetuple()[3:6])\n heurededebut.append(int(heuredebut.microsecond / 1000))\n\n idl.hd = heurededebut\n\n heuredefin = list(heurefin.utctimetuple()[3:6])\n heuredefin.append(int(heurefin.microsecond / 10000))\n idl.hf = heuredefin\n\n idl.kint = float(integration)\n idl.frequence = frequence\n idl('print,frequence')\n\n idl.npol = 0\n uri = '/var/www/html/Public/Brut/' + uid + '/2d' + \\\n datedeb.strftime(\"%y%m%d\") + '.0'\n idl.url = uri\n\n idl('@rh_common.inc')\n # idl('status=RH_OPEN(fichier,/SEL,/MALAX)')\n\n idl('RH_INTEGRE,fichier,url,kint,hd,hf')\n\n #idl('print, entFI.hdeb ')\n \"\"\"\n DATE-OBS : date de début d’observation ; 2014-02-01\n TIME-OBS : heure de début d’observation ;08 :18 :02.000\n DATE-END : date de fin d’observation\n TIME-END : heure de fin d’observation\n\n PHYSPARA : ‘I+V’ ; paramètres observés (Stokes)\n\n OBSERVATORY : ‘Observatoire de Paris – Nançay’ (a des noms divers selon les fichiers)\n\n OBS-TYPE : ‘radio’\n OBS-SUBTYPE : ‘visibility’\n TELESCOPE : ‘ radio interferometer’\n INSTRUMENT : ‘NRH’\n\n SOURCE : ‘SUN’, ou GYG,CASS,TAUR, VIRG, HYDR, DAB\n FREQ-START : fréquence de départ, en MHZ\n FREQ-STOP : fréquence de fin, MHZ\n BUNITS : ‘SFU’\n EXP-TIME : sampling time, millisec\n TIME-STEP : integration time, millisec\n Entetes fixe\n \"\"\"\n \"\"\"\n idl('.Reset_Session')\n idl('@rh_common.inc')\n\n idl.url='/media/data/Public/Brut/'+uid+'.brut'\n idl('RH_OPEN(url,/SEL,/MONO)')\n idl('print, entFI.hdeb ')\n \"\"\"\n # idl.pro('IMAGE_RH',fichier,h,frequence,polarite,integration,url)\n #idl('rh_hfin, h, klu, heurlu ')\n\n # IDL procedure with Python argument(s)::\n #idl.pro('IMAGE_RH', range(10), range(10), xstyle=True, ystyle=True)\n\n dm = []\n dm = idl.hd.tolist()\n datemoy = datetime.time(dm[0], dm[1], dm[2], dm[3])\n size = os.path.getsize(uri)\n #os.system(\"chmod a+x /media/data/Public/Brut/'+uid+'.brut\")\n\n # print(datemoy.isoformat())\n\n response = {\n \"status\": \"success\",\n \"message\": \"Integration loaded succesfully.\",\n \"type\": \"brut\",\n \"size\":size,\n \"date\": datemoy.isoformat(),\n \"url\": \"/Public/Brut/\" + uid + '/2d' + datedeb.strftime(\"%y%m%d\") + '.0'\n }\n # idl.close()\n idl('exit')\n return response\n\n def image(self, f, sel):\n print(\"getImage\")\n \"\"\"\n ; ----------------------------------------------------------------\n ; RH_IM_2DSEL\n ; ----------------------------------------------------------------\n ; Calcule l'image a la frequence nof non polar ou polar\n ; integree kint fois a partir de klu\n ; Le resultat est dans le tableau ima\n ;\n ; INPUTS:\n ;\tklu : numero de l'acquisition a lire\tLONG\n ;\t debut du fichier = 0L\n ;\tnof : numero de frequence\n ;\tnpol : 0-Non Polar, 1-Polar\n ;\tkint : facteur d'integration\n ;\tsz : taille de l'image\n ;\tlarg : largeur de l'image\n ;\th : Heure de l'image integree (milieu)\n ;\tima : image integree\n ;\n ; OUTPUTS:\n ;\th : Heure de l'image integree (milieu)\n ;\tima : image integree\n ;\n ; COMMON BLOCKS:\n ; RH\n ; MODIFICATIONS:\n ; 03 sep 08 : Appelle RH_MALCROND_IM_2D : qui calcule l'image en\n ; utilisant au mieux tous les harmoniques y compris anti_alias\n ; 05/07/2011: gestion de la panne du correlateur depuis le\n ; 27/11/2009 (antennes AA3 et AA4 avec NS12 a NS16)\n ; on corrige la panne pour l'option image2d de menu_rh\n ; 25/06/2013 : gestion de la panne du correlateur ns08 avec ew01(h1)\n ; -----------------------------------------------------------------\n\n PRO RH_IM_2DSEL, klu, nof, npol, kint, sz, larg, h, ima\n \"\"\"\n filtre = json.loads(f)\n\n # date = datetime.datetime.strptime(filtre[\"date\"], \"%Y-%m-%dT%H:%M:%S.%f\")\n\n # \"2005-12-19T09:18:25.060+0100\"\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n print(datedeb)\n #duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n print(frequence)\n integration = (filtre[\"integration\"])\n print(integration)\n #fRH = [\"2d\" + datedeb.strftime(\"%y%m%d\") +\".01\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01Z\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01ZZ\"]\n\n sel = filtre[\"sel\"]\n print(sel)\n\n \"\"\"\n fichier à utiliser en fonction de la date\n \"\"\"\n\n find = False\n for res in sel:\n nrhfile = res[\"file\"][0]\n #nrhfrq = res[\"frq\"]\n hdeb = (strict_rfc3339.rfc3339_to_timestamp(res[\"hdeb\"][0]))\n hfin = (strict_rfc3339.rfc3339_to_timestamp(res[\"hfin\"][0]))\n\n unixdatedeb = time.mktime(datedeb.timetuple())\n print(hdeb, unixdatedeb, hfin)\n if (unixdatedeb > hdeb and unixdatedeb < hfin):\n find = True\n break\n if find:\n\n fname = \"/data/data_nrh/rh/\" + datedeb.strftime(\"%Y\") + \"/\" + datedeb.strftime(\n \"%m\") + \"/\" + datedeb.strftime(\"%d\") + \"/\" + nrhfile\n\n print(fname)\n uid = str(uuid.uuid4())\n\n idl = pidly.IDL('/usr/local/bin/idl')\n #$IDL = \"/home/user/idl83/bin/idl \".self::$image_rh.\" -quiet -args \".$file.\" \".json_encode($time,JSON_NUMERIC_CHECK).\" \".$frequence.\" \".$polarite.\" \".$integration.\" \".self::$uri.\" \";\n\n idl.fichier = fname\n\n heurededebut = list(datedeb.utctimetuple()[3:6])\n heurededebut.append(int(datedeb.microsecond / 10000))\n idl.h = heurededebut\n\n idl.kint = int(integration)\n idl.frequence = frequence\n idl('print,frequence')\n\n idl.npol = 0\n\n idl.url = '/media/data/Public/Images/' + uid + '.png'\n\n idl('@rh_common.inc')\n idl('status=RH_OPEN(fichier,/SEL,/MALAX)')\n idl('print, entFI.frq')\n idl('nof = WHERE(entFI.frq EQ frequence)')\n idl('print,nof')\n idl('ima=fltarr(300,300)')\n idl('rh_hdeb,h, klu, heurlu')\n\n # idl.pro('IMAGE_RH',fichier,h,frequence,polarite,integration,url)\n #idl('rh_hfin, h, klu, heurlu ')\n\n idl.sz = 200\n idl.LARG = 4\n idl(\"set_plot,'PS'\")\n idl('RH_IM_2DSEL, klu, nof, npol, kint, sz, larg, h, ima')\n idl('loadct,3, /SILENT')\n idl('TVLCT, R, G, B, /GET')\n idl(\"write_image, url,'PNG',bytscl(ima),R, G, B\")\n # IDL procedure with Python argument(s)::\n #idl.pro('IMAGE_RH', range(10), range(10), xstyle=True, ystyle=True)\n\n dm = []\n dm = idl.h.tolist()\n datemoy = datetime.time(dm[0], dm[1], dm[2], dm[3])\n os.system(\"chmod a+x /media/data/Public/Images/'+uid+'.png\")\n\n print(datemoy.isoformat())\n response = {\n \"status\": \"success\",\n \"message\": \"Image loaded succesfully.\",\n \"type\": \"img\",\n \"date\": datemoy.isoformat(),\n \"url\": \"/Public/Images/\" + uid + \".png\"\n }\n # idl.close()\n idl('exit')\n return response\n\n def getVideo(self, f, sel):\n print(\"getVideo\")\n \"\"\"\n une video par jour\n \"\"\"\n filtre = json.loads(f)\n extin = 'mpg'\n extout = 'mp4'\n # \"2005-12-19T09:18:25.060+0100\"\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n print(datedeb)\n #duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n print(frequence)\n integration = filtre[\"integration\"]\n print(integration)\n\n nrhfile = \"nrh\" + datedeb.strftime(\"%d%m%Y\") + \"_\" + str(frequence)\n # nrh01092006_3270.mpg\n print(nrhfile)\n\n \"\"\"\n fichier à utiliser en fonction de la date\n \"\"\"\n\n find = True\n \"\"\"\n for res in sel:\n nrhfile = res[\"file\"][0]\n nrhfrq = res[\"frq\"]\n hdeb = (strict_rfc3339.rfc3339_to_timestamp(res[\"hdeb\"][0]))\n hfin = (strict_rfc3339.rfc3339_to_timestamp(res[\"hfin\"][0]))\n\n unixdatedeb = time.mktime(datedeb.timetuple())\n\n if (unixdatedeb>hdeb and unixdatedeb entfi=mrdfits('nrh2_1509_a00_20100711_082416c03_c.fts',2,h2)\n ;% Attempt to subscript FNAMES with I is out of range.\n ;% Error occurred at: MRD_TABLE 2326 /usr/local/ssw/gen/idl/fits/mrdfits.pro\n ;% MRDFITS 2721 /usr/local/ssw/gen/idl/fits/mrdfits.pro\n ; prevoir un programme pour remplir et completer entfi dans ce cas\n ;-**************************************************************************\n @rh_common.inc\n \"\"\"\n filtre = json.loads(f)\n \"\"\"\n {\n \"datedeb\":\"2007-12-21T09:52:41.010+0100\",\n \"frequence\":1640,\n \"integration\":10,\n \"dure\":72000,\n \"recepteur\":0,\n \"sel\":\n [\n {\n \"hfin\":[\"2007-12-21T14:35:44.089000+00:00\"],\n \"frq\":[2280,3270,4080,1509,4320,0,0,0,0,0],\n \"hdeb\":[\"2007-12-21T11:44:13.005000+00:00\"],\n \"file\":[\"2d071221.01Z\"]\n },{\n \"hdeb\":[\"2007-12-21T14:35:45+00:00\"],\n \"file\":[\"2d071221.01ZZ\"],\n \"hfin\":[\"2007-12-21T15:18:59.036000+00:00\"],\n \"frq\":[2280,3270,4080,1509,4320,0,0,0,0,0]\n },{\n \"hdeb\":[\"2007-12-21T08:52:41.010000+00:00\"],\n \"file\":[\"2d071221.01\"],\n \"hfin\":[\"2007-12-21T11:44:12.094000+00:00\"],\n \"frq\":[2280,3270,4080,1509,4320,0,0,0,0,0]\n }\n ]\n }\n javascript rfc3339 \"2015-01-20T12:52:29.005000+00:00\",\n python \"2007-12-21T09:52:41.010+0100\"\n \"\"\"\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n\n #duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n sel = filtre[\"sel\"]\n\n \"\"\"\n fichier à utiliser en fonction de la date\n \"\"\"\n\n find = False\n for res in sel:\n nrhfile = res[\"file\"][0]\n nrhfrq = res[\"frq\"]\n hdeb = (strict_rfc3339.rfc3339_to_timestamp(res[\"hdeb\"][0]))\n hfin = (strict_rfc3339.rfc3339_to_timestamp(res[\"hfin\"][0]))\n unixdatedeb = time.mktime(datedeb.timetuple())\n\n if (unixdatedeb > hdeb and unixdatedeb < hfin):\n find = True\n break\n\n integration = (filtre[\"integration\"])\n #fRH = [\"2d\" + datedeb.strftime(\"%y%m%d\") +\".01\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01Z\", \"2d\" + datedeb.strftime(\"%y%m%d\") +\".01ZZ\"]\n\n if (find):\n\n fname = \"/data/data_nrh/rh/\" + datedeb.strftime(\"%Y\") + \"/\" + datedeb.strftime(\n \"%m\") + \"/\" + datedeb.strftime(\"%d\") + \"/\" + nrhfile\n\n print(fname)\n uid = str(uuid.uuid4())\n\n idl = pidly.IDL('/usr/local/bin/idl')\n #$IDL = \"/home/user/idl83/bin/idl \".self::$image_rh.\" -quiet -args \".$file.\" \".json_encode($time,JSON_NUMERIC_CHECK).\" \".$frequence.\" \".$polarite.\" \".$integration.\" \".self::$uri.\" \";\n idl('@rh_common.inc')\n idl.fichier = fname\n\n idl.h = list(datedeb.timetuple()[3:7])\n\n idl.kint = int(integration)\n idl.frequence = int(frequence)\n # idl.nof=int(sel)\n\n # idl.npol=0\n\n # idl.url='/media/data/Public/Images/'+uid+'.png'\n dirUrl = '/media/data/Public/Fits/' + uid\n #/media/data/Public/Fits/\n idl.url = dirUrl\n os.mkdir(dirUrl)\n\n idl('status=RH_OPEN(fichier,/SEL,/MONO)')\n idl('nof = WHERE(entFI.frq EQ frequence)')\n # idl('ima=fltarr(200,200)')\n #idl('rh_hdeb,h, klu, heurlu')\n # print(idl.h)\n # print(idl.heurlu)\n # print(idl.klu)\n # print(idl.nof)\n # print(idl.kint)\n # idl.pro('IMAGE_RH',fichier,h,frequence,polarite,integration,url)\n #idl('rh_hfin, h, klu, heurlu ')\n # print(idl.h)\n # print(idl.heurlu)\n # print(idl.klu)\n idl('ch=strarr(10)')\n idl(\"ch=['N','N','N','N','N','N','N','N','N','N']\")\n idl(\"ch[nof]='Y'\")\n idl(\"ch=['Y','N','N','N','N','N','N','N','N','N']\")\n print(idl.ch)\n # idl.sz=200\n # idl.LARG=4\n # idl(\"set_plot,'PS'\")\n #idl('hdeb = JSON_PARSE(time,/TOARRAY)')\n # print(idl.hdeb)\n\n # hfin = JSON_PARSE(Result[1],/TOARRAY)\n #idl('RH_IM_2DSEL, klu, nof, npol, kint, sz, larg, h, ima')\n idl('DPATCHFITSHARM, fichier, 1,[00,00,00,00],[24,00,00,00],url,selec=ch')\n #idl('DPATCHFITSHARM, fichier, 1,[00,00,00,00],[24,00,00,00],url')\n #idl('loadct,3, /SILENT')\n #idl('TVLCT, R, G, B, /GET')\n #idl(\"write_image, url,'PNG',bytscl(ima),R, G, B\")\n # IDL procedure with Python argument(s)::\n #idl.pro('IMAGE_RH', range(10), range(10), xstyle=True, ystyle=True)\n print(\"fin idl\", idl.status)\n # os.cd(dirUrl)\n Files = []\n\n Dirs = os.listdir(dirUrl)\n for filename in Dirs:\n print(filename)\n Files.append(\"/Public/Fits/\" + uid + \"/\" + filename)\n try:\n hdulist = fits.open(dirUrl + \"/\" + filename)\n print(\"coucou\")\n print(hdulist.info())\n for h in hdulist[0].header:\n print(h)\n print(hdulist[0].header[\"DATE_OBS\"])\n print(hdulist[0].header[\"DATE_END\"])\n except:\n print(\"Unexpected error:\", sys.exc_info()[0])\n\n print(Files)\n\n print(\"break\")\n\n print(\"break\")\n response = {\n \"status\": \"success\",\n \"message\": \"Fits done succesfully.\",\n \"type\": \"fits\",\n\n \"date\": datedeb, \"url\": Files[0]\n }\n # idl.close()\n idl('exit')\n print(response)\n return response\n\n def get1d(self, f, sel):\n print(\"get1d\")\n \"\"\"\n ; =======================================================================\n ; IMAGE_1D\n ; -----------------------------------------------------------------------\n ; INPUTS:\n ; \tpath_data:\trepertoire contenant les fichiers de donnees\n ; \tpath_poub:\trepertoire poubelle permettant d'ecrire des\n ; \tdes fichiers temporaires\n ; \tcmd_ps:\t\tcommande d'impression PS\n ; \tliste_scratch: nom du fichier contenant la liste des fichiers\n ; \ttemporaires a detruire a la sortie de 'menu_rh'\n ;\n ; COMMON WIMAGE_1D:\n ; \tnomfich:\tnom du fichier selectionne\n ;\tnom_don:\tnom 'non polar' ou 'polar'\n ;\tt_don:\t\tt_don(i) = 0-non sel, 1-sel\n ;\ttab_don:\tliste de selection\n ;\tcdon:\t\tnombre de polar selectionnees\n ;\tnom_freq:\tnom des frequences\n ;\tt_freq:\t\tt_freq(i) = 0-non sel, 1-sel\n ;\ttab_freq:\tnumeros des frequences selectionnees\n ;\tcfreq:\t\tnombre de frequences selectionnees\n ;\tsel_graf:\t0-Image, 1-Niveaux, 2-Surface\n ;\ts_heure:\tID heure\n ;\theure:\t\theure courante\n ;\tinteg:\t\tfacteur d'integration\n ;\tng:\t\tnombre courant de graphiques\n ;\tsz_ima:\t\ttaille d'une image\n ;\tima:\t\t1 image\n ;\tlec:\t\t=1 si lecture a faire avant de tracer\n ;\tklu:\t\tindice courant de lecture\n ;\tpsmode:\t\t0-Paysage, 1-portrait\n ;\tcmdps:\t\tcommande d'impression PS\n ;\tfpoub:\t\tdirectory poubelle\n ;\tfscratch:\tfichier contenant la liste des fichiers temporaires\n ;\n ; =======================================================================\n\n PRO image_1d, path_data, path_poub, cmd_ps, liste_scratch, GROUP=Gp\n\n \"\"\"\n \"\"\"\n ; ----------------------------------------------------------------\n ; im_1dsel\n ; ----------------------------------------------------------------\n ; Calcule pour la frequence nof l'image EW non polar ou EW polar ou\n ; NS non polar ou NS polar\n ; Image integree kint fois a partir de klu\n ; Le resultat est dans le tableau ima\n ;\n ; INPUTS:\n ;\tklu : numero de l'acquisition a lire\tLONG\n ;\t debut du fichier = 0L\n ;\tnof : numero de frequence\n ;\tnpol : 0-EW Non Polar\n ; 1-EW Polar\n ; 2-NS Non Polar\n ; 3-NS Polar\n ;\tkint : facteur d'integration\n ;\tsz : taille de l'image\n ;\th : Heure de l'image integree (milieu)\n ;\tima : image integree\n ;\n ; OUTPUTS:\n ;\th : Heure de l'image integree (milieu)\n ;\tima : image integree\n ;\n ; COMMON BLOCKS:\n ; RH\n ; 20-09-2004 : correction de position des antennes\n ; -----------------------------------------------------------------\n\n PRO im_1dsel, klu, nof, npol, kint, sz, h, ima\n \"\"\"\n\n filtre = json.loads(f)\n\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n print(datedeb)\n #duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n print(frequence)\n integration = (filtre[\"integration\"])\n if(integration == 0):\n integration = filtre[\"itg\"]\n print(integration)\n fRH = [\"2d\" + datedeb.strftime(\"%y%m%d\") + \".01\", \"2d\" + datedeb.strftime(\n \"%y%m%d\") + \".01Z\", \"2d\" + datedeb.strftime(\"%y%m%d\") + \".01ZZ\"]\n print(fRH)\n for file in fRH:\n fname = \"/data/data_nrh/rh/\" + datedeb.strftime(\"%Y\") + \"/\" + datedeb.strftime(\n \"%m\") + \"/\" + datedeb.strftime(\"%d\") + \"/\" + file\n\n print(fname)\n uid = str(uuid.uuid4())\n\n idl = pidly.IDL('/usr/local/bin/idl')\n #$IDL = \"/home/user/idl83/bin/idl \".self::$image_rh.\" -quiet -args \".$file.\" \".json_encode($time,JSON_NUMERIC_CHECK).\" \".$frequence.\" \".$polarite.\" \".$integration.\" \".self::$uri.\" \";\n\n idl.fichier = fname\n idl.frequence = frequence\n idl.h = list(datedeb.timetuple()[3:7])\n\n idl.kint = int(integration)\n idl('@rh_common.inc')\n idl('status=RH_OPEN(fichier,/SEL,/MONO)')\n idl('print, entFI.frq')\n idl('nof = WHERE(entFI.frq EQ frequence)')\n\n idl.npol = 0\n\n idl.url = '/media/data/Public/Images/' + uid + '.png'\n\n #idl('ima = BINDGEN(1,256,256)')\n idl.sz = 256\n idl('nbmax = min([entFI.klumax+1,1000])')\n idl('nbmax=entFI.klumax+1')\n idl('ima=fltarr(sz,nbmax)')\n idl('rh_hdeb,h, klu, heurlu')\n print(idl.h)\n print(idl.heurlu)\n print(idl.klu)\n # idl.pro('IMAGE_RH',fichier,h,frequence,polarite,integration,url)\n idl('rh_hfin, h, klu, heurlu ')\n print(idl.h)\n print(idl.heurlu)\n print(idl.klu)\n\n idl.LARG = 4\n idl(\"set_plot,'PS'\")\n # PRO im_1dsel, klu, nof, npol, kint, sz, h, ima\n idl('IM_1DSEL, klu, nof, npol, kint, sz, h, ima')\n idl('print,ima')\n idl('loadct,3, /SILENT')\n idl('TVLCT, R, G, B, /GET')\n #idl('p = PLOT(ima)')\n #idl('p.Save, url, BORDER=10, RESOLUTION=300, /TRANSPARENT')\n #idl(\"write_image, url,'PNG',p,R, G, B\")\n # IDL procedure with Python argument(s)::\n #idl.pro('IMAGE_RH', range(10), range(10), xstyle=True, ystyle=True)\n print(idl.status)\n break\n\n response = {\n \"status\": \"success\",\n \"message\": \"Image loaded succesfully.\",\n \"time\": time,\n \"date\": date,\n \"url\": \"/Public/Images/\" + uid + \".png\",\n \"json\": idl.ima.tolist()\n }\n # idl.close()\n idl('exit')\n return response\n\n def getCanvasFromFile(self, f, sel):\n\n filtre = json.loads(f)\n # 2014-01-01T00:00:00.000Z\n\n \"\"\"\n function (doc, meta) {\n var d = dateToArray(doc.datetime);\n for(var k in doc.frequence){\n emit([d[3],d[4],doc.polarite,parseInt(k)],doc.frequence[k] );\n }\n }\n function(keys, values, rereduce) {\n\n var length = values.length;\n var reduceValue=0;\n for(var k in values){\n\n reduceValue = reduceValue + values[k];\n\n }\n\n return reduceValue/length;\n\n\n }\n \"\"\"\n\n # 2014-01-01T00:00:00.000Z\n\n date = datetime.datetime.strptime(\n filtre[\"date\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n time = datetime.datetime.strptime(\n filtre[\"time\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n\n \"\"\"\n cb = Couchbase.connect(bucket='NdaOld')\n r = []\n\n ViewRow objects are simple named tuples with the following fields:\n\n vr.key\n\n The key emitted by the views map function (i.e. first argument to emit.\n\n vr.value\n\n The value emitted by the views map function (i.e. second argument to emit ).\n\n vr.id\n\n The document ID of this row. The ID can be passed to get and set.\n\n vr.doc\n\n cb.query(\"dev_canvas\",\n \"default\",\n limit=3,\n mapkey_range = [\"abbaye\", \"abbaye\" + Query.STRING_RANGE_END],\n descending=True)\n cb.query(\"dev_canvas\", \"default\", query=\"limit=3&skip=1&stale=false\")\n\n\n view = cb.query(\"dev_canvas\",\n \"default\",\n limit=600\n # mapkey_range = [\"abbaye\", \"abbaye\" + Query.STRING_RANGE_END],\n # descending=True\n )\n\n\n view = views.iterator.View(cb, \"dev_canvas\", \"default\",\n limit=600,\n reduce=\"false\")\n\n url\":[{'values':[{'x':0,'y':35},{\"x\":1,\"y\":36...\n\n for result in view:\n r= r + result.value\n\n j = {\"values\":r}\n \"\"\"\n\n #pathtable = date.strftime(\"/an/2/%d/dataset\")\n\n #t = Table.read('/datadam/routine/hdf5/NdaOld.hdf5', path=pathtable)\n\n #d = pd.HDFStore('/datadam/routine/hdf5/NdaOld.hdf5')\n \"\"\"\n with h5py.File(\"/datadam/routine/hdf5/NdaOld3.hdf5\", \"r\") as f:\n print(f)\n dset = f[\"/an/2/1/dataset\"]\n print(dset)\n \"\"\"\n\n fdam = \"S\" + date.strftime(\"%y%m%d\") + \".RT1\"\n rdam = date.strftime(\"%Y\")\n fname = \"/datadam/routine/\" + rdam + \"/\" + fdam\n\n if os.path.isfile(fname):\n\n file = open(fname, \"rb\")\n bytes_read = file.read(405)\n dt = 'c,2S,2S,45S,2S,1S,2S,1S,2S,1S,6S,340S'\n c = np.fromstring(bytes_read, dtype=dt)\n\n jour = int(c[0][4])\n\n mois = int(c[0][6])\n\n an = int(c[0][8])\n\n if 59 <= an <= 99:\n an = int(an + 1900)\n else:\n an = int(an + 2000)\n\n path = \"/\" + str(an) + \"/\" + str(mois) + \"/\" + str(jour)\n\n # dt=np.int8\n dt = '405B'\n\n # dt = np.uint8\n\n data = np.fromfile(fname, dtype=dt)\n\n data = np.delete(data, 0, 0)\n data = np.delete(data, [0, 1, 2, 3], axis=1)\n\n data = np.delete(data, [400], axis=1)\n\n # [57469, 405]\n dd = data[0:2400, :]\n\n # y = np.reshape(dd,(dd.shape[1],dd.shape[0]))\n # z = dd.flatten()\n # ddd = np.zeros((3,z.shape[0]))\n #ddd = np.concatenate((z,z,z,z),axis=1)\n # ddd[0,:]=z\n # ddd[1,:]=z\n # ddd[2,:]=z\n # dddd = np.mat(ddd)\n # e = dddd.getT\n \"\"\"\n R - The color red (from 0-255)\n G - The color green (from 0-255)\n B - The color blue (from 0-255)\n A - The alpha channel (from 0-255; 0 is transparent and 255 is fully visible)\n \"\"\"\n r = np.repeat(dd, 4)\n print(r.shape)\n\n rrr = np.reshape(r, (r.shape[0] / 4, 4))\n rrr[:, 3] = 255\n\n rrr = rrr.flatten()\n print(rrr.shape)\n # dddzzz= np.reshape(ddd,(ddd.shape[1],ddd.shape[0]))\n # zzz = np.concatenate((dd,dd),axis=0)\n\n # dz = np.zeros((data.shape[0],3))\n\n # dz[:,0]= an\n # dz[:,1]= mois\n # dz[:,2]= jour\n\n # l= data.flatten()\n\n # h = np.append(dz, data, axis=1).astype(int)\n\n \"\"\"\n print(f.__contains__(path))\n if not(f.__contains__(path)):\n print(\"create\")\n sb = f.create_group(path)\n else:\n print(\"exist\")\n sb = f.get(path)\n \"\"\"\n\n # sb = f.require_group(path)\n\n #date = datetime.datetime(an,mois,jour,d[0],d[1],d[2],d[3])\n\n # mm = Image.fromarray(data,\"L\")\n\n # zz=ImageOps.colorize(mm, '#000000','#ffffff')\n # zz.show()\n # plt.imshow(data)\n # plt.show()\n # print(img)\n #pil_im = Image.fromarray(np.uint8(l))\n # pil_im.show()\n rr = r.flatten()\n\n response = {\n \"status\": \"success\",\n \"message\": \"Image loaded succesfully.\",\n \"url\": {\"values\": rrr.tolist()}\n }\n return response\n\n def find(self, name, obj):\n\n return obj\n\n def getCanvas(self, f, sel):\n \"\"\"\n :param f:\n :param sel:\n :return:\n \"\"\"\n\n filtre = json.loads(f)\n\n # 2014-01-01T00:00:00.000Z\n\n date = datetime.datetime.strptime(\n filtre[\"date\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n time = datetime.datetime.strptime(\n filtre[\"time\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n\n # path = date.strftime( \"%Y/%m/%d\")\n t = date.timetuple()\n path = \"/\" + str(t[0]) + \"/\" + str(t[1]) + \"/\" + str(t[2])\n\n values = []\n\n out = np.empty((600, 400), dtype=np.uint8)\n\n with h5py.File(\"/datadam/routine/hdf5/NdaOldPython2.hdf5\", \"r\") as hdf5:\n #grp = hdf5[\"/2014/1\"]\n\n # grp.visit(values.append)\n #v = grp.visititems(self.find)\n # v.read_direct(out,np.s_[0:600,1,7:407])\n # print(out)\n\n print(\"open\")\n # print(hdf5.keys())\n e = False\n if path in hdf5.keys():\n print(path)\n g = hdf5[path]\n e = True\n print(g)\n v = int(g[0:600, 1, 7:407])\n\n response = {\n \"status\": \"success\",\n \"message\": \"Image loaded succesfully.\",\n \"url\": [{\"values\": v.tolist()}]\n }\n\n else:\n print(path)\n g = hdf5[path]\n e = True\n # Image.fromarray(data,\"L\")\n\n v = g[:, 1, 7:407]\n vv = v.astype(np.uint8)\n print(vv.dtype)\n im = cv2.imdecode(vv, 0)\n cv2.imshow(\"test\", vv)\n w = cv2.cvtColor(vv, cv2.COLOR_GRAY2BGR)\n cv2.imwrite('image.png', w)\n response = {\n \"status\": \"error\",\n \"message\": \"Not Found\",\n \"url\": ima.tolist()\n }\n\n return response\n\n def getViewFromFile(self, f, sel):\n \"\"\"\n :param f:\n :param sel:\n :return:\n \"\"\"\n\n filtre = json.loads(f)\n\n # 2014-01-01T00:00:00.000Z\n\n date = datetime.datetime.strptime(\n filtre[\"date\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n time = datetime.datetime.strptime(\n filtre[\"time\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = filtre[\"dure\"]\n frequence = filtre[\"frequence\"]\n\n \"\"\"\n function (doc, meta) {\n var d = dateToArray(doc.datetime);\n for(var k in doc.data){\n emit([d[3],d[4],d[5],doc.polarite,parseInt(k)],{'x':parseInt(k),'y':doc.data[k]} );\n }\n }\n \"\"\"\n fdam = \"S\" + date.strftime(\"%y%m%d\") + \".RT1\"\n rdam = date.strftime(\"%Y\")\n fname = \"/datadam/routine/\" + rdam + \"/\" + fdam\n\n if os.path.isfile(fname):\n\n file = open(fname, \"rb\")\n bytes_read = file.read(405)\n dt = 'c,2S,2S,45S,2S,1S,2S,1S,2S,1S,6S,340S'\n c = np.fromstring(bytes_read, dtype=dt)\n\n jour = int(c[0][4])\n\n mois = int(c[0][6])\n\n an = int(c[0][8])\n\n if 59 <= an <= 99:\n an = int(an + 1900)\n else:\n an = int(an + 2000)\n\n path = \"/\" + str(an) + \"/\" + str(mois) + \"/\" + str(jour)\n\n # dt=np.int8\n dt = '405B'\n\n # dt = np.uint8\n\n data = np.fromfile(fname, dtype=dt)\n\n data = np.delete(data, 0, 0)\n data = np.delete(data, [0, 1, 2, 3], axis=1)\n\n data = np.delete(data, [400], axis=1)\n\n # [57469, 405]\n dd = data[0:600, :]\n\n line = data[100, :]\n\n \"\"\"\n np.set_printoptions(formatter={'all':lambda x: 'x: '+str(x)})\n\n print(line)\n n = 0\n\n values = np.array2string(\n\n dd,formatter={\n \"all\":lambda x:\n\n \"{'y': \"+str(x)+\"},\"\n }\n )\n \"\"\"\n values = []\n it = np.nditer([dd, None])\n for x, y in it:\n\n values.append({'x': int(x), 'y': int(y)})\n # print(it.operands[1])\n\n \"\"\"\n all=[]\n m=0\n for v in np.nditer(data):\n values = []\n n = 0\n for y in np.nditer(data[v,:]):\n n +=1\n values.append({'x':int(n),'y':int(y)})\n all.append({\"values\":values})\n print(all)\n \"\"\"\n \"\"\"\n cb = Couchbase.connect(bucket='NdaOld')\n r = []\n view = views.iterator.View(cb, \"dev_default\", \"spectre\",\n limit=400,\n reduce=\"false\")\n\n url\":[{'values':[{'x':0,'y':35},{\"x\":1,\"y\":36...\n\n for result in view:\n r.append(result.value)\n j = [{\"values\":r}]\n \"\"\"\n response = {\n \"status\": \"success\",\n \"message\": \"Image loaded succesfully.\",\n \"url\": [{\"values\": values}]\n }\n\n return response\n\n def setImportLog(self, f, sel):\n\n logging.basicConfig(filename='/data/var_nrh/nrh.import.log', level=logging.DEBUG,\n format='%(asctime)s -- %(name)s -- %(levelname)s -- %(message)s')\n filtre = json.loads(f)\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n duree = int(filtre[\"dure\"])\n frequence = filtre[\"frequence\"]\n recepteur = filtre[\"recepteur\"]\n\n oneyear = datetime.timedelta(days=365)\n\n DU = datetime.timedelta(days=duree * 365)\n\n #DD = datedeb\n #DF = datedeb + DU\n print(\"boucle\")\n counterYear = 0\n\n es = Elasticsearch(['master-rsdb'])\n es.indices.create(index='test', ignore=400)\n print(\"boucle date debut : \", datedeb)\n\n while counterYear < duree * 365:\n\n print(\"Import brut\", counterYear)\n data_dir = self.DATA_DIR\n file = sel + \"_\" + datedeb.strftime(\"%Y\")\n fname = data_dir + \"suivi/\" + file\n print(fname)\n Year = int(datedeb.strftime(\"%Y\"))\n pattern = '^(\\d{2})\\/(\\d{2})\\s|\\t'\n if os.path.isfile(fname):\n f = open(fname, \"r\")\n n = 0\n\n t = f.tell()\n p = f.seek(1100)\n print(\"coucou\")\n try:\n for line in f.read().split('\\n'):\n\n if re.match(pattern, line):\n print(pattern, line)\n l = re.split(pattern, line, 0)\n\n if l[0] == \"\":\n if l[1]:\n\n if n:\n body = {\"query\":\n {\"bool\":\n {\"must\":\n [{\"match\":\n {\"@timestamp\": dd}\n }]\n }\n }\n }\n\n res = es.search(index=\"test\",\n doc_type=sel,\n body=body, fields='_id')\n\n if res['hits']['total'] > 0:\n print(\"update\")\n _id = res['hits']['hits'][0]['_id']\n res = es.index(\n index=\"test\", doc_type=sel, id=_id, body=entete)\n\n else:\n\n print(\"new\")\n res = es.index(\n index=\"test\", doc_type=sel, body=entete)\n n = 0\n\n entete = {}\n try:\n dd = datetime.datetime(\n Year, int(l[2]), int(l[1]), 0, 0, 0, 0, pytz.UTC)\n\n entete['file'] = fname\n entete['name'] = file\n entete['@timestamp'] = dd\n entete['french'] = \"\"\n entete['english'] = \"\"\n for i in range(3, len(l)):\n if l[i]:\n entete['french'] += l[i]\n n = n + 1\n print(entete['french'])\n\n #self.translate('bonjour')\n #entete['english'] = translate.translator('fr', 'en','bonjour')\n\n print(entete['english'])\n # pass\n except Exception as e:\n raise\n print(\"error\")\n\n else:\n n = n + 1\n entete['french'] += line\n\n #entete['english'] = translator('fr', 'en',entete['french'])\n print(entete)\n #exit()\n\n # res = es.search(index=\"nrh\", doc_type='entFI', body={\"query\": { \"term\":{\n #'@timestamp':datedeb}}},fields='_id')\n except:\n\n print(\"Unexpected error:\", sys.exc_info()[0])\n pass\n f.close()\n counterYear += 365\n\n datedeb = datedeb + oneyear\n\n response = {\n \"status\": \"success\",\n \"message\": \"Import succesful\",\n\n \"time\": t,\n \"date\": d\n\n }\n return response\n\n def setImport1d(self, f, sel):\n print(\"boucle\")\n logging.basicConfig(filename='/data/var_nrh/nrh.import.log', level=logging.DEBUG,\n format='%(asctime)s -- %(name)s -- %(levelname)s -- %(message)s')\n #filtre = json.loads(f)\n filtre = f\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n datefin = datetime.datetime.strptime(\n filtre[\"datefin\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n #datefin = datetime.datetime.strptime(filtre[\"datefin\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = int(filtre[\"dure\"])\n frequence = filtre[\"frequence\"]\n recepteur = filtre[\"recepteur\"]\n session = requests.Session()\n headers = {\n \"Content-Type\": \"application/json\", \"Content-Encoding\": \"gzip\"\n }\n session.headers.update(headers)\n oneday = datetime.timedelta(days=1)\n #counterDay = 0\n #DU = datetime.timedelta(days=duree)\n\n DD = datedeb\n #DF = datedeb + DU\n DF = datefin\n\n #oneday = datetime.timedelta(days=1)\n\n counterDay = 1\n\n es = Elasticsearch(['master-rsdb'])\n es.indices.create(index='1d', ignore=400)\n idl = pidly.IDL('/usr/local/bin/idl', long_delay=0.05)\n #sudo ssh -N -f -L222:radio-monitoring.obspm.fr:22 services@ambari-rsdb.obs-nancay.fr\n\n\n print(\"boucle date debut : \", datedeb)\n try:\n client = paramiko.SSHClient()\n client.set_missing_host_key_policy(paramiko.AutoAddPolicy() )\n #client.connect('ambari-rsdb.obs-nancay.fr', username='services', password='oulid12mdlp')\n # ssh -N -f -L2222:radio-monitoring.obspm.fr:22 services@ambari-rsdb.obs-nancay.fr\n #\n client.connect('127.0.0.1', username='ac-nrhmeudon', password='dep0uilleuz',port=2222)\n print(client)\n sftpclient = client.open_sftp()\n except:\n\n print(\"Unexpected error:\", sys.exc_info()[0])\n\n #while counterDay < duree:\n print(\"coucou\")\n\n while DD < DF:\n\n print(\"Import 1d\", DD)\n f = frequence\n\n integration = (filtre[\"integration\"])\n # ext = [\"01\",\"01Z\",\"01ZZ\"]\n #ext = []\n\n #ext.append(\"fts\")\n #print(f)\n #fRH = \"2d\" + date.strftime(\"%y%m%d\") +\".01\"\n #fname = \"/data/data_nrh/rh/\"+date.strftime(\"%Y\")+\"/\"+date.strftime(\"%m\")+\"/\"+ date.strftime(\"%d\")+\"/\"+fRH\n #sftp://ac-nrhmeudon@radio-monitoring.obspm.fr/nas_nrh/festival/fits/2002/0201\n fserveur = \"ac-nrhmeudon@radio-monitoring.obspm.fr\"\n fddir = \"/nas_nrh/festival/fits_120s/\"+DD.strftime(\"%Y\")+\"/\"+DD.strftime(\"%y%m\")\n #print(fddir)\n fldir = \"/tmp\"\n n = 0\n #stdin, stdout, stderr = client.exec_command(\"cd \"+fddir)\n #response = stdout.readlines()\n #errormsg = stderr.read()\n #stdin, stdout, stderr = client.exec_command(\"ls -x \"+fddir)\n sftpclient.chdir(fddir)\n #response = stdout.readlines()\n #errormsg = stderr.read()\n\n response = sftpclient.listdir(path='.')\n\n #(^nrh2_1509_h[0-9]0_20120306_[a-zA-Z0-9_.]+fts$)\n filen = \"(^nrh2_\" + str(f) + \"_h[0-9]0_\" + \\\n DD.strftime(\"%Y%m%d\") + \"_[a-zA-Z0-9_.]+fts$)\"\n print(filen)\n #nrh2_1509_h60_20120306_102553c05_i.fts\n #nrh2_1509_h60_20120306_083336c05_i.fts\n\n fname = fddir +\"/\"+ filen\n #print(fname)\n\n #for file in os.listdir(fdir):\n for index, item in enumerate(response):\n print (index, item)\n #for file in response:\n # print(fil)\n\n #if fnmatch.fnmatch(item, filen):\n if re.search(filen, item):\n #print(item,\"@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@\")\n\n remotepath = item\n localpath = fldir+\"/\"+item\n isdir = sftpclient.getcwd()\n #print(remotepath,localpath,isdir)\n try:\n res = sftpclient.get(remotepath, localpath)\n\n idl('.Reset_Session')\n\n idl('@rh_common.inc')\n\n idl.filename = localpath\n idl.tstart = 0\n idl.tstop = 36468000000\n idl.tbg0 = 1\n idl.tbg1 = 1\n idl('nrh1d_rd, filename, tew, tsn, taxis, ewaxis, snaxis, tstart, tstop ,tbg0, tbg1, label, date, freq,/dalmiro')\n idl('tew=tew')\n idl('tsn=tsn')\n idl('taxis=taxis')\n idl('ewaxis=ewaxis')\n idl('snaxis=snaxis')\n\n dd = datetime.date(idl.date[0],idl.date[1],idl.date[2])\n\n tew = np.array(idl.tew)\n tsn = np.array(idl.tsn)\n taxis = np.array(idl.taxis)\n\n ewaxis = np.array(idl.ewaxis)\n snaxis = np.array(idl.snaxis)\n\n ntew = np.transpose(tew)\n ntsn = np.transpose(tsn)\n\n itaxis = np.around(taxis)\n\n ds = int(dd.strftime(\"%s\"))\n\n ntaxis = (itaxis / 1000 + ds)\n\n INDEX = pd.to_datetime(ntaxis, unit='s', utc=False)\n ii = INDEX.to_pydatetime()\n data = {}\n #print(\"coucou\")\n ori = \"ew\"\n data[\"ew\"]= pd.DataFrame(tew, index=ewaxis, columns=ii)\n data[\"sn\"]= pd.DataFrame(tsn, index=snaxis, columns=ii)\n #print(\"coucou\")\n\n url = 'http://opentsdb-rsdb:4242/api/put'\n z = 0\n\n\n for d in data:\n print(d)\n ori = d\n\n for i, v in data[d].iterrows():\n\n stream = []\n for index, value in v.iteritems():\n #print(index)\n try:\n dt = datetime.datetime.strptime(\n str(index) + \"+0000\", \"%Y-%m-%d %H:%M:%S.%f%z\")\n except:\n dt = datetime.datetime.strptime(\n str(index) + \".000000+0000\", \"%Y-%m-%d %H:%M:%S.%f%z\")\n\n #print(dt)\n if z == 0:\n fdatdeb = dt\n\n fdatfin = dt\n z = z + 1\n #print(z,fdatdeb)\n #exit()\n val = {}\n val[\"metric\"] = \"nrh.\"+d\n\n val[\"timestamp\"] = round(\n dt.timestamp())\n val[\"value\"] = float(value)\n\n tags = {\"coor\": float(i),\"int\":120,\"f\":frequence,\"ori\":d}\n val[\"tags\"] = tags\n #mean = mean + int(data[i, k + 4])\n #count = count + 1\n #tab[int(k)] = int(data[i, k + 4])\n # print(\"--------------------------------\",val)\n stream.append(val)\n\n #print(stream ,\"+++++++++++++++\")\n #print(stream)\n cc = json.dumps(stream)\n #print(cc)\n s_in = binascii.a2b_qp(\n json.dumps(stream))\n\n s_out = gzip.compress(s_in)\n\n rep = session.post(url, data=s_out)\n s = rep.status_code\n #print(s)\n #entete['@timestamp'] = datedebut\n\n # res = es.index(\n # index=\"nda\", doc_type='old', body=entete,\n # timeout=30)\n\n\n entete = {}\n\n entete['typ'] = \"1d\"\n\n\n\n entete['hdeb'] = fdatdeb\n\n entete['hfin'] = fdatfin\n\n\n entete['@timestamp'] = fdatdeb\n entete['file'] = item\n entete['ori'] = d\n entete['frq'] = str(f)\n entete['int'] = 120\n #print(entete)\n #entete['descrip'] = idl.desc\n\n\n body = {\"query\":\n {\"bool\":\n {\"must\":\n [{\"match\":\n {\"hdeb\": fdatdeb}\n },\n {\"term\":\n {\"file\": item.lower()}\n },\n {\"term\":\n {\"ori\": d}\n }\n ]\n }\n }\n }\n\n #print(body)\n # res = es.search(index=\"nrh\", doc_type='entFI', body={\"query\": { \"term\":{\n #'@timestamp':datedeb}}},fields='_id')\n res = es.search(index=\"1d\",\n doc_type='1d',\n body=body, fields='_id')\n #print(res)\n if res['hits']['total'] > 0:\n print(\"update\")\n _id = res['hits']['hits'][0]['_id']\n res = es.index(\n index=\"1d\", doc_type='1d', id=_id, body=entete)\n\n else:\n\n print(\"new\")\n res = es.index(\n index=\"1d\", doc_type='1d', body=entete)\n os.remove(localpath)\n except:\n\n print(\"Unexpected error:\", sys.exc_info()[0])\n idl('exit')\n\n counterDay += 1\n\n DD = DD + oneday\n\n response = {\n \"status\": \"success\",\n \"message\": \"Import succesful\",\n \"date\": DD.strftime(\"%Y%m%d\")\n\n }\n# idl.close()\n idl('exit')\n return response\n\n def setImport(self, f, sel):\n\n logging.basicConfig(filename='/data/var_nrh/nrh.import.log', level=logging.DEBUG,\n format='%(asctime)s -- %(name)s -- %(levelname)s -- %(message)s')\n filtre = json.loads(f)\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%f%z\")\n\n #datefin = datetime.datetime.strptime(filtre[\"datefin\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = int(filtre[\"dure\"])\n frequence = filtre[\"frequence\"]\n recepteur = filtre[\"recepteur\"]\n\n oneday = datetime.timedelta(days=1)\n counterDay = 0\n DU = datetime.timedelta(days=duree)\n\n DD = datedeb\n DF = datedeb + DU\n\n oneday = datetime.timedelta(days=1)\n\n counterDay = 1\n\n es = Elasticsearch(['master-rsdb'])\n es.indices.create(index='nrh', ignore=400)\n idl = pidly.IDL('/usr/local/bin/idl', long_delay=0.05)\n\n print(\"boucle date debut : \", datedeb)\n\n while counterDay < duree:\n\n print(\"Import brut\", counterDay)\n\n f = frequence\n\n integration = (filtre[\"integration\"])\n # ext = [\"01\",\"01Z\",\"01ZZ\"]\n ext = []\n\n for i in list(range(1, 10, 1)):\n\n ext.append(\"0\" + str(i))\n ext.append(\"0\" + str(i) + \"Z\")\n ext.append(\"0\" + str(i) + \"ZZ\")\n\n #fRH = \"2d\" + date.strftime(\"%y%m%d\") +\".01\"\n #fname = \"/data/data_nrh/rh/\"+date.strftime(\"%Y\")+\"/\"+date.strftime(\"%m\")+\"/\"+ date.strftime(\"%d\")+\"/\"+fRH\n fdir = \"/data/data_nrh/nrh/\" + \\\n datedeb.strftime(\"%Y\") + \"/\" + datedeb.strftime(\"%m\") + \\\n \"/\" + datedeb.strftime(\"%d\") + \"/\"\n n = 0\n for e in ext:\n\n fil = \"2d\" + datedeb.strftime(\"%y%m%d\") + \".\" + e\n fname = fdir + fil\n\n if os.path.isfile(fname):\n\n print(e, fname)\n #idl.h = list(time.timetuple()[3:7])\n\n # idl.kint=int(integration)\n # idl.nof=int(f)\n # idl.npol=0\n\n try:\n\n idl('.Reset_Session')\n\n idl('@rh_common.inc')\n\n idl.fichier = fname\n idl('status = RH_OPEN(fichier,/SEL,/MONO)')\n\n idl('s = status')\n\n if idl.s:\n print(\"ok\", fname)\n else:\n print(\"PB\", fname)\n continue\n\n idl('IF status THEN BEGIN')\n\n idl('typ=entFI.typ')\n\n idl('dat=entFI.dat')\n\n d = datetime.date(idl.dat[2], idl.dat[1], idl.dat[0])\n idl('hdeb=entFI.hdeb')\n idl('hfin=entFI.hfin')\n\n try:\n\n fdatedeb = datetime.datetime(idl.dat[2], idl.dat[1], idl.dat[0], idl.hdeb[\n 0], idl.hdeb[1], idl.hdeb[2], idl.hdeb[3] * 10000, pytz.UTC)\n\n #sdeb = ((idl.hdeb[0]*60 + idl.hdeb[1])*60 + idl.hdeb[2])*60 + idl.hdeb[3]/100\n fdatefin = datetime.datetime(idl.dat[2], idl.dat[1], idl.dat[0], idl.hfin[\n 0], idl.hfin[1], idl.hfin[2], idl.hfin[3] * 10000, pytz.UTC)\n\n #sfin = ((idl.hfin[0]*60 + idl.hfin[1])*60 + idl.hfin[2])*60 + idl.hfin[3]/100\n\n except ValueError as e:\n print(e)\n logging.error(\n 'IDL %s: %s %s %s', idl.hdeb.tolist(), idl.hfin.tolist(), file, e)\n continue\n except:\n print(\"problem break\")\n continue\n\n idl('frq=entFI.frq')\n idl('itg=entFI.itg')\n idl('dec=entFI.dec')\n idl('hg=entFI.hg')\n idl('trj=entFI.trj')\n idl('comp=entFI.comp')\n idl('cyclms=entFI.cyclms')\n idl('d_obs=entFI.d_obs')\n\n idl('corel=entFI.corel')\n\n entete = {}\n entete['typ'] = \"entFI\"\n entete['typ'] = idl.typ.tolist()\n entete['frq'] = idl.frq.tolist()\n entete['itg'] = idl.itg.tolist()\n entete['dec'] = idl.dec.tolist()\n entete['hg'] = idl.hg.tolist()\n entete['hdeb'] = fdatedeb\n #entete['sdeb']= sdeb\n entete['hfin'] = fdatefin\n #entete['sfin']= sfin\n #entete['hdeb']= timedeb.strftime(\"%H:%M:%S.%f\")\n #entete['hfin']= timefin.strftime(\"%H:%M:%S.%f\")\n entete['trj'] = idl.trj.tolist()\n entete['comp'] = idl.comp.tolist()\n entete['cyclms'] = idl.cyclms.tolist()\n entete['d_obs'] = idl.d_obs.tolist()\n\n entete['corel'] = idl.corel.tolist()\n\n entete['@timestamp'] = fdatedeb\n entete['file'] = fil\n entete['ext'] = e\n entete['ord'] = n\n #entete['descrip'] = idl.descrip.tolist()\n #entete['activ'] = idl.activ.tolist()\n #entete['pannes'] = idl.pannes.tolist()\n\n \"\"\"\n DATE-OBS : date de début d’observation ; 2014-02-01\n TIME-OBS : heure de début d’observation ;08 :18 :02.000\n DATE-END : date de fin d’observation\n TIME-END : heure de fin d’observation\n\n PHYSPARA : ‘I+V’ ; paramètres observés (Stokes)\n\n OBSERVATORY : ‘Observatoire de Paris – Nançay’ (a des noms divers selon les fichiers)\n\n OBS-TYPE : ‘radio’\n OBS-SUBTYPE : ‘visibility’\n TELESCOPE : ‘ radio interferometer’\n INSTRUMENT : ‘NRH’\n\n SOURCE : ‘SUN’, ou GYG,CASS,TAUR, VIRG, HYDR, DAB\n FREQ-START : fréquence de départ, en MHZ\n FREQ-STOP : fréquence de fin, MHZ\n BUNITS : ‘SFU’\n EXP-TIME : sampling time, millisec\n TIME-STEP : integration time, millisec\n Entetes fixe\n\n \"\"\"\n entete['physpara'] = 'I+V'\n entete['observatory'] = 'Observatoire de Paris – Nançay'\n entete['obs-type'] = 'radio'\n entete['obs-subtype'] = 'visibilité'\n entete['telescope'] = 'radio interferometer'\n entete['instrument'] = 'nrh'\n entete['source'] = 'sun'\n entete['bunits'] = 'sfu'\n\n except:\n idl('exit')\n print(\"Unexpected error:\", sys.exc_info()[0])\n #dd = datetime.datetime(idl.dat[2],idl.dat[1], idl.dat[0],idl.hdeb[0],idl.hdeb[1], idl.hdeb[2], idl.hdeb[3]*1000, pytz.UTC)\n\n body = {\"query\":\n {\"bool\":\n {\"must\":\n [{\"match\":\n {\"hdeb\": fdatedeb}\n },\n {\"term\":\n {\"file\": fil.lower()}\n }]\n }\n }\n }\n\n # print(entete)\n \"\"\"\n body ={\"query\": { \"term\":{\n '@timestamp':fdatedeb}}}\n\n GET _search\n {\"bool\": {\n \"must\": [{\n \"term\": {\n \"timestamp\": \"2013/12/06T08:12:38\"\n }\n }\n , {\n \"term\": {\n \"file\": \"2d131206.01\"\n\n }\n }\n ]\n }}\n \"\"\"\n # res = es.search(index=\"nrh\", doc_type='entFI', body={\"query\": { \"term\":{\n #'@timestamp':datedeb}}},fields='_id')\n res = es.search(index=\"nrh\",\n doc_type='entFI',\n body=body, fields='_id')\n\n if res['hits']['total'] > 0:\n print(\"update\")\n _id = res['hits']['hits'][0]['_id']\n res = es.index(\n index=\"nrh\", doc_type='entFI', id=_id, body=entete)\n\n else:\n\n print(\"new\")\n res = es.index(\n index=\"nrh\", doc_type='entFI', body=entete)\n\n n += 1\n\n print(\"Import file\")\n\n if os.path.isdir(fdir):\n files = os.listdir(fdir)\n print(files)\n\n # ['1v030504.01', '2c030504.01', '2d030504.01', '2d030504.01Z', '2d030504.01ZZ', '2i030504.01', '2q030504.01', 'bilan_030504.01', 'nrh04052003_1640.mpg', 'nrh04052003_3270.mpg']\n for f in files:\n\n dispo = {}\n dispo['file'] = f.lower()\n m = f.split('.')\n dispo['ext'] = m[1]\n dispo['name'] = m[0]\n t = dispo['name'].split('_')\n if dispo['ext'] == \"mpg\":\n dispo['type'] = dispo['ext']\n dispo['typ'] = 2\n #dispo['frq'] = {0: t[1]}\n dispo['frq'] = []\n dispo['frq'].append(t[1])\n\n elif dispo['ext'] == \"fts\":\n dispo['type'] = \"fits\"\n dispo['typ'] = 4\n elif t[0] == \"bilan\":\n dispo['type'] = \"txt\"\n dispo['typ'] = 3\n elif t[0] == \"tt\":\n dispo['type'] = \"tt\"\n dispo['typ'] = 5\n else:\n dispo['type'] = m[0][:2]\n dispo['typ'] = int(m[0][:1])\n\n dispo['@timestamp'] = fdatedeb\n # print(datedeb)\n body = {\"query\":\n {\"bool\":\n {\"must\":\n [{\"match\":\n {\"@timestamp\": fdatedeb}\n },\n {\"match\":\n {\"file\": f.lower()}\n }]\n }\n }\n }\n res = es.search(index=\"nrh\",\n doc_type='files',\n body=body, fields='_id')\n\n if res['hits']['total'] > 0:\n print(\"----------------------------------- update\")\n _id = res['hits']['hits'][0]['_id']\n res = es.index(\n index=\"nrh\", doc_type='files', id=_id, body=dispo)\n\n else:\n\n print(\"----------------------------------------------- new\")\n\n res = es.index(\n index=\"nrh\", doc_type='files', body=dispo)\n\n counterDay += 1\n\n datedeb = datedeb + oneday\n\n response = {\n \"status\": \"success\",\n \"message\": \"Import succesful\",\n\n \"time\": t,\n \"date\": d\n\n }\n # idl.close()\n idl('exit')\n return response\n\n def setVideo(self, f, sel):\n logging.basicConfig(filename='/var/www/backend/nrh.import2.log', level=logging.DEBUG,\n format='%(asctime)s -- %(name)s -- %(levelname)s -- %(message)s')\n filtre = json.loads(f)\n datedeb = datetime.datetime.strptime(\n filtre[\"datedeb\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n #datefin = datetime.datetime.strptime(filtre[\"datefin\"], \"%Y-%m-%dT%H:%M:%S.%fZ\")\n duree = int(filtre[\"dure\"])\n frequence = filtre[\"frequence\"]\n recepteur = filtre[\"recepteur\"]\n\n oneday = datetime.timedelta(days=1)\n counterDay = 0\n DU = datetime.timedelta(days=duree)\n DD = date\n DF = date + DU\n\n oneday = datetime.timedelta(days=1)\n\n counterDay = 1\n\n es = Elasticsearch(['master-rsdb'])\n es.indices.create(index='nrh', ignore=400, timeout=30)\n idl = pidly.IDL('/usr/local/bin/idl', long_delay=0.05)\n\n print(\"boucle date debut : \", date)\n\n while counterDay < duree:\n\n print(\"Import \", counterDay)\n\n f = frequence\n integration = (filtre[\"integration\"])\n ext = [\"1640\"]\n #fRH = \"2d\" + date.strftime(\"%y%m%d\") +\".01\"\n #fname = \"/data/data_nrh/rh/\"+date.strftime(\"%Y\")+\"/\"+date.strftime(\"%m\")+\"/\"+ date.strftime(\"%d\")+\"/\"+fRH\n # nrh09092006_1640.mpg\n n = 0\n for e in ext:\n file = \"nrh\" + date.strftime(\"%d%m%Y\") + \"_\" + e + \".mpg\"\n\n # file = \"2d\" + date.strftime(\"%y%m%d\") + \".\" + e\n fname = \"/data/data_nrh/rh/\" + \\\n date.strftime(\"%Y\") + \"/\" + date.strftime(\"%m\") + \\\n \"/\" + date.strftime(\"%d\") + \"/\" + file\n\n if os.path.isfile(fname):\n\n #idl.h = list(time.timetuple()[3:7])\n\n # idl.kint=int(integration)\n # idl.nof=int(f)\n # idl.npol=0\n try:\n idl('.Reset_Session')\n idl('@rh_common.inc')\n idl.fichier = fname\n idl('status = RH_OPEN(fichier,/SEL,/MONO)')\n idl('s = status')\n\n if idl.s:\n print(\"ok\", fname)\n else:\n print(\"PB\", fname)\n break\n idl('IF status THEN BEGIN')\n\n idl('typ=entFI.typ')\n\n idl('dat=entFI.dat')\n\n d = datetime.date(idl.dat[2], idl.dat[1], idl.dat[0])\n idl('hdeb=entFI.hdeb')\n idl('hfin=entFI.hfin')\n\n try:\n\n datedeb = datetime.datetime(idl.dat[2], idl.dat[1], idl.dat[0], idl.hdeb[\n 0], idl.hdeb[1], idl.hdeb[2], idl.hdeb[3] * 1000, pytz.UTC)\n sdeb = ((idl.hdeb[0] * 60 + idl.hdeb[1]) *\n 60 + idl.hdeb[2]) * 60 + idl.hdeb[3] / 100\n datefin = datetime.datetime(idl.dat[2], idl.dat[1], idl.dat[0], idl.hfin[\n 0], idl.hfin[1], idl.hfin[2], idl.hfin[3] * 1000, pytz.UTC)\n\n sfin = ((idl.hfin[0] * 60 + idl.hfin[1]) *\n 60 + idl.hfin[2]) * 60 + idl.hfin[3] / 100\n\n except ValueError as e:\n print(e)\n logging.error(\n 'IDL %s: %s %s %s', idl.hdeb.tolist(), idl.hfin.tolist(), file, e)\n break\n except:\n print(\"problem break\")\n break\n idl('frq=entFI.frq')\n idl('itg=entFI.itg')\n idl('dec=entFI.dec')\n idl('hg=entFI.hg')\n idl('trj=entFI.trj')\n idl('comp=entFI.comp')\n idl('cyclms=entFI.cyclms')\n idl('d_obs=entFI.d_obs')\n\n idl('corel=entFI.corel')\n\n entete = {}\n entete['typ'] = \"entFI\"\n entete['typ'] = idl.typ.tolist()\n entete['frq'] = idl.frq.tolist()\n entete['itg'] = idl.itg.tolist()\n entete['dec'] = idl.dec.tolist()\n entete['hg'] = idl.hg.tolist()\n entete['hdeb'] = datedeb\n entete['sdeb'] = sdeb\n entete['hfin'] = datefin\n entete['sfin'] = sfin\n #entete['hdeb']= timedeb.strftime(\"%H:%M:%S.%f\")\n #entete['hfin']= timefin.strftime(\"%H:%M:%S.%f\")\n entete['trj'] = idl.trj.tolist()\n entete['comp'] = idl.comp.tolist()\n entete['cyclms'] = idl.cyclms.tolist()\n entete['d_obs'] = idl.d_obs.tolist()\n\n entete['corel'] = idl.corel.tolist()\n\n entete['@timestamp'] = datedeb\n entete['file'] = file\n entete['ext'] = e\n entete['ord'] = n\n except:\n idl('exit')\n print(\"Unexpected error:\", sys.exc_info()[0])\n\n body = {\"query\":\n {\"bool\":\n {\"must\":\n [{\"term\":\n {\"@timestamp\": datedeb}\n },\n {\"term\":\n {\"file\": file.lower()}\n }]\n }\n }\n }\n # print(body)\n # print(entete)\n \"\"\"\n body ={\"query\": { \"term\":{\n '@timestamp':datedeb}}}\n\n GET _search\n {\"bool\": {\n \"must\": [{\n \"term\": {\n \"timestamp\": \"2013/12/06T08:12:38\"\n }\n }\n , {\n \"term\": {\n \"file\": \"2d131206.01\"\n\n }\n }\n ]\n }}\n \"\"\"\n # res = es.search(index=\"nrh\", doc_type='entFI', body={\"query\": { \"term\":{\n #'@timestamp':datedeb}}},fields='_id')\n res = es.search(index=\"nrh\",\n doc_type='entFI',\n body=body, fields='_id')\n print(res)\n\n if res['hits']['total'] > 0:\n print(\"update\")\n _id = res['hits']['hits'][0]['_id']\n res = es.index(index=\"nrh\", doc_type='entFI',\n id=_id, body=entete, timeout=30)\n # print(res)\n else:\n\n # print(\"new\")\n res = es.index(\n index=\"nrh\", doc_type='entFI', body=entete, timeout=30)\n\n n += 1\n\n counterDay += 1\n\n date = date + oneday\n\n response = {\n \"status\": \"success\",\n \"message\": \"Import succesful\",\n\n \"time\": t,\n \"date\": d\n\n }\n # idl.close()\n idl('exit')\n return response\n", "sub_path": "back_nsa/app/Recepteur/nrh.py", "file_name": "nrh.py", "file_ext": "py", "file_size_in_byte": 83888, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "77", "api": [{"api_name": "selenium.webdriver.Firefox", "line_number": 86, "usage_type": "call"}, {"api_name": "selenium.webdriver", "line_number": 86, "usage_type": "name"}, {"api_name": "time.sleep", "line_number": 95, "usage_type": "call"}, {"api_name": "bs4.BeautifulSoup", "line_number": 101, "usage_type": "call"}, {"api_name": "app.Serveur.opentsdb.opentsdb", "line_number": 138, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 145, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 196, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 196, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 199, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 199, "usage_type": "attribute"}, {"api_name": "time.mktime", "line_number": 218, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 219, "usage_type": "call"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 224, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 227, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 227, "usage_type": "attribute"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 232, "usage_type": "call"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 233, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 233, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 245, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 245, "usage_type": "attribute"}, {"api_name": "datetime.datetime.fromtimestamp", "line_number": 250, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 250, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 255, "usage_type": "call"}, {"api_name": "os.path.mkdir", "line_number": 256, "usage_type": "call"}, {"api_name": "os.path", "line_number": 256, "usage_type": "name"}, {"api_name": "pidly.IDL", "line_number": 257, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 325, "usage_type": "call"}, {"api_name": "os.path.path.getsize", "line_number": 326, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 326, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 326, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 381, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 386, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 386, "usage_type": "attribute"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 407, "usage_type": "call"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 408, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 410, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 421, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 423, "usage_type": "call"}, {"api_name": "datetime.time", "line_number": 463, "usage_type": "call"}, {"api_name": "os.path.system", "line_number": 464, "usage_type": "call"}, {"api_name": "os.path", "line_number": 464, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 483, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 487, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 487, "usage_type": "attribute"}, {"api_name": "os.path.path.isfile", "line_number": 521, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 521, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 521, "usage_type": "name"}, {"api_name": "cv2.VideoCapture", "line_number": 579, "usage_type": "call"}, {"api_name": "os.path.popen", "line_number": 586, "usage_type": "call"}, {"api_name": "os.path", "line_number": 586, "usage_type": "name"}, {"api_name": "json.loads", "line_number": 678, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 709, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 709, "usage_type": "attribute"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 724, "usage_type": "call"}, {"api_name": "strict_rfc3339.rfc3339_to_timestamp", "line_number": 725, "usage_type": "call"}, {"api_name": "time.mktime", "line_number": 726, "usage_type": "call"}, {"api_name": "uuid.uuid4", "line_number": 741, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 743, "usage_type": "call"}, {"api_name": "os.path.mkdir", "line_number": 760, "usage_type": "call"}, {"api_name": "os.path", "line_number": 760, "usage_type": "name"}, {"api_name": "os.path.listdir", "line_number": 800, "usage_type": "call"}, {"api_name": "os.path", "line_number": 800, "usage_type": "name"}, {"api_name": "astropy.io.fits.open", "line_number": 805, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 805, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 813, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 909, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 911, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 911, "usage_type": "attribute"}, {"api_name": "uuid.uuid4", "line_number": 929, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 931, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 992, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1020, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1020, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1022, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1022, "usage_type": "attribute"}, {"api_name": "os.path.path.isfile", "line_number": 1091, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 1091, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 1091, "usage_type": "name"}, {"api_name": "numpy.fromstring", "line_number": 1096, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 1116, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1118, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1119, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1121, "usage_type": "call"}, {"api_name": "numpy.repeat", "line_number": 1141, "usage_type": "call"}, {"api_name": "numpy.reshape", "line_number": 1144, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 1205, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1209, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1209, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1211, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1211, "usage_type": "attribute"}, {"api_name": "numpy.empty", "line_number": 1222, "usage_type": "call"}, {"api_name": "numpy.uint8", "line_number": 1222, "usage_type": "attribute"}, {"api_name": "numpy.uint8", "line_number": 1255, "usage_type": "attribute"}, {"api_name": "cv2.imdecode", "line_number": 1257, "usage_type": "call"}, {"api_name": "cv2.imshow", "line_number": 1258, "usage_type": "call"}, {"api_name": "cv2.cvtColor", "line_number": 1259, "usage_type": "call"}, {"api_name": "cv2.COLOR_GRAY2BGR", "line_number": 1259, "usage_type": "attribute"}, {"api_name": "cv2.imwrite", "line_number": 1260, "usage_type": "call"}, {"api_name": "json.loads", "line_number": 1276, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1280, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1280, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1282, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1282, "usage_type": "attribute"}, {"api_name": "os.path.path.isfile", "line_number": 1299, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 1299, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 1299, "usage_type": "name"}, {"api_name": "numpy.fromstring", "line_number": 1304, "usage_type": "call"}, {"api_name": "numpy.fromfile", "line_number": 1324, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1326, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1327, "usage_type": "call"}, {"api_name": "numpy.delete", "line_number": 1329, "usage_type": "call"}, {"api_name": "numpy.nditer", "line_number": 1352, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 1393, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 1393, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 1395, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1396, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1396, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1402, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1404, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 1411, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 1424, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 1424, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 1424, "usage_type": "name"}, {"api_name": "re.match", "line_number": 1434, "usage_type": "call"}, {"api_name": "re.split", "line_number": 1436, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1471, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 1472, "usage_type": "attribute"}, {"api_name": "sys.exc_info", "line_number": 1506, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 1525, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 1525, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1529, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1529, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1531, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1531, "usage_type": "attribute"}, {"api_name": "requests.Session", "line_number": 1537, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1542, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 1554, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 1556, "usage_type": "call"}, {"api_name": "paramiko.SSHClient", "line_number": 1562, "usage_type": "call"}, {"api_name": "paramiko.AutoAddPolicy", "line_number": 1563, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 1572, "usage_type": "call"}, {"api_name": "re.search", "line_number": 1623, "usage_type": "call"}, {"api_name": "datetime.date", "line_number": 1649, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1651, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1652, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1653, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1655, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 1656, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1658, "usage_type": "call"}, {"api_name": "numpy.transpose", "line_number": 1659, "usage_type": "call"}, {"api_name": "numpy.around", "line_number": 1661, "usage_type": "call"}, {"api_name": "pandas.to_datetime", "line_number": 1667, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1672, "usage_type": "call"}, {"api_name": "pandas.DataFrame", "line_number": 1673, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1690, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1690, "usage_type": "attribute"}, {"api_name": "datetime.datetime.strptime", "line_number": 1693, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1693, "usage_type": "attribute"}, {"api_name": "json.dumps", "line_number": 1721, "usage_type": "call"}, {"api_name": "binascii.a2b_qp", "line_number": 1723, "usage_type": "call"}, {"api_name": "json.dumps", "line_number": 1724, "usage_type": "call"}, {"api_name": "gzip.compress", "line_number": 1726, "usage_type": "call"}, {"api_name": "os.path.remove", "line_number": 1793, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1793, "usage_type": "name"}, {"api_name": "sys.exc_info", "line_number": 1796, "usage_type": "call"}, {"api_name": "logging.basicConfig", "line_number": 1815, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 1815, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 1817, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 1818, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1818, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 1826, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1828, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 1833, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 1837, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 1839, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 1870, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 1870, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 1870, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 1902, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 1908, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 1909, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 1912, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 1913, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 1919, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 2000, "usage_type": "call"}, {"api_name": "os.path.path.isdir", "line_number": 2059, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 2059, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 2059, "usage_type": "name"}, {"api_name": "os.path.listdir", "line_number": 2060, "usage_type": "call"}, {"api_name": "os.path", "line_number": 2060, "usage_type": "name"}, {"api_name": "logging.basicConfig", "line_number": 2140, "usage_type": "call"}, {"api_name": "logging.DEBUG", "line_number": 2140, "usage_type": "attribute"}, {"api_name": "json.loads", "line_number": 2142, "usage_type": "call"}, {"api_name": "datetime.datetime.strptime", "line_number": 2143, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2143, "usage_type": "attribute"}, {"api_name": "datetime.timedelta", "line_number": 2150, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 2152, "usage_type": "call"}, {"api_name": "datetime.timedelta", "line_number": 2156, "usage_type": "call"}, {"api_name": "elasticsearch.Elasticsearch", "line_number": 2160, "usage_type": "call"}, {"api_name": "pidly.IDL", "line_number": 2162, "usage_type": "call"}, {"api_name": "os.path.path.isfile", "line_number": 2185, "usage_type": "call"}, {"api_name": "os.path.path", "line_number": 2185, "usage_type": "attribute"}, {"api_name": "os.path", "line_number": 2185, "usage_type": "name"}, {"api_name": "datetime.date", "line_number": 2210, "usage_type": "call"}, {"api_name": "datetime.datetime", "line_number": 2216, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 2217, "usage_type": "attribute"}, {"api_name": "datetime.datetime", "line_number": 2220, "usage_type": "call"}, {"api_name": "pytz.UTC", "line_number": 2221, "usage_type": "attribute"}, {"api_name": "logging.error", "line_number": 2228, "usage_type": "call"}, {"api_name": "sys.exc_info", "line_number": 2271, "usage_type": "call"}]} +{"seq_id": "74293512", "text": "\"\"\"\nThe gui module was created by typing\n from PyQt5.uic import pyuic\n !pyuic5 PyMovie.ui -o gui.py\nin the IPython console while in src/pymovie directory\n\nThe helpDialog module was created by typing\n !pyuic5 helpDialog.ui -o helpDialog.py\nin the IPython console while in src/pymovie directory\n\nThe apertureEditDialog module was created by typing\n !pyuic5 apertureEditDialog.ui -o apertureEditDialog.py\nin the IPython console while in src/pymovie directory\n\nThe apertureNameDialog module was created by typing\n !pyuic5 apertureNameDialog.ui -o apertureNameDialog.py\nin the IPython console while in src/pymovie directory\n\nThe ocrProfileNameDialog module was created by typing\n !pyuic5 ocrProfileNameDialog.ui -o ocrProfileNameDialog.py\nin the IPython console while in src/pymovie directory\n\nThe selectProfile module was created by typing\n !pyuic5 selectProfile.ui -o selectProfile.py\nin the IPython console while in src/pymovie directory\n\nThe starPositionDialog module was created by typing\n !pyuic5 starPositionDialog.ui -o starPositionDialog.py\nin the IPython console while in src/pymovie directory\n\"\"\"\n\nimport matplotlib\n\nmatplotlib.use('Qt5Agg')\n\n# from matplotlib.figure import Figure\nfrom matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas\n# from matplotlib.backends.backend_qt5agg import NavigationToolbar2QT as NavigationToolbar\n\n# Leave the following import in place, even though PyCharm thinks it is unused. Apparently\n# there is a side effect of this import that is needed to make 3d plots work even though\n# Axes3D is never directly referenced\nfrom mpl_toolkits.mplot3d import Axes3D # !!!! Don't take me out\n\nimport matplotlib.pyplot as plt\n\nfrom more_itertools import sort_together\n\n# from resource import getrusage, RUSAGE_SELF\n# import gc\n\ntry:\n from pyoteapp import pyote\n # print('PyOTE installation found')\n pyote_available = True\nexcept ImportError:\n # print('No PyOTE installation found')\n pyote_available = False\n\nimport site\nimport warnings\nfrom astropy.utils.exceptions import AstropyWarning\nimport sys\nimport os\nimport errno\nimport platform\nimport pickle\nfrom pathlib import Path\nfrom urllib.request import urlopen\nimport numpy as np\nfrom pymovie.checkForNewerVersion import getMostRecentVersionOfPyMovie\nfrom pymovie.checkForNewerVersion import upgradePyMovie\nfrom pymovie import starPositionDialog\nfrom pymovie import ocrProfileNameDialog\nfrom pymovie import selectProfile\nfrom pymovie import astrometry_client\nfrom pymovie import wcs_helper_functions\nfrom pymovie import stacker\nimport pyqtgraph.exporters as pex\nfrom numpy import sqrt, arcsin\nfrom numpy import pi as PI\nimport PyQt5\nfrom PyQt5.QtCore import *\nfrom PyQt5.QtWidgets import QFileDialog, QGraphicsRectItem, QButtonGroup, QMessageBox, QTableWidgetItem\nfrom PyQt5.QtCore import QSettings, QSize, QPoint, QRectF, QTimer\nfrom PyQt5.QtCore import pyqtSlot\nfrom PyQt5.QtGui import QPainter\nfrom pymovie import gui, helpDialog, version, apertureEditDialog\nimport cv2\nimport glob\nimport astropy.io.fits as pyfits # Used for reading/writing FITS files\nfrom astropy import wcs\nfrom astropy import units as u\nfrom astropy.coordinates import SkyCoord\nfrom astroquery.vizier import Vizier\nfrom skimage import measure, exposure\nimport skimage\nimport subprocess\n\n\nfrom pymovie.aperture import *\nfrom pymovie.ocrCharacterBox import *\nfrom pymovie.ocr import *\nfrom pymovie.apertureEdit import *\n# from scipy.signal import savgol_filter\nfrom pymovie import alias_lnk_resolver\nimport pathlib\n\nif not os.name == 'posix':\n import winshell\n\nwarnings.filterwarnings(\"ignore\", category=FutureWarning)\nwarnings.filterwarnings(\"ignore\", category=RuntimeWarning)\n\n\ndef log_gray(x, a=None, b=None):\n if a is None:\n a = np.min(x)\n if b is None:\n b = np.max(x)\n\n # If the range of pixel values exceeds what will fit in an int16, we\n # need to abort this calculation because (b - a) will overflow short_scalars\n if float(b) - float(a) > 32767:\n return x\n\n linval = 10.0 + 990.0 * (x-float(a))/(b-a)\n return (np.log10(linval)-1.0)*0.5 * 255.0\n\n\nclass FixedImageExporter(pex.ImageExporter):\n def __init__(self, item):\n pex.ImageExporter.__init__(self, item)\n\n def makeWidthHeightInts(self):\n self.params['height'] = int(self.params['height'] + 1) # The +1 is needed\n self.params['width'] = int(self.params['width'] + 1)\n\n def widthChanged(self):\n sr = self.getSourceRect()\n ar = float(sr.height()) / sr.width()\n self.params.param('height').setValue(int(self.params['width'] * ar),\n blockSignal=self.heightChanged)\n\n def heightChanged(self):\n sr = self.getSourceRect()\n ar = float(sr.width()) / sr.height()\n self.params.param('width').setValue(int(self.params['height'] * ar),\n blockSignal=self.widthChanged)\n\n\nclass HelpDialog(QDialog, helpDialog.Ui_Dialog):\n def __init__(self):\n super(HelpDialog, self).__init__()\n self.setupUi(self)\n\n\nclass OcrProfileNameDialog(QDialog, ocrProfileNameDialog.Ui_ocrNameDialog):\n def __init__(self):\n super(OcrProfileNameDialog, self).__init__()\n self.setupUi(self)\n\n\nclass SelectProfileDialog(QDialog, selectProfile.Ui_Dialog):\n def __init__(self, msger, profile_dict_list, current_profile_dict):\n super(SelectProfileDialog, self).__init__()\n self.setupUi(self)\n self.msger = msger\n self.profiles = profile_dict_list\n self.currentProfile = current_profile_dict\n self.resultCode = -1 # Load profile was not performed\n\n self.exitButton.clicked.connect(self.exitProcedure)\n\n self.fillTableFromProfileList()\n\n # We do this so as to erase the default selection of row 0. Don't know why\n # this works, but it seems reliable.\n profile_selected = self.selectionTable.currentIndex()\n self.selectionTable.setCurrentIndex(profile_selected)\n\n self.deleteButton.clicked.connect(self.deleteSelection)\n\n self.addProfileButton.clicked.connect(self.addCurrentProfile)\n\n self.loadButton.clicked.connect(self.loadSelectedProfile)\n\n def loadSelectedProfile(self):\n profile_selected = self.selectionTable.currentIndex()\n row = profile_selected.row()\n self.resultCode = row\n self.close()\n\n def getResult(self):\n return self.resultCode\n\n def addCurrentProfile(self):\n self.currentProfile['id'] = self.profileNameEdit.text()\n self.profiles.append(self.currentProfile)\n self.selectionTable.setRowCount(0)\n self.fillTableFromProfileList()\n\n def fillTableFromProfileList(self):\n for profile in self.profiles:\n title = profile['id']\n numRows = self.selectionTable.rowCount()\n self.selectionTable.insertRow(numRows)\n item = QTableWidgetItem(str(title))\n self.selectionTable.setItem(numRows, 0, item)\n\n def deleteSelection(self):\n profile_selected = self.selectionTable.currentIndex()\n row = profile_selected.row()\n # self.msger(f'deleting row: {row}')\n self.profiles.pop(row) # Remove from dictionary\n self.selectionTable.setRowCount(0)\n self.fillTableFromProfileList() # Update table display\n\n def exitProcedure(self):\n for i in range(self.selectionTable.rowCount()):\n new_id = self.selectionTable.item(i, 0).text()\n self.profiles[i]['id'] = new_id\n self.close()\n\n\nclass StarPositionDialog(QDialog, starPositionDialog.Ui_Dialog):\n def __init__(self):\n super(StarPositionDialog, self).__init__()\n self.setupUi(self)\n\n\nclass Qt5MplCanvas(FigureCanvas):\n def __init__(self, img, title='Bobs plot', invert=False):\n # self.fig = Figure()\n # self.fig = Figure((5.0, 4.0), dpi=100) # 5x4 inches at 100 dpi\n self.fig = plt.figure()\n # super(FigureCanvas, self).__init__(self.fig)\n # self.ax = self.fig.add_subplot(111, projection='3d')\n self.ax = self.fig.gca(projection='3d')\n\n self.ax.set_xlabel('x', fontsize=20)\n self.ax.set_ylabel('y', fontsize=20)\n self.ax.set_title(title)\n self.ax.mouse_init()\n self.x = range(img.shape[0])\n\n if invert:\n self.y = range(img.shape[1])\n else:\n self.y = range(img.shape[1]-1, -1, -1)\n\n self.x, self.y = np.meshgrid(self.x, self.y)\n self.surf = self.ax.plot_surface(self.x, self.y, img, rstride=1, cstride=1,\n cmap='viridis', linewidth=0)\n\n # The positioning of the next two lines was found to be super-critical. If\n # these are moved, it will break mouse drag of the 3D image for MacOS or\n # Windows or both. You've been warned.\n FigureCanvas.__init__(self, self.fig)\n # super(FigureCanvas, self).__init__(self.fig)\n self.ax.mouse_init()\n\n\nclass PyMovie(QtGui.QMainWindow, gui.Ui_MainWindow):\n def __init__(self):\n super(PyMovie, self).__init__()\n\n # self.setFont(QtGui.QFont(\"Courier New\")) # Had no effect\n\n # Change pyqtgraph plots to be black on white\n pg.setConfigOption('background',\n (255, 255, 255)) # Do before any widgets drawn\n pg.setConfigOption('foreground', 'k') # Do before any widgets drawn\n pg.setConfigOptions(imageAxisOrder='row-major')\n\n # Build our GUI by calling the setupUi() function that defined/built\n # by pyuic5 from our PyMovie.ui and is found in gui.py\n self.setupUi(self)\n\n # This object is used to display tooltip help in a separate\n # modeless dialog box.\n self.helperThing = HelpDialog()\n\n self.homeDir = os.path.split(__file__)[0]\n self.ocrBoxesDir = self.homeDir\n self.ocrDigitsDir = self.homeDir\n\n self.clearTextBox()\n title = f'PyMovie Version: {version.version()}'\n self.setWindowTitle(title)\n\n self.showMsg(f'pyote available: {pyote_available}')\n\n if pyote_available:\n self.runPyote.setEnabled(True)\n\n self.runPyote.installEventFilter(self)\n\n # Open (or create) file for holding 'sticky' stuff\n self.settings = QSettings('PyMovie.ini', QSettings.IniFormat)\n self.settings.setFallbacksEnabled(False)\n\n # Use 'sticky' settings (from earlier session) to size and position the main screen\n self.resize(self.settings.value('size', QSize(800, 800)))\n self.move(self.settings.value('pos', QPoint(50, 50)))\n self.cascadeCheckBox.setChecked(self.settings.value('cascade', False) == 'true')\n self.plotSymbolSizeSpinBox.setValue(int(self.settings.value('plot_symbol_size', 4)))\n\n # splitterOne is the vertical splitter in the lower panel.\n # splitterTwo is the vertical splitter in the upper panel\n # splitterThree is the horizontal splitter between the top and bottom panel\n\n if self.settings.value('splitterOne') is not None:\n self.splitterOne.restoreState(self.settings.value('splitterOne'))\n self.splitterTwo.restoreState(self.settings.value('splitterTwo'))\n self.splitterThree.restoreState(self.settings.value('splitterThree'))\n\n self.api_key = self.settings.value('api_key', '')\n\n # This is a 'secret' switch that I use for experimental purposes. It causes\n # an extended context menu to be generated for ocr character selection boxes.\n # However, if one or modelDigits are found missing, the menu will appear for\n # normal users too.\n self.enableOcrTemplateSampling = self.settings.value('ocrsamplemenu', 'false') == 'true'\n\n self.modelDigits = [None] * 10\n\n # Clean up the frame display by hiding the 'extras' that pyqtgraph\n # standardly includes in an ImageView widget\n self.frameView.ui.menuBtn.hide()\n self.frameView.ui.roiBtn.hide()\n self.frameView.ui.histogram.hide()\n\n view = self.frameView.getView()\n # add new actions to the ViewBox context menu:\n view.menu.addSeparator()\n addSnapApp = view.menu.addAction(\"Add snap-to-blob aperture\")\n addFixedApp = view.menu.addAction('Add static aperture (no snap)')\n addAppStack = view.menu.addAction('Add stack of 5 apertures')\n addSnapApp.triggered.connect(self.addSnapAperture)\n addFixedApp.triggered.connect(self.addNamedStaticAperture)\n addAppStack.triggered.connect(self.addApertureStack)\n\n # We use mouse movements to dynamically display in the status bar the mouse\n # coordinates and pixel value under the mouse cursor.\n self.frameView.scene.sigMouseMoved.connect(self.mouseMovedInFrameView)\n self.thumbOneView.scene.sigMouseMoved.connect(self.mouseMovedInThumbOne)\n self.thumbTwoView.scene.sigMouseMoved.connect(self.mouseMovedInThumbTwo)\n self.frameView.ui.histogram.sigLevelsChanged.connect(self.levelChangedInImageControl)\n\n # Clean up thumbOneView by hiding the 'extras' that pyqtgraph\n # standardly includes in an ImageView widget\n self.thumbOneView.ui.menuBtn.hide()\n self.thumbOneView.ui.roiBtn.hide()\n self.thumbOneView.ui.histogram.hide()\n\n # add cross hairs\n self.hair1 = pg.InfiniteLine(angle=-45, movable=False)\n self.hair2 = pg.InfiniteLine(angle=45, movable=False)\n self.thumbOneView.addItem(self.hair1)\n self.thumbOneView.addItem(self.hair2)\n\n # Clean up thumbTwoView by hiding the 'extras' that pyqtgraph\n # standardly includes in an ImageView widget\n self.thumbTwoView.ui.menuBtn.hide()\n self.thumbTwoView.ui.roiBtn.hide()\n self.thumbTwoView.ui.histogram.hide()\n\n # The initial value must be coordinated with instance variable initiation\n self.roiComboBox.addItem(\"51\")\n self.roiComboBox.addItem(\"41\")\n self.roiComboBox.addItem(\"31\")\n self.roiComboBox.addItem(\"21\")\n\n self.vtiSelectLabel.installEventFilter(self)\n\n self.roiComboBox.currentIndexChanged.connect(self.setRoiFromComboBox)\n self.roiComboBox.installEventFilter(self)\n self.selectApertureSizeLabel.installEventFilter(self)\n\n # We need to change to a different vtiList pickle name with each version\n # change in order to capture any changes we make to the list --- we cannot\n # expect a user to find and delete that file on their own.\n vtiListfn = f'vtiList-{version.version()}.p'\n vtiListFilename = os.path.join(self.homeDir, vtiListfn)\n if os.path.exists(vtiListFilename):\n self.VTIlist = pickle.load(open(vtiListFilename, \"rb\"))\n self.showMsg(f'VTIlist loaded from {vtiListFilename}')\n else:\n # Create initial list --- a new installation\n self.VTIlist = [\n {'name': 'None'},\n {'name': 'IOTA VTI 3: one line (with F)'},\n {'name': 'IOTA VTI 3: two line (with F)'},\n {'name': 'IOTA VTI 2: one line (with P)'},\n {'name': 'IOTA VTI 2: two line (with P)'},\n {'name': 'BoxSprite: one-line'},\n {'name': 'Kiwi (left)'},\n {'name': 'Kiwi (right)'}\n ]\n pickle.dump(self.VTIlist, open(vtiListFilename, \"wb\"))\n self.showMsg(f'pickled self.VTIlist to {vtiListFilename}')\n\n for vtiDict in self.VTIlist:\n self.vtiSelectComboBox.addItem(vtiDict['name'])\n\n self.currentVTIindex = 0\n self.timestampFormatter = None\n self.upperTimestamp = ''\n self.lowerTimestamp = ''\n self.ocrboxBasePath = None\n self.modelDigitsFilename = None\n\n self.vtiSelectComboBox.installEventFilter(self)\n self.vtiSelectComboBox.currentIndexChanged.connect(self.vtiSelected)\n\n self.saveApertureState.clicked.connect(self.saveApertureGroup)\n self.saveApertureState.installEventFilter(self)\n\n self.restoreApertureState.clicked.connect(self.restoreApertureGroup)\n self.restoreApertureState.installEventFilter(self)\n\n self.createAVIWCSfolderButton.clicked.connect(self.createAviWcsFolder)\n self.createAVIWCSfolderButton.installEventFilter(self)\n self.createAVIWCSfolderButton.setEnabled(False)\n\n self.loadCustomProfilesButton.clicked.connect(self.loadCustomOcrProfiles)\n self.loadCustomProfilesButton.installEventFilter(self)\n self.loadCustomProfilesButton.setEnabled(False)\n\n # For now, we will save OCR profiles in the users home directory. If\n # later we find a better place, this is the only line we need to change\n self.profilesDir = os.path.expanduser('~')\n\n # We will need the user name when we write a pickled list of profile dictionaries.\n # We name them: pymovie-ocr-profiles-username.p to facilitate sharing among users.\n # Actually, we have changed our mind and will only use a single dictionary, but we might\n # need the user's name for some other reason.\n self.userName = os.path.basename(self.profilesDir)\n\n # Initialize all instance variables as a block (to satisfy PEP 8 standard)\n\n self.printKeyCodes = False\n self.consecutiveKcount = 0\n\n self.savedStateApertures = []\n self.savedStateFrameNumber = None\n self.savedPositions = []\n self.saveStateNeeded = True\n\n self.pixelAspectRatio = None\n\n self.upper_left_count = 0 # When Kiwi used: accumulate count ot times t2 was at left in upper field\n self.upper_right_count = 0 # When Kiwi used: accumulate count ot times t2 was at the right in upper field\n\n self.lower_left_count = 0 # When Kiwi used: accumulate count ot times t2 was at left in lower field\n self.lower_right_count = 0 # When Kiwi used: accumulate count ot times t2 was at the right in lower field\n\n self.currentUpperBoxPos = '' # Used by Kiwi timestamp extraction\n self.currentLowerBoxPos = '' # Used by Kiwi timestamp extraction\n\n self.kiwiInUse = False\n\n # Workspace for self.placeOcrBoxesOnImage()\n self.newLowerOcrBoxes = []\n\n # Standard return list for self.getApertureList()\n self.appList = []\n\n # Standard return list for self.getOcrBoxList()\n self.ocrBoxList = []\n\n self.suppressExtractTimestampCallInSpinnerResponder = False\n self.timestampReadingEnabled = False\n self.detectFieldTimeOrder = False\n\n self.acceptAviFolderDirectoryWithoutUserIntervention = False\n\n self.savedApertures = None\n\n self.upperOcrBoxesLeft = []\n self.lowerOcrBoxesLeft = []\n\n # These boxes come into play only when Kiwi is in use\n self.upperOcrBoxesRight = []\n self.lowerOcrBoxesRight = []\n\n self.kiwiUpperOcrBoxes = None\n self.kiwiLowerOcrBoxes = None\n self.kiwiAltUpperOcrBoxes = None\n self.kiwiAltLowerOcrBoxes = None\n\n self.frameJumpSmall = 25\n self.frameJumpBig = 200\n\n self.avi_location = None\n\n self.big_thresh = 9999\n self.one_time_suppress_stats = False\n\n self.analysisInProgress = False\n self.analysisRequested = False\n self.analysisPaused = True\n self.playPaused = True\n\n self.record_target_aperture = False\n\n self.plot_symbol_size = 1\n\n self.fits_folder_in_use = False\n self.avi_wcs_folder_in_use = False\n self.folder_dir = None\n\n self.timer = QTimer(self)\n self.timer.timeout.connect(self.setDoTestFlag)\n self.do_test = False\n\n # self.filename is set to the full path of the selected image file (or folder) once\n # the user has made a valid selection\n self.filename = None\n\n self.fourcc = ''\n\n # We use this variable to automatically number apertures as they are added. It is set\n # to zero when the user makes a valid selection of a file (or folder)\n self.apertureId = None\n\n # If an avi file was selected, these variables come into play\n self.cap = None\n self.avi_in_use = False\n self.preserve_apertures = False\n\n # If a FITS file folder was selected, this variable gets filled with a list\n # of the filenames ending in .fits found within the selected FITS folder.\n self.fits_filenames = []\n\n self.image = None\n self.upper_field = None\n self.lower_field = None\n self.image_fields = None\n self.thumbOneImage = None\n self.thumbTwoImage = None\n\n # A True/False to indicate when a first frame has been read and displayed. This\n # is used in self.showFrame() and set in self.readFitsFile() and self.readAviFile()\n self.initialFrame = None\n\n # This variable not yet used. It will come into play when we implement timestamp reading\n self.vti_trim = 120\n\n self.fits_timestamp = None\n self.fits_date = None\n\n # This 'state' variable controls the writing of reference star data files\n # during manual WCS calibration. The method handleSetRaDecSignal uses this\n # to determine which reference file to write. The meanings of the values\n # are: state == 0 Do nothing except warn the user that a manual process is not started\n # state == 1 Write data file for reference star 1 and advance state to 2\n # state == 2 Write data file for star 2, set state to 0, call calibration routine\n # which will set the target aperture (if location set)\n self.manual_wcs_state = 0\n\n self.num_yellow_apertures = None\n\n self.setRoiFromComboBox()\n\n # For target aperture(s), a blob must be within 8 pixels of the aperture center to\n # be considered valid.\n self.allowed_centroid_delta = 8\n\n self.gaussian_blur = (5, 5)\n\n # The following two variables are used by MeasurementAperture to keep apertures completely\n # within the image boundary. They are initialized when the first frame is read (in self.showFrame())\n self.roi_max_x = None\n self.roi_max_y = None\n\n self.show_stats = True\n self.img_max = None\n self.img_min = None\n\n # We track mouse movement whenever the cursor is in the main image or\n # either of the thumbnails. That lets us display x,y coordinates and pixel values\n # in the staus bar at the bottom of the app window.\n self.mousex = None\n self.mousey = None\n\n # When the mouse cursor is moved over an aperture, this variable is set to that\n # aperture. This allows for selection of aperture-to-be-reported while an\n # analysis is in progress.\n self.pointed_at_aperture = None\n\n self.yellow_mask = None # Holds mask created from 'tracking star'\n self.use_yellow_mask = False\n\n self.yellow_x = None\n self.yellow_y = None\n self.delta_theta = 0.0\n\n self.avi_wcs_folder_in_use = False\n self.wcs_solution_available = False\n self.wcs_frame_num = None\n self.wcs = None # This holds the active WCS solution (if any)\n\n # Keeps track of all pyqtgraph plot windows that have been created so that they\n # can be gracefully closed when the user closes this app.\n self.plots = []\n\n # We keep track of the aperture name that is being displayed in Thumbnail One\n # so that we can add that info to the 3D plots\n self.thumbnail_one_aperture_name = None\n\n self.levels = []\n self.frame_at_level_set = None\n\n # These are part of an experiment and obsolete now. They are left in place in\n # case we resurrect the idea of shrinking or expanding a mask by erosion or inflation.\n # This concept is likely only useful when a 'yellow_mask' is being used\n # self.erode_mask = False\n # self.inflate_mask = False\n\n self.apertureEditor = None\n\n # end instance variable declarations\n\n self.transportMaxLeft.installEventFilter(self)\n self.transportMaxLeft.clicked.connect(self.seekMaxLeft)\n\n self.transportBigLeft.installEventFilter(self)\n self.transportSmallLeft.installEventFilter(self)\n\n self.transportMinusOneFrame.clicked.connect(self.moveOneFrameLeft)\n self.transportMinusOneFrame.installEventFilter(self)\n\n self.transportPlusOneFrame.clicked.connect(self.moveOneFrameRight)\n self.transportPlusOneFrame.installEventFilter(self)\n\n self.transportPlayLeft.installEventFilter(self)\n self.transportPlayLeft.clicked.connect(self.playLeft)\n\n self.transportPause.installEventFilter(self)\n self.transportPause.clicked.connect(self.pauseAnalysis)\n\n self.transportAnalyze.installEventFilter(self)\n self.transportAnalyze.clicked.connect(self.startAnalysis)\n\n self.transportPlayRight.installEventFilter(self)\n self.transportPlayRight.clicked.connect(self.playRight)\n\n self.transportSmallRight.installEventFilter(self)\n self.transportBigRight.installEventFilter(self)\n\n self.transportMaxRight.installEventFilter(self)\n self.transportMaxRight.clicked.connect(self.seekMaxRight)\n\n self.transportCurrentFrameLabel.installEventFilter(self)\n self.transportStopAtFrameLabel.installEventFilter(self)\n\n self.invertImagesCheckBox.clicked.connect(self.invertImages)\n self.invertImagesCheckBox.installEventFilter(self)\n\n self.showImageControlCheckBox.clicked.connect(self.toggleImageControl)\n self.showImageControlCheckBox.installEventFilter(self)\n\n self.editAperturesButton.clicked.connect(self.editApertures)\n self.editAperturesButton.installEventFilter(self)\n\n # Captures the toolTip info and displays it in our own helpDialog\n self.textOutLabel.installEventFilter(self)\n\n self.frameView.installEventFilter(self)\n # self.mainImageLabel.installEventFilter(self)\n\n self.transportHelp.installEventFilter(self)\n self.transportHelp.clicked.connect(self.mainImageHelp)\n\n # self.viewFieldsCheckBox.clicked.connect(self.showFrame)\n self.viewFieldsCheckBox.toggled.connect(self.handleChangeOfDisplayMode)\n self.viewFieldsCheckBox.installEventFilter(self)\n\n self.useYellowMaskCheckBox.clicked.connect(self.handleYellowMaskClick)\n self.useYellowMaskCheckBox.installEventFilter(self)\n\n self.readFitsFolderButton.clicked.connect(self.readFitsFile)\n self.readFitsFolderButton.installEventFilter(self)\n\n self.openBmpPushButton.clicked.connect(self.readFinderImage)\n self.openBmpPushButton.installEventFilter(self)\n\n self.readAviFileButton.clicked.connect(self.readAviFile)\n self.readAviFileButton.installEventFilter(self)\n\n self.selectAviWcsFolderButton.clicked.connect(self.selectAviFolder)\n self.selectAviWcsFolderButton.installEventFilter(self)\n\n self.currentFrameSpinBox.valueChanged.connect(self.updateFrameWithTracking)\n\n self.bg2 = QButtonGroup()\n self.bg2.addButton(self.topFieldFirstRadioButton)\n self.bg2.addButton(self.bottomFieldFirstRadioButton)\n self.topFieldFirstRadioButton.setChecked(True)\n\n self.topFieldFirstRadioButton.clicked.connect(self.fieldTimeOrderChanged)\n self.bottomFieldFirstRadioButton.clicked.connect(self.fieldTimeOrderChanged)\n\n self.queryVizierButton.clicked.connect(self.queryVizier)\n self.queryVizierButton.installEventFilter(self)\n\n self.ucac4Label.installEventFilter(self)\n self.starIdEdit.textChanged.connect(self.clearCoordinatesEdit)\n\n self.saveTargetLocButton.clicked.connect(self.saveTargetInFolder)\n self.saveTargetLocButton.installEventFilter(self)\n\n self.threshValueEdit.valueChanged.connect(self.changeThreshold)\n self.setMskthLabel.installEventFilter(self)\n\n self.metadataButton.clicked.connect(self.showFitsMetadata)\n self.metadataButton.installEventFilter(self)\n\n # self.clearAppDataButton.clicked.connect(self.clearApertureData)\n # self.clearAppDataButton.installEventFilter(self)\n\n # self.writeCsvButton.clicked.connect(self.writeCsvFile)\n # self.writeCsvButton.installEventFilter(self)\n\n self.infoButton.clicked.connect(self.showInfo)\n self.infoButton.installEventFilter(self)\n\n self.documentationPushButton.clicked.connect(self.showDocumentation)\n self.documentationPushButton.installEventFilter(self)\n\n self.demoMeanPushButton.clicked.connect(self.showRobustMeanDemo)\n self.demoMeanPushButton.installEventFilter(self)\n\n self.plotSymbolSizeSpinBox.valueChanged.connect(self.changePlotSymbolSize)\n self.plotSymbolSizeLabel.installEventFilter(self)\n\n # self.displayPlotsButton.clicked.connect(self.showLightcurves)\n # self.displayPlotsButton.installEventFilter(self)\n\n self.cascadeCheckBox.installEventFilter(self)\n\n self.manualWcsButton.clicked.connect(self.manualWcsCalibration)\n self.manualWcsButton.installEventFilter(self)\n\n self.stackFramesButton.clicked.connect(self.performFrameStacking)\n self.stackFramesButton.installEventFilter(self)\n\n self.finderRedactLinesLabel.installEventFilter(self)\n self.finderNumFramesLabel.installEventFilter(self)\n\n self.frameToFitsButton.clicked.connect(self.getWCSsolution)\n self.frameToFitsButton.installEventFilter(self)\n\n self.thumbnailOneLabel.installEventFilter(self)\n self.thumbnailTwoLabel.installEventFilter(self)\n\n self.transportSmallLeft.clicked.connect(self.jumpSmallFramesBack)\n\n self.transportBigLeft.clicked.connect(self.jumpBigFramesBack)\n\n self.transportSmallRight.clicked.connect(self.jumpSmallFramesForward)\n\n self.transportBigRight.clicked.connect(self.jumpBigFramesForward)\n\n self.view3DButton.clicked.connect(self.show3DThumbnail)\n self.view3DButton.installEventFilter(self)\n\n self.transportReturnToMark.clicked.connect(self.restoreSavedState)\n self.transportReturnToMark.installEventFilter(self)\n\n self.transportClearData.clicked.connect(self.clearApertureData)\n self.transportClearData.installEventFilter(self)\n\n self.transportMark.clicked.connect(self.saveCurrentState)\n self.transportMark.installEventFilter(self)\n\n self.transportPlot.clicked.connect(self.showLightcurves)\n self.transportPlot.installEventFilter(self)\n\n self.transportCsv.clicked.connect(self.writeCsvFile)\n self.transportCsv.installEventFilter(self)\n\n self.pixelHeightLabel.installEventFilter(self)\n self.pixelWidthLabel.installEventFilter(self)\n\n self.changePlotSymbolSize()\n\n self.disableControlsWhenNoData()\n\n QtGui.QGuiApplication.processEvents()\n self.checkForNewerVersion()\n\n self.copy_desktop_icon_file_to_home_directory()\n\n def mainImageHelp(self):\n msg = self.transportHelp.toolTip()\n self.helperThing.textEdit.clear()\n self.helperThing.textEdit.insertHtml(msg)\n self.helperThing.raise_()\n self.helperThing.show()\n\n def addApertureStack(self):\n self.showMsg('Not yet implemented')\n for i in range(5):\n self.addStaticAperture(askForName=False)\n for app in self.getApertureList():\n if app.color == 'green':\n app.setRed()\n\n def composeApertureStateDictionary(self, aperture):\n my_dict = {}\n my_dict.update({'name': aperture.name})\n my_dict.update({'thresh': aperture.thresh})\n my_dict.update({'color': aperture.color})\n my_dict.update({'x0': aperture.x0})\n my_dict.update({'y0': aperture.y0})\n my_dict.update({'xsize': aperture.xsize})\n my_dict.update({'ysize': aperture.ysize})\n my_dict.update({'jogging_enabled': aperture.jogging_enabled})\n my_dict.update({'auto_display': aperture.auto_display})\n my_dict.update({'thumbnail_source': aperture.thumbnail_source})\n my_dict.update({'default_mask_radius': aperture.default_mask_radius})\n my_dict.update({'order_number': aperture.order_number})\n my_dict.update({'defaultMask': aperture.defaultMask})\n my_dict.update({'defaultMaskPixelCount': aperture.defaultMaskPixelCount})\n my_dict.update({'theta': aperture.theta})\n my_dict.update({'dx': aperture.dx})\n my_dict.update({'dy': aperture.dy})\n my_dict.update({'xc': aperture.xc})\n my_dict.update({'yc': aperture.yc})\n my_dict.update({'max_xpos': aperture.max_xpos})\n my_dict.update({'max_ypos': aperture.max_ypos})\n return my_dict\n\n\n def restoreApertureGroup(self):\n # Set to the correct frame first (if present).\n frameFn = self.folder_dir + '/markedFrameNumber.p'\n if os.path.exists(frameFn):\n markedFrameNumber = pickle.load(open(frameFn, 'rb'))\n self.showMsg(f'marked frame number is: {markedFrameNumber}')\n self.currentFrameSpinBox.setValue(markedFrameNumber)\n else:\n return\n\n # Force frame view\n self.viewFieldsCheckBox.setChecked(False)\n\n # Remove all apertures that have been already placed (particularly the target\n # aperture that is automatically placed when a WCS solution was present)\n self.clearApertures()\n\n\n # Then place all the apertures with complete state\n aperturesFn = self.folder_dir + '/markedApertures.p'\n if os.path.exists(aperturesFn):\n savedApertureDicts = pickle.load(open(aperturesFn, \"rb\"))\n self.showMsg(f'Num saved apertures: {len(savedApertureDicts)}')\n\n for dict in savedApertureDicts:\n try:\n x0 = dict['x0']\n y0 = dict['y0']\n xsize = dict['xsize']\n ysize = dict['ysize']\n bbox = (x0, y0, xsize, ysize)\n name = dict['name']\n max_xpos = dict['max_xpos']\n max_ypos = dict['max_ypos']\n\n # Create an aperture object (box1) and connect it to us (self)\n aperture = MeasurementAperture(name, bbox, max_xpos, max_ypos)\n\n aperture.thresh = dict['thresh']\n\n color = dict['color']\n if color == 'red':\n aperture.setRed()\n elif color == 'green':\n aperture.setGreen()\n elif color == 'white':\n aperture.setWhite()\n elif color == 'yellow':\n aperture.setYellowNoCheck()\n else:\n self.showMsg(f'Unexpected color (color) found while restoring marked apertures')\n\n aperture.jogging_enabled = dict['jogging_enabled']\n aperture.auto_display = dict['auto_display']\n aperture.thumbnail_source = dict['thumbnail_source']\n aperture.default_mask_radius = dict['default_mask_radius']\n aperture.order_number = dict['order_number']\n aperture.defaultMask = dict['defaultMask']\n aperture.defaultMaskPixelCount = dict['defaultMaskPixelCount']\n aperture.theta = dict['theta']\n aperture.dx = dict['dx']\n aperture.dy = dict['dy']\n aperture.xc = dict['xc']\n aperture.yc = dict['yc']\n\n self.connectAllSlots(aperture)\n\n view = self.frameView.getView()\n view.addItem(aperture)\n\n except Exception as e:\n self.showMsg(f'While restoring aperture constellation exception: {e}')\n\n def saveApertureGroup(self):\n # We need to have the apertures visible before we can save them\n if self.viewFieldsCheckBox.isChecked():\n self.viewFieldsCheckBox.setChecked(False)\n self.savedStateApertures = self.getApertureList()\n savedApertureDicts = []\n for aperture in self.savedStateApertures:\n dict = self.composeApertureStateDictionary(aperture)\n savedApertureDicts.append(dict)\n\n # Pickle the saved aperture dictionaries for use during opening of file/folder\n pickle.dump(savedApertureDicts, open(self.folder_dir + '/markedApertures.p', \"wb\"))\n\n self.savedStateFrameNumber = self.currentFrameSpinBox.value()\n pickle.dump(self.savedStateFrameNumber, open(self.folder_dir + '/markedFrameNumber.p', \"wb\"))\n\n self.showMsg(f'Current aperture group and frame number saved.')\n\n def saveCurrentState(self):\n # We need to have the apertures visible before we can save them\n if self.viewFieldsCheckBox.isChecked():\n self.viewFieldsCheckBox.setChecked(False)\n self.savedStateApertures = self.getApertureList()\n self.savedPositions = []\n self.savedStateFrameNumber = self.currentFrameSpinBox.value()\n for aperture in self.savedStateApertures:\n self.savedPositions.append(aperture.getBbox())\n\n self.transportReturnToMark.setEnabled(True)\n\n self.showMsg(f'Configuration marked.')\n\n def restoreSavedState(self):\n # We should be showing full frame before adding back in the save apertures\n if self.viewFieldsCheckBox.isChecked():\n self.viewFieldsCheckBox.setChecked(False)\n self.clearOcrBoxes()\n\n if not self.savedStateFrameNumber is None:\n self.currentFrameSpinBox.setValue(self.savedStateFrameNumber)\n\n # restore any saved apertures\n if self.savedStateApertures:\n view = self.frameView.getView()\n for i, aperture in enumerate(self.savedStateApertures):\n view.addItem(aperture)\n aperture.setPos(self.savedPositions[i])\n self.connectAllSlots(aperture)\n\n def moveOneFrameLeft(self):\n curFrame = self.currentFrameSpinBox.value()\n curFrame -= 1\n self.currentFrameSpinBox.setValue(curFrame)\n\n def moveOneFrameRight(self):\n curFrame = self.currentFrameSpinBox.value()\n curFrame += 1\n self.currentFrameSpinBox.setValue(curFrame)\n\n def seekMaxLeft(self):\n self.currentFrameSpinBox.setValue(0)\n\n def seekMaxRight(self):\n maxFrame = self.stopAtFrameSpinBox.maximum()\n self.stopAtFrameSpinBox.setValue(maxFrame)\n self.currentFrameSpinBox.setValue(maxFrame)\n\n def playRight(self):\n self.playPaused = False\n self.autoPlayRight()\n\n def playLeft(self):\n self.playPaused = False\n self.autoPlayLeft()\n\n def pauseAnalysis(self):\n self.analysisPaused = True\n self.playPaused = True\n self.analysisRequested = False\n self.setTransportButtonEnableState(True)\n\n def startAnalysis(self):\n if self.saveStateNeeded:\n self.saveStateNeeded = False\n self.saveCurrentState()\n self.analysisRequested = True\n self.analysisPaused = False\n self.setTransportButtonEnableState(False)\n self.transportPause.setEnabled(True)\n self.autoRun()\n\n @staticmethod\n def queryWhetherNewVersionShouldBeInstalled():\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText('A newer version of PyMovie is available. Do you wish to install it?')\n msg.setWindowTitle('Get latest version of PyMovie query')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n return retval\n\n def checkForNewerVersion(self):\n gotVersion, latestVersion = getMostRecentVersionOfPyMovie()\n if gotVersion:\n if latestVersion <= version.version():\n self.showMsg(f'Found the latest version is: {latestVersion}')\n self.showMsg('You are running the most recent version of PyMovie')\n else:\n self.showMsg('Version ' + latestVersion + ' is available')\n if self.queryWhetherNewVersionShouldBeInstalled() == QMessageBox.Yes:\n self.showMsg('You have opted to install latest version of PyMovie')\n self.installLatestVersion(f'pymovie=={latestVersion}')\n else:\n self.showMsg('You have declined the opportunity to install latest PyMovie')\n else:\n self.showMsg(f'latestVersion found: {latestVersion}')\n\n def installLatestVersion(self, pymovieversion):\n self.showMsg(f'Asking to upgrade to: {pymovieversion}')\n pipResult = upgradePyMovie(pymovieversion)\n for line in pipResult:\n self.showMsg(line, blankLine=False)\n\n self.showMsg('', blankLine=False)\n self.showMsg('The new version is installed but not yet running.')\n self.showMsg('Close and reopen PyMovie to start the new version running.')\n\n def createAviWcsFolder(self):\n options = QFileDialog.Options()\n # options |= QFileDialog.DontUseNativeDialog\n options |= QFileDialog.DirectoryOnly\n\n dirname = QFileDialog.getExistingDirectory(\n self, # parent\n \"Select directory where AVI-WCS folder should be placed\", # title for dialog\n self.settings.value('avidir', \"./\"), # starting directory\n options=options\n )\n if dirname:\n\n self.showMsg(f'AVI-WCS folder will be created in: {dirname}', blankLine=False)\n base_with_ext = os.path.basename(self.filename)\n base, _ = os.path.splitext(base_with_ext)\n self.showMsg(f'and the folder will be named {base}')\n full_dir_path = os.path.join(dirname, base)\n\n msg = f'AVI-WCS folder will be created in: {dirname}\\n\\n'\n msg += f'Folder name: {base}'\n self.showMsgPopup(msg)\n\n self.settings.setValue('avidir', full_dir_path) # Make dir 'sticky'\"\n\n pathlib.Path(full_dir_path).mkdir(parents=True, exist_ok=True)\n if sys.platform == 'darwin':\n ok, file, dir, retval, source = alias_lnk_resolver.create_osx_alias_in_dir(self.filename, full_dir_path)\n if not ok:\n self.showMsg('Failed to create and populate AVI-WCS folder')\n else:\n self.showMsg('AVI-WCS folder created and populated')\n # self.showMsg(f' file: {file}\\n dir: {dir}\\n retval: {retval}\\n source: {source}')\n\n elif sys.platform == 'linux':\n src = self.filename\n dst = os.path.join(dirname, base, base_with_ext)\n try:\n os.symlink(src,dst)\n self.showMsgPopup('AVI-WCS folder created and populated')\n except OSError as e:\n if e.errno == errno.EEXIST:\n os.remove(dst)\n os.symlink(src,dst)\n self.showMsgPopup('AVI-WCS folder created and old symlink overwritten')\n else:\n self.showMsgPopup('Failed to create and populate AVI-WCS folder')\n\n else:\n # self.showMsg(f'os.name={os.name} not yet fully supported for AVI-WCS folder creation.')\n # Make sure that there is a directory waiting for the shortcut file\n os.makedirs(full_dir_path, exist_ok=True)\n\n shortcut = winshell.shortcut(self.filename)\n base_lnk_name = os.path.basename(shortcut.lnk_filepath)\n dest_path = os.path.join(full_dir_path, base_lnk_name)\n shortcut.lnk_filepath = dest_path\n shortcut.write()\n\n self.acceptAviFolderDirectoryWithoutUserIntervention = True\n self.selectAviFolder()\n else:\n self.showMsg(f'Operation was cancelled.')\n\n def readSavedOcrProfiles(self):\n\n available_profiles = glob.glob(self.profilesDir + '/pymovie-ocr-profiles.p')\n\n dictionary_list = []\n if len(available_profiles) == 0:\n return dictionary_list\n else:\n for file in available_profiles:\n # self.showMsg(f'{file}', blankLine=False)\n # unpickle the list of profile dictionaries ---\n # {'id': 'profile info', 'upper-boxes': upperOcrBoxes[], 'lower-boxes': lowerOcrBoxes[],\n # 'digits': modelDigits}, 'formatter-code': 'iota'}[]\n # Keep apending until all profile files have been read\n dict_list = pickle.load(open(file, \"rb\"))\n for entry in dict_list:\n dictionary_list.append(entry)\n return dictionary_list\n\n def handleChangeOfDisplayMode(self):\n # self.showMsg(f'View avi fields: {self.viewFieldsCheckBox.isChecked()}')\n if self.viewFieldsCheckBox.isChecked():\n # preserve all apertures\n self.savedApertures = self.getApertureList()\n # clear all apertures\n self.clearApertures()\n self.placeOcrBoxesOnImage()\n self.showFrame()\n else:\n # clear ocr boxes (if any)\n # if self.lowerOcrBoxes:\n self.clearOcrBoxes()\n # restore any saved apertures\n if self.savedApertures:\n view = self.frameView.getView()\n for aperture in self.savedApertures:\n view.addItem(aperture)\n self.connectAllSlots(aperture)\n self.showFrame()\n\n def fieldTimeOrderChanged(self):\n self.showMsg(f'top field earlist is {self.topFieldFirstRadioButton.isChecked()}')\n self.vtiSelected()\n\n def jogSingleOcrBox(self, dx, dy, boxnum, position, ocr):\n\n # Frame 0 is often messed up (somehow). So we protect the user by not\n # letting him change ocr box positions while on frame 0\n if self.currentFrameSpinBox.value() == 0:\n self.showMsg(f'!!!! Move past frame 0 first. It is not representative. !!!!')\n return\n\n assert(position == 'upper' or position == 'lower')\n if position == 'upper':\n if self.currentUpperBoxPos == 'left':\n selected_box = self.upperOcrBoxesLeft[boxnum]\n xL, xR, yU, yL = selected_box\n self.upperOcrBoxesLeft[boxnum] = (xL + dx, xR + dx, yU + dy, yL + dy)\n ocr.setBox(self.upperOcrBoxesLeft[boxnum])\n else:\n selected_box = self.upperOcrBoxesRight[boxnum]\n xL, xR, yU, yL = selected_box\n self.upperOcrBoxesRight[boxnum] = (xL + dx, xR + dx, yU + dy, yL + dy)\n ocr.setBox(self.upperOcrBoxesRight[boxnum])\n else:\n yadj = int(self.image.shape[0] / 2)\n if self.currentLowerBoxPos == 'left':\n selected_box = self.lowerOcrBoxesLeft[boxnum]\n xL, xR, yU, yL = selected_box\n self.lowerOcrBoxesLeft[boxnum] = (xL + dx, xR + dx, yU + dy, yL + dy)\n ocr.setBox((xL + dx, xR + dx, yU + dy + yadj, yL + dy + yadj))\n else:\n selected_box = self.lowerOcrBoxesRight[boxnum]\n xL, xR, yU, yL = selected_box\n self.lowerOcrBoxesRight[boxnum] = (xL + dx, xR + dx, yU + dy, yL + dy)\n ocr.setBox((xL + dx, xR + dx, yU + dy + yadj, yL + dy + yadj))\n\n self.pickleOcrBoxes()\n\n def placeOcrBoxesOnImage(self):\n\n if not self.upperOcrBoxesLeft:\n return\n\n y_adjust = int(self.image.shape[0] / 2)\n\n self.newLowerOcrBoxes = []\n if self.currentLowerBoxPos == 'left':\n for ocrbox in self.lowerOcrBoxesLeft:\n xL, xR, yU, yL = ocrbox\n self.newLowerOcrBoxes.append((xL, xR, yU + y_adjust, yL + y_adjust))\n else:\n for ocrbox in self.lowerOcrBoxesRight:\n xL, xR, yU, yL = ocrbox\n self.newLowerOcrBoxes.append((xL, xR, yU + y_adjust, yL + y_adjust))\n\n boxnum = 0\n if self.currentUpperBoxPos == 'left':\n for ocrbox in self.upperOcrBoxesLeft:\n self.addOcrAperture(ocrbox, boxnum, 'upper')\n boxnum += 1\n else:\n for ocrbox in self.upperOcrBoxesRight:\n self.addOcrAperture(ocrbox, boxnum, 'upper')\n boxnum += 1\n\n boxnum = 0\n for ocrbox in self.newLowerOcrBoxes:\n self.addOcrAperture(ocrbox, boxnum, 'lower')\n boxnum += 1\n\n def pickleOcrBoxes(self):\n base_path = self.ocrboxBasePath\n upper_boxes_fn = f'{base_path}-upper.p'\n lower_boxes_fn = f'{base_path}-lower.p'\n\n upper_boxes_right_fn = f'{base_path}-upper-right.p'\n lower_boxes_right_fn = f'{base_path}-lower-right.p'\n\n upper_boxes = os.path.join(self.ocrBoxesDir, upper_boxes_fn)\n lower_boxes = os.path.join(self.ocrBoxesDir, lower_boxes_fn)\n\n upper_boxes_right = os.path.join(self.ocrBoxesDir, upper_boxes_right_fn)\n lower_boxes_right = os.path.join(self.ocrBoxesDir, lower_boxes_right_fn)\n\n pickle.dump(self.upperOcrBoxesLeft, open(upper_boxes, \"wb\"))\n pickle.dump(self.lowerOcrBoxesLeft, open(lower_boxes, \"wb\"))\n\n pickle.dump(self.upperOcrBoxesRight, open(upper_boxes_right, \"wb\"))\n pickle.dump(self.lowerOcrBoxesRight, open(lower_boxes_right, \"wb\"))\n\n return\n\n def loadPickledOcrBoxes(self):\n base_path = self.ocrboxBasePath\n\n upper_boxes_fn = f'{base_path}-upper.p'\n lower_boxes_fn = f'{base_path}-lower.p'\n\n # These files are only present for Kiwi\n upper_boxes_right_fn = f'{base_path}-upper-right.p'\n lower_boxes_right_fn = f'{base_path}-lower-right.p'\n\n upper_boxes = os.path.join(self.ocrBoxesDir, upper_boxes_fn)\n lower_boxes = os.path.join(self.ocrBoxesDir, lower_boxes_fn)\n\n # These files are only present for Kiwiw\n upper_boxes_right = os.path.join(self.ocrBoxesDir, upper_boxes_right_fn)\n lower_boxes_right = os.path.join(self.ocrBoxesDir, lower_boxes_right_fn)\n\n if os.path.exists(upper_boxes) and os.path.exists(lower_boxes):\n self.upperOcrBoxesLeft = pickle.load(open(upper_boxes, \"rb\"))\n # self.showMsg(f'upper OCR boxes loaded from {upper_boxes}')\n self.lowerOcrBoxesLeft = pickle.load(open(lower_boxes, \"rb\"))\n # self.showMsg(f'lower OCR boxes loaded from {lower_boxes}')\n else:\n self.upperOcrBoxesLeft = None\n self.lowerOcrBoxesLeft = None\n\n if os.path.exists(upper_boxes_right) and os.path.exists(lower_boxes_right):\n self.upperOcrBoxesRight = pickle.load(open(upper_boxes_right, \"rb\"))\n # self.showMsg(f'upper OCR boxes loaded from {upper_boxes}')\n self.lowerOcrBoxesRight = pickle.load(open(lower_boxes_right, \"rb\"))\n # self.showMsg(f'lower OCR boxes loaded from {lower_boxes}')\n else:\n self.upperOcrBoxesRight = []\n self.lowerOcrBoxesRight = []\n\n def showMissingModelDigits(self):\n missing_model_digits = ''\n for i in range(10):\n if self.modelDigits[i] is None:\n missing_model_digits += f'{i} '\n if missing_model_digits:\n self.showMsg(f'!!! Model digits {missing_model_digits}are missing !!!')\n self.timestampReadingEnabled = False\n self.enableOcrTemplateSampling = True\n return True\n else:\n # self.showMsg(f'All model digits (0...9) are present.')\n # self.timestampReadingEnabled = True\n self.enableOcrTemplateSampling = self.settings.value('ocrsamplemenu', 'false') == 'true'\n return False\n\n def saveModelDigits(self):\n pickled_digits_fn = self.modelDigitsFilename\n pickled_digits_path = os.path.join(self.ocrDigitsDir, pickled_digits_fn)\n pickle.dump(self.modelDigits, open(pickled_digits_path, \"wb\"))\n\n def deleteModelDigits(self):\n digits_fn = self.modelDigitsFilename\n digits_path = os.path.join(self.ocrDigitsDir, digits_fn)\n if os.path.exists(digits_path):\n os.remove(digits_path)\n for i in range(10):\n self.modelDigits[i] = None\n\n def loadModelDigits(self):\n pickled_digits_fn = self.modelDigitsFilename\n pickled_digits_path= os.path.join(self.ocrDigitsDir, pickled_digits_fn)\n\n if os.path.exists(pickled_digits_path):\n self.modelDigits = pickle.load(open(pickled_digits_path, \"rb\"))\n self.showMissingModelDigits()\n else:\n self.modelDigits = [None] * 10\n self.showMissingModelDigits()\n\n def extractTimestamps(self, printresults = True):\n if not self.timestampReadingEnabled:\n return None, None, None, None, None, None, None, None, None, None\n\n # kb = getrusage(RUSAGE_SELF).ru_maxrss\n # self.showMsg(f'Mem usage: {kb / 1024 / 1024:.2f} (mb)')\n\n thresh = 0\n\n # if self.formatterCode == 'kiwi-left' or self.formatterCode == 'kiwi-right':\n if self.kiwiInUse:\n\n if self.upper_left_count + self.upper_right_count > 3:\n use_left = self.upper_left_count > self.upper_right_count\n else:\n use_left = None\n\n # Note: left_used is only useful when kiwi is True\n # reg_* means the left box position\n # alt_* means the right box position\n reg_upper_timestamp, reg_upper_time, \\\n reg_upper_ts, reg_upper_scores, reg_upper_cum_score, reg_upper_left_used = \\\n extract_timestamp(\n self.upper_field, self.upperOcrBoxesLeft, self.modelDigits, self.timestampFormatter,\n thresh, kiwi=True, t2fromleft=use_left)\n alt_upper_timestamp, alt_upper_time, \\\n alt_upper_ts, alt_upper_scores, alt_upper_cum_score, alt_upper_left_used = \\\n extract_timestamp(\n self.upper_field, self.upperOcrBoxesRight, self.modelDigits, self.timestampFormatter,\n thresh, kiwi=True, t2fromleft=use_left)\n\n if self.lower_left_count + self.lower_right_count > 3:\n use_left = self.lower_left_count > self.lower_right_count\n else:\n use_left = None\n\n reg_lower_timestamp, reg_lower_time, \\\n reg_lower_ts, reg_lower_scores, reg_lower_cum_score, reg_lower_left_used = \\\n extract_timestamp(\n self.lower_field, self.lowerOcrBoxesLeft, self.modelDigits, self.timestampFormatter,\n thresh, kiwi=True, t2fromleft=use_left)\n alt_lower_timestamp, alt_lower_time, \\\n alt_lower_ts, alt_lower_scores, alt_lower_cum_score, alt_lower_left_used = \\\n extract_timestamp(\n self.lower_field, self.lowerOcrBoxesRight, self.modelDigits, self.timestampFormatter,\n thresh, kiwi=True, t2fromleft=use_left)\n\n need_to_redisplay_ocr_boxes = False\n if reg_upper_cum_score > alt_upper_cum_score: # lefthand boxes score better than righthand boxes\n if self.currentUpperBoxPos == 'right':\n need_to_redisplay_ocr_boxes = True\n self.currentUpperBoxPos = 'left'\n self.upper_timestamp = reg_upper_timestamp\n self.upper_time = reg_upper_time\n self.upper_ts = reg_upper_ts\n self.upper_scores = reg_upper_scores\n self.upper_cum_score = reg_upper_cum_score\n upper_left_used = reg_upper_left_used\n else:\n if self.currentUpperBoxPos == 'left':\n need_to_redisplay_ocr_boxes = True\n self.currentUpperBoxPos = 'right'\n self.upper_timestamp = alt_upper_timestamp\n self.upper_time = alt_upper_time\n self.upper_ts = alt_upper_ts\n self.upper_scores = alt_upper_scores\n self.upper_cum_score = alt_upper_cum_score\n upper_left_used = alt_upper_left_used\n\n if reg_lower_cum_score > alt_lower_cum_score:\n if self.currentLowerBoxPos == 'right':\n need_to_redisplay_ocr_boxes = True\n self.currentLowerBoxPos = 'left'\n self.lower_timestamp = reg_lower_timestamp\n self.lower_time = reg_lower_time\n self.lower_ts = reg_lower_ts\n self.lower_scores = reg_lower_scores\n self.lower_cum_score = reg_lower_cum_score\n lower_left_used = reg_lower_left_used\n\n else:\n if self.currentLowerBoxPos == 'left':\n need_to_redisplay_ocr_boxes = True\n self.currentLowerBoxPos = 'right'\n self.lower_timestamp = alt_lower_timestamp\n self.lower_time = alt_lower_time\n self.lower_ts = alt_lower_ts\n self.lower_scores = alt_lower_scores\n self.lower_cum_score = alt_lower_cum_score\n lower_left_used = alt_lower_left_used\n\n if self.analysisPaused:\n # When we're manually stepping through an avi, we need to see\n # the actual box placements.\n need_to_redisplay_ocr_boxes = True\n\n if need_to_redisplay_ocr_boxes and self.viewFieldsCheckBox.isChecked():\n self.clearOcrBoxes()\n self.placeOcrBoxesOnImage()\n\n else: # handle non-kiwi VTI here\n # Note: left_used is only useful when kiwi=TRUE\n self.upper_timestamp, self.upper_time, \\\n self.upper_ts, self.upper_scores, self.upper_cum_score, upper_left_used = extract_timestamp(\n self.upper_field, self.upperOcrBoxesLeft, self.modelDigits, self.timestampFormatter, thresh)\n self.lower_timestamp, self.lower_time,\\\n self.lower_ts, self.lower_scores, self.lower_cum_score, lower_left_used = extract_timestamp(\n self.lower_field, self.lowerOcrBoxesLeft, self.modelDigits, self.timestampFormatter, thresh)\n\n if upper_left_used is not None and upper_left_used:\n self.upper_left_count += 1\n else:\n self.upper_right_count += 1\n\n if lower_left_used is not None and lower_left_used:\n self.lower_left_count += 1\n else:\n self.lower_right_count += 1\n\n if printresults:\n if self.kiwiInUse:\n self.showMsg(f'upper field timestamp:{self.upper_timestamp} '\n f'time:{self.upper_time:0.4f} scores:{self.upper_scores} '\n f'{self.upper_left_count}/{self.upper_right_count}',\n blankLine=False)\n self.showMsg(f'lower field timestamp:{self.lower_timestamp} '\n f'time:{self.lower_time:0.4f} scores:{self.lower_scores} '\n f'{self.lower_left_count}/{self.lower_right_count}')\n else:\n self.showMsg(f'upper field timestamp:{self.upper_timestamp} '\n f'time:{self.upper_time:0.4f} scores:{self.upper_scores} ',\n blankLine=False)\n self.showMsg(f'lower field timestamp:{self.lower_timestamp} '\n f'time:{self.lower_time:0.4f} scores:{self.lower_scores} ')\n\n\n if self.detectFieldTimeOrder:\n if self.lower_time >= 0 and self.upper_time >= 0:\n if self.lower_time < self.upper_time:\n self.showMsg(f'Detected bottom field is first in time')\n self.bottomFieldFirstRadioButton.setChecked(True)\n else:\n self.showMsg(f'Detected top field is first in time')\n self.topFieldFirstRadioButton.setChecked(True)\n self.detectFieldTimeOrder = False\n\n return self.upper_timestamp, self.upper_time, self.upper_scores, self.upper_cum_score, \\\n self.lower_timestamp, self.lower_time, self.lower_scores, self.lower_cum_score\n\n def writeFormatTypeFile(self, format_type):\n f_path = os.path.join(self.folder_dir, 'formatter.txt')\n with open(f_path, 'w') as f:\n f.writelines(f'{format_type}')\n\n def vtiSelected(self):\n\n # Clear the flag that we use to automatically detect which field is earliest in time.\n self.detectFieldTimeOrder = False\n\n self.currentVTIindex = self.vtiSelectComboBox.currentIndex()\n\n # dictionaryOfSelection = repr(self.VTIlist[self.currentVTIindex])\n # self.showMsg(f'VTI: {dictionaryOfSelection}')\n\n if not self.avi_in_use or self.image is None:\n return\n\n if not self.avi_wcs_folder_in_use:\n if not self.vtiSelectComboBox.currentIndex() == 0:\n self.showMsg(f'VTI timestamp extraction only supported for AVI-WCS folders')\n self.vtiSelectComboBox.setCurrentIndex(0)\n\n if self.currentVTIindex == 0: # None\n return\n\n self.kiwiInUse = False\n\n self.viewFieldsCheckBox.setChecked(True)\n\n # There is often something messed up with frame 0, so we protect the user\n # by automatically moving to frame 1 in that case\n if self.currentFrameSpinBox.value() == 0:\n self.currentFrameSpinBox.setValue(1)\n\n # Set the flag that we use to automatically detect which field is earliest in time.\n # We only want to do this test once.\n\n self.detectFieldTimeOrder = True\n\n self.showFrame()\n\n self.clearOcrBoxes()\n\n width = self.image.shape[1]\n\n if not (width == 640 or width == 720):\n self.showMsg(f'Unexpected image width of {width}')\n return\n\n self.ocrBoxesDir = self.folder_dir\n self.ocrDigitsDir = self.folder_dir\n\n # Only when Kiwi is in use do the following variables take on any different vales\n self.currentUpperBoxPos = 'left'\n self.currentLowerBoxPos = 'left'\n self.upperOcrBoxesRight = []\n self.lowerOcrBoxesRight = []\n\n if self.currentVTIindex == 1: # IOTA-3 w=640 or 720 full screen mode\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_640_full_screen_mode3()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_720_full_screen_mode3()\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_iota_timestamp\n self.writeFormatTypeFile('iota')\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 2: # IOTA-3 w=640 or 720 safe mode\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_640_safe_mode3()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_720_safe_mode3()\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_iota_timestamp\n self.writeFormatTypeFile('iota')\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 3: # IOTA-2 w=640 and 720 full screen mode\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_640_full_screen_mode2()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_720_full_screen_mode2()\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_iota_timestamp\n self.writeFormatTypeFile('iota')\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 4: # IOTA-2 w=640 and 720 safe mode\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft= setup_for_iota_640_safe_mode2()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_iota_720_safe_mode2()\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_iota_timestamp\n self.writeFormatTypeFile('iota')\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 5: # BoxSprite 3 w=640 and 720\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft= setup_for_boxsprite3_640()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_boxsprite3_720()\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_boxsprite3_timestamp\n self.writeFormatTypeFile('boxsprite')\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 6: # Kiwi w=720 and 640 (left position)\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft= setup_for_kiwi_vti_640_left()\n self.upperOcrBoxesRight, self.lowerOcrBoxesRight= setup_for_kiwi_vti_640_right()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_kiwi_vti_720_left()\n self.upperOcrBoxesRight, self.lowerOcrBoxesRight = setup_for_kiwi_vti_720_right()\n\n self.currentUpperBoxPos = 'left'\n self.currentLowerBoxPos = 'left'\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.kiwiInUse = True\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_kiwi_timestamp\n self.writeFormatTypeFile('kiwi-left')\n self.formatterCode = 'kiwi-left'\n self.extractTimestamps()\n return\n\n if self.currentVTIindex == 7: # Kiwi w=720 and 640 (right position)\n\n if width == 640:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_kiwi_vti_640_left()\n self.upperOcrBoxesRight, self.lowerOcrBoxesRight = setup_for_kiwi_vti_640_right()\n else:\n self.upperOcrBoxesLeft, self.lowerOcrBoxesLeft = setup_for_kiwi_vti_720_left()\n self.upperOcrBoxesRight, self.lowerOcrBoxesRight = setup_for_kiwi_vti_720_right()\n\n self.currentUpperBoxPos = 'right'\n self.currentLowerBoxPos = 'right'\n\n self.ocrboxBasePath = 'custom-boxes'\n self.pickleOcrBoxes()\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.loadModelDigits()\n self.saveModelDigits()\n\n self.kiwiInUse = True\n self.placeOcrBoxesOnImage()\n self.timestampFormatter = format_kiwi_timestamp\n self.writeFormatTypeFile('kiwi-right')\n self.formatterCode = 'kiwi-right'\n self.extractTimestamps()\n return\n\n\n self.showMsg('Not yet implemented')\n return\n\n def loadCustomOcrProfiles(self):\n if not self.avi_wcs_folder_in_use:\n self.showMsg(f'This function only available when an AVI-WCS folder is in use.')\n return\n # all = self.readSavedOcrProfiles(pattern='/pymovie-ocr-profiles*.p')\n profile_dict = self.readSavedOcrProfiles()\n\n code_to_save = self.formatterCode\n\n current_profile = {'id': 'default',\n 'upper-boxes-left': self.upperOcrBoxesLeft,\n 'lower-boxes-left': self.lowerOcrBoxesLeft,\n 'upper-boxes-right': self.upperOcrBoxesRight,\n 'lower-boxes-right': self.lowerOcrBoxesRight,\n 'digits': self.modelDigits,\n 'formatter-code': code_to_save}\n\n selector = SelectProfileDialog(self.showMsg, profile_dict, current_profile)\n result = selector.exec_()\n\n result_code = selector.getResult()\n\n # self.showMsg(f'Selector dialog returned: {result_code}')\n\n # We assume that some change to the profile dictionary may have been made and\n # so simply always re-pickle that dictionary\n my_profile_fn = '/pymovie-ocr-profiles.p'\n pickle.dump(profile_dict, open(self.profilesDir + my_profile_fn, \"wb\"))\n\n if result_code >= 0:\n # self.showMsg(f'Load profile was asked for...')\n profile_selected = result_code\n ocr_dict = profile_dict[profile_selected]\n id_found = ocr_dict['id']\n self.showMsg(f'Loading profile: {id_found}')\n self.clearOcrBoxes()\n self.upperOcrBoxesLeft = ocr_dict['upper-boxes-left']\n self.lowerOcrBoxesLeft = ocr_dict['lower-boxes-left']\n self.upperOcrBoxesRight = ocr_dict['upper-boxes-right']\n self.lowerOcrBoxesRight = ocr_dict['lower-boxes-right']\n self.modelDigits = ocr_dict['digits']\n self.formatterCode = ocr_dict['formatter-code']\n\n # Next we pickle boxes, digits, and write format code txt file and start reading timestamps\n self.pickleOcrBoxes()\n self.saveModelDigits()\n self.writeFormatTypeFile(self.formatterCode)\n\n self.startTimestampReading()\n # self.showFrame()\n\n def generateKiwiOcrBoxesAtRight(self):\n self.showMsg(f'We are now generating the kiwi specific OcrBoxes')\n\n # Compute alternate (right position)\n newUpperBoxes = []\n dx = 11\n for ocrbox in self.upperOcrBoxesLeft:\n xL, xR, yU, yL = ocrbox\n newUpperBoxes.append((xL + dx, xR + dx, yU, yL))\n newLowerBoxes = []\n for ocrbox in self.lowerOcrBoxesLeft:\n xL, xR, yU, yL = ocrbox\n newLowerBoxes.append((xL + dx, xR + dx, yU, yL))\n self.upperOcrBoxesRight = newUpperBoxes[:]\n self.lowerOcrBoxesRight = newLowerBoxes[:]\n\n def changeNavButtonTitles(self):\n if self.frameJumpBig == 200: # FITS titling needed\n self.transportSmallLeft.setText(f'< {self.frameJumpSmall} frames')\n self.transportSmallRight.setText(f'{self.frameJumpSmall} frames >')\n self.transportBigLeft.setText(f'< {self.frameJumpBig} frames')\n self.transportBigRight.setText(f'{self.frameJumpBig} frames >')\n else:\n self.transportSmallLeft.setText(f'- 1 sec')\n self.transportSmallRight.setText(f'+ 1 sec')\n self.transportBigLeft.setText(f'- 10 sec')\n self.transportBigRight.setText(f'+ 10 sec')\n\n def fillApertureDictionaries(self):\n # This will become a list of dictionaries, one for each aperture. The customer\n # for this list is fillApertureTable()\n self.appDictList = []\n for app in self.getApertureList():\n appDict = dict(\n appRef = app,\n name = app.name,\n threshDelta = app.thresh,\n xy = app.getCenter(),\n defMskRadius = app.default_mask_radius,\n color = app.color,\n joggable = app.jogging_enabled,\n autoTextOut = app.auto_display,\n thumbnailSource = app.thumbnail_source,\n outputOrder = app.order_number,\n )\n self.appDictList.append(appDict)\n\n # self.showMsg('appDictList has been filled')\n\n def setThumbnails(self, aperture, showDefaultMaskInThumbnail2):\n # self.showMsg(f'We will execute a thumbnail update on {aperture.name}', blankLine=False)\n # self.showMsg(f'... showDefaultMaskInThumbnail2 is {showDefaultMaskInThumbnail2}')\n self.centerAperture(aperture, show_stats=False)\n if showDefaultMaskInThumbnail2:\n self.getApertureStats(aperture, show_stats=True)\n mask = aperture.defaultMask\n self.thumbTwoView.setImage(mask)\n else:\n self.getApertureStats(aperture, show_stats=True)\n QtGui.QGuiApplication.processEvents()\n\n def editApertures(self):\n # Fill self.appDictList from apertures --- this will be passed to EditApertureDialog\n self.fillApertureDictionaries()\n\n self.apertureEditor = EditApertureDialog(\n self.showMsg,\n saver=self.settings,\n dictList=self.appDictList,\n appSize=self.roi_size,\n threshSpinner=self.threshValueEdit,\n imageUpdate=self.frameView.getView().update,\n setThumbnails=self.setThumbnails\n )\n\n # Set size and position of the dialog window to last known...\n newSize = self.settings.value('appEditDialogSize')\n newPos = self.settings.value('appEditDialogPos')\n if newSize is not None:\n self.apertureEditor.resize(newSize)\n if newPos is not None:\n self.apertureEditor.move(newPos)\n\n self.apertureEditor.show()\n\n def copy_desktop_icon_file_to_home_directory(self):\n if sys.platform == 'linux':\n pass\n elif platform.mac_ver()[0]:\n icon_dest_path = f\"{os.environ['HOME']}{r'/Desktop/run-pymovie'}\"\n if not os.path.exists(icon_dest_path):\n # Here is where the .bat file will be when running an installed pymovie\n icon_src_path = f\"{os.environ['HOME']}\" + r\"/Anaconda3/Lib/site-packages/pymovie/run-pymovie-mac.bat\"\n if not os.path.exists(icon_src_path):\n # But here is where the .bat file is during a development run\n icon_src_path = os.path.join(os.path.split(__file__)[0], 'run-pymovie-mac.bat')\n with open(icon_src_path) as src, open(icon_dest_path, 'w') as dest:\n dest.writelines(src.readlines())\n os.chmod(icon_dest_path, 0o755) # Make it executable\n else:\n # We must be on a Windows machine because Mac version number was empty\n icon_dest_path = r\"C:\\Anaconda3\\PyMovie.bat\"\n\n if not os.path.exists(icon_dest_path):\n # Here is where the .bat file will be when running an installed pymovie\n icon_src_path = r\"C:\\Anaconda3\\Lib\\site-packages\\pymovie\\PyMovie.bat\"\n if not os.path.exists(icon_src_path):\n # But here is where the .bat file is during a development run\n icon_src_path = os.path.join(os.path.split(__file__)[0], 'PyMovie.bat')\n with open(icon_src_path) as src, open(icon_dest_path, 'w') as dest:\n dest.writelines(src.readlines())\n\n def performFrameStacking(self):\n if not (self.avi_wcs_folder_in_use or self.fits_folder_in_use):\n self.showMsg(f'This function can only be performed in the context of an AVI-WCS or FITS folder.')\n return\n\n # Deal with timestamp redaction first.\n # Get a robust mean from near the center of the current image\n y0 = int(self.image.shape[0]/2)\n x0 = int(self.image.shape[1]/2)\n ny = 51\n nx = 51\n thumbnail = self.image[y0:y0 + ny, x0:x0 + nx]\n mean, *_ = robustMeanStd(thumbnail, outlier_fraction=.5)\n\n image_height = self.image.shape[0] # number of rows\n image_width = self.image.shape[1] # number of columns\n\n num_lines_to_redact = 0\n\n early_exit = False\n\n if self.redactLinesEdit.text():\n try:\n num_lines_to_redact = int(self.redactLinesEdit.text())\n except ValueError:\n self.showMsg(f'invalid numeric entry: {self.redactLinesEdit.text()}')\n return\n else:\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText(f'It is necessary to remove any timestamp overlay that may be '\n f'present as such an overlay will keep the image registration '\n f'from working properly.'\n f'\\n\\nPlease enter a number in the redact lines edit box. '\n f'Enter 0 if there is no timestamp.')\n msg.setWindowTitle('Please fill in redact lines')\n msg.setStandardButtons(QMessageBox.Ok)\n msg.exec()\n early_exit = True\n\n if not self.numFramesToStackEdit.text():\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText(f'Please specify the number of frames to stack. '\n f'\\n\\nA number in the range of 100 to 400 would be usual.')\n msg.setWindowTitle('Please fill in num frames')\n msg.setStandardButtons(QMessageBox.Ok)\n msg.exec()\n early_exit = True\n\n if early_exit:\n return\n\n if abs(num_lines_to_redact) > image_height / 2:\n self.showMsg(f'{num_lines_to_redact} is an unreasonable number of lines to redact.')\n self.showMsg(f'Operation aborted.')\n return\n\n redacted_image = self.image[:,:].astype('int16')\n if num_lines_to_redact > 0:\n for i in range(image_height - num_lines_to_redact, image_height):\n for j in range(0, image_width):\n redacted_image[i, j] = mean\n else:\n for i in range(0, abs(num_lines_to_redact)):\n for j in range(0, image_width):\n redacted_image[i, j] = mean\n\n self.image = redacted_image\n self.frameView.setImage(self.image)\n if self.levels:\n self.frameView.setLevels(min=self.levels[0], max=self.levels[1])\n\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText('Is the timestamp completely removed?')\n msg.setWindowTitle('Is timestamp removed')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n ready_for_submission = retval == QMessageBox.Yes\n\n if not ready_for_submission:\n self.showFrame()\n return\n\n first_frame = self.currentFrameSpinBox.value()\n\n try:\n txt = self.numFramesToStackEdit.text()\n num_frames_to_stack = int(txt)\n except ValueError:\n self.showMsg(f'\" {txt} \" is an invalid specification of number of frames to stack')\n return\n\n if num_frames_to_stack > 400:\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText(f'{num_frames_to_stack} is rather large.'\n f'\\n\\nDo you wish to proceed anyway?')\n msg.setWindowTitle('Num frames to stack ok')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n if retval == QMessageBox.No:\n return\n\n last_frame = first_frame + num_frames_to_stack - 1\n last_frame = min(last_frame, self.stopAtFrameSpinBox.maximum())\n\n # Remove the current enhanced-image.fit and associated frame num file\n try:\n os.remove(self.folder_dir + r'/enhanced-image.fit')\n except FileNotFoundError:\n pass\n\n try:\n os.remove(self.folder_dir + r'/enhanced-image-frame-num.txt')\n except FileNotFoundError:\n pass\n\n if self.fits_folder_in_use:\n fitsReader = self.getFitsFrame\n else:\n fitsReader = None\n\n stacker.frameStacker(\n self.showMsg, self.stackerProgressBar, QtGui.QGuiApplication.processEvents,\n first_frame=first_frame, last_frame=last_frame,\n timestamp_trim=num_lines_to_redact,\n fitsReader = fitsReader,\n avi_location=self.avi_location, out_dir_path=self.folder_dir)\n\n # Now that we're back, if we got a new enhanced-image.fit, display it.\n if os.path.isfile(self.folder_dir + r'/enhanced-image.fit'):\n # And now is time to write the frame number of the corresponding reference frame\n with open(self.folder_dir + r'/enhanced-image-frame-num.txt', 'w') as f:\n f.write(f'{first_frame}')\n self.clearApertures()\n self.readFinderImage()\n\n def clearCoordinatesEdit(self):\n self.coordinatesEdit.setText('')\n\n def queryVizier(self):\n self.coordinatesEdit.setText('waiting for response')\n for i in range(10):\n QtGui.QGuiApplication.processEvents()\n\n id_constraint = f'=={self.starIdEdit.text()}'\n star_id = f'UCAC4 {self.starIdEdit.text()}'\n v = Vizier(columns=['_RAJ2000', '_DEJ2000', 'f.mag'],\n column_filters={'UCAC4': id_constraint})\n result = v.query_object(star_id, catalog=['I/322A'])\n if not len(result) == 0:\n ans = result[0]\n c = SkyCoord(ans['_RAJ2000'], ans['_DEJ2000'], frame='icrs')\n loc = c.to_string('hmsdms')\n self.coordinatesEdit.setText(loc[0])\n else:\n self.coordinatesEdit.setText('star not found')\n\n def saveTargetInFolder(self):\n with open(self.folder_dir + r'/target-location.txt', 'w') as f:\n f.writelines(self.coordinatesEdit.text())\n self.showMsg(f'{self.coordinatesEdit.text()} written to target-location.txt')\n self.doManualWcsCalibration()\n\n def yellowAperturePresent(self):\n for app in self.getApertureList():\n if app.color == 'yellow':\n return True\n\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText('You have not designated any yellow apertures. Failing to do so' +\n ' will cause the current apertures to reposition themselves (probably NOT what you want)' +\n ' when the first frame of the video is loaded and each aperture tries to \"snap\" to' +\n ' better locations. Answer NO to get a second chance.')\n msg.setWindowTitle('!!! No yellow aperture(s) set !!!')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n return retval == QMessageBox.Yes\n\n def show3DThumbnail(self):\n if self.thumbOneImage is not None:\n title = f'Frame={self.currentFrameSpinBox.value()} Aperture( {self.thumbnail_one_aperture_name} ) '\n mpl = Qt5MplCanvas(self.thumbOneImage, title=title, invert=self.invertImagesCheckBox.isChecked())\n self.plots.append(mpl)\n mpl.show()\n else:\n self.showMsg(f'There is no Thumbnail One image to show')\n\n def displayKeystroke(self, event):\n if not self.printKeyCodes:\n return\n\n key = event.key()\n modifiers = int(event.modifiers())\n if (key != Qt.Key_Shift and key != Qt.Key_Alt and\n key != Qt.Key_Control and key != Qt.Key_Meta):\n keyname = PyQt5.QtGui.QKeySequence(modifiers + key).toString()\n self.showMsg(f'key(s) pressed: {keyname} raw: {key}')\n\n def processKeystroke(self, event):\n\n key = event.key()\n modifiers = int(event.modifiers())\n\n self.displayKeystroke(event)\n\n if key == ord('K'): # Could be 'k' or 'K'\n if modifiers & Qt.SHIFT == Qt.SHIFT: # it's 'K'\n self.consecutiveKcount += 1\n if self.consecutiveKcount >= 2:\n self.printKeyCodes = True\n elif modifiers == 0:\n self.printKeyCodes = False\n self.consecutiveKcount = 0\n\n joggable_aperture_available = False\n app_list = self.getApertureList()\n\n joggable_ocr_box_available = False\n ocr_list = self.getOcrBoxList()\n\n for app in app_list:\n if app.jogging_enabled:\n joggable_aperture_available = True\n break\n\n for ocr in ocr_list:\n if ocr.joggable:\n joggable_ocr_box_available = True\n break\n\n if not joggable_aperture_available and not joggable_ocr_box_available:\n return True\n\n got_arrow_key = False\n dx = 0\n dy = 0\n if key == Qt.Key_Up:\n if self.printKeyCodes:\n self.showMsg(f'Jogging up')\n dy = -1\n got_arrow_key = True\n elif key == Qt.Key_Down:\n if self.printKeyCodes:\n self.showMsg(f'Jogging down')\n dy = 1\n got_arrow_key = True\n elif key == Qt.Key_Left:\n if self.printKeyCodes:\n self.showMsg(f'Jogging left')\n dx = -1\n got_arrow_key = True\n elif key == Qt.Key_Right:\n if self.printKeyCodes:\n self.showMsg(f'Jogging right')\n dx = 1\n got_arrow_key = True\n\n if not got_arrow_key:\n return False\n\n for app in app_list:\n if app.jogging_enabled:\n # self.showMsg(f'The jog will be applied to {app.name}', blankLine=False)\n jogAperture(app, -dx, -dy)\n if app.auto_display:\n self.getApertureStats(app, show_stats=True)\n\n for ocr in ocr_list:\n if ocr.joggable:\n # The following call also calls pickleOcrBoxes\n self.jogSingleOcrBox(dx=dx, dy=dy,\n boxnum=ocr.boxnum,\n position=ocr.position, ocr=ocr)\n\n self.frameView.getView().update()\n\n return True\n\n # Diagnostic/debug/exploratory code\n # MOD_MASK = (Qt.CTRL | Qt.ALT | Qt.SHIFT | Qt.META)\n #\n # keyname = ''\n # key = event.key()\n # modifiers = int(event.modifiers())\n # if (modifiers and modifiers & MOD_MASK == modifiers and\n # key > 0 and key != Qt.Key_Shift and key != Qt.Key_Alt and\n # key != Qt.Key_Control and key != Qt.Key_Meta):\n # keyname = PyQt5.QtGui.QKeySequence(modifiers + key).toString()\n #\n # self.showMsg(f'event.text(): {event.text()}')\n # self.showMsg(f'event.key(): {keyname}')\n #\n # self.showMsg(f'key pressed was: {key}')\n\n def invertImages(self):\n self.frameView.view.invertY(not self.invertImagesCheckBox.isChecked())\n self.thumbOneView.view.invertY(not self.invertImagesCheckBox.isChecked())\n self.thumbTwoView.view.invertY(not self.invertImagesCheckBox.isChecked())\n\n def toggleImageControl(self):\n if self.showImageControlCheckBox.isChecked():\n self.frameView.ui.histogram.show()\n self.frame_at_level_set = self.currentFrameSpinBox.value()\n else:\n self.frame_at_level_set = None\n self.frameView.ui.histogram.hide()\n self.levels = self.frameView.ui.histogram.getLevels()\n self.showMsg(f'New scaling levels: black={self.levels[0]:0.1f} white={self.levels[1]:0.1f}')\n\n def jumpSmallFramesBack(self):\n newFrame = self.currentFrameSpinBox.value() - self.frameJumpSmall\n self.currentFrameSpinBox.setValue(max(0, newFrame))\n\n def jumpBigFramesBack(self):\n newFrame = self.currentFrameSpinBox.value() - self.frameJumpBig\n self.currentFrameSpinBox.setValue(max(0, newFrame))\n\n def jumpSmallFramesForward(self):\n newFrame = self.currentFrameSpinBox.value() + self.frameJumpSmall\n maxFrame = self.stopAtFrameSpinBox.maximum()\n self.currentFrameSpinBox.setValue(min(maxFrame, newFrame))\n\n def jumpBigFramesForward(self):\n newFrame = self.currentFrameSpinBox.value() + self.frameJumpBig\n maxFrame = self.stopAtFrameSpinBox.maximum()\n self.currentFrameSpinBox.setValue(min(maxFrame, newFrame))\n\n def changePlotSymbolSize(self):\n self.plot_symbol_size = self.plotSymbolSizeSpinBox.value()\n\n def updateFrameWithTracking(self):\n if not self.analysisRequested:\n self.initializeTracking()\n self.showFrame()\n\n def disableControlsWhenNoData(self):\n self.savedStateFrameNumber = None\n\n self.saveApertureState.setEnabled(False)\n self.restoreApertureState.setEnabled(False)\n\n self.viewFieldsCheckBox.setEnabled(False)\n self.currentFrameSpinBox.setEnabled(False)\n\n self.setTransportButtonEnableState(False)\n self.transportReturnToMark.setEnabled(False)\n\n self.processAsFieldsCheckBox.setEnabled(False)\n self.topFieldFirstRadioButton.setEnabled(False)\n self.bottomFieldFirstRadioButton.setEnabled(False)\n\n def setTransportButtonEnableState(self, state):\n self.transportMaxLeft.setEnabled(state)\n self.transportBigLeft.setEnabled(state)\n self.transportSmallLeft.setEnabled(state)\n self.transportMinusOneFrame.setEnabled(state)\n self.transportPlayLeft.setEnabled(state)\n self.transportPause.setEnabled(state)\n self.transportAnalyze.setEnabled(state)\n self.transportPlayRight.setEnabled(state)\n self.transportPlusOneFrame.setEnabled(state)\n self.transportSmallRight.setEnabled(state)\n self.transportBigRight.setEnabled(state)\n self.transportMaxRight.setEnabled(state)\n # self.transportReturnToMark.setEnabled(state)\n self.transportMark.setEnabled(state)\n\n def enableControlsForAviData(self):\n\n self.setTransportButtonEnableState(True)\n self.transportReturnToMark.setEnabled(False)\n\n self.saveApertureState.setEnabled(True)\n\n self.viewFieldsCheckBox.setEnabled(True)\n self.currentFrameSpinBox.setEnabled(True)\n self.processAsFieldsCheckBox.setEnabled(True)\n self.topFieldFirstRadioButton.setEnabled(True)\n self.bottomFieldFirstRadioButton.setEnabled(True)\n\n def enableControlsForFitsData(self):\n\n self.setTransportButtonEnableState(True)\n self.transportReturnToMark.setEnabled(False)\n\n self.saveApertureState.setEnabled(True)\n\n self.currentFrameSpinBox.setEnabled(True)\n self.viewFieldsCheckBox.setChecked(False)\n self.viewFieldsCheckBox.setEnabled(False)\n\n def getStarPositionString(self):\n starPos = StarPositionDialog()\n starPos.RaHours.setFocus()\n starPos.apiKeyEdit.setText(self.settings.value('api_key'))\n\n result = starPos.exec_()\n\n if result == QDialog.Accepted:\n # Now we extract all the fields\n if not starPos.singleLineEdit.text():\n valid_entry = True\n if not starPos.RaHours.text():\n valid_entry = False\n if not starPos.RaMinutes.text():\n valid_entry = False\n if not starPos.RaSeconds.text():\n valid_entry = False\n if not starPos.DecDegrees.text():\n valid_entry = False\n if not starPos.DecMinutes.text():\n valid_entry = False\n if not starPos.DecSeconds.text():\n valid_entry = False\n if not valid_entry:\n self.settings.setValue('api_key', starPos.apiKeyEdit.text())\n return ''\n ss = starPos.RaHours.text() + 'h'\n ss += starPos.RaMinutes.text() + 'm'\n ss += starPos.RaSeconds.text() + 's '\n dec_degrees = starPos.DecDegrees.text()\n if not (dec_degrees.startswith('+') or dec_degrees.startswith('-')):\n ss += '+' + dec_degrees + 'd'\n else:\n ss += dec_degrees + 'd'\n ss += starPos.DecMinutes.text() + 'm'\n ss += starPos.DecSeconds.text() + 's'\n self.settings.setValue('api_key', starPos.apiKeyEdit.text())\n return ss\n else:\n self.settings.setValue('api_key', starPos.apiKeyEdit.text())\n return starPos.singleLineEdit.text()\n\n else:\n return ''\n\n def nameAperture(self, aperture):\n appNamerThing = AppNameDialog()\n appNamerThing.apertureNameEdit.setText(aperture.name)\n appNamerThing.apertureNameEdit.setFocus()\n result = appNamerThing.exec_()\n\n if result == QDialog.Accepted:\n aperture.name = appNamerThing.apertureNameEdit.text()\n\n def setRoiFromComboBox(self):\n self.clearApertures()\n self.roi_size = int(self.roiComboBox.currentText())\n self.roi_center = int(self.roi_size / 2)\n if self.image is not None:\n height, width = self.image.shape\n self.roi_max_x = width - self.roi_size\n self.roi_max_y = height - self.roi_size\n\n def buildDefaultMask(self, radius=4.5):\n # Create the default mask\n self.defaultMask = np.zeros((self.roi_size, self.roi_size), 'int16')\n self.defaultMaskPixelCount = 0\n c = self.roi_center\n r = int(np.ceil(radius))\n for i in range(c - r - 1, c + r + 2):\n for j in range(c - r - 1, c + r + 2):\n if (i - c)**2 + (j - c)**2 <= radius**2:\n self.defaultMaskPixelCount += 1\n self.defaultMask[i,j] = 1\n # self.showMsg(f'The current default mask contains {self.defaultMaskPixelCount} pixels')\n\n def resetMaxStopAtFrameValue(self):\n self.stopAtFrameSpinBox.setValue(self.stopAtFrameSpinBox.maximum())\n\n def showFitsMetadata(self):\n if self.fits_filenames:\n frame = self.currentFrameSpinBox.value()\n\n # The following line prints to console --- use for diagnostics only\n # pyfits.info(self.fits_filenames[frame])\n\n file_name = self.fits_filenames[frame]\n hdr = pyfits.getheader(file_name, 0)\n msg = repr(hdr)\n self.showMsg(f'############### Start frame {frame}:{file_name} data ###############')\n self.showMsg(msg)\n self.showMsg(f'################# End frame {frame}:{file_name} data ###############')\n\n # pyfits.info(file_name) # This prints to the console only\n\n self.num_yellow_apertures = 0\n for app in self.getApertureList():\n if app.color == 'yellow':\n self.num_yellow_apertures += 1\n\n # self.num_yellow_apertures can only take on values of 0, 1, and 2. This is enforced\n # by self.handleSetYellowSignal()\n\n # If there are two yellow apertures, we need to record the initial geometries\n if self.num_yellow_apertures == 2:\n yellow_count = 0\n for app in self.getApertureList():\n if app.color == 'yellow' and yellow_count == 0: # This our yellow #1\n yellow_count += 1\n if self.use_yellow_mask:\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False, save_yellow_mask=True)\n else:\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n app.dx = 0\n app.dy = 0\n app.theta = 0.0\n\n # Save the coordinates of yellow #1 aperture\n self.yellow_x = xc_world\n self.yellow_y = yc_world\n\n elif app.color == 'yellow' and yellow_count == 1: # This is our yellow #2\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n\n # Get the distance and angle measurements back to yellow #1\n dy = yc_world - self.yellow_y\n dx = xc_world - self.yellow_x\n\n app.dx = dx\n app.dy = dy\n app.theta, _ = calcTheta(dx, dy)\n\n # Set the current field rotation angle\n self.delta_theta = 0.0\n\n for app in self.getApertureList():\n if not app.color == 'yellow':\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n app.xc = xc_world\n app.yc = yc_world\n\n # Get the distance measurements back to yellow #1\n dy = yc_world - self.yellow_y\n dx = xc_world - self.yellow_x\n\n app.dx = dx\n app.dy = dy\n app.theta = None # We don't use this value during tracking\n\n def autoPlayLeft(self):\n self.setTransportButtonEnableState(False)\n self.transportPause.setEnabled(True)\n self.transportReturnToMark.setEnabled(False)\n\n self.initializeTracking()\n\n currentFrame = self.currentFrameSpinBox.value()\n lastFrame = self.stopAtFrameSpinBox.value()\n while not self.playPaused:\n if currentFrame == 0:\n self.playPaused = True\n self.setTransportButtonEnableState(True)\n mark_available = not self.savedStateFrameNumber is None\n self.transportReturnToMark.setEnabled(mark_available)\n return\n else:\n currentFrame -= 1\n self.currentFrameSpinBox.setValue(currentFrame)\n QtGui.QGuiApplication.processEvents()\n\n self.setTransportButtonEnableState(True)\n mark_available = not self.savedStateFrameNumber is None\n self.transportReturnToMark.setEnabled(mark_available)\n\n def autoPlayRight(self):\n self.setTransportButtonEnableState(False)\n self.transportPause.setEnabled(True)\n self.transportReturnToMark.setEnabled(False)\n\n self.initializeTracking()\n\n currentFrame = self.currentFrameSpinBox.value()\n lastFrame = self.stopAtFrameSpinBox.value()\n while not self.playPaused:\n if currentFrame == lastFrame:\n self.playPaused = True\n self.setTransportButtonEnableState(True)\n mark_available = not self.savedStateFrameNumber is None\n self.transportReturnToMark.setEnabled(mark_available)\n return\n else:\n currentFrame += 1\n self.currentFrameSpinBox.setValue(currentFrame)\n QtGui.QGuiApplication.processEvents()\n\n self.setTransportButtonEnableState(True)\n mark_available = not self.savedStateFrameNumber is None\n self.transportReturnToMark.setEnabled(mark_available)\n\n def autoRun(self):\n if self.analysisRequested:\n\n # We need to not record the current frame if we got here following\n # a pause.\n if self.analysisInProgress:\n pass\n else:\n self.analysisInProgress = True\n if self.viewFieldsCheckBox.isChecked():\n # This toggles the checkbox and so causes a call to self.showFrame()\n self.viewFieldsCheckBox.setChecked(False)\n self.viewFieldsCheckBox.setEnabled(False)\n else:\n # We make this call so that we record the frame data for the current frame.\n self.showFrame()\n\n # Go count yellow apertures to determine type of tracking that we'll be doing.\n # This will initialize the aperture geometries (distances to yellow #1)\n # if we have two yellow tracking apertures in use.\n self.initializeTracking()\n\n currentFrame = self.currentFrameSpinBox.value()\n lastFrame = self.stopAtFrameSpinBox.value()\n stop_offset = 0\n if currentFrame > lastFrame:\n stop_offset = 1\n\n while self.analysisRequested:\n currentFrame = self.currentFrameSpinBox.value()\n lastFrame = self.stopAtFrameSpinBox.value()\n\n if currentFrame == lastFrame + stop_offset:\n self.analysisPaused = True\n self.analysisRequested = False\n self.setTransportButtonEnableState(True)\n mark_available = not self.savedStateFrameNumber is None\n self.transportReturnToMark.setEnabled(mark_available)\n return\n else:\n if currentFrame > lastFrame:\n currentFrame -= 1\n else:\n currentFrame += 1\n # The value change that we do here will automatically trigger\n # a call to self.showFrame() which causes data to be recorded\n self.currentFrameSpinBox.setValue(currentFrame)\n QtGui.QGuiApplication.processEvents()\n else:\n self.viewFieldsCheckBox.setEnabled(True)\n\n def clearApertureData(self):\n self.analysisInProgress = False\n for app in self.getApertureList():\n app.data = []\n app.last_theta = None\n self.showMsg(f'All aperture data has been removed.')\n\n def prepareAutorunPyoteFile(self, csv_file):\n with open(self.folder_dir + '/auto_run_pyote.py', \"w\") as f:\n f.writelines('import sys\\n\\n')\n f.writelines('# The following path is needed to locate pyoteapp\\n')\n f.writelines(f'sys.path.append(r\"{Path(site.getusersitepackages())}\")\\n\\n')\n f.writelines('# The following path(s) is/are needed to locate standard packages\\n')\n for path in site.getsitepackages():\n f.writelines(f'sys.path.append(r\"{Path(path)}\")\\n')\n f.writelines('\\n')\n f.writelines('from pyoteapp import pyote\\n')\n f.writelines(f'pyote.main(r\"{Path(csv_file)}\")\\n')\n\n def writeCsvFile(self):\n\n def sortOnFrame(val):\n return val[8]\n\n options = QFileDialog.Options()\n # options |= QFileDialog.DontUseNativeDialog\n\n if self.fits_folder_in_use:\n filename, _ = QFileDialog.getSaveFileName(\n self, # parent\n \"Select video file\", # title for dialog\n self.settings.value('fitsdir', \"./\"), # starting directory\n \"csv files (*.csv);; all files (*.*)\",\n options=options\n )\n else:\n filename, _ = QFileDialog.getSaveFileName(\n self, # parent\n \"Select video file\", # title for dialog\n self.settings.value('avidir', \"./\"), # starting directory\n \"csv files (*.csv);; all files (*.*)\",\n options=options\n )\n\n QtGui.QGuiApplication.processEvents()\n\n if filename:\n self.showMsg(f'Output file selected: {filename}')\n\n appdata = [] # Will become a list of list of lists\n names = [] # A simple list of aperture names\n order = []\n num_data_pts = None\n\n for app in self.getApertureList():\n names.append(app.name)\n order.append(app.order_number)\n # Sort the data points into frame order (to support running backwards)\n app.data.sort(key=sortOnFrame)\n # app.data is a list of lists, so appdata will become a list of list of lists\n appdata.append(app.data)\n num_data_pts = len(app.data)\n\n num_apps = len(names) # Number of apertures\n\n # Sort names and appData in user specified order\n answer = sort_together([order, names, appdata], key_list=[0])\n names = answer[1]\n appdata = answer[2]\n\n with open(filename, 'w') as f:\n # Standard header (single line)\n f.write(f'# PyMovie Version {version.version()}\\n')\n f.write(f'# source: {self.filename}\\n')\n\n if not self.avi_in_use:\n f.write(f'# date at frame 0: {self.fits_date}\\n')\n\n # csv column headers with aperture names in entry order\n f.write(f'FrameNum,timeInfo')\n # Put all signals in the first columns so that R-OTE and PyOTE can read the file\n for name in names:\n f.write(f',signal-{name}')\n for name in names:\n f.write(f',appsum-{name},avgbkg-{name},stdbkg-{name},nmaskpx-{name},'\n f'maxpx-{name},xcentroid-{name},ycentroid-{name}')\n f.write('\\n')\n\n # Now we add the data lines\n for i in range(num_data_pts):\n frame = appdata[0][i][8] # [aperture index][data group][data id]\n\n timestamp = appdata[0][i][12]\n\n f.write(f'{frame:0.2f},{timestamp}')\n for k in range(num_apps):\n signal = appdata[k][i][4]\n f.write(f',{signal:0.2f}')\n\n for k in range(num_apps):\n appsum = appdata[k][i][5]\n bkgnd = appdata[k][i][6]\n std = appdata[k][i][11]\n nmskpx = appdata[k][i][7]\n maxpx = appdata[k][i][10]\n xcentroid = appdata[k][i][2]\n ycentroid = appdata[k][i][3]\n\n f.write(f',{appsum:0.2f},{bkgnd:0.2f},{std:0.2f},{nmskpx},{maxpx}')\n if xcentroid is not None:\n f.write(f',{xcentroid:0.2f},{ycentroid:0.2f}')\n else:\n f.write(f',,')\n f.write('\\n')\n f.flush()\n\n if self.runPyote.isChecked():\n # We need to prepare a script that is unique to the user's platform\n # and to include a path to the csv file to be given to PyOTE\n self.prepareAutorunPyoteFile(filename)\n\n # Next, we run that script.\n # We use Popen so that we don't have to wait for the process to complete (i.e.,\n # for the user to quit using PyOTE) and so that multiple PyOTE processes\n # can be running at the same time.\n subprocess.Popen(f'python \"{self.folder_dir + \"/auto_run_pyote.py\"}\" ', shell=True)\n self.showMsg(f'##### PyOTE is starting up --- this takes a few seconds #####')\n\n def trackerPresent(self):\n for app in self.getApertureList():\n if app.color == 'yellow':\n return True\n return False\n\n def changeThreshold(self):\n new_thresh = int(self.threshValueEdit.value())\n for app in self.getApertureList():\n if app.color == 'green':\n app.thresh = new_thresh\n if self.trackerPresent():\n self.getApertureStats(app)\n else:\n self.centerAperture(app, show_stats=True)\n\n def handleYellowMaskClick(self):\n self.use_yellow_mask = self.useYellowMaskCheckBox.isChecked()\n\n def computeInitialThreshold(self, aperture):\n\n # This method is called by a click on an item in a context menu.\n # Calling .processEvents() gives the GUI an opportunity to close that menu.\n QtGui.QGuiApplication.processEvents()\n\n # Grap the properties that we need from the aperture object\n bbox = aperture.getBbox()\n x0, y0, nx, ny = bbox\n\n # img is the portion of the main image that is covered by the aperture bounding box\n img = self.image[y0:y0 + ny, x0:x0 + nx]\n\n bkavg, std, *_ = robustMeanStd(img)\n\n background = int(np.ceil(bkavg))\n\n thresh = background + int(np.ceil(std))\n\n aperture.thresh = thresh - background\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(aperture.thresh)\n\n def eventFilter(self, obj, event):\n if event.type() == QtCore.QEvent.KeyPress:\n handled = self.processKeystroke(event)\n if handled:\n return True\n else:\n return super(PyMovie, self).eventFilter(obj, event)\n\n if event.type() == QtCore.QEvent.MouseButtonPress:\n if event.button() == Qt.RightButton:\n if obj.toolTip():\n self.helperThing.textEdit.clear()\n self.helperThing.textEdit.insertHtml(obj.toolTip())\n self.helperThing.raise_()\n self.helperThing.show()\n return True\n return super(PyMovie, self).eventFilter(obj, event)\n # return False\n\n if event.type() == QtCore.QEvent.ToolTip:\n return True\n\n return super(PyMovie, self).eventFilter(obj, event)\n # return False\n\n @pyqtSlot('PyQt_PyObject')\n def handleAppSignal(self, aperture): # aperture is an instance of MeasurementAperture\n self.getApertureStats(aperture)\n\n @pyqtSlot('PyQt_PyObject')\n def handleRecenterSignal(self, aperture):\n self.centerAperture(aperture)\n self.frameView.getView().update()\n\n @pyqtSlot('PyQt_PyObject')\n def handleSetGreenSignal(self, aperture):\n for app in self.getApertureList():\n if app.color == 'green':\n app.setRed()\n aperture.setGreen()\n if aperture.thresh is not None:\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(aperture.thresh)\n\n\n @pyqtSlot('PyQt_PyObject')\n def handleSetYellowSignal(self, aperture):\n num_yellow_apertures = 0\n for app in self.getApertureList():\n if app.color == 'yellow':\n num_yellow_apertures += 1\n if num_yellow_apertures < 2:\n aperture.pen = pg.mkPen('y')\n aperture.color = 'yellow'\n self.frameView.getView().update()\n else:\n self.showMsg(f' !!!! Only two yellow apertures are allowed at a time !!!!')\n\n @pyqtSlot('PyQt_PyObject')\n def handleSetThreshSignal(self, aperture):\n try:\n thresh = int(self.threshValueEdit.text())\n aperture.thresh = thresh\n self.getApertureStats(aperture)\n except ValueError:\n self.showMsg(f'Bad input string for thresh')\n\n @pyqtSlot('PyQt_PyObject')\n def handleDeleteSignal(self, aperture):\n # self.showMsg(f'Aperture {aperture.name} has asked to be removed')\n self.removeAperture(aperture)\n\n @pyqtSlot('PyQt_PyObject')\n def handleSetThumbnailSourceSignal(self, aperture):\n for app in self.getApertureList():\n app.thumbnail_source = False\n aperture.thumbnail_source = True\n\n def makeApertureSignalToSlotConnection(self, app_object):\n app_object.sendAperture.connect(self.handleAppSignal)\n\n def disconnectApertureSignalToSlot(self, app_object):\n app_object.sendAperture.disconnect(self.handleAppSignal)\n\n def makeRecenterSignalToSlotConnection(self, app_object):\n app_object.sendRecenter.connect(self.handleRecenterSignal)\n\n def disconnectRecenterSignalToSlot(self, app_object):\n app_object.sendRecenter.disconnect(self.handleRecenterSignal)\n\n def makeSetGreenSignalToSlotConnection(self, app_object):\n app_object.sendSetGreen.connect(self.handleSetGreenSignal)\n\n def disconnectSetGreenSignalToSlot(self, app_object):\n app_object.sendSetGreen.disconnect(self.handleSetGreenSignal)\n\n def makeSetYellowSignalToSlotConnection(self, app_object):\n app_object.sendSetYellow.connect(self.handleSetYellowSignal)\n\n def disconnectSetYellowSignalToSlot(self, app_object):\n app_object.sendSetYellow.disconnect(self.handleSetYellowSignal)\n\n def makeDeleteSignalToSlotConnection(self, app_object):\n app_object.sendDelete.connect(self.handleDeleteSignal)\n\n def disconnectDeleteSignalToSlot(self, app_object):\n app_object.sendDelete.disconnect(self.handleDeleteSignal)\n\n def makeSetThreshSignalToSlotConnection(self, app_object):\n app_object.sendSetThresh.connect(self.handleSetThreshSignal)\n\n def disconnectSetThreshSignalToSlot(self, app_object):\n app_object.sendSetThresh.disconnect(self.handleSetThreshSignal)\n\n def makeSetThumbnailSourceSignalToSlotConnection(self, app_object):\n app_object.sendThumbnailSource.connect(self.handleSetThumbnailSourceSignal)\n\n def disconnectSetThumbnailSourceSignalToSlot(self, app_object):\n app_object.sendThumbnailSource.disconnect(self.handleSetThumbnailSourceSignal)\n\n def makeSetRaDecSignalToSlotConnection(self, app_object):\n app_object.sendSetRaDec.connect(self.handleSetRaDecSignal)\n\n def disconnectSetRaDecSignalToSlot(self, app_object):\n app_object.sendSetRaDec.disconnect(self.handleSetRaDecSignal)\n\n def handleSetRaDecSignal(self, aperture):\n if self.manual_wcs_state == 0:\n self.showMsg(f'There is no manual WCS procedure active at the moment!')\n return\n\n # Grab the coordinates and validate them\n ss = self.coordinatesEdit.text()\n x = aperture.getCenter()[0]\n y = aperture.getCenter()[1]\n\n xy_loc = f'x={x} y={y}'\n self.showMsg(f'aperture {aperture.name} icrs coord: ({ss}) @ {xy_loc}')\n try:\n coord = SkyCoord(ss, frame='icrs')\n except Exception as e:\n self.showMsg(f'Bad coordinate string: {e}')\n return\n\n if self.manual_wcs_state == 1:\n with open(self.folder_dir + r'/ref1-data.txt', 'w') as f:\n f.write(ss + '\\n')\n f.write(str(x) + '\\n')\n f.write(str(y) + '\\n')\n self.showMsg(f'Reference star 1 data recorded: waiting for aperture 2 to be placed and RA DEC assigned.')\n self.manual_wcs_state +=1\n return\n\n if self.manual_wcs_state == 2:\n with open(self.folder_dir + r'/ref2-data.txt', 'w') as f:\n f.write(ss + '\\n')\n f.write(str(x) + '\\n')\n f.write(str(y) + '\\n')\n self.showMsg(f'Reference star 2 data recorded. Frame calibration started.')\n self.manual_wcs_state = 0\n self.doManualWcsCalibration()\n return\n\n def readManualCalibrationDataFile(self, filename):\n lines = []\n try:\n with open(filename) as f:\n for line in f:\n lines.append(line)\n except FileNotFoundError:\n return False, 0.0, 0.0, 0, 0\n\n if not len(lines) == 3:\n self.showMsg(f'{len(lines)} lines were read when 3 expected')\n return False, 0.0, 0.0, 0, 0\n\n c = SkyCoord(lines[0], frame='icrs')\n ra = c.ra.degree\n dec = c.dec.deg\n x = int(lines[1])\n y = int(lines[2])\n return True, ra, dec, x, y\n\n def doManualWcsCalibration(self):\n\n self.getPixelAspectRatio()\n if self.pixelAspectRatio is None:\n self.showMsg(f'Failed to compute a valid pixel aspect ratio. Cannot continue')\n self.showMsgDialog(f'You must fill in pixel height and width in order to continue.')\n return\n\n file_missing = False\n fpath = self.folder_dir + r'/ref1-data.txt'\n ok, ra1, dec1, x1, y1 = self.readManualCalibrationDataFile(fpath)\n if ok:\n # Make the dictionary item solve_triangle() will want to see\n ref1 = {'ra': ra1, 'dec': dec1, 'x': x1, 'y': y1}\n # self.showMsg(f'RA: {ra1:0.5f} Dec: {dec1:0.5f} x: {x1} y: {y1}')\n self.showMsg(f'ref1 data= {repr(ref1)}')\n else:\n self.showMsg(f'reference 1 data file not found.')\n file_missing = True\n\n fpath = self.folder_dir + r'/ref2-data.txt'\n ok, ra2, dec2, x2, y2 = self.readManualCalibrationDataFile(fpath)\n if ok:\n # Make the dictionary item solve_triangle() will want to see\n ref2 = {'ra': ra2, 'dec': dec2, 'x': x2, 'y': y2}\n # self.showMsg(f'RA: {ra2:0.5f} Dec: {dec2:0.5f} x: {x2} y: {y2}')\n self.showMsg(f'ref2 data= {repr(ref2)}')\n else:\n self.showMsg(f'reference 2 data file not found.')\n file_missing = True\n\n if file_missing:\n self.showMsg(f'Cannot place target aperture because of missing data.')\n return\n\n # So far so good. Now let's get target icrs coords\n try:\n with open(self.folder_dir + r'/target-location.txt') as f:\n ss = f.readline()\n except FileNotFoundError:\n self.showMsg(f'You need to set a target location now.')\n return\n\n c = SkyCoord(ss, frame='icrs')\n ra_target = c.ra.degree\n dec_target = c.dec.deg\n\n # self.showMsg(f'{ra_target} {dec_target}')\n # Make targ dictionary that solve_triangle will need\n targ = {'ra': ra_target, 'dec': dec_target, 'x': None, 'y': None}\n # self.showMsg(repr(targ))\n\n plate_scale = None\n # plate_scale_str = self.plateScaleEdit.text()\n # if plate_scale_str:\n # try:\n # plate_scale = float(plate_scale_str)\n # except ValueError:\n # self.showMsg(f'{plate_scale_str} is an invalid entry.')\n # return\n\n if ref1['x'] > ref2['x']:\n flip_x = ref1['ra'] < ref2['ra']\n else:\n flip_x = ref1['ra'] > ref2['ra']\n\n if ref1['y'] > ref2['y']:\n flip_y = ref1['dec'] < ref2['dec']\n else:\n flip_y = ref1['dec'] > ref2['dec']\n\n self.showMsg(f'flip_x: {flip_x} flip_y: {flip_y}')\n\n solution, plate_scale, targ_theta, ra_dec_x_y_rotation = wcs_helper_functions.solve_triangle(\n ref1, ref2, targ, self.pixelAspectRatio, plate_scale=plate_scale, xflipped=flip_x, yflipped=flip_y\n )\n\n self.showMsg(f'solution: {repr(solution)}', blankLine=False)\n self.showMsg(f'plate_scale: {plate_scale:0.5f} arc-seconds/pixel'\n f' ref1-to-target angle: {targ_theta:0.1f} degrees', blankLine=False)\n self.showMsg(f'ra_dec_x_y angle: {ra_dec_x_y_rotation:0.1f} degrees')\n self.showMsg(\"\", blankLine=False)\n\n # The -0.5 is meant to correct for the fact that RA DEC coords are associated with\n # the upper left corner of a pixel. But it seems to make sense to associate RA DEC\n # coords with the center of a pixel. The 0.5 'moves' the pixel a half step to the left\n # and a half step up to simulate the association of RA DEC with center pixel\n x_calc = int(round(solution['x'] - 0.5))\n y_calc = int(round(solution['y'] - 0.5))\n\n target_app = self.addApertureAtPosition(x_calc, y_calc)\n target_app.thresh = self.big_thresh\n target_app.name = 'target'\n target_app.setRed()\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(self.big_thresh) # Causes call to self.changeThreshold()\n return\n\n def addSnapAperture(self):\n if self.image is None: # Don't add an aperture if there is no image showing yet.\n return\n\n self.one_time_suppress_stats = True\n aperture = self.addGenericAperture()\n\n self.nameAperture(aperture)\n\n self.computeInitialThreshold(aperture)\n\n def addNamedStaticAperture(self):\n self.addStaticAperture(askForName=True)\n\n def addStaticAperture(self, askForName = True):\n if self.image is None: # Don't add an aperture if there is no image showing yet.\n return\n\n aperture = self.addGenericAperture() # This adds a green aperture\n aperture.thresh = self.big_thresh\n\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(self.big_thresh) # Causes call to self.changeThreshold()\n\n if askForName:\n self.nameAperture(aperture)\n\n def addOcrAperture(self, fieldbox, boxnum, position):\n\n aperture = OcrAperture(\n fieldbox,\n boxnum,\n position,\n msgRoutine=self.showMsg,\n templater=self.processOcrTemplate,\n jogcontroller=self.setAllOcrBoxJogging,\n showcharacter=self.showOcrCharacter,\n showtemplates=self.showDigitTemplates,\n neededdigits=self.needDigits,\n kiwi=self.kiwiInUse,\n # samplemenu=self.enableOcrTemplateSampling\n samplemenu=True\n )\n view = self.frameView.getView()\n view.addItem(aperture)\n\n def needDigits(self):\n needs_list = []\n for img in self.modelDigits:\n needs_list.append(img is None)\n return needs_list\n\n def showDigitTemplates(self, retrain=False):\n x_size = None\n y_size = None\n max_pixel = None\n\n if retrain:\n for i, _ in enumerate(self.modelDigits):\n self.modelDigits[i] = None\n self.saveModelDigits()\n self.acceptAviFolderDirectoryWithoutUserIntervention = True\n self.showMissingModelDigits()\n return\n\n for img in self.modelDigits:\n if not img is None:\n y_size, x_size = img.shape\n max_pixel = img.max()\n break\n\n if x_size is None:\n self.showMsg(f'There are no model digits to display.')\n return\n else:\n self.showMsg(f'model digits height:{y_size} width:{x_size}')\n\n if max_pixel == 1:\n border_value = 1\n else:\n border_value = 255\n\n blank = np.zeros((y_size, x_size), dtype='uint8')\n\n ok_to_print_confusion_matrix = True\n\n max_px_value = 0\n digits = self.modelDigits.copy()\n spaced_digits = []\n for i, digit in enumerate(digits):\n if digit is None:\n digits[i] = blank\n ok_to_print_confusion_matrix = False\n else:\n max_px = np.max(digit)\n if max_px > max_px_value:\n max_px_value = max_px\n\n blk_border = cv2.copyMakeBorder(digits[i], 1, 1, 1, 1, cv2.BORDER_CONSTANT, value=0)\n wht_border = cv2.copyMakeBorder(blk_border, 1, 1, 1, 1, cv2.BORDER_CONSTANT, value=border_value)\n spaced_digits.append(wht_border)\n digits_strip = cv2.hconcat(spaced_digits[:])\n\n p = pg.image(digits_strip)\n p.ui.menuBtn.hide()\n p.ui.roiBtn.hide()\n p.ui.histogram.hide()\n self.showMsg(f'max pixel value: {max_px_value}')\n p.ui.histogram.setLevels(0, max_px_value)\n\n if ok_to_print_confusion_matrix:\n print_confusion_matrix(self.modelDigits, self.showMsg)\n\n def showOcrCharacter(self, ocrbox):\n self.currentOcrBox = ocrbox\n self.showOcrboxInThumbnails(ocrbox)\n\n def showOcrboxInThumbnails(self, ocrbox):\n img = timestamp_box_image(self.image_fields, ocrbox, kiwi=self.kiwiInUse)\n self.thumbOneImage = img\n self.thumbOneView.setImage(img)\n self.thumbTwoImage = img\n self.thumbTwoView.setImage(img)\n return img\n\n def processOcrTemplate(self, digit, ocrbox):\n self.showMsg(f'Recording digit {digit} from pixels in {ocrbox}')\n t_img = self.showOcrboxInThumbnails(ocrbox)\n\n if self.formatterCode == 'kiwi-left' or self.formatterCode == 'kiwi-right':\n # blurred_t_img = cv2.GaussianBlur(t_img, ksize=(5, 5), sigmaX=0)\n # self.modelDigits[digit] = blurred_t_img\n self.modelDigits[digit] = t_img\n else:\n self.modelDigits[digit] = t_img\n\n self.saveModelDigits()\n if not self.showMissingModelDigits():\n self.acceptAviFolderDirectoryWithoutUserIntervention = True\n self.startTimestampReading()\n self.showMsg(f'Training completed.')\n self.showFrame()\n\n\n def addApertureAtPosition(self, x, y):\n x0 = x - self.roi_center\n y0 = y - self.roi_center\n xsize = self.roi_size\n ysize = self.roi_size\n bbox = (x0, y0, xsize, ysize)\n\n # Create an aperture object (box1) and connect it to us (self)\n # Give it a default name. The user can change it later with a context menu\n aperture = MeasurementAperture(f'app{self.apertureId:02d}', bbox, self.roi_max_x, self.roi_max_y)\n\n aperture.order_number = self.apertureId\n\n self.connectAllSlots(aperture)\n\n self.apertureId += 1\n view = self.frameView.getView()\n view.addItem(aperture)\n\n # Make an aperture specific default mask\n self.buildDefaultMask(aperture.default_mask_radius)\n aperture.defaultMask = self.defaultMask[:, :]\n aperture.defaultMaskPixelCount = self.defaultMaskPixelCount\n\n aperture.auto_display = True\n aperture.thresh = self.big_thresh\n self.handleSetGreenSignal(aperture)\n\n return aperture\n\n def connectAllSlots(self, aperture):\n self.makeApertureSignalToSlotConnection(aperture)\n self.makeRecenterSignalToSlotConnection(aperture)\n self.makeSetThreshSignalToSlotConnection(aperture)\n self.makeSetGreenSignalToSlotConnection(aperture)\n self.makeSetYellowSignalToSlotConnection(aperture)\n self.makeDeleteSignalToSlotConnection(aperture)\n self.makeSetThumbnailSourceSignalToSlotConnection(aperture)\n self.makeSetRaDecSignalToSlotConnection(aperture)\n\n def disconnectAllSlots(self, aperture):\n self.disconnectApertureSignalToSlot(aperture)\n self.disconnectRecenterSignalToSlot(aperture)\n self.disconnectSetThreshSignalToSlot(aperture)\n self.disconnectSetGreenSignalToSlot(aperture)\n self.disconnectSetYellowSignalToSlot(aperture)\n self.disconnectDeleteSignalToSlot(aperture)\n self.disconnectSetThumbnailSourceSignalToSlot(aperture)\n self.disconnectSetRaDecSignalToSlot(aperture)\n\n def addGenericAperture(self):\n # self.mousex and self.mousey are continuously updated by mouseMovedInFrameView()\n # self.showMsg(f'placing generic aperture at {self.mousex} {self.mousey}')\n return self.addApertureAtPosition(self.mousex, self.mousey)\n\n def positionApertureAtCentroid(self, aperture, xc, yc):\n bbox = aperture.getBbox()\n x0, y0, xsize, ysize = bbox\n x0_new = int(round(xc - self.roi_center))\n y0_new = int(round(yc - self.roi_center))\n\n # Move the bbox to center the centroid.\n bbox = (x0_new, y0_new, xsize, ysize)\n\n # The setPos() method will intervene, if necessary, to keep the total extent of\n # the aperture inside the image\n aperture.setPos(bbox)\n\n def trackCentroid(self, aperture, xc_roi, yc_roi):\n # Quietly reposition the aperture so that it remains centered on the blob\n bbox = aperture.getBbox()\n x0, y0, xsize, ysize = bbox\n if xc_roi is not None: # The aperture had enough of a blob to calc centroid\n delta_xc = self.roi_center - int(round(xc_roi))\n delta_yc = self.roi_center - int(round(yc_roi))\n\n xpos = x0\n ypos = y0\n w = xsize\n h = ysize\n\n # Move the bbox to center the centroid.\n bbox = (xpos - delta_xc, ypos - delta_yc, w, h)\n\n # The setPos() method will intervene, if necessary, to keep the total extent of\n # the aperture inside the image\n aperture.setPos(bbox)\n\n def initializeTracking(self):\n self.num_yellow_apertures = 0\n for app in self.getApertureList():\n if app.color == 'yellow':\n self.num_yellow_apertures += 1\n\n # self.num_yellow_apertures can only take on values of 0, 1, and 2. This is enforced\n # by self.handleSetYellowSignal()\n\n # If there are two yellow apertures, we need to record the initial geometries\n if self.num_yellow_apertures == 2:\n yellow_count = 0\n for app in self.getApertureList():\n if app.color == 'yellow' and yellow_count == 0: # This our yellow #1\n yellow_count += 1\n if self.use_yellow_mask:\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False, save_yellow_mask=True)\n else:\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n app.dx = 0\n app.dy = 0\n app.theta = 0.0\n\n # Save the coordinates of yellow #1 aperture\n self.yellow_x = xc_world\n self.yellow_y = yc_world\n\n elif app.color == 'yellow' and yellow_count == 1: # This is our yellow #2\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n\n # Get the distance and angle measurements back to yellow #1\n dy = yc_world - self.yellow_y\n dx = xc_world - self.yellow_x\n\n app.dx = dx\n app.dy = dy\n app.theta, _ = calcTheta(dx, dy)\n\n # Set the current field rotation angle\n self.delta_theta = 0.0\n\n for app in self.getApertureList():\n if not app.color == 'yellow':\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n app.xc = xc_world\n app.yc = yc_world\n\n # Get the distance measurements back to yellow #1\n dy = yc_world - self.yellow_y\n dx = xc_world - self.yellow_x\n\n app.dx = dx\n app.dy = dy\n app.theta = None # We don't use this value during tracking\n\n def centerAllApertures(self):\n\n if self.preserve_apertures:\n return\n\n num_yellow_apertures = 0\n delta_xc = 0\n delta_yc = 0\n\n # Look for yellow apertures. If present, we use those to adjust the others.\n # If there is just one, we will use its centroid change to adjust all others.\n # If there is a second yellow, we will use rotations around the first yellow\n # to rotate all others.\n if self.num_yellow_apertures > 0: # this variable is set by self.initializeTracking()\n # We need to find yellow #1 to use for translation tracking\n for app in self.getApertureList():\n if app.color == 'yellow':\n num_yellow_apertures += 1\n if num_yellow_apertures == 1:\n # if self.use_yellow_mask:\n # xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n # self.getApertureStats(app, show_stats=False, save_yellow_mask=True)\n # else:\n # Find out where the centroid of this yellow aperture is located\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n app.dx = 0\n app.dy = 0\n app.theta = 0.0\n\n # Save the current coordinates of the number 1 yellow aperture\n self.yellow_x = xc_world\n self.yellow_y = yc_world\n\n # Compute the needed jog values (will be used/needed if there is but one yellow aperture)\n delta_xc = self.roi_center - int(round(xc_roi))\n delta_yc = self.roi_center - int(round(yc_roi))\n\n # If we're referencing everything off of yellow #1, we need to jog it\n # so that translations are followed by the aperture when we are in field\n # rotation tracking configuration\n if self.num_yellow_apertures == 2:\n jogAperture(app, delta_xc, delta_yc)\n # If we are going to use the mask of this aperture for all the others,\n # now that it's properly positioned, we need to recalculate and save\n # that mask.\n if self.use_yellow_mask:\n self.getApertureStats(app, show_stats=False, save_yellow_mask=True)\n\n elif num_yellow_apertures == 2:\n # We've found a second yellow aperture\n\n # We're referencing everything off of yellow #1, we need to jog yellow #2\n # so that translations are followed and we can get a good angle calculation\n jogAperture(app, delta_xc, delta_yc)\n\n # Note that if we're in 'use yellow mask mode', the mask computed from\n # the 'already jogged into position' yellow 1 will be used here.\n xc_roi, yc_roi, xc_world, yc_world, *_ = \\\n self.getApertureStats(app, show_stats=False)\n\n app.xc = xc_world\n app.yc = yc_world\n\n dx = xc_world - self.yellow_x\n dy = yc_world - self.yellow_y\n\n # Compute new angular position of yellow #2\n new_theta, _ = calcTheta(dx, dy)\n\n # Compute the field rotation that has ocurred since this run started\n self.delta_theta = new_theta - app.theta\n\n self.positionApertureAtCentroid(app, app.xc, app.yc)\n\n # self.showMsg(f'delta_theta={self.delta_theta:7.4f}')\n\n if self.num_yellow_apertures == 2:\n\n cosdt = np.cos(self.delta_theta)\n sindt = np.sin(self.delta_theta)\n for appnew in self.getApertureList():\n if not (appnew.color == 'yellow' or appnew.color == 'white'):\n dx = appnew.dx # These are the original distances to yellow #1 at tracking start\n dy = appnew.dy # These are the original distances to yellow #1 at tracking start\n appnew.xc = dx * cosdt - dy * sindt + self.yellow_x\n appnew.yc = dx * sindt + dy * cosdt + self.yellow_y\n self.positionApertureAtCentroid(appnew, appnew.xc, appnew.yc)\n\n if self.analysisRequested:\n for aperture in self.getApertureList():\n data = self.getApertureStats(aperture, show_stats=False)\n if self.processAsFieldsCheckBox.isChecked():\n aperture.addData(self.field1_data)\n aperture.addData(self.field2_data)\n else:\n aperture.addData(data)\n return\n\n if self.num_yellow_apertures == 1:\n # We simply jog all the apertures (non-white)\n for eachapp in self.getApertureList():\n if not eachapp.color == 'white':\n jogAperture(eachapp, delta_xc, delta_yc)\n\n # Find the first yellow aperture (now jogged into correct position) and compute\n # the mask that will be used by the other apertures\n if self.use_yellow_mask:\n for eachapp in self.getApertureList():\n if eachapp.color == 'yellow':\n self.getApertureStats(eachapp, show_stats=False, save_yellow_mask=True)\n break\n else:\n # There were no yellow apertures, so just center all apertures using the centroid of their mask\n for app in self.getApertureList():\n if not app.color == 'white':\n self.centerAperture(app)\n\n if self.analysisRequested:\n for aperture in self.getApertureList():\n try:\n data = self.getApertureStats(aperture, show_stats=False)\n if self.processAsFieldsCheckBox.isChecked():\n aperture.addData(self.field1_data)\n aperture.addData(self.field2_data)\n else:\n aperture.addData(data)\n except Exception as e:\n self.showMsg(f'while attempting to addData: {repr(e)}')\n\n def centerAperture(self, aperture, show_stats=False):\n # Quietly get the stats for this aperture placement. We are interested in\n # the centroid position (if any) so that we can 'snap to centroid'\n self.one_time_suppress_stats = False\n xc_roi, yc_roi, xc_world, yc_world, *_ = self.getApertureStats(aperture, show_stats=False)\n\n aperture.xc = xc_world\n aperture.yc = yc_world\n\n self.trackCentroid(aperture, xc_roi, yc_roi)\n\n # Display the thumbnails if the caller request show_stats\n self.getApertureStats(aperture, show_stats=show_stats)\n self.frameView.getView().update() # because the bounding box may have shifted\n\n def levelChangedInImageControl(self, pos):\n if self.showImageControlCheckBox.isChecked():\n if self.frame_at_level_set == self.currentFrameSpinBox.value():\n self.levels = self.frameView.ui.histogram.getLevels()\n # self.showMsg(f'Detected level change in histogram widget {self.levels}')\n\n def mouseMovedInFrameView(self, pos):\n\n # inBbox determines whether or not the point x, y is in\n # the bounding box bbox. Used to determine if the cursor is inside an aperture\n def inBbox(x, y, bbox):\n x0, y0, w, h = bbox\n xin = x0 < x < x0 + w\n yin = y0 < y < y0 + h\n return xin and yin\n\n def statusMsg(aperture):\n msg = f' For aperture( {aperture.name} ):'\n if aperture.jogging_enabled:\n msg += f' jogging is ON,'\n else:\n msg += f' jogging is OFF,'\n if aperture.auto_display:\n msg += f' auto_display is ON'\n else:\n msg += f' auto_display is OFF'\n if aperture.thumbnail_source:\n msg += f' (default source for Thumbnail One during run)'\n if aperture.color == 'green':\n msg += f' (responds to threshold spinner)'\n return msg\n\n mousePoint = self.frameView.getView().mapSceneToView(pos)\n x = int(mousePoint.x())\n y = int(mousePoint.y())\n self.mousex = x\n self.mousey = y\n\n if self.viewFieldsCheckBox.isChecked():\n ylim, xlim = self.image.shape\n if 0 <= y < ylim and 0 <= x < xlim:\n self.statusbar.showMessage(f'x={x} y={y} intensity={self.image_fields[y, x]}')\n else:\n self.statusbar.showMessage(f'')\n return\n\n add_on = ''\n if self.wcs_solution_available:\n add_on = 'WCS coords:'\n\n if self.pixelAspectRatio <= 1.0:\n newx = x * self.pixelAspectRatio\n newy = y\n else:\n newx = x\n newy = y / self.pixelAspectRatio\n\n if self.wcs_frame_num == self.currentFrameSpinBox.value():\n pixcrd = np.array([[newx, newy]], dtype='float')\n world = self.wcs.wcs_pix2world(pixcrd, 0)\n thing = SkyCoord(world[0][0] * u.deg, world[0][1] * u.deg, frame='icrs')\n add_on += f' {thing.to_string(style=\"hmsdms\")} {world[0]}'\n else:\n add_on += f' (only available for frame {self.wcs_frame_num})'\n\n if self.image is not None:\n ylim, xlim = self.image.shape\n if 0 <= y < ylim and 0 <= x < xlim:\n # Compose a list of all apertures at the current cursor position\n # for output to the status bar.\n appsStacked = \"\"\n\n for app in self.getApertureList():\n if inBbox(x, y, app.getBbox()):\n appsStacked += f'\"{app.name}\" '\n aperture_to_point_at = app\n\n if appsStacked:\n # set pointed_at to last aperture in the list (should be\n # the most recent addition) If it was None, we have entered\n # for the first time and should show stats\n if self.pointed_at_aperture is None:\n self.pointed_at_aperture = aperture_to_point_at\n if self.analysisPaused:\n self.getApertureStats(self.pointed_at_aperture)\n else:\n # Cursor is not in any aperture so reset pointed_at_aperture\n self.pointed_at_aperture = None\n\n if appsStacked: # The cursor was one or more apertures\n # status = statusMsg(app)\n # self.statusbar.showMessage(f'x={x} y={y} intensity={self.image[y,x]} {status} {add_on}')\n self.statusbar.showMessage(\n f'x={x} y={y} intensity={self.image[y,x]} Apertures under cursor: {appsStacked} {add_on}')\n else:\n self.pointed_at_aperture = None\n self.statusbar.showMessage(f'x={x} y={y} intensity={self.image[y,x]} {add_on}')\n\n else:\n self.statusbar.showMessage(f'')\n\n def mouseMovedInThumbOne(self, pos):\n mousePoint = self.thumbOneView.getView().mapSceneToView(pos)\n x = int(mousePoint.x())\n y = int(mousePoint.y())\n if self.thumbOneImage is not None:\n ylim, xlim = self.thumbOneImage.shape\n if 0 <= y < ylim and 0 <= x < xlim:\n self.statusbar.showMessage(f'x={x} y={y} intensity={self.thumbOneImage[y,x]}')\n else:\n self.statusbar.showMessage(f'x={x} y={y}')\n\n def mouseMovedInThumbTwo(self, pos):\n mousePoint = self.thumbTwoView.getView().mapSceneToView(pos)\n x = int(mousePoint.x())\n y = int(mousePoint.y())\n if self.thumbTwoImage is not None:\n ylim, xlim = self.thumbTwoImage.shape\n if 0 <= y < ylim and 0 <= x < xlim:\n self.statusbar.showMessage(f'x={x} y={y} intensity={self.thumbTwoImage[y, x]}')\n else:\n self.statusbar.showMessage(f'x={x} y={y}')\n\n def getApertureStats(self, aperture, show_stats=True, save_yellow_mask=False):\n # This routine is dual purpose. When self.show_stats is True, there is output to\n # the information text box, and to the the two thumbnail ImageViews.\n # But sometime we use this routine just to get the measurements that it returns.\n\n if self.one_time_suppress_stats:\n self.one_time_suppress_stats = False\n return None\n\n # Grab the properties that we need from the aperture object\n bbox = aperture.getBbox()\n x0, y0, nx, ny = bbox\n name = aperture.name\n\n # thumbnail is the portion of the main image that is covered by the aperture bounding box\n thumbnail = self.image[y0:y0+ny, x0:x0+nx]\n mean, std, sorted_data, *_ = robustMeanStd(thumbnail, outlier_fraction=.5)\n\n maxpx = sorted_data[-1]\n\n # We computed the initial aperture.thresh as an offset from the background value present\n # in the frame used for the initial threshold determination. Now we add the current\n # value of the background so that we can respond to a general change in background dynamically.\n background = int(round(mean))\n threshold = aperture.thresh + background\n\n default_mask_used = False\n\n if aperture.color == 'yellow':\n max_area, mask, t_mask, centroid, cvxhull, nblobs, extent = \\\n get_mask(thumbnail, ksize=self.gaussian_blur, cut=threshold, outlier_fraction=0.5,\n apply_centroid_distance_constraint=False, max_centroid_distance=self.allowed_centroid_delta)\n elif aperture.color == 'white':\n max_area = self.roi_size * self.roi_size\n centroid = (self.roi_center, self.roi_center)\n cvxhull = max_area\n mask = np.ones((self.roi_size, self.roi_size), dtype='int')\n for i in range(self.roi_size):\n # Create a black border one pixel wide around the edges\n mask[0, i] = 0\n mask[i, 0] = 0\n mask[self.roi_size-1, i] = 0\n mask[i, self.roi_size-1] = 0\n max_area = np.sum(mask)\n else:\n # This handles 'red' and 'green' apertures\n max_area, mask, t_mask, centroid, cvxhull, nblobs, extent = \\\n get_mask(thumbnail, ksize=self.gaussian_blur, cut=threshold, outlier_fraction=0.5,\n apply_centroid_distance_constraint=True, max_centroid_distance=self.allowed_centroid_delta)\n\n if save_yellow_mask:\n self.yellow_mask = mask.copy()\n\n comment = \"\"\n\n if max_area == 0:\n default_mask_used = True\n mask = aperture.defaultMask\n max_area = aperture.defaultMaskPixelCount\n\n centroid = (self.roi_center, self.roi_center)\n comment = f'default mask used'\n\n if show_stats:\n if self.pointed_at_aperture is not None:\n if aperture == self.pointed_at_aperture:\n self.thumbnail_one_aperture_name = aperture.name\n self.thumbOneImage = thumbnail\n self.thumbOneView.setImage(thumbnail)\n self.thumbTwoView.setImage(mask)\n else:\n priority_aperture_present = False\n for app in self.getApertureList():\n if app.thumbnail_source:\n priority_aperture_present = True\n break\n\n if priority_aperture_present:\n if aperture.thumbnail_source:\n self.thumbnail_one_aperture_name = aperture.name\n self.thumbOneImage = thumbnail\n self.thumbOneView.setImage(thumbnail)\n self.thumbTwoView.setImage(mask)\n else:\n self.thumbnail_one_aperture_name = aperture.name\n self.thumbOneImage = thumbnail\n self.thumbOneView.setImage(thumbnail)\n self.thumbTwoView.setImage(mask)\n\n self.hair1.setPos((0,self.roi_size))\n self.hair2.setPos((0,0))\n\n if self.levels:\n self.thumbOneView.setLevels(min=self.levels[0], max=self.levels[1])\n\n # Show the mask itself\n if self.use_yellow_mask:\n self.thumbTwoImage = self.yellow_mask\n else:\n self.thumbTwoImage = mask\n\n if self.thumbTwoImage is not None:\n self.thumbTwoView.setImage(self.thumbTwoImage)\n\n if self.use_yellow_mask and self.yellow_mask is not None:\n default_mask_used = False\n appsum = np.sum(self.yellow_mask * thumbnail)\n max_area = int(np.sum(self.yellow_mask))\n signal = appsum - int(round(max_area * mean))\n else:\n try:\n appsum = np.sum(mask * thumbnail)\n if aperture.color == 'white':\n signal = appsum\n else:\n signal = appsum - int(round(max_area * mean))\n except Exception as e:\n self.showMsg(f'in showApertureStats: {e}')\n appsum = 0\n signal = 0\n\n if not centroid == (None, None):\n xc_roi = centroid[1]\n yc_roi = centroid[0]\n xc_world = xc_roi + x0 # x0 and y0 are ints that give the corner position of the aperture\n yc_world = yc_roi + y0\n else:\n xc_roi = yc_roi = xc_world = yc_world = None\n\n frame_num = float(self.currentFrameSpinBox.text())\n\n if default_mask_used:\n # A negative value for mask pixel count indicates that a default mask was used in the measurement\n # This will appear in the csv file. In our plots, will use the negative value to\n # add visual annotation that a default mask was employed\n max_area = -max_area\n\n if show_stats:\n minpx = sorted_data[0]\n maxpx = sorted_data[-1]\n xpos = int(round(xc_world))\n ypos = int(round(yc_world))\n\n self.showMsg(f'{name}:{comment} frame:{frame_num:0.1f}', blankLine=False)\n self.showMsg(f'signal appsum bkavg bkstd mskth mskpx cvxhull xpos ypos minpx maxpx',\n blankLine=False)\n\n if xpos is not None:\n line = '%6d%7d%9.3f%7.2f%7d%7d%9d%6d%6d%6d%6d' % \\\n (signal, appsum, mean, std, threshold, max_area, cvxhull, xpos, ypos, minpx, maxpx)\n else:\n line = '%6d%7d%9.3f%7.2f%7d%7d%9d%6s%6s%6d%6d' % \\\n (signal, appsum, mean, std, threshold, max_area, cvxhull, ' NA', ' NA', minpx, maxpx)\n self.showMsg(line)\n\n # xc_roi and yc_roi are used by centerAperture() to recenter the aperture\n # The remaining outputs are used in writing the lightcurve information\n # !!! ANY CHANGE TO THE TYPE OR ORDERING OF THIS OUTPUT MUST BE REFLECTED IN writeCsvFile() !!!\n if self.processAsFieldsCheckBox.isChecked():\n top_mask = mask[0::2,:]\n top_mask_pixel_count = np.sum(top_mask)\n top_thumbnail = thumbnail[0::2,:]\n top_appsum = np.sum(top_mask * top_thumbnail)\n top_signal = top_appsum - int(round(top_mask_pixel_count * mean))\n if default_mask_used:\n top_mask_pixel_count = -top_mask_pixel_count\n\n bottom_mask = mask[1::2,:]\n bottom_mask_pixel_count = np.sum(bottom_mask)\n bottom_thumbnail = thumbnail[1::2,:]\n bottom_appsum = np.sum(bottom_mask * bottom_thumbnail)\n bottom_signal = bottom_appsum - int(round(bottom_mask_pixel_count * mean))\n if default_mask_used:\n bottom_mask_pixel_count = -bottom_mask_pixel_count\n\n if aperture.color == 'white':\n top_signal = top_appsum\n bottom_signal = bottom_appsum\n\n if self.topFieldFirstRadioButton.isChecked():\n timestamp = self.upperTimestamp\n self.field1_data = (xc_roi, yc_roi, xc_world, yc_world,\n top_signal, top_appsum, mean, top_mask_pixel_count,\n frame_num, cvxhull, maxpx, std, timestamp)\n timestamp = self.lowerTimestamp\n self.field2_data = (xc_roi, yc_roi, xc_world, yc_world,\n bottom_signal, bottom_appsum, mean, bottom_mask_pixel_count,\n frame_num + 0.5, cvxhull, maxpx, std, timestamp)\n else:\n timestamp = self.lowerTimestamp\n self.field1_data = (xc_roi, yc_roi, xc_world, yc_world,\n bottom_signal, bottom_appsum, mean, bottom_mask_pixel_count,\n frame_num, cvxhull, maxpx, std, timestamp)\n timestamp = self.upperTimestamp\n self.field2_data = (xc_roi, yc_roi, xc_world, yc_world,\n top_signal, top_appsum, mean, top_mask_pixel_count,\n frame_num + 0.5, cvxhull, maxpx, std, timestamp)\n\n if not self.avi_in_use:\n timestamp = self.fits_timestamp\n else:\n if self.topFieldFirstRadioButton.isChecked():\n if self.upperTimestamp:\n timestamp = self.upperTimestamp\n else:\n timestamp = ''\n else:\n if self.lowerTimestamp:\n timestamp = self.lowerTimestamp\n else:\n timestamp = ''\n\n return (xc_roi, yc_roi, xc_world, yc_world, signal,\n appsum, mean, max_area, frame_num, cvxhull, maxpx, std, timestamp)\n\n def clearApertures(self):\n # Remove measurement apertures (if any)\n apertures = self.getApertureList()\n if apertures:\n for aperture in apertures:\n self.removeAperture(aperture)\n\n def clearOcrBoxes(self):\n # Remove OcrBoxes (if any)\n ocrboxes = self.getOcrBoxList()\n if ocrboxes:\n for ocrbox in ocrboxes:\n self.removeOcrBox(ocrbox)\n self.frameView.getView().update()\n\n def setAllOcrBoxJogging(self, enable, position):\n ocrboxes = self.getOcrBoxList()\n if ocrboxes:\n for ocrbox in ocrboxes:\n if ocrbox.position == position:\n ocrbox.joggable = enable\n\n def readFitsFile(self):\n\n # If a bitmap has just been loaded, it is assumed that the user is employing\n # a RegiStax star locator to place his apertures. It is critical to maintaing the correct\n # offsets between the apertures that at least one of them is yellow, otherwise\n # the positioning will be lost when the first fits file loads and the apertures try to\n # 'snap' to better positions. Here we remind the user to do so.\n if self.preserve_apertures:\n ok = self.yellowAperturePresent()\n if not ok:\n # self.showMsg(f'No yellow aperture(s)!!! Need to add query to confirm')\n return\n\n options = QFileDialog.Options()\n options |= QFileDialog.ShowDirsOnly\n # options |= QFileDialog.DontUseNativeDialog\n dir_path = QFileDialog.getExistingDirectory(\n self,\n \"Select directory\",\n self.settings.value('fitsdir', \"./\"), # starting directory,\n options=options\n )\n\n QtGui.QGuiApplication.processEvents()\n\n if dir_path:\n self.saveStateNeeded = True\n self.avi_wcs_folder_in_use = False\n self.fits_folder_in_use = True\n self.clearTextBox()\n self.saveTargetLocButton.setEnabled(True)\n self.loadCustomProfilesButton.setEnabled(False)\n\n\n self.createAVIWCSfolderButton.setEnabled(False)\n self.vtiSelectComboBox.setEnabled(False)\n\n self.levels = []\n # remove the star rectangles (possibly) left from previous file\n self.clearApertures()\n self.filename = dir_path\n self.apertureId = 0\n self.num_yellow_apertures = 0\n self.avi_in_use = False\n self.showMsg(f'Opened FITS folder: {dir_path}', blankLine=False)\n self.settings.setValue('fitsdir', dir_path) # Make dir 'sticky'\"\n self.folder_dir = dir_path\n self.fits_filenames = sorted(glob.glob(dir_path + '/*.fits'))\n\n if os.path.exists(self.folder_dir + '/pixel-dimensions.p'):\n self.readPixelDimensions()\n else:\n self.pixelAspectRatio = 1.0\n self.pixelWidthEdit.setText('1.00')\n self.pixelHeightEdit.setText('1.00')\n\n self.fourcc = ''\n\n self.disableControlsWhenNoData()\n self.enableControlsForFitsData()\n\n # self.showMsg('Changing navigation buttons to 25 frames')\n self.frameJumpSmall = 25\n self.frameJumpBig = 200\n self.changeNavButtonTitles()\n\n frame_count = len(self.fits_filenames)\n self.currentFrameSpinBox.setMaximum(frame_count - 1)\n self.currentFrameSpinBox.setValue(0)\n self.stopAtFrameSpinBox.setMaximum(frame_count - 1)\n self.stopAtFrameSpinBox.setValue(frame_count - 1)\n\n _, file_id = os.path.split(self.fits_filenames[0])\n self.showMsg(f'... and found {frame_count} files of the form: {file_id}')\n # This will get our image display initialized with default pan/zoom state.\n # It will also capture to the date from the first timestamp. We add that as a comment\n # to the csv file\n self.initialFrame = True\n self.currentOcrBox = None\n self.showFrame()\n\n self.thumbOneView.clear()\n self.thumbTwoView.clear()\n\n self.processTargetAperturePlacementFiles()\n\n # Check for the presence of a 'saved aperture group' and enable the Restore group\n # button accordingly\n file1 = self.folder_dir + '/markedApertures.p'\n file2 = self.folder_dir + '/markedFrameNumber.p'\n\n if os.path.exists(file1) and os.path.exists(file2):\n self.restoreApertureState.setEnabled(True)\n\n def showMsgDialog(self, msg):\n msg_box = QMessageBox()\n msg_box.setText(msg)\n msg_box.exec()\n\n def showMsgPopup(self, msg):\n self.helperThing.textEdit.clear()\n self.helperThing.textEdit.setText(msg)\n self.helperThing.raise_()\n self.helperThing.show()\n # msg_box = QMessageBox()\n # msg_box.setText(msg)\n # msg_box.exec()\n\n\n def openFitsImageFile(self, fpath):\n self.image = pyfits.getdata(fpath).astype('int16', casting='unsafe')\n self.frameView.setImage(self.image)\n msg_box = QMessageBox()\n msg_box.setText(f'Always use a single static (no-snap) aperture to designate the target!'\n f'\\n\\nThis technique forces the aperture to use a default mask (by setting '\n f'a very high mskth) which '\n f'means that this aperture will not move (snap) when you switch back '\n f'to the avi.'\n f'\\n\\nThe selected location will automatically be saved when a '\n f'frame change is made that returns the view to the avi.'\n f'\\n\\nThe avi will be automatically positioned to the frame '\n f'that was used as the reference frame for the enhanced image stack.')\n msg_box.exec()\n\n def readFinderImage(self):\n\n if self.avi_wcs_folder_in_use or self.fits_folder_in_use:\n # Look for enhanced-image.fit and if present, open it and return\n # otherwise let the user find a .bmp file whereever.\n fullpath = self.folder_dir + r'/enhanced-image.fit'\n if os.path.isfile(fullpath):\n self.showMsg(f'Found an enhanced image file')\n self.openFitsImageFile(fullpath)\n self.record_target_aperture = True\n return\n\n options = QFileDialog.Options()\n # options |= QFileDialog.DontUseNativeDialog\n\n self.filename, _ = QFileDialog.getOpenFileName(\n self, # parent\n \"Select bmp image\", # title for dialog\n self.settings.value('bmpdir', \"./\"), # starting directory\n \"bmp images (*.bmp);; all files (*.*)\",\n options=options\n )\n\n QtGui.QGuiApplication.processEvents()\n\n if self.filename:\n self.createAVIWCSfolderButton.setEnabled(False)\n self.clearTextBox()\n self.preserve_apertures = True\n # remove the apertures (possibly) left from previous file\n self.clearApertures()\n self.apertureId = 0\n self.num_yellow_apertures = 0\n self.levels = []\n\n dirpath, _ = os.path.split(self.filename)\n self.settings.setValue('bmpdir', dirpath) # Make dir 'sticky'\"\n self.showMsg(f'Opened: {self.filename}')\n img = cv2.imread(self.filename)\n self.image = img[:, :, 0]\n\n self.frameView.setImage(self.image)\n height, width = self.image.shape\n\n # The following variables are used by MeasurementAperture to limit\n # aperture placement so that it stays within the image at all times\n self.roi_max_x = width - self.roi_size\n self.roi_max_y = height - self.roi_size\n\n def getFrame(self, fr_num):\n\n trace = False\n\n if self.cap is None or not self.cap.isOpened():\n return False, None\n\n next_frame = self.cap.get(cv2.CAP_PROP_POS_FRAMES)\n if trace:\n self.showMsg(f'requested frame: {fr_num} next in line for cap.read(): {next_frame}')\n\n if fr_num == next_frame - 1:\n # User is asking for the frame that is currently being displayed\n return True, self.image\n\n if fr_num == next_frame:\n if trace:\n self.showMsg('frame requested is next to be read by cap.read()')\n success, frame = self.cap.read()\n if not success:\n if trace:\n self.showMsg('read() failed')\n return success, frame\n\n if fr_num > next_frame:\n frames_to_read = fr_num - next_frame + 1\n if trace:\n self.showMsg(f'We will read forward {frames_to_read} frames')\n while frames_to_read > 0:\n frames_to_read -= 1\n success, frame = self.cap.read()\n return success, frame\n\n if fr_num < next_frame:\n if trace:\n self.showMsg(f'Closing and reopening avi_file: {self.filename}')\n self.cap.release()\n self.cap = cv2.VideoCapture(self.filename, cv2.CAP_FFMPEG)\n next_frame = self.cap.get(cv2.CAP_PROP_POS_FRAMES)\n if trace:\n self.showMsg(f'requested frame: {fr_num} next in line for cap.read(): {next_frame}')\n frames_to_read = fr_num - next_frame + 1\n if trace:\n self.showMsg(f'We will read forward {frames_to_read} frames')\n\n while frames_to_read > 0:\n frames_to_read -= 1\n success, frame = self.cap.read()\n return success, frame\n\n return False, None\n\n def readAviFile(self):\n\n options = QFileDialog.Options()\n # options |= QFileDialog.DontUseNativeDialog\n\n self.filename, _ = QFileDialog.getOpenFileName(\n self, # parent\n \"Select avi file\", # title for dialog\n self.settings.value('avidir', \"./\"), # starting directory\n \"avi files (*.avi);; all files (*.*)\",\n options=options\n )\n\n QtGui.QGuiApplication.processEvents()\n\n if self.filename:\n self.saveStateNeeded = True\n self.wcs_solution_available = False\n self.wcs_frame_num = None\n self.avi_wcs_folder_in_use = False\n self.fits_folder_in_use = False\n self.saveTargetLocButton.setEnabled(False)\n self.loadCustomProfilesButton.setEnabled(False)\n\n self.pixelAspectRatio = None\n\n self.createAVIWCSfolderButton.setEnabled(True)\n self.vtiSelectComboBox.setEnabled(False)\n\n dirpath, _ = os.path.split(self.filename)\n self.folder_dir = dirpath\n self.settings.setValue('avidir', dirpath) # Make dir 'sticky'\"\n self.clearTextBox()\n\n # remove the star rectangles (possibly) left from previous file\n if not self.preserve_apertures:\n self.clearApertures()\n\n self.apertureId = 0\n self.num_yellow_apertures = 0\n self.levels = []\n\n self.showMsg(f'Opened: {self.filename}')\n if self.cap:\n self.cap.release()\n self.cap = cv2.VideoCapture(self.filename, cv2.CAP_FFMPEG)\n if not self.cap.isOpened():\n self.showMsg(f' {self.filename} could not be opened!')\n self.fourcc = ''\n else:\n self.avi_in_use = True\n self.savedApertures = None\n self.enableControlsForAviData()\n self.saveApertureState.setEnabled(False)\n # Let's get the FOURCC code\n fourcc = int(self.cap.get(cv2.CAP_PROP_FOURCC))\n fourcc_str = f'{fourcc & 0xff:c}{fourcc >> 8 & 0xff:c}{fourcc >> 16 & 0xff:c}{fourcc >> 24 & 0xff:c}'\n self.fourcc = fourcc_str\n self.showMsg(f'FOURCC codec ID: {fourcc_str}')\n\n fps = self.cap.get(cv2.CAP_PROP_FPS)\n if fps > 29.0:\n # self.showMsg('Changing navigation buttons to 30 frames')\n self.frameJumpSmall = 30\n self.frameJumpBig = 300\n self.changeNavButtonTitles()\n else:\n # self.showMsg('Changing navigation buttons to 25 frames')\n self.frameJumpSmall = 25\n self.frameJumpBig = 250\n self.changeNavButtonTitles()\n\n\n self.showMsg(f'frames per second:{fps:0.6f}')\n\n frame_count = int(self.cap.get(cv2.CAP_PROP_FRAME_COUNT))\n self.showMsg(f'There are {frame_count} frames in the file.')\n\n # We need to do this before self.showFrame() is called either directly\n # or indirectly (when self.currentFrameSpinBox is changed, it invokes\n # self.showFrame())\n\n self.currentOcrBox = None\n\n self.vtiSelectComboBox.setCurrentIndex(0)\n self.vtiSelected()\n\n self.currentFrameSpinBox.setMaximum(frame_count-1)\n self.currentFrameSpinBox.setValue(0)\n self.stopAtFrameSpinBox.setMaximum(frame_count - 1)\n self.stopAtFrameSpinBox.setValue(frame_count - 1)\n\n # This will get our image display initialized with default pan/zoom state\n self.initialFrame = True\n self.showFrame()\n self.clearOcrBoxes()\n\n self.thumbOneView.clear()\n self.thumbTwoView.clear()\n\n def setTimestampFormatter(self):\n self.kiwiInUse = False\n if self.formatterCode is None:\n self.showMsg(f'Timestamp formatter code was missing.')\n self.timestampFormatter = None\n elif self.formatterCode == 'iota':\n self.timestampFormatter = format_iota_timestamp\n elif self.formatterCode == 'boxsprite':\n self.timestampFormatter = format_boxsprite3_timestamp\n elif self.formatterCode == 'kiwi-left' or self.formatterCode == 'kiwi-right':\n self.timestampFormatter = format_kiwi_timestamp\n self.kiwiInUse = True\n else:\n self.showMsg(f'Unknown timestamp formatter code: {self.formatterCode}. Defaulting to Iota')\n self.timestampFormatter = format_iota_timestamp\n\n def readFormatTypeFile(self):\n f_path = os.path.join(self.folder_dir, 'formatter.txt')\n if not os.path.exists(f_path):\n return None\n with open(f_path, 'r') as f:\n code = f.readline()\n return code\n\n def selectAviFolder(self):\n\n if not self.acceptAviFolderDirectoryWithoutUserIntervention:\n options = QFileDialog.Options()\n options |= QFileDialog.ShowDirsOnly\n # options |= QFileDialog.DontUseNativeDialog\n\n dir_path = QFileDialog.getExistingDirectory(\n self, # parent\n \"Select directory\", # title for dialog\n self.settings.value('avidir', \"./\"), # starting directory\n options=options\n )\n\n QtGui.QGuiApplication.processEvents()\n\n if dir_path:\n self.showMsg(f'dir_path= {dir_path}')\n else:\n self.showMsg(f'User cancelled')\n return\n else:\n dir_path = self.settings.value('avidir', \"./\")\n self.acceptAviFolderDirectoryWithoutUserIntervention = False\n\n if dir_path:\n\n self.saveStateNeeded = True\n self.upper_left_count = 0 # When Kiwi used: accumulate count ot times t2 was at left in upper field\n self.upper_right_count = 0 # When Kiwi used: accumulate count ot times t2 was at the right in upper field\n\n self.lower_left_count = 0 # When Kiwi used: accumulate count ot times t2 was at left in lower field\n self.lower_right_count = 0 # When Kiwi used: accumulate count ot times t2 was at the right in lower field\n\n self.wcs_solution_available = False\n self.wcs_frame_num = None\n self.avi_wcs_folder_in_use = True\n self.fits_folder_in_use = False\n self.saveTargetLocButton.setEnabled(True)\n self.loadCustomProfilesButton.setEnabled(True)\n\n self.createAVIWCSfolderButton.setEnabled(False)\n self.vtiSelectComboBox.setEnabled(False)\n\n self.settings.setValue('avidir', dir_path) # Make dir 'sticky'\"\n self.folder_dir = dir_path\n\n self.clearTextBox()\n self.readPixelDimensions()\n\n self.disableControlsWhenNoData()\n try:\n self.frameView.clear()\n QtGui.QGuiApplication.processEvents()\n if self.cap:\n self.cap.release()\n except Exception as e:\n self.showMsg(f'While trying to clear FrameView got following exception:',\n blankLine=False)\n self.showMsg(f'{e}')\n\n # We need to know what OS we're running under in order to look for\n # either 'aliases' (MacOs) or 'shortcuts' (Windows) to the avi file\n\n # if os.name == 'posix':\n # # self.showMsg(f'os: MacOS')\n # macOS = True\n # windows = False\n # else:\n # macOS = False\n # windows = True\n # # self.showMsg(f'os: Windows')\n\n # use `sys.platform` to distinguish macOS from Linux\n if sys.platform == 'linux':\n linux, macOS, windows = True, False, False\n elif sys.platform == 'darwin':\n linux, macOS, windows = False, True, False\n else:\n linux, macOS, windows = False, False, True\n\n\n # Find a .avi file in the given directory. Enforce that there be only one.\n # Note: this picks up alias (mac) and shortcut (Windows) files too.\n avi_filenames = glob.glob(dir_path + '/*.avi*')\n\n avi_location = ''\n num_avifiles = len(avi_filenames)\n\n if num_avifiles == 1: # one avi (or alias or shortcut) is in the folder)\n avi_location = avi_filenames[0]\n if macOS:\n avi_location = alias_lnk_resolver.resolve_osx_alias(avi_location)\n\n elif linux:\n avi_location = os.readlink(avi_location)\n else:\n target = winshell.shortcut(avi_location)\n avi_location = target.path\n # Save as instance variable for use in stacker\n self.avi_location = avi_location\n self.filename = avi_location\n elif num_avifiles > 1:\n self.showMsg(f'{num_avifiles} avi files were found. Only one is allowed in an AVI-WCS folder')\n return\n else:\n self.showMsg(f'No avi files were found in that folder.')\n return\n\n # remove the apertures (possibly) left from previous file\n if not self.preserve_apertures:\n self.clearApertures()\n\n self.apertureId = 0\n self.num_yellow_apertures = 0\n self.levels = []\n\n self.showMsg(f'Opened: {avi_location}')\n if self.cap:\n self.cap.release()\n self.cap = cv2.VideoCapture(avi_location)\n if not self.cap.isOpened():\n self.showMsg(f' {avi_location} could not be opened!')\n else:\n self.timestampReadingEnabled = False\n self.vtiSelectComboBox.setCurrentIndex(0)\n self.avi_in_use = True\n self.savedApertures = None\n self.enableControlsForAviData()\n # Let's get the FOURCC code\n fourcc = int(self.cap.get(cv2.CAP_PROP_FOURCC))\n fourcc_str = f'{fourcc & 0xff:c}{fourcc >> 8 & 0xff:c}{fourcc >> 16 & 0xff:c}{fourcc >> 24 & 0xff:c}'\n self.showMsg(f'FOURCC codec ID: {fourcc_str}')\n self.showMsg(f'frames per second:{self.cap.get(cv2.CAP_PROP_FPS):0.6f}')\n\n frame_count = int(self.cap.get(cv2.CAP_PROP_FRAME_COUNT))\n self.showMsg(f'There are {frame_count} frames in the file.')\n\n fps = self.cap.get(cv2.CAP_PROP_FPS)\n if fps > 29.0:\n self.frameJumpSmall = 30\n self.frameJumpBig = 300\n self.changeNavButtonTitles()\n else:\n self.frameJumpSmall = 25\n self.frameJumpBig = 250\n self.changeNavButtonTitles()\n\n self.currentFrameSpinBox.setMaximum(frame_count-1)\n self.currentFrameSpinBox.setValue(0)\n self.stopAtFrameSpinBox.setMaximum(frame_count - 1)\n self.stopAtFrameSpinBox.setValue(frame_count - 1)\n\n # This will get our image display initialized with default pan/zoom state\n self.initialFrame = True\n self.showFrame()\n\n # Initialize ocr related directories\n self.ocrDigitsDir = self.folder_dir\n self.ocrBoxesDir = self.folder_dir\n self.currentOcrBox = None\n self.clearOcrBoxes() # From any previous ocr setup\n\n self.modelDigitsFilename = 'custom-digits.p'\n self.ocrboxBasePath = 'custom-boxes'\n\n self.processTargetAperturePlacementFiles()\n\n # Check for the presence of a 'saved aperture group' and enable the Restore group\n # button accordingly\n file1 = self.folder_dir + '/markedApertures.p'\n file2 = self.folder_dir + '/markedFrameNumber.p'\n\n if os.path.exists(file1) and os.path.exists(file2):\n self.restoreApertureState.setEnabled(True)\n\n self.startTimestampReading()\n self.showFrame() # So that we get the first frame timestamp (if possible)\n\n self.thumbOneView.clear()\n self.thumbTwoView.clear()\n\n def startTimestampReading(self):\n # This is how we starup timestamp extraction.\n\n # We assume that if a valid timestamp formatter selection code is\n # present, then timestamp reading should be attempted\n formatter_code = self.readFormatTypeFile()\n self.formatterCode = formatter_code\n processTimestampProfile = not self.formatterCode is None\n\n if processTimestampProfile:\n self.loadPickledOcrBoxes() # if any\n self.loadModelDigits() # if any\n self.detectFieldTimeOrder = True\n # Reset the Kiwi special counters that record where t2 has been found\n self.upper_left_count = 0\n self.upper_right_count = 0\n self.lower_left_count = 0\n self.lower_right_count = 0\n self.setTimestampFormatter()\n self.currentLowerBoxPos = 'left'\n self.currentUpperBoxPos = 'left'\n if self.formatterCode == 'kiwi-right':\n self.currentLowerBoxPos = 'right'\n self.currentUpperBoxPos = 'right'\n self.viewFieldsCheckBox.setChecked(True)\n self.placeOcrBoxesOnImage()\n self.currentFrameSpinBox.setValue(1) # This triggers a self.showFrame() call\n self.timestampReadingEnabled = not self.showMissingModelDigits()\n self.vtiSelectComboBox.setEnabled(not self.timestampReadingEnabled)\n else:\n self.vtiSelectComboBox.setEnabled(True)\n\n\n\n def getFrameNumberFromFile(self, filename):\n fullpath = self.folder_dir + r'/' + filename\n if not os.path.isfile(fullpath):\n return False, 0\n try:\n with open(fullpath, 'r') as f:\n text = f.read()\n frame_num = int(text)\n return True, frame_num\n except ValueError:\n return True, None\n\n def readPixelDimensions(self):\n # Check for presence of pixel dimensions file\n matching_name = glob.glob(self.folder_dir + '/pixel-dimensions.p')\n if matching_name:\n pixHstr, pixWstr = pickle.load(open(matching_name[0], \"rb\"))\n self.pixelHeightEdit.setText(pixHstr)\n self.pixelWidthEdit.setText(pixWstr)\n self.pixelAspectRatio = float(pixWstr) / float(pixHstr)\n self.showMsg(f'Found pixel dimensions of {pixHstr}(H) and {pixWstr}(W)')\n else:\n self.pixelAspectRatio = None\n self.pixelHeightEdit.setText('')\n self.pixelWidthEdit.setText('')\n self.showMsg(f'No pixel dimensions were found.')\n\n def processTargetAperturePlacementFiles(self):\n # If enhanced image target positioning files are found, it is the priority\n # method for automatically placing the target aperture. It came from stacking\n # frames from the video to get an enhanced video from which the user selected\n # the target star from a star chart. It is given first priority because it\n # is so directly connected to the pobservation data.\n frame_file = 'enhanced-image-frame-num.txt'\n file_found, frame_num = self.getFrameNumberFromFile(frame_file)\n if file_found:\n if frame_num is None:\n self.showMsg(f'Content error in: {frame_file}')\n return\n else:\n got_frame_number = True\n self.currentFrameSpinBox.setValue(frame_num)\n\n matching_name = glob.glob(self.folder_dir + '/target-aperture-xy.txt')\n if matching_name:\n # We read the file and place the aperture.\n with open(matching_name[0], 'r') as f:\n xy_str = f.readline()\n parts = xy_str.split()\n try:\n x = int(parts[0])\n y = int(parts[1])\n self.showMsg(f'Target aperture was placed from \"enhanced image\" data.')\n aperture = self.addApertureAtPosition(x, y)\n aperture.setRed()\n aperture.name = 'target'\n aperture.thresh = self.big_thresh\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(self.big_thresh) # Causes call to self.changeThreshold()\n return\n except ValueError:\n self.showMsg(f'Invalid target-aperture-xy.txt contents: {xy_str}')\n\n # Check for presence of pixel dimensions file\n matching_name = glob.glob(self.folder_dir + '/pixel-dimensions.p')\n if matching_name:\n pixHstr, pixWstr = pickle.load(open(matching_name[0], \"rb\"))\n self.pixelHeightEdit.setText(pixHstr)\n self.pixelWidthEdit.setText(pixWstr)\n\n # Check for presence of target-location.txt This file is needed for both\n # the manual WCS placement and the nova.astrometry.net placement\n matching_name = sorted(glob.glob(self.folder_dir + '/target-location.txt'))\n\n got_star_position = False\n if not matching_name:\n self.showMsg(f'No target star location found in the folder.')\n return\n\n # ss = self.getStarPositionString()\n # if ss:\n # self.showMsg(f'star position string provided: \"{ss}\"')\n #\n # try:\n # _ = SkyCoord(ss, frame='icrs')\n # except Exception as e:\n # self.showMsg(f'star location string is invalid: {e}')\n # return\n #\n # with open(self.folder_dir + r'/target-location.txt', 'w') as f:\n # f.writelines(ss)\n # got_star_position = True\n # else:\n # self.showMsg(f'No star position was provided.')\n # # Both the manual WCS and the nova.astrometry.net WCS aperture placements\n # # depend on this file, so we can exit immediately\n # return\n else:\n with open(self.folder_dir + r'/target-location.txt', 'r') as f:\n ss = f.read()\n self.showMsg(f'target star position is: {ss}')\n got_star_position = True\n\n got_fits_wcs_calibration = False\n got_manual_wcs_calibration = False\n\n # Check for presence of wcs*.fits file\n wcs_fits = sorted(glob.glob(self.folder_dir + '/wcs*.fit'))\n\n if wcs_fits:\n self.showMsg(f'nova.astrometry.net WCS calibration file found in the folder.')\n got_fits_wcs_calibration = True\n\n # Check for presence of wcs-frame-num.txt file\n frame_file = 'wcs-frame-num.txt'\n file_found, frame_num_of_wcs = self.getFrameNumberFromFile(frame_file)\n got_frame_number = False\n if not file_found:\n self.showMsg(f'No WCS calibration frame number found in the folder.')\n got_fits_wcs_calibration = False\n else:\n if frame_num_of_wcs is None:\n self.showMsg(f'Content error in: {frame_file}')\n return\n else:\n got_frame_number = True\n got_fits_wcs_calibration = True\n self.currentFrameSpinBox.setValue(frame_num_of_wcs)\n\n if not got_fits_wcs_calibration: # try for manual WCS placement\n frame_file = 'manual-wcs-frame-num.txt'\n file_found, frame_num_of_wcs = self.getFrameNumberFromFile(frame_file)\n if not file_found:\n self.showMsg(f'No manual WCS calibration frame number found in the folder.')\n return\n else:\n if frame_num_of_wcs is None:\n self.showMsg(f'Content error in: {frame_file}')\n return\n else:\n got_frame_number = True\n\n ref_names = glob.glob(self.folder_dir + '/ref*.txt')\n if len(ref_names) == 2:\n self.showMsg(f'manual WCS calibration files found in the folder.')\n got_manual_wcs_calibration = True\n self.currentFrameSpinBox.setValue(frame_num_of_wcs)\n\n if got_fits_wcs_calibration and got_star_position and got_frame_number:\n self.wcs_frame_num = frame_num_of_wcs\n self.setApertureFromWcsData(ss, wcs_fits[0])\n\n if got_manual_wcs_calibration and got_star_position and got_frame_number:\n self.wcs_frame_num = frame_num_of_wcs\n self.doManualWcsCalibration()\n\n def extractUpperFieldFromImage(self):\n self.upper_field = self.image[0::2,:]\n\n def extractLowerFieldFromImage(self):\n self.lower_field = self.image[1::2,:]\n\n def createImageFields(self):\n self.extractLowerFieldFromImage()\n self.extractUpperFieldFromImage()\n try:\n self.image_fields = np.concatenate((self.upper_field, self.lower_field))\n except Exception as e:\n self.showMsg(f'shape of lower_field: {self.lower_field.shape}')\n self.showMsg(f'shape of upper_field: {self.upper_field.shape}')\n self.showMsg(f'in createImageFields: {e}')\n\n # This routine is only used by the frame stacker program --- it is passed as a parameter\n def getFitsFrame(self, frame_to_read):\n try:\n image = pyfits.getdata(\n self.fits_filenames[frame_to_read], 0).astype('float32', casting='unsafe')\n # self.showMsg(f'image shape: {self.image.shape}')\n except:\n image = None\n return image\n\n def showFrame(self):\n\n if self.record_target_aperture:\n self.showMsg(f'We will save the aperture location for enhanced placement')\n self.record_target_aperture = False\n app_list = self.getApertureList()\n if len(app_list) > 1:\n self.showMsg(f'!!!! Only a single target may be designated !!!!')\n self.clearApertures()\n elif len(app_list) == 1:\n aperture = app_list[0]\n x0, y0, _, _ = aperture.getBbox()\n xc = x0 + self.roi_center\n yc = y0 + self.roi_center\n\n # Save the aperture coordinates...\n self.showMsg(f'recorded: x:{xc} y:{yc}')\n with open(self.folder_dir + r'/target-aperture-xy.txt', 'w') as f:\n f.writelines(f'{xc} {yc}')\n\n # and set the current frame to the proper reference frame\n frame_file = 'enhanced-image-frame-num.txt'\n file_found, frame_num = self.getFrameNumberFromFile(frame_file)\n if file_found:\n if frame_num is None:\n self.showMsg(f'Content error in: {frame_file}')\n return\n else:\n self.showMsg(f'Set current frame to reference frame {frame_num}')\n self.currentFrameSpinBox.setValue(frame_num)\n\n # Local variables used to save and restore the pan/zoom state of the main image\n state = None\n view_box = None\n\n try:\n if not self.initialFrame:\n # We want to maintain whatever pan/zoom is in effect ...\n view_box = self.frameView.getView()\n # ... so we read and save the current state of the view box of our frameView\n state = view_box.getState()\n\n frame_to_show = self.currentFrameSpinBox.value() # Get the desired frame number from the spinner\n\n if self.avi_in_use:\n try:\n if self.fourcc == 'dvsd':\n success, frame = self.getFrame(frame_to_show)\n if len(frame.shape) == 3:\n self.image = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n else:\n self.cap.set(cv2.CAP_PROP_POS_FRAMES, frame_to_show)\n status, frame = self.cap.read()\n self.image = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)\n\n except Exception as e:\n self.showMsg(f'Problem reading avi file: {e}')\n else: # We're dealing with FITS files\n try:\n try:\n self.image = pyfits.getdata(\n self.fits_filenames[frame_to_show], 0).astype('int16', casting='unsafe')\n # self.showMsg(f'image shape: {self.image.shape}')\n except:\n self.image = None\n\n hdr = pyfits.getheader(self.fits_filenames[frame_to_show], 0)\n\n try:\n date_time = hdr['DATE-OBS']\n # The form of DATE-ObS is '2018-08-21T05:21:02.4561235' so we can simply 'split' at the T\n parts = date_time.split('T')\n self.showMsg(f'Timestamp found: {parts[0]} @ {parts[1]}')\n # We only want to save the date from the first file (to add to the csv file)...\n if self.initialFrame:\n self.fits_date = parts[0]\n\n # ...but we need the time from every new frame.\n self.fits_timestamp = f'[{parts[1]}]'\n except Exception as e:\n self.showMsg(f'{e}')\n pass\n # This scaling was used to be able to read a file from Joel --- not generally useful\n # except as an example\n # self.image = (pyfits.getdata(self.fits_filenames[frame_to_show], 0) / 3.0).astype('int16', casting='safe')\n except:\n self.showMsg(f'Cannot convert image to int16 safely')\n return\n # self.image = (pyfits.getdata(self.fits_filenames[frame_to_show], 0) / 3.0).astype('int16')\n # self.showMsg(f'image shape: {self.image.shape} type: {type(self.image[0,0])}')\n # self.showMsg(f'max:{np.max(self.image)} min:{np.min(self.image)}')\n\n if self.viewFieldsCheckBox.isChecked():\n self.createImageFields()\n self.frameView.setImage(self.image_fields)\n else:\n self.frameView.setImage(self.image)\n self.createImageFields()\n\n try:\n if self.avi_wcs_folder_in_use and self.timestampReadingEnabled:\n if self.timestampFormatter is not None:\n self.upperTimestamp, time1, score1, _, self.lowerTimestamp, time2, score2, _ = \\\n self.extractTimestamps()\n except Exception as e:\n self.showMsg(f'The following exception occurred while trying to read timestamp:',\n blankLine=False)\n self.showMsg(repr(e))\n\n if self.levels:\n self.frameView.setLevels(min=self.levels[0], max=self.levels[1])\n self.thumbOneView.setLevels(min=self.levels[0], max=self.levels[1])\n\n if not self.initialFrame:\n # Displaying the new image resets the pan/zoom to none ..\n # ... so here we restore the view box to the state extracted above.\n view_box.setState(state)\n else:\n self.initialFrame = False\n height, width = self.image.shape\n\n self.showMsg(f'image shape: width={width} height={height}')\n\n # The following variables are used by MeasurementAperture to limit\n # aperture placement so that it stays within the image at all times\n self.roi_max_x = width - self.roi_size\n self.roi_max_y = height - self.roi_size\n\n # We save these for use in displaying thumbnails at the same scaling as\n # the main image\n self.img_max = np.max(self.image)\n self.img_min = np.min(self.image)\n\n # if self.collectDataCheckBox.isChecked():\n # centerAllApertures() calls centerAperture() and\n # that routine will add the data to the aperture when collectDataCheckBox.isChecked()\n # We've changed philosophy: now apertures always 'track'\n try:\n self.centerAllApertures()\n except Exception as e:\n self.showMsg(f'during centerAllApertures(): {repr(e)} ')\n self.frameView.getView().update()\n\n # Find the auto_display (if any). We do dynamic thumbnail\n # display on such an aperture but let a 'pointed-at-aperture' trump all\n if self.pointed_at_aperture is not None:\n self.getApertureStats(self.pointed_at_aperture)\n else:\n for app in self.getApertureList():\n if app.auto_display and not app.thumbnail_source:\n self.getApertureStats(app)\n for app in self.getApertureList():\n if app.thumbnail_source:\n self.getApertureStats(app)\n\n except Exception as e:\n self.showMsg(repr(e))\n self.showMsg(f'There are no frames to display. Have you read a file?')\n\n def removeAperture(self, aperture):\n self.disconnectAllSlots(aperture)\n self.frameView.getView().removeItem(aperture)\n\n def removeOcrBox(self, ocrbox):\n self.frameView.getView().removeItem(ocrbox)\n\n def getApertureList(self):\n \"\"\"\n Returns all of the aperture objects that have been added\n to frameView\n \"\"\"\n\n # Get all objects that have been added to frameView\n items = self.frameView.getView().allChildItems()\n self.appList = []\n\n # Not all objects in frameView are apertures, so we need to filter the list\n for item in items:\n if type(item) is MeasurementAperture:\n self.appList.append(item)\n\n return self.appList\n\n def getOcrBoxList(self):\n\n # Get all objects that have been added to frameView\n items = self.frameView.getView().allChildItems()\n self.ocrBoxList = []\n\n # Not all objects in frameView are ocr boxes, so we need to filter the list\n for item in items:\n if type(item) is OcrAperture:\n self.ocrBoxList.append(item)\n\n return self.ocrBoxList\n\n def showInfo(self):\n self.openInfoFile()\n\n def showDocumentation(self):\n self.openDocFile()\n\n def setDoTestFlag(self):\n self.do_test = True\n\n def removePreviousWcsFiles(self):\n self.showMsg(f'A new WCS solution has been requested so we will', blankLine=False)\n self.showMsg(f'so there may be some pre-existing WCS related files to be removed.')\n files_to_delete = glob.glob(self.folder_dir + f'/frame*img.fit')\n for file in files_to_delete:\n self.showMsg(f'....deleting: {file}', blankLine=False)\n os.remove(file)\n files_to_delete = glob.glob(self.folder_dir + f'/wcs*.fit')\n for file in files_to_delete:\n self.showMsg(f'....deleting: {file}', blankLine=False)\n os.remove(file)\n self.showMsg(f'\\nWCS related files have been cleared out.')\n\n def getPixelAspectRatio(self):\n try:\n pixHeight = float(self.pixelHeightEdit.text())\n pixWidth = float(self.pixelWidthEdit.text())\n if not (pixWidth < 0.0 or pixHeight <= 0.0):\n self.pixelAspectRatio = pixWidth / pixHeight\n self.showMsg(f'pixel aspect ratio: {self.pixelAspectRatio:0.4f} (W/H)')\n # Write the pixel-dimensions.p file\n dims = (self.pixelHeightEdit.text(), self.pixelWidthEdit.text())\n pickle.dump(dims, open(self.folder_dir + '/pixel-dimensions.p', \"wb\"))\n except ValueError as e:\n self.pixelAspectRatio = None\n self.showMsg(f'in calculation of pixel aspect ratio: {e}', blankLine=False)\n self.showMsg(f'Possibly an empty field?')\n\n def resizeImage(self, image, aspect_ratio):\n self.showMsg(f'image shape: {image.shape}')\n height, width = image.shape\n if aspect_ratio <= 1.0:\n width = round(width * aspect_ratio)\n else:\n height = round(height / aspect_ratio)\n\n try:\n image_resized = skimage.transform.resize(image, (height, width), mode='edge',\n anti_aliasing=False, anti_aliasing_sigma=None,\n preserve_range=True, order=0)\n status = True\n self.showMsg(f'image_resized shape: {image_resized.shape}')\n except Exception as e:\n status = False\n image_resized = None\n self.showMsg(f'Resizing failed: {e}')\n\n return status, image_resized\n\n def getWCSsolution(self):\n\n if not (self.avi_wcs_folder_in_use or self.fits_folder_in_use):\n self.showMsg(f'No AVI-WCS or FITS folder is currently in use.', blankLine=False)\n self.showMsg(f'That is a requirement for this operation.')\n return\n\n self.getPixelAspectRatio()\n\n if self.pixelAspectRatio is None:\n self.showMsg(f'Failed to compute a valid pixel aspect ratio. Cannot continue')\n self.showMsgDialog(f'You must fill in pixel height and width in order to continue.')\n return\n\n # This is set in the selectAviFolder() or readFitsFile()method.\n dir_path = self.folder_dir\n\n # Check for presence of target-location.txt\n matching_name = sorted(glob.glob(dir_path + '/target-location.txt'))\n\n if not matching_name or not self.api_key:\n self.showMsg(f'No target location and/or api-key file found in the folder.')\n star_icrs = self.getStarPositionString()\n self.showMsg(f'star position string provided: \"{star_icrs}\"')\n if not star_icrs:\n self.showMsg(f'Cannot proceed without a star/target position entry.')\n return\n\n try:\n star_loc = SkyCoord(star_icrs, frame='icrs')\n except Exception as e:\n self.showMsg(f'star location string is invalid: {e}')\n return\n\n self.showMsg(f'RA: {star_loc.ra.value}')\n self.showMsg(f'Dec: {star_loc.dec.value}')\n\n with open(dir_path + r'/target-location.txt', 'w') as f:\n f.writelines(star_icrs)\n else:\n with open(dir_path + r'/target-location.txt', 'r') as f:\n star_icrs = f.read()\n self.showMsg(f'Star/target position is: {star_icrs}')\n\n try:\n star_loc = SkyCoord(star_icrs, frame='icrs')\n except Exception as e:\n self.showMsg(f'star location string is invalid: {e}')\n return\n\n self.showMsg(f'RA: {star_loc.ra.value}')\n self.showMsg(f'Dec: {star_loc.dec.value}')\n\n self.clearApertures()\n self.showFrame()\n\n # Get a robust mean from near the center of the current image\n y0 = int(self.image.shape[0]/2)\n x0 = int(self.image.shape[1]/2)\n ny = 51\n nx = 51\n thumbnail = self.image[y0:y0 + ny, x0:x0 + nx]\n mean, *_ = robustMeanStd(thumbnail, outlier_fraction=.5)\n\n image_height = self.image.shape[0] # number of rows\n image_width = self.image.shape[1] # number of columns\n\n num_lines_to_redact = 0\n\n if self.redactLinesEdit.text():\n try:\n num_lines_to_redact = int(self.redactLinesEdit.text())\n except ValueError:\n self.showMsg(f'invalid numeric entry: {self.redactLinesEdit.text()}')\n return\n else:\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText(f'It is good practice to remove any timestamp overlay that may be '\n f'present as such an overlay may impair star field identification. '\n f'\\n\\nPlease enter the number of lines from the bottom to remove '\n f'in the redact lines edit box. '\n f'Enter 0 if there is no timestamp.')\n msg.setWindowTitle('Please fill in redact lines')\n msg.setStandardButtons(QMessageBox.Ok)\n msg.exec()\n return\n\n if num_lines_to_redact < 0 or num_lines_to_redact > image_height / 2:\n self.showMsg(f'{num_lines_to_redact} is an unreasonable number of lines to redact.')\n self.showMsg(f'Operation aborted.')\n return\n\n redacted_image = self.image[:,:].astype('int16')\n for i in range(image_height - num_lines_to_redact, image_height):\n for j in range(0, image_width):\n redacted_image[i, j] = mean\n\n self.image = redacted_image\n self.frameView.setImage(self.image)\n if self.levels:\n self.frameView.setLevels(min=self.levels[0], max=self.levels[1])\n\n # Tests with nova.astrometry.net show that you should always give them the the original\n # (possibly redacted) image. DO NOT CLIP AND SCALE. It confuses their star\n # extractor which is extremely robust. So we comment out that little 'fiddle'\n # if self.levels:\n # processed_image = exposure.rescale_intensity(redacted_image, in_range=self.levels)\n # else:\n # processed_image = redacted_image\n processed_image = redacted_image\n\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText('Is the image suitable for submission to nova.astrometry.net for WCS calibration?')\n msg.setWindowTitle('Is image ready for submission')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n ready_for_submission = retval == QMessageBox.Yes\n\n if not ready_for_submission:\n self.showFrame()\n return\n\n # If this point is reached, we have a satisfactory image and a star position file,\n # so we are ready to try to make a submission to Astrometry.net\n\n if not self.pixelAspectRatio == 1.0:\n self.showMsg(f'The image will be resized from ...')\n # Here we will send processed_image out for resizing\n status, resized_image = self.resizeImage(processed_image, self.pixelAspectRatio)\n if not status:\n self.showMsg(f'Resizing failed.')\n return\n else:\n resized_image = processed_image\n\n self.removePreviousWcsFiles()\n\n frame_num = self.currentFrameSpinBox.value()\n with open(dir_path + r'/wcs-frame-num.txt', 'w') as f:\n f.writelines(f'{frame_num}')\n\n hdr = pyfits.Header()\n hdr['OBSERVER'] = 'PyMovie ' + version.version()\n hdr['FROMDIR'] = dir_path\n\n cal_image_path = dir_path + f'/frame-{frame_num}-img.fit'\n\n pyfits.writeto(cal_image_path, resized_image.astype('int16'), hdr, overwrite=True)\n\n # Login in to nova.astrometry.net usingthe supplied api key. We will need\n # each user to apply for his own.\n\n self.showMsg(f'Attempting to login to nova.astrometry.net using supplied api key.')\n QtGui.QGuiApplication.processEvents()\n\n c = astrometry_client.Client(tracer=self.showMsg, trace=False)\n try:\n # This will create a new session. There is apparently no need to close\n # a session --- no API call provided to do so anyway.\n c.login(self.api_key)\n except astrometry_client.RequestError as e:\n self.showMsg(f'Login attempt failed: {e}')\n return\n\n self.showMsg(f'Login to nova.astrometry.net succeeded.')\n\n QtGui.QGuiApplication.processEvents()\n\n # Set up the source file (image to calibrate) and the complete filepath for\n # any solution found.\n image_to_calibrate = cal_image_path\n calibration_file_dest = dir_path + f'/wcs-{frame_num}.fit'\n\n # These are the parameters/arguments that the 'solver' uses to work a little faster ...\n kwargs = dict()\n kwargs['center_ra'] = star_loc.ra.value\n kwargs['center_dec'] = star_loc.dec.value\n kwargs['crpix_center'] = True\n kwargs['radius'] = 1.0\n kwargs['scale_units'] = 'degwidth'\n kwargs['scale_lower'] = 0.1\n kwargs['scale_upper'] = 20.0\n\n self.showMsg(f'Submitting image for WCS calibration...')\n QtGui.QGuiApplication.processEvents()\n\n upload_result = c.upload(image_to_calibrate, **kwargs)\n\n # Wait for the upload to be accepted and submission id to be returned (subid)\n # and then start waiting for a job number to be assigned (in the 'jobs' dict entry)\n sub_id = str(upload_result['subid'])\n self.showMsg(f'...submission ID returned is {sub_id}.')\n self.showMsg(f'Waiting for job number to be assigned...')\n\n pass_counter = 0\n self.do_test = False\n self.timer.start(5000) # When this timer elapses, it sets self.do_test to True\n\n while True:\n QtGui.QGuiApplication.processEvents()\n if self.do_test:\n stat = c.sub_status(sub_id, justdict=True)\n # self.showMsg(f'Got status: {stat}')\n jobs = stat.get('jobs', [])\n if len(jobs):\n for j in jobs:\n if j is not None:\n break\n if j is not None:\n self.showMsg(\"\", blankLine=False)\n self.showMsg(f'...received job id {j}.')\n job_id = j\n break\n pass_counter += 1\n if pass_counter % 10 == 0:\n self.showMsg(f'\\nGot status: {stat}')\n self.showMsg(f'...waiting for job id (wait count is {pass_counter})', blankLine=False)\n self.do_test = False\n self.timer.stop()\n\n self.showMsg(f'Waiting for WCS solution...')\n\n self.do_test = False\n pass_counter = 0\n self.timer.start(5000) # When this timer elapses, it sets self.do_test to True\n\n while True:\n QtGui.QGuiApplication.processEvents()\n if self.do_test:\n stat = c.job_status(job_id, justdict=True)\n # self.showMsg(f'Got job status: {stat}')\n if stat.get('status', '') in ['success', 'failure']:\n success = (stat['status'] == 'success')\n solved_id = int(job_id)\n self.showMsg(\"\", blankLine=False)\n break\n pass_counter += 1\n if pass_counter % 10 == 0:\n self.showMsg(f'\\nGot job status: {stat}')\n self.showMsg(f'...still solving (wait count is {pass_counter})', blankLine=False)\n self.do_test = False\n self.timer.stop()\n\n if success:\n self.showMsg(f'A WCS solution was found.')\n # We don't need the API for this, just construct URL\n url = astrometry_client.Client.default_url.replace(\n '/api/', '/wcs_file/%i' % solved_id) # solved_id\n # self.showMsg(f'url: {url}')\n\n self.showMsg(f'Retrieving solution file from {url}')\n f = urlopen(url)\n txt = f.read()\n w = open(calibration_file_dest, 'wb')\n w.write(txt)\n w.close()\n self.showMsg(f'Wrote solution file to {calibration_file_dest}')\n\n self.wcs_frame_num = frame_num\n self.setApertureFromWcsData(star_icrs, calibration_file_dest)\n else:\n self.showMsg(f'WCS calibration failed.')\n\n def runExperimentalCode(self):\n\n # exporter = FixedImageExporter(self.save_p1.sceneObj)\n # exporter.makeWidthHeightInts()\n # targetFile = self.folder_dir + '/composite.png'\n # exporter.export(targetFile)\n # self.showMsg(f'A work in progress')\n pass\n\n def manualWcsCalibration(self):\n if not (self.avi_wcs_folder_in_use or self.fits_folder_in_use):\n self.showMsg(f'There is no WCS folder open.')\n return\n\n # Don't start manual WCS until self.pixelAspectRatio is known\n self.getPixelAspectRatio()\n if self.pixelAspectRatio is None:\n self.showMsg(f'Failed to compute a valid pixel aspect ratio. Cannot continue')\n self.showMsgDialog(f'You must fill in pixel height and width in order to continue.')\n return\n\n # if self.manual_wcs_state is None or self.manual_wcs_state > 0: # Initial state\n ref_filenames = sorted(glob.glob(self.folder_dir + '/ref*.txt'))\n\n if len(ref_filenames) > 0:\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Question)\n msg.setText('This operation will erase reference star information files' +\n ' from the previous manual calibration.' +\n ' Do you wish to continue?')\n msg.setWindowTitle('Confirmation requested')\n msg.setStandardButtons(QMessageBox.Yes | QMessageBox.No)\n retval = msg.exec_()\n if retval == QMessageBox.No:\n return\n else:\n self.showMsg(f'Proceeding ...')\n # Delete the existing reference star data files\n for file in ref_filenames:\n os.remove(file)\n self.showMsg(f'Deleted: {file}', blankLine=False)\n self.showMsg(\"\", blankLine=False)\n\n # Write the frame number file\n frame_num = self.currentFrameSpinBox.value()\n with open(self.folder_dir + r'/manual-wcs-frame-num.txt', 'w') as f:\n f.writelines(f'{frame_num}')\n\n self.showMsg(f'Manual WCS calibration process activated. Waiting for aperture 1 to be placed and RA DEC assigned.')\n self.manual_wcs_state = 1\n\n def setApertureFromWcsData(self, star_location, wcs_fits):\n\n try:\n star_loc = SkyCoord(star_location, frame='icrs')\n except Exception as e:\n self.showMsg(f'star location string is invalid: {e}')\n return\n\n # This context capture of AstropyWarning is to suppress the innocuous warning\n # FITSFixedWarning: The WCS transformation has more axes(2) than ....\n with warnings.catch_warnings():\n warnings.simplefilter('ignore', AstropyWarning)\n hdulist = pyfits.open(wcs_fits)\n w = wcs.WCS(hdulist[0].header)\n\n # Make the solution available for the cursor move routine\n self.wcs = w\n # self.showMsg(f'w.wcs.name={w.wcs.name}')\n pixcrd = np.array([[200, 200]], dtype='float')\n world = w.wcs_pix2world(pixcrd, 0)\n # self.showMsg(f'{world}')\n # self.showMsg(f'star_loc: {star_loc}')\n pixcrd2 = star_loc.to_pixel(w)\n # self.showMsg(f'{pixcrd2}')\n xcoord = pixcrd2[0].tolist()\n ycoord = pixcrd2[1].tolist()\n x = xcoord\n y = ycoord\n\n\n # Correct for pixel aspect ratio\n if not self.pixelAspectRatio == 1.0:\n if self.pixelAspectRatio < 1.0:\n x = x / self.pixelAspectRatio\n else:\n # This has never been tested, but should be correct\n y = y * self.pixelAspectRatio\n\n self.showMsg(f'astrometry.net: x={x:0.2f} y={y:0.2f}')\n target_app = self.addApertureAtPosition(round(x), round(y))\n target_app.thresh = self.big_thresh\n target_app.name = 'target'\n target_app.setRed()\n\n self.one_time_suppress_stats = True\n self.threshValueEdit.setValue(self.big_thresh) # Causes call to self.changeThreshold()\n\n self.wcs_solution_available = True\n\n def showRobustMeanDemo(self):\n\n dark_gray = (50, 50, 50)\n\n if self.thumbOneImage is None:\n self.showMsg(f'No image in Thumbnail One to use for demo')\n return\n\n good_mean, sigma, sorted_data, window, data_size, left, right = robustMeanStd(self.thumbOneImage)\n # self.showMsg(f'{good_mean} {sigma} {window} {data_size} {left} {right}')\n\n # Start a new plot\n self.plots.append(pg.GraphicsWindow(title=\"Robust Mean Calculation\"))\n self.plots[-1].resize(1000, 600)\n self.plots[-1].setWindowTitle(f'PyMovie {version.version()} Robust Mean Calculation')\n\n p1 = self.plots[-1].addPlot(\n row=0, col=0,\n y= self.thumbOneImage.flatten(),\n title=f'pixel values in thumbnail image (mean: green line; +/- sigma: red lines)',\n pen=dark_gray\n )\n hLineMean = pg.InfiniteLine(angle=0, movable=False, pen='g')\n p1.addItem(hLineMean, ignoreBounds=True)\n hLineMean.setPos(good_mean)\n\n hLineUpperStd = pg.InfiniteLine(angle=0, movable=False, pen='r')\n p1.addItem(hLineUpperStd, ignoreBounds=True)\n hLineUpperStd.setPos(good_mean + sigma)\n\n hLineLowerStd = pg.InfiniteLine(angle=0, movable=False, pen='r')\n p1.addItem(hLineLowerStd, ignoreBounds=True)\n hLineLowerStd.setPos(good_mean - sigma)\n\n self.plots[-1].nextRow() # Tell GraphicsWindow that we want another row of plots\n\n p2 = self.plots[-1].addPlot(\n row=1, col=0,\n y=sorted_data,\n title=f'sorted pixel values (red lines enclose \"non-outliers\")',\n pen=dark_gray\n )\n vLineLeft = pg.InfiniteLine(angle=90, movable=False, pen='r')\n vLineRight = pg.InfiniteLine(angle=90, movable=False, pen='r')\n p2.addItem(vLineLeft, ignoreBounds=True)\n p2.addItem(vLineRight, ignoreBounds=True)\n vLineLeft.setPos(left)\n vLineRight.setPos(right)\n\n self.plots[-1].show() # Let everyone see the results\n\n def showLightcurves(self):\n\n def mouseMovedFactory(p1, vb, label, vLine_p1, vLine_p2, xvalues, yvalues, pvalues, tvalues):\n def mouseMoved(evt):\n pos = evt\n if p1.sceneBoundingRect().contains(pos):\n mousePoint = vb.mapSceneToView(pos)\n dx = xvalues[1] - xvalues[0]\n if dx == 1.0:\n index = int(mousePoint.x() + 0.5)\n else:\n index = int(2 * mousePoint.x() + 0.5)\n # if xvalues[0] <= index <= xvalues[-1]:\n if xvalues[0] <= mousePoint.x() <= xvalues[-1]:\n try:\n # k = index - int(xvalues[0])\n if dx == 1.0:\n k = int(mousePoint.x() - xvalues[0] + 0.5)\n else:\n k = int((mousePoint.x() - xvalues[0]) * 2 + 0.5)\n\n p1.setTitle(f'{label} at frame {xvalues[k]}: intensity={yvalues[k]} '\n f'mask_pixels={pvalues[k]} timestamp={tvalues[k]}')\n except Exception as e:\n pass\n vLine_p1.setPos(mousePoint.x())\n vLine_p2.setPos(mousePoint.x())\n return mouseMoved\n\n def sortOnFrame(val):\n return val[8]\n\n # Create a color list (we use it circularly --- i.e., after teal we return to red)\n my_colors = [\n (200, 0, 0), # red\n (0, 200, 0), # green\n (0, 0, 200), # blue\n (200, 200, 0), # red-green (yellow)\n (200, 0, 200), # red-blue (purple)\n (0, 200, 200) # blue-green (teal)\n ]\n\n light_gray = (200, 200, 200)\n dark_gray = (50, 50, 50)\n\n # Enable antialiasing for prettier plots\n pg.setConfigOptions(antialias=True)\n\n appList = self.getApertureList()\n\n # Trap users asking for plots before there are even any apertures\n if len(appList) == 0:\n self.showMsg(f'There are no measurement apertures defined yet.')\n return\n\n cascadePosition = 50\n cascadeDelta = 26\n\n color_index = 0\n for app in appList:\n # Trap user asking for plots before data is present\n if len(app.data) == 0:\n self.showMsg(f'There is no data available to plot.')\n return\n\n app.data.sort(key = sortOnFrame)\n\n # Start a new plot for each aperture\n self.plots.append(pg.GraphicsWindow(title=\"PyMovie lightcurve plot\"))\n self.plots[-1].resize(1000, 600)\n if self.cascadeCheckBox.isChecked():\n self.plots[-1].move(QPoint(cascadePosition, cascadePosition))\n cascadePosition += cascadeDelta\n self.plots[-1].setWindowTitle(f'PyMovie {version.version()} lightcurve for aperture: {app.name}')\n\n yvalues = []\n xvalues = []\n for entry in app.data:\n yvalues.append(entry[4]) # signal==4 appsum==5 frame_num == 8\n xvalues.append(entry[8]) # signal==4 appsum==5 frame_num == 8 timestamp == 12\n\n # Here's how to add filtering if that ever becomes a desired feature\n # self.p3 = self.win.addPlot(values, pen=(200, 200, 200), symbolBrush=(255, 0, 0), symbolPen='w')\n # smooth_values = savgol_filter(values, 9 , 2)\n\n tvalues = [] # timestamps\n pvalues = []\n for entry in app.data:\n pvalues.append(entry[7]) # max_area (num pixels in aperture)\n tvalues.append(entry[12])\n\n pens = [pg.mkPen('r') if x > 0 else pg.mkPen('k') for x in pvalues]\n brushes = [pg.mkBrush('r') if x > 0 else pg.mkBrush('k') for x in pvalues]\n symbols = ['o' if x > 0 else 't' for x in pvalues]\n\n p1 = self.plots[-1].addPlot(\n row=0, col=0,\n x=xvalues, y=yvalues, title=f'{app.name} signal (background subtracted)',\n pen=light_gray, symbolBrush=brushes, name='plot1',\n symbolSize=self.plot_symbol_size, pxMode=True, symbolPen=pens, symbol=symbols\n )\n\n p1.setYRange(min(0, min(yvalues)), max(yvalues))\n p1.showGrid(y=True, alpha=1.0)\n p1.setXRange(xvalues[0] - 1, xvalues[-1] + 1)\n\n # TODO remove this experiment\n p1.setMouseEnabled(x=True, y=False)\n\n self.plots[-1].nextRow() # Tell GraphicsWindow that we want another row of plots\n\n pvalues = []\n for entry in app.data:\n pvalues.append(abs(entry[7])) # max_area (num pixels in aperture)\n\n p2 = self.plots[-1].addPlot(\n row=1, col=0,\n title=\"Number of pixels in aperture \",\n y=pvalues, x=xvalues,\n pen=dark_gray #, symbol='o', symbolSize=self.plot_symbol_size, symbolBrush='k', symbolPen='k'\n )\n p2.setYRange(min(min(pvalues),0), max(pvalues))\n p2.setXRange(xvalues[0] - 1, xvalues[-1] + 1)\n p2.setXLink('plot1')\n p2.showGrid(y=True, alpha=1.0)\n p2.setMouseEnabled(x=True, y=False)\n\n\n vLine_p1 = pg.InfiniteLine(angle=90, movable=False)\n vLine_p2 = pg.InfiniteLine(angle=90, movable=False)\n p1.addItem(vLine_p1, ignoreBounds=True)\n p2.addItem(vLine_p2, ignoreBounds=True)\n vb = p1.vb\n mouseMoved = mouseMovedFactory(p1, vb, f'{app.name} signal (background subtracted)', vLine_p1, vLine_p2,\n xvalues[:], yvalues[:], pvalues[:], tvalues[:])\n p1.scene().sigMouseMoved.connect(mouseMoved)\n\n qGraphicsGridLayout = self.plots[-1].ci.layout\n qGraphicsGridLayout.setRowStretchFactor(0, 2)\n qGraphicsGridLayout.setRowStretchFactor(1, 1)\n\n self.plots[-1].show() # Let everyone see the results\n\n # Move to the next color, wrapping if end of available unique colors\n # has been reached.\n color_index += 1\n if color_index >= len(my_colors):\n color_index = 0\n\n # Add a composite plot of all lightcurves\n self.plots.append(pg.GraphicsWindow(title=f'PyMovie {version.version()} composite lightcurve'))\n # pw = PlotWidget(title=f'PyMovie {version.version()} composite lightcurve')\n # self.plots.append(pw.getPlotItem())\n self.plots[-1].resize(1000, 600)\n if self.cascadeCheckBox.isChecked():\n self.plots[-1].move(QPoint(cascadePosition, cascadePosition))\n p1 = self.plots[-1].addPlot(title=f'Composite lightcurve plot')\n p1.addLegend()\n p1.setMouseEnabled(x=True, y=False)\n\n max_max = 0\n color_index = 0\n min_min = 0\n for app in appList:\n yvalues = []\n xvalues = []\n for entry in app.data:\n yvalues.append(entry[4]) # signal==4 appsum==5 frame_num == 8\n xvalues.append(entry[8]) # signal==4 appsum==5 frame_num == 8\n p1.plot(\n x=xvalues, y=yvalues, title=\"Aperture intensity\",\n pen=light_gray, symbolBrush=my_colors[color_index],\n symbolSize=self.plot_symbol_size, pxMode=True, symbolPen=my_colors[color_index],\n name=f'    {app.name}'\n )\n max_max = max(max_max, max(yvalues))\n min_min = min(min_min, min(yvalues))\n p1.setYRange(min(0, min_min), max_max)\n\n # Move to the next color, wrapping if end of available unique colors\n # has been reached.\n color_index += 1\n if color_index >= len(my_colors):\n color_index = 0\n\n p1.showGrid(y=True)\n\n self.plots[-1].show() # Let everyone see the results\n\n QtGui.QGuiApplication.processEvents()\n\n self.save_p1 = p1\n\n\n def clearTextBox(self):\n self.textOut.clear()\n title = f'PyMovie Version: {version.version()}'\n self.showMsg(title)\n self.showMsg(f'Home directory: {self.homeDir}')\n\n def showMsg(self, msg, blankLine=True):\n self.textOut.append(msg)\n self.textOut.moveCursor(QtGui.QTextCursor.End)\n\n if blankLine:\n self.textOut.append(\"\")\n self.textOut.moveCursor(QtGui.QTextCursor.End)\n\n self.textOut.ensureCursorVisible()\n\n def closeEvent(self, event):\n # Capture the close request and update 'sticky' settings\n self.settings.setValue('size', self.size())\n self.settings.setValue('pos', self.pos())\n # self.settings.setValue('logscale', self.logScalingCheckBox.isChecked())\n self.settings.setValue('cascade', self.cascadeCheckBox.isChecked())\n self.settings.setValue('plot_symbol_size', self.plotSymbolSizeSpinBox.value())\n self.settings.setValue('splitterOne', self.splitterOne.saveState())\n self.settings.setValue('splitterTwo', self.splitterTwo.saveState())\n self.settings.setValue('splitterThree', self.splitterThree.saveState())\n\n if self.apertureEditor:\n self.apertureEditor.close()\n\n if self.helperThing:\n self.helperThing.close()\n\n if self.cap:\n self.cap.release()\n\n self.timer.stop()\n\n if self.plots:\n for plot in self.plots:\n plot.close()\n\n event.accept()\n\n def openInfoFile(self):\n infoFilePath = os.path.join(os.path.split(__file__)[0], 'PyMovie-info.pdf')\n\n url = QtCore.QUrl.fromLocalFile(infoFilePath)\n fileOpened = QtGui.QDesktopServices.openUrl(url)\n\n if not fileOpened:\n self.showMsg('Failed to open PyMovie version-info file', blankLine=False)\n self.showMsg('Location of PyMovie version-info file: ' + infoFilePath)\n\n def openDocFile(self):\n docFilePath = os.path.join(os.path.split(__file__)[0], 'PyMovie-doc.pdf')\n\n url = QtCore.QUrl.fromLocalFile(docFilePath)\n fileOpened = QtGui.QDesktopServices.openUrl(url)\n\n if not fileOpened:\n self.showMsg('Failed to open PyMovie documentation file', blankLine=False)\n self.showMsg('Location of PyMovie documentaion file: ' + docFilePath)\n\n def showBbox(self, bbox, border=0, color=PyQt5.QtCore.Qt.darkYellow):\n ymin, xmin, ymax, xmax = bbox\n ymin -= border\n ymax += border\n xmin -= border\n xmax += border\n\n view_box = self.frameView.getView()\n pen = QtGui.QPen(color)\n rect_item = QGraphicsRectItem(QRectF(xmin, ymin, xmax - xmin + 1, ymax - ymin + 1))\n rect_item.setPen(pen)\n view_box.addItem(rect_item)\n self.rect_list.append(rect_item)\n\n\ndef jogAperture(aperture, delta_xc, delta_yc):\n # Get coordinate info of this aperture\n bbox = aperture.getBbox()\n x0, y0, xsize, ysize = bbox\n\n # Jog the bbox by the amounts given.\n bbox = (x0 - delta_xc, y0 - delta_yc, xsize, ysize)\n\n # The setPos() method will intervene, if necessary, to keep the total extent of\n # the aperture inside the image\n aperture.setPos(bbox)\n\n\n# noinspection PyChainedComparisons,PyChainedComparisons\ndef calcTheta(dx, dy):\n d = sqrt(dx * dx + dy * dy)\n a = arcsin(dy / d)\n if dx >= 0 and dy >= 0:\n theta = a\n elif dx <= 0 and dy >= 0:\n theta = PI - a\n elif dx <= 0 and dy <= 0:\n theta = PI - a\n elif dx >= 0 and dy <= 0:\n theta = PI + PI + a\n else:\n return None, None\n return theta, theta * 180 / PI\n\n\ndef get_mask(\n img, ksize=(5, 5), cut=None, min_pixels=9,\n outlier_fraction=0.5,\n apply_centroid_distance_constraint=False, max_centroid_distance=None):\n\n blurred_img = cv2.GaussianBlur(img, ksize=ksize, sigmaX=0)\n\n # cut is threshold\n ret, t_mask = cv2.threshold(blurred_img, cut, 1, cv2.THRESH_BINARY)\n labels = measure.label(t_mask, neighbors=4, background=0)\n blob_count = np.max(labels)\n\n centroid = (None, None)\n max_area = 0\n max_signal = 0\n cvxhull = 0\n extent = 0\n bbox = None\n\n # We assume/require that measurement apertures be square. Without that 'truth',\n # the following calculation will be invalid\n roi_center = int(img.shape[0] / 2)\n\n bkavg, *_ = robustMeanStd(img, outlier_fraction=outlier_fraction)\n blob_signals = []\n\n if blob_count > 0:\n max_area = 0\n cvxhull = 0\n props = measure.regionprops(labels)\n coords = []\n for prop in props:\n if apply_centroid_distance_constraint:\n xc, yc = prop.centroid\n distance_to_center = np.sqrt((xc - roi_center)**2 + (yc - roi_center)**2)\n if distance_to_center > max_centroid_distance:\n continue\n\n # Here we compute the net signal that is contained in this particular blob\n signal = 0\n for point in prop.coords:\n signal += img[point[0], point[1]] - bkavg\n blob_signals.append(signal)\n\n if signal > max_signal:\n max_signal = signal\n max_area = prop.area\n coords = prop.coords\n centroid = prop.centroid\n cvxhull = prop.convex_area\n bbox = prop.bbox\n\n # Here is how we use a prop (from a label) and create a unit mask (0 or 1)\n mask = np.zeros(img.shape, 'int16')\n\n # calculate extent\n if bbox:\n min_row, min_col, max_row, max_col = bbox\n extent = max(max_col - min_col, max_row - min_row)\n\n # TODO Do we still want to consider min_pixels?\n if max_area >= min_pixels:\n for point in coords:\n mask[point[0], point[1]] = 1\n else:\n max_area = 0\n\n else:\n # We get here if number of blobs found was zero\n mask = np.zeros(img.shape, 'int16')\n\n return max_area, mask, t_mask, centroid, cvxhull, blob_count, extent\n\n\ndef robustMeanStd(data, outlier_fraction=0.5, max_pts=10000, assume_gaussian=True):\n # Note: it is expected that type(data) is numpy.darray\n\n # Protect the user against accidentally running this procedure with an\n # excessively large number of data points (which could take too long)\n if data.size > max_pts:\n raise Exception(\n f'In robustMean(): data.size limit of {max_pts} exceeded. (Change max_pts if needed)'\n )\n\n if outlier_fraction > 1:\n raise Exception(\n f'In robustMean(): {outlier_fraction} was given as outlier_fraction. This value must be <= 1.0'\n )\n\n # The None 'flattens' data automatically so sorted_data will be 1D\n sorted_data = np.sort(data, None)\n\n if outlier_fraction > 0:\n # window is the number points to be included in the 'mean' calculation\n window = int(sorted_data.size * (1 - outlier_fraction))\n\n # Handle the case of outlier_fraction too close to zero\n if window == data.size:\n window -= 1\n\n # nout is the number of outliers to exclude\n nout = sorted_data.size - window\n diffs = sorted_data[window:window + nout] - sorted_data[0:nout]\n\n min_diff_pts = np.where(diffs == min(diffs))\n\n j = min_diff_pts[0][0]\n k = min_diff_pts[0][-1]\n data_used = sorted_data[j:k + window]\n first_index = j\n last_index = k + window - 1\n else:\n first_index = 0\n last_index = data.size - 1\n data_used = sorted_data\n window = data.size\n\n good_mean = np.mean(data_used)\n\n # MAD means: Median Absolute Deviation This is a robust estimator of 'scale' (measure of data dispersion)\n # It can be related to standard deviation by a correction factor if the data can be assumed to be drawn\n # from a gaussian distribution.\n # med = np.median(sorted_data)\n sigma = np.median(np.abs(sorted_data - good_mean)) # This is the MAD estimator\n if assume_gaussian:\n sigma = sigma * 1.486 # sigma(gaussian) can be proved to equal 1.486*MAD\n\n return good_mean, sigma, sorted_data, window, data.size, first_index, last_index\n\n\ndef main():\n if sys.version_info < (3,7):\n sys.exit('Sorry, this program requires Python 3.7+')\n\n import traceback\n import os\n # QtGui.QApplication.setStyle('windows')\n QtGui.QApplication.setStyle('fusion')\n app = QtGui.QApplication(sys.argv)\n\n # if os.name == 'posix':\n # print(f'os: MacOS')\n # else:\n # print(f'os: Windows')\n # app.setStyleSheet(\"QLabel, QPushButton, QToolButton, QCheckBox, QRadioButton {font-size: 8pt}\")\n\n if sys.platform == 'linux':\n print(f'os: Linux')\n elif sys.platform == 'darwin':\n print(f'os: MacOS')\n else:\n print(f'os: Windows')\n app.setStyleSheet(\"QLabel, QPushButton, QToolButton, QCheckBox, QRadioButton, QLineEdit {font-size: 8pt}\")\n\n # Save the current/proper sys.excepthook object\n # sys._excepthook = sys.excepthook\n saved_excepthook = sys.excepthook\n\n def exception_hook(exctype, value, tb):\n # The next lines are a horrible hack to deal with the pyqtgraph Histogram widget.\n # It cannot be disabled, but if given an image containing pixels of exactly one value,\n # it throws an exception that is harmless but disturbing to have printed out in the\n # console all the time. Here I intercept that an qietly suppress the normal display\n # of an uncaught (in my code) exception.\n s = str(value)\n if s.startswith('arange:'):\n return None\n # End horrible hack\n\n print('')\n print('=' * 30)\n print(value)\n print('=' * 30)\n print('')\n\n traceback.print_tb(tb)\n # Call the usual exception processor\n # sys._excepthook(exctype, value, tb)\n saved_excepthook(exctype, value, tb)\n # Exit if you prefer...\n # sys.exit(1)\n\n sys.excepthook = exception_hook\n\n main_window = PyMovie()\n main_window.show()\n app.exec_()\n\n\nif __name__ == '__main__':\n main()\n", "sub_path": "src/pymovie/main.py", "file_name": "main.py", "file_ext": "py", "file_size_in_byte": 233493, "program_lang": "python", "lang": "en", "doc_type": "code", "dataset": "code-starcoder2", "pt": "63", "api": [{"api_name": "matplotlib.use", "line_number": 34, "usage_type": "call"}, {"api_name": "os.name", "line_number": 109, "usage_type": "attribute"}, {"api_name": "warnings.filterwarnings", "line_number": 112, "usage_type": "call"}, {"api_name": "warnings.filterwarnings", "line_number": 113, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 118, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 120, "usage_type": "call"}, {"api_name": "numpy.log10", "line_number": 128, "usage_type": "call"}, {"api_name": "pyqtgraph.exporters.ImageExporter", "line_number": 131, "usage_type": "attribute"}, {"api_name": "pyqtgraph.exporters", "line_number": 131, "usage_type": "name"}, {"api_name": "pyqtgraph.exporters.ImageExporter.__init__", "line_number": 133, "usage_type": "call"}, {"api_name": "pyqtgraph.exporters.ImageExporter", "line_number": 133, "usage_type": "attribute"}, {"api_name": "pyqtgraph.exporters", "line_number": 133, "usage_type": "name"}, {"api_name": "pymovie.helpDialog.Ui_Dialog", "line_number": 152, "usage_type": "attribute"}, {"api_name": "pymovie.helpDialog", "line_number": 152, "usage_type": "name"}, {"api_name": "pymovie.ocrProfileNameDialog.Ui_ocrNameDialog", "line_number": 158, "usage_type": "attribute"}, {"api_name": "pymovie.ocrProfileNameDialog", "line_number": 158, "usage_type": "name"}, {"api_name": "pymovie.selectProfile.Ui_Dialog", "line_number": 164, "usage_type": "attribute"}, {"api_name": "pymovie.selectProfile", "line_number": 164, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QTableWidgetItem", "line_number": 208, "usage_type": "call"}, {"api_name": "pymovie.starPositionDialog.Ui_Dialog", "line_number": 226, "usage_type": "attribute"}, {"api_name": "pymovie.starPositionDialog", "line_number": 226, "usage_type": "name"}, {"api_name": "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg", "line_number": 232, "usage_type": "name"}, {"api_name": "matplotlib.pyplot.figure", "line_number": 236, "usage_type": "call"}, {"api_name": "matplotlib.pyplot", "line_number": 236, "usage_type": "name"}, {"api_name": "numpy.meshgrid", "line_number": 252, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg.__init__", "line_number": 259, "usage_type": "call"}, {"api_name": "matplotlib.backends.backend_qt5agg.FigureCanvasQTAgg", "line_number": 259, "usage_type": "name"}, {"api_name": "pymovie.gui.Ui_MainWindow", "line_number": 264, "usage_type": "attribute"}, {"api_name": "pymovie.gui", "line_number": 264, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 284, "usage_type": "call"}, {"api_name": "os.path", "line_number": 284, "usage_type": "attribute"}, {"api_name": "pymovie.version.version", "line_number": 289, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 289, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QSettings", "line_number": 300, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QSettings.IniFormat", "line_number": 300, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QSize", "line_number": 304, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QPoint", "line_number": 305, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 384, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 384, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 385, "usage_type": "call"}, {"api_name": "os.path", "line_number": 385, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 386, "usage_type": "call"}, {"api_name": "os.path", "line_number": 386, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 387, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 401, "usage_type": "call"}, {"api_name": "os.path.expanduser", "line_number": 433, "usage_type": "call"}, {"api_name": "os.path", "line_number": 433, "usage_type": "attribute"}, {"api_name": "os.path.basename", "line_number": 439, "usage_type": "call"}, {"api_name": "os.path", "line_number": 439, "usage_type": "attribute"}, {"api_name": "PyQt5.QtCore.QTimer", "line_number": 514, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QButtonGroup", "line_number": 698, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 841, "usage_type": "call"}, {"api_name": "os.path", "line_number": 841, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 842, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 858, "usage_type": "call"}, {"api_name": "os.path", "line_number": 858, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 859, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 922, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 925, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1004, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 1005, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1005, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1008, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1008, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1008, "usage_type": "attribute"}, {"api_name": "pymovie.checkForNewerVersion.getMostRecentVersionOfPyMovie", "line_number": 1013, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 1015, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 1015, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1020, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1020, "usage_type": "name"}, {"api_name": "pymovie.checkForNewerVersion.upgradePyMovie", "line_number": 1030, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 1039, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 1039, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.DirectoryOnly", "line_number": 1041, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 1041, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 1043, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 1043, "usage_type": "name"}, {"api_name": "os.path.basename", "line_number": 1052, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1052, "usage_type": "attribute"}, {"api_name": "os.path.splitext", "line_number": 1053, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1053, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1055, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1055, "usage_type": "attribute"}, {"api_name": "pathlib.Path", "line_number": 1063, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 1064, "usage_type": "attribute"}, {"api_name": "pymovie.alias_lnk_resolver.create_osx_alias_in_dir", "line_number": 1065, "usage_type": "call"}, {"api_name": "pymovie.alias_lnk_resolver", "line_number": 1065, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 1072, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1074, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1074, "usage_type": "attribute"}, {"api_name": "os.symlink", "line_number": 1076, "usage_type": "call"}, {"api_name": "errno.EEXIST", "line_number": 1079, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1080, "usage_type": "call"}, {"api_name": "os.symlink", "line_number": 1081, "usage_type": "call"}, {"api_name": "os.makedirs", "line_number": 1089, "usage_type": "call"}, {"api_name": "winshell.shortcut", "line_number": 1091, "usage_type": "call"}, {"api_name": "os.path.basename", "line_number": 1092, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1092, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1093, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1093, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 1104, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 1116, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1221, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1221, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1222, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1222, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1224, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1224, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1225, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1225, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 1227, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 1228, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 1230, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 1231, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1245, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1245, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1246, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1246, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1249, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1249, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1250, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1250, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1252, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1252, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 1253, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 1255, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1261, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1261, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 1262, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 1264, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1288, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1288, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 1289, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1293, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1293, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1294, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1294, "usage_type": "attribute"}, {"api_name": "os.remove", "line_number": 1295, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1301, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1301, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1303, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1303, "usage_type": "attribute"}, {"api_name": "pickle.load", "line_number": 1304, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 1460, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1460, "usage_type": "attribute"}, {"api_name": "pickle.dump", "line_number": 1703, "usage_type": "call"}, {"api_name": "sys.platform", "line_number": 1813, "usage_type": "attribute"}, {"api_name": "platform.mac_ver", "line_number": 1815, "usage_type": "call"}, {"api_name": "os.environ", "line_number": 1816, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1817, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1817, "usage_type": "attribute"}, {"api_name": "os.environ", "line_number": 1819, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1820, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1820, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1822, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1822, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 1822, "usage_type": "call"}, {"api_name": "os.chmod", "line_number": 1825, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 1830, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1830, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 1833, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1833, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 1835, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1835, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 1835, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1867, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 1868, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1868, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1875, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1875, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1880, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 1881, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1881, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 1885, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1885, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1912, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 1913, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1913, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1916, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1916, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1916, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1918, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1918, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1934, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 1935, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1935, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 1939, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1939, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1939, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 1941, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 1941, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 1949, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 1954, "usage_type": "call"}, {"api_name": "pymovie.stacker.frameStacker", "line_number": 1963, "usage_type": "call"}, {"api_name": "pymovie.stacker", "line_number": 1963, "usage_type": "name"}, {"api_name": "os.path.isfile", "line_number": 1971, "usage_type": "call"}, {"api_name": "os.path", "line_number": 1971, "usage_type": "attribute"}, {"api_name": "astroquery.vizier.Vizier", "line_number": 1988, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 1993, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 2010, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 2011, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 2011, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 2017, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 2017, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 2017, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 2019, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 2019, "usage_type": "name"}, {"api_name": "PyQt5.QtGui.QKeySequence", "line_number": 2038, "usage_type": "call"}, {"api_name": "PyQt5.QtGui", "line_number": 2038, "usage_type": "attribute"}, {"api_name": "numpy.zeros", "line_number": 2299, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 2302, "usage_type": "call"}, {"api_name": "astropy.io.fits.getheader", "line_number": 2321, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 2321, "usage_type": "name"}, {"api_name": "pathlib.Path", "line_number": 2505, "usage_type": "call"}, {"api_name": "site.getusersitepackages", "line_number": 2505, "usage_type": "call"}, {"api_name": "site.getsitepackages", "line_number": 2507, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 2508, "usage_type": "call"}, {"api_name": "pathlib.Path", "line_number": 2511, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 2518, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 2518, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getSaveFileName", "line_number": 2522, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 2522, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getSaveFileName", "line_number": 2530, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 2530, "usage_type": "name"}, {"api_name": "more_itertools.sort_together", "line_number": 2560, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 2566, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 2566, "usage_type": "name"}, {"api_name": "subprocess.Popen", "line_number": 2619, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 2656, "usage_type": "call"}, {"api_name": "numpy.ceil", "line_number": 2658, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2689, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2693, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2698, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2709, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2722, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2731, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.pyqtSlot", "line_number": 2736, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 2803, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 2840, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 2890, "usage_type": "call"}, {"api_name": "pymovie.wcs_helper_functions.solve_triangle", "line_number": 2920, "usage_type": "call"}, {"api_name": "pymovie.wcs_helper_functions", "line_number": 2920, "usage_type": "name"}, {"api_name": "numpy.zeros", "line_number": 3027, "usage_type": "call"}, {"api_name": "numpy.max", "line_number": 3039, "usage_type": "call"}, {"api_name": "cv2.copyMakeBorder", "line_number": 3043, "usage_type": "call"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 3043, "usage_type": "attribute"}, {"api_name": "cv2.copyMakeBorder", "line_number": 3044, "usage_type": "call"}, {"api_name": "cv2.BORDER_CONSTANT", "line_number": 3044, "usage_type": "attribute"}, {"api_name": "cv2.hconcat", "line_number": 3046, "usage_type": "call"}, {"api_name": "numpy.cos", "line_number": 3324, "usage_type": "call"}, {"api_name": "numpy.sin", "line_number": 3325, "usage_type": "call"}, {"api_name": "numpy.array", "line_number": 3448, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 3450, "usage_type": "call"}, {"api_name": "astropy.units.deg", "line_number": 3450, "usage_type": "attribute"}, {"api_name": "astropy.units", "line_number": 3450, "usage_type": "name"}, {"api_name": "numpy.ones", "line_number": 3549, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3556, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3619, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3620, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3624, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3673, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3675, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3681, "usage_type": "call"}, {"api_name": "numpy.sum", "line_number": 3683, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 3763, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3763, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.ShowDirsOnly", "line_number": 3764, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3764, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 3766, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3766, "usage_type": "name"}, {"api_name": "glob.glob", "line_number": 3797, "usage_type": "call"}, {"api_name": "os.path.exists", "line_number": 3799, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3799, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 3822, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3822, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 3841, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3841, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 3845, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 3860, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 3860, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 3862, "usage_type": "call"}, {"api_name": "os.path.isfile", "line_number": 3880, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3880, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 3886, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3886, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getOpenFileName", "line_number": 3889, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3889, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 3909, "usage_type": "call"}, {"api_name": "os.path", "line_number": 3909, "usage_type": "attribute"}, {"api_name": "cv2.imread", "line_number": 3912, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_POS_FRAMES", "line_number": 3930, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 3960, "usage_type": "call"}, {"api_name": "cv2.CAP_FFMPEG", "line_number": 3960, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_POS_FRAMES", "line_number": 3961, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 3977, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3977, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getOpenFileName", "line_number": 3980, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 3980, "usage_type": "name"}, {"api_name": "os.path.split", "line_number": 4004, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4004, "usage_type": "attribute"}, {"api_name": "cv2.VideoCapture", "line_number": 4020, "usage_type": "call"}, {"api_name": "cv2.CAP_FFMPEG", "line_number": 4020, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FOURCC", "line_number": 4030, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 4035, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 4050, "usage_type": "attribute"}, {"api_name": "os.path.join", "line_number": 4092, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4092, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 4093, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4093, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.Options", "line_number": 4102, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 4102, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.ShowDirsOnly", "line_number": 4103, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 4103, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QFileDialog.getExistingDirectory", "line_number": 4106, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QFileDialog", "line_number": 4106, "usage_type": "name"}, {"api_name": "sys.platform", "line_number": 4173, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 4175, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 4183, "usage_type": "call"}, {"api_name": "pymovie.alias_lnk_resolver.resolve_osx_alias", "line_number": 4191, "usage_type": "call"}, {"api_name": "pymovie.alias_lnk_resolver", "line_number": 4191, "usage_type": "name"}, {"api_name": "os.readlink", "line_number": 4194, "usage_type": "call"}, {"api_name": "winshell.shortcut", "line_number": 4196, "usage_type": "call"}, {"api_name": "cv2.VideoCapture", "line_number": 4219, "usage_type": "call"}, {"api_name": "cv2.CAP_PROP_FOURCC", "line_number": 4229, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 4232, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FRAME_COUNT", "line_number": 4234, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_FPS", "line_number": 4237, "usage_type": "attribute"}, {"api_name": "os.path.exists", "line_number": 4272, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4272, "usage_type": "attribute"}, {"api_name": "os.path.isfile", "line_number": 4317, "usage_type": "call"}, {"api_name": "os.path", "line_number": 4317, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 4329, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 4331, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4358, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4379, "usage_type": "call"}, {"api_name": "pickle.load", "line_number": 4381, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4387, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4422, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4457, "usage_type": "call"}, {"api_name": "numpy.concatenate", "line_number": 4481, "usage_type": "call"}, {"api_name": "astropy.io.fits.getdata", "line_number": 4490, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 4490, "usage_type": "name"}, {"api_name": "cv2.cvtColor", "line_number": 4546, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 4546, "usage_type": "attribute"}, {"api_name": "cv2.CAP_PROP_POS_FRAMES", "line_number": 4548, "usage_type": "attribute"}, {"api_name": "cv2.cvtColor", "line_number": 4550, "usage_type": "call"}, {"api_name": "cv2.COLOR_BGR2GRAY", "line_number": 4550, "usage_type": "attribute"}, {"api_name": "astropy.io.fits.getdata", "line_number": 4557, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 4557, "usage_type": "name"}, {"api_name": "astropy.io.fits.getheader", "line_number": 4563, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 4563, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 4627, "usage_type": "call"}, {"api_name": "numpy.min", "line_number": 4628, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4705, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 4708, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 4709, "usage_type": "call"}, {"api_name": "os.remove", "line_number": 4712, "usage_type": "call"}, {"api_name": "pickle.dump", "line_number": 4724, "usage_type": "call"}, {"api_name": "skimage.transform.resize", "line_number": 4739, "usage_type": "call"}, {"api_name": "skimage.transform", "line_number": 4739, "usage_type": "attribute"}, {"api_name": "glob.glob", "line_number": 4769, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 4780, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 4796, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4827, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 4828, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4828, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Ok", "line_number": 4835, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4835, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4863, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 4864, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4864, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 4867, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4867, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 4867, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 4869, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 4869, "usage_type": "name"}, {"api_name": "astropy.io.fits.Header", "line_number": 4894, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 4894, "usage_type": "name"}, {"api_name": "pymovie.version.version", "line_number": 4895, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 4895, "usage_type": "name"}, {"api_name": "astropy.io.fits.writeto", "line_number": 4900, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 4900, "usage_type": "name"}, {"api_name": "pymovie.astrometry_client.Client", "line_number": 4908, "usage_type": "call"}, {"api_name": "pymovie.astrometry_client", "line_number": 4908, "usage_type": "name"}, {"api_name": "pymovie.astrometry_client.RequestError", "line_number": 4913, "usage_type": "attribute"}, {"api_name": "pymovie.astrometry_client", "line_number": 4913, "usage_type": "name"}, {"api_name": "pymovie.astrometry_client.Client.default_url.replace", "line_number": 4999, "usage_type": "call"}, {"api_name": "pymovie.astrometry_client.Client", "line_number": 4999, "usage_type": "attribute"}, {"api_name": "pymovie.astrometry_client", "line_number": 4999, "usage_type": "name"}, {"api_name": "urllib.request.urlopen", "line_number": 5004, "usage_type": "call"}, {"api_name": "glob.glob", "line_number": 5038, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 5041, "usage_type": "call"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Question", "line_number": 5042, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 5042, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.Yes", "line_number": 5047, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 5047, "usage_type": "name"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 5047, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox.No", "line_number": 5049, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QMessageBox", "line_number": 5049, "usage_type": "name"}, {"api_name": "os.remove", "line_number": 5055, "usage_type": "call"}, {"api_name": "astropy.coordinates.SkyCoord", "line_number": 5070, "usage_type": "call"}, {"api_name": "warnings.catch_warnings", "line_number": 5077, "usage_type": "call"}, {"api_name": "warnings.simplefilter", "line_number": 5078, "usage_type": "call"}, {"api_name": "astropy.utils.exceptions.AstropyWarning", "line_number": 5078, "usage_type": "argument"}, {"api_name": "astropy.io.fits.open", "line_number": 5079, "usage_type": "call"}, {"api_name": "astropy.io.fits", "line_number": 5079, "usage_type": "name"}, {"api_name": "astropy.wcs.WCS", "line_number": 5080, "usage_type": "call"}, {"api_name": "astropy.wcs", "line_number": 5080, "usage_type": "name"}, {"api_name": "numpy.array", "line_number": 5085, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 5130, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 5130, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QPoint", "line_number": 5238, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 5240, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 5240, "usage_type": "name"}, {"api_name": "pymovie.version.version", "line_number": 5317, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 5317, "usage_type": "name"}, {"api_name": "PyQt5.QtCore.QPoint", "line_number": 5322, "usage_type": "call"}, {"api_name": "pymovie.version.version", "line_number": 5363, "usage_type": "call"}, {"api_name": "pymovie.version", "line_number": 5363, "usage_type": "name"}, {"api_name": "os.path.join", "line_number": 5406, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5406, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 5406, "usage_type": "call"}, {"api_name": "os.path.join", "line_number": 5416, "usage_type": "call"}, {"api_name": "os.path", "line_number": 5416, "usage_type": "attribute"}, {"api_name": "os.path.split", "line_number": 5416, "usage_type": "call"}, {"api_name": "PyQt5.QtCore", "line_number": 5425, "usage_type": "attribute"}, {"api_name": "PyQt5.QtWidgets.QGraphicsRectItem", "line_number": 5434, "usage_type": "call"}, {"api_name": "PyQt5.QtCore.QRectF", "line_number": 5434, "usage_type": "call"}, {"api_name": "numpy.sqrt", "line_number": 5455, "usage_type": "call"}, {"api_name": "numpy.arcsin", "line_number": 5456, "usage_type": "call"}, {"api_name": "numpy.pi", "line_number": 5460, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 5462, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 5464, "usage_type": "name"}, {"api_name": "numpy.pi", "line_number": 5467, "usage_type": "name"}, {"api_name": "cv2.GaussianBlur", "line_number": 5475, "usage_type": "call"}, {"api_name": "cv2.threshold", "line_number": 5478, "usage_type": "call"}, {"api_name": "cv2.THRESH_BINARY", "line_number": 5478, "usage_type": "attribute"}, {"api_name": "skimage.measure.label", "line_number": 5479, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 5479, "usage_type": "name"}, {"api_name": "numpy.max", "line_number": 5480, "usage_type": "call"}, {"api_name": "skimage.measure.regionprops", "line_number": 5499, "usage_type": "call"}, {"api_name": "skimage.measure", "line_number": 5499, "usage_type": "name"}, {"api_name": "numpy.sqrt", "line_number": 5504, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 5523, "usage_type": "call"}, {"api_name": "numpy.zeros", "line_number": 5539, "usage_type": "call"}, {"api_name": "numpy.sort", "line_number": 5560, "usage_type": "call"}, {"api_name": "numpy.where", "line_number": 5574, "usage_type": "call"}, {"api_name": "numpy.mean", "line_number": 5587, "usage_type": "call"}, {"api_name": "numpy.median", "line_number": 5593, "usage_type": "call"}, {"api_name": "numpy.abs", "line_number": 5593, "usage_type": "call"}, {"api_name": "sys.version_info", "line_number": 5601, "usage_type": "attribute"}, {"api_name": "sys.exit", "line_number": 5602, "usage_type": "call"}, {"api_name": "sys.argv", "line_number": 5608, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 5616, "usage_type": "attribute"}, {"api_name": "sys.platform", "line_number": 5618, "usage_type": "attribute"}, {"api_name": "sys.excepthook", "line_number": 5626, "usage_type": "attribute"}, {"api_name": "traceback.print_tb", "line_number": 5645, "usage_type": "call"}, {"api_name": "sys.excepthook", "line_number": 5652, "usage_type": "attribute"}]}